milvus/internal/flushcommon/io/io_pool.go
yihao.dai a4439cc911
enhance: Implement flusher in streamingNode (#34942)
- Implement flusher to:
  - Manage the pipelines (creation, deletion, etc.)
  - Manage the segment write buffer
  - Manage sync operation (including receive flushMsg and execute flush)
- Add a new `GetChannelRecoveryInfo` RPC in DataCoord.
- Reorganize packages: `flushcommon` and `datanode`.

issue: https://github.com/milvus-io/milvus/issues/33285

---------

Signed-off-by: bigsheeper <yihao.dai@zilliz.com>
2024-08-02 18:30:23 +08:00

119 lines
2.8 KiB
Go

package io
import (
"context"
"sync"
"sync/atomic"
"go.uber.org/zap"
"github.com/milvus-io/milvus/pkg/config"
"github.com/milvus-io/milvus/pkg/log"
"github.com/milvus-io/milvus/pkg/util/conc"
"github.com/milvus-io/milvus/pkg/util/hardware"
"github.com/milvus-io/milvus/pkg/util/paramtable"
)
var (
ioPool *conc.Pool[any]
ioPoolInitOnce sync.Once
)
var (
statsPool *conc.Pool[any]
statsPoolInitOnce sync.Once
)
var (
bfApplyPool atomic.Pointer[conc.Pool[any]]
bfApplyPoolInitOnce sync.Once
)
func initIOPool() {
capacity := paramtable.Get().DataNodeCfg.IOConcurrency.GetAsInt()
if capacity > 32 {
capacity = 32
}
// error only happens with negative expiry duration or with negative pre-alloc size.
ioPool = conc.NewPool[any](capacity)
}
func GetOrCreateIOPool() *conc.Pool[any] {
ioPoolInitOnce.Do(initIOPool)
return ioPool
}
func initStatsPool() {
poolSize := paramtable.Get().DataNodeCfg.ChannelWorkPoolSize.GetAsInt()
if poolSize <= 0 {
poolSize = hardware.GetCPUNum()
}
statsPool = conc.NewPool[any](poolSize, conc.WithPreAlloc(false), conc.WithNonBlocking(false))
}
func GetOrCreateStatsPool() *conc.Pool[any] {
statsPoolInitOnce.Do(initStatsPool)
return statsPool
}
func initMultiReadPool() {
capacity := paramtable.Get().DataNodeCfg.FileReadConcurrency.GetAsInt()
if capacity > hardware.GetCPUNum() {
capacity = hardware.GetCPUNum()
}
// error only happens with negative expiry duration or with negative pre-alloc size.
ioPool = conc.NewPool[any](capacity)
}
func getMultiReadPool() *conc.Pool[any] {
ioPoolInitOnce.Do(initMultiReadPool)
return ioPool
}
func resizePool(pool *conc.Pool[any], newSize int, tag string) {
log := log.Ctx(context.Background()).
With(
zap.String("poolTag", tag),
zap.Int("newSize", newSize),
)
if newSize <= 0 {
log.Warn("cannot set pool size to non-positive value")
return
}
err := pool.Resize(newSize)
if err != nil {
log.Warn("failed to resize pool", zap.Error(err))
return
}
log.Info("pool resize successfully")
}
func ResizeBFApplyPool(evt *config.Event) {
if evt.HasUpdated {
pt := paramtable.Get()
newSize := hardware.GetCPUNum() * pt.QueryNodeCfg.BloomFilterApplyParallelFactor.GetAsInt()
resizePool(GetBFApplyPool(), newSize, "BFApplyPool")
}
}
func initBFApplyPool() {
bfApplyPoolInitOnce.Do(func() {
pt := paramtable.Get()
poolSize := hardware.GetCPUNum() * pt.QueryNodeCfg.BloomFilterApplyParallelFactor.GetAsInt()
log.Info("init BFApplyPool", zap.Int("poolSize", poolSize))
pool := conc.NewPool[any](
poolSize,
)
bfApplyPool.Store(pool)
pt.Watch(pt.QueryNodeCfg.BloomFilterApplyParallelFactor.Key, config.NewHandler("dn.bfapply.parallel", ResizeBFApplyPool))
})
}
func GetBFApplyPool() *conc.Pool[any] {
initBFApplyPool()
return bfApplyPool.Load()
}