// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use this file except in compliance // with the License. You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. // Package datanode implements data persistence logic. // // Data node persists insert logs into persistent storage like minIO/S3. package datanode import ( "context" "fmt" "io" "math/rand" "os" "sync" "sync/atomic" "time" "github.com/cockroachdb/errors" "github.com/tidwall/gjson" clientv3 "go.etcd.io/etcd/client/v3" "go.uber.org/zap" "github.com/milvus-io/milvus-proto/go-api/v2/commonpb" "github.com/milvus-io/milvus-proto/go-api/v2/milvuspb" "github.com/milvus-io/milvus/internal/datanode/allocator" "github.com/milvus-io/milvus/internal/datanode/channel" "github.com/milvus-io/milvus/internal/datanode/compaction" "github.com/milvus-io/milvus/internal/datanode/importv2" "github.com/milvus-io/milvus/internal/datanode/util" "github.com/milvus-io/milvus/internal/flushcommon/broker" "github.com/milvus-io/milvus/internal/flushcommon/pipeline" "github.com/milvus-io/milvus/internal/flushcommon/syncmgr" util2 "github.com/milvus-io/milvus/internal/flushcommon/util" "github.com/milvus-io/milvus/internal/flushcommon/writebuffer" etcdkv "github.com/milvus-io/milvus/internal/kv/etcd" "github.com/milvus-io/milvus/internal/storage" "github.com/milvus-io/milvus/internal/types" "github.com/milvus-io/milvus/internal/util/dependency" "github.com/milvus-io/milvus/internal/util/sessionutil" "github.com/milvus-io/milvus/internal/util/streamingutil" "github.com/milvus-io/milvus/pkg/kv" "github.com/milvus-io/milvus/pkg/log" "github.com/milvus-io/milvus/pkg/metrics" "github.com/milvus-io/milvus/pkg/mq/msgdispatcher" "github.com/milvus-io/milvus/pkg/util/conc" "github.com/milvus-io/milvus/pkg/util/expr" "github.com/milvus-io/milvus/pkg/util/logutil" "github.com/milvus-io/milvus/pkg/util/metricsinfo" "github.com/milvus-io/milvus/pkg/util/paramtable" "github.com/milvus-io/milvus/pkg/util/retry" "github.com/milvus-io/milvus/pkg/util/typeutil" ) const ( // ConnectEtcdMaxRetryTime is used to limit the max retry time for connection etcd ConnectEtcdMaxRetryTime = 100 ) // makes sure DataNode implements types.DataNode var _ types.DataNode = (*DataNode)(nil) // Params from config.yaml var Params *paramtable.ComponentParam = paramtable.Get() // DataNode communicates with outside services and unioun all // services in datanode package. // // DataNode implements `types.Component`, `types.DataNode` interfaces. // // `etcdCli` is a connection of etcd // `rootCoord` is a grpc client of root coordinator. // `dataCoord` is a grpc client of data service. // `stateCode` is current statement of this data node, indicating whether it's healthy. // // `clearSignal` is a signal channel for releasing the flowgraph resources. // `segmentCache` stores all flushing and flushed segments. type DataNode struct { ctx context.Context cancel context.CancelFunc Role string stateCode atomic.Value // commonpb.StateCode_Initializing flowgraphManager pipeline.FlowgraphManager channelManager channel.ChannelManager syncMgr syncmgr.SyncManager writeBufferManager writebuffer.BufferManager importTaskMgr importv2.TaskManager importScheduler importv2.Scheduler segmentCache *util.Cache compactionExecutor compaction.Executor timeTickSender *util2.TimeTickSender channelCheckpointUpdater *util2.ChannelCheckpointUpdater etcdCli *clientv3.Client address string rootCoord types.RootCoordClient dataCoord types.DataCoordClient broker broker.Broker // call once initOnce sync.Once startOnce sync.Once stopOnce sync.Once sessionMu sync.Mutex // to fix data race session *sessionutil.Session watchKv kv.WatchKV chunkManager storage.ChunkManager allocator allocator.Allocator closer io.Closer dispClient msgdispatcher.Client factory dependency.Factory reportImportRetryTimes uint // unitest set this value to 1 to save time, default is 10 pool *conc.Pool[any] metricsRequest *metricsinfo.MetricsRequest } // NewDataNode will return a DataNode with abnormal state. func NewDataNode(ctx context.Context, factory dependency.Factory) *DataNode { rand.Seed(time.Now().UnixNano()) ctx2, cancel2 := context.WithCancel(ctx) node := &DataNode{ ctx: ctx2, cancel: cancel2, Role: typeutil.DataNodeRole, rootCoord: nil, dataCoord: nil, factory: factory, segmentCache: util.NewCache(), compactionExecutor: compaction.NewExecutor(), reportImportRetryTimes: 10, metricsRequest: metricsinfo.NewMetricsRequest(), } node.UpdateStateCode(commonpb.StateCode_Abnormal) expr.Register("datanode", node) return node } func (node *DataNode) SetAddress(address string) { node.address = address } func (node *DataNode) GetAddress() string { return node.address } // SetEtcdClient sets etcd client for DataNode func (node *DataNode) SetEtcdClient(etcdCli *clientv3.Client) { node.etcdCli = etcdCli } // SetRootCoordClient sets RootCoord's grpc client, error is returned if repeatedly set. func (node *DataNode) SetRootCoordClient(rc types.RootCoordClient) error { switch { case rc == nil, node.rootCoord != nil: return errors.New("nil parameter or repeatedly set") default: node.rootCoord = rc return nil } } // SetDataCoordClient sets data service's grpc client, error is returned if repeatedly set. func (node *DataNode) SetDataCoordClient(ds types.DataCoordClient) error { switch { case ds == nil, node.dataCoord != nil: return errors.New("nil parameter or repeatedly set") default: node.dataCoord = ds return nil } } // Register register datanode to etcd func (node *DataNode) Register() error { log.Debug("node begin to register to etcd", zap.String("serverName", node.session.ServerName), zap.Int64("ServerID", node.session.ServerID)) node.session.Register() metrics.NumNodes.WithLabelValues(fmt.Sprint(node.GetNodeID()), typeutil.DataNodeRole).Inc() log.Info("DataNode Register Finished") // Start liveness check node.session.LivenessCheck(node.ctx, func() { log.Error("Data Node disconnected from etcd, process will exit", zap.Int64("Server Id", node.GetSession().ServerID)) os.Exit(1) }) return nil } func (node *DataNode) initSession() error { node.session = sessionutil.NewSession(node.ctx) if node.session == nil { return errors.New("failed to initialize session") } node.session.Init(typeutil.DataNodeRole, node.address, false, true) sessionutil.SaveServerInfo(typeutil.DataNodeRole, node.session.ServerID) return nil } func (node *DataNode) GetNodeID() int64 { if node.session != nil { return node.session.ServerID } return paramtable.GetNodeID() } func (node *DataNode) Init() error { var initError error node.initOnce.Do(func() { node.registerMetricsRequest() logutil.Logger(node.ctx).Info("DataNode server initializing", zap.String("TimeTickChannelName", Params.CommonCfg.DataCoordTimeTick.GetValue()), ) if err := node.initSession(); err != nil { log.Error("DataNode server init session failed", zap.Error(err)) initError = err return } serverID := node.GetNodeID() log := log.Ctx(node.ctx).With(zap.String("role", typeutil.DataNodeRole), zap.Int64("nodeID", serverID)) node.broker = broker.NewCoordBroker(node.dataCoord, serverID) node.dispClient = msgdispatcher.NewClient(node.factory, typeutil.DataNodeRole, serverID) log.Info("DataNode server init dispatcher client done") alloc, err := allocator.New(context.Background(), node.rootCoord, serverID) if err != nil { log.Error("failed to create id allocator", zap.Error(err)) initError = err return } node.allocator = alloc node.factory.Init(Params) log.Info("DataNode server init succeeded", zap.String("MsgChannelSubName", Params.CommonCfg.DataNodeSubName.GetValue())) chunkManager, err := node.factory.NewPersistentStorageChunkManager(node.ctx) if err != nil { initError = err return } node.chunkManager = chunkManager syncMgr := syncmgr.NewSyncManager(node.chunkManager) node.syncMgr = syncMgr node.writeBufferManager = writebuffer.NewManager(syncMgr) node.importTaskMgr = importv2.NewTaskManager() node.importScheduler = importv2.NewScheduler(node.importTaskMgr) node.channelCheckpointUpdater = util2.NewChannelCheckpointUpdater(node.broker) node.flowgraphManager = pipeline.NewFlowgraphManager() log.Info("init datanode done", zap.String("Address", node.address)) }) return initError } func (node *DataNode) registerMetricsRequest() { node.metricsRequest.RegisterMetricsRequest(metricsinfo.SystemInfoMetrics, func(ctx context.Context, req *milvuspb.GetMetricsRequest, jsonReq gjson.Result) (string, error) { return node.getSystemInfoMetrics(ctx, req) }) node.metricsRequest.RegisterMetricsRequest(metricsinfo.SyncTaskKey, func(ctx context.Context, req *milvuspb.GetMetricsRequest, jsonReq gjson.Result) (string, error) { return node.syncMgr.TaskStatsJSON(), nil }) node.metricsRequest.RegisterMetricsRequest(metricsinfo.SegmentKey, func(ctx context.Context, req *milvuspb.GetMetricsRequest, jsonReq gjson.Result) (string, error) { return node.flowgraphManager.GetSegmentsJSON(), nil }) node.metricsRequest.RegisterMetricsRequest(metricsinfo.ChannelKey, func(ctx context.Context, req *milvuspb.GetMetricsRequest, jsonReq gjson.Result) (string, error) { return node.flowgraphManager.GetChannelsJSON(), nil }) log.Info("register metrics actions finished") } // tryToReleaseFlowgraph tries to release a flowgraph func (node *DataNode) tryToReleaseFlowgraph(channel string) { log.Info("try to release flowgraph", zap.String("channel", channel)) if node.compactionExecutor != nil { node.compactionExecutor.DiscardPlan(channel) } if node.flowgraphManager != nil { node.flowgraphManager.RemoveFlowgraph(channel) } if node.writeBufferManager != nil { node.writeBufferManager.RemoveChannel(channel) } } // Start will update DataNode state to HEALTHY func (node *DataNode) Start() error { var startErr error node.startOnce.Do(func() { if err := node.allocator.Start(); err != nil { log.Error("failed to start id allocator", zap.Error(err), zap.String("role", typeutil.DataNodeRole)) startErr = err return } log.Info("start id allocator done", zap.String("role", typeutil.DataNodeRole)) connectEtcdFn := func() error { etcdKV := etcdkv.NewEtcdKV(node.etcdCli, Params.EtcdCfg.MetaRootPath.GetValue(), etcdkv.WithRequestTimeout(paramtable.Get().ServiceParam.EtcdCfg.RequestTimeout.GetAsDuration(time.Millisecond))) node.watchKv = etcdKV return nil } err := retry.Do(node.ctx, connectEtcdFn, retry.Attempts(ConnectEtcdMaxRetryTime)) if err != nil { startErr = errors.New("DataNode fail to connect etcd") return } if !streamingutil.IsStreamingServiceEnabled() { node.writeBufferManager.Start() node.timeTickSender = util2.NewTimeTickSender(node.broker, node.session.ServerID, retry.Attempts(20), retry.Sleep(time.Millisecond*100)) node.timeTickSender.Start() node.channelManager = channel.NewChannelManager(getPipelineParams(node), node.flowgraphManager) node.channelManager.Start() go node.channelCheckpointUpdater.Start() } go node.compactionExecutor.Start(node.ctx) go node.importScheduler.Start() node.UpdateStateCode(commonpb.StateCode_Healthy) }) return startErr } // UpdateStateCode updates datanode's state code func (node *DataNode) UpdateStateCode(code commonpb.StateCode) { node.stateCode.Store(code) } // GetStateCode return datanode's state code func (node *DataNode) GetStateCode() commonpb.StateCode { return node.stateCode.Load().(commonpb.StateCode) } func (node *DataNode) isHealthy() bool { return node.GetStateCode() == commonpb.StateCode_Healthy } // ReadyToFlush tells whether DataNode is ready for flushing func (node *DataNode) ReadyToFlush() error { if !node.isHealthy() { return errors.New("DataNode not in HEALTHY state") } return nil } // Stop will release DataNode resources and shutdown datanode func (node *DataNode) Stop() error { node.stopOnce.Do(func() { // https://github.com/milvus-io/milvus/issues/12282 node.UpdateStateCode(commonpb.StateCode_Abnormal) if node.channelManager != nil { node.channelManager.Close() } if node.flowgraphManager != nil { node.flowgraphManager.ClearFlowgraphs() node.flowgraphManager.Close() } if node.writeBufferManager != nil { node.writeBufferManager.Stop() } if node.allocator != nil { log.Info("close id allocator", zap.String("role", typeutil.DataNodeRole)) node.allocator.Close() } if node.closer != nil { node.closer.Close() } if node.session != nil { node.session.Stop() } if node.timeTickSender != nil { node.timeTickSender.Stop() } if node.channelCheckpointUpdater != nil { node.channelCheckpointUpdater.Close() } if node.importScheduler != nil { node.importScheduler.Close() } // Delay the cancellation of ctx to ensure that the session is automatically recycled after closed the flow graph node.cancel() }) return nil } // SetSession to fix data race func (node *DataNode) SetSession(session *sessionutil.Session) { node.sessionMu.Lock() defer node.sessionMu.Unlock() node.session = session } // GetSession to fix data race func (node *DataNode) GetSession() *sessionutil.Session { node.sessionMu.Lock() defer node.sessionMu.Unlock() return node.session } func getPipelineParams(node *DataNode) *util2.PipelineParams { return &util2.PipelineParams{ Ctx: node.ctx, Broker: node.broker, SyncMgr: node.syncMgr, TimeTickSender: node.timeTickSender, CompactionExecutor: node.compactionExecutor, MsgStreamFactory: node.factory, DispClient: node.dispClient, ChunkManager: node.chunkManager, Session: node.session, WriteBufferManager: node.writeBufferManager, CheckpointUpdater: node.channelCheckpointUpdater, Allocator: node.allocator, } }