mirror of
https://gitee.com/milvus-io/milvus.git
synced 2024-12-03 04:19:18 +08:00
5737e0075f
Signed-off-by: Congqi Xia <congqi.xia@zilliz.com>
339 lines
9.5 KiB
Go
339 lines
9.5 KiB
Go
// Licensed to the LF AI & Data foundation under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package datanode
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
|
|
miniokv "github.com/milvus-io/milvus/internal/kv/minio"
|
|
"github.com/milvus-io/milvus/internal/log"
|
|
"github.com/milvus-io/milvus/internal/msgstream"
|
|
"github.com/milvus-io/milvus/internal/proto/commonpb"
|
|
"github.com/milvus-io/milvus/internal/proto/datapb"
|
|
"github.com/milvus-io/milvus/internal/types"
|
|
"github.com/milvus-io/milvus/internal/util/flowgraph"
|
|
|
|
"go.uber.org/zap"
|
|
)
|
|
|
|
// dataSyncService controls a flowgraph for a specific collection
|
|
type dataSyncService struct {
|
|
ctx context.Context
|
|
cancelFn context.CancelFunc
|
|
fg *flowgraph.TimeTickedFlowGraph
|
|
flushCh chan flushMsg
|
|
replica Replica
|
|
idAllocator allocatorInterface
|
|
msFactory msgstream.Factory
|
|
collectionID UniqueID
|
|
dataCoord types.DataCoord
|
|
clearSignal chan<- UniqueID
|
|
|
|
flushingSegCache *Cache
|
|
flushManager flushManager
|
|
}
|
|
|
|
func newDataSyncService(ctx context.Context,
|
|
flushCh chan flushMsg,
|
|
replica Replica,
|
|
alloc allocatorInterface,
|
|
factory msgstream.Factory,
|
|
vchan *datapb.VchannelInfo,
|
|
clearSignal chan<- UniqueID,
|
|
dataCoord types.DataCoord,
|
|
flushingSegCache *Cache,
|
|
|
|
) (*dataSyncService, error) {
|
|
|
|
if replica == nil {
|
|
return nil, errors.New("Nil input")
|
|
}
|
|
|
|
ctx1, cancel := context.WithCancel(ctx)
|
|
|
|
service := &dataSyncService{
|
|
ctx: ctx1,
|
|
cancelFn: cancel,
|
|
fg: nil,
|
|
flushCh: flushCh,
|
|
replica: replica,
|
|
idAllocator: alloc,
|
|
msFactory: factory,
|
|
collectionID: vchan.GetCollectionID(),
|
|
dataCoord: dataCoord,
|
|
clearSignal: clearSignal,
|
|
flushingSegCache: flushingSegCache,
|
|
}
|
|
|
|
if err := service.initNodes(vchan); err != nil {
|
|
return nil, err
|
|
}
|
|
return service, nil
|
|
}
|
|
|
|
type parallelConfig struct {
|
|
maxQueueLength int32
|
|
maxParallelism int32
|
|
}
|
|
|
|
type nodeConfig struct {
|
|
msFactory msgstream.Factory // msgStream factory
|
|
collectionID UniqueID
|
|
vChannelName string
|
|
replica Replica // Segment replica
|
|
allocator allocatorInterface
|
|
|
|
// defaults
|
|
parallelConfig
|
|
}
|
|
|
|
func newParallelConfig() parallelConfig {
|
|
return parallelConfig{Params.FlowGraphMaxQueueLength, Params.FlowGraphMaxParallelism}
|
|
}
|
|
|
|
// start starts the flowgraph in datasyncservice
|
|
func (dsService *dataSyncService) start() {
|
|
if dsService.fg != nil {
|
|
log.Debug("Data Sync Service starting flowgraph")
|
|
dsService.fg.Start()
|
|
} else {
|
|
log.Debug("Data Sync Service flowgraph nil")
|
|
}
|
|
}
|
|
|
|
func (dsService *dataSyncService) close() {
|
|
if dsService.fg != nil {
|
|
log.Debug("Data Sync Service closing flowgraph")
|
|
dsService.fg.Close()
|
|
}
|
|
|
|
dsService.cancelFn()
|
|
}
|
|
|
|
// initNodes inits a TimetickedFlowGraph
|
|
func (dsService *dataSyncService) initNodes(vchanInfo *datapb.VchannelInfo) error {
|
|
dsService.fg = flowgraph.NewTimeTickedFlowGraph(dsService.ctx)
|
|
|
|
m := map[string]interface{}{
|
|
"PulsarAddress": Params.PulsarAddress,
|
|
"ReceiveBufSize": 1024,
|
|
"PulsarBufSize": 1024,
|
|
}
|
|
|
|
err := dsService.msFactory.SetParams(m)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// MinIO
|
|
option := &miniokv.Option{
|
|
Address: Params.MinioAddress,
|
|
AccessKeyID: Params.MinioAccessKeyID,
|
|
SecretAccessKeyID: Params.MinioSecretAccessKey,
|
|
UseSSL: Params.MinioUseSSL,
|
|
CreateBucket: true,
|
|
BucketName: Params.MinioBucketName,
|
|
}
|
|
|
|
minIOKV, err := miniokv.NewMinIOKV(dsService.ctx, option)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
dsService.flushManager = NewRendezvousFlushManager(dsService.idAllocator, minIOKV, dsService.replica, func(pack *segmentFlushPack) error {
|
|
fieldInsert := []*datapb.FieldBinlog{}
|
|
fieldStats := []*datapb.FieldBinlog{}
|
|
checkPoints := []*datapb.CheckPoint{}
|
|
for k, v := range pack.insertLogs {
|
|
fieldInsert = append(fieldInsert, &datapb.FieldBinlog{FieldID: k, Binlogs: []string{v}})
|
|
}
|
|
for k, v := range pack.statsLogs {
|
|
fieldStats = append(fieldStats, &datapb.FieldBinlog{FieldID: k, Binlogs: []string{v}})
|
|
}
|
|
// only current segment checkpoint info,
|
|
updates, _ := dsService.replica.getSegmentStatisticsUpdates(pack.segmentID)
|
|
checkPoints = append(checkPoints, &datapb.CheckPoint{
|
|
SegmentID: pack.segmentID,
|
|
NumOfRows: updates.GetNumRows(),
|
|
Position: pack.pos,
|
|
})
|
|
|
|
log.Debug("SaveBinlogPath",
|
|
zap.Int64("SegmentID", pack.segmentID),
|
|
zap.Int64("CollectionID", dsService.collectionID),
|
|
zap.Int("Length of Field2BinlogPaths", len(fieldInsert)),
|
|
)
|
|
|
|
req := &datapb.SaveBinlogPathsRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: 0, //TODO msg type
|
|
MsgID: 0, //TODO msg id
|
|
Timestamp: 0, //TODO time stamp
|
|
SourceID: Params.NodeID,
|
|
},
|
|
SegmentID: pack.segmentID,
|
|
CollectionID: dsService.collectionID,
|
|
Field2BinlogPaths: fieldInsert,
|
|
//TODO WIP add statslog and deltalog
|
|
CheckPoints: checkPoints,
|
|
|
|
StartPositions: dsService.replica.listNewSegmentsStartPositions(),
|
|
Flushed: pack.flushed,
|
|
}
|
|
rsp, err := dsService.dataCoord.SaveBinlogPaths(dsService.ctx, req)
|
|
if err != nil {
|
|
return fmt.Errorf(err.Error())
|
|
}
|
|
if rsp.ErrorCode != commonpb.ErrorCode_Success {
|
|
return fmt.Errorf("data service save bin log path failed, reason = %s", rsp.Reason)
|
|
}
|
|
return nil
|
|
|
|
})
|
|
|
|
c := &nodeConfig{
|
|
msFactory: dsService.msFactory,
|
|
collectionID: vchanInfo.GetCollectionID(),
|
|
vChannelName: vchanInfo.GetChannelName(),
|
|
replica: dsService.replica,
|
|
allocator: dsService.idAllocator,
|
|
|
|
parallelConfig: newParallelConfig(),
|
|
}
|
|
|
|
var dmStreamNode Node
|
|
dmStreamNode, err = newDmInputNode(dsService.ctx, vchanInfo.GetSeekPosition(), c)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
var ddNode Node = newDDNode(dsService.clearSignal, dsService.collectionID, vchanInfo)
|
|
var insertBufferNode Node
|
|
insertBufferNode, err = newInsertBufferNode(
|
|
dsService.ctx,
|
|
dsService.flushCh,
|
|
dsService.flushManager,
|
|
dsService.flushingSegCache,
|
|
c,
|
|
)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
var deleteNode Node
|
|
deleteNode, err = newDeleteNode(dsService.ctx, dsService.flushManager, c)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// recover segment checkpoints
|
|
for _, us := range vchanInfo.GetUnflushedSegments() {
|
|
if us.CollectionID != dsService.collectionID ||
|
|
us.GetInsertChannel() != vchanInfo.ChannelName {
|
|
log.Warn("Collection ID or ChannelName not compact",
|
|
zap.Int64("Wanted ID", dsService.collectionID),
|
|
zap.Int64("Actual ID", us.CollectionID),
|
|
zap.String("Wanted Channel Name", vchanInfo.ChannelName),
|
|
zap.String("Actual Channel Name", us.GetInsertChannel()),
|
|
)
|
|
continue
|
|
}
|
|
|
|
log.Info("Recover Segment NumOfRows form checkpoints",
|
|
zap.String("InsertChannel", us.GetInsertChannel()),
|
|
zap.Int64("SegmentID", us.GetID()),
|
|
zap.Int64("NumOfRows", us.GetNumOfRows()),
|
|
)
|
|
|
|
if err := dsService.replica.addNormalSegment(us.GetID(), us.CollectionID, us.PartitionID, us.GetInsertChannel(),
|
|
us.GetNumOfRows(), &segmentCheckPoint{us.GetNumOfRows(), *us.GetDmlPosition()}); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
for _, fs := range vchanInfo.GetFlushedSegments() {
|
|
if fs.CollectionID != dsService.collectionID ||
|
|
fs.GetInsertChannel() != vchanInfo.ChannelName {
|
|
log.Warn("Collection ID or ChannelName not compact",
|
|
zap.Int64("Wanted ID", dsService.collectionID),
|
|
zap.Int64("Actual ID", fs.CollectionID),
|
|
zap.String("Wanted Channel Name", vchanInfo.ChannelName),
|
|
zap.String("Actual Channel Name", fs.GetInsertChannel()),
|
|
)
|
|
continue
|
|
}
|
|
|
|
log.Info("Recover Segment NumOfRows form checkpoints",
|
|
zap.String("InsertChannel", fs.GetInsertChannel()),
|
|
zap.Int64("SegmentID", fs.GetID()),
|
|
zap.Int64("NumOfRows", fs.GetNumOfRows()),
|
|
)
|
|
if err := dsService.replica.addFlushedSegment(fs.GetID(), fs.CollectionID,
|
|
fs.PartitionID, fs.GetInsertChannel(), fs.GetNumOfRows()); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
dsService.fg.AddNode(dmStreamNode)
|
|
dsService.fg.AddNode(ddNode)
|
|
dsService.fg.AddNode(insertBufferNode)
|
|
dsService.fg.AddNode(deleteNode)
|
|
|
|
// ddStreamNode
|
|
err = dsService.fg.SetEdges(dmStreamNode.Name(),
|
|
[]string{},
|
|
[]string{ddNode.Name()},
|
|
)
|
|
if err != nil {
|
|
log.Error("set edges failed in node", zap.String("name", dmStreamNode.Name()), zap.Error(err))
|
|
return err
|
|
}
|
|
|
|
// ddNode
|
|
err = dsService.fg.SetEdges(ddNode.Name(),
|
|
[]string{dmStreamNode.Name()},
|
|
[]string{insertBufferNode.Name()},
|
|
)
|
|
if err != nil {
|
|
log.Error("set edges failed in node", zap.String("name", ddNode.Name()), zap.Error(err))
|
|
return err
|
|
}
|
|
|
|
// insertBufferNode
|
|
err = dsService.fg.SetEdges(insertBufferNode.Name(),
|
|
[]string{ddNode.Name()},
|
|
[]string{deleteNode.Name()},
|
|
)
|
|
if err != nil {
|
|
log.Error("set edges failed in node", zap.String("name", insertBufferNode.Name()), zap.Error(err))
|
|
return err
|
|
}
|
|
|
|
//deleteNode
|
|
err = dsService.fg.SetEdges(deleteNode.Name(),
|
|
[]string{insertBufferNode.Name()},
|
|
[]string{},
|
|
)
|
|
if err != nil {
|
|
log.Error("set edges failed in node", zap.String("name", deleteNode.Name()), zap.Error(err))
|
|
return err
|
|
}
|
|
return nil
|
|
}
|