// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use this file except in compliance // with the License. You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package grpcrootcoord import ( "context" "encoding/json" "fmt" "math/rand" "path" "sync" "testing" "time" clientv3 "go.etcd.io/etcd/client/v3" "github.com/golang/protobuf/proto" "github.com/stretchr/testify/assert" rcc "github.com/milvus-io/milvus/internal/distributed/rootcoord/client" "github.com/milvus-io/milvus/internal/proto/commonpb" "github.com/milvus-io/milvus/internal/proto/datapb" "github.com/milvus-io/milvus/internal/proto/etcdpb" "github.com/milvus-io/milvus/internal/proto/internalpb" "github.com/milvus-io/milvus/internal/proto/milvuspb" "github.com/milvus-io/milvus/internal/proto/proxypb" "github.com/milvus-io/milvus/internal/proto/rootcoordpb" "github.com/milvus-io/milvus/internal/proto/schemapb" "github.com/milvus-io/milvus/internal/rootcoord" "github.com/milvus-io/milvus/internal/types" "github.com/milvus-io/milvus/internal/util/dependency" "github.com/milvus-io/milvus/internal/util/etcd" "github.com/milvus-io/milvus/internal/util/retry" "github.com/milvus-io/milvus/internal/util/sessionutil" "github.com/milvus-io/milvus/internal/util/typeutil" ) type proxyMock struct { types.Proxy invalidateCollectionMetaCache func(ctx context.Context, request *proxypb.InvalidateCollMetaCacheRequest) (*commonpb.Status, error) } func (p *proxyMock) InvalidateCollectionMetaCache(ctx context.Context, request *proxypb.InvalidateCollMetaCacheRequest) (*commonpb.Status, error) { return p.invalidateCollectionMetaCache(ctx, request) } func TestGrpcService(t *testing.T) { const ( dbName = "testDB" collName = "testColl" collName2 = "testColl-again" partName = "testPartition" fieldName = "vector" fieldID = 100 segID = 1001 ) rand.Seed(time.Now().UnixNano()) randVal := rand.Int() Params.InitOnce(typeutil.RootCoordRole) Params.Port = (randVal % 100) + 10000 t.Log("newParams.Address:", Params.GetAddress()) ctx := context.Background() factory := dependency.NewDefaultFactory(true) svr, err := NewServer(ctx, factory) assert.Nil(t, err) rootcoord.Params.Init() rootcoord.Params.EtcdCfg.MetaRootPath = fmt.Sprintf("/%d/test/meta", randVal) rootcoord.Params.EtcdCfg.KvRootPath = fmt.Sprintf("/%d/test/kv", randVal) rootcoord.Params.CommonCfg.RootCoordSubName = fmt.Sprintf("msgChannel%d", randVal) rootcoord.Params.CommonCfg.RootCoordTimeTick = fmt.Sprintf("timeTick%d", randVal) rootcoord.Params.CommonCfg.RootCoordStatistics = fmt.Sprintf("stateChannel%d", randVal) rootcoord.Params.RootCoordCfg.MaxPartitionNum = 64 rootcoord.Params.CommonCfg.DefaultPartitionName = "_default" rootcoord.Params.CommonCfg.DefaultIndexName = "_default" t.Logf("service port = %d", Params.Port) core, ok := (svr.rootCoord).(*rootcoord.Core) assert.True(t, ok) err = svr.startGrpc(Params.Port) assert.Nil(t, err) svr.rootCoord.UpdateStateCode(internalpb.StateCode_Initializing) etcdCli, err := etcd.GetEtcdClient(&Params.EtcdCfg) assert.Nil(t, err) sessKey := path.Join(rootcoord.Params.EtcdCfg.MetaRootPath, sessionutil.DefaultServiceRoot) _, err = etcdCli.Delete(ctx, sessKey, clientv3.WithPrefix()) assert.Nil(t, err) pnb, err := json.Marshal( &sessionutil.Session{ ServerID: 100, }, ) assert.Nil(t, err) _, err = etcdCli.Put(ctx, path.Join(sessKey, typeutil.ProxyRole+"-100"), string(pnb)) assert.Nil(t, err) rootcoord.Params.RootCoordCfg.Address = Params.GetAddress() core.SetEtcdClient(etcdCli) err = core.Init() assert.Nil(t, err) timeTickArray := make([]typeutil.Timestamp, 0, 16) timeTickLock := sync.Mutex{} core.SendTimeTick = func(ts typeutil.Timestamp, reason string) error { timeTickLock.Lock() defer timeTickLock.Unlock() t.Logf("send time tick %d", ts) timeTickArray = append(timeTickArray, ts) return nil } core.SendDdCreateCollectionReq = func(ctx context.Context, req *internalpb.CreateCollectionRequest, channelNames []string) (map[string][]byte, error) { return map[string][]byte{}, nil } dropCollectionArray := make([]*internalpb.DropCollectionRequest, 0, 16) core.SendDdDropCollectionReq = func(ctx context.Context, req *internalpb.DropCollectionRequest, channelNames []string) error { t.Logf("Drop Collection %s", req.CollectionName) dropCollectionArray = append(dropCollectionArray, req) return nil } createPartitionArray := make([]*internalpb.CreatePartitionRequest, 0, 16) core.SendDdCreatePartitionReq = func(ctx context.Context, req *internalpb.CreatePartitionRequest, channelNames []string) error { t.Logf("Create Partition %s", req.PartitionName) createPartitionArray = append(createPartitionArray, req) return nil } dropPartitionArray := make([]*internalpb.DropPartitionRequest, 0, 16) core.SendDdDropPartitionReq = func(ctx context.Context, req *internalpb.DropPartitionRequest, channelNames []string) error { t.Logf("Drop Partition %s", req.PartitionName) dropPartitionArray = append(dropPartitionArray, req) return nil } core.CallGetBinlogFilePathsService = func(ctx context.Context, segID typeutil.UniqueID, fieldID typeutil.UniqueID) ([]string, error) { return []string{"file1", "file2", "file3"}, nil } core.CallGetNumRowsService = func(ctx context.Context, segID typeutil.UniqueID, isFromFlushedChan bool) (int64, error) { return rootcoord.Params.RootCoordCfg.MinSegmentSizeToEnableIndex, nil } core.CallWatchChannels = func(ctx context.Context, collectionID int64, channelNames []string) error { return nil } var segs []typeutil.UniqueID segLock := sync.Mutex{} core.CallGetFlushedSegmentsService = func(ctx context.Context, collID, partID typeutil.UniqueID) ([]typeutil.UniqueID, error) { segLock.Lock() defer segLock.Unlock() var ret []typeutil.UniqueID ret = append(ret, segs...) return ret, nil } var binlogLock sync.Mutex binlogPathArray := make([]string, 0, 16) core.CallBuildIndexService = func(ctx context.Context, binlog []string, field *schemapb.FieldSchema, idxInfo *etcdpb.IndexInfo, numRows int64) (typeutil.UniqueID, error) { binlogLock.Lock() defer binlogLock.Unlock() binlogPathArray = append(binlogPathArray, binlog...) return 2000, nil } var dropIDLock sync.Mutex dropID := make([]typeutil.UniqueID, 0, 16) core.CallDropIndexService = func(ctx context.Context, indexID typeutil.UniqueID) error { dropIDLock.Lock() defer dropIDLock.Unlock() dropID = append(dropID, indexID) return nil } collectionMetaCache := make([]string, 0, 16) pnm := proxyMock{} core.NewProxyClient = func(*sessionutil.Session) (types.Proxy, error) { return &pnm, nil } pnm.invalidateCollectionMetaCache = func(ctx context.Context, request *proxypb.InvalidateCollMetaCacheRequest) (*commonpb.Status, error) { collectionMetaCache = append(collectionMetaCache, request.CollectionName) return &commonpb.Status{ ErrorCode: commonpb.ErrorCode_Success, }, nil } core.CallReleaseCollectionService = func(ctx context.Context, ts typeutil.Timestamp, dbID typeutil.UniqueID, collectionID typeutil.UniqueID) error { return nil } core.CallReleasePartitionService = func(ctx context.Context, ts typeutil.Timestamp, dbID, collectionID typeutil.UniqueID, partitionIDs []typeutil.UniqueID) error { return nil } core.CallImportService = func(ctx context.Context, req *datapb.ImportTaskRequest) *datapb.ImportTaskResponse { return nil } err = svr.start() assert.Nil(t, err) svr.rootCoord.UpdateStateCode(internalpb.StateCode_Healthy) cli, err := rcc.NewClient(context.Background(), rootcoord.Params.EtcdCfg.MetaRootPath, etcdCli) assert.Nil(t, err) err = cli.Init() assert.Nil(t, err) err = cli.Start() assert.Nil(t, err) t.Run("get component states", func(t *testing.T) { req := &internalpb.GetComponentStatesRequest{} rsp, err := svr.GetComponentStates(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) }) t.Run("get time tick channel", func(t *testing.T) { req := &internalpb.GetTimeTickChannelRequest{} rsp, err := svr.GetTimeTickChannel(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) }) t.Run("get statistics channel", func(t *testing.T) { req := &internalpb.GetStatisticsChannelRequest{} rsp, err := svr.GetStatisticsChannel(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) }) t.Run("alloc time stamp", func(t *testing.T) { req := &rootcoordpb.AllocTimestampRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_RequestTSO, }, Count: 1, } rsp, err := svr.AllocTimestamp(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) }) t.Run("alloc id", func(t *testing.T) { req := &rootcoordpb.AllocIDRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_RequestID, }, Count: 1, } rsp, err := svr.AllocID(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) }) t.Run("update channel timetick", func(t *testing.T) { req := &internalpb.ChannelTimeTickMsg{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_TimeTick, }, } status, err := svr.UpdateChannelTimeTick(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, status.ErrorCode) }) t.Run("release DQL msg stream", func(t *testing.T) { req := &proxypb.ReleaseDQLMessageStreamRequest{} assert.Panics(t, func() { svr.ReleaseDQLMessageStream(ctx, req) }) }) t.Run("get metrics", func(t *testing.T) { req := &milvuspb.GetMetricsRequest{} rsp, err := svr.GetMetrics(ctx, req) assert.Nil(t, err) assert.NotEqual(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) }) t.Run("create collection", func(t *testing.T) { schema := schemapb.CollectionSchema{ Name: collName, AutoID: true, Fields: []*schemapb.FieldSchema{ { FieldID: fieldID, Name: fieldName, IsPrimaryKey: false, DataType: schemapb.DataType_FloatVector, TypeParams: nil, IndexParams: []*commonpb.KeyValuePair{ { Key: "ik1", Value: "iv1", }, }, }, }, } sbf, err := proto.Marshal(&schema) assert.Nil(t, err) req := &milvuspb.CreateCollectionRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_CreateCollection, MsgID: 100, Timestamp: 100, SourceID: 100, }, DbName: dbName, CollectionName: collName, Schema: sbf, } status, err := cli.CreateCollection(ctx, req) assert.Nil(t, err) colls, err := core.MetaTable.ListCollections(0) assert.Nil(t, err) assert.Equal(t, 1, len(colls)) assert.Equal(t, commonpb.ErrorCode_Success, status.ErrorCode) //assert.Equal(t, commonpb.MsgType_CreateCollection, createCollectionArray[0].Base.MsgType) _, has := colls[collName] assert.True(t, has) req.Base.MsgID = 101 req.Base.Timestamp = 101 req.Base.SourceID = 101 status, err = cli.CreateCollection(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_UnexpectedError, status.ErrorCode) req.Base.MsgID = 102 req.Base.Timestamp = 102 req.Base.SourceID = 102 req.CollectionName = collName2 status, err = cli.CreateCollection(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_UnexpectedError, status.ErrorCode) schema.Name = req.CollectionName sbf, err = proto.Marshal(&schema) assert.Nil(t, err) req.Schema = sbf req.Base.MsgID = 103 req.Base.Timestamp = 103 req.Base.SourceID = 103 status, err = cli.CreateCollection(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, status.ErrorCode) colls, err = core.MetaTable.ListCollections(0) assert.Nil(t, err) assert.Equal(t, 2, len(colls)) _, has = colls[collName2] assert.True(t, has) //time stamp go back, master response to add the timestamp, so the time tick will never go back //schema.Name = "testColl-goback" //sbf, err = proto.Marshal(&schema) //assert.Nil(t, err) //req.CollectionName = schema.Name //req.Schema = sbf //req.Base.MsgID = 103 //req.Base.Timestamp = 103 //req.Base.SourceID = 103 //status, err = cli.CreateCollection(ctx, req) //assert.Nil(t, err) //assert.Equal(t, status.ErrorCode, commonpb.ErrorCode_UnexpectedError) //matched, err := regexp.MatchString("input timestamp = [0-9]+, last dd time stamp = [0-9]+", status.Reason) //assert.Nil(t, err) //assert.True(t, matched) }) t.Run("has collection", func(t *testing.T) { req := &milvuspb.HasCollectionRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_HasCollection, MsgID: 110, Timestamp: 110, SourceID: 110, }, DbName: "testDb", CollectionName: collName, } rsp, err := cli.HasCollection(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) assert.Equal(t, true, rsp.Value) req = &milvuspb.HasCollectionRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_HasCollection, MsgID: 111, Timestamp: 111, SourceID: 111, }, DbName: "testDb", CollectionName: "testColl2", } rsp, err = cli.HasCollection(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) assert.Equal(t, false, rsp.Value) // test time stamp go back req = &milvuspb.HasCollectionRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_HasCollection, MsgID: 111, Timestamp: 111, SourceID: 111, }, DbName: "testDb", CollectionName: "testColl2", } rsp, err = cli.HasCollection(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) assert.Equal(t, false, rsp.Value) }) t.Run("describe collection", func(t *testing.T) { collMeta, err := core.MetaTable.GetCollectionByName(collName, 0) assert.Nil(t, err) req := &milvuspb.DescribeCollectionRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_DescribeCollection, MsgID: 120, Timestamp: 120, SourceID: 120, }, DbName: "testDb", CollectionName: collName, } rsp, err := cli.DescribeCollection(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) assert.Equal(t, collName, rsp.Schema.Name) assert.Equal(t, collMeta.ID, rsp.CollectionID) }) t.Run("show collection", func(t *testing.T) { req := &milvuspb.ShowCollectionsRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_ShowCollections, MsgID: 130, Timestamp: 130, SourceID: 130, }, DbName: "testDb", } rsp, err := cli.ShowCollections(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) assert.ElementsMatch(t, rsp.CollectionNames, []string{collName, collName2}) assert.Equal(t, 2, len(rsp.CollectionNames)) }) t.Run("create partition", func(t *testing.T) { req := &milvuspb.CreatePartitionRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_CreatePartition, MsgID: 140, Timestamp: 140, SourceID: 140, }, DbName: dbName, CollectionName: collName, PartitionName: partName, } status, err := cli.CreatePartition(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, status.ErrorCode) collMeta, err := core.MetaTable.GetCollectionByName(collName, 0) assert.Nil(t, err) assert.Equal(t, 2, len(collMeta.PartitionIDs)) partName2, err := core.MetaTable.GetPartitionNameByID(collMeta.ID, collMeta.PartitionIDs[1], 0) assert.Nil(t, err) assert.Equal(t, partName, partName2) assert.Equal(t, 1, len(collectionMetaCache)) }) t.Run("has partition", func(t *testing.T) { req := &milvuspb.HasPartitionRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_HasPartition, MsgID: 150, Timestamp: 150, SourceID: 150, }, DbName: dbName, CollectionName: collName, PartitionName: partName, } rsp, err := cli.HasPartition(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) assert.Equal(t, true, rsp.Value) }) t.Run("show partition", func(t *testing.T) { coll, err := core.MetaTable.GetCollectionByName(collName, 0) assert.Nil(t, err) req := &milvuspb.ShowPartitionsRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_ShowPartitions, MsgID: 160, Timestamp: 160, SourceID: 160, }, DbName: "testDb", CollectionName: collName, CollectionID: coll.ID, } rsp, err := cli.ShowPartitions(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) assert.Equal(t, 2, len(rsp.PartitionNames)) assert.Equal(t, 2, len(rsp.PartitionIDs)) }) t.Run("show segment", func(t *testing.T) { coll, err := core.MetaTable.GetCollectionByName(collName, 0) assert.Nil(t, err) partID := coll.PartitionIDs[1] _, err = core.MetaTable.GetPartitionNameByID(coll.ID, partID, 0) assert.Nil(t, err) segLock.Lock() segs = []typeutil.UniqueID{1000} segLock.Unlock() req := &milvuspb.ShowSegmentsRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_ShowSegments, MsgID: 170, Timestamp: 170, SourceID: 170, }, CollectionID: coll.ID, PartitionID: partID, } rsp, err := cli.ShowSegments(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) assert.Equal(t, int64(1000), rsp.SegmentIDs[0]) assert.Equal(t, 1, len(rsp.SegmentIDs)) }) t.Run("create index", func(t *testing.T) { req := &milvuspb.CreateIndexRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_CreateIndex, MsgID: 180, Timestamp: 180, SourceID: 180, }, DbName: dbName, CollectionName: collName, FieldName: fieldName, ExtraParams: []*commonpb.KeyValuePair{ { Key: "ik1", Value: "iv1", }, }, } collMeta, err := core.MetaTable.GetCollectionByName(collName, 0) assert.Nil(t, err) assert.Zero(t, len(collMeta.FieldIndexes)) rsp, err := cli.CreateIndex(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.ErrorCode) collMeta, err = core.MetaTable.GetCollectionByName(collName, 0) assert.Nil(t, err) assert.Equal(t, 1, len(collMeta.FieldIndexes)) binlogLock.Lock() defer binlogLock.Unlock() assert.Equal(t, 3, len(binlogPathArray)) assert.ElementsMatch(t, binlogPathArray, []string{"file1", "file2", "file3"}) req.FieldName = "no field" rsp, err = cli.CreateIndex(ctx, req) assert.Nil(t, err) assert.NotEqual(t, commonpb.ErrorCode_Success, rsp.ErrorCode) }) t.Run("describe segment", func(t *testing.T) { coll, err := core.MetaTable.GetCollectionByName(collName, 0) assert.Nil(t, err) req := &milvuspb.DescribeSegmentRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_DescribeSegment, MsgID: 190, Timestamp: 190, SourceID: 190, }, CollectionID: coll.ID, SegmentID: 1000, } rsp, err := cli.DescribeSegment(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) t.Logf("index id = %d", rsp.IndexID) }) t.Run("describe index", func(t *testing.T) { req := &milvuspb.DescribeIndexRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_DescribeIndex, MsgID: 200, Timestamp: 200, SourceID: 200, }, DbName: dbName, CollectionName: collName, FieldName: fieldName, IndexName: "", } rsp, err := cli.DescribeIndex(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) assert.Equal(t, 1, len(rsp.IndexDescriptions)) assert.Equal(t, rootcoord.Params.CommonCfg.DefaultIndexName, rsp.IndexDescriptions[0].IndexName) }) t.Run("flush segment", func(t *testing.T) { coll, err := core.MetaTable.GetCollectionByName(collName, 0) assert.Nil(t, err) partID := coll.PartitionIDs[1] _, err = core.MetaTable.GetPartitionNameByID(coll.ID, partID, 0) assert.Nil(t, err) segLock.Lock() segs = append(segs, segID) segLock.Unlock() flushReq := &datapb.SegmentFlushCompletedMsg{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_SegmentFlushDone, }, Segment: &datapb.SegmentInfo{ ID: segID, CollectionID: coll.ID, PartitionID: partID, }, } flushRsp, err := cli.SegmentFlushCompleted(ctx, flushReq) assert.Nil(t, err) assert.Equal(t, flushRsp.ErrorCode, commonpb.ErrorCode_Success) flushRsp, err = cli.SegmentFlushCompleted(ctx, flushReq) assert.Nil(t, err) assert.Equal(t, flushRsp.ErrorCode, commonpb.ErrorCode_Success) req := &milvuspb.DescribeIndexRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_DescribeIndex, MsgID: 210, Timestamp: 210, SourceID: 210, }, DbName: dbName, CollectionName: collName, FieldName: fieldName, IndexName: "", } rsp, err := cli.DescribeIndex(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode) assert.Equal(t, 1, len(rsp.IndexDescriptions)) assert.Equal(t, rootcoord.Params.CommonCfg.DefaultIndexName, rsp.IndexDescriptions[0].IndexName) }) t.Run("drop index", func(t *testing.T) { req := &milvuspb.DropIndexRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_DropIndex, MsgID: 215, Timestamp: 215, SourceID: 215, }, DbName: dbName, CollectionName: collName, FieldName: fieldName, IndexName: rootcoord.Params.CommonCfg.DefaultIndexName, } _, idx, err := core.MetaTable.GetIndexByName(collName, rootcoord.Params.CommonCfg.DefaultIndexName) assert.Nil(t, err) assert.Equal(t, len(idx), 1) rsp, err := cli.DropIndex(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, rsp.ErrorCode) dropIDLock.Lock() assert.Equal(t, 1, len(dropID)) assert.Equal(t, idx[0].IndexID, dropID[0]) dropIDLock.Unlock() }) t.Run("drop partition", func(t *testing.T) { req := &milvuspb.DropPartitionRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_DropPartition, MsgID: 220, Timestamp: 220, SourceID: 220, }, DbName: dbName, CollectionName: collName, PartitionName: partName, } status, err := cli.DropPartition(ctx, req) assert.Nil(t, err) assert.Equal(t, commonpb.ErrorCode_Success, status.ErrorCode) collMeta, err := core.MetaTable.GetCollectionByName(collName, 0) assert.Nil(t, err) assert.Equal(t, 1, len(collMeta.PartitionIDs)) partName, err := core.MetaTable.GetPartitionNameByID(collMeta.ID, collMeta.PartitionIDs[0], 0) assert.Nil(t, err) assert.Equal(t, rootcoord.Params.CommonCfg.DefaultPartitionName, partName) assert.Equal(t, 2, len(collectionMetaCache)) }) t.Run("drop collection", func(t *testing.T) { req := &milvuspb.DropCollectionRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_DropCollection, MsgID: 230, Timestamp: 230, SourceID: 230, }, DbName: "testDb", CollectionName: collName, } status, err := cli.DropCollection(ctx, req) assert.Nil(t, err) assert.Equal(t, 1, len(dropCollectionArray)) assert.Equal(t, commonpb.MsgType_DropCollection, dropCollectionArray[0].Base.MsgType) assert.Equal(t, commonpb.ErrorCode_Success, status.ErrorCode) assert.Equal(t, collName, dropCollectionArray[0].CollectionName) assert.Equal(t, 3, len(collectionMetaCache)) assert.Equal(t, collName, collectionMetaCache[0]) req = &milvuspb.DropCollectionRequest{ Base: &commonpb.MsgBase{ MsgType: commonpb.MsgType_DropCollection, MsgID: 231, Timestamp: 231, SourceID: 231, }, DbName: "testDb", CollectionName: collName, } status, err = cli.DropCollection(ctx, req) assert.Nil(t, err) assert.Equal(t, 1, len(dropCollectionArray)) assert.Equal(t, commonpb.ErrorCode_UnexpectedError, status.ErrorCode) }) err = cli.Stop() assert.Nil(t, err) err = svr.Stop() assert.Nil(t, err) _, err = etcdCli.Delete(ctx, sessKey, clientv3.WithPrefix()) assert.Nil(t, err) } type mockCore struct { types.RootCoordComponent } func (m *mockCore) UpdateStateCode(internalpb.StateCode) { } func (m *mockCore) SetEtcdClient(etcdClient *clientv3.Client) { } func (m *mockCore) SetDataCoord(context.Context, types.DataCoord) error { return nil } func (m *mockCore) SetIndexCoord(types.IndexCoord) error { return nil } func (m *mockCore) SetQueryCoord(types.QueryCoord) error { return nil } func (m *mockCore) Register() error { return nil } func (m *mockCore) Init() error { return nil } func (m *mockCore) Start() error { return nil } func (m *mockCore) Stop() error { return fmt.Errorf("stop error") } func (m *mockCore) SetNewProxyClient(func(sess *sessionutil.Session) (types.Proxy, error)) { } type mockDataCoord struct { types.DataCoord } func (m *mockDataCoord) Init() error { return nil } func (m *mockDataCoord) Start() error { return nil } func (m *mockDataCoord) GetComponentStates(ctx context.Context) (*internalpb.ComponentStates, error) { return &internalpb.ComponentStates{ State: &internalpb.ComponentInfo{ StateCode: internalpb.StateCode_Healthy, }, Status: &commonpb.Status{ ErrorCode: commonpb.ErrorCode_Success, }, SubcomponentStates: []*internalpb.ComponentInfo{ { StateCode: internalpb.StateCode_Healthy, }, }, }, nil } func (m *mockDataCoord) Stop() error { return fmt.Errorf("stop error") } type mockIndex struct { types.IndexCoord } func (m *mockIndex) Init() error { return nil } func (m *mockIndex) Stop() error { return fmt.Errorf("stop error") } type mockQuery struct { types.QueryCoord } func (m *mockQuery) Init() error { return nil } func (m *mockQuery) Start() error { return nil } func (m *mockQuery) Stop() error { return fmt.Errorf("stop error") } func TestRun(t *testing.T) { ctx, cancel := context.WithCancel(context.Background()) svr := Server{ rootCoord: &mockCore{}, ctx: ctx, cancel: cancel, grpcErrChan: make(chan error), } Params.InitOnce(typeutil.RootCoordRole) Params.Port = 1000000 err := svr.Run() assert.NotNil(t, err) assert.EqualError(t, err, "listen tcp: address 1000000: invalid port") svr.newDataCoordClient = func(string, *clientv3.Client) types.DataCoord { return &mockDataCoord{} } svr.newIndexCoordClient = func(string, *clientv3.Client) types.IndexCoord { return &mockIndex{} } svr.newQueryCoordClient = func(string, *clientv3.Client) types.QueryCoord { return &mockQuery{} } Params.Port = rand.Int()%100 + 10000 rand.Seed(time.Now().UnixNano()) randVal := rand.Int() rootcoord.Params.Init() rootcoord.Params.EtcdCfg.MetaRootPath = fmt.Sprintf("/%d/test/meta", randVal) etcdCli, err := etcd.GetEtcdClient(&Params.EtcdCfg) assert.Nil(t, err) sessKey := path.Join(rootcoord.Params.EtcdCfg.MetaRootPath, sessionutil.DefaultServiceRoot) _, err = etcdCli.Delete(ctx, sessKey, clientv3.WithPrefix()) assert.Nil(t, err) err = svr.Run() assert.Nil(t, err) err = svr.Stop() assert.Nil(t, err) } func initEtcd(etcdEndpoints []string) (*clientv3.Client, error) { var etcdCli *clientv3.Client connectEtcdFn := func() error { etcd, err := clientv3.New(clientv3.Config{Endpoints: etcdEndpoints, DialTimeout: 5 * time.Second}) if err != nil { return err } etcdCli = etcd return nil } err := retry.Do(context.TODO(), connectEtcdFn, retry.Attempts(100)) if err != nil { return nil, err } return etcdCli, nil }