mirror of
https://gitee.com/milvus-io/milvus.git
synced 2024-12-04 12:59:23 +08:00
1c4b949a1d
Signed-off-by: groot <yihua.mo@zilliz.com>
952 lines
28 KiB
Go
952 lines
28 KiB
Go
// Licensed to the LF AI & Data foundation under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package grpcrootcoord
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"fmt"
|
|
"math/rand"
|
|
"path"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
|
|
clientv3 "go.etcd.io/etcd/client/v3"
|
|
|
|
"github.com/golang/protobuf/proto"
|
|
rcc "github.com/milvus-io/milvus/internal/distributed/rootcoord/client"
|
|
"github.com/milvus-io/milvus/internal/mq/msgstream"
|
|
"github.com/milvus-io/milvus/internal/proto/commonpb"
|
|
"github.com/milvus-io/milvus/internal/proto/datapb"
|
|
"github.com/milvus-io/milvus/internal/proto/etcdpb"
|
|
"github.com/milvus-io/milvus/internal/proto/internalpb"
|
|
"github.com/milvus-io/milvus/internal/proto/milvuspb"
|
|
"github.com/milvus-io/milvus/internal/proto/proxypb"
|
|
"github.com/milvus-io/milvus/internal/proto/rootcoordpb"
|
|
"github.com/milvus-io/milvus/internal/proto/schemapb"
|
|
"github.com/milvus-io/milvus/internal/rootcoord"
|
|
"github.com/milvus-io/milvus/internal/types"
|
|
"github.com/milvus-io/milvus/internal/util/etcd"
|
|
"github.com/milvus-io/milvus/internal/util/retry"
|
|
"github.com/milvus-io/milvus/internal/util/sessionutil"
|
|
"github.com/milvus-io/milvus/internal/util/typeutil"
|
|
"github.com/stretchr/testify/assert"
|
|
)
|
|
|
|
type proxyMock struct {
|
|
types.Proxy
|
|
invalidateCollectionMetaCache func(ctx context.Context, request *proxypb.InvalidateCollMetaCacheRequest) (*commonpb.Status, error)
|
|
}
|
|
|
|
func (p *proxyMock) InvalidateCollectionMetaCache(ctx context.Context, request *proxypb.InvalidateCollMetaCacheRequest) (*commonpb.Status, error) {
|
|
return p.invalidateCollectionMetaCache(ctx, request)
|
|
}
|
|
|
|
func TestGrpcService(t *testing.T) {
|
|
const (
|
|
dbName = "testDB"
|
|
collName = "testColl"
|
|
collName2 = "testColl-again"
|
|
partName = "testPartition"
|
|
fieldName = "vector"
|
|
fieldID = 100
|
|
segID = 1001
|
|
)
|
|
rand.Seed(time.Now().UnixNano())
|
|
randVal := rand.Int()
|
|
|
|
Params.InitOnce(typeutil.RootCoordRole)
|
|
Params.Port = (randVal % 100) + 10000
|
|
t.Log("newParams.Address:", Params.GetAddress())
|
|
|
|
ctx := context.Background()
|
|
msFactory := msgstream.NewPmsFactory()
|
|
svr, err := NewServer(ctx, msFactory)
|
|
assert.Nil(t, err)
|
|
|
|
rootcoord.Params.Init()
|
|
rootcoord.Params.EtcdCfg.MetaRootPath = fmt.Sprintf("/%d/test/meta", randVal)
|
|
rootcoord.Params.EtcdCfg.KvRootPath = fmt.Sprintf("/%d/test/kv", randVal)
|
|
rootcoord.Params.CommonCfg.RootCoordSubName = fmt.Sprintf("msgChannel%d", randVal)
|
|
rootcoord.Params.CommonCfg.RootCoordTimeTick = fmt.Sprintf("timeTick%d", randVal)
|
|
rootcoord.Params.CommonCfg.RootCoordStatistics = fmt.Sprintf("stateChannel%d", randVal)
|
|
|
|
rootcoord.Params.RootCoordCfg.MaxPartitionNum = 64
|
|
rootcoord.Params.CommonCfg.DefaultPartitionName = "_default"
|
|
rootcoord.Params.CommonCfg.DefaultIndexName = "_default"
|
|
|
|
t.Logf("service port = %d", Params.Port)
|
|
|
|
core, ok := (svr.rootCoord).(*rootcoord.Core)
|
|
assert.True(t, ok)
|
|
|
|
err = svr.startGrpc(Params.Port)
|
|
assert.Nil(t, err)
|
|
svr.rootCoord.UpdateStateCode(internalpb.StateCode_Initializing)
|
|
|
|
etcdCli, err := etcd.GetEtcdClient(&Params.EtcdCfg)
|
|
assert.Nil(t, err)
|
|
sessKey := path.Join(rootcoord.Params.EtcdCfg.MetaRootPath, sessionutil.DefaultServiceRoot)
|
|
_, err = etcdCli.Delete(ctx, sessKey, clientv3.WithPrefix())
|
|
assert.Nil(t, err)
|
|
|
|
pnb, err := json.Marshal(
|
|
&sessionutil.Session{
|
|
ServerID: 100,
|
|
},
|
|
)
|
|
assert.Nil(t, err)
|
|
_, err = etcdCli.Put(ctx, path.Join(sessKey, typeutil.ProxyRole+"-100"), string(pnb))
|
|
assert.Nil(t, err)
|
|
|
|
rootcoord.Params.RootCoordCfg.Address = Params.GetAddress()
|
|
|
|
core.SetEtcdClient(etcdCli)
|
|
err = core.Init()
|
|
assert.Nil(t, err)
|
|
|
|
timeTickArray := make([]typeutil.Timestamp, 0, 16)
|
|
timeTickLock := sync.Mutex{}
|
|
core.SendTimeTick = func(ts typeutil.Timestamp, reason string) error {
|
|
timeTickLock.Lock()
|
|
defer timeTickLock.Unlock()
|
|
t.Logf("send time tick %d", ts)
|
|
timeTickArray = append(timeTickArray, ts)
|
|
return nil
|
|
}
|
|
core.SendDdCreateCollectionReq = func(ctx context.Context, req *internalpb.CreateCollectionRequest, channelNames []string) (map[string][]byte, error) {
|
|
return map[string][]byte{}, nil
|
|
}
|
|
|
|
dropCollectionArray := make([]*internalpb.DropCollectionRequest, 0, 16)
|
|
core.SendDdDropCollectionReq = func(ctx context.Context, req *internalpb.DropCollectionRequest, channelNames []string) error {
|
|
t.Logf("Drop Collection %s", req.CollectionName)
|
|
dropCollectionArray = append(dropCollectionArray, req)
|
|
return nil
|
|
}
|
|
|
|
createPartitionArray := make([]*internalpb.CreatePartitionRequest, 0, 16)
|
|
core.SendDdCreatePartitionReq = func(ctx context.Context, req *internalpb.CreatePartitionRequest, channelNames []string) error {
|
|
t.Logf("Create Partition %s", req.PartitionName)
|
|
createPartitionArray = append(createPartitionArray, req)
|
|
return nil
|
|
}
|
|
|
|
dropPartitionArray := make([]*internalpb.DropPartitionRequest, 0, 16)
|
|
core.SendDdDropPartitionReq = func(ctx context.Context, req *internalpb.DropPartitionRequest, channelNames []string) error {
|
|
t.Logf("Drop Partition %s", req.PartitionName)
|
|
dropPartitionArray = append(dropPartitionArray, req)
|
|
return nil
|
|
}
|
|
|
|
core.CallGetBinlogFilePathsService = func(ctx context.Context, segID typeutil.UniqueID, fieldID typeutil.UniqueID) ([]string, error) {
|
|
return []string{"file1", "file2", "file3"}, nil
|
|
}
|
|
core.CallGetNumRowsService = func(ctx context.Context, segID typeutil.UniqueID, isFromFlushedChan bool) (int64, error) {
|
|
return rootcoord.Params.RootCoordCfg.MinSegmentSizeToEnableIndex, nil
|
|
}
|
|
core.CallWatchChannels = func(ctx context.Context, collectionID int64, channelNames []string) error {
|
|
return nil
|
|
}
|
|
|
|
segs := []typeutil.UniqueID{}
|
|
segLock := sync.Mutex{}
|
|
core.CallGetFlushedSegmentsService = func(ctx context.Context, collID, partID typeutil.UniqueID) ([]typeutil.UniqueID, error) {
|
|
segLock.Lock()
|
|
defer segLock.Unlock()
|
|
ret := []typeutil.UniqueID{}
|
|
ret = append(ret, segs...)
|
|
return ret, nil
|
|
}
|
|
|
|
var binlogLock sync.Mutex
|
|
binlogPathArray := make([]string, 0, 16)
|
|
core.CallBuildIndexService = func(ctx context.Context, binlog []string, field *schemapb.FieldSchema, idxInfo *etcdpb.IndexInfo, numRows int64) (typeutil.UniqueID, error) {
|
|
binlogLock.Lock()
|
|
defer binlogLock.Unlock()
|
|
binlogPathArray = append(binlogPathArray, binlog...)
|
|
return 2000, nil
|
|
}
|
|
|
|
var dropIDLock sync.Mutex
|
|
dropID := make([]typeutil.UniqueID, 0, 16)
|
|
core.CallDropIndexService = func(ctx context.Context, indexID typeutil.UniqueID) error {
|
|
dropIDLock.Lock()
|
|
defer dropIDLock.Unlock()
|
|
dropID = append(dropID, indexID)
|
|
return nil
|
|
}
|
|
|
|
collectionMetaCache := make([]string, 0, 16)
|
|
pnm := proxyMock{}
|
|
core.NewProxyClient = func(*sessionutil.Session) (types.Proxy, error) {
|
|
return &pnm, nil
|
|
}
|
|
pnm.invalidateCollectionMetaCache = func(ctx context.Context, request *proxypb.InvalidateCollMetaCacheRequest) (*commonpb.Status, error) {
|
|
collectionMetaCache = append(collectionMetaCache, request.CollectionName)
|
|
return &commonpb.Status{
|
|
ErrorCode: commonpb.ErrorCode_Success,
|
|
}, nil
|
|
}
|
|
|
|
core.CallReleaseCollectionService = func(ctx context.Context, ts typeutil.Timestamp, dbID typeutil.UniqueID, collectionID typeutil.UniqueID) error {
|
|
return nil
|
|
}
|
|
core.CallReleasePartitionService = func(ctx context.Context, ts typeutil.Timestamp, dbID, collectionID typeutil.UniqueID, partitionIDs []typeutil.UniqueID) error {
|
|
return nil
|
|
}
|
|
core.CallImportService = func(ctx context.Context, req *datapb.ImportTask) *datapb.ImportTaskResponse {
|
|
return nil
|
|
}
|
|
|
|
err = svr.start()
|
|
assert.Nil(t, err)
|
|
|
|
svr.rootCoord.UpdateStateCode(internalpb.StateCode_Healthy)
|
|
|
|
cli, err := rcc.NewClient(context.Background(), rootcoord.Params.EtcdCfg.MetaRootPath, etcdCli)
|
|
assert.Nil(t, err)
|
|
|
|
err = cli.Init()
|
|
assert.Nil(t, err)
|
|
|
|
err = cli.Start()
|
|
assert.Nil(t, err)
|
|
|
|
t.Run("get component states", func(t *testing.T) {
|
|
req := &internalpb.GetComponentStatesRequest{}
|
|
rsp, err := svr.GetComponentStates(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
})
|
|
|
|
t.Run("get time tick channel", func(t *testing.T) {
|
|
req := &internalpb.GetTimeTickChannelRequest{}
|
|
rsp, err := svr.GetTimeTickChannel(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
})
|
|
|
|
t.Run("get statistics channel", func(t *testing.T) {
|
|
req := &internalpb.GetStatisticsChannelRequest{}
|
|
rsp, err := svr.GetStatisticsChannel(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
})
|
|
|
|
t.Run("alloc time stamp", func(t *testing.T) {
|
|
req := &rootcoordpb.AllocTimestampRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_RequestTSO,
|
|
},
|
|
Count: 1,
|
|
}
|
|
rsp, err := svr.AllocTimestamp(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
})
|
|
|
|
t.Run("alloc id", func(t *testing.T) {
|
|
req := &rootcoordpb.AllocIDRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_RequestID,
|
|
},
|
|
Count: 1,
|
|
}
|
|
rsp, err := svr.AllocID(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
})
|
|
|
|
t.Run("update channel timetick", func(t *testing.T) {
|
|
req := &internalpb.ChannelTimeTickMsg{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_TimeTick,
|
|
},
|
|
}
|
|
status, err := svr.UpdateChannelTimeTick(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, status.ErrorCode)
|
|
})
|
|
|
|
t.Run("release DQL msg stream", func(t *testing.T) {
|
|
req := &proxypb.ReleaseDQLMessageStreamRequest{}
|
|
assert.Panics(t, func() { svr.ReleaseDQLMessageStream(ctx, req) })
|
|
})
|
|
|
|
t.Run("get metrics", func(t *testing.T) {
|
|
req := &milvuspb.GetMetricsRequest{}
|
|
rsp, err := svr.GetMetrics(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.NotEqual(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
})
|
|
|
|
t.Run("create collection", func(t *testing.T) {
|
|
schema := schemapb.CollectionSchema{
|
|
Name: collName,
|
|
AutoID: true,
|
|
Fields: []*schemapb.FieldSchema{
|
|
{
|
|
FieldID: fieldID,
|
|
Name: fieldName,
|
|
IsPrimaryKey: false,
|
|
DataType: schemapb.DataType_FloatVector,
|
|
TypeParams: nil,
|
|
IndexParams: []*commonpb.KeyValuePair{
|
|
{
|
|
Key: "ik1",
|
|
Value: "iv1",
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
sbf, err := proto.Marshal(&schema)
|
|
assert.Nil(t, err)
|
|
|
|
req := &milvuspb.CreateCollectionRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_CreateCollection,
|
|
MsgID: 100,
|
|
Timestamp: 100,
|
|
SourceID: 100,
|
|
},
|
|
DbName: dbName,
|
|
CollectionName: collName,
|
|
Schema: sbf,
|
|
}
|
|
|
|
status, err := cli.CreateCollection(ctx, req)
|
|
assert.Nil(t, err)
|
|
colls, err := core.MetaTable.ListCollections(0)
|
|
assert.Nil(t, err)
|
|
|
|
assert.Equal(t, 1, len(colls))
|
|
assert.Equal(t, commonpb.ErrorCode_Success, status.ErrorCode)
|
|
//assert.Equal(t, commonpb.MsgType_CreateCollection, createCollectionArray[0].Base.MsgType)
|
|
_, has := colls[collName]
|
|
assert.True(t, has)
|
|
|
|
req.Base.MsgID = 101
|
|
req.Base.Timestamp = 101
|
|
req.Base.SourceID = 101
|
|
status, err = cli.CreateCollection(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_UnexpectedError, status.ErrorCode)
|
|
|
|
req.Base.MsgID = 102
|
|
req.Base.Timestamp = 102
|
|
req.Base.SourceID = 102
|
|
req.CollectionName = collName2
|
|
status, err = cli.CreateCollection(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_UnexpectedError, status.ErrorCode)
|
|
|
|
schema.Name = req.CollectionName
|
|
sbf, err = proto.Marshal(&schema)
|
|
assert.Nil(t, err)
|
|
req.Schema = sbf
|
|
req.Base.MsgID = 103
|
|
req.Base.Timestamp = 103
|
|
req.Base.SourceID = 103
|
|
status, err = cli.CreateCollection(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, status.ErrorCode)
|
|
colls, err = core.MetaTable.ListCollections(0)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 2, len(colls))
|
|
_, has = colls[collName2]
|
|
assert.True(t, has)
|
|
//time stamp go back, master response to add the timestamp, so the time tick will never go back
|
|
//schema.Name = "testColl-goback"
|
|
//sbf, err = proto.Marshal(&schema)
|
|
//assert.Nil(t, err)
|
|
//req.CollectionName = schema.Name
|
|
//req.Schema = sbf
|
|
//req.Base.MsgID = 103
|
|
//req.Base.Timestamp = 103
|
|
//req.Base.SourceID = 103
|
|
//status, err = cli.CreateCollection(ctx, req)
|
|
//assert.Nil(t, err)
|
|
//assert.Equal(t, status.ErrorCode, commonpb.ErrorCode_UnexpectedError)
|
|
//matched, err := regexp.MatchString("input timestamp = [0-9]+, last dd time stamp = [0-9]+", status.Reason)
|
|
//assert.Nil(t, err)
|
|
//assert.True(t, matched)
|
|
})
|
|
|
|
t.Run("has collection", func(t *testing.T) {
|
|
req := &milvuspb.HasCollectionRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_HasCollection,
|
|
MsgID: 110,
|
|
Timestamp: 110,
|
|
SourceID: 110,
|
|
},
|
|
DbName: "testDb",
|
|
CollectionName: collName,
|
|
}
|
|
rsp, err := cli.HasCollection(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
assert.Equal(t, true, rsp.Value)
|
|
|
|
req = &milvuspb.HasCollectionRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_HasCollection,
|
|
MsgID: 111,
|
|
Timestamp: 111,
|
|
SourceID: 111,
|
|
},
|
|
DbName: "testDb",
|
|
CollectionName: "testColl2",
|
|
}
|
|
rsp, err = cli.HasCollection(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
assert.Equal(t, false, rsp.Value)
|
|
|
|
// test time stamp go back
|
|
req = &milvuspb.HasCollectionRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_HasCollection,
|
|
MsgID: 111,
|
|
Timestamp: 111,
|
|
SourceID: 111,
|
|
},
|
|
DbName: "testDb",
|
|
CollectionName: "testColl2",
|
|
}
|
|
rsp, err = cli.HasCollection(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
assert.Equal(t, false, rsp.Value)
|
|
})
|
|
|
|
t.Run("describe collection", func(t *testing.T) {
|
|
collMeta, err := core.MetaTable.GetCollectionByName(collName, 0)
|
|
assert.Nil(t, err)
|
|
req := &milvuspb.DescribeCollectionRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_DescribeCollection,
|
|
MsgID: 120,
|
|
Timestamp: 120,
|
|
SourceID: 120,
|
|
},
|
|
DbName: "testDb",
|
|
CollectionName: collName,
|
|
}
|
|
rsp, err := cli.DescribeCollection(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
assert.Equal(t, collName, rsp.Schema.Name)
|
|
assert.Equal(t, collMeta.ID, rsp.CollectionID)
|
|
})
|
|
|
|
t.Run("show collection", func(t *testing.T) {
|
|
req := &milvuspb.ShowCollectionsRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_ShowCollections,
|
|
MsgID: 130,
|
|
Timestamp: 130,
|
|
SourceID: 130,
|
|
},
|
|
DbName: "testDb",
|
|
}
|
|
rsp, err := cli.ShowCollections(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
assert.ElementsMatch(t, rsp.CollectionNames, []string{collName, collName2})
|
|
assert.Equal(t, 2, len(rsp.CollectionNames))
|
|
})
|
|
|
|
t.Run("create partition", func(t *testing.T) {
|
|
req := &milvuspb.CreatePartitionRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_CreatePartition,
|
|
MsgID: 140,
|
|
Timestamp: 140,
|
|
SourceID: 140,
|
|
},
|
|
DbName: dbName,
|
|
CollectionName: collName,
|
|
PartitionName: partName,
|
|
}
|
|
status, err := cli.CreatePartition(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, status.ErrorCode)
|
|
collMeta, err := core.MetaTable.GetCollectionByName(collName, 0)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 2, len(collMeta.PartitionIDs))
|
|
partName2, err := core.MetaTable.GetPartitionNameByID(collMeta.ID, collMeta.PartitionIDs[1], 0)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, partName, partName2)
|
|
assert.Equal(t, 1, len(collectionMetaCache))
|
|
})
|
|
|
|
t.Run("has partition", func(t *testing.T) {
|
|
req := &milvuspb.HasPartitionRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_HasPartition,
|
|
MsgID: 150,
|
|
Timestamp: 150,
|
|
SourceID: 150,
|
|
},
|
|
DbName: dbName,
|
|
CollectionName: collName,
|
|
PartitionName: partName,
|
|
}
|
|
rsp, err := cli.HasPartition(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
assert.Equal(t, true, rsp.Value)
|
|
})
|
|
|
|
t.Run("show partition", func(t *testing.T) {
|
|
coll, err := core.MetaTable.GetCollectionByName(collName, 0)
|
|
assert.Nil(t, err)
|
|
req := &milvuspb.ShowPartitionsRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_ShowPartitions,
|
|
MsgID: 160,
|
|
Timestamp: 160,
|
|
SourceID: 160,
|
|
},
|
|
DbName: "testDb",
|
|
CollectionName: collName,
|
|
CollectionID: coll.ID,
|
|
}
|
|
rsp, err := cli.ShowPartitions(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
assert.Equal(t, 2, len(rsp.PartitionNames))
|
|
assert.Equal(t, 2, len(rsp.PartitionIDs))
|
|
})
|
|
|
|
t.Run("show segment", func(t *testing.T) {
|
|
coll, err := core.MetaTable.GetCollectionByName(collName, 0)
|
|
assert.Nil(t, err)
|
|
partID := coll.PartitionIDs[1]
|
|
_, err = core.MetaTable.GetPartitionNameByID(coll.ID, partID, 0)
|
|
assert.Nil(t, err)
|
|
|
|
segLock.Lock()
|
|
segs = []typeutil.UniqueID{1000}
|
|
segLock.Unlock()
|
|
|
|
req := &milvuspb.ShowSegmentsRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_ShowSegments,
|
|
MsgID: 170,
|
|
Timestamp: 170,
|
|
SourceID: 170,
|
|
},
|
|
CollectionID: coll.ID,
|
|
PartitionID: partID,
|
|
}
|
|
rsp, err := cli.ShowSegments(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
assert.Equal(t, int64(1000), rsp.SegmentIDs[0])
|
|
assert.Equal(t, 1, len(rsp.SegmentIDs))
|
|
})
|
|
|
|
t.Run("create index", func(t *testing.T) {
|
|
req := &milvuspb.CreateIndexRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_CreateIndex,
|
|
MsgID: 180,
|
|
Timestamp: 180,
|
|
SourceID: 180,
|
|
},
|
|
DbName: dbName,
|
|
CollectionName: collName,
|
|
FieldName: fieldName,
|
|
ExtraParams: []*commonpb.KeyValuePair{
|
|
{
|
|
Key: "ik1",
|
|
Value: "iv1",
|
|
},
|
|
},
|
|
}
|
|
collMeta, err := core.MetaTable.GetCollectionByName(collName, 0)
|
|
assert.Nil(t, err)
|
|
assert.Zero(t, len(collMeta.FieldIndexes))
|
|
rsp, err := cli.CreateIndex(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.ErrorCode)
|
|
collMeta, err = core.MetaTable.GetCollectionByName(collName, 0)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, len(collMeta.FieldIndexes))
|
|
|
|
binlogLock.Lock()
|
|
defer binlogLock.Unlock()
|
|
assert.Equal(t, 3, len(binlogPathArray))
|
|
assert.ElementsMatch(t, binlogPathArray, []string{"file1", "file2", "file3"})
|
|
|
|
req.FieldName = "no field"
|
|
rsp, err = cli.CreateIndex(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.NotEqual(t, commonpb.ErrorCode_Success, rsp.ErrorCode)
|
|
})
|
|
|
|
t.Run("describe segment", func(t *testing.T) {
|
|
coll, err := core.MetaTable.GetCollectionByName(collName, 0)
|
|
assert.Nil(t, err)
|
|
|
|
req := &milvuspb.DescribeSegmentRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_DescribeSegment,
|
|
MsgID: 190,
|
|
Timestamp: 190,
|
|
SourceID: 190,
|
|
},
|
|
CollectionID: coll.ID,
|
|
SegmentID: 1000,
|
|
}
|
|
rsp, err := cli.DescribeSegment(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
t.Logf("index id = %d", rsp.IndexID)
|
|
})
|
|
|
|
t.Run("describe index", func(t *testing.T) {
|
|
req := &milvuspb.DescribeIndexRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_DescribeIndex,
|
|
MsgID: 200,
|
|
Timestamp: 200,
|
|
SourceID: 200,
|
|
},
|
|
DbName: dbName,
|
|
CollectionName: collName,
|
|
FieldName: fieldName,
|
|
IndexName: "",
|
|
}
|
|
rsp, err := cli.DescribeIndex(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
assert.Equal(t, 1, len(rsp.IndexDescriptions))
|
|
assert.Equal(t, rootcoord.Params.CommonCfg.DefaultIndexName, rsp.IndexDescriptions[0].IndexName)
|
|
})
|
|
|
|
t.Run("flush segment", func(t *testing.T) {
|
|
coll, err := core.MetaTable.GetCollectionByName(collName, 0)
|
|
assert.Nil(t, err)
|
|
partID := coll.PartitionIDs[1]
|
|
_, err = core.MetaTable.GetPartitionNameByID(coll.ID, partID, 0)
|
|
assert.Nil(t, err)
|
|
|
|
segLock.Lock()
|
|
segs = append(segs, segID)
|
|
segLock.Unlock()
|
|
|
|
flushReq := &datapb.SegmentFlushCompletedMsg{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_SegmentFlushDone,
|
|
},
|
|
Segment: &datapb.SegmentInfo{
|
|
ID: segID,
|
|
CollectionID: coll.ID,
|
|
PartitionID: partID,
|
|
},
|
|
}
|
|
flushRsp, err := cli.SegmentFlushCompleted(ctx, flushReq)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, flushRsp.ErrorCode, commonpb.ErrorCode_Success)
|
|
|
|
flushRsp, err = cli.SegmentFlushCompleted(ctx, flushReq)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, flushRsp.ErrorCode, commonpb.ErrorCode_Success)
|
|
|
|
req := &milvuspb.DescribeIndexRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_DescribeIndex,
|
|
MsgID: 210,
|
|
Timestamp: 210,
|
|
SourceID: 210,
|
|
},
|
|
DbName: dbName,
|
|
CollectionName: collName,
|
|
FieldName: fieldName,
|
|
IndexName: "",
|
|
}
|
|
rsp, err := cli.DescribeIndex(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.Status.ErrorCode)
|
|
assert.Equal(t, 1, len(rsp.IndexDescriptions))
|
|
assert.Equal(t, rootcoord.Params.CommonCfg.DefaultIndexName, rsp.IndexDescriptions[0].IndexName)
|
|
|
|
})
|
|
|
|
t.Run("drop index", func(t *testing.T) {
|
|
req := &milvuspb.DropIndexRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_DropIndex,
|
|
MsgID: 215,
|
|
Timestamp: 215,
|
|
SourceID: 215,
|
|
},
|
|
DbName: dbName,
|
|
CollectionName: collName,
|
|
FieldName: fieldName,
|
|
IndexName: rootcoord.Params.CommonCfg.DefaultIndexName,
|
|
}
|
|
_, idx, err := core.MetaTable.GetIndexByName(collName, rootcoord.Params.CommonCfg.DefaultIndexName)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, len(idx), 1)
|
|
rsp, err := cli.DropIndex(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, rsp.ErrorCode)
|
|
|
|
dropIDLock.Lock()
|
|
assert.Equal(t, 1, len(dropID))
|
|
assert.Equal(t, idx[0].IndexID, dropID[0])
|
|
dropIDLock.Unlock()
|
|
})
|
|
|
|
t.Run("drop partition", func(t *testing.T) {
|
|
req := &milvuspb.DropPartitionRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_DropPartition,
|
|
MsgID: 220,
|
|
Timestamp: 220,
|
|
SourceID: 220,
|
|
},
|
|
DbName: dbName,
|
|
CollectionName: collName,
|
|
PartitionName: partName,
|
|
}
|
|
status, err := cli.DropPartition(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, status.ErrorCode)
|
|
collMeta, err := core.MetaTable.GetCollectionByName(collName, 0)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, len(collMeta.PartitionIDs))
|
|
partName, err := core.MetaTable.GetPartitionNameByID(collMeta.ID, collMeta.PartitionIDs[0], 0)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, rootcoord.Params.CommonCfg.DefaultPartitionName, partName)
|
|
assert.Equal(t, 2, len(collectionMetaCache))
|
|
})
|
|
|
|
t.Run("drop collection", func(t *testing.T) {
|
|
req := &milvuspb.DropCollectionRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_DropCollection,
|
|
MsgID: 230,
|
|
Timestamp: 230,
|
|
SourceID: 230,
|
|
},
|
|
DbName: "testDb",
|
|
CollectionName: collName,
|
|
}
|
|
|
|
status, err := cli.DropCollection(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, len(dropCollectionArray))
|
|
assert.Equal(t, commonpb.MsgType_DropCollection, dropCollectionArray[0].Base.MsgType)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, status.ErrorCode)
|
|
assert.Equal(t, collName, dropCollectionArray[0].CollectionName)
|
|
assert.Equal(t, 3, len(collectionMetaCache))
|
|
assert.Equal(t, collName, collectionMetaCache[0])
|
|
|
|
req = &milvuspb.DropCollectionRequest{
|
|
Base: &commonpb.MsgBase{
|
|
MsgType: commonpb.MsgType_DropCollection,
|
|
MsgID: 231,
|
|
Timestamp: 231,
|
|
SourceID: 231,
|
|
},
|
|
DbName: "testDb",
|
|
CollectionName: collName,
|
|
}
|
|
status, err = cli.DropCollection(ctx, req)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, len(dropCollectionArray))
|
|
assert.Equal(t, commonpb.ErrorCode_UnexpectedError, status.ErrorCode)
|
|
})
|
|
|
|
err = cli.Stop()
|
|
assert.Nil(t, err)
|
|
|
|
err = svr.Stop()
|
|
assert.Nil(t, err)
|
|
|
|
_, err = etcdCli.Delete(ctx, sessKey, clientv3.WithPrefix())
|
|
assert.Nil(t, err)
|
|
|
|
}
|
|
|
|
type mockCore struct {
|
|
types.RootCoordComponent
|
|
}
|
|
|
|
func (m *mockCore) UpdateStateCode(internalpb.StateCode) {
|
|
}
|
|
|
|
func (m *mockCore) SetEtcdClient(etcdClient *clientv3.Client) {
|
|
}
|
|
|
|
func (m *mockCore) SetDataCoord(context.Context, types.DataCoord) error {
|
|
return nil
|
|
}
|
|
func (m *mockCore) SetIndexCoord(types.IndexCoord) error {
|
|
return nil
|
|
}
|
|
|
|
func (m *mockCore) SetQueryCoord(types.QueryCoord) error {
|
|
return nil
|
|
}
|
|
|
|
func (m *mockCore) Register() error {
|
|
return nil
|
|
}
|
|
|
|
func (m *mockCore) Init() error {
|
|
return nil
|
|
}
|
|
|
|
func (m *mockCore) Start() error {
|
|
return nil
|
|
}
|
|
|
|
func (m *mockCore) Stop() error {
|
|
return fmt.Errorf("stop error")
|
|
}
|
|
|
|
func (m *mockCore) SetNewProxyClient(func(sess *sessionutil.Session) (types.Proxy, error)) {
|
|
}
|
|
|
|
type mockDataCoord struct {
|
|
types.DataCoord
|
|
}
|
|
|
|
func (m *mockDataCoord) Init() error {
|
|
return nil
|
|
}
|
|
func (m *mockDataCoord) Start() error {
|
|
return nil
|
|
}
|
|
func (m *mockDataCoord) GetComponentStates(ctx context.Context) (*internalpb.ComponentStates, error) {
|
|
return &internalpb.ComponentStates{
|
|
State: &internalpb.ComponentInfo{
|
|
StateCode: internalpb.StateCode_Healthy,
|
|
},
|
|
Status: &commonpb.Status{
|
|
ErrorCode: commonpb.ErrorCode_Success,
|
|
},
|
|
SubcomponentStates: []*internalpb.ComponentInfo{
|
|
{
|
|
StateCode: internalpb.StateCode_Healthy,
|
|
},
|
|
},
|
|
}, nil
|
|
}
|
|
func (m *mockDataCoord) Stop() error {
|
|
return fmt.Errorf("stop error")
|
|
}
|
|
|
|
type mockIndex struct {
|
|
types.IndexCoord
|
|
}
|
|
|
|
func (m *mockIndex) Init() error {
|
|
return nil
|
|
}
|
|
|
|
func (m *mockIndex) Stop() error {
|
|
return fmt.Errorf("stop error")
|
|
}
|
|
|
|
type mockQuery struct {
|
|
types.QueryCoord
|
|
}
|
|
|
|
func (m *mockQuery) Init() error {
|
|
return nil
|
|
}
|
|
|
|
func (m *mockQuery) Start() error {
|
|
return nil
|
|
}
|
|
|
|
func (m *mockQuery) Stop() error {
|
|
return fmt.Errorf("stop error")
|
|
}
|
|
|
|
func TestRun(t *testing.T) {
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
svr := Server{
|
|
rootCoord: &mockCore{},
|
|
ctx: ctx,
|
|
cancel: cancel,
|
|
grpcErrChan: make(chan error),
|
|
}
|
|
Params.InitOnce(typeutil.RootCoordRole)
|
|
Params.Port = 1000000
|
|
err := svr.Run()
|
|
assert.NotNil(t, err)
|
|
assert.EqualError(t, err, "listen tcp: address 1000000: invalid port")
|
|
|
|
svr.newDataCoordClient = func(string, *clientv3.Client) types.DataCoord {
|
|
return &mockDataCoord{}
|
|
}
|
|
svr.newIndexCoordClient = func(string, *clientv3.Client) types.IndexCoord {
|
|
return &mockIndex{}
|
|
}
|
|
svr.newQueryCoordClient = func(string, *clientv3.Client) types.QueryCoord {
|
|
return &mockQuery{}
|
|
}
|
|
|
|
Params.Port = rand.Int()%100 + 10000
|
|
|
|
rand.Seed(time.Now().UnixNano())
|
|
randVal := rand.Int()
|
|
rootcoord.Params.Init()
|
|
rootcoord.Params.EtcdCfg.MetaRootPath = fmt.Sprintf("/%d/test/meta", randVal)
|
|
|
|
etcdCli, err := etcd.GetEtcdClient(&Params.EtcdCfg)
|
|
assert.Nil(t, err)
|
|
sessKey := path.Join(rootcoord.Params.EtcdCfg.MetaRootPath, sessionutil.DefaultServiceRoot)
|
|
_, err = etcdCli.Delete(ctx, sessKey, clientv3.WithPrefix())
|
|
assert.Nil(t, err)
|
|
err = svr.Run()
|
|
assert.Nil(t, err)
|
|
|
|
err = svr.Stop()
|
|
assert.Nil(t, err)
|
|
|
|
}
|
|
|
|
func initEtcd(etcdEndpoints []string) (*clientv3.Client, error) {
|
|
var etcdCli *clientv3.Client
|
|
connectEtcdFn := func() error {
|
|
etcd, err := clientv3.New(clientv3.Config{Endpoints: etcdEndpoints, DialTimeout: 5 * time.Second})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
etcdCli = etcd
|
|
return nil
|
|
}
|
|
err := retry.Do(context.TODO(), connectEtcdFn, retry.Attempts(300))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
return etcdCli, nil
|
|
}
|