mirror of
https://gitee.com/milvus-io/milvus.git
synced 2024-12-01 03:18:29 +08:00
b5e62023ea
Signed-off-by: groot <yihua.mo@zilliz.com>
560 lines
16 KiB
Go
560 lines
16 KiB
Go
// Licensed to the LF AI & Data foundation under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package rootcoord
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"strconv"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/golang/protobuf/proto"
|
|
"github.com/milvus-io/milvus/internal/kv"
|
|
"github.com/milvus-io/milvus/internal/proto/commonpb"
|
|
"github.com/milvus-io/milvus/internal/proto/datapb"
|
|
"github.com/milvus-io/milvus/internal/proto/milvuspb"
|
|
"github.com/milvus-io/milvus/internal/proto/rootcoordpb"
|
|
"github.com/milvus-io/milvus/internal/util/typeutil"
|
|
"github.com/stretchr/testify/assert"
|
|
)
|
|
|
|
type customKV struct {
|
|
kv.MockMetaKV
|
|
}
|
|
|
|
func TestImportManager_NewImportManager(t *testing.T) {
|
|
var countLock sync.RWMutex
|
|
var globalCount = typeutil.UniqueID(0)
|
|
|
|
var idAlloc = func(count uint32) (typeutil.UniqueID, typeutil.UniqueID, error) {
|
|
countLock.Lock()
|
|
defer countLock.Unlock()
|
|
globalCount++
|
|
return globalCount, 0, nil
|
|
}
|
|
Params.RootCoordCfg.ImportTaskSubPath = "test_import_task"
|
|
Params.RootCoordCfg.ImportTaskExpiration = 100
|
|
checkPendingTasksInterval = 100
|
|
expireOldTasksInterval = 100
|
|
mockKv := &kv.MockMetaKV{}
|
|
mockKv.InMemKv = make(map[string]string)
|
|
ti1 := &datapb.ImportTaskInfo{
|
|
Id: 100,
|
|
State: &datapb.ImportTaskState{
|
|
StateCode: commonpb.ImportState_ImportPending,
|
|
},
|
|
}
|
|
ti2 := &datapb.ImportTaskInfo{
|
|
Id: 200,
|
|
State: &datapb.ImportTaskState{
|
|
StateCode: commonpb.ImportState_ImportPersisted,
|
|
},
|
|
}
|
|
taskInfo1, err := proto.Marshal(ti1)
|
|
assert.NoError(t, err)
|
|
taskInfo2, err := proto.Marshal(ti2)
|
|
assert.NoError(t, err)
|
|
mockKv.SaveWithLease(BuildImportTaskKey(1), "value", 1)
|
|
mockKv.SaveWithLease(BuildImportTaskKey(2), string(taskInfo1), 2)
|
|
mockKv.SaveWithLease(BuildImportTaskKey(3), string(taskInfo2), 3)
|
|
fn := func(ctx context.Context, req *datapb.ImportTaskRequest) *datapb.ImportTaskResponse {
|
|
return &datapb.ImportTaskResponse{
|
|
Status: &commonpb.Status{
|
|
ErrorCode: commonpb.ErrorCode_Success,
|
|
},
|
|
}
|
|
}
|
|
|
|
time.Sleep(1 * time.Second)
|
|
|
|
var wg sync.WaitGroup
|
|
wg.Add(1)
|
|
t.Run("working task expired", func(t *testing.T) {
|
|
defer wg.Done()
|
|
ctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)
|
|
defer cancel()
|
|
mgr := newImportManager(ctx, mockKv, idAlloc, fn, nil)
|
|
assert.NotNil(t, mgr)
|
|
mgr.init(ctx)
|
|
var wgLoop sync.WaitGroup
|
|
wgLoop.Add(2)
|
|
mgr.expireOldTasksLoop(&wgLoop, func(ctx context.Context, int64 int64, int64s []int64) error {
|
|
return nil
|
|
})
|
|
mgr.sendOutTasksLoop(&wgLoop)
|
|
wgLoop.Wait()
|
|
})
|
|
|
|
wg.Add(1)
|
|
t.Run("context done", func(t *testing.T) {
|
|
defer wg.Done()
|
|
ctx, cancel := context.WithTimeout(context.Background(), 1*time.Nanosecond)
|
|
defer cancel()
|
|
mgr := newImportManager(ctx, mockKv, idAlloc, fn, nil)
|
|
assert.NotNil(t, mgr)
|
|
mgr.init(context.TODO())
|
|
var wgLoop sync.WaitGroup
|
|
wgLoop.Add(2)
|
|
mgr.expireOldTasksLoop(&wgLoop, func(ctx context.Context, int64 int64, int64s []int64) error {
|
|
return nil
|
|
})
|
|
mgr.sendOutTasksLoop(&wgLoop)
|
|
wgLoop.Wait()
|
|
})
|
|
|
|
wg.Add(1)
|
|
t.Run("pending task expired", func(t *testing.T) {
|
|
defer wg.Done()
|
|
ctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)
|
|
defer cancel()
|
|
mgr := newImportManager(ctx, mockKv, idAlloc, fn, nil)
|
|
assert.NotNil(t, mgr)
|
|
mgr.pendingTasks = append(mgr.pendingTasks, &datapb.ImportTaskInfo{
|
|
Id: 300,
|
|
State: &datapb.ImportTaskState{
|
|
StateCode: commonpb.ImportState_ImportPending,
|
|
},
|
|
CreateTs: time.Now().Unix() - 10,
|
|
})
|
|
mgr.loadFromTaskStore()
|
|
var wgLoop sync.WaitGroup
|
|
wgLoop.Add(2)
|
|
mgr.expireOldTasksLoop(&wgLoop, func(ctx context.Context, int64 int64, int64s []int64) error {
|
|
return nil
|
|
})
|
|
mgr.sendOutTasksLoop(&wgLoop)
|
|
wgLoop.Wait()
|
|
})
|
|
|
|
wg.Add(1)
|
|
t.Run("check init", func(t *testing.T) {
|
|
defer wg.Done()
|
|
ctx, cancel := context.WithTimeout(context.Background(), 3*time.Second)
|
|
defer cancel()
|
|
mgr := newImportManager(ctx, mockKv, idAlloc, fn, nil)
|
|
assert.NotNil(t, mgr)
|
|
mgr.init(ctx)
|
|
var wgLoop sync.WaitGroup
|
|
wgLoop.Add(2)
|
|
mgr.expireOldTasksLoop(&wgLoop, func(ctx context.Context, int64 int64, int64s []int64) error {
|
|
return nil
|
|
})
|
|
mgr.sendOutTasksLoop(&wgLoop)
|
|
time.Sleep(500 * time.Millisecond)
|
|
wgLoop.Wait()
|
|
})
|
|
|
|
wg.Wait()
|
|
}
|
|
|
|
func TestImportManager_ImportJob(t *testing.T) {
|
|
var countLock sync.RWMutex
|
|
var globalCount = typeutil.UniqueID(0)
|
|
|
|
var idAlloc = func(count uint32) (typeutil.UniqueID, typeutil.UniqueID, error) {
|
|
countLock.Lock()
|
|
defer countLock.Unlock()
|
|
globalCount++
|
|
return globalCount, 0, nil
|
|
}
|
|
Params.RootCoordCfg.ImportTaskSubPath = "test_import_task"
|
|
colID := int64(100)
|
|
mockKv := &kv.MockMetaKV{}
|
|
mockKv.InMemKv = make(map[string]string)
|
|
mgr := newImportManager(context.TODO(), mockKv, idAlloc, nil, nil)
|
|
resp := mgr.importJob(context.TODO(), nil, colID, 0)
|
|
assert.NotEqual(t, commonpb.ErrorCode_Success, resp.Status.ErrorCode)
|
|
|
|
rowReq := &milvuspb.ImportRequest{
|
|
CollectionName: "c1",
|
|
PartitionName: "p1",
|
|
RowBased: true,
|
|
Files: []string{"f1", "f2", "f3"},
|
|
}
|
|
|
|
resp = mgr.importJob(context.TODO(), rowReq, colID, 0)
|
|
assert.NotEqual(t, commonpb.ErrorCode_Success, resp.Status.ErrorCode)
|
|
|
|
colReq := &milvuspb.ImportRequest{
|
|
CollectionName: "c1",
|
|
PartitionName: "p1",
|
|
RowBased: false,
|
|
Files: []string{"f1", "f2"},
|
|
Options: []*commonpb.KeyValuePair{
|
|
{
|
|
Key: Bucket,
|
|
Value: "mybucket",
|
|
},
|
|
},
|
|
}
|
|
|
|
fn := func(ctx context.Context, req *datapb.ImportTaskRequest) *datapb.ImportTaskResponse {
|
|
return &datapb.ImportTaskResponse{
|
|
Status: &commonpb.Status{
|
|
ErrorCode: commonpb.ErrorCode_UnexpectedError,
|
|
},
|
|
}
|
|
}
|
|
|
|
mgr = newImportManager(context.TODO(), mockKv, idAlloc, fn, nil)
|
|
resp = mgr.importJob(context.TODO(), rowReq, colID, 0)
|
|
assert.Equal(t, len(rowReq.Files), len(mgr.pendingTasks))
|
|
assert.Equal(t, 0, len(mgr.workingTasks))
|
|
|
|
mgr = newImportManager(context.TODO(), mockKv, idAlloc, fn, nil)
|
|
resp = mgr.importJob(context.TODO(), colReq, colID, 0)
|
|
assert.Equal(t, 1, len(mgr.pendingTasks))
|
|
assert.Equal(t, 0, len(mgr.workingTasks))
|
|
|
|
fn = func(ctx context.Context, req *datapb.ImportTaskRequest) *datapb.ImportTaskResponse {
|
|
return &datapb.ImportTaskResponse{
|
|
Status: &commonpb.Status{
|
|
ErrorCode: commonpb.ErrorCode_Success,
|
|
},
|
|
}
|
|
}
|
|
|
|
mgr = newImportManager(context.TODO(), mockKv, idAlloc, fn, nil)
|
|
resp = mgr.importJob(context.TODO(), rowReq, colID, 0)
|
|
assert.Equal(t, 0, len(mgr.pendingTasks))
|
|
assert.Equal(t, len(rowReq.Files), len(mgr.workingTasks))
|
|
|
|
mgr = newImportManager(context.TODO(), mockKv, idAlloc, fn, nil)
|
|
resp = mgr.importJob(context.TODO(), colReq, colID, 0)
|
|
assert.Equal(t, 0, len(mgr.pendingTasks))
|
|
assert.Equal(t, 1, len(mgr.workingTasks))
|
|
|
|
count := 0
|
|
fn = func(ctx context.Context, req *datapb.ImportTaskRequest) *datapb.ImportTaskResponse {
|
|
if count >= 2 {
|
|
return &datapb.ImportTaskResponse{
|
|
Status: &commonpb.Status{
|
|
ErrorCode: commonpb.ErrorCode_UnexpectedError,
|
|
},
|
|
}
|
|
}
|
|
count++
|
|
return &datapb.ImportTaskResponse{
|
|
Status: &commonpb.Status{
|
|
ErrorCode: commonpb.ErrorCode_Success,
|
|
},
|
|
}
|
|
}
|
|
|
|
mgr = newImportManager(context.TODO(), mockKv, idAlloc, fn, nil)
|
|
resp = mgr.importJob(context.TODO(), rowReq, colID, 0)
|
|
assert.Equal(t, len(rowReq.Files)-2, len(mgr.pendingTasks))
|
|
assert.Equal(t, 2, len(mgr.workingTasks))
|
|
|
|
for i := 0; i <= 32; i++ {
|
|
rowReq.Files = append(rowReq.Files, strconv.Itoa(i))
|
|
}
|
|
resp = mgr.importJob(context.TODO(), rowReq, colID, 0)
|
|
assert.NotEqual(t, commonpb.ErrorCode_Success, resp.Status.ErrorCode)
|
|
}
|
|
|
|
func TestImportManager_AllDataNodesBusy(t *testing.T) {
|
|
var countLock sync.RWMutex
|
|
var globalCount = typeutil.UniqueID(0)
|
|
|
|
var idAlloc = func(count uint32) (typeutil.UniqueID, typeutil.UniqueID, error) {
|
|
countLock.Lock()
|
|
defer countLock.Unlock()
|
|
globalCount++
|
|
return globalCount, 0, nil
|
|
}
|
|
Params.RootCoordCfg.ImportTaskSubPath = "test_import_task"
|
|
colID := int64(100)
|
|
mockKv := &kv.MockMetaKV{}
|
|
mockKv.InMemKv = make(map[string]string)
|
|
rowReq := &milvuspb.ImportRequest{
|
|
CollectionName: "c1",
|
|
PartitionName: "p1",
|
|
RowBased: true,
|
|
Files: []string{"f1", "f2", "f3"},
|
|
}
|
|
colReq := &milvuspb.ImportRequest{
|
|
CollectionName: "c1",
|
|
PartitionName: "p1",
|
|
RowBased: false,
|
|
Files: []string{"f1", "f2"},
|
|
Options: []*commonpb.KeyValuePair{
|
|
{
|
|
Key: Bucket,
|
|
Value: "mybucket",
|
|
},
|
|
},
|
|
}
|
|
|
|
dnList := []int64{1, 2, 3}
|
|
count := 0
|
|
fn := func(ctx context.Context, req *datapb.ImportTaskRequest) *datapb.ImportTaskResponse {
|
|
if count < len(dnList) {
|
|
count++
|
|
return &datapb.ImportTaskResponse{
|
|
Status: &commonpb.Status{
|
|
ErrorCode: commonpb.ErrorCode_Success,
|
|
},
|
|
DatanodeId: dnList[count-1],
|
|
}
|
|
}
|
|
return &datapb.ImportTaskResponse{
|
|
Status: &commonpb.Status{
|
|
ErrorCode: commonpb.ErrorCode_UnexpectedError,
|
|
},
|
|
}
|
|
}
|
|
|
|
mgr := newImportManager(context.TODO(), mockKv, idAlloc, fn, nil)
|
|
mgr.importJob(context.TODO(), rowReq, colID, 0)
|
|
assert.Equal(t, 0, len(mgr.pendingTasks))
|
|
assert.Equal(t, len(rowReq.Files), len(mgr.workingTasks))
|
|
|
|
mgr = newImportManager(context.TODO(), mockKv, idAlloc, fn, nil)
|
|
mgr.importJob(context.TODO(), rowReq, colID, 0)
|
|
assert.Equal(t, len(rowReq.Files), len(mgr.pendingTasks))
|
|
assert.Equal(t, 0, len(mgr.workingTasks))
|
|
|
|
// Reset count.
|
|
count = 0
|
|
mgr = newImportManager(context.TODO(), mockKv, idAlloc, fn, nil)
|
|
mgr.importJob(context.TODO(), colReq, colID, 0)
|
|
assert.Equal(t, 0, len(mgr.pendingTasks))
|
|
assert.Equal(t, 1, len(mgr.workingTasks))
|
|
|
|
mgr.importJob(context.TODO(), colReq, colID, 0)
|
|
assert.Equal(t, 0, len(mgr.pendingTasks))
|
|
assert.Equal(t, 2, len(mgr.workingTasks))
|
|
|
|
mgr.importJob(context.TODO(), colReq, colID, 0)
|
|
assert.Equal(t, 0, len(mgr.pendingTasks))
|
|
assert.Equal(t, 3, len(mgr.workingTasks))
|
|
|
|
mgr.importJob(context.TODO(), colReq, colID, 0)
|
|
assert.Equal(t, 1, len(mgr.pendingTasks))
|
|
assert.Equal(t, 3, len(mgr.workingTasks))
|
|
}
|
|
|
|
func TestImportManager_TaskState(t *testing.T) {
|
|
var countLock sync.RWMutex
|
|
var globalCount = typeutil.UniqueID(0)
|
|
|
|
var idAlloc = func(count uint32) (typeutil.UniqueID, typeutil.UniqueID, error) {
|
|
countLock.Lock()
|
|
defer countLock.Unlock()
|
|
globalCount++
|
|
return globalCount, 0, nil
|
|
}
|
|
Params.RootCoordCfg.ImportTaskSubPath = "test_import_task"
|
|
colID := int64(100)
|
|
mockKv := &kv.MockMetaKV{}
|
|
mockKv.InMemKv = make(map[string]string)
|
|
fn := func(ctx context.Context, req *datapb.ImportTaskRequest) *datapb.ImportTaskResponse {
|
|
return &datapb.ImportTaskResponse{
|
|
Status: &commonpb.Status{
|
|
ErrorCode: commonpb.ErrorCode_Success,
|
|
},
|
|
}
|
|
}
|
|
|
|
rowReq := &milvuspb.ImportRequest{
|
|
CollectionName: "c1",
|
|
PartitionName: "p1",
|
|
RowBased: true,
|
|
Files: []string{"f1", "f2", "f3"},
|
|
}
|
|
|
|
mgr := newImportManager(context.TODO(), mockKv, idAlloc, fn, nil)
|
|
mgr.importJob(context.TODO(), rowReq, colID, 0)
|
|
|
|
state := &rootcoordpb.ImportResult{
|
|
TaskId: 10000,
|
|
}
|
|
_, err := mgr.updateTaskState(state)
|
|
assert.NotNil(t, err)
|
|
|
|
state = &rootcoordpb.ImportResult{
|
|
TaskId: 2,
|
|
RowCount: 1000,
|
|
State: commonpb.ImportState_ImportCompleted,
|
|
Infos: []*commonpb.KeyValuePair{
|
|
{
|
|
Key: "key1",
|
|
Value: "value1",
|
|
},
|
|
{
|
|
Key: "failed_reason",
|
|
Value: "some_reason",
|
|
},
|
|
},
|
|
}
|
|
ti, err := mgr.updateTaskState(state)
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, int64(2), ti.GetId())
|
|
assert.Equal(t, int64(100), ti.GetCollectionId())
|
|
assert.Equal(t, int64(100), ti.GetCollectionId())
|
|
assert.Equal(t, int64(0), ti.GetPartitionId())
|
|
assert.Equal(t, true, ti.GetRowBased())
|
|
assert.Equal(t, []string{"f2"}, ti.GetFiles())
|
|
assert.Equal(t, commonpb.ImportState_ImportCompleted, ti.GetState().GetStateCode())
|
|
assert.Equal(t, int64(1000), ti.GetState().GetRowCount())
|
|
|
|
resp := mgr.getTaskState(10000)
|
|
assert.Equal(t, commonpb.ErrorCode_UnexpectedError, resp.Status.ErrorCode)
|
|
|
|
resp = mgr.getTaskState(2)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, resp.Status.ErrorCode)
|
|
assert.Equal(t, commonpb.ImportState_ImportCompleted, resp.State)
|
|
|
|
resp = mgr.getTaskState(1)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, resp.Status.ErrorCode)
|
|
assert.Equal(t, commonpb.ImportState_ImportPending, resp.State)
|
|
}
|
|
|
|
func TestImportManager_AllocFail(t *testing.T) {
|
|
var idAlloc = func(count uint32) (typeutil.UniqueID, typeutil.UniqueID, error) {
|
|
return 0, 0, errors.New("injected failure")
|
|
}
|
|
Params.RootCoordCfg.ImportTaskSubPath = "test_import_task"
|
|
colID := int64(100)
|
|
mockKv := &kv.MockMetaKV{}
|
|
mockKv.InMemKv = make(map[string]string)
|
|
fn := func(ctx context.Context, req *datapb.ImportTaskRequest) *datapb.ImportTaskResponse {
|
|
return &datapb.ImportTaskResponse{
|
|
Status: &commonpb.Status{
|
|
ErrorCode: commonpb.ErrorCode_Success,
|
|
},
|
|
}
|
|
}
|
|
|
|
rowReq := &milvuspb.ImportRequest{
|
|
CollectionName: "c1",
|
|
PartitionName: "p1",
|
|
RowBased: true,
|
|
Files: []string{"f1", "f2", "f3"},
|
|
}
|
|
|
|
mgr := newImportManager(context.TODO(), mockKv, idAlloc, fn, nil)
|
|
mgr.importJob(context.TODO(), rowReq, colID, 0)
|
|
}
|
|
|
|
func TestImportManager_ListAllTasks(t *testing.T) {
|
|
var countLock sync.RWMutex
|
|
var globalCount = typeutil.UniqueID(0)
|
|
|
|
var idAlloc = func(count uint32) (typeutil.UniqueID, typeutil.UniqueID, error) {
|
|
countLock.Lock()
|
|
defer countLock.Unlock()
|
|
globalCount++
|
|
return globalCount, 0, nil
|
|
}
|
|
|
|
Params.RootCoordCfg.ImportTaskSubPath = "test_import_task"
|
|
colID := int64(100)
|
|
mockKv := &kv.MockMetaKV{}
|
|
mockKv.InMemKv = make(map[string]string)
|
|
|
|
// reject some tasks so there are 3 tasks left in pending list
|
|
fn := func(ctx context.Context, req *datapb.ImportTaskRequest) *datapb.ImportTaskResponse {
|
|
return &datapb.ImportTaskResponse{
|
|
Status: &commonpb.Status{
|
|
ErrorCode: commonpb.ErrorCode_UnexpectedError,
|
|
},
|
|
}
|
|
}
|
|
|
|
rowReq := &milvuspb.ImportRequest{
|
|
CollectionName: "c1",
|
|
PartitionName: "p1",
|
|
RowBased: true,
|
|
Files: []string{"f1", "f2", "f3"},
|
|
}
|
|
|
|
mgr := newImportManager(context.TODO(), mockKv, idAlloc, fn, nil)
|
|
mgr.importJob(context.TODO(), rowReq, colID, 0)
|
|
|
|
tasks := mgr.listAllTasks()
|
|
assert.Equal(t, len(rowReq.Files), len(tasks))
|
|
|
|
resp := mgr.getTaskState(1)
|
|
assert.Equal(t, commonpb.ErrorCode_Success, resp.Status.ErrorCode)
|
|
assert.Equal(t, commonpb.ImportState_ImportPending, resp.State)
|
|
assert.Equal(t, int64(1), resp.Id)
|
|
|
|
// accept tasks to working list
|
|
mgr.callImportService = func(ctx context.Context, req *datapb.ImportTaskRequest) *datapb.ImportTaskResponse {
|
|
return &datapb.ImportTaskResponse{
|
|
Status: &commonpb.Status{
|
|
ErrorCode: commonpb.ErrorCode_Success,
|
|
},
|
|
}
|
|
}
|
|
|
|
mgr.importJob(context.TODO(), rowReq, colID, 0)
|
|
tasks = mgr.listAllTasks()
|
|
assert.Equal(t, len(rowReq.Files)*2, len(tasks))
|
|
|
|
// the id of tasks must be 1,2,3,4,5,6(sequence not guaranteed)
|
|
ids := make(map[int64]struct{})
|
|
for i := 0; i < len(tasks); i++ {
|
|
ids[int64(i)+1] = struct{}{}
|
|
}
|
|
for i := 0; i < len(tasks); i++ {
|
|
delete(ids, tasks[i].Id)
|
|
}
|
|
assert.Equal(t, 0, len(ids))
|
|
}
|
|
|
|
func TestImportManager_getCollectionPartitionName(t *testing.T) {
|
|
mgr := &importManager{
|
|
getCollectionName: func(collID, partitionID typeutil.UniqueID) (string, string, error) {
|
|
return "c1", "p1", nil
|
|
},
|
|
}
|
|
|
|
task := &datapb.ImportTaskInfo{
|
|
CollectionId: 1,
|
|
PartitionId: 2,
|
|
}
|
|
resp := &milvuspb.GetImportStateResponse{
|
|
Infos: make([]*commonpb.KeyValuePair, 0),
|
|
}
|
|
mgr.getCollectionPartitionName(task, resp)
|
|
assert.Equal(t, "c1", resp.Infos[0].Value)
|
|
assert.Equal(t, "p1", resp.Infos[1].Value)
|
|
}
|
|
|
|
func TestImportManager_rearrangeTasks(t *testing.T) {
|
|
tasks := make([]*milvuspb.GetImportStateResponse, 0)
|
|
tasks = append(tasks, &milvuspb.GetImportStateResponse{
|
|
Id: 100,
|
|
})
|
|
tasks = append(tasks, &milvuspb.GetImportStateResponse{
|
|
Id: 1,
|
|
})
|
|
tasks = append(tasks, &milvuspb.GetImportStateResponse{
|
|
Id: 50,
|
|
})
|
|
rearrangeTasks(tasks)
|
|
assert.Equal(t, 3, len(tasks))
|
|
assert.Equal(t, int64(1), tasks[0].GetId())
|
|
assert.Equal(t, int64(50), tasks[1].GetId())
|
|
assert.Equal(t, int64(100), tasks[2].GetId())
|
|
}
|