mirror of
https://gitee.com/milvus-io/milvus.git
synced 2024-12-05 13:28:49 +08:00
a434d33e75
This PR introduces novel managerial roles for importv2: 1. ImportMeta: To manage all the import tasks; 2. ImportScheduler: To process tasks and modify their states; 3. ImportChecker: To ascertain the completion of all tasks and instigate relevant operations. issue: https://github.com/milvus-io/milvus/issues/28521 --------- Signed-off-by: bigsheeper <yihao.dai@zilliz.com>
208 lines
6.9 KiB
Go
208 lines
6.9 KiB
Go
// Licensed to the LF AI & Data foundation under one
|
|
// or more contributor license agreementassert. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package datacoord
|
|
|
|
import (
|
|
"testing"
|
|
|
|
"github.com/cockroachdb/errors"
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/mock"
|
|
|
|
"github.com/milvus-io/milvus/internal/metastore/mocks"
|
|
"github.com/milvus-io/milvus/internal/proto/datapb"
|
|
"github.com/milvus-io/milvus/internal/proto/internalpb"
|
|
)
|
|
|
|
func TestImportMeta_Restore(t *testing.T) {
|
|
catalog := mocks.NewDataCoordCatalog(t)
|
|
catalog.EXPECT().ListImportJobs().Return([]*datapb.ImportJob{{JobID: 0}}, nil)
|
|
catalog.EXPECT().ListPreImportTasks().Return([]*datapb.PreImportTask{{TaskID: 1}}, nil)
|
|
catalog.EXPECT().ListImportTasks().Return([]*datapb.ImportTaskV2{{TaskID: 2}}, nil)
|
|
|
|
im, err := NewImportMeta(catalog)
|
|
assert.NoError(t, err)
|
|
|
|
jobs := im.GetJobBy()
|
|
assert.Equal(t, 1, len(jobs))
|
|
assert.Equal(t, int64(0), jobs[0].GetJobID())
|
|
tasks := im.GetTaskBy()
|
|
assert.Equal(t, 2, len(tasks))
|
|
tasks = im.GetTaskBy(WithType(PreImportTaskType))
|
|
assert.Equal(t, 1, len(tasks))
|
|
assert.Equal(t, int64(1), tasks[0].GetTaskID())
|
|
tasks = im.GetTaskBy(WithType(ImportTaskType))
|
|
assert.Equal(t, 1, len(tasks))
|
|
assert.Equal(t, int64(2), tasks[0].GetTaskID())
|
|
|
|
// new meta failed
|
|
mockErr := errors.New("mock error")
|
|
catalog = mocks.NewDataCoordCatalog(t)
|
|
catalog.EXPECT().ListPreImportTasks().Return([]*datapb.PreImportTask{{TaskID: 1}}, mockErr)
|
|
_, err = NewImportMeta(catalog)
|
|
assert.Error(t, err)
|
|
|
|
catalog = mocks.NewDataCoordCatalog(t)
|
|
catalog.EXPECT().ListImportTasks().Return([]*datapb.ImportTaskV2{{TaskID: 2}}, mockErr)
|
|
catalog.EXPECT().ListPreImportTasks().Return([]*datapb.PreImportTask{{TaskID: 1}}, nil)
|
|
_, err = NewImportMeta(catalog)
|
|
assert.Error(t, err)
|
|
|
|
catalog = mocks.NewDataCoordCatalog(t)
|
|
catalog.EXPECT().ListImportJobs().Return([]*datapb.ImportJob{{JobID: 0}}, mockErr)
|
|
catalog.EXPECT().ListPreImportTasks().Return([]*datapb.PreImportTask{{TaskID: 1}}, nil)
|
|
catalog.EXPECT().ListImportTasks().Return([]*datapb.ImportTaskV2{{TaskID: 2}}, nil)
|
|
_, err = NewImportMeta(catalog)
|
|
assert.Error(t, err)
|
|
}
|
|
|
|
func TestImportMeta_ImportJob(t *testing.T) {
|
|
catalog := mocks.NewDataCoordCatalog(t)
|
|
catalog.EXPECT().ListImportJobs().Return(nil, nil)
|
|
catalog.EXPECT().ListPreImportTasks().Return(nil, nil)
|
|
catalog.EXPECT().ListImportTasks().Return(nil, nil)
|
|
catalog.EXPECT().SaveImportJob(mock.Anything).Return(nil)
|
|
catalog.EXPECT().DropImportJob(mock.Anything).Return(nil)
|
|
|
|
im, err := NewImportMeta(catalog)
|
|
assert.NoError(t, err)
|
|
|
|
var job ImportJob = &importJob{
|
|
ImportJob: &datapb.ImportJob{
|
|
JobID: 0,
|
|
CollectionID: 1,
|
|
PartitionIDs: []int64{2},
|
|
Vchannels: []string{"ch0"},
|
|
State: internalpb.ImportJobState_Pending,
|
|
},
|
|
}
|
|
|
|
err = im.AddJob(job)
|
|
assert.NoError(t, err)
|
|
jobs := im.GetJobBy()
|
|
assert.Equal(t, 1, len(jobs))
|
|
err = im.AddJob(job)
|
|
assert.NoError(t, err)
|
|
jobs = im.GetJobBy()
|
|
assert.Equal(t, 1, len(jobs))
|
|
|
|
assert.Nil(t, job.GetSchema())
|
|
err = im.UpdateJob(job.GetJobID(), UpdateJobState(internalpb.ImportJobState_Completed))
|
|
assert.NoError(t, err)
|
|
job2 := im.GetJob(job.GetJobID())
|
|
assert.Equal(t, internalpb.ImportJobState_Completed, job2.GetState())
|
|
assert.Equal(t, job.GetJobID(), job2.GetJobID())
|
|
assert.Equal(t, job.GetCollectionID(), job2.GetCollectionID())
|
|
assert.Equal(t, job.GetPartitionIDs(), job2.GetPartitionIDs())
|
|
assert.Equal(t, job.GetVchannels(), job2.GetVchannels())
|
|
|
|
err = im.RemoveJob(job.GetJobID())
|
|
assert.NoError(t, err)
|
|
jobs = im.GetJobBy()
|
|
assert.Equal(t, 0, len(jobs))
|
|
|
|
// test failed
|
|
mockErr := errors.New("mock err")
|
|
catalog = mocks.NewDataCoordCatalog(t)
|
|
catalog.EXPECT().SaveImportJob(mock.Anything).Return(mockErr)
|
|
catalog.EXPECT().DropImportJob(mock.Anything).Return(mockErr)
|
|
im.(*importMeta).catalog = catalog
|
|
|
|
err = im.AddJob(job)
|
|
assert.Error(t, err)
|
|
im.(*importMeta).jobs[job.GetJobID()] = job
|
|
err = im.UpdateJob(job.GetJobID())
|
|
assert.Error(t, err)
|
|
err = im.RemoveJob(job.GetJobID())
|
|
assert.Error(t, err)
|
|
}
|
|
|
|
func TestImportMeta_ImportTask(t *testing.T) {
|
|
catalog := mocks.NewDataCoordCatalog(t)
|
|
catalog.EXPECT().ListImportJobs().Return(nil, nil)
|
|
catalog.EXPECT().ListPreImportTasks().Return(nil, nil)
|
|
catalog.EXPECT().ListImportTasks().Return(nil, nil)
|
|
catalog.EXPECT().SaveImportTask(mock.Anything).Return(nil)
|
|
catalog.EXPECT().DropImportTask(mock.Anything).Return(nil)
|
|
|
|
im, err := NewImportMeta(catalog)
|
|
assert.NoError(t, err)
|
|
|
|
task1 := &importTask{
|
|
ImportTaskV2: &datapb.ImportTaskV2{
|
|
JobID: 1,
|
|
TaskID: 2,
|
|
CollectionID: 3,
|
|
SegmentIDs: []int64{5, 6},
|
|
NodeID: 7,
|
|
State: datapb.ImportTaskStateV2_Pending,
|
|
},
|
|
}
|
|
err = im.AddTask(task1)
|
|
assert.NoError(t, err)
|
|
err = im.AddTask(task1)
|
|
assert.NoError(t, err)
|
|
res := im.GetTask(task1.GetTaskID())
|
|
assert.Equal(t, task1, res)
|
|
|
|
task2 := task1.Clone()
|
|
task2.(*importTask).TaskID = 8
|
|
task2.(*importTask).State = datapb.ImportTaskStateV2_Completed
|
|
err = im.AddTask(task2)
|
|
assert.NoError(t, err)
|
|
|
|
tasks := im.GetTaskBy(WithJob(task1.GetJobID()))
|
|
assert.Equal(t, 2, len(tasks))
|
|
tasks = im.GetTaskBy(WithType(ImportTaskType), WithStates(datapb.ImportTaskStateV2_Completed))
|
|
assert.Equal(t, 1, len(tasks))
|
|
assert.Equal(t, task2.GetTaskID(), tasks[0].GetTaskID())
|
|
|
|
err = im.UpdateTask(task1.GetTaskID(), UpdateNodeID(9),
|
|
UpdateState(datapb.ImportTaskStateV2_Failed),
|
|
UpdateFileStats([]*datapb.ImportFileStats{1: {
|
|
FileSize: 100,
|
|
}}))
|
|
assert.NoError(t, err)
|
|
task := im.GetTask(task1.GetTaskID())
|
|
assert.Equal(t, int64(9), task.GetNodeID())
|
|
assert.Equal(t, datapb.ImportTaskStateV2_Failed, task.GetState())
|
|
|
|
err = im.RemoveTask(task1.GetTaskID())
|
|
assert.NoError(t, err)
|
|
tasks = im.GetTaskBy()
|
|
assert.Equal(t, 1, len(tasks))
|
|
err = im.RemoveTask(10)
|
|
assert.NoError(t, err)
|
|
tasks = im.GetTaskBy()
|
|
assert.Equal(t, 1, len(tasks))
|
|
|
|
// test failed
|
|
mockErr := errors.New("mock err")
|
|
catalog = mocks.NewDataCoordCatalog(t)
|
|
catalog.EXPECT().SaveImportTask(mock.Anything).Return(mockErr)
|
|
catalog.EXPECT().DropImportTask(mock.Anything).Return(mockErr)
|
|
im.(*importMeta).catalog = catalog
|
|
|
|
err = im.AddTask(task1)
|
|
assert.Error(t, err)
|
|
im.(*importMeta).tasks[task1.GetTaskID()] = task1
|
|
err = im.UpdateTask(task1.GetTaskID(), UpdateNodeID(9))
|
|
assert.Error(t, err)
|
|
err = im.RemoveTask(task1.GetTaskID())
|
|
assert.Error(t, err)
|
|
}
|