mirror of
https://gitee.com/milvus-io/milvus.git
synced 2024-12-02 03:48:37 +08:00
b0bd290a6e
Some checks are pending
Code Checker / Code Checker AMD64 Ubuntu 22.04 (push) Waiting to run
Code Checker / Code Checker Amazonlinux 2023 (push) Waiting to run
Code Checker / Code Checker rockylinux8 (push) Waiting to run
Mac Code Checker / Code Checker MacOS 12 (push) Waiting to run
Build and test / Build and test AMD64 Ubuntu 22.04 (push) Waiting to run
Build and test / UT for Cpp (push) Blocked by required conditions
Build and test / UT for Go (push) Blocked by required conditions
Build and test / Integration Test (push) Blocked by required conditions
Build and test / Upload Code Coverage (push) Blocked by required conditions
Related to #35020 Signed-off-by: Congqi Xia <congqi.xia@zilliz.com>
131 lines
4.0 KiB
Go
131 lines
4.0 KiB
Go
// Licensed to the LF AI & Data foundation under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package datacoord
|
|
|
|
import (
|
|
"context"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/stretchr/testify/mock"
|
|
"github.com/stretchr/testify/suite"
|
|
|
|
"github.com/milvus-io/milvus/internal/json"
|
|
"github.com/milvus-io/milvus/internal/metastore/mocks"
|
|
"github.com/milvus-io/milvus/internal/proto/datapb"
|
|
"github.com/milvus-io/milvus/pkg/util/metricsinfo"
|
|
)
|
|
|
|
func TestCompactionTaskMetaSuite(t *testing.T) {
|
|
suite.Run(t, new(CompactionTaskMetaSuite))
|
|
}
|
|
|
|
type CompactionTaskMetaSuite struct {
|
|
suite.Suite
|
|
catalog *mocks.DataCoordCatalog
|
|
meta *compactionTaskMeta
|
|
}
|
|
|
|
func (suite *CompactionTaskMetaSuite) SetupTest() {
|
|
catalog := mocks.NewDataCoordCatalog(suite.T())
|
|
catalog.EXPECT().ListCompactionTask(mock.Anything).Return(nil, nil)
|
|
catalog.EXPECT().SaveCompactionTask(mock.Anything, mock.Anything).Return(nil)
|
|
suite.catalog = catalog
|
|
meta, err := newCompactionTaskMeta(context.TODO(), catalog)
|
|
suite.NoError(err)
|
|
suite.meta = meta
|
|
}
|
|
|
|
func newTestCompactionTaskMeta(t *testing.T) *compactionTaskMeta {
|
|
catalog := mocks.NewDataCoordCatalog(t)
|
|
catalog.EXPECT().ListCompactionTask(mock.Anything).Return(nil, nil).Maybe()
|
|
catalog.EXPECT().SaveCompactionTask(mock.Anything, mock.Anything).Return(nil).Maybe()
|
|
meta, _ := newCompactionTaskMeta(context.TODO(), catalog)
|
|
return meta
|
|
}
|
|
|
|
func (suite *CompactionTaskMetaSuite) TestGetCompactionTasksByCollection() {
|
|
suite.meta.SaveCompactionTask(&datapb.CompactionTask{
|
|
TriggerID: 1,
|
|
PlanID: 10,
|
|
CollectionID: 100,
|
|
})
|
|
res := suite.meta.GetCompactionTasksByCollection(100)
|
|
suite.Equal(1, len(res))
|
|
}
|
|
|
|
func (suite *CompactionTaskMetaSuite) TestGetCompactionTasksByCollectionAbnormal() {
|
|
suite.meta.SaveCompactionTask(&datapb.CompactionTask{
|
|
TriggerID: 1,
|
|
PlanID: 10,
|
|
CollectionID: 100,
|
|
})
|
|
suite.meta.SaveCompactionTask(&datapb.CompactionTask{
|
|
TriggerID: 2,
|
|
PlanID: 11,
|
|
CollectionID: 101,
|
|
})
|
|
res := suite.meta.GetCompactionTasksByCollection(101)
|
|
suite.Equal(1, len(res))
|
|
}
|
|
|
|
func (suite *CompactionTaskMetaSuite) TestTaskStatsJSON() {
|
|
task1 := &datapb.CompactionTask{
|
|
PlanID: 1,
|
|
CollectionID: 100,
|
|
Type: datapb.CompactionType_MergeCompaction,
|
|
State: datapb.CompactionTaskState_completed,
|
|
FailReason: "",
|
|
StartTime: time.Now().Unix(),
|
|
EndTime: time.Now().Add(time.Hour).Unix(),
|
|
TotalRows: 1000,
|
|
InputSegments: []int64{1, 2},
|
|
ResultSegments: []int64{3},
|
|
}
|
|
task2 := &datapb.CompactionTask{
|
|
PlanID: 2,
|
|
CollectionID: 101,
|
|
Type: datapb.CompactionType_MergeCompaction,
|
|
State: datapb.CompactionTaskState_completed,
|
|
FailReason: "",
|
|
StartTime: time.Now().Unix(),
|
|
EndTime: time.Now().Add(time.Hour).Unix(),
|
|
TotalRows: 2000,
|
|
InputSegments: []int64{4, 5},
|
|
ResultSegments: []int64{6},
|
|
}
|
|
|
|
// testing return empty string
|
|
actualJSON := suite.meta.TaskStatsJSON()
|
|
suite.Equal("[]", actualJSON)
|
|
|
|
err := suite.meta.SaveCompactionTask(task1)
|
|
suite.NoError(err)
|
|
err = suite.meta.SaveCompactionTask(task2)
|
|
suite.NoError(err)
|
|
|
|
expectedTasks := []*metricsinfo.CompactionTask{
|
|
newCompactionTaskStats(task1),
|
|
newCompactionTaskStats(task2),
|
|
}
|
|
expectedJSON, err := json.Marshal(expectedTasks)
|
|
suite.NoError(err)
|
|
|
|
actualJSON = suite.meta.TaskStatsJSON()
|
|
suite.JSONEq(string(expectedJSON), actualJSON)
|
|
}
|