// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use this file except in compliance // with the License. You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package datanode import ( "context" "path" "testing" "github.com/cockroachdb/errors" "github.com/samber/lo" "github.com/stretchr/testify/mock" "github.com/stretchr/testify/suite" "go.uber.org/zap" "github.com/milvus-io/milvus-proto/go-api/v2/schemapb" "github.com/milvus-io/milvus/internal/datanode/allocator" "github.com/milvus-io/milvus/internal/datanode/io" iter "github.com/milvus-io/milvus/internal/datanode/iterators" "github.com/milvus-io/milvus/internal/datanode/metacache" "github.com/milvus-io/milvus/internal/mocks" "github.com/milvus-io/milvus/internal/proto/datapb" "github.com/milvus-io/milvus/internal/storage" "github.com/milvus-io/milvus/pkg/common" "github.com/milvus-io/milvus/pkg/log" "github.com/milvus-io/milvus/pkg/util/metautil" "github.com/milvus-io/milvus/pkg/util/paramtable" "github.com/milvus-io/milvus/pkg/util/timerecord" ) func TestLevelZeroCompactionTaskSuite(t *testing.T) { suite.Run(t, new(LevelZeroCompactionTaskSuite)) } type LevelZeroCompactionTaskSuite struct { suite.Suite mockBinlogIO *io.MockBinlogIO mockAlloc *allocator.MockAllocator task *levelZeroCompactionTask dData *storage.DeleteData dBlob []byte } func (s *LevelZeroCompactionTaskSuite) SetupTest() { s.mockAlloc = allocator.NewMockAllocator(s.T()) s.mockBinlogIO = io.NewMockBinlogIO(s.T()) // plan of the task is unset s.task = newLevelZeroCompactionTask(context.Background(), s.mockBinlogIO, s.mockAlloc, nil, nil) pk2ts := map[int64]uint64{ 1: 20000, 2: 20001, 3: 20002, } s.dData = storage.NewDeleteData([]storage.PrimaryKey{}, []Timestamp{}) for pk, ts := range pk2ts { s.dData.Append(storage.NewInt64PrimaryKey(pk), ts) } dataCodec := storage.NewDeleteCodec() blob, err := dataCodec.Serialize(0, 0, 0, s.dData) s.Require().NoError(err) s.dBlob = blob.GetValue() } func (s *LevelZeroCompactionTaskSuite) TestLinearBatchLoadDeltaFail() { plan := &datapb.CompactionPlan{ PlanID: 19530, Type: datapb.CompactionType_Level0DeleteCompaction, SegmentBinlogs: []*datapb.CompactionSegmentBinlogs{ { SegmentID: 100, Level: datapb.SegmentLevel_L0, Deltalogs: []*datapb.FieldBinlog{ { Binlogs: []*datapb.Binlog{ {LogPath: "a/b/c1", LogSize: 100}, {LogPath: "a/b/c2", LogSize: 100}, {LogPath: "a/b/c3", LogSize: 100}, {LogPath: "a/b/c4", LogSize: 100}, }, }, }, }, {SegmentID: 200, Level: datapb.SegmentLevel_L1}, }, Schema: &schemapb.CollectionSchema{ Fields: []*schemapb.FieldSchema{ { IsPrimaryKey: true, }, }, }, } s.task.plan = plan s.task.tr = timerecord.NewTimeRecorder("test") s.mockBinlogIO.EXPECT().Download(mock.Anything, mock.Anything).Return(nil, errors.New("mock download fail")).Twice() targetSegments := lo.Filter(plan.SegmentBinlogs, func(s *datapb.CompactionSegmentBinlogs, _ int) bool { return s.Level == datapb.SegmentLevel_L1 }) deltaLogs := map[int64][]string{100: {"a/b/c1"}} segments, err := s.task.linearProcess(context.Background(), targetSegments, deltaLogs) s.Error(err) s.Empty(segments) segments, err = s.task.batchProcess(context.Background(), targetSegments, lo.Values(deltaLogs)...) s.Error(err) s.Empty(segments) } func (s *LevelZeroCompactionTaskSuite) TestLinearBatchUploadByCheckFail() { plan := &datapb.CompactionPlan{ PlanID: 19530, Type: datapb.CompactionType_Level0DeleteCompaction, SegmentBinlogs: []*datapb.CompactionSegmentBinlogs{ { SegmentID: 100, Level: datapb.SegmentLevel_L0, Deltalogs: []*datapb.FieldBinlog{ { Binlogs: []*datapb.Binlog{ {LogPath: "a/b/c1", LogSize: 100}, {LogPath: "a/b/c2", LogSize: 100}, {LogPath: "a/b/c3", LogSize: 100}, {LogPath: "a/b/c4", LogSize: 100}, }, }, }, }, {SegmentID: 200, Level: datapb.SegmentLevel_L1, Field2StatslogPaths: []*datapb.FieldBinlog{ { Binlogs: []*datapb.Binlog{ {LogID: 9999, LogSize: 100}, }, }, }}, }, Schema: &schemapb.CollectionSchema{ Fields: []*schemapb.FieldSchema{ { IsPrimaryKey: true, }, }, }, } s.task.plan = plan s.task.tr = timerecord.NewTimeRecorder("test") data := &storage.Int64FieldData{ Data: []int64{1, 2, 3, 4, 5, 6, 7, 8, 9}, } sw := &storage.StatsWriter{} err := sw.GenerateByData(common.RowIDField, schemapb.DataType_Int64, data) s.NoError(err) cm := mocks.NewChunkManager(s.T()) cm.EXPECT().MultiRead(mock.Anything, mock.Anything).Return([][]byte{sw.GetBuffer()}, nil) s.task.cm = cm s.mockBinlogIO.EXPECT().Download(mock.Anything, mock.Anything).Return([][]byte{s.dBlob}, nil).Times(2) mockAlloc := allocator.NewMockAllocator(s.T()) mockAlloc.EXPECT().AllocOne().Return(0, errors.New("mock alloc err")) s.task.allocator = mockAlloc targetSegments := lo.Filter(plan.SegmentBinlogs, func(s *datapb.CompactionSegmentBinlogs, _ int) bool { return s.Level == datapb.SegmentLevel_L1 }) deltaLogs := map[int64][]string{100: {"a/b/c1"}} segments, err := s.task.linearProcess(context.Background(), targetSegments, deltaLogs) s.Error(err) s.Empty(segments) segments, err = s.task.batchProcess(context.Background(), targetSegments, lo.Values(deltaLogs)...) s.Error(err) s.Empty(segments) } func (s *LevelZeroCompactionTaskSuite) TestCompactLinear() { plan := &datapb.CompactionPlan{ PlanID: 19530, Type: datapb.CompactionType_Level0DeleteCompaction, SegmentBinlogs: []*datapb.CompactionSegmentBinlogs{ { CollectionID: 1, SegmentID: 100, Level: datapb.SegmentLevel_L0, Deltalogs: []*datapb.FieldBinlog{ { Binlogs: []*datapb.Binlog{ {LogPath: "a/b/c1", LogSize: 100}, {LogPath: "a/b/c2", LogSize: 100}, {LogPath: "a/b/c3", LogSize: 100}, {LogPath: "a/b/c4", LogSize: 100}, }, }, }, }, { CollectionID: 1, SegmentID: 101, Level: datapb.SegmentLevel_L0, Deltalogs: []*datapb.FieldBinlog{ { Binlogs: []*datapb.Binlog{ {LogPath: "a/d/c1", LogSize: 100}, {LogPath: "a/d/c2", LogSize: 100}, {LogPath: "a/d/c3", LogSize: 100}, {LogPath: "a/d/c4", LogSize: 100}, }, }, }, }, { CollectionID: 1, SegmentID: 200, Level: datapb.SegmentLevel_L1, Field2StatslogPaths: []*datapb.FieldBinlog{ { Binlogs: []*datapb.Binlog{ {LogID: 9999, LogSize: 100}, }, }, }, }, { CollectionID: 1, SegmentID: 201, Level: datapb.SegmentLevel_L1, Field2StatslogPaths: []*datapb.FieldBinlog{ { Binlogs: []*datapb.Binlog{ {LogID: 9999, LogSize: 100}, }, }, }, }, }, Schema: &schemapb.CollectionSchema{ Fields: []*schemapb.FieldSchema{ { IsPrimaryKey: true, }, }, }, } s.task.plan = plan s.task.tr = timerecord.NewTimeRecorder("test") data := &storage.Int64FieldData{ Data: []int64{1, 2, 3, 4, 5, 6, 7, 8, 9}, } sw := &storage.StatsWriter{} err := sw.GenerateByData(common.RowIDField, schemapb.DataType_Int64, data) s.NoError(err) cm := mocks.NewChunkManager(s.T()) cm.EXPECT().MultiRead(mock.Anything, mock.Anything).Return([][]byte{sw.GetBuffer()}, nil) s.task.cm = cm s.mockBinlogIO.EXPECT().Download(mock.Anything, mock.Anything).Return([][]byte{s.dBlob}, nil).Times(2) s.mockAlloc.EXPECT().AllocOne().Return(19530, nil).Times(2) s.mockBinlogIO.EXPECT().JoinFullPath(mock.Anything, mock.Anything). RunAndReturn(func(paths ...string) string { return path.Join(paths...) }).Times(2) s.mockBinlogIO.EXPECT().Upload(mock.Anything, mock.Anything).Return(nil).Once() s.Require().Equal(plan.GetPlanID(), s.task.GetPlanID()) s.Require().Equal(plan.GetChannel(), s.task.GetChannelName()) s.Require().EqualValues(1, s.task.GetCollection()) l0Segments := lo.Filter(s.task.plan.GetSegmentBinlogs(), func(s *datapb.CompactionSegmentBinlogs, _ int) bool { return s.Level == datapb.SegmentLevel_L0 }) targetSegments := lo.Filter(s.task.plan.GetSegmentBinlogs(), func(s *datapb.CompactionSegmentBinlogs, _ int) bool { return s.Level == datapb.SegmentLevel_L1 }) totalDeltalogs := make(map[UniqueID][]string) for _, s := range l0Segments { paths := []string{} for _, d := range s.GetDeltalogs() { for _, l := range d.GetBinlogs() { paths = append(paths, l.GetLogPath()) } } if len(paths) > 0 { totalDeltalogs[s.GetSegmentID()] = paths } } segments, err := s.task.linearProcess(context.Background(), targetSegments, totalDeltalogs) s.NoError(err) s.NotEmpty(segments) s.Equal(2, len(segments)) s.ElementsMatch([]int64{200, 201}, lo.Map(segments, func(seg *datapb.CompactionSegment, _ int) int64 { return seg.GetSegmentID() })) for _, segment := range segments { s.NotNil(segment.GetDeltalogs()) } log.Info("test segment results", zap.Any("result", segments)) } func (s *LevelZeroCompactionTaskSuite) TestCompactBatch() { plan := &datapb.CompactionPlan{ PlanID: 19530, Type: datapb.CompactionType_Level0DeleteCompaction, SegmentBinlogs: []*datapb.CompactionSegmentBinlogs{ { SegmentID: 100, Level: datapb.SegmentLevel_L0, Deltalogs: []*datapb.FieldBinlog{ { Binlogs: []*datapb.Binlog{ {LogPath: "a/b/c1", LogSize: 100}, {LogPath: "a/b/c2", LogSize: 100}, {LogPath: "a/b/c3", LogSize: 100}, {LogPath: "a/b/c4", LogSize: 100}, }, }, }, }, { SegmentID: 101, Level: datapb.SegmentLevel_L0, Deltalogs: []*datapb.FieldBinlog{ { Binlogs: []*datapb.Binlog{ {LogPath: "a/d/c1", LogSize: 100}, {LogPath: "a/d/c2", LogSize: 100}, {LogPath: "a/d/c3", LogSize: 100}, {LogPath: "a/d/c4", LogSize: 100}, }, }, }, }, {SegmentID: 200, Level: datapb.SegmentLevel_L1, Field2StatslogPaths: []*datapb.FieldBinlog{ { Binlogs: []*datapb.Binlog{ {LogID: 9999, LogSize: 100}, }, }, }}, {SegmentID: 201, Level: datapb.SegmentLevel_L1, Field2StatslogPaths: []*datapb.FieldBinlog{ { Binlogs: []*datapb.Binlog{ {LogID: 9999, LogSize: 100}, }, }, }}, }, Schema: &schemapb.CollectionSchema{ Fields: []*schemapb.FieldSchema{ { IsPrimaryKey: true, }, }, }, } s.task.plan = plan s.task.tr = timerecord.NewTimeRecorder("test") data := &storage.Int64FieldData{ Data: []int64{1, 2, 3, 4, 5, 6, 7, 8, 9}, } sw := &storage.StatsWriter{} err := sw.GenerateByData(common.RowIDField, schemapb.DataType_Int64, data) s.NoError(err) cm := mocks.NewChunkManager(s.T()) cm.EXPECT().MultiRead(mock.Anything, mock.Anything).Return([][]byte{sw.GetBuffer()}, nil) s.task.cm = cm s.mockBinlogIO.EXPECT().Download(mock.Anything, mock.Anything).Return([][]byte{s.dBlob}, nil).Once() s.mockAlloc.EXPECT().AllocOne().Return(19530, nil).Times(2) s.mockBinlogIO.EXPECT().JoinFullPath(mock.Anything, mock.Anything). RunAndReturn(func(paths ...string) string { return path.Join(paths...) }).Times(2) s.mockBinlogIO.EXPECT().Upload(mock.Anything, mock.Anything).Return(nil).Once() l0Segments := lo.Filter(s.task.plan.GetSegmentBinlogs(), func(s *datapb.CompactionSegmentBinlogs, _ int) bool { return s.Level == datapb.SegmentLevel_L0 }) targetSegments := lo.Filter(s.task.plan.GetSegmentBinlogs(), func(s *datapb.CompactionSegmentBinlogs, _ int) bool { return s.Level == datapb.SegmentLevel_L1 }) totalDeltalogs := make(map[UniqueID][]string) for _, s := range l0Segments { paths := []string{} for _, d := range s.GetDeltalogs() { for _, l := range d.GetBinlogs() { paths = append(paths, l.GetLogPath()) } } if len(paths) > 0 { totalDeltalogs[s.GetSegmentID()] = paths } } segments, err := s.task.batchProcess(context.TODO(), targetSegments, lo.Values(totalDeltalogs)...) s.NoError(err) s.NotEmpty(segments) s.Equal(2, len(segments)) s.ElementsMatch([]int64{200, 201}, lo.Map(segments, func(seg *datapb.CompactionSegment, _ int) int64 { return seg.GetSegmentID() })) for _, segment := range segments { s.NotNil(segment.GetDeltalogs()) } log.Info("test segment results", zap.Any("result", segments)) } func (s *LevelZeroCompactionTaskSuite) TestUploadByCheck() { ctx := context.Background() plan := &datapb.CompactionPlan{ SegmentBinlogs: []*datapb.CompactionSegmentBinlogs{ { SegmentID: 100, }, }, } s.Run("uploadByCheck directly composeDeltalog failed", func() { s.SetupTest() s.task.plan = plan mockAlloc := allocator.NewMockAllocator(s.T()) mockAlloc.EXPECT().AllocOne().Return(0, errors.New("mock alloc err")) s.task.allocator = mockAlloc segments := map[int64]*storage.DeleteData{100: s.dData} results := make(map[int64]*datapb.CompactionSegment) err := s.task.uploadByCheck(ctx, false, segments, results) s.Error(err) s.Equal(0, len(results)) }) s.Run("uploadByCheck directly Upload failed", func() { s.SetupTest() s.task.plan = plan s.mockBinlogIO.EXPECT().Upload(mock.Anything, mock.Anything).Return(errors.New("mock upload failed")) s.mockAlloc.EXPECT().AllocOne().Return(19530, nil) blobKey := metautil.JoinIDPath(1, 10, 100, 19530) blobPath := path.Join(common.SegmentDeltaLogPath, blobKey) s.mockBinlogIO.EXPECT().JoinFullPath(mock.Anything, mock.Anything).Return(blobPath) segments := map[int64]*storage.DeleteData{100: s.dData} results := make(map[int64]*datapb.CompactionSegment) err := s.task.uploadByCheck(ctx, false, segments, results) s.Error(err) s.Equal(0, len(results)) }) s.Run("upload directly", func() { s.SetupTest() s.task.plan = plan s.mockBinlogIO.EXPECT().Upload(mock.Anything, mock.Anything).Return(nil) s.mockAlloc.EXPECT().AllocOne().Return(19530, nil) blobKey := metautil.JoinIDPath(1, 10, 100, 19530) blobPath := path.Join(common.SegmentDeltaLogPath, blobKey) s.mockBinlogIO.EXPECT().JoinFullPath(mock.Anything, mock.Anything).Return(blobPath) segments := map[int64]*storage.DeleteData{100: s.dData} results := make(map[int64]*datapb.CompactionSegment) err := s.task.uploadByCheck(ctx, false, segments, results) s.NoError(err) s.Equal(1, len(results)) seg1, ok := results[100] s.True(ok) s.EqualValues(100, seg1.GetSegmentID()) s.Equal(1, len(seg1.GetDeltalogs())) s.Equal(1, len(seg1.GetDeltalogs()[0].GetBinlogs())) }) s.Run("check without upload", func() { s.SetupTest() segments := map[int64]*storage.DeleteData{100: s.dData} results := make(map[int64]*datapb.CompactionSegment) s.Require().Empty(results) err := s.task.uploadByCheck(ctx, true, segments, results) s.NoError(err) s.Empty(results) }) s.Run("check with upload", func() { s.task.plan = plan blobKey := metautil.JoinIDPath(1, 10, 100, 19530) blobPath := path.Join(common.SegmentDeltaLogPath, blobKey) s.mockBinlogIO.EXPECT().Upload(mock.Anything, mock.Anything).Return(nil) s.mockAlloc.EXPECT().AllocOne().Return(19530, nil) s.mockBinlogIO.EXPECT().JoinFullPath(mock.Anything, mock.Anything).Return(blobPath) segments := map[int64]*storage.DeleteData{100: s.dData} results := map[int64]*datapb.CompactionSegment{ 100: {SegmentID: 100, Deltalogs: []*datapb.FieldBinlog{{Binlogs: []*datapb.Binlog{{LogID: 1}}}}}, } s.Require().Equal(1, len(results)) paramtable.Get().Save(paramtable.Get().DataNodeCfg.FlushDeleteBufferBytes.Key, "1") defer paramtable.Get().Reset(paramtable.Get().DataNodeCfg.FlushDeleteBufferBytes.Key) err := s.task.uploadByCheck(ctx, true, segments, results) s.NoError(err) s.NotEmpty(results) s.Equal(1, len(results)) seg1, ok := results[100] s.True(ok) s.EqualValues(100, seg1.GetSegmentID()) s.Equal(1, len(seg1.GetDeltalogs())) s.Equal(2, len(seg1.GetDeltalogs()[0].GetBinlogs())) }) } func (s *LevelZeroCompactionTaskSuite) TestComposeDeltalog() { plan := &datapb.CompactionPlan{ SegmentBinlogs: []*datapb.CompactionSegmentBinlogs{ { SegmentID: 100, }, { SegmentID: 101, }, }, } s.task.plan = plan s.mockAlloc.EXPECT().AllocOne().Return(19530, nil) blobKey := metautil.JoinIDPath(1, 10, 100, 19530) blobPath := path.Join(common.SegmentDeltaLogPath, blobKey) s.mockBinlogIO.EXPECT().JoinFullPath(mock.Anything, mock.Anything).Return(blobPath) kvs, binlog, err := s.task.composeDeltalog(100, s.dData) s.NoError(err) s.Equal(1, len(kvs)) v, ok := kvs[blobPath] s.True(ok) s.NotNil(v) s.Equal(blobPath, binlog.LogPath) kvs, _, err = s.task.composeDeltalog(101, s.dData) s.NoError(err) s.Equal(1, len(kvs)) v, ok = kvs[blobPath] s.True(ok) s.NotNil(v) s.Equal(blobPath, binlog.LogPath) } func (s *LevelZeroCompactionTaskSuite) TestSplitDelta() { bfs1 := metacache.NewBloomFilterSetWithBatchSize(100) bfs1.UpdatePKRange(&storage.Int64FieldData{Data: []int64{1, 3}}) bfs2 := metacache.NewBloomFilterSetWithBatchSize(100) bfs2.UpdatePKRange(&storage.Int64FieldData{Data: []int64{3}}) bfs3 := metacache.NewBloomFilterSetWithBatchSize(100) bfs3.UpdatePKRange(&storage.Int64FieldData{Data: []int64{3}}) predicted := []int64{100, 101, 102} diter := iter.NewDeltalogIterator([][]byte{s.dBlob}, nil) s.Require().NotNil(diter) targetSegBuffer := make(map[int64]*storage.DeleteData) segmentBFs := map[int64]*metacache.BloomFilterSet{ 100: bfs1, 101: bfs2, 102: bfs3, } s.task.splitDelta(context.TODO(), []*iter.DeltalogIterator{diter}, targetSegBuffer, segmentBFs) s.NotEmpty(targetSegBuffer) s.ElementsMatch(predicted, lo.Keys(targetSegBuffer)) s.EqualValues(2, targetSegBuffer[100].RowCount) s.EqualValues(1, targetSegBuffer[101].RowCount) s.EqualValues(1, targetSegBuffer[102].RowCount) s.ElementsMatch([]storage.PrimaryKey{storage.NewInt64PrimaryKey(1), storage.NewInt64PrimaryKey(3)}, targetSegBuffer[100].Pks) s.Equal(storage.NewInt64PrimaryKey(3), targetSegBuffer[101].Pks[0]) s.Equal(storage.NewInt64PrimaryKey(3), targetSegBuffer[102].Pks[0]) } func (s *LevelZeroCompactionTaskSuite) TestLoadDelta() { ctx := context.TODO() s.mockBinlogIO.EXPECT().Download(mock.Anything, mock.MatchedBy( func(paths []string) bool { return len(paths) > 0 && paths[0] == "correct" })).Return([][]byte{s.dBlob}, nil).Once() s.mockBinlogIO.EXPECT().Download(mock.Anything, mock.MatchedBy( func(paths []string) bool { return len(paths) > 0 && paths[0] == "error" })).Return(nil, errors.New("mock err")).Once() s.mockBinlogIO.EXPECT().Download(mock.Anything, mock.MatchedBy( func(paths []string) bool { return len(paths) > 0 && paths[0] == "invalid-blobs" })).Return([][]byte{{1}}, nil).Once() tests := []struct { description string paths []string expectNilIter bool expectError bool }{ {"no error", []string{"correct"}, false, false}, {"download error", []string{"error"}, true, true}, {"new iter error", []string{"invalid-blobs"}, true, false}, } for _, test := range tests { iters, err := s.task.loadDelta(ctx, test.paths) if test.expectNilIter { if len(iters) > 0 { for _, iter := range iters { s.False(iter.HasNext()) } } else { s.Nil(iters) } } else { s.NotNil(iters) s.Equal(1, len(iters)) s.True(iters[0].HasNext()) iter := iters[0] var pks []storage.PrimaryKey var tss []storage.Timestamp for iter.HasNext() { labeled, err := iter.Next() s.NoError(err) pks = append(pks, labeled.GetPk()) tss = append(tss, labeled.GetTimestamp()) } s.ElementsMatch(pks, s.dData.Pks) s.ElementsMatch(tss, s.dData.Tss) } if test.expectError { s.Error(err) } else { s.NoError(err) } } } func (s *LevelZeroCompactionTaskSuite) TestLoadBF() { plan := &datapb.CompactionPlan{ PlanID: 19530, Type: datapb.CompactionType_Level0DeleteCompaction, SegmentBinlogs: []*datapb.CompactionSegmentBinlogs{ {SegmentID: 201, Level: datapb.SegmentLevel_L1, Field2StatslogPaths: []*datapb.FieldBinlog{ { Binlogs: []*datapb.Binlog{ {LogID: 9999, LogSize: 100}, }, }, }}, }, Schema: &schemapb.CollectionSchema{ Fields: []*schemapb.FieldSchema{ { IsPrimaryKey: true, }, }, }, } s.task.plan = plan data := &storage.Int64FieldData{ Data: []int64{1, 2, 3, 4, 5, 6, 7, 8, 9}, } sw := &storage.StatsWriter{} err := sw.GenerateByData(common.RowIDField, schemapb.DataType_Int64, data) s.NoError(err) cm := mocks.NewChunkManager(s.T()) cm.EXPECT().MultiRead(mock.Anything, mock.Anything).Return([][]byte{sw.GetBuffer()}, nil) s.task.cm = cm bfs, err := s.task.loadBF(context.Background(), plan.SegmentBinlogs) s.NoError(err) s.Len(bfs, 1) for _, pk := range s.dData.Pks { lc := storage.NewLocationsCache(pk) s.True(bfs[201].PkExists(lc)) } } func (s *LevelZeroCompactionTaskSuite) TestFailed() { s.Run("no primary key", func() { plan := &datapb.CompactionPlan{ PlanID: 19530, Type: datapb.CompactionType_Level0DeleteCompaction, SegmentBinlogs: []*datapb.CompactionSegmentBinlogs{ {SegmentID: 201, Level: datapb.SegmentLevel_L1, Field2StatslogPaths: []*datapb.FieldBinlog{ { Binlogs: []*datapb.Binlog{ {LogID: 9999, LogSize: 100}, }, }, }}, }, Schema: &schemapb.CollectionSchema{ Fields: []*schemapb.FieldSchema{ { IsPrimaryKey: false, }, }, }, } s.task.plan = plan _, err := s.task.loadBF(context.Background(), plan.SegmentBinlogs) s.Error(err) }) s.Run("no l1 segments", func() { plan := &datapb.CompactionPlan{ PlanID: 19530, Type: datapb.CompactionType_Level0DeleteCompaction, SegmentBinlogs: []*datapb.CompactionSegmentBinlogs{ {SegmentID: 201, Level: datapb.SegmentLevel_L0}, }, } s.task.plan = plan _, err := s.task.Compact() s.Error(err) }) }