// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use this file except in compliance // with the License. You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. package datanode import ( "context" "errors" "path" "testing" "time" "github.com/milvus-io/milvus/internal/common" "github.com/milvus-io/milvus/internal/log" "github.com/milvus-io/milvus/internal/proto/schemapb" "github.com/milvus-io/milvus/internal/storage" "github.com/milvus-io/milvus/internal/util/typeutil" "github.com/stretchr/testify/assert" "github.com/stretchr/testify/require" "go.uber.org/zap" ) var binlogTestDir = "/tmp/milvus_test/test_binlog_io" func TestBinlogIOInterfaceMethods(t *testing.T) { alloc := NewAllocatorFactory() cm := storage.NewLocalChunkManager(storage.RootPath(binlogTestDir)) defer cm.RemoveWithPrefix("") b := &binlogIO{cm, alloc} t.Run("Test upload", func(t *testing.T) { f := &MetaFactory{} meta := f.GetCollectionMeta(UniqueID(10001), "uploads", schemapb.DataType_Int64) iData := genInsertData() pk := newInt64PrimaryKey(888) dData := &DeleteData{ RowCount: 1, Pks: []primaryKey{pk}, Tss: []uint64{666666}, } p, err := b.upload(context.TODO(), 1, 10, []*InsertData{iData}, dData, meta) assert.NoError(t, err) assert.Equal(t, 12, len(p.inPaths)) assert.Equal(t, 1, len(p.statsPaths)) assert.Equal(t, 1, len(p.inPaths[0].GetBinlogs())) assert.Equal(t, 1, len(p.statsPaths[0].GetBinlogs())) assert.NotNil(t, p.deltaInfo) p, err = b.upload(context.TODO(), 1, 10, []*InsertData{iData, iData}, dData, meta) assert.NoError(t, err) assert.Equal(t, 12, len(p.inPaths)) assert.Equal(t, 1, len(p.statsPaths)) assert.Equal(t, 2, len(p.inPaths[0].GetBinlogs())) assert.Equal(t, 2, len(p.statsPaths[0].GetBinlogs())) assert.NotNil(t, p.deltaInfo) ctx, cancel := context.WithCancel(context.Background()) cancel() p, err = b.upload(ctx, 1, 10, []*InsertData{iData}, dData, meta) assert.EqualError(t, err, errUploadToBlobStorage.Error()) assert.Nil(t, p) }) t.Run("Test upload error", func(t *testing.T) { f := &MetaFactory{} meta := f.GetCollectionMeta(UniqueID(10001), "uploads", schemapb.DataType_Int64) dData := &DeleteData{ Pks: []primaryKey{}, Tss: []uint64{}, } iData := genEmptyInsertData() p, err := b.upload(context.TODO(), 1, 10, []*InsertData{iData}, dData, meta) assert.NoError(t, err) assert.Empty(t, p.inPaths) assert.Empty(t, p.statsPaths) assert.Empty(t, p.deltaInfo) iData = &InsertData{Data: make(map[int64]storage.FieldData)} p, err = b.upload(context.TODO(), 1, 10, []*InsertData{iData}, dData, meta) assert.NoError(t, err) assert.Empty(t, p.inPaths) assert.Empty(t, p.statsPaths) assert.Empty(t, p.deltaInfo) iData = genInsertData() dData = &DeleteData{ Pks: []primaryKey{}, Tss: []uint64{1}, RowCount: 1, } p, err = b.upload(context.TODO(), 1, 10, []*InsertData{iData}, dData, meta) assert.Error(t, err) assert.Empty(t, p) mkc := &mockCm{errMultiSave: true} bin := &binlogIO{mkc, alloc} iData = genInsertData() pk := newInt64PrimaryKey(1) dData = &DeleteData{ Pks: []primaryKey{pk}, Tss: []uint64{1}, RowCount: 1, } ctx, cancel := context.WithTimeout(context.TODO(), 20*time.Millisecond) p, err = bin.upload(ctx, 1, 10, []*InsertData{iData}, dData, meta) assert.Error(t, err) assert.Empty(t, p) cancel() }) t.Run("Test download", func(t *testing.T) { tests := []struct { isvalid bool ks []string // for preparation inctx context.Context description string }{ {true, []string{"a", "b", "c"}, context.TODO(), "valid input"}, {false, nil, context.Background(), "cancel by context"}, } for _, test := range tests { t.Run(test.description, func(t *testing.T) { if test.isvalid { inkeys := []string{} for _, k := range test.ks { blob, key, err := prepareBlob(cm, k) require.NoError(t, err) assert.NotEmpty(t, blob) inkeys = append(inkeys, key) loaded, err := b.download(test.inctx, []string{key}) assert.NoError(t, err) assert.ElementsMatch(t, blob, loaded[0].GetValue()) } loaded, err := b.download(test.inctx, inkeys) assert.NoError(t, err) assert.Equal(t, len(test.ks), len(loaded)) } else { ctx, cancel := context.WithCancel(test.inctx) cancel() _, err := b.download(ctx, nil) assert.EqualError(t, err, errDownloadFromBlobStorage.Error()) } }) } }) t.Run("Test download twice", func(t *testing.T) { mkc := &mockCm{errMultiLoad: true} b := &binlogIO{mkc, alloc} ctx, cancel := context.WithTimeout(context.TODO(), time.Millisecond*20) blobs, err := b.download(ctx, []string{"a"}) assert.Error(t, err) assert.Empty(t, blobs) cancel() }) } func prepareBlob(cm storage.ChunkManager, key string) ([]byte, string, error) { k := path.Join("test_prepare_blob", key) blob := []byte{1, 2, 3, 255, 188} err := cm.Write(k, blob[:]) if err != nil { return nil, "", err } return blob, k, nil } func TestBinlogIOInnerMethods(t *testing.T) { alloc := NewAllocatorFactory() cm := storage.NewLocalChunkManager(storage.RootPath(binlogTestDir)) defer cm.RemoveWithPrefix("") b := &binlogIO{ cm, alloc, } t.Run("Test genDeltaBlobs", func(t *testing.T) { f := &MetaFactory{} meta := f.GetCollectionMeta(UniqueID(10002), "test_gen_blobs", schemapb.DataType_Int64) tests := []struct { isvalid bool deletepk primaryKey ts uint64 description string }{ {true, newInt64PrimaryKey(1), 1111111, "valid input"}, } for _, test := range tests { t.Run(test.description, func(t *testing.T) { if test.isvalid { k, v, err := b.genDeltaBlobs(&DeleteData{ Pks: []primaryKey{test.deletepk}, Tss: []uint64{test.ts}, }, meta.GetID(), 10, 1) assert.NoError(t, err) assert.NotEmpty(t, k) assert.NotEmpty(t, v) log.Debug("genDeltaBlobs returns", zap.String("key", k)) } }) } }) t.Run("Test genDeltaBlobs error", func(t *testing.T) { pk := newInt64PrimaryKey(1) k, v, err := b.genDeltaBlobs(&DeleteData{Pks: []primaryKey{pk}, Tss: []uint64{}}, 1, 1, 1) assert.Error(t, err) assert.Empty(t, k) assert.Empty(t, v) errAlloc := NewAllocatorFactory() errAlloc.isvalid = false bin := binlogIO{cm, errAlloc} k, v, err = bin.genDeltaBlobs(&DeleteData{Pks: []primaryKey{pk}, Tss: []uint64{1}}, 1, 1, 1) assert.Error(t, err) assert.Empty(t, k) assert.Empty(t, v) }) t.Run("Test genInsertBlobs", func(t *testing.T) { f := &MetaFactory{} tests := []struct { pkType schemapb.DataType description string }{ {schemapb.DataType_Int64, "int64PrimaryField"}, {schemapb.DataType_VarChar, "varCharPrimaryField"}, } for _, test := range tests { t.Run(test.description, func(t *testing.T) { meta := f.GetCollectionMeta(UniqueID(10001), "test_gen_blobs", test.pkType) helper, err := typeutil.CreateSchemaHelper(meta.Schema) assert.NoError(t, err) primaryKeyFieldSchema, err := helper.GetPrimaryKeyField() assert.NoError(t, err) primaryKeyFieldID := primaryKeyFieldSchema.GetFieldID() kvs, pin, pstats, err := b.genInsertBlobs(genInsertData(), 10, 1, meta) assert.NoError(t, err) assert.Equal(t, 1, len(pstats)) assert.Equal(t, 12, len(pin)) assert.Equal(t, 13, len(kvs)) log.Debug("test paths", zap.Any("kvs no.", len(kvs)), zap.String("insert paths field0", pin[common.TimeStampField].GetBinlogs()[0].GetLogPath()), zap.String("stats paths field0", pstats[primaryKeyFieldID].GetBinlogs()[0].GetLogPath())) }) } }) t.Run("Test genInsertBlobs error", func(t *testing.T) { kvs, pin, pstats, err := b.genInsertBlobs(&InsertData{}, 1, 1, nil) assert.Error(t, err) assert.Empty(t, kvs) assert.Empty(t, pin) assert.Empty(t, pstats) f := &MetaFactory{} meta := f.GetCollectionMeta(UniqueID(10001), "test_gen_blobs", schemapb.DataType_Int64) kvs, pin, pstats, err = b.genInsertBlobs(genEmptyInsertData(), 10, 1, meta) assert.Error(t, err) assert.Empty(t, kvs) assert.Empty(t, pin) assert.Empty(t, pstats) errAlloc := NewAllocatorFactory() errAlloc.errAllocBatch = true bin := &binlogIO{cm, errAlloc} kvs, pin, pstats, err = bin.genInsertBlobs(genInsertData(), 10, 1, meta) assert.Error(t, err) assert.Empty(t, kvs) assert.Empty(t, pin) assert.Empty(t, pstats) }) t.Run("Test idxGenerator", func(t *testing.T) { tests := []struct { isvalid bool innumber int expectedNo int description string }{ {false, 0, 0, "Invalid input count n"}, {true, 1, 1, "valid input n 1"}, {true, 3, 3, "valid input n 3 with cancel"}, } for _, test := range tests { t.Run(test.description, func(t *testing.T) { done := make(chan struct{}) if test.isvalid { gen, err := b.idxGenerator(test.innumber, done) assert.NoError(t, err) r := make([]UniqueID, 0) for i := range gen { r = append(r, i) } assert.Equal(t, test.expectedNo, len(r)) if test.innumber > 1 { donedone := make(chan struct{}) gen, err := b.idxGenerator(test.innumber, donedone) assert.NoError(t, err) _, ok := <-gen assert.True(t, ok) donedone <- struct{}{} _, ok = <-gen assert.False(t, ok) } } else { gen, err := b.idxGenerator(test.innumber, done) assert.Error(t, err) assert.Nil(t, gen) } }) } }) } type mockCm struct { storage.ChunkManager errMultiLoad bool errMultiSave bool } var _ storage.ChunkManager = (*mockCm)(nil) func (mk *mockCm) Write(filePath string, content []byte) error { return nil } func (mk *mockCm) MultiWrite(contents map[string][]byte) error { if mk.errMultiSave { return errors.New("mockKv multisave error") } return nil } func (mk *mockCm) Read(filePath string) ([]byte, error) { return nil, nil } func (mk *mockCm) MultiRead(filePaths []string) ([][]byte, error) { if mk.errMultiLoad { return nil, errors.New("mockKv multiload error") } return [][]byte{[]byte("a")}, nil } func (mk *mockCm) ReadWithPrefix(prefix string) ([]string, [][]byte, error) { return nil, nil, nil } func (mk *mockCm) Remove(key string) error { return nil } func (mk *mockCm) MultiRemove(keys []string) error { return nil } func (mk *mockCm) RemoveWithPrefix(key string) error { return nil } func (mk *mockCm) Close() {}