mirror of
https://gitee.com/milvus-io/milvus.git
synced 2024-12-05 05:18:52 +08:00
2599a3ece0
Signed-off-by: yhmo <yihua.mo@zilliz.com> Signed-off-by: yhmo <yihua.mo@zilliz.com>
1133 lines
38 KiB
Go
1133 lines
38 KiB
Go
// Licensed to the LF AI & Data foundation under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
package importutil
|
|
|
|
import (
|
|
"encoding/json"
|
|
"errors"
|
|
"strconv"
|
|
"testing"
|
|
|
|
"github.com/milvus-io/milvus/api/schemapb"
|
|
"github.com/milvus-io/milvus/internal/common"
|
|
"github.com/milvus-io/milvus/internal/storage"
|
|
"github.com/stretchr/testify/assert"
|
|
)
|
|
|
|
const (
|
|
baseTimestamp = 43757345
|
|
)
|
|
|
|
func createDeltalogBuf(t *testing.T, deleteList interface{}, varcharType bool) []byte {
|
|
deleteData := &storage.DeleteData{
|
|
Pks: make([]storage.PrimaryKey, 0),
|
|
Tss: make([]storage.Timestamp, 0),
|
|
RowCount: 0,
|
|
}
|
|
|
|
if varcharType {
|
|
deltaData := deleteList.([]string)
|
|
assert.NotNil(t, deltaData)
|
|
for i, id := range deltaData {
|
|
deleteData.Pks = append(deleteData.Pks, storage.NewVarCharPrimaryKey(id))
|
|
deleteData.Tss = append(deleteData.Tss, baseTimestamp+uint64(i))
|
|
deleteData.RowCount++
|
|
}
|
|
} else {
|
|
deltaData := deleteList.([]int64)
|
|
assert.NotNil(t, deltaData)
|
|
for i, id := range deltaData {
|
|
deleteData.Pks = append(deleteData.Pks, storage.NewInt64PrimaryKey(id))
|
|
deleteData.Tss = append(deleteData.Tss, baseTimestamp+uint64(i))
|
|
deleteData.RowCount++
|
|
}
|
|
}
|
|
|
|
deleteCodec := storage.NewDeleteCodec()
|
|
blob, err := deleteCodec.Serialize(1, 1, 1, deleteData)
|
|
assert.Nil(t, err)
|
|
assert.NotNil(t, blob)
|
|
|
|
return blob.Value
|
|
}
|
|
|
|
// this function create fields data for the sampleSchame()
|
|
func createFieldsData(rowCount int) map[storage.FieldID]interface{} {
|
|
fieldsData := make(map[storage.FieldID]interface{})
|
|
|
|
rowIDData := make([]int64, 0)
|
|
timestampData := make([]int64, 0)
|
|
boolData := make([]bool, 0)
|
|
int8Data := make([]int8, 0)
|
|
int16Data := make([]int16, 0)
|
|
int32Data := make([]int32, 0)
|
|
int64Data := make([]int64, 0)
|
|
floatData := make([]float32, 0)
|
|
doubleData := make([]float64, 0)
|
|
varcharData := make([]string, 0)
|
|
binVecData := make([][]byte, 0)
|
|
floatVecData := make([][]float32, 0)
|
|
|
|
boolFunc := func(i int) bool {
|
|
return i%3 != 0
|
|
}
|
|
|
|
for i := 0; i < rowCount; i++ {
|
|
rowIDData = append(rowIDData, int64(i))
|
|
timestampData = append(timestampData, baseTimestamp+int64(i))
|
|
boolData = append(boolData, boolFunc(i))
|
|
int8Data = append(int8Data, int8(i%256))
|
|
int16Data = append(int16Data, int16(i%65536))
|
|
int32Data = append(int32Data, int32(i%1000))
|
|
int64Data = append(int64Data, int64(i)) // this is primary key
|
|
floatData = append(floatData, float32(i/2))
|
|
doubleData = append(doubleData, float64(i/5))
|
|
varcharData = append(varcharData, "no."+strconv.Itoa(i))
|
|
binVecData = append(binVecData, []byte{byte(i % 256), byte(i % 256)}) // dim = 16
|
|
floatVecData = append(floatVecData, []float32{float32(i / 2), float32(i / 4), float32(i / 5), float32(i / 8)}) // dim = 4
|
|
}
|
|
|
|
fieldsData[0] = rowIDData
|
|
fieldsData[1] = timestampData
|
|
fieldsData[102] = boolData
|
|
fieldsData[103] = int8Data
|
|
fieldsData[104] = int16Data
|
|
fieldsData[105] = int32Data
|
|
fieldsData[106] = int64Data
|
|
fieldsData[107] = floatData
|
|
fieldsData[108] = doubleData
|
|
fieldsData[109] = varcharData
|
|
fieldsData[110] = binVecData
|
|
fieldsData[111] = floatVecData
|
|
|
|
return fieldsData
|
|
}
|
|
|
|
func createSegmentsData(fieldsData map[storage.FieldID]interface{}, shardNum int32) []map[storage.FieldID]storage.FieldData {
|
|
segmentsData := make([]map[storage.FieldID]storage.FieldData, 0, shardNum)
|
|
for i := 0; i < int(shardNum); i++ {
|
|
segData := initSegmentData(sampleSchema())
|
|
fieldID := int64(0)
|
|
segData[fieldID].(*storage.Int64FieldData).Data = append(segData[fieldID].(*storage.Int64FieldData).Data, fieldsData[fieldID].([]int64)...) // RoeIDField
|
|
fieldID = int64(102)
|
|
segData[fieldID].(*storage.BoolFieldData).Data = append(segData[fieldID].(*storage.BoolFieldData).Data, fieldsData[fieldID].([]bool)...)
|
|
fieldID = int64(103)
|
|
segData[fieldID].(*storage.Int8FieldData).Data = append(segData[fieldID].(*storage.Int8FieldData).Data, fieldsData[fieldID].([]int8)...)
|
|
fieldID = int64(104)
|
|
segData[fieldID].(*storage.Int16FieldData).Data = append(segData[fieldID].(*storage.Int16FieldData).Data, fieldsData[fieldID].([]int16)...)
|
|
fieldID = int64(105)
|
|
segData[fieldID].(*storage.Int32FieldData).Data = append(segData[fieldID].(*storage.Int32FieldData).Data, fieldsData[fieldID].([]int32)...)
|
|
fieldID = int64(106)
|
|
segData[fieldID].(*storage.Int64FieldData).Data = append(segData[fieldID].(*storage.Int64FieldData).Data, fieldsData[fieldID].([]int64)...) // primary key
|
|
fieldID = int64(107)
|
|
segData[fieldID].(*storage.FloatFieldData).Data = append(segData[fieldID].(*storage.FloatFieldData).Data, fieldsData[fieldID].([]float32)...)
|
|
fieldID = int64(108)
|
|
segData[fieldID].(*storage.DoubleFieldData).Data = append(segData[fieldID].(*storage.DoubleFieldData).Data, fieldsData[fieldID].([]float64)...)
|
|
fieldID = int64(109)
|
|
segData[fieldID].(*storage.StringFieldData).Data = append(segData[fieldID].(*storage.StringFieldData).Data, fieldsData[fieldID].([]string)...)
|
|
fieldID = int64(110)
|
|
binVectors := fieldsData[fieldID].([][]byte)
|
|
for _, vec := range binVectors {
|
|
segData[fieldID].(*storage.BinaryVectorFieldData).Data = append(segData[fieldID].(*storage.BinaryVectorFieldData).Data, vec...)
|
|
}
|
|
fieldID = int64(111)
|
|
floatVectors := fieldsData[fieldID].([][]float32)
|
|
for _, vec := range floatVectors {
|
|
segData[fieldID].(*storage.FloatVectorFieldData).Data = append(segData[fieldID].(*storage.FloatVectorFieldData).Data, vec...)
|
|
}
|
|
segmentsData = append(segmentsData, segData)
|
|
}
|
|
return segmentsData
|
|
}
|
|
|
|
func Test_NewBinlogAdapter(t *testing.T) {
|
|
// nil schema
|
|
adapter, err := NewBinlogAdapter(nil, 2, 1024, 2048, nil, nil, 0)
|
|
assert.Nil(t, adapter)
|
|
assert.NotNil(t, err)
|
|
|
|
// nil chunkmanager
|
|
adapter, err = NewBinlogAdapter(sampleSchema(), 2, 1024, 2048, nil, nil, 0)
|
|
assert.Nil(t, adapter)
|
|
assert.NotNil(t, err)
|
|
|
|
// nil flushfunc
|
|
adapter, err = NewBinlogAdapter(sampleSchema(), 2, 1024, 2048, &MockChunkManager{}, nil, 0)
|
|
assert.Nil(t, adapter)
|
|
assert.NotNil(t, err)
|
|
|
|
// succeed
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
return nil
|
|
}
|
|
adapter, err = NewBinlogAdapter(sampleSchema(), 2, 2048, 1024, &MockChunkManager{}, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
// no primary key
|
|
schema := &schemapb.CollectionSchema{
|
|
Name: "schema",
|
|
Description: "schema",
|
|
AutoID: true,
|
|
Fields: []*schemapb.FieldSchema{
|
|
{
|
|
FieldID: 101,
|
|
Name: "id",
|
|
IsPrimaryKey: false,
|
|
DataType: schemapb.DataType_Int64,
|
|
},
|
|
},
|
|
}
|
|
adapter, err = NewBinlogAdapter(schema, 2, 1024, 2048, &MockChunkManager{}, flushFunc, 0)
|
|
assert.Nil(t, adapter)
|
|
assert.NotNil(t, err)
|
|
}
|
|
|
|
func Test_BinlogAdapterVerify(t *testing.T) {
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
return nil
|
|
}
|
|
adapter, err := NewBinlogAdapter(sampleSchema(), 2, 1024, 2048, &MockChunkManager{}, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
// nil input
|
|
err = adapter.verify(nil)
|
|
assert.NotNil(t, err)
|
|
|
|
// empty holder
|
|
holder := &SegmentFilesHolder{}
|
|
err = adapter.verify(holder)
|
|
assert.NotNil(t, err)
|
|
|
|
// row id field missed
|
|
holder.fieldFiles = make(map[int64][]string)
|
|
for i := int64(102); i <= 111; i++ {
|
|
holder.fieldFiles[i] = make([]string, 0)
|
|
}
|
|
err = adapter.verify(holder)
|
|
assert.NotNil(t, err)
|
|
|
|
// timestamp field missed
|
|
holder.fieldFiles[common.RowIDField] = []string{
|
|
"a",
|
|
}
|
|
|
|
err = adapter.verify(holder)
|
|
assert.NotNil(t, err)
|
|
|
|
// binlog file count of each field must be equal
|
|
holder.fieldFiles[common.TimeStampField] = []string{
|
|
"a",
|
|
}
|
|
err = adapter.verify(holder)
|
|
assert.NotNil(t, err)
|
|
|
|
// succeed
|
|
for i := int64(102); i <= 111; i++ {
|
|
holder.fieldFiles[i] = []string{
|
|
"a",
|
|
}
|
|
}
|
|
err = adapter.verify(holder)
|
|
assert.Nil(t, err)
|
|
}
|
|
|
|
func Test_BinlogAdapterReadDeltalog(t *testing.T) {
|
|
deleteItems := []int64{1001, 1002, 1003}
|
|
buf := createDeltalogBuf(t, deleteItems, false)
|
|
chunkManager := &MockChunkManager{
|
|
readBuf: map[string][]byte{
|
|
"dummy": buf,
|
|
},
|
|
}
|
|
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
return nil
|
|
}
|
|
|
|
adapter, err := NewBinlogAdapter(sampleSchema(), 2, 1024, 2048, chunkManager, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
// succeed
|
|
deleteLogs, err := adapter.readDeltalog("dummy")
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, len(deleteItems), len(deleteLogs))
|
|
|
|
// failed to init BinlogFile
|
|
adapter.chunkManager = nil
|
|
deleteLogs, err = adapter.readDeltalog("dummy")
|
|
assert.NotNil(t, err)
|
|
assert.Nil(t, deleteLogs)
|
|
|
|
// failed to open binlog file
|
|
chunkManager.readErr = errors.New("error")
|
|
adapter.chunkManager = chunkManager
|
|
deleteLogs, err = adapter.readDeltalog("dummy")
|
|
assert.NotNil(t, err)
|
|
assert.Nil(t, deleteLogs)
|
|
}
|
|
|
|
func Test_BinlogAdapterDecodeDeleteLogs(t *testing.T) {
|
|
deleteItems := []int64{1001, 1002, 1003, 1004, 1005}
|
|
buf := createDeltalogBuf(t, deleteItems, false)
|
|
chunkManager := &MockChunkManager{
|
|
readBuf: map[string][]byte{
|
|
"dummy": buf,
|
|
},
|
|
}
|
|
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
return nil
|
|
}
|
|
|
|
adapter, err := NewBinlogAdapter(sampleSchema(), 2, 1024, 2048, chunkManager, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
holder := &SegmentFilesHolder{
|
|
deltaFiles: []string{
|
|
"dummy",
|
|
},
|
|
}
|
|
|
|
// use timetamp to filter the no.1 and no.2 deletions
|
|
adapter.tsEndPoint = baseTimestamp + 1
|
|
deletions, err := adapter.decodeDeleteLogs(holder)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 2, len(deletions))
|
|
|
|
// wrong data type of delta log
|
|
chunkManager.readBuf = map[string][]byte{
|
|
"dummy": createDeltalogBuf(t, []string{"1001", "1002"}, true),
|
|
}
|
|
|
|
adapter, err = NewBinlogAdapter(sampleSchema(), 2, 1024, 2048, chunkManager, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
adapter.tsEndPoint = baseTimestamp
|
|
deletions, err = adapter.decodeDeleteLogs(holder)
|
|
assert.NotNil(t, err)
|
|
assert.Nil(t, deletions)
|
|
}
|
|
|
|
func Test_BinlogAdapterDecodeDeleteLog(t *testing.T) {
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
return nil
|
|
}
|
|
|
|
adapter, err := NewBinlogAdapter(sampleSchema(), 2, 1024, 2048, &MockChunkManager{}, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
// v2.1 format
|
|
st := &storage.DeleteLog{
|
|
Pk: &storage.Int64PrimaryKey{
|
|
Value: 100,
|
|
},
|
|
Ts: uint64(450000),
|
|
PkType: 5,
|
|
}
|
|
|
|
m, _ := json.Marshal(st)
|
|
|
|
del, err := adapter.decodeDeleteLog(string(m))
|
|
assert.Nil(t, err)
|
|
assert.NotNil(t, del)
|
|
assert.True(t, del.Pk.EQ(st.Pk))
|
|
assert.Equal(t, st.Ts, del.Ts)
|
|
assert.Equal(t, st.PkType, del.PkType)
|
|
|
|
// v2.0 format
|
|
del, err = adapter.decodeDeleteLog("")
|
|
assert.Nil(t, del)
|
|
assert.NotNil(t, err)
|
|
|
|
del, err = adapter.decodeDeleteLog("a,b")
|
|
assert.Nil(t, del)
|
|
assert.NotNil(t, err)
|
|
|
|
del, err = adapter.decodeDeleteLog("5,b")
|
|
assert.Nil(t, del)
|
|
assert.NotNil(t, err)
|
|
|
|
del, err = adapter.decodeDeleteLog("5,1000")
|
|
assert.Nil(t, err)
|
|
assert.NotNil(t, del)
|
|
assert.True(t, del.Pk.EQ(&storage.Int64PrimaryKey{
|
|
Value: 5,
|
|
}))
|
|
tt, _ := strconv.ParseUint("1000", 10, 64)
|
|
assert.Equal(t, del.Ts, tt)
|
|
assert.Equal(t, del.PkType, int64(schemapb.DataType_Int64))
|
|
}
|
|
|
|
func Test_BinlogAdapterReadDeltalogs(t *testing.T) {
|
|
deleteItems := []int64{1001, 1002, 1003, 1004, 1005}
|
|
buf := createDeltalogBuf(t, deleteItems, false)
|
|
chunkManager := &MockChunkManager{
|
|
readBuf: map[string][]byte{
|
|
"dummy": buf,
|
|
},
|
|
}
|
|
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
return nil
|
|
}
|
|
|
|
adapter, err := NewBinlogAdapter(sampleSchema(), 2, 1024, 2048, chunkManager, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
holder := &SegmentFilesHolder{
|
|
deltaFiles: []string{
|
|
"dummy",
|
|
},
|
|
}
|
|
|
|
// 1. int64 primary key, succeed, return the no.1 and no.2 deletion
|
|
adapter.tsEndPoint = baseTimestamp + 1
|
|
intDeletions, strDeletions, err := adapter.readDeltalogs(holder)
|
|
assert.Nil(t, err)
|
|
assert.Nil(t, strDeletions)
|
|
assert.NotNil(t, intDeletions)
|
|
|
|
ts, ok := intDeletions[deleteItems[0]]
|
|
assert.True(t, ok)
|
|
assert.Equal(t, uint64(baseTimestamp), ts)
|
|
|
|
ts, ok = intDeletions[deleteItems[1]]
|
|
assert.True(t, ok)
|
|
assert.Equal(t, uint64(baseTimestamp+1), ts)
|
|
|
|
// 2. varchar primary key, succeed, return the no.1 and no.2 deletetion
|
|
schema := &schemapb.CollectionSchema{
|
|
Name: "schema",
|
|
Fields: []*schemapb.FieldSchema{
|
|
{
|
|
FieldID: 101,
|
|
Name: "ID",
|
|
IsPrimaryKey: true,
|
|
DataType: schemapb.DataType_VarChar,
|
|
},
|
|
},
|
|
}
|
|
|
|
chunkManager.readBuf = map[string][]byte{
|
|
"dummy": createDeltalogBuf(t, []string{"1001", "1002"}, true),
|
|
}
|
|
|
|
adapter, err = NewBinlogAdapter(schema, 2, 1024, 2048, chunkManager, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
// 2.1 all deletion have been filtered out
|
|
intDeletions, strDeletions, err = adapter.readDeltalogs(holder)
|
|
assert.Nil(t, err)
|
|
assert.Nil(t, intDeletions)
|
|
assert.Nil(t, strDeletions)
|
|
|
|
// 2.2 filter the no.1 and no.2 deletion
|
|
adapter.tsEndPoint = baseTimestamp + 1
|
|
intDeletions, strDeletions, err = adapter.readDeltalogs(holder)
|
|
assert.Nil(t, err)
|
|
assert.Nil(t, intDeletions)
|
|
assert.NotNil(t, strDeletions)
|
|
|
|
ts, ok = strDeletions["1001"]
|
|
assert.True(t, ok)
|
|
assert.Equal(t, uint64(baseTimestamp), ts)
|
|
|
|
ts, ok = strDeletions["1002"]
|
|
assert.True(t, ok)
|
|
assert.Equal(t, uint64(baseTimestamp+1), ts)
|
|
|
|
// 3. unsupported primary key type
|
|
schema = &schemapb.CollectionSchema{
|
|
Name: "schema",
|
|
Fields: []*schemapb.FieldSchema{
|
|
{
|
|
FieldID: 101,
|
|
Name: "ID",
|
|
IsPrimaryKey: true,
|
|
DataType: schemapb.DataType_Float,
|
|
},
|
|
},
|
|
}
|
|
|
|
adapter, err = NewBinlogAdapter(schema, 2, 1024, 2048, chunkManager, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
adapter.tsEndPoint = baseTimestamp + 1
|
|
intDeletions, strDeletions, err = adapter.readDeltalogs(holder)
|
|
assert.NotNil(t, err)
|
|
assert.Nil(t, intDeletions)
|
|
assert.Nil(t, strDeletions)
|
|
}
|
|
|
|
func Test_BinlogAdapterReadTimestamp(t *testing.T) {
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
return nil
|
|
}
|
|
adapter, err := NewBinlogAdapter(sampleSchema(), 2, 1024, 2048, &MockChunkManager{}, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
// new BinglogFile error
|
|
adapter.chunkManager = nil
|
|
ts, err := adapter.readTimestamp("dummy")
|
|
assert.Nil(t, ts)
|
|
assert.NotNil(t, err)
|
|
|
|
// open binlog file error
|
|
chunkManager := &MockChunkManager{
|
|
readBuf: make(map[string][]byte),
|
|
}
|
|
adapter.chunkManager = chunkManager
|
|
ts, err = adapter.readTimestamp("dummy")
|
|
assert.Nil(t, ts)
|
|
assert.NotNil(t, err)
|
|
|
|
// succeed
|
|
rowCount := 10
|
|
fieldsData := createFieldsData(rowCount)
|
|
chunkManager.readBuf["dummy"] = createBinlogBuf(t, schemapb.DataType_Int64, fieldsData[106].([]int64))
|
|
ts, err = adapter.readTimestamp("dummy")
|
|
assert.Nil(t, err)
|
|
assert.NotNil(t, ts)
|
|
assert.Equal(t, rowCount, len(ts))
|
|
}
|
|
|
|
func Test_BinlogAdapterReadPrimaryKeys(t *testing.T) {
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
return nil
|
|
}
|
|
adapter, err := NewBinlogAdapter(sampleSchema(), 2, 1024, 2048, &MockChunkManager{}, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
// new BinglogFile error
|
|
adapter.chunkManager = nil
|
|
intList, strList, err := adapter.readPrimaryKeys("dummy")
|
|
assert.Nil(t, intList)
|
|
assert.Nil(t, strList)
|
|
assert.NotNil(t, err)
|
|
|
|
// open binlog file error
|
|
chunkManager := &MockChunkManager{
|
|
readBuf: make(map[string][]byte),
|
|
}
|
|
adapter.chunkManager = chunkManager
|
|
intList, strList, err = adapter.readPrimaryKeys("dummy")
|
|
assert.Nil(t, intList)
|
|
assert.Nil(t, strList)
|
|
assert.NotNil(t, err)
|
|
|
|
// wrong primary key type
|
|
rowCount := 10
|
|
fieldsData := createFieldsData(rowCount)
|
|
chunkManager.readBuf["dummy"] = createBinlogBuf(t, schemapb.DataType_Bool, fieldsData[102].([]bool))
|
|
|
|
adapter.primaryType = schemapb.DataType_Bool
|
|
intList, strList, err = adapter.readPrimaryKeys("dummy")
|
|
assert.Nil(t, intList)
|
|
assert.Nil(t, strList)
|
|
assert.NotNil(t, err)
|
|
|
|
// succeed int64
|
|
adapter.primaryType = schemapb.DataType_Int64
|
|
chunkManager.readBuf["dummy"] = createBinlogBuf(t, schemapb.DataType_Int64, fieldsData[106].([]int64))
|
|
intList, strList, err = adapter.readPrimaryKeys("dummy")
|
|
assert.NotNil(t, intList)
|
|
assert.Nil(t, strList)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, rowCount, len(intList))
|
|
|
|
// succeed varchar
|
|
adapter.primaryType = schemapb.DataType_VarChar
|
|
chunkManager.readBuf["dummy"] = createBinlogBuf(t, schemapb.DataType_VarChar, fieldsData[109].([]string))
|
|
intList, strList, err = adapter.readPrimaryKeys("dummy")
|
|
assert.Nil(t, intList)
|
|
assert.NotNil(t, strList)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, rowCount, len(strList))
|
|
}
|
|
|
|
func Test_BinlogAdapterShardListInt64(t *testing.T) {
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
return nil
|
|
}
|
|
|
|
shardNum := int32(2)
|
|
adapter, err := NewBinlogAdapter(sampleSchema(), shardNum, 1024, 2048, &MockChunkManager{}, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
segmentsData := make([]map[storage.FieldID]storage.FieldData, 0, shardNum)
|
|
for i := 0; i < int(shardNum); i++ {
|
|
segmentData := initSegmentData(sampleSchema())
|
|
segmentsData = append(segmentsData, segmentData)
|
|
}
|
|
|
|
// wrong input
|
|
shardList, err := adapter.getShardingListByPrimaryInt64([]int64{1}, []int64{1, 2}, segmentsData, map[int64]uint64{})
|
|
assert.Nil(t, shardList)
|
|
assert.NotNil(t, err)
|
|
|
|
// succeed
|
|
// 5 ids, delete two items, the ts end point is 25, there shardList should be [-1, 0, 1, -1, -1]
|
|
adapter.tsEndPoint = 30
|
|
idList := []int64{1, 2, 3, 4, 5}
|
|
tsList := []int64{10, 20, 30, 40, 50}
|
|
deletion := map[int64]uint64{
|
|
1: 23,
|
|
4: 36,
|
|
}
|
|
shardList, err = adapter.getShardingListByPrimaryInt64(idList, tsList, segmentsData, deletion)
|
|
assert.Nil(t, err)
|
|
assert.NotNil(t, shardList)
|
|
correctShardList := []int32{-1, 0, 1, -1, -1}
|
|
assert.Equal(t, len(correctShardList), len(shardList))
|
|
for i := 0; i < len(shardList); i++ {
|
|
assert.Equal(t, correctShardList[i], shardList[i])
|
|
}
|
|
}
|
|
|
|
func Test_BinlogAdapterShardListVarchar(t *testing.T) {
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
return nil
|
|
}
|
|
|
|
shardNum := int32(2)
|
|
adapter, err := NewBinlogAdapter(strKeySchema(), shardNum, 1024, 2048, &MockChunkManager{}, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
segmentsData := make([]map[storage.FieldID]storage.FieldData, 0, shardNum)
|
|
for i := 0; i < int(shardNum); i++ {
|
|
segmentData := initSegmentData(strKeySchema())
|
|
segmentsData = append(segmentsData, segmentData)
|
|
}
|
|
|
|
// wrong input
|
|
shardList, err := adapter.getShardingListByPrimaryVarchar([]string{"1"}, []int64{1, 2}, segmentsData, map[string]uint64{})
|
|
assert.Nil(t, shardList)
|
|
assert.NotNil(t, err)
|
|
|
|
// succeed
|
|
// 5 ids, delete two items, the ts end point is 25, there shardList should be [-1, 1, 1, -1, -1]
|
|
adapter.tsEndPoint = 30
|
|
idList := []string{"1", "2", "3", "4", "5"}
|
|
tsList := []int64{10, 20, 30, 40, 50}
|
|
deletion := map[string]uint64{
|
|
"1": 23,
|
|
"4": 36,
|
|
}
|
|
shardList, err = adapter.getShardingListByPrimaryVarchar(idList, tsList, segmentsData, deletion)
|
|
assert.Nil(t, err)
|
|
assert.NotNil(t, shardList)
|
|
correctShardList := []int32{-1, 1, 1, -1, -1}
|
|
assert.Equal(t, len(correctShardList), len(shardList))
|
|
for i := 0; i < len(shardList); i++ {
|
|
assert.Equal(t, correctShardList[i], shardList[i])
|
|
}
|
|
}
|
|
|
|
func Test_BinlogAdapterReadInt64PK(t *testing.T) {
|
|
chunkManager := &MockChunkManager{}
|
|
|
|
flushCounter := 0
|
|
flushRowCount := 0
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
flushCounter++
|
|
rowCount := 0
|
|
for _, v := range fields {
|
|
rowCount = v.RowNum()
|
|
break
|
|
}
|
|
flushRowCount += rowCount
|
|
for _, v := range fields {
|
|
assert.Equal(t, rowCount, v.RowNum())
|
|
}
|
|
return nil
|
|
}
|
|
|
|
shardNum := int32(2)
|
|
adapter, err := NewBinlogAdapter(sampleSchema(), shardNum, 1024, 2048, chunkManager, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
adapter.tsEndPoint = baseTimestamp + 1
|
|
|
|
// nil holder
|
|
err = adapter.Read(nil)
|
|
assert.NotNil(t, err)
|
|
|
|
// verify failed
|
|
holder := &SegmentFilesHolder{}
|
|
err = adapter.Read(holder)
|
|
assert.NotNil(t, err)
|
|
|
|
// failed to read delta log
|
|
holder.fieldFiles = map[int64][]string{
|
|
int64(0): {"0_insertlog"},
|
|
int64(1): {"1_insertlog"},
|
|
int64(102): {"102_insertlog"},
|
|
int64(103): {"103_insertlog"},
|
|
int64(104): {"104_insertlog"},
|
|
int64(105): {"105_insertlog"},
|
|
int64(106): {"106_insertlog"},
|
|
int64(107): {"107_insertlog"},
|
|
int64(108): {"108_insertlog"},
|
|
int64(109): {"109_insertlog"},
|
|
int64(110): {"110_insertlog"},
|
|
int64(111): {"111_insertlog"},
|
|
}
|
|
holder.deltaFiles = []string{"deltalog"}
|
|
err = adapter.Read(holder)
|
|
assert.NotNil(t, err)
|
|
|
|
// prepare binlog data
|
|
rowCount := 1000
|
|
fieldsData := createFieldsData(rowCount)
|
|
deletedItems := []int64{41, 51, 100, 400, 600}
|
|
|
|
chunkManager.readBuf = map[string][]byte{
|
|
"102_insertlog": createBinlogBuf(t, schemapb.DataType_Bool, fieldsData[102].([]bool)),
|
|
"103_insertlog": createBinlogBuf(t, schemapb.DataType_Int8, fieldsData[103].([]int8)),
|
|
"104_insertlog": createBinlogBuf(t, schemapb.DataType_Int16, fieldsData[104].([]int16)),
|
|
"105_insertlog": createBinlogBuf(t, schemapb.DataType_Int32, fieldsData[105].([]int32)),
|
|
"106_insertlog": createBinlogBuf(t, schemapb.DataType_Int64, fieldsData[106].([]int64)), // this is primary key
|
|
"107_insertlog": createBinlogBuf(t, schemapb.DataType_Float, fieldsData[107].([]float32)),
|
|
"108_insertlog": createBinlogBuf(t, schemapb.DataType_Double, fieldsData[108].([]float64)),
|
|
"109_insertlog": createBinlogBuf(t, schemapb.DataType_VarChar, fieldsData[109].([]string)),
|
|
"110_insertlog": createBinlogBuf(t, schemapb.DataType_BinaryVector, fieldsData[110].([][]byte)),
|
|
"111_insertlog": createBinlogBuf(t, schemapb.DataType_FloatVector, fieldsData[111].([][]float32)),
|
|
"deltalog": createDeltalogBuf(t, deletedItems, false),
|
|
}
|
|
|
|
// failed to read primary keys
|
|
err = adapter.Read(holder)
|
|
assert.NotNil(t, err)
|
|
|
|
// failed to read timestamp field
|
|
chunkManager.readBuf["0_insertlog"] = createBinlogBuf(t, schemapb.DataType_Int64, fieldsData[0].([]int64))
|
|
err = adapter.Read(holder)
|
|
assert.NotNil(t, err)
|
|
|
|
// succeed flush
|
|
chunkManager.readBuf["1_insertlog"] = createBinlogBuf(t, schemapb.DataType_Int64, fieldsData[1].([]int64))
|
|
|
|
adapter.tsEndPoint = baseTimestamp + uint64(499) // 4 entities deleted, 500 entities excluded
|
|
err = adapter.Read(holder)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, shardNum, int32(flushCounter))
|
|
assert.Equal(t, rowCount-4-500, flushRowCount)
|
|
}
|
|
|
|
func Test_BinlogAdapterReadVarcharPK(t *testing.T) {
|
|
chunkManager := &MockChunkManager{}
|
|
|
|
flushCounter := 0
|
|
flushRowCount := 0
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
flushCounter++
|
|
rowCount := 0
|
|
for _, v := range fields {
|
|
rowCount = v.RowNum()
|
|
break
|
|
}
|
|
flushRowCount += rowCount
|
|
for _, v := range fields {
|
|
assert.Equal(t, rowCount, v.RowNum())
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// prepare data
|
|
holder := &SegmentFilesHolder{}
|
|
holder.fieldFiles = map[int64][]string{
|
|
int64(0): {"0_insertlog"},
|
|
int64(1): {"1_insertlog"},
|
|
int64(101): {"101_insertlog"},
|
|
int64(102): {"102_insertlog"},
|
|
int64(103): {"103_insertlog"},
|
|
int64(104): {"104_insertlog"},
|
|
int64(105): {"105_insertlog"},
|
|
int64(106): {"106_insertlog"},
|
|
}
|
|
holder.deltaFiles = []string{"deltalog"}
|
|
|
|
rowIDData := make([]int64, 0)
|
|
timestampData := make([]int64, 0)
|
|
pkData := make([]string, 0)
|
|
int32Data := make([]int32, 0)
|
|
floatData := make([]float32, 0)
|
|
varcharData := make([]string, 0)
|
|
boolData := make([]bool, 0)
|
|
floatVecData := make([][]float32, 0)
|
|
|
|
boolFunc := func(i int) bool {
|
|
return i%3 != 0
|
|
}
|
|
|
|
rowCount := 1000
|
|
for i := 0; i < rowCount; i++ {
|
|
rowIDData = append(rowIDData, int64(i))
|
|
timestampData = append(timestampData, baseTimestamp+int64(i))
|
|
pkData = append(pkData, strconv.Itoa(i)) // primary key
|
|
int32Data = append(int32Data, int32(i%1000))
|
|
floatData = append(floatData, float32(i/2))
|
|
varcharData = append(varcharData, "no."+strconv.Itoa(i))
|
|
boolData = append(boolData, boolFunc(i))
|
|
floatVecData = append(floatVecData, []float32{float32(i / 2), float32(i / 4), float32(i / 5), float32(i / 8)}) // dim = 4
|
|
}
|
|
|
|
deletedItems := []string{"1", "100", "999"}
|
|
|
|
chunkManager.readBuf = map[string][]byte{
|
|
"0_insertlog": createBinlogBuf(t, schemapb.DataType_Int64, rowIDData),
|
|
"1_insertlog": createBinlogBuf(t, schemapb.DataType_Int64, timestampData),
|
|
"101_insertlog": createBinlogBuf(t, schemapb.DataType_VarChar, pkData),
|
|
"102_insertlog": createBinlogBuf(t, schemapb.DataType_Int32, int32Data),
|
|
"103_insertlog": createBinlogBuf(t, schemapb.DataType_Float, floatData),
|
|
"104_insertlog": createBinlogBuf(t, schemapb.DataType_VarChar, varcharData),
|
|
"105_insertlog": createBinlogBuf(t, schemapb.DataType_Bool, boolData),
|
|
"106_insertlog": createBinlogBuf(t, schemapb.DataType_FloatVector, floatVecData),
|
|
"deltalog": createDeltalogBuf(t, deletedItems, true),
|
|
}
|
|
|
|
// succeed
|
|
shardNum := int32(3)
|
|
adapter, err := NewBinlogAdapter(strKeySchema(), shardNum, 1024, 2048, chunkManager, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
adapter.tsEndPoint = baseTimestamp + uint64(499) // 3 entities deleted, 500 entities excluded, the "999" is excluded, so totally 502 entities skipped
|
|
err = adapter.Read(holder)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, shardNum, int32(flushCounter))
|
|
assert.Equal(t, rowCount-502, flushRowCount)
|
|
}
|
|
|
|
func Test_BinlogAdapterTryFlush(t *testing.T) {
|
|
flushCounter := 0
|
|
flushRowCount := 0
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
flushCounter++
|
|
rowCount := 0
|
|
for _, v := range fields {
|
|
rowCount = v.RowNum()
|
|
break
|
|
}
|
|
flushRowCount += rowCount
|
|
for _, v := range fields {
|
|
assert.Equal(t, rowCount, v.RowNum())
|
|
}
|
|
return nil
|
|
}
|
|
|
|
segmentSize := int64(1024)
|
|
maxTotalSize := int64(2048)
|
|
shardNum := int32(3)
|
|
adapter, err := NewBinlogAdapter(sampleSchema(), shardNum, segmentSize, maxTotalSize, &MockChunkManager{}, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
// prepare flush data, 3 shards, each shard 10 rows
|
|
rowCount := 10
|
|
fieldsData := createFieldsData(rowCount)
|
|
|
|
// non-force flush
|
|
segmentsData := createSegmentsData(fieldsData, shardNum)
|
|
err = adapter.tryFlushSegments(segmentsData, false)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 0, flushCounter)
|
|
assert.Equal(t, 0, flushRowCount)
|
|
|
|
// force flush
|
|
err = adapter.tryFlushSegments(segmentsData, true)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, int(shardNum), flushCounter)
|
|
assert.Equal(t, rowCount*int(shardNum), flushRowCount)
|
|
|
|
// after force flush, no data left
|
|
flushCounter = 0
|
|
flushRowCount = 0
|
|
err = adapter.tryFlushSegments(segmentsData, true)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 0, flushCounter)
|
|
assert.Equal(t, 0, flushRowCount)
|
|
|
|
// flush when segment size exceeds segmentSize
|
|
segmentsData = createSegmentsData(fieldsData, shardNum)
|
|
adapter.segmentSize = 100 // segmentSize is 100 bytes, less than the 10 rows size
|
|
err = adapter.tryFlushSegments(segmentsData, false)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, int(shardNum), flushCounter)
|
|
assert.Equal(t, rowCount*int(shardNum), flushRowCount)
|
|
|
|
flushCounter = 0
|
|
flushRowCount = 0
|
|
err = adapter.tryFlushSegments(segmentsData, true) // no data left
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 0, flushCounter)
|
|
assert.Equal(t, 0, flushRowCount)
|
|
|
|
// flush when segments total size exceeds maxTotalSize
|
|
segmentsData = createSegmentsData(fieldsData, shardNum)
|
|
adapter.segmentSize = 4096 // segmentSize is 4096 bytes, larger than the 10 rows size
|
|
adapter.maxTotalSize = 100 // maxTotalSize is 100 bytes, less than the 30 rows size
|
|
err = adapter.tryFlushSegments(segmentsData, false)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, flushCounter) // only the max segment is flushed
|
|
assert.Equal(t, 10, flushRowCount)
|
|
|
|
flushCounter = 0
|
|
flushRowCount = 0
|
|
err = adapter.tryFlushSegments(segmentsData, true) // two segments left
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 2, flushCounter)
|
|
assert.Equal(t, 20, flushRowCount)
|
|
}
|
|
|
|
func Test_BinlogAdapterDispatch(t *testing.T) {
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
return nil
|
|
}
|
|
shardNum := int32(3)
|
|
adapter, err := NewBinlogAdapter(sampleSchema(), shardNum, 1024, 2048, &MockChunkManager{}, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
// prepare empty in-memory segments data
|
|
segmentsData := make([]map[storage.FieldID]storage.FieldData, 0, shardNum)
|
|
for i := 0; i < int(shardNum); i++ {
|
|
segmentData := initSegmentData(sampleSchema())
|
|
segmentsData = append(segmentsData, segmentData)
|
|
}
|
|
|
|
// dispatch bool data
|
|
fieldID := int64(102)
|
|
shardList := []int32{0, -1, 1}
|
|
err = adapter.dispatchBoolToShards([]bool{true}, segmentsData, shardList, fieldID) // row count mismatch
|
|
assert.NotNil(t, err)
|
|
for _, segment := range segmentsData {
|
|
assert.Equal(t, 0, segment[fieldID].RowNum())
|
|
}
|
|
err = adapter.dispatchBoolToShards([]bool{true, false, false}, segmentsData, shardList, fieldID) // succeed
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, segmentsData[0][fieldID].RowNum())
|
|
assert.Equal(t, 1, segmentsData[1][fieldID].RowNum())
|
|
assert.Equal(t, 0, segmentsData[2][fieldID].RowNum())
|
|
|
|
// dispatch int8 data
|
|
fieldID = int64(103)
|
|
err = adapter.dispatchInt8ToShards([]int8{1, 2, 3, 4}, segmentsData, shardList, fieldID) // row count mismatch
|
|
assert.NotNil(t, err)
|
|
for _, segment := range segmentsData {
|
|
assert.Equal(t, 0, segment[fieldID].RowNum())
|
|
}
|
|
err = adapter.dispatchInt8ToShards([]int8{1, 2, 3}, segmentsData, shardList, fieldID) // succeed
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, segmentsData[0][fieldID].RowNum())
|
|
assert.Equal(t, 1, segmentsData[1][fieldID].RowNum())
|
|
assert.Equal(t, 0, segmentsData[2][fieldID].RowNum())
|
|
|
|
// dispatch int16 data
|
|
fieldID = int64(104)
|
|
err = adapter.dispatchInt16ToShards([]int16{1, 2, 3, 4}, segmentsData, shardList, fieldID) // row count mismatch
|
|
assert.NotNil(t, err)
|
|
for _, segment := range segmentsData {
|
|
assert.Equal(t, 0, segment[fieldID].RowNum())
|
|
}
|
|
err = adapter.dispatchInt16ToShards([]int16{1, 2, 3}, segmentsData, shardList, fieldID) // succeed
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, segmentsData[0][fieldID].RowNum())
|
|
assert.Equal(t, 1, segmentsData[1][fieldID].RowNum())
|
|
assert.Equal(t, 0, segmentsData[2][fieldID].RowNum())
|
|
|
|
// dispatch int32 data
|
|
fieldID = int64(105)
|
|
err = adapter.dispatchInt32ToShards([]int32{1, 2, 3, 4}, segmentsData, shardList, fieldID) // row count mismatch
|
|
assert.NotNil(t, err)
|
|
for _, segment := range segmentsData {
|
|
assert.Equal(t, 0, segment[fieldID].RowNum())
|
|
}
|
|
err = adapter.dispatchInt32ToShards([]int32{1, 2, 3}, segmentsData, shardList, fieldID) // succeed
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, segmentsData[0][fieldID].RowNum())
|
|
assert.Equal(t, 1, segmentsData[1][fieldID].RowNum())
|
|
assert.Equal(t, 0, segmentsData[2][fieldID].RowNum())
|
|
|
|
// dispatch int64 data
|
|
fieldID = int64(106)
|
|
err = adapter.dispatchInt64ToShards([]int64{1, 2, 3, 4}, segmentsData, shardList, fieldID) // row count mismatch
|
|
assert.NotNil(t, err)
|
|
for _, segment := range segmentsData {
|
|
assert.Equal(t, 0, segment[fieldID].RowNum())
|
|
}
|
|
err = adapter.dispatchInt64ToShards([]int64{1, 2, 3}, segmentsData, shardList, fieldID) // succeed
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, segmentsData[0][fieldID].RowNum())
|
|
assert.Equal(t, 1, segmentsData[1][fieldID].RowNum())
|
|
assert.Equal(t, 0, segmentsData[2][fieldID].RowNum())
|
|
|
|
// dispatch float data
|
|
fieldID = int64(107)
|
|
err = adapter.dispatchFloatToShards([]float32{1, 2, 3, 4}, segmentsData, shardList, fieldID) // row count mismatch
|
|
assert.NotNil(t, err)
|
|
for _, segment := range segmentsData {
|
|
assert.Equal(t, 0, segment[fieldID].RowNum())
|
|
}
|
|
err = adapter.dispatchFloatToShards([]float32{1, 2, 3}, segmentsData, shardList, fieldID) // succeed
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, segmentsData[0][fieldID].RowNum())
|
|
assert.Equal(t, 1, segmentsData[1][fieldID].RowNum())
|
|
assert.Equal(t, 0, segmentsData[2][fieldID].RowNum())
|
|
|
|
// dispatch double data
|
|
fieldID = int64(108)
|
|
err = adapter.dispatchDoubleToShards([]float64{1, 2, 3, 4}, segmentsData, shardList, fieldID) // row count mismatch
|
|
assert.NotNil(t, err)
|
|
for _, segment := range segmentsData {
|
|
assert.Equal(t, 0, segment[fieldID].RowNum())
|
|
}
|
|
err = adapter.dispatchDoubleToShards([]float64{1, 2, 3}, segmentsData, shardList, fieldID) // succeed
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, segmentsData[0][fieldID].RowNum())
|
|
assert.Equal(t, 1, segmentsData[1][fieldID].RowNum())
|
|
assert.Equal(t, 0, segmentsData[2][fieldID].RowNum())
|
|
|
|
// dispatch varchar data
|
|
fieldID = int64(109)
|
|
err = adapter.dispatchVarcharToShards([]string{"a", "b", "c", "d"}, segmentsData, shardList, fieldID) // row count mismatch
|
|
assert.NotNil(t, err)
|
|
for _, segment := range segmentsData {
|
|
assert.Equal(t, 0, segment[fieldID].RowNum())
|
|
}
|
|
err = adapter.dispatchVarcharToShards([]string{"a", "b", "c"}, segmentsData, shardList, fieldID) // succeed
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, segmentsData[0][fieldID].RowNum())
|
|
assert.Equal(t, 1, segmentsData[1][fieldID].RowNum())
|
|
assert.Equal(t, 0, segmentsData[2][fieldID].RowNum())
|
|
|
|
// dispatch binary vector data
|
|
fieldID = int64(110)
|
|
err = adapter.dispatchBinaryVecToShards([]byte{1, 2, 3, 4}, 16, segmentsData, shardList, fieldID) // row count mismatch
|
|
assert.NotNil(t, err)
|
|
for _, segment := range segmentsData {
|
|
assert.Equal(t, 0, segment[fieldID].RowNum())
|
|
}
|
|
err = adapter.dispatchBinaryVecToShards([]byte{1, 2, 3, 4, 5, 6}, 16, segmentsData, shardList, fieldID) // succeed
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, segmentsData[0][fieldID].RowNum())
|
|
assert.Equal(t, 1, segmentsData[1][fieldID].RowNum())
|
|
assert.Equal(t, 0, segmentsData[2][fieldID].RowNum())
|
|
|
|
// dispatch float vector data
|
|
fieldID = int64(111)
|
|
err = adapter.dispatchFloatVecToShards([]float32{1, 2, 3, 4}, 4, segmentsData, shardList, fieldID) // row count mismatch
|
|
assert.NotNil(t, err)
|
|
for _, segment := range segmentsData {
|
|
assert.Equal(t, 0, segment[fieldID].RowNum())
|
|
}
|
|
err = adapter.dispatchFloatVecToShards([]float32{1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12}, 4, segmentsData, shardList, fieldID) // succeed
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, segmentsData[0][fieldID].RowNum())
|
|
assert.Equal(t, 1, segmentsData[1][fieldID].RowNum())
|
|
assert.Equal(t, 0, segmentsData[2][fieldID].RowNum())
|
|
}
|
|
|
|
func Test_BinlogAdapterReadInsertlog(t *testing.T) {
|
|
flushFunc := func(fields map[storage.FieldID]storage.FieldData, shardID int) error {
|
|
return nil
|
|
}
|
|
adapter, err := NewBinlogAdapter(sampleSchema(), 2, 1024, 2048, &MockChunkManager{}, flushFunc, 0)
|
|
assert.NotNil(t, adapter)
|
|
assert.Nil(t, err)
|
|
|
|
// new BinglogFile error
|
|
segmentsData := make([]map[storage.FieldID]storage.FieldData, 0, 2)
|
|
adapter.chunkManager = nil
|
|
err = adapter.readInsertlog(1, "dummy", segmentsData, []int32{1})
|
|
assert.NotNil(t, err)
|
|
|
|
// open binlog file error
|
|
chunkManager := &MockChunkManager{
|
|
readBuf: make(map[string][]byte),
|
|
}
|
|
adapter.chunkManager = chunkManager
|
|
err = adapter.readInsertlog(1, "dummy", segmentsData, []int32{1})
|
|
assert.NotNil(t, err)
|
|
|
|
// prepare binlog data
|
|
rowCount := 10
|
|
fieldsData := createFieldsData(rowCount)
|
|
|
|
// failed to dispatch bool data
|
|
chunkManager.readBuf["bool"] = createBinlogBuf(t, schemapb.DataType_Bool, fieldsData[102].([]bool))
|
|
err = adapter.readInsertlog(1, "bool", segmentsData, []int32{1})
|
|
assert.NotNil(t, err)
|
|
|
|
// failed to dispatch int8 data
|
|
chunkManager.readBuf["int8"] = createBinlogBuf(t, schemapb.DataType_Int8, fieldsData[103].([]int8))
|
|
err = adapter.readInsertlog(1, "int8", segmentsData, []int32{1})
|
|
assert.NotNil(t, err)
|
|
|
|
// failed to dispatch int16 data
|
|
chunkManager.readBuf["int16"] = createBinlogBuf(t, schemapb.DataType_Int16, fieldsData[104].([]int16))
|
|
err = adapter.readInsertlog(1, "int16", segmentsData, []int32{1})
|
|
assert.NotNil(t, err)
|
|
|
|
// failed to dispatch int32 data
|
|
chunkManager.readBuf["int32"] = createBinlogBuf(t, schemapb.DataType_Int32, fieldsData[105].([]int32))
|
|
err = adapter.readInsertlog(1, "int32", segmentsData, []int32{1})
|
|
assert.NotNil(t, err)
|
|
|
|
// failed to dispatch int64 data
|
|
chunkManager.readBuf["int64"] = createBinlogBuf(t, schemapb.DataType_Int64, fieldsData[106].([]int64))
|
|
err = adapter.readInsertlog(1, "int64", segmentsData, []int32{1})
|
|
assert.NotNil(t, err)
|
|
|
|
// failed to dispatch float data
|
|
chunkManager.readBuf["float"] = createBinlogBuf(t, schemapb.DataType_Float, fieldsData[107].([]float32))
|
|
err = adapter.readInsertlog(1, "float", segmentsData, []int32{1})
|
|
assert.NotNil(t, err)
|
|
|
|
// failed to dispatch double data
|
|
chunkManager.readBuf["double"] = createBinlogBuf(t, schemapb.DataType_Double, fieldsData[108].([]float64))
|
|
err = adapter.readInsertlog(1, "double", segmentsData, []int32{1})
|
|
assert.NotNil(t, err)
|
|
|
|
// failed to dispatch varchar data
|
|
chunkManager.readBuf["varchar"] = createBinlogBuf(t, schemapb.DataType_VarChar, fieldsData[109].([]string))
|
|
err = adapter.readInsertlog(1, "varchar", segmentsData, []int32{1})
|
|
assert.NotNil(t, err)
|
|
|
|
// failed to dispatch binvector data
|
|
chunkManager.readBuf["binvector"] = createBinlogBuf(t, schemapb.DataType_BinaryVector, fieldsData[110].([][]byte))
|
|
err = adapter.readInsertlog(1, "binvector", segmentsData, []int32{1})
|
|
assert.NotNil(t, err)
|
|
|
|
// failed to dispatch floatvector data
|
|
chunkManager.readBuf["floatvector"] = createBinlogBuf(t, schemapb.DataType_FloatVector, fieldsData[111].([][]float32))
|
|
err = adapter.readInsertlog(1, "floatvector", segmentsData, []int32{1})
|
|
assert.NotNil(t, err)
|
|
}
|