mirror of
https://gitee.com/milvus-io/milvus.git
synced 2024-11-30 10:59:32 +08:00
cd19d99ad7
Signed-off-by: bigsheeper <yihao.dai@zilliz.com> Signed-off-by: bigsheeper <yihao.dai@zilliz.com>
1106 lines
32 KiB
Go
1106 lines
32 KiB
Go
// Licensed to the LF AI & Data foundation under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package datacoord
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"path/filepath"
|
|
"strconv"
|
|
"strings"
|
|
"testing"
|
|
|
|
"github.com/milvus-io/milvus/internal/common"
|
|
|
|
"github.com/golang/protobuf/proto"
|
|
"github.com/milvus-io/milvus-proto/go-api/commonpb"
|
|
"github.com/milvus-io/milvus/internal/kv"
|
|
memkv "github.com/milvus-io/milvus/internal/kv/mem"
|
|
"github.com/milvus-io/milvus/internal/metastore/kv/datacoord"
|
|
"github.com/milvus-io/milvus/internal/proto/datapb"
|
|
"github.com/milvus-io/milvus/internal/proto/internalpb"
|
|
"github.com/milvus-io/milvus/internal/util"
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
)
|
|
|
|
type mockEtcdKv struct {
|
|
kv.TxnKV
|
|
}
|
|
|
|
func (mek *mockEtcdKv) LoadWithPrefix(key string) ([]string, []string, error) {
|
|
var val []byte
|
|
switch {
|
|
case strings.Contains(key, datacoord.SegmentPrefix):
|
|
segInfo := &datapb.SegmentInfo{ID: 1, Binlogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "log1")}}
|
|
val, _ = proto.Marshal(segInfo)
|
|
case strings.Contains(key, datacoord.SegmentBinlogPathPrefix):
|
|
segInfo := getFieldBinlogPaths(1, "binlog1")
|
|
val, _ = proto.Marshal(segInfo)
|
|
case strings.Contains(key, datacoord.SegmentDeltalogPathPrefix):
|
|
segInfo := getFieldBinlogPaths(1, "deltalog1")
|
|
val, _ = proto.Marshal(segInfo)
|
|
case strings.Contains(key, datacoord.SegmentStatslogPathPrefix):
|
|
segInfo := getFieldBinlogPaths(1, "statslog1")
|
|
val, _ = proto.Marshal(segInfo)
|
|
case strings.Contains(key, datacoord.ChannelCheckpointPrefix):
|
|
channelCP := &internalpb.MsgPosition{
|
|
Timestamp: 1000,
|
|
}
|
|
val, _ = proto.Marshal(channelCP)
|
|
default:
|
|
return nil, nil, fmt.Errorf("invalid key")
|
|
}
|
|
|
|
return nil, []string{string(val)}, nil
|
|
}
|
|
|
|
type mockKvLoadSegmentError struct {
|
|
kv.TxnKV
|
|
}
|
|
|
|
func (mek *mockKvLoadSegmentError) LoadWithPrefix(key string) ([]string, []string, error) {
|
|
if strings.Contains(key, datacoord.SegmentPrefix) {
|
|
return nil, nil, fmt.Errorf("segment LoadWithPrefix error")
|
|
}
|
|
return nil, nil, nil
|
|
}
|
|
|
|
type mockKvLoadBinlogError struct {
|
|
kv.TxnKV
|
|
}
|
|
|
|
func (mek *mockKvLoadBinlogError) LoadWithPrefix(key string) ([]string, []string, error) {
|
|
var val []byte
|
|
switch {
|
|
case strings.Contains(key, datacoord.SegmentPrefix):
|
|
segInfo := &datapb.SegmentInfo{ID: 1, Deltalogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "delta_log_1")}}
|
|
val, _ = proto.Marshal(segInfo)
|
|
case strings.Contains(key, datacoord.SegmentBinlogPathPrefix):
|
|
return nil, nil, fmt.Errorf("LoadWithPrefix for binlogs error")
|
|
}
|
|
return nil, []string{string(val)}, nil
|
|
}
|
|
|
|
type mockKvLoadDeltaBinlogError struct {
|
|
kv.TxnKV
|
|
}
|
|
|
|
func (mek *mockKvLoadDeltaBinlogError) LoadWithPrefix(key string) ([]string, []string, error) {
|
|
var val []byte
|
|
switch {
|
|
case strings.Contains(key, datacoord.SegmentPrefix):
|
|
segInfo := &datapb.SegmentInfo{ID: 1, Binlogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "binlog_1")}}
|
|
val, _ = proto.Marshal(segInfo)
|
|
case strings.Contains(key, datacoord.SegmentDeltalogPathPrefix):
|
|
return nil, nil, fmt.Errorf("LoadWithPrefix for deltalog error")
|
|
}
|
|
return nil, []string{string(val)}, nil
|
|
}
|
|
|
|
type mockKvLoadStatsBinlogError struct {
|
|
kv.TxnKV
|
|
}
|
|
|
|
func (mek *mockKvLoadStatsBinlogError) LoadWithPrefix(key string) ([]string, []string, error) {
|
|
var val []byte
|
|
switch {
|
|
case strings.Contains(key, datacoord.SegmentPrefix+"/"):
|
|
segInfo := &datapb.SegmentInfo{ID: 1, Binlogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "binlog_1")}}
|
|
val, _ = proto.Marshal(segInfo)
|
|
case strings.Contains(key, datacoord.SegmentStatslogPathPrefix):
|
|
return nil, nil, fmt.Errorf("LoadWithPrefix for statslog error")
|
|
}
|
|
return nil, []string{string(val)}, nil
|
|
}
|
|
|
|
type mockKvIllegalSegment struct {
|
|
kv.TxnKV
|
|
}
|
|
|
|
func (mek *mockKvIllegalSegment) LoadWithPrefix(key string) ([]string, []string, error) {
|
|
var val []byte
|
|
switch {
|
|
case strings.Contains(key, datacoord.SegmentPrefix):
|
|
val = []byte{'i', 'l', 'l', 'e', 'g', 'a', 'l'}
|
|
}
|
|
|
|
return nil, []string{string(val)}, nil
|
|
}
|
|
|
|
type mockKvIllegalBinlog struct {
|
|
kv.TxnKV
|
|
}
|
|
|
|
func (mek *mockKvIllegalBinlog) LoadWithPrefix(key string) ([]string, []string, error) {
|
|
var val []byte
|
|
switch {
|
|
case strings.Contains(key, datacoord.SegmentBinlogPathPrefix):
|
|
val = []byte{'i', 'l', 'l', 'e', 'g', 'a', 'l'}
|
|
}
|
|
|
|
return nil, []string{string(val)}, nil
|
|
}
|
|
|
|
type mockKvIllegalDeltalog struct {
|
|
kv.TxnKV
|
|
}
|
|
|
|
func (mek *mockKvIllegalDeltalog) LoadWithPrefix(key string) ([]string, []string, error) {
|
|
var val []byte
|
|
switch {
|
|
case strings.Contains(key, datacoord.SegmentDeltalogPathPrefix):
|
|
val = []byte{'i', 'l', 'l', 'e', 'g', 'a', 'l'}
|
|
}
|
|
|
|
return nil, []string{string(val)}, nil
|
|
}
|
|
|
|
type mockKvIllegalStatslog struct {
|
|
kv.TxnKV
|
|
}
|
|
|
|
func (mek *mockKvIllegalStatslog) LoadWithPrefix(key string) ([]string, []string, error) {
|
|
var val []byte
|
|
switch {
|
|
case strings.Contains(key, datacoord.SegmentStatslogPathPrefix):
|
|
val = []byte{'i', 'l', 'l', 'e', 'g', 'a', 'l'}
|
|
}
|
|
|
|
return nil, []string{string(val)}, nil
|
|
}
|
|
|
|
func TestMetaReloadFromKV(t *testing.T) {
|
|
t.Run("Test ReloadFromKV success", func(t *testing.T) {
|
|
fkv := &mockEtcdKv{}
|
|
_, err := newMeta(context.TODO(), fkv, "")
|
|
assert.Nil(t, err)
|
|
})
|
|
|
|
// load segment error
|
|
t.Run("Test ReloadFromKV load segment fails", func(t *testing.T) {
|
|
fkv := &mockKvLoadSegmentError{}
|
|
_, err := newMeta(context.TODO(), fkv, "")
|
|
assert.NotNil(t, err)
|
|
})
|
|
|
|
// illegal segment info
|
|
t.Run("Test ReloadFromKV unmarshal segment fails", func(t *testing.T) {
|
|
fkv := &mockKvIllegalSegment{}
|
|
_, err := newMeta(context.TODO(), fkv, "")
|
|
assert.NotNil(t, err)
|
|
})
|
|
|
|
// load binlog/deltalog/statslog error
|
|
t.Run("Test ReloadFromKV load binlog fails", func(t *testing.T) {
|
|
fkv := &mockKvLoadBinlogError{}
|
|
_, err := newMeta(context.TODO(), fkv, "")
|
|
assert.NotNil(t, err)
|
|
})
|
|
t.Run("Test ReloadFromKV load deltalog fails", func(t *testing.T) {
|
|
fkv := &mockKvLoadDeltaBinlogError{}
|
|
_, err := newMeta(context.TODO(), fkv, "")
|
|
assert.NotNil(t, err)
|
|
})
|
|
t.Run("Test ReloadFromKV load statslog fails", func(t *testing.T) {
|
|
fkv := &mockKvLoadStatsBinlogError{}
|
|
_, err := newMeta(context.TODO(), fkv, "")
|
|
assert.NotNil(t, err)
|
|
})
|
|
|
|
// illegal binlog/deltalog/statslog info
|
|
t.Run("Test ReloadFromKV unmarshal binlog fails", func(t *testing.T) {
|
|
fkv := &mockKvIllegalBinlog{}
|
|
_, err := newMeta(context.TODO(), fkv, "")
|
|
assert.NotNil(t, err)
|
|
})
|
|
t.Run("Test ReloadFromKV unmarshal deltalog fails", func(t *testing.T) {
|
|
fkv := &mockKvIllegalDeltalog{}
|
|
_, err := newMeta(context.TODO(), fkv, "")
|
|
assert.NotNil(t, err)
|
|
})
|
|
t.Run("Test ReloadFromKV unmarshal statslog fails", func(t *testing.T) {
|
|
fkv := &mockKvIllegalStatslog{}
|
|
_, err := newMeta(context.TODO(), fkv, "")
|
|
assert.NotNil(t, err)
|
|
})
|
|
}
|
|
|
|
func TestMeta_Basic(t *testing.T) {
|
|
const collID = UniqueID(0)
|
|
const partID0 = UniqueID(100)
|
|
const partID1 = UniqueID(101)
|
|
const channelName = "c1"
|
|
ctx := context.Background()
|
|
|
|
mockAllocator := newMockAllocator()
|
|
meta, err := newMemoryMeta()
|
|
assert.Nil(t, err)
|
|
|
|
testSchema := newTestSchema()
|
|
|
|
Params.Init()
|
|
|
|
collInfo := &collectionInfo{
|
|
ID: collID,
|
|
Schema: testSchema,
|
|
Partitions: []UniqueID{partID0, partID1},
|
|
StartPositions: []*commonpb.KeyDataPair{},
|
|
}
|
|
collInfoWoPartition := &collectionInfo{
|
|
ID: collID,
|
|
Schema: testSchema,
|
|
Partitions: []UniqueID{},
|
|
}
|
|
|
|
t.Run("Test Collection", func(t *testing.T) {
|
|
meta.AddCollection(collInfo)
|
|
// check has collection
|
|
collInfo := meta.GetCollection(collID)
|
|
assert.NotNil(t, collInfo)
|
|
|
|
// check partition info
|
|
assert.EqualValues(t, collID, collInfo.ID)
|
|
assert.EqualValues(t, testSchema, collInfo.Schema)
|
|
assert.EqualValues(t, 2, len(collInfo.Partitions))
|
|
assert.EqualValues(t, partID0, collInfo.Partitions[0])
|
|
assert.EqualValues(t, partID1, collInfo.Partitions[1])
|
|
})
|
|
|
|
t.Run("Test Segment", func(t *testing.T) {
|
|
meta.AddCollection(collInfoWoPartition)
|
|
// create seg0 for partition0, seg0/seg1 for partition1
|
|
segID0_0, err := mockAllocator.allocID(ctx)
|
|
assert.Nil(t, err)
|
|
segInfo0_0 := buildSegment(collID, partID0, segID0_0, channelName, true)
|
|
segID1_0, err := mockAllocator.allocID(ctx)
|
|
assert.Nil(t, err)
|
|
segInfo1_0 := buildSegment(collID, partID1, segID1_0, channelName, false)
|
|
segID1_1, err := mockAllocator.allocID(ctx)
|
|
assert.Nil(t, err)
|
|
segInfo1_1 := buildSegment(collID, partID1, segID1_1, channelName, false)
|
|
|
|
// check AddSegment
|
|
err = meta.AddSegment(segInfo0_0)
|
|
assert.Nil(t, err)
|
|
err = meta.AddSegment(segInfo1_0)
|
|
assert.Nil(t, err)
|
|
err = meta.AddSegment(segInfo1_1)
|
|
assert.Nil(t, err)
|
|
|
|
// check GetSegment
|
|
info0_0 := meta.GetSegment(segID0_0)
|
|
assert.NotNil(t, info0_0)
|
|
assert.True(t, proto.Equal(info0_0, segInfo0_0))
|
|
info1_0 := meta.GetSegment(segID1_0)
|
|
assert.NotNil(t, info1_0)
|
|
assert.True(t, proto.Equal(info1_0, segInfo1_0))
|
|
|
|
// check GetSegmentsOfCollection
|
|
segIDs := meta.GetSegmentsIDOfCollection(collID)
|
|
assert.EqualValues(t, 3, len(segIDs))
|
|
assert.Contains(t, segIDs, segID0_0)
|
|
assert.Contains(t, segIDs, segID1_0)
|
|
assert.Contains(t, segIDs, segID1_1)
|
|
|
|
// check GetSegmentsOfPartition
|
|
segIDs = meta.GetSegmentsIDOfPartition(collID, partID0)
|
|
assert.EqualValues(t, 1, len(segIDs))
|
|
assert.Contains(t, segIDs, segID0_0)
|
|
segIDs = meta.GetSegmentsIDOfPartition(collID, partID1)
|
|
assert.EqualValues(t, 2, len(segIDs))
|
|
assert.Contains(t, segIDs, segID1_0)
|
|
assert.Contains(t, segIDs, segID1_1)
|
|
|
|
// check DropSegment
|
|
err = meta.DropSegment(segID1_0)
|
|
assert.Nil(t, err)
|
|
segIDs = meta.GetSegmentsIDOfPartition(collID, partID1)
|
|
assert.EqualValues(t, 1, len(segIDs))
|
|
assert.Contains(t, segIDs, segID1_1)
|
|
|
|
err = meta.SetState(segID0_0, commonpb.SegmentState_Sealed)
|
|
assert.Nil(t, err)
|
|
err = meta.SetState(segID0_0, commonpb.SegmentState_Flushed)
|
|
assert.Nil(t, err)
|
|
|
|
info0_0 = meta.GetSegment(segID0_0)
|
|
assert.NotNil(t, info0_0)
|
|
assert.EqualValues(t, commonpb.SegmentState_Flushed, info0_0.State)
|
|
|
|
info0_0 = meta.GetSegment(segID0_0)
|
|
assert.NotNil(t, info0_0)
|
|
assert.Equal(t, true, info0_0.GetIsImporting())
|
|
err = meta.UnsetIsImporting(segID0_0)
|
|
assert.NoError(t, err)
|
|
info0_0 = meta.GetSegment(segID0_0)
|
|
assert.NotNil(t, info0_0)
|
|
assert.Equal(t, false, info0_0.GetIsImporting())
|
|
|
|
// UnsetIsImporting on segment that does not exist.
|
|
err = meta.UnsetIsImporting(segID1_0)
|
|
assert.Error(t, err)
|
|
|
|
info1_1 := meta.GetSegment(segID1_1)
|
|
assert.NotNil(t, info1_1)
|
|
assert.Equal(t, false, info1_1.GetIsImporting())
|
|
err = meta.UnsetIsImporting(segID1_1)
|
|
assert.NoError(t, err)
|
|
info1_1 = meta.GetSegment(segID1_1)
|
|
assert.NotNil(t, info1_1)
|
|
assert.Equal(t, false, info1_1.GetIsImporting())
|
|
|
|
})
|
|
|
|
t.Run("Test segment with kv fails", func(t *testing.T) {
|
|
// inject error for `Save`
|
|
memoryKV := memkv.NewMemoryKV()
|
|
fkv := &saveFailKV{TxnKV: memoryKV}
|
|
meta, err := newMeta(context.TODO(), fkv, "")
|
|
assert.Nil(t, err)
|
|
|
|
err = meta.AddSegment(NewSegmentInfo(&datapb.SegmentInfo{}))
|
|
assert.NotNil(t, err)
|
|
|
|
fkv2 := &removeFailKV{TxnKV: memoryKV}
|
|
meta, err = newMeta(context.TODO(), fkv2, "")
|
|
assert.Nil(t, err)
|
|
// nil, since no segment yet
|
|
err = meta.DropSegment(0)
|
|
assert.Nil(t, err)
|
|
// nil, since Save error not injected
|
|
err = meta.AddSegment(NewSegmentInfo(&datapb.SegmentInfo{}))
|
|
assert.Nil(t, err)
|
|
// error injected
|
|
err = meta.DropSegment(0)
|
|
assert.NotNil(t, err)
|
|
|
|
meta, err = newMeta(context.TODO(), fkv, "")
|
|
assert.Nil(t, err)
|
|
})
|
|
|
|
t.Run("Test GetCount", func(t *testing.T) {
|
|
const rowCount0 = 100
|
|
const rowCount1 = 300
|
|
|
|
// no segment
|
|
nums := meta.GetNumRowsOfCollection(collID)
|
|
assert.EqualValues(t, 0, nums)
|
|
|
|
// add seg1 with 100 rows
|
|
segID0, err := mockAllocator.allocID(ctx)
|
|
assert.Nil(t, err)
|
|
segInfo0 := buildSegment(collID, partID0, segID0, channelName, false)
|
|
segInfo0.NumOfRows = rowCount0
|
|
err = meta.AddSegment(segInfo0)
|
|
assert.Nil(t, err)
|
|
|
|
// add seg2 with 300 rows
|
|
segID1, err := mockAllocator.allocID(ctx)
|
|
assert.Nil(t, err)
|
|
segInfo1 := buildSegment(collID, partID0, segID1, channelName, false)
|
|
segInfo1.NumOfRows = rowCount1
|
|
err = meta.AddSegment(segInfo1)
|
|
assert.Nil(t, err)
|
|
|
|
// check partition/collection statistics
|
|
nums = meta.GetNumRowsOfPartition(collID, partID0)
|
|
assert.EqualValues(t, (rowCount0 + rowCount1), nums)
|
|
nums = meta.GetNumRowsOfCollection(collID)
|
|
assert.EqualValues(t, (rowCount0 + rowCount1), nums)
|
|
})
|
|
|
|
t.Run("Test GetSegmentsChanPart", func(t *testing.T) {
|
|
result := meta.GetSegmentsChanPart(func(*SegmentInfo) bool { return true })
|
|
assert.Equal(t, 2, len(result))
|
|
for _, entry := range result {
|
|
assert.Equal(t, "c1", entry.channelName)
|
|
if entry.partitionID == UniqueID(100) {
|
|
assert.Equal(t, 3, len(entry.segments))
|
|
}
|
|
if entry.partitionID == UniqueID(101) {
|
|
assert.Equal(t, 1, len(entry.segments))
|
|
}
|
|
}
|
|
result = meta.GetSegmentsChanPart(func(seg *SegmentInfo) bool { return seg.GetCollectionID() == 10 })
|
|
assert.Equal(t, 0, len(result))
|
|
})
|
|
|
|
t.Run("GetClonedCollectionInfo", func(t *testing.T) {
|
|
// collection does not exist
|
|
ret := meta.GetClonedCollectionInfo(-1)
|
|
assert.Nil(t, ret)
|
|
|
|
collInfo.Properties = map[string]string{
|
|
common.CollectionTTLConfigKey: "3600",
|
|
}
|
|
meta.AddCollection(collInfo)
|
|
ret = meta.GetClonedCollectionInfo(collInfo.ID)
|
|
equalCollectionInfo(t, collInfo, ret)
|
|
|
|
collInfo.StartPositions = []*commonpb.KeyDataPair{
|
|
{
|
|
Key: "k",
|
|
Data: []byte("v"),
|
|
},
|
|
}
|
|
meta.AddCollection(collInfo)
|
|
ret = meta.GetClonedCollectionInfo(collInfo.ID)
|
|
equalCollectionInfo(t, collInfo, ret)
|
|
})
|
|
|
|
t.Run("Test GetTotalBinlogSize", func(t *testing.T) {
|
|
const size0 = 1024
|
|
const size1 = 2048
|
|
|
|
// no binlog
|
|
size := meta.GetTotalBinlogSize()
|
|
assert.EqualValues(t, 0, size)
|
|
|
|
// add seg0 with size0
|
|
segID0, err := mockAllocator.allocID(ctx)
|
|
assert.Nil(t, err)
|
|
segInfo0 := buildSegment(collID, partID0, segID0, channelName, false)
|
|
segInfo0.size = size0
|
|
err = meta.AddSegment(segInfo0)
|
|
assert.Nil(t, err)
|
|
|
|
// add seg1 with size1
|
|
segID1, err := mockAllocator.allocID(ctx)
|
|
assert.Nil(t, err)
|
|
segInfo1 := buildSegment(collID, partID0, segID1, channelName, false)
|
|
segInfo1.size = size1
|
|
err = meta.AddSegment(segInfo1)
|
|
assert.Nil(t, err)
|
|
|
|
// check TotalBinlogSize
|
|
size = meta.GetTotalBinlogSize()
|
|
assert.Equal(t, int64(size0+size1), size)
|
|
})
|
|
}
|
|
|
|
func TestGetUnFlushedSegments(t *testing.T) {
|
|
meta, err := newMemoryMeta()
|
|
assert.Nil(t, err)
|
|
s1 := &datapb.SegmentInfo{
|
|
ID: 0,
|
|
CollectionID: 0,
|
|
PartitionID: 0,
|
|
State: commonpb.SegmentState_Growing,
|
|
}
|
|
err = meta.AddSegment(NewSegmentInfo(s1))
|
|
assert.Nil(t, err)
|
|
s2 := &datapb.SegmentInfo{
|
|
ID: 1,
|
|
CollectionID: 0,
|
|
PartitionID: 0,
|
|
State: commonpb.SegmentState_Flushed,
|
|
}
|
|
err = meta.AddSegment(NewSegmentInfo(s2))
|
|
assert.Nil(t, err)
|
|
|
|
segments := meta.GetUnFlushedSegments()
|
|
assert.Nil(t, err)
|
|
|
|
assert.EqualValues(t, 1, len(segments))
|
|
assert.EqualValues(t, 0, segments[0].ID)
|
|
assert.NotEqualValues(t, commonpb.SegmentState_Flushed, segments[0].State)
|
|
}
|
|
|
|
func TestUpdateFlushSegmentsInfo(t *testing.T) {
|
|
t.Run("normal", func(t *testing.T) {
|
|
meta, err := newMeta(context.TODO(), memkv.NewMemoryKV(), "")
|
|
assert.Nil(t, err)
|
|
|
|
segment1 := &SegmentInfo{SegmentInfo: &datapb.SegmentInfo{ID: 1, State: commonpb.SegmentState_Growing, Binlogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "binlog0")},
|
|
Statslogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "statslog0")}}}
|
|
err = meta.AddSegment(segment1)
|
|
assert.Nil(t, err)
|
|
|
|
err = meta.UpdateFlushSegmentsInfo(1, true, false, true, []*datapb.FieldBinlog{getFieldBinlogPaths(1, "binlog1")},
|
|
[]*datapb.FieldBinlog{getFieldBinlogPaths(1, "statslog1")},
|
|
[]*datapb.FieldBinlog{{Binlogs: []*datapb.Binlog{{EntriesNum: 1, TimestampFrom: 100, TimestampTo: 200, LogSize: 1000}}}},
|
|
[]*datapb.CheckPoint{{SegmentID: 1, NumOfRows: 10}}, []*datapb.SegmentStartPosition{{SegmentID: 1, StartPosition: &internalpb.MsgPosition{MsgID: []byte{1, 2, 3}}}})
|
|
assert.Nil(t, err)
|
|
|
|
updated := meta.GetSegment(1)
|
|
expected := &SegmentInfo{SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 1, State: commonpb.SegmentState_Flushing, NumOfRows: 10,
|
|
StartPosition: &internalpb.MsgPosition{MsgID: []byte{1, 2, 3}},
|
|
Binlogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "binlog0", "binlog1")},
|
|
Statslogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "statslog0", "statslog1")},
|
|
Deltalogs: []*datapb.FieldBinlog{{Binlogs: []*datapb.Binlog{{EntriesNum: 1, TimestampFrom: 100, TimestampTo: 200, LogSize: 1000}}}},
|
|
}}
|
|
|
|
fmt.Println(updated)
|
|
fmt.Println(expected)
|
|
assert.Equal(t, updated.StartPosition, expected.StartPosition)
|
|
assert.Equal(t, updated.DmlPosition, expected.DmlPosition)
|
|
assert.Equal(t, updated.DmlPosition, expected.DmlPosition)
|
|
assert.Equal(t, len(updated.Binlogs[0].Binlogs), len(expected.Binlogs[0].Binlogs))
|
|
assert.Equal(t, len(updated.Statslogs[0].Binlogs), len(expected.Statslogs[0].Binlogs))
|
|
assert.Equal(t, len(updated.Deltalogs[0].Binlogs), len(expected.Deltalogs[0].Binlogs))
|
|
assert.Equal(t, updated.State, expected.State)
|
|
assert.Equal(t, updated.size, expected.size)
|
|
assert.Equal(t, updated.NumOfRows, expected.NumOfRows)
|
|
|
|
})
|
|
|
|
t.Run("update non-existed segment", func(t *testing.T) {
|
|
meta, err := newMeta(context.TODO(), memkv.NewMemoryKV(), "")
|
|
assert.Nil(t, err)
|
|
|
|
err = meta.UpdateFlushSegmentsInfo(1, false, false, false, nil, nil, nil, nil, nil)
|
|
assert.Nil(t, err)
|
|
})
|
|
|
|
t.Run("update checkpoints and start position of non existed segment", func(t *testing.T) {
|
|
meta, err := newMeta(context.TODO(), memkv.NewMemoryKV(), "")
|
|
assert.Nil(t, err)
|
|
|
|
segment1 := &SegmentInfo{SegmentInfo: &datapb.SegmentInfo{ID: 1, State: commonpb.SegmentState_Growing}}
|
|
err = meta.AddSegment(segment1)
|
|
assert.Nil(t, err)
|
|
|
|
err = meta.UpdateFlushSegmentsInfo(1, false, false, false, nil, nil, nil, []*datapb.CheckPoint{{SegmentID: 2, NumOfRows: 10}},
|
|
|
|
[]*datapb.SegmentStartPosition{{SegmentID: 2, StartPosition: &internalpb.MsgPosition{MsgID: []byte{1, 2, 3}}}})
|
|
assert.Nil(t, err)
|
|
assert.Nil(t, meta.GetSegment(2))
|
|
})
|
|
|
|
t.Run("test save etcd failed", func(t *testing.T) {
|
|
kv := memkv.NewMemoryKV()
|
|
failedKv := &saveFailKV{kv}
|
|
meta, err := newMeta(context.TODO(), failedKv, "")
|
|
assert.Nil(t, err)
|
|
|
|
segmentInfo := &SegmentInfo{
|
|
SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 1,
|
|
NumOfRows: 0,
|
|
State: commonpb.SegmentState_Growing,
|
|
},
|
|
}
|
|
meta.segments.SetSegment(1, segmentInfo)
|
|
|
|
err = meta.UpdateFlushSegmentsInfo(1, true, false, false, []*datapb.FieldBinlog{getFieldBinlogPaths(1, "binlog")},
|
|
[]*datapb.FieldBinlog{getFieldBinlogPaths(1, "statslog")},
|
|
[]*datapb.FieldBinlog{{Binlogs: []*datapb.Binlog{{EntriesNum: 1, TimestampFrom: 100, TimestampTo: 200, LogSize: 1000}}}},
|
|
[]*datapb.CheckPoint{{SegmentID: 1, NumOfRows: 10}}, []*datapb.SegmentStartPosition{{SegmentID: 1, StartPosition: &internalpb.MsgPosition{MsgID: []byte{1, 2, 3}}}})
|
|
assert.NotNil(t, err)
|
|
assert.Equal(t, "mocked fail", err.Error())
|
|
segmentInfo = meta.GetSegment(1)
|
|
assert.EqualValues(t, 0, segmentInfo.NumOfRows)
|
|
assert.Equal(t, commonpb.SegmentState_Growing, segmentInfo.State)
|
|
assert.Nil(t, segmentInfo.Binlogs)
|
|
assert.Nil(t, segmentInfo.StartPosition)
|
|
})
|
|
}
|
|
|
|
func TestSaveHandoffMeta(t *testing.T) {
|
|
kvClient := memkv.NewMemoryKV()
|
|
meta, err := newMeta(context.TODO(), kvClient, "")
|
|
assert.Nil(t, err)
|
|
|
|
info := &datapb.SegmentInfo{
|
|
ID: 100,
|
|
State: commonpb.SegmentState_Flushing,
|
|
}
|
|
segmentInfo := &SegmentInfo{
|
|
SegmentInfo: info,
|
|
}
|
|
|
|
err = meta.catalog.AddSegment(context.TODO(), segmentInfo.SegmentInfo)
|
|
assert.Nil(t, err)
|
|
|
|
keys, _, err := kvClient.LoadWithPrefix(util.FlushedSegmentPrefix)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 0, len(keys))
|
|
|
|
newInfo := &datapb.SegmentInfo{
|
|
ID: 100,
|
|
State: commonpb.SegmentState_Flushed,
|
|
}
|
|
|
|
err = meta.catalog.AlterSegment(context.TODO(), newInfo, segmentInfo.SegmentInfo)
|
|
assert.Nil(t, err)
|
|
|
|
keys, _, err = kvClient.LoadWithPrefix(util.FlushedSegmentPrefix)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 1, len(keys))
|
|
|
|
segmentID, err := strconv.ParseInt(filepath.Base(keys[0]), 10, 64)
|
|
assert.Nil(t, err)
|
|
assert.Equal(t, 100, int(segmentID))
|
|
}
|
|
|
|
func TestMeta_alterMetaStore(t *testing.T) {
|
|
toAlter := []*datapb.SegmentInfo{
|
|
{
|
|
CollectionID: 100,
|
|
PartitionID: 10,
|
|
ID: 1,
|
|
NumOfRows: 10,
|
|
},
|
|
}
|
|
|
|
newSeg := &datapb.SegmentInfo{
|
|
Binlogs: []*datapb.FieldBinlog{
|
|
{
|
|
FieldID: 101,
|
|
Binlogs: []*datapb.Binlog{},
|
|
},
|
|
},
|
|
Deltalogs: []*datapb.FieldBinlog{
|
|
{
|
|
FieldID: 101,
|
|
Binlogs: []*datapb.Binlog{},
|
|
},
|
|
},
|
|
CollectionID: 100,
|
|
PartitionID: 10,
|
|
ID: 2,
|
|
NumOfRows: 15,
|
|
}
|
|
|
|
m := &meta{
|
|
catalog: &datacoord.Catalog{Txn: memkv.NewMemoryKV()},
|
|
segments: &SegmentsInfo{map[int64]*SegmentInfo{
|
|
1: {SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 1,
|
|
Binlogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "log1", "log2")},
|
|
Statslogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "statlog1", "statlog2")},
|
|
Deltalogs: []*datapb.FieldBinlog{getFieldBinlogPaths(0, "deltalog1", "deltalog2")},
|
|
}},
|
|
}},
|
|
}
|
|
|
|
err := m.alterMetaStoreAfterCompaction(toAlter, newSeg)
|
|
assert.NoError(t, err)
|
|
|
|
err = m.revertAlterMetaStoreAfterCompaction(toAlter, newSeg)
|
|
assert.NoError(t, err)
|
|
}
|
|
|
|
func TestMeta_alterInMemoryMetaAfterCompaction(t *testing.T) {
|
|
m := &meta{
|
|
catalog: &datacoord.Catalog{Txn: memkv.NewMemoryKV()},
|
|
segments: &SegmentsInfo{make(map[UniqueID]*SegmentInfo)},
|
|
}
|
|
|
|
tests := []struct {
|
|
description string
|
|
compactToSeg *SegmentInfo
|
|
}{
|
|
{
|
|
"numRows>0", &SegmentInfo{
|
|
SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 1,
|
|
NumOfRows: 10,
|
|
},
|
|
},
|
|
},
|
|
{
|
|
"numRows=0", &SegmentInfo{
|
|
SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 1,
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
compactFrom := []*SegmentInfo{{}, {}}
|
|
|
|
for _, test := range tests {
|
|
t.Run(test.description, func(t *testing.T) {
|
|
m.alterInMemoryMetaAfterCompaction(test.compactToSeg, compactFrom)
|
|
})
|
|
}
|
|
|
|
}
|
|
|
|
func TestMeta_PrepareCompleteCompactionMutation(t *testing.T) {
|
|
prepareSegments := &SegmentsInfo{
|
|
map[UniqueID]*SegmentInfo{
|
|
1: {SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 1,
|
|
CollectionID: 100,
|
|
PartitionID: 10,
|
|
State: commonpb.SegmentState_Flushed,
|
|
Binlogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "log1", "log2")},
|
|
Statslogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "statlog1", "statlog2")},
|
|
Deltalogs: []*datapb.FieldBinlog{getFieldBinlogPaths(0, "deltalog1", "deltalog2")},
|
|
}},
|
|
2: {SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 2,
|
|
CollectionID: 100,
|
|
PartitionID: 10,
|
|
State: commonpb.SegmentState_Flushed,
|
|
Binlogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "log3", "log4")},
|
|
Statslogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "statlog3", "statlog4")},
|
|
Deltalogs: []*datapb.FieldBinlog{getFieldBinlogPaths(0, "deltalog3", "deltalog4")},
|
|
}},
|
|
},
|
|
}
|
|
|
|
m := &meta{
|
|
catalog: &datacoord.Catalog{Txn: memkv.NewMemoryKV()},
|
|
segments: prepareSegments,
|
|
}
|
|
|
|
inCompactionLogs := []*datapb.CompactionSegmentBinlogs{
|
|
{
|
|
SegmentID: 1,
|
|
FieldBinlogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "log1", "log2")},
|
|
Field2StatslogPaths: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "statlog1", "statlog2")},
|
|
Deltalogs: []*datapb.FieldBinlog{getFieldBinlogPaths(0, "deltalog1", "deltalog2")},
|
|
},
|
|
{
|
|
SegmentID: 2,
|
|
FieldBinlogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "log3", "log4")},
|
|
Field2StatslogPaths: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "statlog3", "statlog4")},
|
|
Deltalogs: []*datapb.FieldBinlog{getFieldBinlogPaths(0, "deltalog3", "deltalog4")},
|
|
},
|
|
}
|
|
|
|
inCompactionResult := &datapb.CompactionResult{
|
|
SegmentID: 3,
|
|
InsertLogs: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "log5")},
|
|
Field2StatslogPaths: []*datapb.FieldBinlog{getFieldBinlogPaths(1, "statlog5")},
|
|
Deltalogs: []*datapb.FieldBinlog{getFieldBinlogPaths(0, "deltalog5")},
|
|
NumOfRows: 1,
|
|
}
|
|
beforeCompact, afterCompact, newSegment := m.PrepareCompleteCompactionMutation(inCompactionLogs, inCompactionResult)
|
|
assert.NotNil(t, beforeCompact)
|
|
assert.NotNil(t, afterCompact)
|
|
assert.NotNil(t, newSegment)
|
|
|
|
require.Equal(t, 2, len(beforeCompact))
|
|
assert.Equal(t, commonpb.SegmentState_Flushed, beforeCompact[0].GetState())
|
|
assert.Equal(t, commonpb.SegmentState_Flushed, beforeCompact[1].GetState())
|
|
assert.Zero(t, beforeCompact[0].GetDroppedAt())
|
|
assert.Zero(t, beforeCompact[1].GetDroppedAt())
|
|
|
|
require.Equal(t, 2, len(afterCompact))
|
|
assert.Equal(t, commonpb.SegmentState_Dropped, afterCompact[0].GetState())
|
|
assert.Equal(t, commonpb.SegmentState_Dropped, afterCompact[1].GetState())
|
|
assert.NotZero(t, afterCompact[0].GetDroppedAt())
|
|
assert.NotZero(t, afterCompact[1].GetDroppedAt())
|
|
|
|
assert.Equal(t, inCompactionResult.SegmentID, newSegment.GetID())
|
|
assert.Equal(t, UniqueID(100), newSegment.GetCollectionID())
|
|
assert.Equal(t, UniqueID(10), newSegment.GetPartitionID())
|
|
assert.Equal(t, inCompactionResult.NumOfRows, newSegment.GetNumOfRows())
|
|
assert.Equal(t, commonpb.SegmentState_Flushing, newSegment.GetState())
|
|
|
|
assert.EqualValues(t, inCompactionResult.GetInsertLogs(), newSegment.GetBinlogs())
|
|
assert.EqualValues(t, inCompactionResult.GetField2StatslogPaths(), newSegment.GetStatslogs())
|
|
assert.EqualValues(t, inCompactionResult.GetDeltalogs(), newSegment.GetDeltalogs())
|
|
assert.NotZero(t, newSegment.lastFlushTime)
|
|
}
|
|
|
|
func Test_meta_SetSegmentCompacting(t *testing.T) {
|
|
type fields struct {
|
|
client kv.TxnKV
|
|
segments *SegmentsInfo
|
|
}
|
|
type args struct {
|
|
segmentID UniqueID
|
|
compacting bool
|
|
}
|
|
tests := []struct {
|
|
name string
|
|
fields fields
|
|
args args
|
|
}{
|
|
{
|
|
"test set segment compacting",
|
|
fields{
|
|
memkv.NewMemoryKV(),
|
|
&SegmentsInfo{
|
|
map[int64]*SegmentInfo{
|
|
1: {
|
|
SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 1,
|
|
State: commonpb.SegmentState_Flushed,
|
|
},
|
|
isCompacting: false,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
args{
|
|
segmentID: 1,
|
|
compacting: true,
|
|
},
|
|
},
|
|
}
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
m := &meta{
|
|
catalog: &datacoord.Catalog{Txn: tt.fields.client},
|
|
segments: tt.fields.segments,
|
|
}
|
|
m.SetSegmentCompacting(tt.args.segmentID, tt.args.compacting)
|
|
segment := m.GetSegment(tt.args.segmentID)
|
|
assert.Equal(t, tt.args.compacting, segment.isCompacting)
|
|
})
|
|
}
|
|
}
|
|
|
|
func Test_meta_SetSegmentImporting(t *testing.T) {
|
|
type fields struct {
|
|
client kv.TxnKV
|
|
segments *SegmentsInfo
|
|
}
|
|
type args struct {
|
|
segmentID UniqueID
|
|
importing bool
|
|
}
|
|
tests := []struct {
|
|
name string
|
|
fields fields
|
|
args args
|
|
}{
|
|
{
|
|
"test set segment importing",
|
|
fields{
|
|
memkv.NewMemoryKV(),
|
|
&SegmentsInfo{
|
|
map[int64]*SegmentInfo{
|
|
1: {
|
|
SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 1,
|
|
State: commonpb.SegmentState_Flushed,
|
|
IsImporting: false,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
args{
|
|
segmentID: 1,
|
|
importing: true,
|
|
},
|
|
},
|
|
}
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
m := &meta{
|
|
catalog: &datacoord.Catalog{Txn: tt.fields.client},
|
|
segments: tt.fields.segments,
|
|
}
|
|
m.SetSegmentCompacting(tt.args.segmentID, tt.args.importing)
|
|
segment := m.GetSegment(tt.args.segmentID)
|
|
assert.Equal(t, tt.args.importing, segment.isCompacting)
|
|
})
|
|
}
|
|
}
|
|
|
|
func Test_meta_GetSegmentsOfCollection(t *testing.T) {
|
|
type fields struct {
|
|
segments *SegmentsInfo
|
|
}
|
|
type args struct {
|
|
collectionID UniqueID
|
|
}
|
|
tests := []struct {
|
|
name string
|
|
fields fields
|
|
args args
|
|
expect []*SegmentInfo
|
|
}{
|
|
{
|
|
"test get segments",
|
|
fields{
|
|
&SegmentsInfo{
|
|
map[int64]*SegmentInfo{
|
|
1: {
|
|
SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 1,
|
|
CollectionID: 1,
|
|
State: commonpb.SegmentState_Flushed,
|
|
},
|
|
},
|
|
2: {
|
|
SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 2,
|
|
CollectionID: 1,
|
|
State: commonpb.SegmentState_Growing,
|
|
},
|
|
},
|
|
3: {
|
|
SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 3,
|
|
CollectionID: 2,
|
|
State: commonpb.SegmentState_Flushed,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
args{
|
|
collectionID: 1,
|
|
},
|
|
[]*SegmentInfo{
|
|
{
|
|
SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 1,
|
|
CollectionID: 1,
|
|
State: commonpb.SegmentState_Flushed,
|
|
},
|
|
},
|
|
{
|
|
SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 2,
|
|
CollectionID: 1,
|
|
State: commonpb.SegmentState_Growing,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
m := &meta{
|
|
segments: tt.fields.segments,
|
|
}
|
|
got := m.GetSegmentsOfCollection(tt.args.collectionID)
|
|
assert.ElementsMatch(t, tt.expect, got)
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestMeta_HasSegments(t *testing.T) {
|
|
m := &meta{
|
|
segments: &SegmentsInfo{
|
|
segments: map[UniqueID]*SegmentInfo{
|
|
1: {
|
|
SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 1,
|
|
},
|
|
currRows: 100,
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
has, err := m.HasSegments([]UniqueID{1})
|
|
assert.Equal(t, true, has)
|
|
assert.NoError(t, err)
|
|
|
|
has, err = m.HasSegments([]UniqueID{2})
|
|
assert.Equal(t, false, has)
|
|
assert.Error(t, err)
|
|
}
|
|
|
|
func TestMeta_GetAllSegments(t *testing.T) {
|
|
m := &meta{
|
|
segments: &SegmentsInfo{
|
|
segments: map[UniqueID]*SegmentInfo{
|
|
1: {
|
|
SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 1,
|
|
State: commonpb.SegmentState_Growing,
|
|
},
|
|
},
|
|
2: {
|
|
SegmentInfo: &datapb.SegmentInfo{
|
|
ID: 2,
|
|
State: commonpb.SegmentState_Dropped,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
seg1 := m.GetSegment(1)
|
|
seg1All := m.GetSegmentUnsafe(1)
|
|
seg2 := m.GetSegment(2)
|
|
seg2All := m.GetSegmentUnsafe(2)
|
|
assert.NotNil(t, seg1)
|
|
assert.NotNil(t, seg1All)
|
|
assert.Nil(t, seg2)
|
|
assert.NotNil(t, seg2All)
|
|
}
|
|
|
|
func TestMeta_isSegmentHealthy_issue17823_panic(t *testing.T) {
|
|
var seg *SegmentInfo
|
|
|
|
assert.False(t, isSegmentHealthy(seg))
|
|
}
|
|
|
|
func equalCollectionInfo(t *testing.T, a *collectionInfo, b *collectionInfo) {
|
|
assert.Equal(t, a.ID, b.ID)
|
|
assert.Equal(t, a.Partitions, b.Partitions)
|
|
assert.Equal(t, a.Schema, b.Schema)
|
|
assert.Equal(t, a.Properties, b.Properties)
|
|
assert.Equal(t, a.StartPositions, b.StartPositions)
|
|
}
|
|
|
|
func TestChannelCP(t *testing.T) {
|
|
mockVChannel := "fake-by-dev-rootcoord-dml-1-testchannelcp-v0"
|
|
mockPChannel := "fake-by-dev-rootcoord-dml-1"
|
|
|
|
pos := &internalpb.MsgPosition{
|
|
ChannelName: mockPChannel,
|
|
MsgID: []byte{},
|
|
Timestamp: 1000,
|
|
}
|
|
|
|
t.Run("UpdateChannelCheckpoint", func(t *testing.T) {
|
|
meta, err := newMeta(context.TODO(), memkv.NewMemoryKV(), "")
|
|
assert.NoError(t, err)
|
|
|
|
// nil position
|
|
err = meta.UpdateChannelCheckpoint(mockVChannel, nil)
|
|
assert.Error(t, err)
|
|
|
|
err = meta.UpdateChannelCheckpoint(mockVChannel, pos)
|
|
assert.NoError(t, err)
|
|
})
|
|
|
|
t.Run("GetChannelCheckpoint", func(t *testing.T) {
|
|
meta, err := newMeta(context.TODO(), memkv.NewMemoryKV(), "")
|
|
assert.NoError(t, err)
|
|
|
|
position := meta.GetChannelCheckpoint(mockVChannel)
|
|
assert.Nil(t, position)
|
|
|
|
err = meta.UpdateChannelCheckpoint(mockVChannel, pos)
|
|
assert.NoError(t, err)
|
|
position = meta.GetChannelCheckpoint(mockVChannel)
|
|
assert.NotNil(t, position)
|
|
assert.True(t, position.ChannelName == pos.ChannelName)
|
|
assert.True(t, position.Timestamp == pos.Timestamp)
|
|
})
|
|
|
|
t.Run("DropChannelCheckpoint", func(t *testing.T) {
|
|
meta, err := newMeta(context.TODO(), memkv.NewMemoryKV(), "")
|
|
assert.NoError(t, err)
|
|
|
|
err = meta.DropChannelCheckpoint(mockVChannel)
|
|
assert.NoError(t, err)
|
|
|
|
err = meta.UpdateChannelCheckpoint(mockVChannel, pos)
|
|
assert.NoError(t, err)
|
|
err = meta.DropChannelCheckpoint(mockVChannel)
|
|
assert.NoError(t, err)
|
|
})
|
|
}
|