mirror of
https://gitee.com/milvus-io/milvus.git
synced 2024-12-05 05:18:52 +08:00
29c72cad15
Signed-off-by: Congqi Xia <congqi.xia@zilliz.com> Signed-off-by: Congqi Xia <congqi.xia@zilliz.com>
994 lines
30 KiB
Go
994 lines
30 KiB
Go
// Licensed to the LF AI & Data foundation under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package datanode
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"fmt"
|
|
"math/rand"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/bits-and-blooms/bloom/v3"
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
"github.com/stretchr/testify/suite"
|
|
|
|
"github.com/milvus-io/milvus/internal/common"
|
|
"github.com/milvus-io/milvus/internal/proto/datapb"
|
|
"github.com/milvus-io/milvus/internal/proto/internalpb"
|
|
"github.com/milvus-io/milvus/internal/proto/schemapb"
|
|
"github.com/milvus-io/milvus/internal/storage"
|
|
)
|
|
|
|
var segmentReplicaNodeTestDir = "/tmp/milvus_test/segment_replica"
|
|
|
|
func TestNewReplica(t *testing.T) {
|
|
rc := &RootCoordFactory{}
|
|
cm := storage.NewLocalChunkManager(storage.RootPath(segmentReplicaNodeTestDir))
|
|
defer cm.RemoveWithPrefix("")
|
|
replica, err := newReplica(context.Background(), rc, cm, 0)
|
|
assert.Nil(t, err)
|
|
assert.NotNil(t, replica)
|
|
}
|
|
|
|
type mockDataCM struct {
|
|
storage.ChunkManager
|
|
}
|
|
|
|
func (kv *mockDataCM) MultiRead(keys []string) ([][]byte, error) {
|
|
stats := &storage.PrimaryKeyStats{
|
|
FieldID: common.RowIDField,
|
|
Min: 0,
|
|
Max: 10,
|
|
BF: bloom.NewWithEstimates(bloomFilterSize, maxBloomFalsePositive),
|
|
}
|
|
buffer, _ := json.Marshal(stats)
|
|
return [][]byte{buffer}, nil
|
|
}
|
|
|
|
type mockPkfilterMergeError struct {
|
|
storage.ChunkManager
|
|
}
|
|
|
|
func (kv *mockPkfilterMergeError) MultiRead(keys []string) ([][]byte, error) {
|
|
stats := &storage.PrimaryKeyStats{
|
|
FieldID: common.RowIDField,
|
|
Min: 0,
|
|
Max: 10,
|
|
BF: bloom.NewWithEstimates(1, 0.0001),
|
|
}
|
|
buffer, _ := json.Marshal(stats)
|
|
return [][]byte{buffer}, nil
|
|
}
|
|
|
|
type mockDataCMError struct {
|
|
storage.ChunkManager
|
|
}
|
|
|
|
func (kv *mockDataCMError) MultiRead(keys []string) ([][]byte, error) {
|
|
return nil, fmt.Errorf("mock error")
|
|
}
|
|
|
|
type mockDataCMStatsError struct {
|
|
storage.ChunkManager
|
|
}
|
|
|
|
func (kv *mockDataCMStatsError) MultiRead(keys []string) ([][]byte, error) {
|
|
return [][]byte{[]byte("3123123,error,test")}, nil
|
|
}
|
|
|
|
func getSimpleFieldBinlog() *datapb.FieldBinlog {
|
|
return &datapb.FieldBinlog{
|
|
FieldID: 106,
|
|
Binlogs: []*datapb.Binlog{{LogPath: "test"}},
|
|
}
|
|
}
|
|
|
|
func TestSegmentReplica_getCollectionAndPartitionID(te *testing.T) {
|
|
tests := []struct {
|
|
segInNew UniqueID
|
|
segInNormal UniqueID
|
|
segInFlushed UniqueID
|
|
|
|
inCollID UniqueID
|
|
inParID UniqueID
|
|
description string
|
|
}{
|
|
{100, 0, 0, 1, 10, "Segment 100 in NewSegments"},
|
|
{0, 200, 0, 2, 20, "Segment 200 in NormalSegments"},
|
|
{0, 0, 300, 3, 30, "Segment 300 in FlushedSegments"},
|
|
{0, 0, 0, 4, 40, "No Segment in replica"},
|
|
}
|
|
|
|
for _, test := range tests {
|
|
te.Run(test.description, func(t *testing.T) {
|
|
if test.segInNew != 0 {
|
|
sr := &SegmentReplica{
|
|
newSegments: map[UniqueID]*Segment{
|
|
test.segInNew: {
|
|
collectionID: test.inCollID,
|
|
partitionID: test.inParID,
|
|
segmentID: test.segInNew,
|
|
}},
|
|
}
|
|
|
|
collID, parID, err := sr.getCollectionAndPartitionID(test.segInNew)
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, test.inCollID, collID)
|
|
assert.Equal(t, test.inParID, parID)
|
|
} else if test.segInNormal != 0 {
|
|
sr := &SegmentReplica{
|
|
normalSegments: map[UniqueID]*Segment{
|
|
test.segInNormal: {
|
|
collectionID: test.inCollID,
|
|
partitionID: test.inParID,
|
|
segmentID: test.segInNormal,
|
|
}},
|
|
}
|
|
|
|
collID, parID, err := sr.getCollectionAndPartitionID(test.segInNormal)
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, test.inCollID, collID)
|
|
assert.Equal(t, test.inParID, parID)
|
|
} else if test.segInFlushed != 0 {
|
|
sr := &SegmentReplica{
|
|
flushedSegments: map[UniqueID]*Segment{
|
|
test.segInFlushed: {
|
|
collectionID: test.inCollID,
|
|
partitionID: test.inParID,
|
|
segmentID: test.segInFlushed,
|
|
}},
|
|
}
|
|
|
|
collID, parID, err := sr.getCollectionAndPartitionID(test.segInFlushed)
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, test.inCollID, collID)
|
|
assert.Equal(t, test.inParID, parID)
|
|
} else {
|
|
sr := &SegmentReplica{}
|
|
collID, parID, err := sr.getCollectionAndPartitionID(1000)
|
|
assert.Error(t, err)
|
|
assert.Zero(t, collID)
|
|
assert.Zero(t, parID)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestSegmentReplica(t *testing.T) {
|
|
rc := &RootCoordFactory{
|
|
pkType: schemapb.DataType_Int64,
|
|
}
|
|
collID := UniqueID(1)
|
|
cm := storage.NewLocalChunkManager(storage.RootPath(segmentReplicaNodeTestDir))
|
|
defer cm.RemoveWithPrefix("")
|
|
|
|
t.Run("Test coll mot match", func(t *testing.T) {
|
|
replica, err := newReplica(context.Background(), rc, cm, collID)
|
|
assert.Nil(t, err)
|
|
|
|
err = replica.addNewSegment(1, collID+1, 0, "", nil, nil)
|
|
assert.NotNil(t, err)
|
|
})
|
|
|
|
t.Run("Test segmentFlushed", func(t *testing.T) {
|
|
testReplica := &SegmentReplica{
|
|
newSegments: make(map[UniqueID]*Segment),
|
|
normalSegments: make(map[UniqueID]*Segment),
|
|
flushedSegments: make(map[UniqueID]*Segment),
|
|
}
|
|
|
|
type Test struct {
|
|
inisNew bool
|
|
inisFlushed bool
|
|
inSegID UniqueID
|
|
|
|
expectedisNew bool
|
|
expectedisFlushed bool
|
|
expectedSegID UniqueID
|
|
}
|
|
|
|
tests := []Test{
|
|
// new segment
|
|
{true, false, 1, false, true, 1},
|
|
{true, false, 2, false, true, 2},
|
|
{true, false, 3, false, true, 3},
|
|
// normal segment
|
|
{false, false, 10, false, true, 10},
|
|
{false, false, 20, false, true, 20},
|
|
{false, false, 30, false, true, 30},
|
|
// flushed segment
|
|
{false, true, 100, false, true, 100},
|
|
{false, true, 200, false, true, 200},
|
|
{false, true, 300, false, true, 300},
|
|
}
|
|
|
|
newSeg := func(sr *SegmentReplica, isNew, isFlushed bool, id UniqueID) {
|
|
ns := &Segment{segmentID: id}
|
|
ns.isNew.Store(isNew)
|
|
ns.isFlushed.Store(isFlushed)
|
|
|
|
if isNew && !isFlushed {
|
|
sr.newSegments[id] = ns
|
|
return
|
|
}
|
|
|
|
if !isNew && !isFlushed {
|
|
sr.normalSegments[id] = ns
|
|
return
|
|
}
|
|
|
|
if !isNew && isFlushed {
|
|
sr.flushedSegments[id] = ns
|
|
return
|
|
}
|
|
}
|
|
|
|
for _, te := range tests {
|
|
// prepare case
|
|
newSeg(testReplica, te.inisNew, te.inisFlushed, te.inSegID)
|
|
|
|
testReplica.segmentFlushed(te.inSegID)
|
|
|
|
flushedSeg := testReplica.flushedSegments[te.inSegID]
|
|
assert.Equal(t, te.expectedSegID, flushedSeg.segmentID)
|
|
assert.Equal(t, te.expectedisNew, flushedSeg.isNew.Load().(bool))
|
|
assert.Equal(t, te.expectedisFlushed, flushedSeg.isFlushed.Load().(bool))
|
|
}
|
|
|
|
})
|
|
}
|
|
|
|
func TestSegmentReplica_InterfaceMethod(t *testing.T) {
|
|
rc := &RootCoordFactory{
|
|
pkType: schemapb.DataType_Int64,
|
|
}
|
|
cm := storage.NewLocalChunkManager(storage.RootPath(segmentReplicaNodeTestDir))
|
|
defer cm.RemoveWithPrefix("")
|
|
|
|
t.Run("Test addFlushedSegmentWithPKs", func(t *testing.T) {
|
|
tests := []struct {
|
|
isvalid bool
|
|
|
|
incollID UniqueID
|
|
replicaCollID UniqueID
|
|
description string
|
|
}{
|
|
{true, 1, 1, "valid input collection with replica collection"},
|
|
{false, 1, 2, "invalid input collection with replica collection"},
|
|
}
|
|
|
|
primaryKeyData := &storage.Int64FieldData{
|
|
Data: []int64{9},
|
|
}
|
|
for _, test := range tests {
|
|
t.Run(test.description, func(t *testing.T) {
|
|
replica, err := newReplica(context.TODO(), rc, cm, test.replicaCollID)
|
|
require.NoError(t, err)
|
|
if test.isvalid {
|
|
replica.addFlushedSegmentWithPKs(100, test.incollID, 10, "a", 1, primaryKeyData)
|
|
|
|
assert.True(t, replica.hasSegment(100, true))
|
|
assert.False(t, replica.hasSegment(100, false))
|
|
} else {
|
|
replica.addFlushedSegmentWithPKs(100, test.incollID, 10, "a", 1, primaryKeyData)
|
|
assert.False(t, replica.hasSegment(100, true))
|
|
assert.False(t, replica.hasSegment(100, false))
|
|
}
|
|
})
|
|
}
|
|
})
|
|
|
|
t.Run("Test_addNewSegment", func(t *testing.T) {
|
|
tests := []struct {
|
|
isValidCase bool
|
|
replicaCollID UniqueID
|
|
inCollID UniqueID
|
|
inSegID UniqueID
|
|
|
|
instartPos *internalpb.MsgPosition
|
|
|
|
expectdIsNew bool
|
|
expectedIsFlushed bool
|
|
|
|
description string
|
|
}{
|
|
{isValidCase: false, replicaCollID: 1, inCollID: 2, inSegID: 300, description: "input CollID 2 mismatch with Replica collID"},
|
|
{true, 1, 1, 200, new(internalpb.MsgPosition), true, false, "nill address for startPos"},
|
|
{true, 1, 1, 200, &internalpb.MsgPosition{}, true, false, "empty struct for startPos"},
|
|
}
|
|
|
|
for _, test := range tests {
|
|
t.Run(test.description, func(t *testing.T) {
|
|
sr, err := newReplica(context.Background(), rc, cm, test.replicaCollID)
|
|
assert.Nil(t, err)
|
|
require.False(t, sr.hasSegment(test.inSegID, true))
|
|
err = sr.addNewSegment(test.inSegID,
|
|
test.inCollID, 1, "", test.instartPos, &internalpb.MsgPosition{})
|
|
if test.isValidCase {
|
|
assert.NoError(t, err)
|
|
assert.True(t, sr.hasSegment(test.inSegID, true))
|
|
assert.Equal(t, test.expectdIsNew, sr.newSegments[test.inSegID].isNew.Load().(bool))
|
|
assert.Equal(t, test.expectedIsFlushed, sr.newSegments[test.inSegID].isFlushed.Load().(bool))
|
|
} else {
|
|
assert.Error(t, err)
|
|
assert.False(t, sr.hasSegment(test.inSegID, true))
|
|
}
|
|
})
|
|
}
|
|
})
|
|
|
|
t.Run("Test_addNormalSegment", func(t *testing.T) {
|
|
tests := []struct {
|
|
isValidCase bool
|
|
replicaCollID UniqueID
|
|
inCollID UniqueID
|
|
inSegID UniqueID
|
|
|
|
expectdIsNew bool
|
|
expectedIsFlushed bool
|
|
|
|
description string
|
|
}{
|
|
{isValidCase: false, replicaCollID: 1, inCollID: 2, inSegID: 300, description: "input CollID 2 mismatch with Replica collID"},
|
|
{true, 1, 1, 200, false, false, "normal case"},
|
|
}
|
|
|
|
for _, test := range tests {
|
|
t.Run(test.description, func(t *testing.T) {
|
|
sr, err := newReplica(context.Background(), rc, &mockDataCM{}, test.replicaCollID)
|
|
assert.Nil(t, err)
|
|
require.False(t, sr.hasSegment(test.inSegID, true))
|
|
err = sr.addNormalSegment(test.inSegID, test.inCollID, 1, "", 0, []*datapb.FieldBinlog{getSimpleFieldBinlog()}, &segmentCheckPoint{}, 0)
|
|
if test.isValidCase {
|
|
assert.NoError(t, err)
|
|
assert.True(t, sr.hasSegment(test.inSegID, true))
|
|
assert.Equal(t, test.expectdIsNew, sr.normalSegments[test.inSegID].isNew.Load().(bool))
|
|
assert.Equal(t, test.expectedIsFlushed, sr.normalSegments[test.inSegID].isFlushed.Load().(bool))
|
|
} else {
|
|
assert.Error(t, err)
|
|
assert.False(t, sr.hasSegment(test.inSegID, true))
|
|
}
|
|
})
|
|
}
|
|
})
|
|
|
|
t.Run("Test_addNormalSegmentWithNilDml", func(t *testing.T) {
|
|
sr, err := newReplica(context.Background(), rc, &mockDataCM{}, 1)
|
|
require.NoError(t, err)
|
|
segID := int64(101)
|
|
require.False(t, sr.hasSegment(segID, true))
|
|
assert.NotPanics(t, func() {
|
|
err = sr.addNormalSegment(segID, 1, 10, "empty_dml_chan", 0, []*datapb.FieldBinlog{}, nil, 0)
|
|
assert.NoError(t, err)
|
|
})
|
|
})
|
|
|
|
t.Run("Test_listSegmentsCheckPoints", func(t *testing.T) {
|
|
tests := []struct {
|
|
newSegID UniqueID
|
|
newSegCP *segmentCheckPoint
|
|
|
|
normalSegID UniqueID
|
|
normalSegCP *segmentCheckPoint
|
|
|
|
flushedSegID UniqueID
|
|
flushedSegCP *segmentCheckPoint
|
|
|
|
description string
|
|
}{
|
|
{newSegID: 100, newSegCP: new(segmentCheckPoint),
|
|
description: "Only contain new Seg 100"},
|
|
{normalSegID: 200, normalSegCP: new(segmentCheckPoint),
|
|
description: "Only contain normal Seg 200"},
|
|
{flushedSegID: 300, flushedSegCP: new(segmentCheckPoint),
|
|
description: "Only contain flushed Seg 300"},
|
|
{100, new(segmentCheckPoint), 200, new(segmentCheckPoint), 0, new(segmentCheckPoint),
|
|
"New seg 100 and normal seg 200"},
|
|
{100, new(segmentCheckPoint), 0, new(segmentCheckPoint), 300, new(segmentCheckPoint),
|
|
"New seg 100 and flushed seg 300"},
|
|
{0, new(segmentCheckPoint), 200, new(segmentCheckPoint), 300, new(segmentCheckPoint),
|
|
"Normal seg 200 and flushed seg 300"},
|
|
{100, new(segmentCheckPoint), 200, new(segmentCheckPoint), 300, new(segmentCheckPoint),
|
|
"New seg 100, normal seg 200 and flushed seg 300"},
|
|
}
|
|
|
|
for _, test := range tests {
|
|
t.Run(test.description, func(t *testing.T) {
|
|
sr := SegmentReplica{
|
|
newSegments: make(map[UniqueID]*Segment),
|
|
normalSegments: make(map[UniqueID]*Segment),
|
|
flushedSegments: make(map[UniqueID]*Segment),
|
|
}
|
|
|
|
expectdCount := 0
|
|
if test.newSegID != 0 {
|
|
sr.newSegments[test.newSegID] = &Segment{checkPoint: *test.newSegCP}
|
|
expectdCount++
|
|
}
|
|
if test.normalSegID != 0 {
|
|
sr.normalSegments[test.normalSegID] = &Segment{checkPoint: *test.normalSegCP}
|
|
expectdCount++
|
|
}
|
|
if test.flushedSegID != 0 {
|
|
sr.flushedSegments[test.flushedSegID] = &Segment{checkPoint: *test.flushedSegCP}
|
|
}
|
|
|
|
scp := sr.listSegmentsCheckPoints()
|
|
assert.Equal(t, expectdCount, len(scp))
|
|
})
|
|
}
|
|
})
|
|
|
|
t.Run("Test_updateSegmentEndPosition", func(t *testing.T) {
|
|
tests := []struct {
|
|
newSegID UniqueID
|
|
normalSegID UniqueID
|
|
flushedSegID UniqueID
|
|
|
|
inSegID UniqueID
|
|
description string
|
|
}{
|
|
{newSegID: 100, inSegID: 100,
|
|
description: "input seg 100 in newSegments"},
|
|
{newSegID: 100, inSegID: 101,
|
|
description: "input seg 101 not in newSegments"},
|
|
{normalSegID: 200, inSegID: 200,
|
|
description: "input seg 200 in normalSegments"},
|
|
{normalSegID: 200, inSegID: 201,
|
|
description: "input seg 201 not in normalSegments"},
|
|
{flushedSegID: 300, inSegID: 300,
|
|
description: "input seg 300 in flushedSegments"},
|
|
{flushedSegID: 300, inSegID: 301,
|
|
description: "input seg 301 not in flushedSegments"},
|
|
}
|
|
|
|
for _, test := range tests {
|
|
t.Run(test.description, func(t *testing.T) {
|
|
sr := SegmentReplica{
|
|
newSegments: make(map[UniqueID]*Segment),
|
|
normalSegments: make(map[UniqueID]*Segment),
|
|
flushedSegments: make(map[UniqueID]*Segment),
|
|
}
|
|
|
|
if test.newSegID != 0 {
|
|
sr.newSegments[test.newSegID] = &Segment{}
|
|
}
|
|
if test.normalSegID != 0 {
|
|
sr.normalSegments[test.normalSegID] = &Segment{}
|
|
}
|
|
if test.flushedSegID != 0 {
|
|
sr.flushedSegments[test.flushedSegID] = &Segment{}
|
|
}
|
|
sr.updateSegmentEndPosition(test.inSegID, new(internalpb.MsgPosition))
|
|
sr.removeSegments(0)
|
|
|
|
})
|
|
}
|
|
})
|
|
|
|
t.Run("Test_updateStatistics", func(t *testing.T) {
|
|
tests := []struct {
|
|
isvalidCase bool
|
|
|
|
newSegID UniqueID
|
|
normalSegID UniqueID
|
|
flushedSegID UniqueID
|
|
|
|
inSegID UniqueID
|
|
inNumRows int64
|
|
description string
|
|
}{
|
|
{isvalidCase: true, newSegID: 100, inSegID: 100, inNumRows: 100,
|
|
description: "input seg 100 in newSegments with numRows 100"},
|
|
{isvalidCase: false, newSegID: 100, inSegID: 101, inNumRows: 100,
|
|
description: "input seg 101 not in newSegments with numRows 100"},
|
|
{isvalidCase: true, normalSegID: 200, inSegID: 200, inNumRows: 200,
|
|
description: "input seg 200 in normalSegments with numRows 200"},
|
|
{isvalidCase: false, normalSegID: 200, inSegID: 201, inNumRows: 200,
|
|
description: "input seg 201 not in normalSegments with numRows 200"},
|
|
{isvalidCase: true, flushedSegID: 300, inSegID: 300, inNumRows: 300,
|
|
description: "input seg 300 in flushedSegments"},
|
|
{isvalidCase: false, flushedSegID: 300, inSegID: 301, inNumRows: 300,
|
|
description: "input seg 301 not in flushedSegments"},
|
|
}
|
|
for _, test := range tests {
|
|
t.Run(test.description, func(t *testing.T) {
|
|
sr := SegmentReplica{
|
|
newSegments: make(map[UniqueID]*Segment),
|
|
normalSegments: make(map[UniqueID]*Segment),
|
|
flushedSegments: make(map[UniqueID]*Segment),
|
|
}
|
|
|
|
if test.newSegID != 0 {
|
|
sr.newSegments[test.newSegID] = &Segment{}
|
|
}
|
|
if test.normalSegID != 0 {
|
|
sr.normalSegments[test.normalSegID] = &Segment{}
|
|
}
|
|
if test.flushedSegID != 0 { // not update flushed num rows
|
|
sr.flushedSegments[test.flushedSegID] = &Segment{
|
|
numRows: test.inNumRows,
|
|
}
|
|
}
|
|
|
|
sr.updateStatistics(test.inSegID, test.inNumRows)
|
|
if test.isvalidCase {
|
|
|
|
updates, err := sr.getSegmentStatisticsUpdates(test.inSegID)
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, test.inNumRows, updates.GetNumRows())
|
|
assert.Equal(t, test.inSegID, updates.GetSegmentID())
|
|
|
|
sr.updateSegmentCheckPoint(10000)
|
|
} else {
|
|
updates, err := sr.getSegmentStatisticsUpdates(test.inSegID)
|
|
assert.Error(t, err)
|
|
assert.Nil(t, updates)
|
|
}
|
|
})
|
|
}
|
|
})
|
|
|
|
t.Run("Test_getCollectionSchema", func(t *testing.T) {
|
|
tests := []struct {
|
|
isValid bool
|
|
replicaCollID UniqueID
|
|
inputCollID UniqueID
|
|
|
|
metaServiceErr bool
|
|
description string
|
|
}{
|
|
{true, 1, 1, false, "Normal case"},
|
|
{false, 1, 2, false, "Input collID 2 mismatch with replicaCollID 1"},
|
|
{false, 1, 1, true, "RPC call fails"},
|
|
}
|
|
|
|
for _, test := range tests {
|
|
t.Run(test.description, func(t *testing.T) {
|
|
sr, err := newReplica(context.Background(), rc, cm, test.replicaCollID)
|
|
assert.Nil(t, err)
|
|
|
|
if test.metaServiceErr {
|
|
rc.setCollectionID(-1)
|
|
} else {
|
|
rc.setCollectionID(1)
|
|
}
|
|
|
|
s, err := sr.getCollectionSchema(test.inputCollID, Timestamp(0))
|
|
if test.isValid {
|
|
assert.NoError(t, err)
|
|
assert.NotNil(t, s)
|
|
} else {
|
|
assert.Error(t, err)
|
|
assert.Nil(t, s)
|
|
}
|
|
})
|
|
}
|
|
rc.setCollectionID(1)
|
|
})
|
|
|
|
t.Run("Test listAllSegmentIDs", func(t *testing.T) {
|
|
sr := &SegmentReplica{
|
|
newSegments: map[UniqueID]*Segment{1: {segmentID: 1}},
|
|
normalSegments: map[UniqueID]*Segment{2: {segmentID: 2}},
|
|
flushedSegments: map[UniqueID]*Segment{3: {segmentID: 3}},
|
|
}
|
|
|
|
ids := sr.listAllSegmentIDs()
|
|
assert.ElementsMatch(t, []UniqueID{1, 2, 3}, ids)
|
|
})
|
|
|
|
t.Run("Test listPartitionSegments", func(t *testing.T) {
|
|
sr := &SegmentReplica{
|
|
newSegments: map[UniqueID]*Segment{1: {segmentID: 1, partitionID: 1}, 4: {segmentID: 4, partitionID: 2}},
|
|
normalSegments: map[UniqueID]*Segment{2: {segmentID: 2, partitionID: 1}, 5: {segmentID: 5, partitionID: 2}},
|
|
flushedSegments: map[UniqueID]*Segment{3: {segmentID: 3, partitionID: 1}, 6: {segmentID: 6, partitionID: 2}},
|
|
}
|
|
|
|
ids := sr.listPartitionSegments(1)
|
|
assert.ElementsMatch(t, []UniqueID{1, 2, 3}, ids)
|
|
})
|
|
|
|
t.Run("Test_addSegmentMinIOLoadError", func(t *testing.T) {
|
|
sr, err := newReplica(context.Background(), rc, cm, 1)
|
|
assert.Nil(t, err)
|
|
sr.chunkManager = &mockDataCMError{}
|
|
|
|
cpPos := &internalpb.MsgPosition{ChannelName: "insert-01", Timestamp: Timestamp(10)}
|
|
cp := &segmentCheckPoint{int64(10), *cpPos}
|
|
err = sr.addNormalSegment(1, 1, 2, "insert-01", int64(10), []*datapb.FieldBinlog{getSimpleFieldBinlog()}, cp, 0)
|
|
assert.NotNil(t, err)
|
|
err = sr.addFlushedSegment(1, 1, 2, "insert-01", int64(0), []*datapb.FieldBinlog{getSimpleFieldBinlog()}, 0)
|
|
assert.NotNil(t, err)
|
|
})
|
|
|
|
t.Run("Test_addSegmentStatsError", func(t *testing.T) {
|
|
sr, err := newReplica(context.Background(), rc, cm, 1)
|
|
assert.Nil(t, err)
|
|
sr.chunkManager = &mockDataCMStatsError{}
|
|
|
|
cpPos := &internalpb.MsgPosition{ChannelName: "insert-01", Timestamp: Timestamp(10)}
|
|
cp := &segmentCheckPoint{int64(10), *cpPos}
|
|
err = sr.addNormalSegment(1, 1, 2, "insert-01", int64(10), []*datapb.FieldBinlog{getSimpleFieldBinlog()}, cp, 0)
|
|
assert.NotNil(t, err)
|
|
err = sr.addFlushedSegment(1, 1, 2, "insert-01", int64(0), []*datapb.FieldBinlog{getSimpleFieldBinlog()}, 0)
|
|
assert.NotNil(t, err)
|
|
})
|
|
|
|
t.Run("Test_addSegmentPkfilterError", func(t *testing.T) {
|
|
sr, err := newReplica(context.Background(), rc, cm, 1)
|
|
assert.Nil(t, err)
|
|
sr.chunkManager = &mockPkfilterMergeError{}
|
|
|
|
cpPos := &internalpb.MsgPosition{ChannelName: "insert-01", Timestamp: Timestamp(10)}
|
|
cp := &segmentCheckPoint{int64(10), *cpPos}
|
|
err = sr.addNormalSegment(1, 1, 2, "insert-01", int64(10), []*datapb.FieldBinlog{getSimpleFieldBinlog()}, cp, 0)
|
|
assert.NotNil(t, err)
|
|
err = sr.addFlushedSegment(1, 1, 2, "insert-01", int64(0), []*datapb.FieldBinlog{getSimpleFieldBinlog()}, 0)
|
|
assert.NotNil(t, err)
|
|
})
|
|
|
|
t.Run("Test_mergeFlushedSegments", func(t *testing.T) {
|
|
sr, err := newReplica(context.Background(), rc, cm, 1)
|
|
assert.Nil(t, err)
|
|
|
|
primaryKeyData := &storage.Int64FieldData{
|
|
Data: []UniqueID{1},
|
|
}
|
|
sr.addFlushedSegmentWithPKs(1, 1, 0, "channel", 10, primaryKeyData)
|
|
sr.addFlushedSegmentWithPKs(2, 1, 0, "channel", 10, primaryKeyData)
|
|
require.True(t, sr.hasSegment(1, true))
|
|
require.True(t, sr.hasSegment(2, true))
|
|
|
|
sr.mergeFlushedSegments(3, 1, 0, 100, []UniqueID{1, 2}, "channel", 15)
|
|
assert.True(t, sr.hasSegment(3, true))
|
|
assert.False(t, sr.hasSegment(1, true))
|
|
assert.False(t, sr.hasSegment(2, true))
|
|
|
|
to2from := sr.listCompactedSegmentIDs()
|
|
assert.NotEmpty(t, to2from)
|
|
|
|
from, ok := to2from[3]
|
|
assert.True(t, ok)
|
|
assert.ElementsMatch(t, []UniqueID{1, 2}, from)
|
|
})
|
|
|
|
}
|
|
func TestInnerFunctionSegment(t *testing.T) {
|
|
rc := &RootCoordFactory{
|
|
pkType: schemapb.DataType_Int64,
|
|
}
|
|
collID := UniqueID(1)
|
|
cm := storage.NewLocalChunkManager(storage.RootPath(segmentReplicaNodeTestDir))
|
|
defer cm.RemoveWithPrefix("")
|
|
replica, err := newReplica(context.Background(), rc, cm, collID)
|
|
assert.Nil(t, err)
|
|
replica.chunkManager = &mockDataCM{}
|
|
assert.False(t, replica.hasSegment(0, true))
|
|
assert.False(t, replica.hasSegment(0, false))
|
|
|
|
startPos := &internalpb.MsgPosition{ChannelName: "insert-01", Timestamp: Timestamp(100)}
|
|
endPos := &internalpb.MsgPosition{ChannelName: "insert-01", Timestamp: Timestamp(200)}
|
|
err = replica.addNewSegment(0, 1, 2, "insert-01", startPos, endPos)
|
|
assert.NoError(t, err)
|
|
assert.True(t, replica.hasSegment(0, true))
|
|
assert.Equal(t, 1, len(replica.newSegments))
|
|
|
|
seg, ok := replica.newSegments[UniqueID(0)]
|
|
assert.True(t, ok)
|
|
require.NotNil(t, seg)
|
|
assert.Equal(t, UniqueID(0), seg.segmentID)
|
|
assert.Equal(t, UniqueID(1), seg.collectionID)
|
|
assert.Equal(t, UniqueID(2), seg.partitionID)
|
|
assert.Equal(t, "insert-01", seg.channelName)
|
|
assert.Equal(t, Timestamp(100), seg.startPos.Timestamp)
|
|
assert.Equal(t, Timestamp(200), seg.endPos.Timestamp)
|
|
assert.Equal(t, startPos.ChannelName, seg.checkPoint.pos.ChannelName)
|
|
assert.Equal(t, startPos.Timestamp, seg.checkPoint.pos.Timestamp)
|
|
assert.Equal(t, int64(0), seg.numRows)
|
|
assert.True(t, seg.isNew.Load().(bool))
|
|
assert.False(t, seg.isFlushed.Load().(bool))
|
|
|
|
replica.updateStatistics(0, 10)
|
|
assert.Equal(t, int64(10), seg.numRows)
|
|
|
|
cpPos := &internalpb.MsgPosition{ChannelName: "insert-01", Timestamp: Timestamp(10)}
|
|
cp := &segmentCheckPoint{int64(10), *cpPos}
|
|
err = replica.addNormalSegment(1, 1, 2, "insert-01", int64(10), []*datapb.FieldBinlog{getSimpleFieldBinlog()}, cp, 0)
|
|
assert.NoError(t, err)
|
|
assert.True(t, replica.hasSegment(1, true))
|
|
assert.Equal(t, 1, len(replica.normalSegments))
|
|
seg, ok = replica.normalSegments[UniqueID(1)]
|
|
assert.True(t, ok)
|
|
require.NotNil(t, seg)
|
|
assert.Equal(t, UniqueID(1), seg.segmentID)
|
|
assert.Equal(t, UniqueID(1), seg.collectionID)
|
|
assert.Equal(t, UniqueID(2), seg.partitionID)
|
|
assert.Equal(t, "insert-01", seg.channelName)
|
|
assert.Equal(t, cpPos.ChannelName, seg.checkPoint.pos.ChannelName)
|
|
assert.Equal(t, cpPos.Timestamp, seg.checkPoint.pos.Timestamp)
|
|
assert.Equal(t, int64(10), seg.numRows)
|
|
assert.False(t, seg.isNew.Load().(bool))
|
|
assert.False(t, seg.isFlushed.Load().(bool))
|
|
|
|
err = replica.addNormalSegment(1, 100000, 2, "invalid", int64(0), []*datapb.FieldBinlog{getSimpleFieldBinlog()}, &segmentCheckPoint{}, 0)
|
|
assert.Error(t, err)
|
|
|
|
replica.updateStatistics(1, 10)
|
|
assert.Equal(t, int64(20), seg.numRows)
|
|
|
|
segPos := replica.listNewSegmentsStartPositions()
|
|
assert.Equal(t, 1, len(segPos))
|
|
assert.Equal(t, UniqueID(0), segPos[0].SegmentID)
|
|
assert.Equal(t, "insert-01", segPos[0].StartPosition.ChannelName)
|
|
assert.Equal(t, Timestamp(100), segPos[0].StartPosition.Timestamp)
|
|
|
|
assert.Equal(t, 0, len(replica.newSegments))
|
|
assert.Equal(t, 2, len(replica.normalSegments))
|
|
|
|
cps := replica.listSegmentsCheckPoints()
|
|
assert.Equal(t, 2, len(cps))
|
|
assert.Equal(t, startPos.Timestamp, cps[UniqueID(0)].pos.Timestamp)
|
|
assert.Equal(t, int64(0), cps[UniqueID(0)].numRows)
|
|
assert.Equal(t, cp.pos.Timestamp, cps[UniqueID(1)].pos.Timestamp)
|
|
assert.Equal(t, int64(10), cps[UniqueID(1)].numRows)
|
|
|
|
updates, err := replica.getSegmentStatisticsUpdates(0)
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, int64(10), updates.NumRows)
|
|
|
|
updates, err = replica.getSegmentStatisticsUpdates(1)
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, int64(20), updates.NumRows)
|
|
|
|
replica.updateSegmentCheckPoint(0)
|
|
assert.Equal(t, int64(10), replica.normalSegments[UniqueID(0)].checkPoint.numRows)
|
|
replica.updateSegmentCheckPoint(1)
|
|
assert.Equal(t, int64(20), replica.normalSegments[UniqueID(1)].checkPoint.numRows)
|
|
|
|
err = replica.addFlushedSegment(1, 1, 2, "insert-01", int64(0), []*datapb.FieldBinlog{getSimpleFieldBinlog()}, 0)
|
|
assert.Nil(t, err)
|
|
|
|
totalSegments := replica.filterSegments("insert-01", common.InvalidPartitionID)
|
|
assert.Equal(t, len(totalSegments), 3)
|
|
}
|
|
|
|
func TestSegmentReplica_UpdatePKRange(t *testing.T) {
|
|
seg := &Segment{
|
|
pkFilter: bloom.NewWithEstimates(100000, 0.005),
|
|
}
|
|
|
|
cases := make([]int64, 0, 100)
|
|
for i := 0; i < 100; i++ {
|
|
cases = append(cases, rand.Int63())
|
|
}
|
|
buf := make([]byte, 8)
|
|
for _, c := range cases {
|
|
seg.updatePKRange(&storage.Int64FieldData{
|
|
Data: []int64{c},
|
|
})
|
|
|
|
pk := newInt64PrimaryKey(c)
|
|
|
|
assert.Equal(t, true, seg.minPK.LE(pk))
|
|
assert.Equal(t, true, seg.maxPK.GE(pk))
|
|
|
|
common.Endian.PutUint64(buf, uint64(c))
|
|
assert.True(t, seg.pkFilter.Test(buf))
|
|
}
|
|
}
|
|
|
|
func TestSegment_getSegmentStatslog(t *testing.T) {
|
|
rand.Seed(time.Now().UnixNano())
|
|
|
|
cases := make([][]int64, 0, 100)
|
|
for i := 0; i < 100; i++ {
|
|
tc := make([]int64, 0, 10)
|
|
for j := 0; j < 100; j++ {
|
|
tc = append(tc, rand.Int63())
|
|
}
|
|
cases = append(cases, tc)
|
|
}
|
|
buf := make([]byte, 8)
|
|
for _, tc := range cases {
|
|
seg := &Segment{
|
|
pkFilter: bloom.NewWithEstimates(100000, 0.005),
|
|
}
|
|
|
|
seg.updatePKRange(&storage.Int64FieldData{
|
|
Data: tc,
|
|
})
|
|
|
|
statBytes, err := seg.getSegmentStatslog(1, schemapb.DataType_Int64)
|
|
assert.NoError(t, err)
|
|
|
|
pks := storage.PrimaryKeyStats{}
|
|
err = json.Unmarshal(statBytes, &pks)
|
|
require.NoError(t, err)
|
|
|
|
assert.Equal(t, int64(1), pks.FieldID)
|
|
assert.Equal(t, int64(schemapb.DataType_Int64), pks.PkType)
|
|
|
|
for _, v := range tc {
|
|
pk := newInt64PrimaryKey(v)
|
|
assert.True(t, pks.MinPk.LE(pk))
|
|
assert.True(t, pks.MaxPk.GE(pk))
|
|
|
|
common.Endian.PutUint64(buf, uint64(v))
|
|
assert.True(t, seg.pkFilter.Test(buf))
|
|
}
|
|
}
|
|
|
|
pks := &storage.PrimaryKeyStats{}
|
|
_, err := json.Marshal(pks)
|
|
assert.NoError(t, err)
|
|
}
|
|
|
|
func TestReplica_UpdatePKRange(t *testing.T) {
|
|
rc := &RootCoordFactory{
|
|
pkType: schemapb.DataType_Int64,
|
|
}
|
|
collID := UniqueID(1)
|
|
partID := UniqueID(2)
|
|
chanName := "insert-02"
|
|
startPos := &internalpb.MsgPosition{ChannelName: chanName, Timestamp: Timestamp(100)}
|
|
endPos := &internalpb.MsgPosition{ChannelName: chanName, Timestamp: Timestamp(200)}
|
|
cpPos := &internalpb.MsgPosition{ChannelName: chanName, Timestamp: Timestamp(10)}
|
|
cp := &segmentCheckPoint{int64(10), *cpPos}
|
|
|
|
cm := storage.NewLocalChunkManager(storage.RootPath(segmentReplicaNodeTestDir))
|
|
defer cm.RemoveWithPrefix("")
|
|
replica, err := newReplica(context.Background(), rc, cm, collID)
|
|
assert.Nil(t, err)
|
|
replica.chunkManager = &mockDataCM{}
|
|
|
|
err = replica.addNewSegment(1, collID, partID, chanName, startPos, endPos)
|
|
assert.Nil(t, err)
|
|
err = replica.addNormalSegment(2, collID, partID, chanName, 100, []*datapb.FieldBinlog{getSimpleFieldBinlog()}, cp, 0)
|
|
assert.Nil(t, err)
|
|
|
|
segNew := replica.newSegments[1]
|
|
segNormal := replica.normalSegments[2]
|
|
|
|
cases := make([]int64, 0, 100)
|
|
for i := 0; i < 100; i++ {
|
|
cases = append(cases, rand.Int63())
|
|
}
|
|
buf := make([]byte, 8)
|
|
for _, c := range cases {
|
|
replica.updateSegmentPKRange(1, &storage.Int64FieldData{Data: []int64{c}}) // new segment
|
|
replica.updateSegmentPKRange(2, &storage.Int64FieldData{Data: []int64{c}}) // normal segment
|
|
replica.updateSegmentPKRange(3, &storage.Int64FieldData{Data: []int64{c}}) // non-exist segment
|
|
|
|
pk := newInt64PrimaryKey(c)
|
|
|
|
assert.Equal(t, true, segNew.minPK.LE(pk))
|
|
assert.Equal(t, true, segNew.maxPK.GE(pk))
|
|
|
|
assert.Equal(t, true, segNormal.minPK.LE(pk))
|
|
assert.Equal(t, true, segNormal.maxPK.GE(pk))
|
|
|
|
common.Endian.PutUint64(buf, uint64(c))
|
|
assert.True(t, segNew.pkFilter.Test(buf))
|
|
assert.True(t, segNormal.pkFilter.Test(buf))
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// SegmentReplicaSuite setup test suite for SegmentReplica
|
|
type SegmentReplicaSuite struct {
|
|
suite.Suite
|
|
sr *SegmentReplica
|
|
|
|
collID UniqueID
|
|
partID UniqueID
|
|
vchanName string
|
|
cm *storage.LocalChunkManager
|
|
}
|
|
|
|
func (s *SegmentReplicaSuite) SetupSuite() {
|
|
rc := &RootCoordFactory{
|
|
pkType: schemapb.DataType_Int64,
|
|
}
|
|
s.collID = 1
|
|
s.cm = storage.NewLocalChunkManager(storage.RootPath(segmentReplicaNodeTestDir))
|
|
var err error
|
|
s.sr, err = newReplica(context.Background(), rc, s.cm, s.collID)
|
|
s.Require().NoError(err)
|
|
}
|
|
|
|
func (s *SegmentReplicaSuite) TearDownSuite() {
|
|
s.cm.RemoveWithPrefix("")
|
|
|
|
}
|
|
|
|
func (s *SegmentReplicaSuite) SetupTest() {
|
|
var err error
|
|
err = s.sr.addNewSegment(1, s.collID, s.partID, s.vchanName, &internalpb.MsgPosition{}, nil)
|
|
s.Require().NoError(err)
|
|
err = s.sr.addNormalSegment(2, s.collID, s.partID, s.vchanName, 10, nil, nil, 0)
|
|
s.Require().NoError(err)
|
|
err = s.sr.addFlushedSegment(3, s.collID, s.partID, s.vchanName, 10, nil, 0)
|
|
s.Require().NoError(err)
|
|
}
|
|
|
|
func (s *SegmentReplicaSuite) TearDownTest() {
|
|
s.sr.removeSegments(1, 2, 3)
|
|
}
|
|
|
|
func (s *SegmentReplicaSuite) TestGetSegmentStatslog() {
|
|
bs, err := s.sr.getSegmentStatslog(1)
|
|
s.NoError(err)
|
|
|
|
segment, ok := s.getSegmentByID(1)
|
|
s.Require().True(ok)
|
|
expected, err := segment.getSegmentStatslog(106, schemapb.DataType_Int64)
|
|
s.Require().NoError(err)
|
|
s.Equal(expected, bs)
|
|
|
|
bs, err = s.sr.getSegmentStatslog(2)
|
|
s.NoError(err)
|
|
|
|
segment, ok = s.getSegmentByID(2)
|
|
s.Require().True(ok)
|
|
expected, err = segment.getSegmentStatslog(106, schemapb.DataType_Int64)
|
|
s.Require().NoError(err)
|
|
s.Equal(expected, bs)
|
|
|
|
bs, err = s.sr.getSegmentStatslog(3)
|
|
s.NoError(err)
|
|
|
|
segment, ok = s.getSegmentByID(3)
|
|
s.Require().True(ok)
|
|
expected, err = segment.getSegmentStatslog(106, schemapb.DataType_Int64)
|
|
s.Require().NoError(err)
|
|
s.Equal(expected, bs)
|
|
|
|
_, err = s.sr.getSegmentStatslog(4)
|
|
s.Error(err)
|
|
}
|
|
|
|
func (s *SegmentReplicaSuite) getSegmentByID(id UniqueID) (*Segment, bool) {
|
|
s.sr.segMu.RLock()
|
|
defer s.sr.segMu.RUnlock()
|
|
|
|
seg, ok := s.sr.newSegments[id]
|
|
if ok {
|
|
return seg, true
|
|
}
|
|
|
|
seg, ok = s.sr.normalSegments[id]
|
|
if ok {
|
|
return seg, true
|
|
}
|
|
|
|
seg, ok = s.sr.flushedSegments[id]
|
|
if ok {
|
|
return seg, true
|
|
}
|
|
|
|
return nil, false
|
|
}
|
|
|
|
func TestSegmentReplicaSuite(t *testing.T) {
|
|
suite.Run(t, new(SegmentReplicaSuite))
|
|
}
|