mirror of
https://gitee.com/milvus-io/milvus.git
synced 2024-11-30 02:48:45 +08:00
2c9bb4dfa3
issue: #33744 This PR includes the following changes: 1. Added a new task type to the task scheduler in datacoord: stats task, which sorts segments by primary key. 2. Implemented segment sorting in indexnode. 3. Added a new field `FieldStatsLog` to SegmentInfo to store token index information. --------- Signed-off-by: Cai Zhang <cai.zhang@zilliz.com>
207 lines
5.9 KiB
Go
207 lines
5.9 KiB
Go
// Licensed to the LF AI & Data foundation under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package session
|
|
|
|
import (
|
|
"context"
|
|
"testing"
|
|
|
|
"github.com/cockroachdb/errors"
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/mock"
|
|
|
|
"github.com/milvus-io/milvus/internal/mocks"
|
|
"github.com/milvus-io/milvus/internal/proto/workerpb"
|
|
"github.com/milvus-io/milvus/internal/types"
|
|
"github.com/milvus-io/milvus/pkg/util/lock"
|
|
"github.com/milvus-io/milvus/pkg/util/merr"
|
|
"github.com/milvus-io/milvus/pkg/util/paramtable"
|
|
typeutil "github.com/milvus-io/milvus/pkg/util/typeutil"
|
|
)
|
|
|
|
func TestIndexNodeManager_AddNode(t *testing.T) {
|
|
paramtable.Init()
|
|
nm := NewNodeManager(context.Background(), defaultIndexNodeCreatorFunc)
|
|
|
|
t.Run("success", func(t *testing.T) {
|
|
err := nm.AddNode(1, "indexnode-1")
|
|
assert.NoError(t, err)
|
|
})
|
|
|
|
t.Run("fail", func(t *testing.T) {
|
|
err := nm.AddNode(2, "")
|
|
assert.Error(t, err)
|
|
})
|
|
}
|
|
|
|
func TestIndexNodeManager_PickClient(t *testing.T) {
|
|
paramtable.Init()
|
|
getMockedGetJobStatsClient := func(resp *workerpb.GetJobStatsResponse, err error) types.IndexNodeClient {
|
|
ic := mocks.NewMockIndexNodeClient(t)
|
|
ic.EXPECT().GetJobStats(mock.Anything, mock.Anything, mock.Anything).Return(resp, err)
|
|
return ic
|
|
}
|
|
|
|
err := errors.New("error")
|
|
|
|
t.Run("multiple unavailable IndexNode", func(t *testing.T) {
|
|
nm := &IndexNodeManager{
|
|
ctx: context.TODO(),
|
|
nodeClients: map[typeutil.UniqueID]types.IndexNodeClient{
|
|
1: getMockedGetJobStatsClient(&workerpb.GetJobStatsResponse{
|
|
Status: merr.Status(err),
|
|
}, err),
|
|
2: getMockedGetJobStatsClient(&workerpb.GetJobStatsResponse{
|
|
Status: merr.Status(err),
|
|
}, err),
|
|
3: getMockedGetJobStatsClient(&workerpb.GetJobStatsResponse{
|
|
Status: merr.Status(err),
|
|
}, err),
|
|
4: getMockedGetJobStatsClient(&workerpb.GetJobStatsResponse{
|
|
Status: merr.Status(err),
|
|
}, err),
|
|
5: getMockedGetJobStatsClient(&workerpb.GetJobStatsResponse{
|
|
Status: merr.Status(err),
|
|
}, nil),
|
|
6: getMockedGetJobStatsClient(&workerpb.GetJobStatsResponse{
|
|
Status: merr.Status(err),
|
|
}, nil),
|
|
7: getMockedGetJobStatsClient(&workerpb.GetJobStatsResponse{
|
|
Status: merr.Status(err),
|
|
}, nil),
|
|
8: getMockedGetJobStatsClient(&workerpb.GetJobStatsResponse{
|
|
TaskSlots: 1,
|
|
Status: merr.Success(),
|
|
}, nil),
|
|
9: getMockedGetJobStatsClient(&workerpb.GetJobStatsResponse{
|
|
TaskSlots: 10,
|
|
Status: merr.Success(),
|
|
}, nil),
|
|
},
|
|
}
|
|
|
|
selectNodeID, client := nm.PickClient()
|
|
assert.NotNil(t, client)
|
|
assert.Contains(t, []typeutil.UniqueID{8, 9}, selectNodeID)
|
|
})
|
|
}
|
|
|
|
func TestIndexNodeManager_ClientSupportDisk(t *testing.T) {
|
|
paramtable.Init()
|
|
getMockedGetJobStatsClient := func(resp *workerpb.GetJobStatsResponse, err error) types.IndexNodeClient {
|
|
ic := mocks.NewMockIndexNodeClient(t)
|
|
ic.EXPECT().GetJobStats(mock.Anything, mock.Anything, mock.Anything).Return(resp, err)
|
|
return ic
|
|
}
|
|
|
|
err := errors.New("error")
|
|
|
|
t.Run("support", func(t *testing.T) {
|
|
nm := &IndexNodeManager{
|
|
ctx: context.Background(),
|
|
lock: lock.RWMutex{},
|
|
nodeClients: map[typeutil.UniqueID]types.IndexNodeClient{
|
|
1: getMockedGetJobStatsClient(&workerpb.GetJobStatsResponse{
|
|
Status: merr.Success(),
|
|
TaskSlots: 1,
|
|
JobInfos: nil,
|
|
EnableDisk: true,
|
|
}, nil),
|
|
},
|
|
}
|
|
|
|
support := nm.ClientSupportDisk()
|
|
assert.True(t, support)
|
|
})
|
|
|
|
t.Run("not support", func(t *testing.T) {
|
|
nm := &IndexNodeManager{
|
|
ctx: context.Background(),
|
|
lock: lock.RWMutex{},
|
|
nodeClients: map[typeutil.UniqueID]types.IndexNodeClient{
|
|
1: getMockedGetJobStatsClient(&workerpb.GetJobStatsResponse{
|
|
Status: merr.Success(),
|
|
TaskSlots: 1,
|
|
JobInfos: nil,
|
|
EnableDisk: false,
|
|
}, nil),
|
|
},
|
|
}
|
|
|
|
support := nm.ClientSupportDisk()
|
|
assert.False(t, support)
|
|
})
|
|
|
|
t.Run("no indexnode", func(t *testing.T) {
|
|
nm := &IndexNodeManager{
|
|
ctx: context.Background(),
|
|
lock: lock.RWMutex{},
|
|
nodeClients: map[typeutil.UniqueID]types.IndexNodeClient{},
|
|
}
|
|
|
|
support := nm.ClientSupportDisk()
|
|
assert.False(t, support)
|
|
})
|
|
|
|
t.Run("error", func(t *testing.T) {
|
|
nm := &IndexNodeManager{
|
|
ctx: context.Background(),
|
|
lock: lock.RWMutex{},
|
|
nodeClients: map[typeutil.UniqueID]types.IndexNodeClient{
|
|
1: getMockedGetJobStatsClient(nil, err),
|
|
},
|
|
}
|
|
|
|
support := nm.ClientSupportDisk()
|
|
assert.False(t, support)
|
|
})
|
|
|
|
t.Run("fail reason", func(t *testing.T) {
|
|
nm := &IndexNodeManager{
|
|
ctx: context.Background(),
|
|
lock: lock.RWMutex{},
|
|
nodeClients: map[typeutil.UniqueID]types.IndexNodeClient{
|
|
1: getMockedGetJobStatsClient(&workerpb.GetJobStatsResponse{
|
|
Status: merr.Status(err),
|
|
TaskSlots: 0,
|
|
JobInfos: nil,
|
|
EnableDisk: false,
|
|
}, nil),
|
|
},
|
|
}
|
|
|
|
support := nm.ClientSupportDisk()
|
|
assert.False(t, support)
|
|
})
|
|
}
|
|
|
|
func TestNodeManager_StoppingNode(t *testing.T) {
|
|
paramtable.Init()
|
|
nm := NewNodeManager(context.Background(), defaultIndexNodeCreatorFunc)
|
|
err := nm.AddNode(1, "indexnode-1")
|
|
assert.NoError(t, err)
|
|
assert.Equal(t, 1, len(nm.GetAllClients()))
|
|
|
|
nm.StoppingNode(1)
|
|
assert.Equal(t, 0, len(nm.GetAllClients()))
|
|
assert.Equal(t, 1, len(nm.stoppingNodes))
|
|
|
|
nm.RemoveNode(1)
|
|
assert.Equal(t, 0, len(nm.GetAllClients()))
|
|
assert.Equal(t, 0, len(nm.stoppingNodes))
|
|
}
|