mirror of
https://gitee.com/milvus-io/milvus.git
synced 2024-12-02 20:09:57 +08:00
feat: add a balancer based on both of row count and segment count (#30188)
issue: https://github.com/milvus-io/milvus/issues/30039 --------- Signed-off-by: sunby <sunbingyi1992@gmail.com>
This commit is contained in:
parent
b111f3b110
commit
715f042965
561
internal/querycoordv2/balance/multi_target_balance.go
Normal file
561
internal/querycoordv2/balance/multi_target_balance.go
Normal file
@ -0,0 +1,561 @@
|
||||
package balance
|
||||
|
||||
import (
|
||||
"math"
|
||||
"math/rand"
|
||||
"sort"
|
||||
"time"
|
||||
|
||||
"github.com/samber/lo"
|
||||
"go.uber.org/zap"
|
||||
|
||||
"github.com/milvus-io/milvus/internal/proto/datapb"
|
||||
"github.com/milvus-io/milvus/internal/querycoordv2/meta"
|
||||
"github.com/milvus-io/milvus/internal/querycoordv2/params"
|
||||
"github.com/milvus-io/milvus/internal/querycoordv2/session"
|
||||
"github.com/milvus-io/milvus/internal/querycoordv2/task"
|
||||
"github.com/milvus-io/milvus/pkg/log"
|
||||
"github.com/milvus-io/milvus/pkg/util/paramtable"
|
||||
)
|
||||
|
||||
func init() {
|
||||
rand.Seed(time.Now().UnixNano())
|
||||
}
|
||||
|
||||
type rowCountCostModel struct {
|
||||
nodeSegments map[int64][]*meta.Segment
|
||||
}
|
||||
|
||||
func (m *rowCountCostModel) cost() float64 {
|
||||
nodeCount := len(m.nodeSegments)
|
||||
if nodeCount == 0 {
|
||||
return 0
|
||||
}
|
||||
totalRowCount := 0
|
||||
nodesRowCount := make(map[int64]int)
|
||||
for node, segments := range m.nodeSegments {
|
||||
rowCount := 0
|
||||
for _, segment := range segments {
|
||||
rowCount += int(segment.GetNumOfRows())
|
||||
}
|
||||
totalRowCount += rowCount
|
||||
nodesRowCount[node] = rowCount
|
||||
}
|
||||
expectAvg := float64(totalRowCount) / float64(nodeCount)
|
||||
|
||||
// calculate worst case, all rows are allocated to only one node
|
||||
worst := float64(nodeCount-1)*expectAvg + float64(totalRowCount) - expectAvg
|
||||
// calculate best case, all rows are allocated meanly
|
||||
nodeWithMoreRows := totalRowCount % nodeCount
|
||||
best := float64(nodeWithMoreRows)*(math.Ceil(expectAvg)-expectAvg) + float64(nodeCount-nodeWithMoreRows)*(expectAvg-math.Floor(expectAvg))
|
||||
|
||||
if worst == best {
|
||||
return 0
|
||||
}
|
||||
var currCost float64
|
||||
for _, rowCount := range nodesRowCount {
|
||||
currCost += math.Abs(float64(rowCount) - expectAvg)
|
||||
}
|
||||
|
||||
// normalization
|
||||
return (currCost - best) / (worst - best)
|
||||
}
|
||||
|
||||
type segmentCountCostModel struct {
|
||||
nodeSegments map[int64][]*meta.Segment
|
||||
}
|
||||
|
||||
func (m *segmentCountCostModel) cost() float64 {
|
||||
nodeCount := len(m.nodeSegments)
|
||||
if nodeCount == 0 {
|
||||
return 0
|
||||
}
|
||||
totalSegmentCount := 0
|
||||
nodeSegmentCount := make(map[int64]int)
|
||||
for node, segments := range m.nodeSegments {
|
||||
totalSegmentCount += len(segments)
|
||||
nodeSegmentCount[node] = len(segments)
|
||||
}
|
||||
expectAvg := float64(totalSegmentCount) / float64(nodeCount)
|
||||
// calculate worst case, all segments are allocated to only one node
|
||||
worst := float64(nodeCount-1)*expectAvg + float64(totalSegmentCount) - expectAvg
|
||||
// calculate best case, all segments are allocated meanly
|
||||
nodeWithMoreRows := totalSegmentCount % nodeCount
|
||||
best := float64(nodeWithMoreRows)*(math.Ceil(expectAvg)-expectAvg) + float64(nodeCount-nodeWithMoreRows)*(expectAvg-math.Floor(expectAvg))
|
||||
|
||||
var currCost float64
|
||||
for _, count := range nodeSegmentCount {
|
||||
currCost += math.Abs(float64(count) - expectAvg)
|
||||
}
|
||||
|
||||
if worst == best {
|
||||
return 0
|
||||
}
|
||||
// normalization
|
||||
return (currCost - best) / (worst - best)
|
||||
}
|
||||
|
||||
func cmpFloat(f1, f2 float64) int {
|
||||
if math.Abs(f1-f2) < 0.001 {
|
||||
return 0
|
||||
}
|
||||
if f1 < f2 {
|
||||
return -1
|
||||
}
|
||||
return 1
|
||||
}
|
||||
|
||||
type generator interface {
|
||||
setPlans(plans []SegmentAssignPlan)
|
||||
setReplicaNodeSegments(replicaNodeSegments map[int64][]*meta.Segment)
|
||||
setGlobalNodeSegments(globalNodeSegments map[int64][]*meta.Segment)
|
||||
setCost(cost float64)
|
||||
getReplicaNodeSegments() map[int64][]*meta.Segment
|
||||
getGlobalNodeSegments() map[int64][]*meta.Segment
|
||||
getCost() float64
|
||||
generatePlans() []SegmentAssignPlan
|
||||
}
|
||||
|
||||
type basePlanGenerator struct {
|
||||
plans []SegmentAssignPlan
|
||||
currClusterCost float64
|
||||
replicaNodeSegments map[int64][]*meta.Segment
|
||||
globalNodeSegments map[int64][]*meta.Segment
|
||||
rowCountCostWeight float64
|
||||
globalRowCountCostWeight float64
|
||||
segmentCountCostWeight float64
|
||||
globalSegmentCountCostWeight float64
|
||||
}
|
||||
|
||||
func newBasePlanGenerator() *basePlanGenerator {
|
||||
return &basePlanGenerator{
|
||||
rowCountCostWeight: params.Params.QueryCoordCfg.RowCountFactor.GetAsFloat(),
|
||||
globalRowCountCostWeight: params.Params.QueryCoordCfg.GlobalRowCountFactor.GetAsFloat(),
|
||||
segmentCountCostWeight: params.Params.QueryCoordCfg.SegmentCountFactor.GetAsFloat(),
|
||||
globalSegmentCountCostWeight: params.Params.QueryCoordCfg.GlobalSegmentCountFactor.GetAsFloat(),
|
||||
}
|
||||
}
|
||||
|
||||
func (g *basePlanGenerator) setPlans(plans []SegmentAssignPlan) {
|
||||
g.plans = plans
|
||||
}
|
||||
|
||||
func (g *basePlanGenerator) setReplicaNodeSegments(replicaNodeSegments map[int64][]*meta.Segment) {
|
||||
g.replicaNodeSegments = replicaNodeSegments
|
||||
}
|
||||
|
||||
func (g *basePlanGenerator) setGlobalNodeSegments(globalNodeSegments map[int64][]*meta.Segment) {
|
||||
g.globalNodeSegments = globalNodeSegments
|
||||
}
|
||||
|
||||
func (g *basePlanGenerator) setCost(cost float64) {
|
||||
g.currClusterCost = cost
|
||||
}
|
||||
|
||||
func (g *basePlanGenerator) getReplicaNodeSegments() map[int64][]*meta.Segment {
|
||||
return g.replicaNodeSegments
|
||||
}
|
||||
|
||||
func (g *basePlanGenerator) getGlobalNodeSegments() map[int64][]*meta.Segment {
|
||||
return g.globalNodeSegments
|
||||
}
|
||||
|
||||
func (g *basePlanGenerator) getCost() float64 {
|
||||
return g.currClusterCost
|
||||
}
|
||||
|
||||
func (g *basePlanGenerator) applyPlans(nodeSegments map[int64][]*meta.Segment, plans []SegmentAssignPlan) map[int64][]*meta.Segment {
|
||||
newCluster := make(map[int64][]*meta.Segment)
|
||||
for k, v := range nodeSegments {
|
||||
newCluster[k] = append(newCluster[k], v...)
|
||||
}
|
||||
for _, p := range plans {
|
||||
for i, s := range newCluster[p.From] {
|
||||
if s.GetID() == p.Segment.ID {
|
||||
newCluster[p.From] = append(newCluster[p.From][:i], newCluster[p.From][i+1:]...)
|
||||
break
|
||||
}
|
||||
}
|
||||
newCluster[p.To] = append(newCluster[p.To], p.Segment)
|
||||
}
|
||||
return newCluster
|
||||
}
|
||||
|
||||
func (g *basePlanGenerator) calClusterCost(replicaNodeSegments, globalNodeSegments map[int64][]*meta.Segment) float64 {
|
||||
replicaRowCountCostModel, replicaSegmentCountCostModel := &rowCountCostModel{replicaNodeSegments}, &segmentCountCostModel{replicaNodeSegments}
|
||||
globalRowCountCostModel, globalSegmentCountCostModel := &rowCountCostModel{globalNodeSegments}, &segmentCountCostModel{globalNodeSegments}
|
||||
replicaCost1, replicaCost2 := replicaRowCountCostModel.cost(), replicaSegmentCountCostModel.cost()
|
||||
globalCost1, globalCost2 := globalRowCountCostModel.cost(), globalSegmentCountCostModel.cost()
|
||||
|
||||
return replicaCost1*g.rowCountCostWeight + replicaCost2*g.segmentCountCostWeight +
|
||||
globalCost1*g.globalRowCountCostWeight + globalCost2*g.globalSegmentCountCostWeight
|
||||
}
|
||||
|
||||
func (g *basePlanGenerator) mergePlans(curr []SegmentAssignPlan, inc []SegmentAssignPlan) []SegmentAssignPlan {
|
||||
// merge plans with the same segment
|
||||
// eg, plan1 is move segment1 from node1 to node2, plan2 is move segment1 from node2 to node3
|
||||
// we should merge plan1 and plan2 to one plan, which is move segment1 from node1 to node3
|
||||
for _, p := range inc {
|
||||
has := false
|
||||
for i := 0; i < len(curr); i++ {
|
||||
if curr[i].Segment.GetID() == p.Segment.GetID() && curr[i].To == p.From {
|
||||
curr[i].To = p.To
|
||||
has = true
|
||||
break
|
||||
}
|
||||
}
|
||||
if !has {
|
||||
curr = append(curr, p)
|
||||
}
|
||||
}
|
||||
return curr
|
||||
}
|
||||
|
||||
type rowCountBasedPlanGenerator struct {
|
||||
*basePlanGenerator
|
||||
maxSteps int
|
||||
isGlobal bool
|
||||
}
|
||||
|
||||
func newRowCountBasedPlanGenerator(maxSteps int, isGlobal bool) *rowCountBasedPlanGenerator {
|
||||
return &rowCountBasedPlanGenerator{
|
||||
basePlanGenerator: newBasePlanGenerator(),
|
||||
maxSteps: maxSteps,
|
||||
isGlobal: isGlobal,
|
||||
}
|
||||
}
|
||||
|
||||
func (g *rowCountBasedPlanGenerator) generatePlans() []SegmentAssignPlan {
|
||||
type nodeWithRowCount struct {
|
||||
id int64
|
||||
count int
|
||||
segments []*meta.Segment
|
||||
}
|
||||
|
||||
if g.currClusterCost == 0 {
|
||||
g.currClusterCost = g.calClusterCost(g.replicaNodeSegments, g.globalNodeSegments)
|
||||
}
|
||||
nodeSegments := g.replicaNodeSegments
|
||||
if g.isGlobal {
|
||||
nodeSegments = g.globalNodeSegments
|
||||
}
|
||||
nodesWithRowCount := make([]*nodeWithRowCount, 0)
|
||||
for node, segments := range g.replicaNodeSegments {
|
||||
rowCount := 0
|
||||
for _, segment := range nodeSegments[node] {
|
||||
rowCount += int(segment.GetNumOfRows())
|
||||
}
|
||||
nodesWithRowCount = append(nodesWithRowCount, &nodeWithRowCount{
|
||||
id: node,
|
||||
count: rowCount,
|
||||
segments: segments,
|
||||
})
|
||||
}
|
||||
|
||||
modified := true
|
||||
for i := 0; i < g.maxSteps; i++ {
|
||||
if modified {
|
||||
sort.Slice(nodesWithRowCount, func(i, j int) bool {
|
||||
return nodesWithRowCount[i].count < nodesWithRowCount[j].count
|
||||
})
|
||||
}
|
||||
maxNode, minNode := nodesWithRowCount[len(nodesWithRowCount)-1], nodesWithRowCount[0]
|
||||
segment := maxNode.segments[rand.Intn(len(maxNode.segments))]
|
||||
plan := SegmentAssignPlan{
|
||||
Segment: segment,
|
||||
From: maxNode.id,
|
||||
To: minNode.id,
|
||||
}
|
||||
newCluster := g.applyPlans(g.replicaNodeSegments, []SegmentAssignPlan{plan})
|
||||
newGlobalCluster := g.applyPlans(g.globalNodeSegments, []SegmentAssignPlan{plan})
|
||||
newCost := g.calClusterCost(newCluster, newGlobalCluster)
|
||||
if cmpFloat(newCost, g.currClusterCost) < 0 {
|
||||
g.currClusterCost = newCost
|
||||
g.replicaNodeSegments = newCluster
|
||||
g.globalNodeSegments = newGlobalCluster
|
||||
maxNode.count -= int(segment.GetNumOfRows())
|
||||
minNode.count += int(segment.GetNumOfRows())
|
||||
for n, segment := range maxNode.segments {
|
||||
if segment.GetID() == plan.Segment.ID {
|
||||
maxNode.segments = append(maxNode.segments[:n], maxNode.segments[n+1:]...)
|
||||
break
|
||||
}
|
||||
}
|
||||
minNode.segments = append(minNode.segments, segment)
|
||||
g.plans = g.mergePlans(g.plans, []SegmentAssignPlan{plan})
|
||||
modified = true
|
||||
} else {
|
||||
modified = false
|
||||
}
|
||||
}
|
||||
return g.plans
|
||||
}
|
||||
|
||||
type segmentCountBasedPlanGenerator struct {
|
||||
*basePlanGenerator
|
||||
maxSteps int
|
||||
isGlobal bool
|
||||
}
|
||||
|
||||
func newSegmentCountBasedPlanGenerator(maxSteps int, isGlobal bool) *segmentCountBasedPlanGenerator {
|
||||
return &segmentCountBasedPlanGenerator{
|
||||
basePlanGenerator: newBasePlanGenerator(),
|
||||
maxSteps: maxSteps,
|
||||
isGlobal: isGlobal,
|
||||
}
|
||||
}
|
||||
|
||||
func (g *segmentCountBasedPlanGenerator) generatePlans() []SegmentAssignPlan {
|
||||
type nodeWithSegmentCount struct {
|
||||
id int64
|
||||
count int
|
||||
segments []*meta.Segment
|
||||
}
|
||||
|
||||
if g.currClusterCost == 0 {
|
||||
g.currClusterCost = g.calClusterCost(g.replicaNodeSegments, g.globalNodeSegments)
|
||||
}
|
||||
|
||||
nodeSegments := g.replicaNodeSegments
|
||||
if g.isGlobal {
|
||||
nodeSegments = g.globalNodeSegments
|
||||
}
|
||||
nodesWithSegmentCount := make([]*nodeWithSegmentCount, 0)
|
||||
for node, segments := range g.replicaNodeSegments {
|
||||
nodesWithSegmentCount = append(nodesWithSegmentCount, &nodeWithSegmentCount{
|
||||
id: node,
|
||||
count: len(nodeSegments[node]),
|
||||
segments: segments,
|
||||
})
|
||||
}
|
||||
|
||||
modified := true
|
||||
for i := 0; i < g.maxSteps; i++ {
|
||||
if modified {
|
||||
sort.Slice(nodesWithSegmentCount, func(i, j int) bool {
|
||||
return nodesWithSegmentCount[i].count < nodesWithSegmentCount[j].count
|
||||
})
|
||||
}
|
||||
maxNode, minNode := nodesWithSegmentCount[len(nodesWithSegmentCount)-1], nodesWithSegmentCount[0]
|
||||
segment := maxNode.segments[rand.Intn(len(maxNode.segments))]
|
||||
plan := SegmentAssignPlan{
|
||||
Segment: segment,
|
||||
From: maxNode.id,
|
||||
To: minNode.id,
|
||||
}
|
||||
newCluster := g.applyPlans(g.replicaNodeSegments, []SegmentAssignPlan{plan})
|
||||
newGlobalCluster := g.applyPlans(g.globalNodeSegments, []SegmentAssignPlan{plan})
|
||||
newCost := g.calClusterCost(newCluster, newGlobalCluster)
|
||||
if cmpFloat(newCost, g.currClusterCost) < 0 {
|
||||
g.currClusterCost = newCost
|
||||
g.replicaNodeSegments = newCluster
|
||||
g.globalNodeSegments = newGlobalCluster
|
||||
maxNode.count -= 1
|
||||
minNode.count += 1
|
||||
for n, segment := range maxNode.segments {
|
||||
if segment.GetID() == plan.Segment.ID {
|
||||
maxNode.segments = append(maxNode.segments[:n], maxNode.segments[n+1:]...)
|
||||
break
|
||||
}
|
||||
}
|
||||
minNode.segments = append(minNode.segments, segment)
|
||||
g.plans = g.mergePlans(g.plans, []SegmentAssignPlan{plan})
|
||||
modified = true
|
||||
} else {
|
||||
modified = false
|
||||
}
|
||||
}
|
||||
return g.plans
|
||||
}
|
||||
|
||||
type planType int
|
||||
|
||||
const (
|
||||
movePlan planType = iota + 1
|
||||
swapPlan
|
||||
)
|
||||
|
||||
type randomPlanGenerator struct {
|
||||
*basePlanGenerator
|
||||
maxSteps int
|
||||
}
|
||||
|
||||
func newRandomPlanGenerator(maxSteps int) *randomPlanGenerator {
|
||||
return &randomPlanGenerator{
|
||||
basePlanGenerator: newBasePlanGenerator(),
|
||||
maxSteps: maxSteps,
|
||||
}
|
||||
}
|
||||
|
||||
func (g *randomPlanGenerator) generatePlans() []SegmentAssignPlan {
|
||||
g.currClusterCost = g.calClusterCost(g.replicaNodeSegments, g.globalNodeSegments)
|
||||
nodes := lo.Keys(g.replicaNodeSegments)
|
||||
for i := 0; i < g.maxSteps; i++ {
|
||||
// random select two nodes and two segments
|
||||
node1 := nodes[rand.Intn(len(nodes))]
|
||||
node2 := nodes[rand.Intn(len(nodes))]
|
||||
if node1 == node2 {
|
||||
continue
|
||||
}
|
||||
segments1 := g.replicaNodeSegments[node1]
|
||||
segments2 := g.replicaNodeSegments[node2]
|
||||
segment1 := segments1[rand.Intn(len(segments1))]
|
||||
segment2 := segments2[rand.Intn(len(segments2))]
|
||||
|
||||
// random select plan type, for move type, we move segment1 to node2; for swap type, we swap segment1 and segment2
|
||||
plans := make([]SegmentAssignPlan, 0)
|
||||
planType := planType(rand.Intn(2) + 1)
|
||||
if planType == movePlan {
|
||||
plan := SegmentAssignPlan{
|
||||
From: node1,
|
||||
To: node2,
|
||||
Segment: segment1,
|
||||
}
|
||||
plans = append(plans, plan)
|
||||
} else {
|
||||
plan1 := SegmentAssignPlan{
|
||||
From: node1,
|
||||
To: node2,
|
||||
Segment: segment1,
|
||||
}
|
||||
plan2 := SegmentAssignPlan{
|
||||
From: node2,
|
||||
To: node1,
|
||||
Segment: segment2,
|
||||
}
|
||||
plans = append(plans, plan1, plan2)
|
||||
}
|
||||
|
||||
// validate the plan, if the plan is valid, we apply the plan and update the cluster cost
|
||||
newCluster := g.applyPlans(g.replicaNodeSegments, plans)
|
||||
newGlobalCluster := g.applyPlans(g.globalNodeSegments, plans)
|
||||
newCost := g.calClusterCost(newCluster, newGlobalCluster)
|
||||
if cmpFloat(newCost, g.currClusterCost) < 0 {
|
||||
g.currClusterCost = newCost
|
||||
g.replicaNodeSegments = newCluster
|
||||
g.globalNodeSegments = newGlobalCluster
|
||||
g.plans = g.mergePlans(g.plans, plans)
|
||||
}
|
||||
}
|
||||
return g.plans
|
||||
}
|
||||
|
||||
type MultiTargetBalancer struct {
|
||||
*ScoreBasedBalancer
|
||||
dist *meta.DistributionManager
|
||||
targetMgr *meta.TargetManager
|
||||
}
|
||||
|
||||
func (b *MultiTargetBalancer) BalanceReplica(replica *meta.Replica) ([]SegmentAssignPlan, []ChannelAssignPlan) {
|
||||
log := log.With(
|
||||
zap.Int64("collection", replica.CollectionID),
|
||||
zap.Int64("replica id", replica.Replica.GetID()),
|
||||
zap.String("replica group", replica.Replica.GetResourceGroup()),
|
||||
)
|
||||
nodes := replica.GetNodes()
|
||||
if len(nodes) == 0 {
|
||||
return nil, nil
|
||||
}
|
||||
|
||||
outboundNodes := b.meta.ResourceManager.CheckOutboundNodes(replica)
|
||||
onlineNodes := make([]int64, 0)
|
||||
offlineNodes := make([]int64, 0)
|
||||
for _, nid := range nodes {
|
||||
if isStopping, err := b.nodeManager.IsStoppingNode(nid); err != nil {
|
||||
log.Info("not existed node", zap.Int64("nid", nid), zap.Error(err))
|
||||
continue
|
||||
} else if isStopping {
|
||||
offlineNodes = append(offlineNodes, nid)
|
||||
} else if outboundNodes.Contain(nid) {
|
||||
// if node is stop or transfer to other rg
|
||||
log.RatedInfo(10, "meet outbound node, try to move out all segment/channel", zap.Int64("node", nid))
|
||||
offlineNodes = append(offlineNodes, nid)
|
||||
} else {
|
||||
onlineNodes = append(onlineNodes, nid)
|
||||
}
|
||||
}
|
||||
|
||||
if len(nodes) == len(offlineNodes) || len(onlineNodes) == 0 {
|
||||
// no available nodes to balance
|
||||
return nil, nil
|
||||
}
|
||||
|
||||
// print current distribution before generating plans
|
||||
segmentPlans, channelPlans := make([]SegmentAssignPlan, 0), make([]ChannelAssignPlan, 0)
|
||||
if len(offlineNodes) != 0 {
|
||||
log.Info("Handle stopping nodes",
|
||||
zap.Any("stopping nodes", offlineNodes),
|
||||
zap.Any("available nodes", onlineNodes),
|
||||
)
|
||||
// handle stopped nodes here, have to assign segments on stopping nodes to nodes with the smallest score
|
||||
channelPlans = append(channelPlans, b.genStoppingChannelPlan(replica, onlineNodes, offlineNodes)...)
|
||||
if len(channelPlans) == 0 {
|
||||
segmentPlans = append(segmentPlans, b.genStoppingSegmentPlan(replica, onlineNodes, offlineNodes)...)
|
||||
}
|
||||
} else {
|
||||
if paramtable.Get().QueryCoordCfg.AutoBalanceChannel.GetAsBool() {
|
||||
channelPlans = append(channelPlans, b.genChannelPlan(replica, onlineNodes)...)
|
||||
}
|
||||
|
||||
if len(channelPlans) == 0 {
|
||||
segmentPlans = b.genSegmentPlan(replica)
|
||||
}
|
||||
}
|
||||
|
||||
return segmentPlans, channelPlans
|
||||
}
|
||||
|
||||
func (b *MultiTargetBalancer) genSegmentPlan(replica *meta.Replica) []SegmentAssignPlan {
|
||||
// get segments distribution on replica level and global level
|
||||
nodeSegments := make(map[int64][]*meta.Segment)
|
||||
globalNodeSegments := make(map[int64][]*meta.Segment)
|
||||
for _, node := range replica.Nodes {
|
||||
dist := b.dist.SegmentDistManager.GetByCollectionAndNode(replica.GetCollectionID(), node)
|
||||
segments := lo.Filter(dist, func(segment *meta.Segment, _ int) bool {
|
||||
return b.targetMgr.GetSealedSegment(segment.GetCollectionID(), segment.GetID(), meta.CurrentTarget) != nil &&
|
||||
b.targetMgr.GetSealedSegment(segment.GetCollectionID(), segment.GetID(), meta.NextTarget) != nil &&
|
||||
segment.GetLevel() != datapb.SegmentLevel_L0
|
||||
})
|
||||
nodeSegments[node] = segments
|
||||
globalNodeSegments[node] = b.dist.SegmentDistManager.GetByNode(node)
|
||||
}
|
||||
|
||||
return b.genPlanByDistributions(nodeSegments, globalNodeSegments)
|
||||
}
|
||||
|
||||
func (b *MultiTargetBalancer) genPlanByDistributions(nodeSegments, globalNodeSegments map[int64][]*meta.Segment) []SegmentAssignPlan {
|
||||
// create generators
|
||||
// we have 3 types of generators: row count, segment count, random
|
||||
// for row count based and segment count based generator, we have 2 types of generators: replica level and global level
|
||||
generators := make([]generator, 0)
|
||||
generators = append(generators,
|
||||
newRowCountBasedPlanGenerator(params.Params.QueryCoordCfg.RowCountMaxSteps.GetAsInt(), false),
|
||||
newRowCountBasedPlanGenerator(params.Params.QueryCoordCfg.RowCountMaxSteps.GetAsInt(), true),
|
||||
newSegmentCountBasedPlanGenerator(params.Params.QueryCoordCfg.SegmentCountMaxSteps.GetAsInt(), false),
|
||||
newSegmentCountBasedPlanGenerator(params.Params.QueryCoordCfg.SegmentCountMaxSteps.GetAsInt(), true),
|
||||
newRandomPlanGenerator(params.Params.QueryCoordCfg.RandomMaxSteps.GetAsInt()),
|
||||
)
|
||||
|
||||
// run generators sequentially to generate plans
|
||||
var cost float64
|
||||
var plans []SegmentAssignPlan
|
||||
for _, generator := range generators {
|
||||
generator.setCost(cost)
|
||||
generator.setPlans(plans)
|
||||
generator.setReplicaNodeSegments(nodeSegments)
|
||||
generator.setGlobalNodeSegments(globalNodeSegments)
|
||||
plans = generator.generatePlans()
|
||||
cost = generator.getCost()
|
||||
nodeSegments = generator.getReplicaNodeSegments()
|
||||
globalNodeSegments = generator.getGlobalNodeSegments()
|
||||
}
|
||||
return plans
|
||||
}
|
||||
|
||||
func NewMultiTargetBalancer(scheduler task.Scheduler, nodeManager *session.NodeManager, dist *meta.DistributionManager, meta *meta.Meta, targetMgr *meta.TargetManager) *MultiTargetBalancer {
|
||||
return &MultiTargetBalancer{
|
||||
ScoreBasedBalancer: NewScoreBasedBalancer(scheduler, nodeManager, dist, meta, targetMgr),
|
||||
dist: dist,
|
||||
targetMgr: targetMgr,
|
||||
}
|
||||
}
|
358
internal/querycoordv2/balance/multi_target_balancer_test.go
Normal file
358
internal/querycoordv2/balance/multi_target_balancer_test.go
Normal file
@ -0,0 +1,358 @@
|
||||
package balance
|
||||
|
||||
import (
|
||||
"math/rand"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"github.com/samber/lo"
|
||||
"github.com/stretchr/testify/suite"
|
||||
|
||||
"github.com/milvus-io/milvus/internal/proto/datapb"
|
||||
"github.com/milvus-io/milvus/internal/querycoordv2/meta"
|
||||
"github.com/milvus-io/milvus/pkg/util/paramtable"
|
||||
"github.com/milvus-io/milvus/pkg/util/typeutil"
|
||||
)
|
||||
|
||||
type MultiTargetBalancerTestSuite struct {
|
||||
suite.Suite
|
||||
}
|
||||
|
||||
func (suite *MultiTargetBalancerTestSuite) SetupSuite() {
|
||||
paramtable.Init()
|
||||
rand.Seed(time.Now().UnixNano())
|
||||
}
|
||||
|
||||
func (suite *MultiTargetBalancerTestSuite) TestRowCountCostModel() {
|
||||
cases := [][]struct {
|
||||
nodeID int64
|
||||
segmentID int64
|
||||
rowCount int64
|
||||
}{
|
||||
// case 1, empty cluster
|
||||
{},
|
||||
// case 2
|
||||
// node 0: 30, node 1: 0
|
||||
{{0, 1, 30}, {1, 0, 0}},
|
||||
// case 3
|
||||
// node 0: 30, node 1: 30
|
||||
{{0, 1, 30}, {1, 2, 30}},
|
||||
// case 4
|
||||
// node 0: 30, node 1: 20, node 2: 10
|
||||
{{0, 1, 30}, {1, 2, 20}, {2, 3, 10}},
|
||||
// case 5
|
||||
{{0, 1, 30}},
|
||||
}
|
||||
|
||||
expects := []float64{
|
||||
0,
|
||||
1,
|
||||
0,
|
||||
0.25,
|
||||
0,
|
||||
}
|
||||
|
||||
for i, c := range cases {
|
||||
nodeSegments := make(map[int64][]*meta.Segment)
|
||||
for _, v := range c {
|
||||
nodeSegments[v.nodeID] = append(nodeSegments[v.nodeID],
|
||||
&meta.Segment{SegmentInfo: &datapb.SegmentInfo{ID: v.segmentID, NumOfRows: v.rowCount}},
|
||||
)
|
||||
}
|
||||
model := &rowCountCostModel{nodeSegments: nodeSegments}
|
||||
suite.InDelta(expects[i], model.cost(), 0.01, "case %d", i+1)
|
||||
}
|
||||
}
|
||||
|
||||
func (suite *MultiTargetBalancerTestSuite) TestSegmentCountCostModel() {
|
||||
cases := [][]struct {
|
||||
nodeID int64
|
||||
segmentCount int
|
||||
}{
|
||||
{},
|
||||
{{0, 10}, {1, 0}},
|
||||
{{0, 10}, {1, 10}},
|
||||
{{0, 30}, {1, 20}, {2, 10}},
|
||||
{{0, 10}},
|
||||
}
|
||||
|
||||
expects := []float64{
|
||||
0,
|
||||
1,
|
||||
0,
|
||||
0.25,
|
||||
0,
|
||||
}
|
||||
for i, c := range cases {
|
||||
nodeSegments := make(map[int64][]*meta.Segment)
|
||||
for _, v := range c {
|
||||
nodeSegments[v.nodeID] = make([]*meta.Segment, 0)
|
||||
for j := 0; j < v.segmentCount; j++ {
|
||||
nodeSegments[v.nodeID] = append(nodeSegments[v.nodeID],
|
||||
&meta.Segment{SegmentInfo: &datapb.SegmentInfo{ID: int64(j)}},
|
||||
)
|
||||
}
|
||||
}
|
||||
model := &segmentCountCostModel{nodeSegments: nodeSegments}
|
||||
suite.InDelta(expects[i], model.cost(), 0.01, "case %d", i+1)
|
||||
}
|
||||
}
|
||||
|
||||
func (suite *MultiTargetBalancerTestSuite) TestBaseGeneratorApplyPlans() {
|
||||
distribution := []struct {
|
||||
nodeID []int64
|
||||
segments [][]int64
|
||||
}{
|
||||
{[]int64{0, 1}, [][]int64{{1}, {2}}},
|
||||
}
|
||||
|
||||
casePlans := []struct {
|
||||
segments []int64
|
||||
from []int64
|
||||
to []int64
|
||||
}{
|
||||
{[]int64{1}, []int64{0}, []int64{1}},
|
||||
}
|
||||
|
||||
expects := []struct {
|
||||
nodeID []int64
|
||||
segments [][]int64
|
||||
}{
|
||||
{[]int64{0, 1}, [][]int64{{}, {1, 2}}},
|
||||
}
|
||||
|
||||
for i := 0; i < len(casePlans); i++ {
|
||||
nodeSegments := make(map[int64][]*meta.Segment)
|
||||
appliedPlans := make([]SegmentAssignPlan, 0)
|
||||
d := distribution[i]
|
||||
for i, nodeID := range d.nodeID {
|
||||
nodeSegments[nodeID] = make([]*meta.Segment, 0)
|
||||
for _, segmentID := range d.segments[i] {
|
||||
nodeSegments[nodeID] = append(nodeSegments[nodeID],
|
||||
&meta.Segment{SegmentInfo: &datapb.SegmentInfo{ID: segmentID}},
|
||||
)
|
||||
}
|
||||
}
|
||||
|
||||
p := casePlans[i]
|
||||
for j := 0; j < len(p.segments); j++ {
|
||||
appliedPlans = append(appliedPlans, SegmentAssignPlan{
|
||||
Segment: &meta.Segment{SegmentInfo: &datapb.SegmentInfo{ID: p.segments[i]}},
|
||||
From: p.from[i],
|
||||
To: p.to[i],
|
||||
})
|
||||
}
|
||||
|
||||
generator := &basePlanGenerator{}
|
||||
newNodeSegments := generator.applyPlans(nodeSegments, appliedPlans)
|
||||
expected := expects[i]
|
||||
for i := 0; i < len(expected.nodeID); i++ {
|
||||
newSegmentIDs := lo.FlatMap(newNodeSegments[int64(i)], func(segment *meta.Segment, _ int) []int64 {
|
||||
return []int64{segment.ID}
|
||||
})
|
||||
suite.ElementsMatch(expected.segments[i], newSegmentIDs)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func (suite *MultiTargetBalancerTestSuite) TestBaseGeneratorMergePlans() {
|
||||
cases := [][2]struct {
|
||||
segment []int64
|
||||
from []int64
|
||||
to []int64
|
||||
}{
|
||||
{{[]int64{1}, []int64{1}, []int64{2}}, {[]int64{1}, []int64{2}, []int64{3}}},
|
||||
{{[]int64{1}, []int64{1}, []int64{2}}, {[]int64{2}, []int64{2}, []int64{3}}},
|
||||
}
|
||||
|
||||
expects := []struct {
|
||||
segment []int64
|
||||
from []int64
|
||||
to []int64
|
||||
}{
|
||||
{[]int64{1}, []int64{1}, []int64{3}},
|
||||
{[]int64{1, 2}, []int64{1, 2}, []int64{2, 3}},
|
||||
}
|
||||
|
||||
for i := 0; i < len(cases); i++ {
|
||||
planGenerator := &basePlanGenerator{}
|
||||
curr := make([]SegmentAssignPlan, 0)
|
||||
inc := make([]SegmentAssignPlan, 0)
|
||||
for j := 0; j < len(cases[i][0].segment); j++ {
|
||||
curr = append(curr, SegmentAssignPlan{
|
||||
Segment: &meta.Segment{SegmentInfo: &datapb.SegmentInfo{ID: cases[i][0].segment[j]}},
|
||||
From: cases[i][0].from[j],
|
||||
To: cases[i][0].to[j],
|
||||
})
|
||||
}
|
||||
for j := 0; j < len(cases[i][1].segment); j++ {
|
||||
inc = append(inc, SegmentAssignPlan{
|
||||
Segment: &meta.Segment{SegmentInfo: &datapb.SegmentInfo{ID: cases[i][1].segment[j]}},
|
||||
From: cases[i][1].from[j],
|
||||
To: cases[i][1].to[j],
|
||||
})
|
||||
}
|
||||
|
||||
res := planGenerator.mergePlans(curr, inc)
|
||||
|
||||
var segment []int64
|
||||
var from []int64
|
||||
var to []int64
|
||||
for _, p := range res {
|
||||
segment = append(segment, p.Segment.ID)
|
||||
from = append(from, p.From)
|
||||
to = append(to, p.To)
|
||||
}
|
||||
suite.ElementsMatch(segment, expects[i].segment, "case %d", i+1)
|
||||
suite.ElementsMatch(from, expects[i].from, "case %d", i+1)
|
||||
suite.ElementsMatch(to, expects[i].to, "case %d", i+1)
|
||||
}
|
||||
}
|
||||
|
||||
func (suite *MultiTargetBalancerTestSuite) TestRowCountPlanGenerator() {
|
||||
cases := []struct {
|
||||
nodeSegments map[int64][]*meta.Segment
|
||||
expectPlanCount int
|
||||
expectCost float64
|
||||
}{
|
||||
// case 1
|
||||
{
|
||||
map[int64][]*meta.Segment{
|
||||
1: {
|
||||
{SegmentInfo: &datapb.SegmentInfo{ID: 1, NumOfRows: 10}},
|
||||
{SegmentInfo: &datapb.SegmentInfo{ID: 2, NumOfRows: 10}},
|
||||
},
|
||||
2: {},
|
||||
},
|
||||
1,
|
||||
0,
|
||||
},
|
||||
// case 2
|
||||
{
|
||||
map[int64][]*meta.Segment{
|
||||
1: {
|
||||
{SegmentInfo: &datapb.SegmentInfo{ID: 1, NumOfRows: 10}},
|
||||
},
|
||||
2: {
|
||||
{SegmentInfo: &datapb.SegmentInfo{ID: 2, NumOfRows: 10}},
|
||||
},
|
||||
},
|
||||
0,
|
||||
0,
|
||||
},
|
||||
}
|
||||
|
||||
for i, c := range cases {
|
||||
generator := newRowCountBasedPlanGenerator(10, false)
|
||||
generator.setReplicaNodeSegments(c.nodeSegments)
|
||||
generator.setGlobalNodeSegments(c.nodeSegments)
|
||||
plans := generator.generatePlans()
|
||||
suite.Len(plans, c.expectPlanCount, "case %d", i+1)
|
||||
suite.InDelta(c.expectCost, generator.currClusterCost, 0.001, "case %d", i+1)
|
||||
}
|
||||
}
|
||||
|
||||
func (suite *MultiTargetBalancerTestSuite) TestSegmentCountPlanGenerator() {
|
||||
cases := []struct {
|
||||
nodeSegments map[int64][]*meta.Segment
|
||||
expectPlanCount int
|
||||
expectCost float64
|
||||
}{
|
||||
// case 1
|
||||
{
|
||||
map[int64][]*meta.Segment{
|
||||
1: {
|
||||
{SegmentInfo: &datapb.SegmentInfo{ID: 1, NumOfRows: 10}},
|
||||
{SegmentInfo: &datapb.SegmentInfo{ID: 2, NumOfRows: 10}},
|
||||
},
|
||||
2: {},
|
||||
},
|
||||
1,
|
||||
0,
|
||||
},
|
||||
// case 2
|
||||
{
|
||||
map[int64][]*meta.Segment{
|
||||
1: {
|
||||
{SegmentInfo: &datapb.SegmentInfo{ID: 1, NumOfRows: 10}},
|
||||
},
|
||||
2: {
|
||||
{SegmentInfo: &datapb.SegmentInfo{ID: 2, NumOfRows: 10}},
|
||||
},
|
||||
},
|
||||
0,
|
||||
0,
|
||||
},
|
||||
}
|
||||
|
||||
for i, c := range cases {
|
||||
generator := newSegmentCountBasedPlanGenerator(10, false)
|
||||
generator.setReplicaNodeSegments(c.nodeSegments)
|
||||
generator.setGlobalNodeSegments(c.nodeSegments)
|
||||
plans := generator.generatePlans()
|
||||
suite.Len(plans, c.expectPlanCount, "case %d", i+1)
|
||||
suite.InDelta(c.expectCost, generator.currClusterCost, 0.001, "case %d", i+1)
|
||||
}
|
||||
}
|
||||
|
||||
func (suite *MultiTargetBalancerTestSuite) TestRandomPlanGenerator() {
|
||||
cases := []struct {
|
||||
nodeSegments map[int64][]*meta.Segment
|
||||
expectCost float64
|
||||
}{
|
||||
// case 1
|
||||
{
|
||||
map[int64][]*meta.Segment{
|
||||
1: {
|
||||
{SegmentInfo: &datapb.SegmentInfo{ID: 1, NumOfRows: 20}}, {SegmentInfo: &datapb.SegmentInfo{ID: 2, NumOfRows: 30}},
|
||||
},
|
||||
2: {
|
||||
{SegmentInfo: &datapb.SegmentInfo{ID: 3, NumOfRows: 20}},
|
||||
{SegmentInfo: &datapb.SegmentInfo{ID: 4, NumOfRows: 10}},
|
||||
},
|
||||
},
|
||||
0,
|
||||
},
|
||||
}
|
||||
|
||||
for _, c := range cases {
|
||||
generator := newRandomPlanGenerator(100) // set a large enough random steps
|
||||
generator.setReplicaNodeSegments(c.nodeSegments)
|
||||
generator.setGlobalNodeSegments(c.nodeSegments)
|
||||
generator.generatePlans()
|
||||
suite.InDelta(c.expectCost, generator.currClusterCost, 0.001)
|
||||
}
|
||||
}
|
||||
|
||||
func (suite *MultiTargetBalancerTestSuite) TestPlanNoConflict() {
|
||||
nodeSegments := make(map[int64][]*meta.Segment)
|
||||
totalCount := 0
|
||||
// 10 nodes, at most 100 segments, at most 1000 rows
|
||||
for i := 0; i < 10; i++ {
|
||||
segNum := rand.Intn(100)
|
||||
for j := 0; j < segNum; j++ {
|
||||
rowCount := rand.Intn(1000)
|
||||
nodeSegments[int64(i)] = append(nodeSegments[int64(i)], &meta.Segment{
|
||||
SegmentInfo: &datapb.SegmentInfo{
|
||||
ID: int64(i*1000 + j),
|
||||
NumOfRows: int64(rowCount),
|
||||
},
|
||||
})
|
||||
totalCount += rowCount
|
||||
}
|
||||
}
|
||||
|
||||
balancer := &MultiTargetBalancer{}
|
||||
plans := balancer.genPlanByDistributions(nodeSegments, nodeSegments)
|
||||
segmentSet := typeutil.NewSet[int64]()
|
||||
for _, p := range plans {
|
||||
suite.False(segmentSet.Contain(p.Segment.ID))
|
||||
segmentSet.Insert(p.Segment.ID)
|
||||
suite.NotEqual(p.From, p.To)
|
||||
}
|
||||
}
|
||||
|
||||
func TestMultiTargetBalancerTestSuite(t *testing.T) {
|
||||
s := new(MultiTargetBalancerTestSuite)
|
||||
suite.Run(t, s)
|
||||
}
|
@ -1362,6 +1362,12 @@ type queryCoordConfig struct {
|
||||
OverloadedMemoryThresholdPercentage ParamItem `refreshable:"true"`
|
||||
BalanceIntervalSeconds ParamItem `refreshable:"true"`
|
||||
MemoryUsageMaxDifferencePercentage ParamItem `refreshable:"true"`
|
||||
RowCountFactor ParamItem `refreshable:"true"`
|
||||
SegmentCountFactor ParamItem `refreshable:"true"`
|
||||
GlobalSegmentCountFactor ParamItem `refreshable:"true"`
|
||||
SegmentCountMaxSteps ParamItem `refreshable:"true"`
|
||||
RowCountMaxSteps ParamItem `refreshable:"true"`
|
||||
RandomMaxSteps ParamItem `refreshable:"true"`
|
||||
GrowingRowCountWeight ParamItem `refreshable:"true"`
|
||||
|
||||
SegmentCheckInterval ParamItem `refreshable:"true"`
|
||||
@ -1481,6 +1487,66 @@ func (p *queryCoordConfig) init(base *BaseTable) {
|
||||
}
|
||||
p.GlobalRowCountFactor.Init(base.mgr)
|
||||
|
||||
p.RowCountFactor = ParamItem{
|
||||
Key: "queryCoord.rowCountFactor",
|
||||
Version: "2.3.0",
|
||||
DefaultValue: "0.4",
|
||||
PanicIfEmpty: true,
|
||||
Doc: "the row count weight used when balancing segments among queryNodes",
|
||||
Export: true,
|
||||
}
|
||||
p.RowCountFactor.Init(base.mgr)
|
||||
|
||||
p.SegmentCountFactor = ParamItem{
|
||||
Key: "queryCoord.segmentCountFactor",
|
||||
Version: "2.3.0",
|
||||
DefaultValue: "0.4",
|
||||
PanicIfEmpty: true,
|
||||
Doc: "the segment count weight used when balancing segments among queryNodes",
|
||||
Export: true,
|
||||
}
|
||||
p.SegmentCountFactor.Init(base.mgr)
|
||||
|
||||
p.GlobalSegmentCountFactor = ParamItem{
|
||||
Key: "queryCoord.globalSegmentCountFactor",
|
||||
Version: "2.3.0",
|
||||
DefaultValue: "0.1",
|
||||
PanicIfEmpty: true,
|
||||
Doc: "the segment count weight used when balancing segments among queryNodes",
|
||||
Export: true,
|
||||
}
|
||||
p.GlobalSegmentCountFactor.Init(base.mgr)
|
||||
|
||||
p.SegmentCountMaxSteps = ParamItem{
|
||||
Key: "queryCoord.segmentCountMaxSteps",
|
||||
Version: "2.3.0",
|
||||
DefaultValue: "50",
|
||||
PanicIfEmpty: true,
|
||||
Doc: "segment count based plan generator max steps",
|
||||
Export: true,
|
||||
}
|
||||
p.SegmentCountMaxSteps.Init(base.mgr)
|
||||
|
||||
p.RowCountMaxSteps = ParamItem{
|
||||
Key: "queryCoord.rowCountMaxSteps",
|
||||
Version: "2.3.0",
|
||||
DefaultValue: "50",
|
||||
PanicIfEmpty: true,
|
||||
Doc: "segment count based plan generator max steps",
|
||||
Export: true,
|
||||
}
|
||||
p.RowCountMaxSteps.Init(base.mgr)
|
||||
|
||||
p.RandomMaxSteps = ParamItem{
|
||||
Key: "queryCoord.randomMaxSteps",
|
||||
Version: "2.3.0",
|
||||
DefaultValue: "10",
|
||||
PanicIfEmpty: true,
|
||||
Doc: "segment count based plan generator max steps",
|
||||
Export: true,
|
||||
}
|
||||
p.RandomMaxSteps.Init(base.mgr)
|
||||
|
||||
p.ScoreUnbalanceTolerationFactor = ParamItem{
|
||||
Key: "queryCoord.scoreUnbalanceTolerationFactor",
|
||||
Version: "2.0.0",
|
||||
|
Loading…
Reference in New Issue
Block a user