mirror of
https://gitee.com/milvus-io/milvus.git
synced 2024-11-30 02:48:45 +08:00
196a7986b3
issue: #34495 --------- Signed-off-by: Cai Zhang <cai.zhang@zilliz.com>
255 lines
10 KiB
Go
255 lines
10 KiB
Go
// Licensed to the LF AI & Data foundation under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package datacoord
|
|
|
|
import (
|
|
"fmt"
|
|
"math/rand"
|
|
"sort"
|
|
"time"
|
|
|
|
"github.com/cockroachdb/errors"
|
|
"github.com/samber/lo"
|
|
|
|
"github.com/milvus-io/milvus-proto/go-api/v2/commonpb"
|
|
"github.com/milvus-io/milvus-proto/go-api/v2/schemapb"
|
|
"github.com/milvus-io/milvus/internal/proto/datapb"
|
|
"github.com/milvus-io/milvus/pkg/util/paramtable"
|
|
"github.com/milvus-io/milvus/pkg/util/tsoutil"
|
|
"github.com/milvus-io/milvus/pkg/util/typeutil"
|
|
)
|
|
|
|
type calUpperLimitPolicy func(schema *schemapb.CollectionSchema) (int, error)
|
|
|
|
func calBySchemaPolicy(schema *schemapb.CollectionSchema) (int, error) {
|
|
if schema == nil {
|
|
return -1, errors.New("nil schema")
|
|
}
|
|
sizePerRecord, err := typeutil.EstimateSizePerRecord(schema)
|
|
if err != nil {
|
|
return -1, err
|
|
}
|
|
// check zero value, preventing panicking
|
|
if sizePerRecord == 0 {
|
|
return -1, errors.New("zero size record schema found")
|
|
}
|
|
threshold := Params.DataCoordCfg.SegmentMaxSize.GetAsFloat() * 1024 * 1024
|
|
return int(threshold / float64(sizePerRecord)), nil
|
|
}
|
|
|
|
func calBySchemaPolicyWithDiskIndex(schema *schemapb.CollectionSchema) (int, error) {
|
|
if schema == nil {
|
|
return -1, errors.New("nil schema")
|
|
}
|
|
sizePerRecord, err := typeutil.EstimateSizePerRecord(schema)
|
|
if err != nil {
|
|
return -1, err
|
|
}
|
|
// check zero value, preventing panicking
|
|
if sizePerRecord == 0 {
|
|
return -1, errors.New("zero size record schema found")
|
|
}
|
|
threshold := Params.DataCoordCfg.DiskSegmentMaxSize.GetAsFloat() * 1024 * 1024
|
|
return int(threshold / float64(sizePerRecord)), nil
|
|
}
|
|
|
|
func calBySegmentSizePolicy(schema *schemapb.CollectionSchema, segmentSize int64) (int, error) {
|
|
if schema == nil {
|
|
return -1, errors.New("nil schema")
|
|
}
|
|
sizePerRecord, err := typeutil.EstimateSizePerRecord(schema)
|
|
if err != nil {
|
|
return -1, err
|
|
}
|
|
// check zero value, preventing panicking
|
|
if sizePerRecord == 0 {
|
|
return -1, errors.New("zero size record schema found")
|
|
}
|
|
return int(segmentSize) / sizePerRecord, nil
|
|
}
|
|
|
|
// AllocatePolicy helper function definition to allocate Segment space
|
|
type AllocatePolicy func(segments []*SegmentInfo, count int64,
|
|
maxCountPerL1Segment int64, level datapb.SegmentLevel) ([]*Allocation, []*Allocation)
|
|
|
|
// alloca policy for L1 segment
|
|
func AllocatePolicyL1(segments []*SegmentInfo, count int64,
|
|
maxCountPerL1Segment int64, level datapb.SegmentLevel,
|
|
) ([]*Allocation, []*Allocation) {
|
|
newSegmentAllocations := make([]*Allocation, 0)
|
|
existedSegmentAllocations := make([]*Allocation, 0)
|
|
// create new segment if count >= max num
|
|
for count >= maxCountPerL1Segment {
|
|
allocation := getAllocation(maxCountPerL1Segment)
|
|
newSegmentAllocations = append(newSegmentAllocations, allocation)
|
|
count -= maxCountPerL1Segment
|
|
}
|
|
|
|
// allocate space for remaining count
|
|
if count == 0 {
|
|
return newSegmentAllocations, existedSegmentAllocations
|
|
}
|
|
for _, segment := range segments {
|
|
var allocSize int64
|
|
for _, allocation := range segment.allocations {
|
|
allocSize += allocation.NumOfRows
|
|
}
|
|
free := segment.GetMaxRowNum() - segment.GetNumOfRows() - allocSize
|
|
if free < count {
|
|
continue
|
|
}
|
|
allocation := getAllocation(count)
|
|
allocation.SegmentID = segment.GetID()
|
|
existedSegmentAllocations = append(existedSegmentAllocations, allocation)
|
|
return newSegmentAllocations, existedSegmentAllocations
|
|
}
|
|
|
|
// allocate new segment for remaining count
|
|
allocation := getAllocation(count)
|
|
newSegmentAllocations = append(newSegmentAllocations, allocation)
|
|
return newSegmentAllocations, existedSegmentAllocations
|
|
}
|
|
|
|
type SegmentSealPolicy interface {
|
|
ShouldSeal(segment *SegmentInfo, ts Timestamp) (bool, string)
|
|
}
|
|
|
|
// segmentSealPolicy seal policy applies to segment
|
|
type segmentSealPolicyFunc func(segment *SegmentInfo, ts Timestamp) (bool, string)
|
|
|
|
func (f segmentSealPolicyFunc) ShouldSeal(segment *SegmentInfo, ts Timestamp) (bool, string) {
|
|
return f(segment, ts)
|
|
}
|
|
|
|
// sealL1SegmentByCapacity get segmentSealPolicy with segment size factor policy
|
|
func sealL1SegmentByCapacity(sizeFactor float64) segmentSealPolicyFunc {
|
|
return func(segment *SegmentInfo, ts Timestamp) (bool, string) {
|
|
jitter := paramtable.Get().DataCoordCfg.SegmentSealProportionJitter.GetAsFloat()
|
|
ratio := (1 - jitter*rand.Float64())
|
|
return float64(segment.currRows) >= sizeFactor*float64(segment.GetMaxRowNum())*ratio,
|
|
fmt.Sprintf("Row count capacity full, current rows: %d, max row: %d, seal factor: %f, jitter ratio: %f", segment.currRows, segment.GetMaxRowNum(), sizeFactor, ratio)
|
|
}
|
|
}
|
|
|
|
// sealL1SegmentByLifetimePolicy get segmentSealPolicy with lifetime limit compares ts - segment.lastExpireTime
|
|
func sealL1SegmentByLifetime(lifetime time.Duration) segmentSealPolicyFunc {
|
|
return func(segment *SegmentInfo, ts Timestamp) (bool, string) {
|
|
pts, _ := tsoutil.ParseTS(ts)
|
|
epts, _ := tsoutil.ParseTS(segment.GetLastExpireTime())
|
|
d := pts.Sub(epts)
|
|
return d >= lifetime,
|
|
fmt.Sprintf("Segment Lifetime expired, segment last expire: %v, now:%v, max lifetime %v",
|
|
pts, epts, lifetime)
|
|
}
|
|
}
|
|
|
|
// sealL1SegmentByBinlogFileNumber seal L1 segment if binlog file number of segment exceed configured max number
|
|
func sealL1SegmentByBinlogFileNumber(maxBinlogFileNumber int) segmentSealPolicyFunc {
|
|
return func(segment *SegmentInfo, ts Timestamp) (bool, string) {
|
|
logFileCounter := 0
|
|
for _, fieldBinlog := range segment.GetStatslogs() {
|
|
logFileCounter += len(fieldBinlog.GetBinlogs())
|
|
}
|
|
|
|
return logFileCounter >= maxBinlogFileNumber,
|
|
fmt.Sprintf("Segment binlog number too large, binlog number: %d, max binlog number: %d", logFileCounter, maxBinlogFileNumber)
|
|
}
|
|
}
|
|
|
|
// sealLongTimeIdlePolicy seal segment if the segment has been written with a high frequency before.
|
|
// serve for this case:
|
|
// If users insert entities into segment continuously within a certain period of time, but they forgot to flush/(seal)
|
|
// it and the size of segment didn't reach the seal proportion. Under this situation, Milvus will wait these segments to
|
|
// be expired and during this period search latency may be a little high. We can assume that entities won't be inserted
|
|
// into this segment anymore, so sealLongTimeIdlePolicy will seal these segments to trigger handoff of query cluster.
|
|
// Q: Why we don't decrease the expiry time directly?
|
|
// A: We don't want to influence segments which are accepting `frequent small` batch entities.
|
|
func sealL1SegmentByIdleTime(idleTimeTolerance time.Duration, minSizeToSealIdleSegment float64, maxSizeOfSegment float64) segmentSealPolicyFunc {
|
|
return func(segment *SegmentInfo, ts Timestamp) (bool, string) {
|
|
limit := (minSizeToSealIdleSegment / maxSizeOfSegment) * float64(segment.GetMaxRowNum())
|
|
return time.Since(segment.lastWrittenTime) > idleTimeTolerance &&
|
|
float64(segment.currRows) > limit,
|
|
fmt.Sprintf("segment idle, segment row number :%d, last written time: %v, max idle duration: %v", segment.currRows, segment.lastWrittenTime, idleTimeTolerance)
|
|
}
|
|
}
|
|
|
|
// channelSealPolicy seal policy applies to channel
|
|
type channelSealPolicy func(string, []*SegmentInfo, Timestamp) ([]*SegmentInfo, string)
|
|
|
|
// getChannelOpenSegCapacityPolicy get channelSealPolicy with channel segment capacity policy
|
|
func getChannelOpenSegCapacityPolicy(limit int) channelSealPolicy {
|
|
return func(channel string, segs []*SegmentInfo, ts Timestamp) ([]*SegmentInfo, string) {
|
|
if len(segs) <= limit {
|
|
return []*SegmentInfo{}, ""
|
|
}
|
|
sortSegmentsByLastExpires(segs)
|
|
offLen := len(segs) - limit
|
|
if offLen > len(segs) {
|
|
offLen = len(segs)
|
|
}
|
|
return segs[0:offLen], fmt.Sprintf("seal by channel segment capacity, len(segs)=%d, limit=%d", len(segs), limit)
|
|
}
|
|
}
|
|
|
|
// sealByTotalGrowingSegmentsSize seals the largest growing segment
|
|
// if the total size of growing segments exceeds the threshold.
|
|
func sealByTotalGrowingSegmentsSize() channelSealPolicy {
|
|
return func(channel string, segments []*SegmentInfo, ts Timestamp) ([]*SegmentInfo, string) {
|
|
growingSegments := lo.Filter(segments, func(segment *SegmentInfo, _ int) bool {
|
|
return segment != nil && segment.GetState() == commonpb.SegmentState_Growing
|
|
})
|
|
|
|
var totalSize int64
|
|
sizeMap := lo.SliceToMap(growingSegments, func(segment *SegmentInfo) (int64, int64) {
|
|
size := segment.getSegmentSize()
|
|
totalSize += size
|
|
return segment.GetID(), size
|
|
})
|
|
|
|
threshold := paramtable.Get().DataCoordCfg.GrowingSegmentsMemSizeInMB.GetAsInt64() * 1024 * 1024
|
|
if totalSize >= threshold {
|
|
target := lo.MaxBy(growingSegments, func(s1, s2 *SegmentInfo) bool {
|
|
return sizeMap[s1.GetID()] > sizeMap[s2.GetID()]
|
|
})
|
|
return []*SegmentInfo{target}, fmt.Sprintf("seal by total growing segments size, "+
|
|
"totalSize=%d, threshold=%d", totalSize, threshold)
|
|
}
|
|
return nil, ""
|
|
}
|
|
}
|
|
|
|
// sortSegmentsByLastExpires sort segmentStatus with lastExpireTime ascending order
|
|
func sortSegmentsByLastExpires(segs []*SegmentInfo) {
|
|
sort.Slice(segs, func(i, j int) bool {
|
|
return segs[i].LastExpireTime < segs[j].LastExpireTime
|
|
})
|
|
}
|
|
|
|
type flushPolicy func(segment *SegmentInfo, t Timestamp) bool
|
|
|
|
func flushPolicyL1(segment *SegmentInfo, t Timestamp) bool {
|
|
return segment.GetState() == commonpb.SegmentState_Sealed &&
|
|
segment.Level != datapb.SegmentLevel_L0 &&
|
|
time.Since(segment.lastFlushTime) >= paramtable.Get().DataCoordCfg.SegmentFlushInterval.GetAsDuration(time.Second) &&
|
|
segment.GetLastExpireTime() <= t &&
|
|
segment.currRows != 0 &&
|
|
// Decoupling the importing segment from the flush process,
|
|
// This check avoids notifying the datanode to flush the
|
|
// importing segment which may not exist.
|
|
!segment.GetIsImporting()
|
|
}
|