mirror of
https://gitee.com/milvus-io/milvus.git
synced 2024-11-29 18:38:44 +08:00
cae9e1c732
issue: #36858 - Start channel manager on datacoord, but with empty assign policy in streaming service. - Make collection at dropping state can be recovered by flusher to make sure that milvus consume the dropCollection message. - Add backoff for flusher lifetime. - remove the proxy watcher from timetick at rootcoord in streaming service. Also see the better fixup: #37176 --------- Signed-off-by: chyezh <chyezh@outlook.com>
420 lines
12 KiB
Go
420 lines
12 KiB
Go
// Licensed to the LF AI & Data foundation under one
|
|
// or more contributor license agreements. See the NOTICE file
|
|
// distributed with this work for additional information
|
|
// regarding copyright ownership. The ASF licenses this file
|
|
// to you under the Apache License, Version 2.0 (the
|
|
// "License"); you may not use this file except in compliance
|
|
// with the License. You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
package rootcoord
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"sync"
|
|
"time"
|
|
|
|
"go.uber.org/zap"
|
|
|
|
"github.com/milvus-io/milvus-proto/go-api/v2/commonpb"
|
|
"github.com/milvus-io/milvus-proto/go-api/v2/msgpb"
|
|
"github.com/milvus-io/milvus/internal/proto/internalpb"
|
|
"github.com/milvus-io/milvus/internal/util/sessionutil"
|
|
"github.com/milvus-io/milvus/internal/util/streamingutil"
|
|
"github.com/milvus-io/milvus/pkg/log"
|
|
"github.com/milvus-io/milvus/pkg/metrics"
|
|
"github.com/milvus-io/milvus/pkg/mq/msgstream"
|
|
"github.com/milvus-io/milvus/pkg/util/commonpbutil"
|
|
"github.com/milvus-io/milvus/pkg/util/timerecord"
|
|
"github.com/milvus-io/milvus/pkg/util/tsoutil"
|
|
"github.com/milvus-io/milvus/pkg/util/typeutil"
|
|
)
|
|
|
|
var (
|
|
// TODO: better to be configurable
|
|
enableTtChecker = true
|
|
timeTickSyncTtInterval = 2 * time.Minute
|
|
ttCheckerName = "rootTtChecker"
|
|
ttCheckerWarnMsg = fmt.Sprintf("RootCoord haven't synchronized the time tick for %f minutes", timeTickSyncTtInterval.Minutes())
|
|
ddlSourceID = UniqueID(-1)
|
|
)
|
|
|
|
type ttHistogram struct {
|
|
*typeutil.ConcurrentMap[string, Timestamp]
|
|
}
|
|
|
|
func newTtHistogram() *ttHistogram {
|
|
return &ttHistogram{
|
|
ConcurrentMap: typeutil.NewConcurrentMap[string, Timestamp](),
|
|
}
|
|
}
|
|
|
|
func (h *ttHistogram) update(channel string, ts Timestamp) {
|
|
h.Insert(channel, ts)
|
|
}
|
|
|
|
func (h *ttHistogram) get(channel string) Timestamp {
|
|
ts, ok := h.Get(channel)
|
|
if !ok {
|
|
return typeutil.ZeroTimestamp
|
|
}
|
|
return ts
|
|
}
|
|
|
|
func (h *ttHistogram) remove(channels ...string) {
|
|
for _, channel := range channels {
|
|
h.GetAndRemove(channel)
|
|
}
|
|
}
|
|
|
|
type timetickSync struct {
|
|
ctx context.Context
|
|
sourceID typeutil.UniqueID
|
|
|
|
dmlChannels *dmlChannels // used for insert
|
|
|
|
lock sync.Mutex
|
|
sess2ChanTsMap map[typeutil.UniqueID]*chanTsMsg
|
|
sendChan chan map[typeutil.UniqueID]*chanTsMsg
|
|
|
|
syncedTtHistogram *ttHistogram
|
|
}
|
|
|
|
type chanTsMsg struct {
|
|
chanTsMap map[string]typeutil.Timestamp
|
|
defaultTs typeutil.Timestamp
|
|
cnt int64
|
|
}
|
|
|
|
func newChanTsMsg(in *internalpb.ChannelTimeTickMsg, cnt int64) *chanTsMsg {
|
|
msg := &chanTsMsg{
|
|
chanTsMap: make(map[string]typeutil.Timestamp),
|
|
defaultTs: in.DefaultTimestamp,
|
|
cnt: cnt,
|
|
}
|
|
for idx := range in.ChannelNames {
|
|
msg.chanTsMap[in.ChannelNames[idx]] = in.Timestamps[idx]
|
|
}
|
|
return msg
|
|
}
|
|
|
|
func (c *chanTsMsg) getTimetick(channelName string) typeutil.Timestamp {
|
|
if ts, ok := c.chanTsMap[channelName]; ok {
|
|
return ts
|
|
}
|
|
return c.defaultTs
|
|
}
|
|
|
|
func newTimeTickSync(ctx context.Context, sourceID int64, factory msgstream.Factory, chanMap map[typeutil.UniqueID][]string) *timetickSync {
|
|
// if the old channels number used by the user is greater than the set default value currently
|
|
// keep the old channels
|
|
chanNum := getNeedChanNum(Params.RootCoordCfg.DmlChannelNum.GetAsInt(), chanMap)
|
|
|
|
// initialize dml channels used for insert
|
|
dmlChannels := newDmlChannels(ctx, factory, Params.CommonCfg.RootCoordDml.GetValue(), int64(chanNum))
|
|
|
|
// recover physical channels for all collections
|
|
for collID, chanNames := range chanMap {
|
|
dmlChannels.addChannels(chanNames...)
|
|
log.Info("recover physical channels", zap.Int64("collectionID", collID), zap.Strings("physical channels", chanNames))
|
|
}
|
|
|
|
return &timetickSync{
|
|
ctx: ctx,
|
|
sourceID: sourceID,
|
|
|
|
dmlChannels: dmlChannels,
|
|
|
|
lock: sync.Mutex{},
|
|
sess2ChanTsMap: make(map[typeutil.UniqueID]*chanTsMsg),
|
|
|
|
// 1 is the most reasonable capacity. In fact, Milvus can only focus on the latest time tick.
|
|
sendChan: make(chan map[typeutil.UniqueID]*chanTsMsg, 1),
|
|
|
|
syncedTtHistogram: newTtHistogram(),
|
|
}
|
|
}
|
|
|
|
// sendToChannel send all channels' timetick to sendChan
|
|
// lock is needed by the invoker
|
|
func (t *timetickSync) sendToChannel() bool {
|
|
if len(t.sess2ChanTsMap) == 0 {
|
|
return false
|
|
}
|
|
|
|
// detect whether rootcoord receives ttMsg from all source sessions
|
|
maxCnt := int64(0)
|
|
idleSessionList := make([]typeutil.UniqueID, 0, len(t.sess2ChanTsMap))
|
|
for id, v := range t.sess2ChanTsMap {
|
|
if v == nil {
|
|
idleSessionList = append(idleSessionList, id)
|
|
} else {
|
|
if maxCnt < v.cnt {
|
|
maxCnt = v.cnt
|
|
}
|
|
}
|
|
}
|
|
|
|
if len(idleSessionList) > 0 {
|
|
// give warning every 2 second if not get ttMsg from source sessions
|
|
if maxCnt%10 == 0 {
|
|
log.Warn("session idle for long time", zap.Any("idle list", idleSessionList),
|
|
zap.Int64("idle time", Params.ProxyCfg.TimeTickInterval.GetAsInt64()*time.Millisecond.Milliseconds()*maxCnt))
|
|
}
|
|
return false
|
|
}
|
|
|
|
// clear sess2ChanTsMap and send a clone
|
|
ptt := make(map[typeutil.UniqueID]*chanTsMsg)
|
|
for k, v := range t.sess2ChanTsMap {
|
|
ptt[k] = v
|
|
t.sess2ChanTsMap[k] = nil
|
|
}
|
|
|
|
select {
|
|
case t.sendChan <- ptt:
|
|
default:
|
|
// The consumer of `sendChan` haven't completed its operation. If we send the `ptt` here, the consumer will
|
|
// always get an older time tick. The older time tick in `sendChan` will block newer time tick in next window.
|
|
// However, in fact the consumer can only focus on the newest.
|
|
|
|
// TODO: maybe a metric should be here.
|
|
}
|
|
|
|
return true
|
|
}
|
|
|
|
// UpdateTimeTick check msg validation and send it to local channel
|
|
func (t *timetickSync) updateTimeTick(in *internalpb.ChannelTimeTickMsg, reason string) error {
|
|
t.lock.Lock()
|
|
defer t.lock.Unlock()
|
|
if len(in.ChannelNames) == 0 && in.DefaultTimestamp == 0 {
|
|
return nil
|
|
}
|
|
if len(in.Timestamps) != len(in.ChannelNames) {
|
|
return fmt.Errorf("invalid TimeTickMsg, timestamp and channelname size mismatch")
|
|
}
|
|
|
|
prev, ok := t.sess2ChanTsMap[in.Base.SourceID]
|
|
if !ok {
|
|
return fmt.Errorf("skip ChannelTimeTickMsg from un-recognized session %d", in.Base.SourceID)
|
|
}
|
|
|
|
if in.Base.SourceID == t.sourceID {
|
|
if prev != nil && in.DefaultTimestamp < prev.defaultTs {
|
|
log.Warn("timestamp go back", zap.Int64("source id", in.Base.SourceID),
|
|
zap.Uint64("curr ts", in.DefaultTimestamp),
|
|
zap.Uint64("prev ts", prev.defaultTs),
|
|
zap.String("reason", reason))
|
|
return nil
|
|
}
|
|
}
|
|
|
|
if prev == nil {
|
|
t.sess2ChanTsMap[in.Base.SourceID] = newChanTsMsg(in, 1)
|
|
} else {
|
|
t.sess2ChanTsMap[in.Base.SourceID] = newChanTsMsg(in, prev.cnt+1)
|
|
}
|
|
t.sendToChannel()
|
|
return nil
|
|
}
|
|
|
|
func (t *timetickSync) addSession(sess *sessionutil.Session) {
|
|
t.lock.Lock()
|
|
defer t.lock.Unlock()
|
|
t.sess2ChanTsMap[sess.ServerID] = nil
|
|
log.Info("Add session for timeticksync", zap.Int64("serverID", sess.ServerID))
|
|
}
|
|
|
|
func (t *timetickSync) delSession(sess *sessionutil.Session) {
|
|
t.lock.Lock()
|
|
defer t.lock.Unlock()
|
|
if _, ok := t.sess2ChanTsMap[sess.ServerID]; ok {
|
|
delete(t.sess2ChanTsMap, sess.ServerID)
|
|
log.Info("Remove session from timeticksync", zap.Int64("serverID", sess.ServerID))
|
|
t.sendToChannel()
|
|
}
|
|
}
|
|
|
|
func (t *timetickSync) initSessions(sess []*sessionutil.Session) {
|
|
t.lock.Lock()
|
|
defer t.lock.Unlock()
|
|
t.sess2ChanTsMap = make(map[typeutil.UniqueID]*chanTsMsg)
|
|
// Init DDL source
|
|
t.sess2ChanTsMap[ddlSourceID] = nil
|
|
for _, s := range sess {
|
|
t.sess2ChanTsMap[s.ServerID] = nil
|
|
log.Info("Init proxy sessions for timeticksync", zap.Int64("serverID", s.ServerID))
|
|
}
|
|
}
|
|
|
|
// StartWatch watches on session changes and processes timeTick messages of all channels.
|
|
func (t *timetickSync) startWatch(wg *sync.WaitGroup) {
|
|
defer wg.Done()
|
|
|
|
var checker *timerecord.LongTermChecker
|
|
if enableTtChecker {
|
|
checker = timerecord.NewLongTermChecker(t.ctx, ttCheckerName, timeTickSyncTtInterval, ttCheckerWarnMsg)
|
|
checker.Start()
|
|
defer checker.Stop()
|
|
}
|
|
|
|
for {
|
|
select {
|
|
case <-t.ctx.Done():
|
|
log.Info("rootcoord context done", zap.Error(t.ctx.Err()))
|
|
return
|
|
case sessTimetick, ok := <-t.sendChan:
|
|
if !ok {
|
|
log.Info("timetickSync sendChan closed")
|
|
return
|
|
}
|
|
if enableTtChecker {
|
|
checker.Check()
|
|
}
|
|
// reduce each channel to get min timestamp
|
|
local := sessTimetick[ddlSourceID]
|
|
if len(local.chanTsMap) == 0 {
|
|
continue
|
|
}
|
|
hdr := fmt.Sprintf("send ts to %d channels", len(local.chanTsMap))
|
|
tr := timerecord.NewTimeRecorder(hdr)
|
|
wg := sync.WaitGroup{}
|
|
for chanName, ts := range local.chanTsMap {
|
|
wg.Add(1)
|
|
go func(chanName string, ts typeutil.Timestamp) {
|
|
mints := ts
|
|
for _, tt := range sessTimetick {
|
|
currTs := tt.getTimetick(chanName)
|
|
if currTs < mints {
|
|
mints = currTs
|
|
}
|
|
}
|
|
if err := t.sendTimeTickToChannel([]string{chanName}, mints); err != nil {
|
|
log.Warn("SendTimeTickToChannel fail", zap.Error(err))
|
|
} else {
|
|
t.syncedTtHistogram.update(chanName, mints)
|
|
}
|
|
wg.Done()
|
|
}(chanName, ts)
|
|
}
|
|
wg.Wait()
|
|
span := tr.ElapseSpan()
|
|
metrics.RootCoordSyncTimeTickLatency.Observe(float64(span.Milliseconds()))
|
|
// rootcoord send tt msg to all channels every 200ms by default
|
|
if span > Params.ProxyCfg.TimeTickInterval.GetAsDuration(time.Millisecond) {
|
|
log.Warn("rootcoord send tt to all channels too slowly",
|
|
zap.Int("chanNum", len(local.chanTsMap)), zap.Int64("span", span.Milliseconds()))
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// SendTimeTickToChannel send each channel's min timetick to msg stream
|
|
func (t *timetickSync) sendTimeTickToChannel(chanNames []string, ts typeutil.Timestamp) error {
|
|
if streamingutil.IsStreamingServiceEnabled() {
|
|
return nil
|
|
}
|
|
func() {
|
|
sub := tsoutil.SubByNow(ts)
|
|
for _, chanName := range chanNames {
|
|
metrics.RootCoordInsertChannelTimeTick.WithLabelValues(chanName).Set(float64(sub))
|
|
}
|
|
}()
|
|
|
|
msgPack := msgstream.MsgPack{}
|
|
|
|
timeTickMsg := &msgstream.TimeTickMsg{
|
|
BaseMsg: msgstream.BaseMsg{
|
|
BeginTimestamp: ts,
|
|
EndTimestamp: ts,
|
|
HashValues: []uint32{0},
|
|
},
|
|
TimeTickMsg: &msgpb.TimeTickMsg{
|
|
Base: commonpbutil.NewMsgBase(
|
|
commonpbutil.WithMsgType(commonpb.MsgType_TimeTick),
|
|
commonpbutil.WithTimeStamp(ts),
|
|
commonpbutil.WithSourceID(t.sourceID),
|
|
),
|
|
},
|
|
}
|
|
msgPack.Msgs = append(msgPack.Msgs, timeTickMsg)
|
|
if err := t.dmlChannels.broadcast(chanNames, &msgPack); err != nil {
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// GetSessionNum return the num of detected sessions
|
|
func (t *timetickSync) getSessionNum() int {
|
|
t.lock.Lock()
|
|
defer t.lock.Unlock()
|
|
return len(t.sess2ChanTsMap)
|
|
}
|
|
|
|
// /////////////////////////////////////////////////////////////////////////////
|
|
// getDmlChannelNames returns list of channel names.
|
|
func (t *timetickSync) getDmlChannelNames(count int) []string {
|
|
return t.dmlChannels.getChannelNames(count)
|
|
}
|
|
|
|
// GetDmlChannelNum return the num of dml channels
|
|
func (t *timetickSync) getDmlChannelNum() int {
|
|
return t.dmlChannels.getChannelNum()
|
|
}
|
|
|
|
// ListDmlChannels return all in-use dml channel names
|
|
func (t *timetickSync) listDmlChannels() []string {
|
|
return t.dmlChannels.listChannels()
|
|
}
|
|
|
|
// AddDmlChannels add dml channels
|
|
func (t *timetickSync) addDmlChannels(names ...string) {
|
|
t.dmlChannels.addChannels(names...)
|
|
log.Info("add dml channels", zap.Strings("channels", names))
|
|
}
|
|
|
|
// RemoveDmlChannels remove dml channels
|
|
func (t *timetickSync) removeDmlChannels(names ...string) {
|
|
t.dmlChannels.removeChannels(names...)
|
|
// t.syncedTtHistogram.remove(names...) // channel ts shouldn't go back.
|
|
log.Info("remove dml channels", zap.Strings("channels", names))
|
|
}
|
|
|
|
// BroadcastDmlChannels broadcasts msg pack into dml channels
|
|
func (t *timetickSync) broadcastDmlChannels(chanNames []string, pack *msgstream.MsgPack) error {
|
|
return t.dmlChannels.broadcast(chanNames, pack)
|
|
}
|
|
|
|
// BroadcastMarkDmlChannels broadcasts msg pack into dml channels
|
|
func (t *timetickSync) broadcastMarkDmlChannels(chanNames []string, pack *msgstream.MsgPack) (map[string][]byte, error) {
|
|
return t.dmlChannels.broadcastMark(chanNames, pack)
|
|
}
|
|
|
|
func (t *timetickSync) getSyncedTimeTick(channel string) Timestamp {
|
|
return t.syncedTtHistogram.get(channel)
|
|
}
|
|
|
|
func minTimeTick(tt ...typeutil.Timestamp) typeutil.Timestamp {
|
|
var ret typeutil.Timestamp
|
|
for _, t := range tt {
|
|
if ret == 0 {
|
|
ret = t
|
|
} else {
|
|
if t < ret {
|
|
ret = t
|
|
}
|
|
}
|
|
}
|
|
return ret
|
|
}
|