enhance: use the key lock and concurrent map in the msg dispatcher client (#34278)

/kind improvement

Signed-off-by: SimFG <bang.fu@zilliz.com>
This commit is contained in:
SimFG 2024-07-05 15:11:31 +08:00 committed by GitHub
parent 7cac20f836
commit 0817802db8
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
2 changed files with 26 additions and 24 deletions

View File

@ -18,7 +18,6 @@ package msgdispatcher
import ( import (
"context" "context"
"sync"
"go.uber.org/zap" "go.uber.org/zap"
@ -27,6 +26,8 @@ import (
"github.com/milvus-io/milvus/pkg/mq/common" "github.com/milvus-io/milvus/pkg/mq/common"
"github.com/milvus-io/milvus/pkg/mq/msgstream" "github.com/milvus-io/milvus/pkg/mq/msgstream"
"github.com/milvus-io/milvus/pkg/util/funcutil" "github.com/milvus-io/milvus/pkg/util/funcutil"
"github.com/milvus-io/milvus/pkg/util/lock"
"github.com/milvus-io/milvus/pkg/util/typeutil"
) )
type ( type (
@ -46,17 +47,18 @@ var _ Client = (*client)(nil)
type client struct { type client struct {
role string role string
nodeID int64 nodeID int64
managers map[string]DispatcherManager managers *typeutil.ConcurrentMap[string, DispatcherManager]
managerMut sync.Mutex managerMut *lock.KeyLock[string]
factory msgstream.Factory factory msgstream.Factory
} }
func NewClient(factory msgstream.Factory, role string, nodeID int64) Client { func NewClient(factory msgstream.Factory, role string, nodeID int64) Client {
return &client{ return &client{
role: role, role: role,
nodeID: nodeID, nodeID: nodeID,
factory: factory, factory: factory,
managers: make(map[string]DispatcherManager), managers: typeutil.NewConcurrentMap[string, DispatcherManager](),
managerMut: lock.NewKeyLock[string](),
} }
} }
@ -64,20 +66,20 @@ func (c *client) Register(ctx context.Context, vchannel string, pos *Pos, subPos
log := log.With(zap.String("role", c.role), log := log.With(zap.String("role", c.role),
zap.Int64("nodeID", c.nodeID), zap.String("vchannel", vchannel)) zap.Int64("nodeID", c.nodeID), zap.String("vchannel", vchannel))
pchannel := funcutil.ToPhysicalChannel(vchannel) pchannel := funcutil.ToPhysicalChannel(vchannel)
c.managerMut.Lock() c.managerMut.Lock(pchannel)
defer c.managerMut.Unlock() defer c.managerMut.Unlock(pchannel)
var manager DispatcherManager var manager DispatcherManager
manager, ok := c.managers[pchannel] manager, ok := c.managers.Get(pchannel)
if !ok { if !ok {
manager = NewDispatcherManager(pchannel, c.role, c.nodeID, c.factory) manager = NewDispatcherManager(pchannel, c.role, c.nodeID, c.factory)
c.managers[pchannel] = manager c.managers.Insert(pchannel, manager)
go manager.Run() go manager.Run()
} }
ch, err := manager.Add(ctx, vchannel, pos, subPos) ch, err := manager.Add(ctx, vchannel, pos, subPos)
if err != nil { if err != nil {
if manager.Num() == 0 { if manager.Num() == 0 {
manager.Close() manager.Close()
delete(c.managers, pchannel) c.managers.Remove(pchannel)
} }
log.Error("register failed", zap.Error(err)) log.Error("register failed", zap.Error(err))
return nil, err return nil, err
@ -88,13 +90,13 @@ func (c *client) Register(ctx context.Context, vchannel string, pos *Pos, subPos
func (c *client) Deregister(vchannel string) { func (c *client) Deregister(vchannel string) {
pchannel := funcutil.ToPhysicalChannel(vchannel) pchannel := funcutil.ToPhysicalChannel(vchannel)
c.managerMut.Lock() c.managerMut.Lock(pchannel)
defer c.managerMut.Unlock() defer c.managerMut.Unlock(pchannel)
if manager, ok := c.managers[pchannel]; ok { if manager, ok := c.managers.Get(pchannel); ok {
manager.Remove(vchannel) manager.Remove(vchannel)
if manager.Num() == 0 { if manager.Num() == 0 {
manager.Close() manager.Close()
delete(c.managers, pchannel) c.managers.Remove(pchannel)
} }
log.Info("deregister done", zap.String("role", c.role), log.Info("deregister done", zap.String("role", c.role),
zap.Int64("nodeID", c.nodeID), zap.String("vchannel", vchannel)) zap.Int64("nodeID", c.nodeID), zap.String("vchannel", vchannel))
@ -104,12 +106,14 @@ func (c *client) Deregister(vchannel string) {
func (c *client) Close() { func (c *client) Close() {
log := log.With(zap.String("role", c.role), log := log.With(zap.String("role", c.role),
zap.Int64("nodeID", c.nodeID)) zap.Int64("nodeID", c.nodeID))
c.managerMut.Lock()
defer c.managerMut.Unlock() c.managers.Range(func(pchannel string, manager DispatcherManager) bool {
for pchannel, manager := range c.managers { c.managerMut.Lock(pchannel)
defer c.managerMut.Unlock(pchannel)
log.Info("close manager", zap.String("channel", pchannel)) log.Info("close manager", zap.String("channel", pchannel))
delete(c.managers, pchannel) c.managers.Remove(pchannel)
manager.Close() manager.Close()
} return true
})
log.Info("dispatcher client closed") log.Info("dispatcher client closed")
} }

View File

@ -79,8 +79,6 @@ func TestClient_Concurrency(t *testing.T) {
expected := int(total - deregisterCount.Load()) expected := int(total - deregisterCount.Load())
c := client1.(*client) c := client1.(*client)
c.managerMut.Lock() n := c.managers.Len()
n := len(c.managers)
c.managerMut.Unlock()
assert.Equal(t, expected, n) assert.Equal(t, expected, n)
} }