Fix pkg codecov & static-check script (#26720)

Signed-off-by: Congqi Xia <congqi.xia@zilliz.com>
This commit is contained in:
congqixia 2023-08-31 09:59:01 +08:00 committed by GitHub
parent a8e5dc3517
commit d343888f3c
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
9 changed files with 99 additions and 55 deletions

View File

@ -72,6 +72,7 @@ else
@GO111MODULE=on env bash $(PWD)/scripts/gofmt.sh internal/
@GO111MODULE=on env bash $(PWD)/scripts/gofmt.sh tests/integration/
@GO111MODULE=on env bash $(PWD)/scripts/gofmt.sh tests/go/
@GO111MODULE=on env bash $(PWD)/scripts/gofmt.sh pkg/
endif
lint: tools/bin/revive
@ -85,7 +86,7 @@ static-check: getdeps
@source $(PWD)/scripts/setenv.sh && GO111MODULE=on $(INSTALL_PATH)/golangci-lint run --timeout=30m --config ./.golangci.yml ./internal/...
@source $(PWD)/scripts/setenv.sh && GO111MODULE=on $(INSTALL_PATH)/golangci-lint run --timeout=30m --config ./.golangci.yml ./cmd/...
@source $(PWD)/scripts/setenv.sh && GO111MODULE=on $(INSTALL_PATH)/golangci-lint run --timeout=30m --config ./.golangci.yml ./tests/integration/...
@source $(PWD)/scripts/setenv.sh && GO111MODULE=on $(INSTALL_PATH)/golangci-lint run --timeout=30m --config ./.golangci.yml ./...
@source $(PWD)/scripts/setenv.sh && cd pkg && GO111MODULE=on $(INSTALL_PATH)/golangci-lint run --timeout=30m --config ../.golangci.yml ./...
verifiers: build-cpp getdeps cppcheck fmt static-check

View File

@ -18,19 +18,24 @@ package metrics
import (
"testing"
"github.com/prometheus/client_golang/prometheus"
"github.com/stretchr/testify/assert"
)
func TestRegisterMetrics(t *testing.T) {
r := NewMilvusRegistry()
// Make sure it doesn't panic.
RegisterRootCoord(r.GoRegistry)
RegisterDataNode(r.GoRegistry)
RegisterDataCoord(r.GoRegistry)
RegisterIndexNode(r.GoRegistry)
RegisterProxy(r.GoRegistry)
RegisterQueryNode(r.GoRegistry)
RegisterQueryCoord(r.GoRegistry)
RegisterMetaMetrics(r.GoRegistry)
RegisterStorageMetrics(r.GoRegistry)
RegisterMsgStreamMetrics(r.GoRegistry)
assert.NotPanics(t, func() {
r := prometheus.NewRegistry()
// Make sure it doesn't panic.
RegisterRootCoord(r)
RegisterDataNode(r)
RegisterDataCoord(r)
RegisterIndexNode(r)
RegisterProxy(r)
RegisterQueryNode(r)
RegisterQueryCoord(r)
RegisterMetaMetrics(r)
RegisterStorageMetrics(r)
RegisterMsgStreamMetrics(r)
})
}

View File

@ -307,25 +307,66 @@ func TestKafkaClient_MsgSerializAndDeserialize(t *testing.T) {
assert.Nil(t, msgID)
}
/*
func createParamItem(v string) paramtable.ParamItem {
item := paramtable.ParamItem{
Formatter: func(originValue string) string { return v },
}
item.Init(&config.Manager{})
return item
}*/
func initParamItem(item *paramtable.ParamItem, v string) {
item.Formatter = func(originValue string) string { return v }
item.Init(&config.Manager{})
}
type kafkaCfgOption func(cfg *paramtable.KafkaConfig)
func withAddr(v string) kafkaCfgOption {
return func(cfg *paramtable.KafkaConfig) {
initParamItem(&cfg.Address, v)
}
}
func withUsername(v string) kafkaCfgOption {
return func(cfg *paramtable.KafkaConfig) {
initParamItem(&cfg.SaslUsername, v)
}
}
func withPasswd(v string) kafkaCfgOption {
return func(cfg *paramtable.KafkaConfig) {
initParamItem(&cfg.SaslPassword, v)
}
}
func withMechanism(v string) kafkaCfgOption {
return func(cfg *paramtable.KafkaConfig) {
initParamItem(&cfg.SaslMechanisms, v)
}
}
func withProtocol(v string) kafkaCfgOption {
return func(cfg *paramtable.KafkaConfig) {
initParamItem(&cfg.SecurityProtocol, v)
}
}
func createKafkaConfig(opts ...kafkaCfgOption) *paramtable.KafkaConfig {
cfg := &paramtable.KafkaConfig{}
for _, opt := range opts {
opt(cfg)
}
return cfg
}
func TestKafkaClient_NewKafkaClientInstanceWithConfig(t *testing.T) {
config1 := &paramtable.KafkaConfig{
Address: createParamItem("addr"),
SaslPassword: createParamItem("password"),
}
config1 := createKafkaConfig(withAddr("addr"), withPasswd("password"))
assert.Panics(t, func() { NewKafkaClientInstanceWithConfig(config1) })
config2 := &paramtable.KafkaConfig{
Address: createParamItem("addr"),
SaslUsername: createParamItem("username"),
}
config2 := createKafkaConfig(withAddr("addr"), withUsername("username"))
assert.Panics(t, func() { NewKafkaClientInstanceWithConfig(config2) })
producerConfig := make(map[string]string)
@ -333,15 +374,10 @@ func TestKafkaClient_NewKafkaClientInstanceWithConfig(t *testing.T) {
consumerConfig := make(map[string]string)
consumerConfig["client.id"] = "dc"
config := &paramtable.KafkaConfig{
Address: createParamItem("addr"),
SaslUsername: createParamItem("username"),
SaslPassword: createParamItem("password"),
SaslMechanisms: createParamItem("sasl"),
SecurityProtocol: createParamItem("plain"),
ConsumerExtraConfig: paramtable.ParamGroup{GetFunc: func() map[string]string { return consumerConfig }},
ProducerExtraConfig: paramtable.ParamGroup{GetFunc: func() map[string]string { return producerConfig }},
}
config := createKafkaConfig(withAddr("addr"), withUsername("username"), withPasswd("password"), withMechanism("sasl"), withProtocol("plain"))
config.ConsumerExtraConfig = paramtable.ParamGroup{GetFunc: func() map[string]string { return consumerConfig }}
config.ProducerExtraConfig = paramtable.ParamGroup{GetFunc: func() map[string]string { return producerConfig }}
client := NewKafkaClientInstanceWithConfig(config)
assert.NotNil(t, client)
assert.NotNil(t, client.basicConfig)

View File

@ -25,7 +25,7 @@ import (
// Meta Prefix consts
const (
MetaStoreTypeEtcd = "etcd"
MetaStoreTypeEtcd = "etcd"
SegmentMetaPrefix = "queryCoord-segmentMeta"
ChangeInfoMetaPrefix = "queryCoord-sealedSegmentChangeInfo"

View File

@ -234,30 +234,30 @@ func TestComponentParam(t *testing.T) {
// })
t.Run("test queryCoordConfig", func(t *testing.T) {
Params := params.QueryCoordCfg
Params := &params.QueryCoordCfg
assert.Equal(t, Params.EnableActiveStandby.GetAsBool(), false)
t.Logf("queryCoord EnableActiveStandby = %t", Params.EnableActiveStandby.GetAsBool())
params.Save("queryCoord.NextTargetSurviveTime", "100")
NextTargetSurviveTime := Params.NextTargetSurviveTime
NextTargetSurviveTime := &Params.NextTargetSurviveTime
assert.Equal(t, int64(100), NextTargetSurviveTime.GetAsInt64())
params.Save("queryCoord.UpdateNextTargetInterval", "100")
UpdateNextTargetInterval := Params.UpdateNextTargetInterval
UpdateNextTargetInterval := &Params.UpdateNextTargetInterval
assert.Equal(t, int64(100), UpdateNextTargetInterval.GetAsInt64())
params.Save("queryCoord.checkNodeInReplicaInterval", "100")
checkNodeInReplicaInterval := Params.CheckNodeInReplicaInterval
checkNodeInReplicaInterval := &Params.CheckNodeInReplicaInterval
assert.Equal(t, 100, checkNodeInReplicaInterval.GetAsInt())
params.Save("queryCoord.checkResourceGroupInterval", "10")
checkResourceGroupInterval := Params.CheckResourceGroupInterval
checkResourceGroupInterval := &Params.CheckResourceGroupInterval
assert.Equal(t, 10, checkResourceGroupInterval.GetAsInt())
enableResourceGroupAutoRecover := Params.EnableRGAutoRecover
enableResourceGroupAutoRecover := &Params.EnableRGAutoRecover
assert.Equal(t, true, enableResourceGroupAutoRecover.GetAsBool())
params.Save("queryCoord.enableRGAutoRecover", "false")
enableResourceGroupAutoRecover = Params.EnableRGAutoRecover
enableResourceGroupAutoRecover = &Params.EnableRGAutoRecover
assert.Equal(t, false, enableResourceGroupAutoRecover.GetAsBool())
checkHealthInterval := Params.CheckHealthInterval.GetAsInt()
@ -284,7 +284,7 @@ func TestComponentParam(t *testing.T) {
})
t.Run("test queryNodeConfig", func(t *testing.T) {
Params := params.QueryNodeCfg
Params := &params.QueryNodeCfg
interval := Params.StatsPublishInterval.GetAsInt()
assert.Equal(t, 1000, interval)
@ -338,14 +338,14 @@ func TestComponentParam(t *testing.T) {
assert.Equal(t, int64(1024), chunkRows)
params.Save("queryNode.gracefulStopTimeout", "100")
gracefulStopTimeout := Params.GracefulStopTimeout
gracefulStopTimeout := &Params.GracefulStopTimeout
assert.Equal(t, int64(100), gracefulStopTimeout.GetAsInt64())
assert.Equal(t, false, Params.EnableWorkerSQCostMetrics.GetAsBool())
})
t.Run("test dataCoordConfig", func(t *testing.T) {
Params := params.DataCoordCfg
Params := &params.DataCoordCfg
assert.Equal(t, 24*60*60*time.Second, Params.SegmentMaxLifetime.GetAsDuration(time.Second))
assert.True(t, Params.EnableGarbageCollection.GetAsBool())
assert.Equal(t, Params.EnableActiveStandby.GetAsBool(), false)
@ -353,7 +353,7 @@ func TestComponentParam(t *testing.T) {
})
t.Run("test dataNodeConfig", func(t *testing.T) {
Params := params.DataNodeCfg
Params := &params.DataNodeCfg
SetNodeID(2)
@ -372,23 +372,23 @@ func TestComponentParam(t *testing.T) {
size := Params.FlushInsertBufferSize.GetAsInt()
t.Logf("FlushInsertBufferSize: %d", size)
period := Params.SyncPeriod
period := &Params.SyncPeriod
t.Logf("SyncPeriod: %v", period)
assert.Equal(t, 10*time.Minute, Params.SyncPeriod.GetAsDuration(time.Second))
bulkinsertTimeout := Params.BulkInsertTimeoutSeconds
bulkinsertTimeout := &Params.BulkInsertTimeoutSeconds
t.Logf("BulkInsertTimeoutSeconds: %v", bulkinsertTimeout)
assert.Equal(t, "18000", Params.BulkInsertTimeoutSeconds.GetValue())
})
t.Run("test indexNodeConfig", func(t *testing.T) {
Params := params.IndexNodeCfg
Params := &params.IndexNodeCfg
params.Save(Params.GracefulStopTimeout.Key, "50")
assert.Equal(t, Params.GracefulStopTimeout.GetAsInt64(), int64(50))
})
t.Run("channel config priority", func(t *testing.T) {
Params := params.CommonCfg
Params := &params.CommonCfg
params.Save(Params.RootCoordDml.Key, "dml1")
params.Save(Params.RootCoordDml.FallbackKeys[0], "dml2")

View File

@ -50,7 +50,7 @@ func TestGrpcServerParams(t *testing.T) {
base.Save("grpc.serverMaxRecvSize", "a")
assert.Equal(t, serverConfig.ServerMaxSendSize.GetAsInt(), DefaultServerMaxRecvSize)
assert.NotZero(t, serverConfig.ServerMaxSendSize)
assert.NotZero(t, serverConfig.ServerMaxSendSize.GetAsInt())
t.Logf("ServerMaxSendSize = %d", serverConfig.ServerMaxSendSize.GetAsInt())
base.Remove(role + ".grpc.serverMaxSendSize")

View File

@ -9,8 +9,8 @@ import (
func TestHTTPConfig_Init(t *testing.T) {
params := ComponentParam{}
params.Init()
cf := params.HTTPCfg
assert.Equal(t, cf.Enabled.GetAsBool(), true)
assert.Equal(t, cf.DebugMode.GetAsBool(), false)
assert.Equal(t, cf.Port.GetValue(), "")
cfg := &params.HTTPCfg
assert.Equal(t, cfg.Enabled.GetAsBool(), true)
assert.Equal(t, cfg.DebugMode.GetAsBool(), false)
assert.Equal(t, cfg.Port.GetValue(), "")
}

View File

@ -285,8 +285,8 @@ func (p *MetaStoreConfig) Init(base *BaseTable) {
Key: "metastore.type",
Version: "2.2.0",
DefaultValue: util.MetaStoreTypeEtcd,
Doc: `Default value: etcd, Valid values: etcd `,
Export: true,
Doc: `Default value: etcd, Valid values: etcd `,
Export: true,
}
p.MetaStoreType.Init(base.mgr)
}

View File

@ -37,13 +37,15 @@ for d in $(go list ./internal/... | grep -v -e vendor -e kafka -e planparserv2/g
rm profile.out
fi
done
for d in $(go list ./pkg/... | grep -v -e vendor -e kafka -e planparserv2/generated -e mocks); do
pushd pkg
for d in $(go list ./... | grep -v -e vendor -e kafka -e planparserv2/generated -e mocks); do
go test -race -tags dynamic -v -coverpkg=./... -coverprofile=profile.out -covermode=atomic "$d"
if [ -f profile.out ]; then
grep -v kafka profile.out | grep -v planparserv2/generated | grep -v mocks | sed '1d' >> ${FILE_COVERAGE_INFO}
grep -v kafka profile.out | grep -v planparserv2/generated | grep -v mocks | sed '1d' >> ../${FILE_COVERAGE_INFO}
rm profile.out
fi
done
popd
endTime=`date +%s`
echo "Total time for go unittest:" $(($endTime-$beginTime)) "s"