1f14ae5b6b
Signed-off-by: jingkl <jingjing.jia@zilliz.com> |
||
---|---|---|
.. | ||
asserts | ||
ci | ||
milvus_benchmark | ||
.gitignore | ||
Dockerfile | ||
README.md | ||
requirements.txt |
The milvus_benchmark is a non-functional testing tool or service which allows users to run tests on k8s cluster or at local, the primary use case is performance/load/stability testing, the objective is to expose problems in milvus project.
Quick start
Description
- Test cases in
milvus_benchmark
can be organized withyaml
- Test can run with local mode or helm mode
- local: install and start your local server, and pass the host/port param when start the tests
- helm: install the server by helm, which will manage the milvus in k8s cluster, and you can interagte the test stage into argo workflow or jenkins pipeline
Usage
-
Using jenkins: Use
ci/main_jenkinsfile
as the jenkins pipeline file -
Using argo: example argo workflow yaml configuration:
ci/argo.yaml
-
Local test:
-
set PYTHONPATH:
export PYTHONPATH=/your/project/path/milvus_benchmark
-
prepare data:
if we need to use the sift/deep dataset as the raw data input, we need to mount NAS and update
RAW_DATA_DIR
inconfig.py
, the example mount command:sudo mount -t cifs -o username=test,vers=1.0 //172.16.70.249/test /test
-
install requirements:
pip install -r requirements.txt
-
write test yaml and run with the yaml param:
cd milvus-benchmark/ && python main.py --local --host=* --port=19530 --suite=suites/2_insert_data.yaml
-
Test suite
Description
Test suite yaml defines the test process, users need to add test suite yaml if adding a customized test into the current test framework.
Example
Take the test file 2_insert_data.yaml
as an example
insert_performance:
collections:
-
milvus:
db_config.primary_path: /test/milvus/db_data_2/cluster/sift_1m_128_l2
wal_enable: true
collection_name: sift_1m_128_l2
ni_per: 50000
build_index: false
index_type: ivf_sq8
index_param:
nlist: 1024
-
insert_performance
The top level is the runner type: the other test types including:
search_performance/build_performance/insert_performance/accuracy/locust_insert/...
, each test type corresponds to the different runner conponent defined in directoryrunnners
-
other fields under runner type
The other parts in the test yaml is the params pass to the runner, such as:
- The field
collection_name
means which kind of collection will be created in milvus - The field
ni_per
means the batch size - The filed
build_index
means that whether to create index during inserting
- The field
While using argo workflow as benchmark pipeline, the test suite is made of both client
and server
configmap, an example will be like this:
server
kind: ConfigMap
apiVersion: v1
metadata:
name: server-cluster-8c16m
namespace: qa
uid: 3752f85c-c840-40c6-a5db-ae44146ad8b5
resourceVersion: '42213135'
creationTimestamp: '2021-05-14T07:00:53Z'
managedFields:
- manager: dashboard
operation: Update
apiVersion: v1
time: '2021-05-14T07:00:53Z'
fieldsType: FieldsV1
fieldsV1:
'f:data':
.: {}
'f:config.yaml': {}
data:
config.yaml: |
server:
server_tag: "8c16m"
milvus:
deploy_mode: "cluster"
client
kind: ConfigMap
apiVersion: v1
metadata:
name: client-insert-batch-1000
namespace: qa
uid: 8604c277-f00f-47c7-8fcb-9b3bc97efa74
resourceVersion: '42988547'
creationTimestamp: '2021-07-09T08:33:02Z'
managedFields:
- manager: dashboard
operation: Update
apiVersion: v1
fieldsType: FieldsV1
fieldsV1:
'f:data':
.: {}
'f:config.yaml': {}
data:
config.yaml: |
insert_performance:
collections:
-
milvus:
wal_enable: true
collection_name: sift_1m_128_l2
ni_per: 1000
build_index: false
index_type: ivf_sq8
index_param:
nlist: 1024
Overview of the benchmark
Components
-
main.py
The entry file: parse the input params and initialize the other components:
metric
,env
,runner
-
metric
The test result can be used to analyze the regression or improvement of the milvus system, so we upload the metrics of the test result when a test suite run finished, and then use
redash
to make sense of our data -
db
Currently we use the
mongodb
to store the test result -
env
The
env
component defines the server environment and environment management, the instance of theenv
corresponds to the run mode of the benchmark-
local
: Only defines the host and port for testing -
helm/docker
: Install and uninstall the server in benchmark stage
-
-
runner
The actual executor in benchmark, each test type defined in test suite will generate the corresponding runner instance, there are three stages in
runner
:-
extract_cases
: There are several test cases defined in each test suite yaml, and each case shares the same server environment and shares the sameprepare
stage, but themetric
for each case is different, so we need to extract cases from the test suite before the cases runs -
prepare
: Prepare the data and operations, for example, before running searching, index needs to be created and data needs to be loaded -
run_case
: Do the core operation and setmetric
value
-
-
suites
: There are two ways to take the content to be tested as input parameters:- Test suite files under
suites
directory - Test suite configmap name including
server_config_map
andclient_config_map
if using argo workflow
- Test suite files under
-
update.py
: While using argo workflow as benchmark pipeline, we have two steps in workflow template:install-milvus
andclient-test
- In stage
install-milvus
,update.py
is used to generate a newvalues.yaml
which will be a param while inhelm install
operation - In stage
client-test
, it runsmain.py
and receives the milvus host and port as the cmd params, with the run modelocal
- In stage
Conceptual overview
The following diagram shows the runtime execution graph of the benchmark (local mode based on argo workflow)