milvus/tests/python_client
2024-11-20 14:22:30 +08:00
..
assets/ann_hdf5 [skip e2e]Add recall test (#18562) 2022-08-09 13:50:36 +08:00
base test: Add full-text search test cases (#36998) 2024-10-23 09:51:27 +08:00
bulk_insert test: fix bulk insert case (#31738) 2024-04-01 10:49:18 +08:00
chaos test: fix tokenizer and monkey patch faker function (#37119) 2024-11-05 08:42:23 +08:00
check test: Add expression template tests and use error msg only as assertion (#37618) 2024-11-13 17:14:30 +08:00
common test: Add expression template tests and use error msg only as assertion (#37618) 2024-11-13 17:14:30 +08:00
config [test]Fix occasional FileExistsError triggered when concurrently executing pytest (#26867) 2023-09-06 12:07:14 +08:00
customize test: add channel exclusive balance test and resource group test (#33093) 2024-05-31 13:55:52 +08:00
deploy enhance: use docker compose instead of docker-compose (#35208) 2024-08-02 19:32:32 +08:00
graphs [skip ci]Update flow chart in chaos test readme (#13537) 2021-12-16 20:27:16 +08:00
load test: Add a new range search test for all indexes and align some index params (#32724) 2024-04-30 17:37:27 +08:00
loadbalance test: add more request type checker for test (#29210) 2023-12-14 19:38:45 +08:00
milvus_client test: update test cases (#37660) 2024-11-14 17:46:39 +08:00
rate_limit Update rate limit test file (#18942) 2022-09-02 10:10:59 +08:00
resource_group test: add channel exclusive balance test and resource group test (#33093) 2024-05-31 13:55:52 +08:00
scale [test] Add load balance verify in querynode scale test (#19028) (#19054) 2022-09-07 19:24:35 +08:00
standby Update milvs helm repo for ci (#28042) 2023-11-01 18:54:16 +08:00
testcases enhance: disallow get raw vector data of a BM25 Function output field (#37800) 2024-11-20 14:22:30 +08:00
utils enhance: add an unify vector index config checker (#36844) 2024-10-28 10:11:37 +08:00
.dockerignore
.gitignore [test]Update bulk insert test case and skip calc_distance testcases (#19855) 2022-10-18 17:21:26 +08:00
conftest.py enhance: update knowhere version (#37688) 2024-11-15 10:10:31 +08:00
Dockerfile enhance: new nightly ci to support testing distributed-streaming (#35672) 2024-08-27 14:24:59 +08:00
pytest.ini test: update import test case to support different dim (#33709) 2024-06-13 17:11:55 +08:00
README_CN.md [skip ci]Fix image size in readme (#13462) 2021-12-15 21:09:45 +08:00
README.md Block creating new error from status reason (#27426) 2023-10-07 11:29:32 +08:00
requirements.txt test: update test cases and sdk version (#37774) 2024-11-19 14:14:32 +08:00
run.sh

Guidelines for Test Framework

This document guides you through the Pytest-based PyMilvus test framework.

You can find the test code on GitHub.

Quick Start

Deploy Milvus

To accommodate the variety of requirements, Milvus offers as many as four deployment methods. PyMilvus supports Milvus deployed with any of the methods below:

  1. Build from source code

  2. Install with Docker Compose

  3. Install on Kunernetes

  4. Install with KinD

For test purposes, we recommend installing Milvus with KinD. KinD supports the ClickOnce deployment of Milvus and its test client. KinD deployment is tailored for scenarios with small data scale, such as development/debugging test cases and functional verification.

  • Prerequisites

  • Install KinD with script

    1. Enter the local directory of the code */milvus/tests/scripts/
    2. Build the KinD environment, and execute CI Regression test cases automatically:
    $ ./e2e-k8s.sh 
    
    • By default, KinD environment will be automatically cleaned up after the execution of the test case. If you need to keep the KinD environment:
    $ ./e2e-k8s.sh --skip-cleanup
    
    • Skip the automatic test case execution and keep the KinD environment:
    $ ./e2e-k8s.sh --skip-cleanup --skip-test --manual
    

Note: You need to log in to the containers of the test client to proceed manual execution and debugging of the test case.

  • See more script parameters:
$ ./e2e-k8s.sh --help
  • Export cluster logs:
$ kind export logs .

PyMilvus Test Environment Deployment and Case Execution

We recommend using Python 3 (3.8 or higher), consistent with the version supported by PyMilvus.

Note: Procedures listed below will be completed automatically if you deployed Milvus using KinD.

  1. Install the Python package prerequisite for the test, enter */milvus/tests/python_client/, and execute:

    $ pip install -r requirements.txt
    
  2. The default test log path is /tmp/ci_logs/ under the config directory. You can add environment variables to change the path before booting up test cases:

    $ export CI_LOG_PATH=/tmp/ci_logs/test/
    
Log Level Log File
debug ci_test_log.debug
info ci_test_log.log
error ci_test_log.err
  1. You can configure default parameters in pytest.ini under the root path. For instance:

    addopts = --host *.*.*.*  --html=/tmp/ci_logs/report.html
    

where host should be set as the IP address of the Milvus service, and *.html is the report generated for the test.

  1. Enter testcases directory, run following command, which is consistent with the command under the pytest framework, to execute the test case:

    $ python3 -W ignore -m pytest <test_file_name>
    

An Introduction to Test Modules

Module Overview

Module Overview

SDK Test Flow Chart

Working directories and files

  • base: stores the encapsulated PyMilvus module files, and setup & teardown functions for pytest framework.
  • check: stores the check module files for returned results from interface.
  • common: stores the files of common methods and parameters for test cases.
  • config: stores the basic configuration file.
  • testcases: stores test case scripts.
  • utils: stores utility programs, such as utility log and environment detection methods.
  • requirements.txt: specifies the python package required for executing test cases
  • conftest.py: you can compile fixture functions or local plugins in this file. These functions and plugins implement within the current folder and its subfolder.
  • pytest.ini: the main configuration file for pytest.

Critical design ideas

  • base/*_wrapper.py encapsulates the tested interface, uniformly processes requests from the interface, abstracts the returned results, and passes the results to check/func_check.py module for checking.
  • check/func_check.py encompasses result checking methods for each interface for invocation from test cases.
  • base/client_base.py uses pytest framework to process setup/teardown functions correspondingly.
  • Test case files in testcases folder should be compiled inheriting the TestcaseBase module from base/client_base.py. Compile the common methods and parameters used by test cases into the Common module for invocation.
  • Add global configurations under config directory, such as log path, etc.
  • Add global implementation methods under utils directory, such as utility log module.

Adding codes

This section specifies references while adding new test cases or framework tools.

Notice and best practices

  1. Coding style
  • Test files: each SDK category corresponds to a test file. So do load and search methods.

  • Test categories: test files fall into two categories

    • TestObjectParams:
      • Indicates the parameter test of corresponding interface. For instance, TestPartitionParams represents the parameter test for Partition interface.
      • Tests the target category/method under different parameter inputs. The parameter test will cover default, empty, none, datatype, maxsize, etc.
    • TestObjectOperations:
      • Indicates the function/operation test of corresponding interface. For instance, TestPartitionOperations represents the function/operation test for Partition interface.
      • Tests the target category/method with legit parameter inputs and interaction with other interfaces.
  • Testcase naming

    • TestObjectParams:

      • Name after the parameter input of the test case. For instance, test_partition_empty_name() represents test on performance with the empty string as the name parameter input.
    • TestObjectOperations

      • Name after the operation procedure of the test case. For instance, test_partition_drop_partition_twice() represents the test on the performance when dropping partitions twice consecutively.
      • Name after assertions. For instance, test_partition_maximum_partitions() represents test on the maximum number of partitions that can be created.
  1. Notice
  • Do not initialize PyMilvus objects in the test case files.
  • Generally, do not add log IDs to test case files.
  • Directly call the encapsulated methods or attributes in test cases, as shown below:

To create multiple partitions objects, call self.init_partition_wrap(), which returns the newly created partition objects. Call self.partition_wrap instead when you do not need multiple objects.

# create partition  -Call the default initialization method
partition_w = self.init_partition_wrap()
assert partition_w.is_empty
# create partition    -Directly call the encapsulated object
self.partition_wrap.init_partition(collection=collection_name, name=partition_name)
assert self.partition_wrap.is_empty
  • To test on the error or exception returned from interfaces:

    • Call check_task=CheckTasks.err_res.
    • Input the expected error ID and message.
    # create partition with collection is None
    self.partition_wrap.init_partition(collection=None, name=partition_name, check_task=CheckTasks.err_res, check_items={ct.err_code: 1, ct.err_msg: "'NoneType' object has no attribute"})
    
  • To test on the normal value returned from interfaces:

    • Call check_task=CheckTasks.check_partition_property. You can build new test methods in CheckTasks for invocation in test cases.
    • Input the expected result for test methods.
    # create partition
    partition_w = self.init_partition_wrap(collection_w, partition_name, check_task=CheckTasks.check_partition_property, check_items={"name": partition_name, "description": description, "is_empty": True, "num_entities": 0})
    
  1. Adding test cases
  • Find the encapsulated tested interface with the same name in the *_wrapper.py files under base directory. Each interface returns a list with two values, among which one is interface returned results of PyMilvus, and the other is the assertion of normal/abnormal results, i.e. True/False. The returned judgment can be used in the extra result checking of test cases.

  • Add the test cases in the corresponding test file of the tested interface in testcases folder. You can refer to all test files under this directory to create your own test cases as shown below:

    @pytest.mark.tags(CaseLabel.L1)
    @pytest.mark.parametrize("partition_name", [cf.gen_unique_str(prefix)])
    def test_partition_dropped_collection(self, partition_name):
        """
        target: verify create partition against a dropped collection
        method: 1. create collection1
                2. drop collection1
                3. create partition in collection1
        expected: raise exception
        """
        # create collection
        collection_w = self.init_collection_wrap()
        # drop collection
        collection_w.drop()
        # create partition failed
        self.partition_wrap.init_partition(collection_w.collection, partition_name, check_task=CheckTasks.err_res, check_items={ct.err_code: 4, ct.err_msg: "collection not found"})
    
  • Tips

    • Case comments encompass three parts: object, test method, and expected result. You should specify each part.
    • Initialize the tested category in the setup method of the Base category in the base/client_base.py file, as shown below:
    self.connection_wrap = ApiConnectionsWrapper()
    self.utility_wrap = ApiUtilityWrapper()
    self.collection_wrap = ApiCollectionWrapper()
    self.partition_wrap = ApiPartitionWrapper()
    self.index_wrap = ApiIndexWrapper()
    self.collection_schema_wrap = ApiCollectionSchemaWrapper()
    self.field_schema_wrap = ApiFieldSchemaWrapper()
    
    • Pass the parameters with corresponding encapsulated methods when calling the interface you need to test on. As shown below, align all parameters with those in PyMilvus interfaces except for check_task and check_items.
    def init_partition(self, collection, name, description="", check_task=None, check_items=None, **kwargs)
    
    • check_task is used to select the corresponding interface test method in the ResponseChecker check category in the check/func_check.py file. You can choose methods under the CheckTasks category in the common/common_type.py file.
    • The specific content of check_items passed to the test method is determined by the implemented test method check_task.
    • The tested interface can return normal results when CheckTasks and check_items are not passed.
  1. Adding framework functions
  • Add global methods or tools under utils directory.

  • Add corresponding configurations under config directory.