mirror of
https://gitee.com/milvus-io/milvus.git
synced 2024-12-03 20:39:36 +08:00
0a2655dba0
* increase waiting time * still apply chaos even timeout Signed-off-by: zhuwenxing <wenxing.zhu@zilliz.com>
132 lines
4.0 KiB
Python
132 lines
4.0 KiB
Python
import glob
|
|
import time
|
|
from yaml import full_load
|
|
import json
|
|
import pandas as pd
|
|
from utils.util_log import test_log as log
|
|
|
|
def gen_experiment_config(yaml):
|
|
"""load the yaml file of chaos experiment"""
|
|
with open(yaml) as f:
|
|
_config = full_load(f)
|
|
f.close()
|
|
return _config
|
|
|
|
|
|
def findkeys(node, kv):
|
|
# refer to https://stackoverflow.com/questions/9807634/find-all-occurrences-of-a-key-in-nested-dictionaries-and-lists
|
|
if isinstance(node, list):
|
|
for i in node:
|
|
for x in findkeys(i, kv):
|
|
yield x
|
|
elif isinstance(node, dict):
|
|
if kv in node:
|
|
yield node[kv]
|
|
for j in node.values():
|
|
for x in findkeys(j, kv):
|
|
yield x
|
|
|
|
|
|
def update_key_value(node, modify_k, modify_v):
|
|
# update the value of modify_k to modify_v
|
|
if isinstance(node, list):
|
|
for i in node:
|
|
update_key_value(i, modify_k, modify_v)
|
|
elif isinstance(node, dict):
|
|
if modify_k in node:
|
|
node[modify_k] = modify_v
|
|
for j in node.values():
|
|
update_key_value(j, modify_k, modify_v)
|
|
return node
|
|
|
|
|
|
def update_key_name(node, modify_k, modify_k_new):
|
|
# update the name of modify_k to modify_k_new
|
|
if isinstance(node, list):
|
|
for i in node:
|
|
update_key_name(i, modify_k, modify_k_new)
|
|
elif isinstance(node, dict):
|
|
if modify_k in node:
|
|
value_backup = node[modify_k]
|
|
del node[modify_k]
|
|
node[modify_k_new] = value_backup
|
|
for j in node.values():
|
|
update_key_name(j, modify_k, modify_k_new)
|
|
return node
|
|
|
|
|
|
def get_collections(file_name="all_collections.json"):
|
|
try:
|
|
with open(f"/tmp/ci_logs/{file_name}", "r") as f:
|
|
data = json.load(f)
|
|
collections = data["all"]
|
|
except Exception as e:
|
|
log.error(f"get_all_collections error: {e}")
|
|
return []
|
|
return collections
|
|
|
|
|
|
def get_deploy_test_collections():
|
|
try:
|
|
with open("/tmp/ci_logs/deploy_test_all_collections.json", "r") as f:
|
|
data = json.load(f)
|
|
collections = data["all"]
|
|
except Exception as e:
|
|
log.error(f"get_all_collections error: {e}")
|
|
return []
|
|
return collections
|
|
|
|
|
|
def get_chaos_test_collections():
|
|
try:
|
|
with open("/tmp/ci_logs/chaos_test_all_collections.json", "r") as f:
|
|
data = json.load(f)
|
|
collections = data["all"]
|
|
except Exception as e:
|
|
log.error(f"get_all_collections error: {e}")
|
|
return []
|
|
return collections
|
|
|
|
|
|
def wait_signal_to_apply_chaos():
|
|
all_db_file = glob.glob("/tmp/ci_logs/event_records*.parquet")
|
|
log.info(f"all files {all_db_file}")
|
|
ready_apply_chaos = True
|
|
timeout = 15*60
|
|
t0 = time.time()
|
|
for f in all_db_file:
|
|
while True and (time.time() - t0 < timeout):
|
|
try:
|
|
df = pd.read_parquet(f)
|
|
log.debug(f"read {f}:result\n {df}")
|
|
result = df[(df['event_name'] == 'init_chaos') & (df['event_status'] == 'ready')]
|
|
if len(result) > 0:
|
|
log.info(f"{f}: {result}")
|
|
ready_apply_chaos = True
|
|
break
|
|
else:
|
|
ready_apply_chaos = False
|
|
except Exception as e:
|
|
log.error(f"read_parquet error: {e}")
|
|
ready_apply_chaos = False
|
|
time.sleep(10)
|
|
|
|
return ready_apply_chaos
|
|
|
|
|
|
if __name__ == "__main__":
|
|
d = { "id" : "abcde",
|
|
"key1" : "blah",
|
|
"key2" : "blah blah",
|
|
"nestedlist" : [
|
|
{ "id" : "qwerty",
|
|
"nestednestedlist" : [
|
|
{ "id" : "xyz", "keyA" : "blah blah blah" },
|
|
{ "id" : "fghi", "keyZ" : "blah blah blah" }],
|
|
"anothernestednestedlist" : [
|
|
{ "id" : "asdf", "keyQ" : "blah blah" },
|
|
{ "id" : "yuiop", "keyW" : "blah" }] } ] }
|
|
print(list(findkeys(d, 'id')))
|
|
update_key_value(d, "none_id", "ccc")
|
|
print(d)
|