Revert "Merge branch 'mysql-0.3.0' into 'branch-0.3.0'"

This reverts merge request !143


Former-commit-id: 8099ff77584b0af7a8d39d0c6289e8e31b2120c8
This commit is contained in:
zhiru.zhu 2019-07-01 15:38:35 +08:00 committed by jinhai
parent c969a6bbb7
commit 4d83c24bd7
5 changed files with 224 additions and 64 deletions

View File

@ -15,7 +15,6 @@ db_config:
index_building_threshold: 1024 # index building trigger threshold, default: 1024, unit: MB index_building_threshold: 1024 # index building trigger threshold, default: 1024, unit: MB
archive_disk_threshold: 512 # triger archive action if storage size exceed this value, unit: GB archive_disk_threshold: 512 # triger archive action if storage size exceed this value, unit: GB
archive_days_threshold: 30 # files older than x days will be archived, unit: day archive_days_threshold: 30 # files older than x days will be archived, unit: day
sql_echo: on # print sql statement in debug log
metric_config: metric_config:
is_startup: off # if monitoring start: on, off is_startup: off # if monitoring start: on, off

View File

@ -1,5 +1,7 @@
#!/bin/bash #!/bin/bash
export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$(pwd)/milvus/lib
LCOV_CMD="lcov" LCOV_CMD="lcov"
LCOV_GEN_CMD="genhtml" LCOV_GEN_CMD="genhtml"
@ -40,7 +42,6 @@ if [ $? -ne 0 ]; then
fi fi
for test in `ls ${DIR_UNITTEST}`; do for test in `ls ${DIR_UNITTEST}`; do
echo $test
case ${test} in case ${test} in
db_test) db_test)
# set run args for db_test # set run args for db_test
@ -57,8 +58,6 @@ for test in `ls ${DIR_UNITTEST}`; do
fi fi
done done
mysql_exc "DROP DATABASE IF EXISTS ${MYSQL_DB_NAME};"
# gen test converage # gen test converage
${LCOV_CMD} -d ${DIR_GCNO} -o "${FILE_INFO_MILVUS}" -c ${LCOV_CMD} -d ${DIR_GCNO} -o "${FILE_INFO_MILVUS}" -c
# merge coverage # merge coverage
@ -71,4 +70,4 @@ ${LCOV_CMD} -r "${FILE_INFO_OUTPUT}" -o "${FILE_INFO_OUTPUT_NEW}" \
"*/cmake_build/*_ep-prefix/*" \ "*/cmake_build/*_ep-prefix/*" \
# gen html report # gen html report
${LCOV_GEN_CMD} "${FILE_INFO_OUTPUT_NEW}" --output-directory ${DIR_LCOV_OUTPUT}/ ${LCOV_GEN_CMD} "${FILE_INFO_OUTPUT_NEW}" --output-directory ${DIR_LCOV_OUTPUT}/

View File

@ -64,10 +64,6 @@ public:
max_idle_time_ = max_idle; max_idle_time_ = max_idle;
} }
std::string getDB() {
return db_;
}
protected: protected:
// Superclass overrides // Superclass overrides

View File

@ -1733,30 +1733,17 @@ namespace meta {
// ENGINE_LOG_WARNING << "MySQLMetaImpl::CleanUp: connection in use = " << mysql_connection_pool_->getConnectionsInUse(); // ENGINE_LOG_WARNING << "MySQLMetaImpl::CleanUp: connection in use = " << mysql_connection_pool_->getConnectionsInUse();
// } // }
ENGINE_LOG_DEBUG << "Remove table file type as NEW";
Query cleanUpQuery = connectionPtr->query(); Query cleanUpQuery = connectionPtr->query();
cleanUpQuery << "SELECT table_name " << cleanUpQuery << "DELETE FROM TableFiles WHERE file_type = " << std::to_string(TableFileSchema::NEW) << ";";
"FROM information_schema.tables " <<
"WHERE table_schema = " << quote << mysql_connection_pool_->getDB() << quote << " " <<
"AND table_name = " << quote << "TableFiles" << quote << ";";
if (options_.sql_echo) { if (options_.sql_echo) {
ENGINE_LOG_DEBUG << "MySQLMetaImpl::CleanUp: " << cleanUpQuery.str(); ENGINE_LOG_DEBUG << "MySQLMetaImpl::CleanUp: " << cleanUpQuery.str();
} }
StoreQueryResult res = cleanUpQuery.store(); if (!cleanUpQuery.exec()) {
assert(res); ENGINE_LOG_ERROR << "QUERY ERROR WHEN CLEANING UP FILES";
if (!res.empty()) { return Status::DBTransactionError("Clean up Error", cleanUpQuery.error());
ENGINE_LOG_DEBUG << "Remove table file type as NEW";
cleanUpQuery << "DELETE FROM TableFiles WHERE file_type = " << std::to_string(TableFileSchema::NEW) << ";";
if (options_.sql_echo) {
ENGINE_LOG_DEBUG << "MySQLMetaImpl::CleanUp: " << cleanUpQuery.str();
}
if (!cleanUpQuery.exec()) {
ENGINE_LOG_ERROR << "QUERY ERROR WHEN CLEANING UP FILES";
return Status::DBTransactionError("Clean up Error", cleanUpQuery.error());
}
} }
} catch (const BadQuery& er) { } catch (const BadQuery& er) {

View File

@ -21,7 +21,184 @@
using namespace zilliz::milvus::engine; using namespace zilliz::milvus::engine;
TEST_F(MySQLTest, TABLE_TEST) { //TEST_F(MySQLTest, InitializeTest) {
// DBMetaOptions options;
// //dialect+driver://username:password@host:port/database
// options.backend_uri = "mysql://root:1234@:/test";
// meta::MySQLMetaImpl impl(options);
// auto status = impl.Initialize();
// std::cout << status.ToString() << std::endl;
// ASSERT_TRUE(status.ok());
//}
TEST_F(MySQLTest, core) {
DBMetaOptions options;
// //dialect+driver://username:password@host:port/database
// options.backend_uri = "mysql://root:1234@:/test";
// options.path = "/tmp/vecwise_test";
try {
options = getDBMetaOptions();
} catch(std::exception& ex) {
ASSERT_TRUE(false);
return;
}
int mode = Options::MODE::SINGLE;
meta::MySQLMetaImpl impl(options, mode);
// auto status = impl.Initialize();
// ASSERT_TRUE(status.ok());
meta::TableSchema schema1;
schema1.table_id_ = "test1";
schema1.dimension_ = 123;
auto status = impl.CreateTable(schema1);
// std::cout << status.ToString() << std::endl;
ASSERT_TRUE(status.ok());
meta::TableSchema schema2;
schema2.table_id_ = "test2";
schema2.dimension_ = 321;
status = impl.CreateTable(schema2);
// std::cout << status.ToString() << std::endl;
ASSERT_TRUE(status.ok());
status = impl.CreateTable(schema2);
// std::cout << status.ToString() << std::endl;
// ASSERT_THROW(impl.CreateTable(schema), mysqlpp::BadQuery);
ASSERT_TRUE(status.ok());
status = impl.DeleteTable(schema2.table_id_);
// std::cout << status.ToString() << std::endl;
ASSERT_TRUE(status.ok());
size_t id1 = schema1.id_;
long created_on1 = schema1.created_on_;
status = impl.DescribeTable(schema1);
ASSERT_TRUE(status.ok());
ASSERT_EQ(schema1.id_, id1);
ASSERT_EQ(schema1.table_id_, "test1");
ASSERT_EQ(schema1.created_on_, created_on1);
ASSERT_EQ(schema1.files_cnt_, 0);
ASSERT_EQ(schema1.engine_type_, 1);
ASSERT_EQ(schema1.store_raw_data_, false);
bool check;
status = impl.HasTable("test1", check);
ASSERT_TRUE(status.ok());
ASSERT_EQ(check, true);
std::vector<meta::TableSchema> table_schema_array;
status = impl.AllTables(table_schema_array);
ASSERT_TRUE(status.ok());
ASSERT_EQ(table_schema_array.size(), 1);
meta::TableSchema resultSchema = table_schema_array[0];
ASSERT_EQ(resultSchema.id_, id1);
ASSERT_EQ(resultSchema.table_id_, "test1");
ASSERT_EQ(resultSchema.dimension_, 123);
ASSERT_EQ(resultSchema.files_cnt_, 0);
ASSERT_EQ(resultSchema.engine_type_, 1);
ASSERT_EQ(resultSchema.store_raw_data_, false);
meta::TableFileSchema tableFileSchema;
tableFileSchema.table_id_ = "test1";
status = impl.CreateTableFile(tableFileSchema);
// std::cout << status.ToString() << std::endl;
ASSERT_TRUE(status.ok());
tableFileSchema.file_type_ = meta::TableFileSchema::TO_INDEX;
status = impl.UpdateTableFile(tableFileSchema);
// std::cout << status.ToString() << std::endl;
ASSERT_TRUE(status.ok());
meta::TableFilesSchema filesToIndex;
status = impl.FilesToIndex(filesToIndex);
ASSERT_TRUE(status.ok());
ASSERT_EQ(filesToIndex.size(), 1);
meta::TableFileSchema fileToIndex = filesToIndex[0];
ASSERT_EQ(fileToIndex.table_id_, "test1");
ASSERT_EQ(fileToIndex.dimension_, 123);
// meta::TableFilesSchema filesToIndex;
// status = impl.FilesToIndex(filesToIndex);
// ASSERT_TRUE(status.ok());
// ASSERT_EQ(filesToIndex.size(), 0);
meta::DatesT partition;
partition.push_back(tableFileSchema.date_);
meta::DatePartionedTableFilesSchema filesToSearch;
status = impl.FilesToSearch(tableFileSchema.table_id_, partition, filesToSearch);
ASSERT_TRUE(status.ok());
ASSERT_EQ(filesToSearch.size(), 1);
ASSERT_EQ(filesToSearch[tableFileSchema.date_].size(), 1);
meta::TableFileSchema fileToSearch = filesToSearch[tableFileSchema.date_][0];
ASSERT_EQ(fileToSearch.table_id_, "test1");
ASSERT_EQ(fileToSearch.dimension_, 123);
tableFileSchema.file_type_ = meta::TableFileSchema::RAW;
status = impl.UpdateTableFile(tableFileSchema);
ASSERT_TRUE(status.ok());
meta::DatePartionedTableFilesSchema filesToMerge;
status = impl.FilesToMerge(tableFileSchema.table_id_, filesToMerge);
// std::cout << status.ToString() << std::endl;
ASSERT_TRUE(status.ok());
ASSERT_EQ(filesToMerge.size(), 1);
ASSERT_EQ(filesToMerge[tableFileSchema.date_].size(), 1);
meta::TableFileSchema fileToMerge = filesToMerge[tableFileSchema.date_][0];
ASSERT_EQ(fileToMerge.table_id_, "test1");
ASSERT_EQ(fileToMerge.dimension_, 123);
meta::TableFilesSchema resultTableFilesSchema;
std::vector<size_t> ids;
ids.push_back(tableFileSchema.id_);
status = impl.GetTableFiles(tableFileSchema.table_id_, ids, resultTableFilesSchema);
ASSERT_TRUE(status.ok());
ASSERT_EQ(resultTableFilesSchema.size(), 1);
meta::TableFileSchema resultTableFileSchema = resultTableFilesSchema[0];
// ASSERT_EQ(resultTableFileSchema.id_, tableFileSchema.id_);
ASSERT_EQ(resultTableFileSchema.table_id_, tableFileSchema.table_id_);
ASSERT_EQ(resultTableFileSchema.file_id_, tableFileSchema.file_id_);
ASSERT_EQ(resultTableFileSchema.file_type_, tableFileSchema.file_type_);
ASSERT_EQ(resultTableFileSchema.size_, tableFileSchema.size_);
ASSERT_EQ(resultTableFileSchema.date_, tableFileSchema.date_);
ASSERT_EQ(resultTableFileSchema.engine_type_, tableFileSchema.engine_type_);
ASSERT_EQ(resultTableFileSchema.dimension_, tableFileSchema.dimension_);
tableFileSchema.size_ = 234;
meta::TableSchema schema3;
schema3.table_id_ = "test3";
schema3.dimension_ = 321;
status = impl.CreateTable(schema3);
ASSERT_TRUE(status.ok());
meta::TableFileSchema tableFileSchema2;
tableFileSchema2.table_id_ = "test3";
tableFileSchema2.size_ = 345;
status = impl.CreateTableFile(tableFileSchema2);
ASSERT_TRUE(status.ok());
meta::TableFilesSchema filesToUpdate;
filesToUpdate.emplace_back(tableFileSchema);
filesToUpdate.emplace_back(tableFileSchema2);
status = impl.UpdateTableFile(tableFileSchema);
ASSERT_TRUE(status.ok());
uint64_t resultSize;
status = impl.Size(resultSize);
// std::cout << status.ToString() << std::endl;
ASSERT_TRUE(status.ok());
ASSERT_EQ(resultSize, tableFileSchema.size_ + tableFileSchema2.size_);
uint64_t countResult;
status = impl.Count(tableFileSchema.table_id_, countResult);
ASSERT_TRUE(status.ok());
status = impl.DropAll();
ASSERT_TRUE(status.ok());
}
TEST_F(MySQLTest, GROUP_TEST) {
DBMetaOptions options; DBMetaOptions options;
try { try {
options = getDBMetaOptions(); options = getDBMetaOptions();
@ -33,37 +210,38 @@ TEST_F(MySQLTest, TABLE_TEST) {
int mode = Options::MODE::SINGLE; int mode = Options::MODE::SINGLE;
meta::MySQLMetaImpl impl(options, mode); meta::MySQLMetaImpl impl(options, mode);
auto table_id = "meta_test_table"; auto table_id = "meta_test_group";
meta::TableSchema table; meta::TableSchema group;
table.table_id_ = table_id; group.table_id_ = table_id;
auto status = impl.CreateTable(table); auto status = impl.CreateTable(group);
ASSERT_TRUE(status.ok()); ASSERT_TRUE(status.ok());
auto gid = table.id_; auto gid = group.id_;
table.id_ = -1; group.id_ = -1;
status = impl.DescribeTable(table); status = impl.DescribeTable(group);
ASSERT_TRUE(status.ok()); ASSERT_TRUE(status.ok());
ASSERT_EQ(table.id_, gid); ASSERT_EQ(group.id_, gid);
ASSERT_EQ(table.table_id_, table_id); ASSERT_EQ(group.table_id_, table_id);
table.table_id_ = "not_found"; group.table_id_ = "not_found";
status = impl.DescribeTable(table); status = impl.DescribeTable(group);
ASSERT_TRUE(!status.ok()); ASSERT_TRUE(!status.ok());
table.table_id_ = table_id; group.table_id_ = table_id;
status = impl.CreateTable(table); status = impl.CreateTable(group);
ASSERT_TRUE(status.ok()); ASSERT_TRUE(status.ok());
table.table_id_ = ""; group.table_id_ = "";
status = impl.CreateTable(table); status = impl.CreateTable(group);
ASSERT_TRUE(status.ok()); ASSERT_TRUE(status.ok());
status = impl.DropAll(); status = impl.DropAll();
ASSERT_TRUE(status.ok()); ASSERT_TRUE(status.ok());
} }
TEST_F(MySQLTest, TABLE_FILE_TEST) { TEST_F(MySQLTest, table_file_TEST) {
DBMetaOptions options; DBMetaOptions options;
try { try {
options = getDBMetaOptions(); options = getDBMetaOptions();
@ -75,16 +253,17 @@ TEST_F(MySQLTest, TABLE_FILE_TEST) {
int mode = Options::MODE::SINGLE; int mode = Options::MODE::SINGLE;
meta::MySQLMetaImpl impl(options, mode); meta::MySQLMetaImpl impl(options, mode);
auto table_id = "meta_test_table"; auto table_id = "meta_test_group";
meta::TableSchema table; meta::TableSchema group;
table.table_id_ = table_id; group.table_id_ = table_id;
table.dimension_ = 256; group.dimension_ = 256;
auto status = impl.CreateTable(table); auto status = impl.CreateTable(group);
meta::TableFileSchema table_file; meta::TableFileSchema table_file;
table_file.table_id_ = table.table_id_; table_file.table_id_ = group.table_id_;
status = impl.CreateTableFile(table_file); status = impl.CreateTableFile(table_file);
// std::cout << status.ToString() << std::endl;
ASSERT_TRUE(status.ok()); ASSERT_TRUE(status.ok());
ASSERT_EQ(table_file.file_type_, meta::TableFileSchema::NEW); ASSERT_EQ(table_file.file_type_, meta::TableFileSchema::NEW);
@ -153,15 +332,15 @@ TEST_F(MySQLTest, ARCHIVE_TEST_DAYS) {
int mode = Options::MODE::SINGLE; int mode = Options::MODE::SINGLE;
meta::MySQLMetaImpl impl(options, mode); meta::MySQLMetaImpl impl(options, mode);
auto table_id = "meta_test_table"; auto table_id = "meta_test_group";
meta::TableSchema table; meta::TableSchema group;
table.table_id_ = table_id; group.table_id_ = table_id;
auto status = impl.CreateTable(table); auto status = impl.CreateTable(group);
meta::TableFilesSchema files; meta::TableFilesSchema files;
meta::TableFileSchema table_file; meta::TableFileSchema table_file;
table_file.table_id_ = table.table_id_; table_file.table_id_ = group.table_id_;
auto cnt = 100; auto cnt = 100;
long ts = utils::GetMicroSecTimeStamp(); long ts = utils::GetMicroSecTimeStamp();
@ -212,13 +391,13 @@ TEST_F(MySQLTest, ARCHIVE_TEST_DISK) {
auto impl = meta::MySQLMetaImpl(options, mode); auto impl = meta::MySQLMetaImpl(options, mode);
auto table_id = "meta_test_group"; auto table_id = "meta_test_group";
meta::TableSchema table; meta::TableSchema group;
table.table_id_ = table_id; group.table_id_ = table_id;
auto status = impl.CreateTable(table); auto status = impl.CreateTable(group);
meta::TableFilesSchema files; meta::TableFilesSchema files;
meta::TableFileSchema table_file; meta::TableFileSchema table_file;
table_file.table_id_ = table.table_id_; table_file.table_id_ = group.table_id_;
auto cnt = 10; auto cnt = 10;
auto each_size = 2UL; auto each_size = 2UL;
@ -266,9 +445,9 @@ TEST_F(MySQLTest, TABLE_FILES_TEST) {
auto table_id = "meta_test_group"; auto table_id = "meta_test_group";
meta::TableSchema table; meta::TableSchema group;
table.table_id_ = table_id; group.table_id_ = table_id;
auto status = impl.CreateTable(table); auto status = impl.CreateTable(group);
int new_files_cnt = 4; int new_files_cnt = 4;
int raw_files_cnt = 5; int raw_files_cnt = 5;
@ -276,7 +455,7 @@ TEST_F(MySQLTest, TABLE_FILES_TEST) {
int index_files_cnt = 7; int index_files_cnt = 7;
meta::TableFileSchema table_file; meta::TableFileSchema table_file;
table_file.table_id_ = table.table_id_; table_file.table_id_ = group.table_id_;
for (auto i=0; i<new_files_cnt; ++i) { for (auto i=0; i<new_files_cnt; ++i) {
status = impl.CreateTableFile(table_file); status = impl.CreateTableFile(table_file);
@ -309,7 +488,7 @@ TEST_F(MySQLTest, TABLE_FILES_TEST) {
ASSERT_EQ(files.size(), to_index_files_cnt); ASSERT_EQ(files.size(), to_index_files_cnt);
meta::DatePartionedTableFilesSchema dated_files; meta::DatePartionedTableFilesSchema dated_files;
status = impl.FilesToMerge(table.table_id_, dated_files); status = impl.FilesToMerge(group.table_id_, dated_files);
ASSERT_TRUE(status.ok()); ASSERT_TRUE(status.ok());
ASSERT_EQ(dated_files[table_file.date_].size(), raw_files_cnt); ASSERT_EQ(dated_files[table_file.date_].size(), raw_files_cnt);