// Copyright (C) 2019-2020 Zilliz. All rights reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance // with the License. You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software distributed under the License // is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express // or implied. See the License for the specific language governing permissions and limitations under the License #include #include "knowhere/index/vector_index/IndexIVF.h" #include "knowhere/index/vector_index/VecIndex.h" #include "knowhere/index/vector_index/adapter/VectorAdapter.h" #include "segcore/SegmentSealedImpl.h" #include "test_utils/DataGen.h" using namespace milvus; using namespace milvus::query; using namespace milvus::segcore; const int64_t ROW_COUNT = 100 * 1000; TEST(Sealed, without_predicate) { using namespace milvus::query; using namespace milvus::segcore; auto schema = std::make_shared(); auto dim = 16; auto topK = 5; auto metric_type = MetricType::METRIC_L2; auto fake_id = schema->AddDebugField("fakevec", DataType::VECTOR_FLOAT, dim, metric_type); auto float_fid = schema->AddDebugField("age", DataType::FLOAT); auto i64_fid = schema->AddDebugField("counter", DataType::INT64); schema->set_primary_field_id(i64_fid); std::string dsl = R"({ "bool": { "must": [ { "vector": { "fakevec": { "metric_type": "L2", "params": { "nprobe": 10 }, "query": "$0", "topk": 5, "round_decimal": 3 } } } ] } })"; auto N = ROW_COUNT; auto dataset = DataGen(schema, N); auto vec_col = dataset.get_col(fake_id); for (int64_t i = 0; i < 1000 * dim; ++i) { vec_col.push_back(0); } auto query_ptr = vec_col.data() + 4200 * dim; auto segment = CreateGrowingSegment(schema); segment->PreInsert(N); segment->Insert(0, N, dataset.row_ids_.data(), dataset.timestamps_.data(), dataset.raw_); auto plan = CreatePlan(*schema, dsl); auto num_queries = 5; auto ph_group_raw = CreatePlaceholderGroupFromBlob(num_queries, 16, query_ptr); auto ph_group = ParsePlaceholderGroup(plan.get(), ph_group_raw.SerializeAsString()); Timestamp time = 1000000; std::vector ph_group_arr = {ph_group.get()}; auto sr = segment->Search(plan.get(), ph_group.get(), time); auto pre_result = SearchResultToJson(*sr); auto indexing = std::make_shared(); auto conf = knowhere::Config{{knowhere::meta::DIM, dim}, {knowhere::meta::TOPK, topK}, {knowhere::IndexParams::nlist, 100}, {knowhere::IndexParams::nprobe, 10}, {knowhere::Metric::TYPE, knowhere::Metric::L2}, {knowhere::meta::DEVICEID, 0}}; auto database = knowhere::GenDataset(N, dim, vec_col.data() + 1000 * dim); indexing->Train(database, conf); indexing->AddWithoutIds(database, conf); EXPECT_EQ(indexing->Count(), N); EXPECT_EQ(indexing->Dim(), dim); auto query_dataset = knowhere::GenDataset(num_queries, dim, query_ptr); auto result = indexing->Query(query_dataset, conf, nullptr); auto ids = result->Get(knowhere::meta::IDS); // for comparison auto dis = result->Get(knowhere::meta::DISTANCE); // for comparison std::vector vec_ids(ids, ids + topK * num_queries); std::vector vec_dis(dis, dis + topK * num_queries); sr->seg_offsets_ = vec_ids; sr->distances_ = vec_dis; auto ref_result = SearchResultToJson(*sr); LoadIndexInfo load_info; load_info.field_id = fake_id.get(); load_info.index = indexing; load_info.index_params["metric_type"] = "L2"; // load index for vec field, load raw data for scalar filed auto sealed_segment = SealedCreator(schema, dataset); sealed_segment->DropFieldData(fake_id); sealed_segment->LoadIndex(load_info); sr = sealed_segment->Search(plan.get(), ph_group.get(), time); auto post_result = SearchResultToJson(*sr); std::cout << "ref_result" << std::endl; std::cout << ref_result.dump(1) << std::endl; std::cout << "post_result" << std::endl; std::cout << post_result.dump(1); // ASSERT_EQ(ref_result.dump(1), post_result.dump(1)); } TEST(Sealed, with_predicate) { using namespace milvus::query; using namespace milvus::segcore; auto schema = std::make_shared(); auto dim = 16; auto topK = 5; auto metric_type = MetricType::METRIC_L2; auto fake_id = schema->AddDebugField("fakevec", DataType::VECTOR_FLOAT, dim, metric_type); auto i64_fid = schema->AddDebugField("counter", DataType::INT64); schema->set_primary_field_id(i64_fid); std::string dsl = R"({ "bool": { "must": [ { "range": { "counter": { "GE": 42000, "LT": 42005 } } }, { "vector": { "fakevec": { "metric_type": "L2", "params": { "nprobe": 10 }, "query": "$0", "topk": 5, "round_decimal": 6 } } } ] } })"; auto N = ROW_COUNT; auto dataset = DataGen(schema, N); auto vec_col = dataset.get_col(fake_id); auto query_ptr = vec_col.data() + 42000 * dim; auto segment = CreateGrowingSegment(schema); segment->PreInsert(N); segment->Insert(0, N, dataset.row_ids_.data(), dataset.timestamps_.data(), dataset.raw_); auto plan = CreatePlan(*schema, dsl); auto num_queries = 5; auto ph_group_raw = CreatePlaceholderGroupFromBlob(num_queries, 16, query_ptr); auto ph_group = ParsePlaceholderGroup(plan.get(), ph_group_raw.SerializeAsString()); Timestamp time = 10000000; std::vector ph_group_arr = {ph_group.get()}; auto sr = segment->Search(plan.get(), ph_group.get(), time); auto indexing = std::make_shared(); auto conf = knowhere::Config{{knowhere::meta::DIM, dim}, {knowhere::meta::TOPK, topK}, {knowhere::IndexParams::nlist, 100}, {knowhere::IndexParams::nprobe, 10}, {knowhere::Metric::TYPE, knowhere::Metric::L2}, {knowhere::meta::DEVICEID, 0}}; auto database = knowhere::GenDataset(N, dim, vec_col.data()); indexing->Train(database, conf); indexing->AddWithoutIds(database, conf); EXPECT_EQ(indexing->Count(), N); EXPECT_EQ(indexing->Dim(), dim); auto query_dataset = knowhere::GenDataset(num_queries, dim, query_ptr); auto result = indexing->Query(query_dataset, conf, nullptr); LoadIndexInfo load_info; load_info.field_id = fake_id.get(); load_info.index = indexing; load_info.index_params["metric_type"] = "L2"; // load index for vec field, load raw data for scalar filed auto sealed_segment = SealedCreator(schema, dataset); sealed_segment->DropFieldData(fake_id); sealed_segment->LoadIndex(load_info); sr = sealed_segment->Search(plan.get(), ph_group.get(), time); for (int i = 0; i < num_queries; ++i) { auto offset = i * topK; ASSERT_EQ(sr->seg_offsets_[offset], 42000 + i); ASSERT_EQ(sr->distances_[offset], 0.0); } } TEST(Sealed, LoadFieldData) { auto dim = 16; auto topK = 5; auto N = ROW_COUNT; auto metric_type = MetricType::METRIC_L2; auto schema = std::make_shared(); auto fakevec_id = schema->AddDebugField("fakevec", DataType::VECTOR_FLOAT, dim, metric_type); auto counter_id = schema->AddDebugField("counter", DataType::INT64); auto double_id = schema->AddDebugField("double", DataType::DOUBLE); auto nothing_id = schema->AddDebugField("nothing", DataType::INT32); auto str_id = schema->AddDebugField("str", DataType::VARCHAR); schema->set_primary_field_id(counter_id); auto dataset = DataGen(schema, N); auto fakevec = dataset.get_col(fakevec_id); auto indexing = GenVecIndexing(N, dim, fakevec.data()); auto segment = CreateSealedSegment(schema); std::string dsl = R"({ "bool": { "must": [ { "range": { "double": { "GE": -1, "LT": 1 } } }, { "vector": { "fakevec": { "metric_type": "L2", "params": { "nprobe": 10 }, "query": "$0", "topk": 5, "round_decimal": 3 } } } ] } })"; Timestamp time = 1000000; auto plan = CreatePlan(*schema, dsl); auto num_queries = 5; auto ph_group_raw = CreatePlaceholderGroup(num_queries, 16, 1024); auto ph_group = ParsePlaceholderGroup(plan.get(), ph_group_raw.SerializeAsString()); ASSERT_ANY_THROW(segment->Search(plan.get(), ph_group.get(), time)); SealedLoadFieldData(dataset, *segment); segment->DropFieldData(nothing_id); segment->Search(plan.get(), ph_group.get(), time); segment->DropFieldData(fakevec_id); ASSERT_ANY_THROW(segment->Search(plan.get(), ph_group.get(), time)); LoadIndexInfo vec_info; vec_info.field_id = fakevec_id.get(); vec_info.index = indexing; vec_info.index_params["metric_type"] = knowhere::Metric::L2; segment->LoadIndex(vec_info); ASSERT_EQ(segment->num_chunk(), 1); ASSERT_EQ(segment->num_chunk_index(double_id), 0); ASSERT_EQ(segment->num_chunk_index(str_id), 0); auto chunk_span1 = segment->chunk_data(counter_id, 0); auto chunk_span2 = segment->chunk_data(double_id, 0); auto chunk_span3 = segment->chunk_data(str_id, 0); auto ref1 = dataset.get_col(counter_id); auto ref2 = dataset.get_col(double_id); auto ref3 = dataset.get_col(str_id)->scalars().string_data().data(); for (int i = 0; i < N; ++i) { ASSERT_EQ(chunk_span1[i], ref1[i]); ASSERT_EQ(chunk_span2[i], ref2[i]); ASSERT_EQ(chunk_span3[i], ref3[i]); } auto sr = segment->Search(plan.get(), ph_group.get(), time); auto json = SearchResultToJson(*sr); std::cout << json.dump(1); segment->DropIndex(fakevec_id); ASSERT_ANY_THROW(segment->Search(plan.get(), ph_group.get(), time)); segment->LoadIndex(vec_info); auto sr2 = segment->Search(plan.get(), ph_group.get(), time); auto json2 = SearchResultToJson(*sr); ASSERT_EQ(json.dump(-2), json2.dump(-2)); segment->DropFieldData(double_id); ASSERT_ANY_THROW(segment->Search(plan.get(), ph_group.get(), time)); #ifdef __linux__ auto std_json = Json::parse(R"( [ [ ["982->0.000000", "25315->4.742000", "57893->4.758000", "48201->6.075000", "53853->6.223000"], ["41772->10.111000", "74859->11.790000", "79777->11.842000", "3785->11.983000", "35888->12.193000"], ["59251->2.543000", "65551->4.454000", "72204->5.332000", "96905->5.479000", "87833->5.765000"], ["59219->5.458000", "21995->6.078000", "97922->6.764000", "25710->7.158000", "14048->7.294000"], ["66353->5.696000", "30664->5.881000", "41087->5.917000", "10393->6.633000", "90215->7.202000"] ] ])"); #else // for mac auto std_json = Json::parse(R"( [ [ ["982->0.000000", "31864->4.270000", "18916->4.651000", "71547->5.125000", "86706->5.991000"], ["96984->4.192000", "65514->6.011000", "89328->6.138000", "80284->6.526000", "68218->6.563000"], ["30119->2.464000", "82365->4.725000", "74834->5.009000", "79995->5.725000", "33359->5.816000"], ["99625->6.129000", "86582->6.900000", "85934->7.792000", "60450->8.087000", "19257->8.530000"], ["37759->3.581000", "31292->5.780000", "98124->6.216000", "63535->6.439000", "11707->6.553000"] ] ])"); #endif ASSERT_EQ(std_json.dump(-2), json.dump(-2)); } TEST(Sealed, LoadScalarIndex) { auto dim = 16; auto N = ROW_COUNT; auto metric_type = MetricType::METRIC_L2; auto schema = std::make_shared(); auto fakevec_id = schema->AddDebugField("fakevec", DataType::VECTOR_FLOAT, dim, metric_type); auto counter_id = schema->AddDebugField("counter", DataType::INT64); auto double_id = schema->AddDebugField("double", DataType::DOUBLE); auto nothing_id = schema->AddDebugField("nothing", DataType::INT32); schema->set_primary_field_id(counter_id); auto dataset = DataGen(schema, N); auto fakevec = dataset.get_col(fakevec_id); auto indexing = GenVecIndexing(N, dim, fakevec.data()); auto segment = CreateSealedSegment(schema); std::string dsl = R"({ "bool": { "must": [ { "range": { "double": { "GE": -1, "LT": 1 } } }, { "vector": { "fakevec": { "metric_type": "L2", "params": { "nprobe": 10 }, "query": "$0", "topk": 5, "round_decimal": 3 } } } ] } })"; Timestamp time = 1000000; auto plan = CreatePlan(*schema, dsl); auto num_queries = 5; auto ph_group_raw = CreatePlaceholderGroup(num_queries, 16, 1024); auto ph_group = ParsePlaceholderGroup(plan.get(), ph_group_raw.SerializeAsString()); LoadFieldDataInfo row_id_info; FieldMeta row_id_field_meta(FieldName("RowID"), RowFieldID, DataType::INT64); auto array = CreateScalarDataArrayFrom(dataset.row_ids_.data(), N, row_id_field_meta); row_id_info.field_data = array.release(); row_id_info.row_count = dataset.row_ids_.size(); row_id_info.field_id = RowFieldID.get(); // field id for RowId segment->LoadFieldData(row_id_info); LoadFieldDataInfo ts_info; FieldMeta ts_field_meta(FieldName("Timestamp"), TimestampFieldID, DataType::INT64); array = CreateScalarDataArrayFrom(dataset.timestamps_.data(), N, ts_field_meta); ts_info.field_data = array.release(); ts_info.row_count = dataset.timestamps_.size(); ts_info.field_id = TimestampFieldID.get(); segment->LoadFieldData(ts_info); LoadIndexInfo vec_info; vec_info.field_id = fakevec_id.get(); vec_info.field_type = CDataType::FloatVector; vec_info.index = indexing; vec_info.index_params["metric_type"] = knowhere::Metric::L2; segment->LoadIndex(vec_info); LoadIndexInfo counter_index; counter_index.field_id = counter_id.get(); counter_index.field_type = CDataType::Int64; counter_index.index_params["index_type"] = "sort"; auto counter_data = dataset.get_col(counter_id); counter_index.index = std::move(GenScalarIndexing(N, counter_data.data())); segment->LoadIndex(counter_index); LoadIndexInfo double_index; double_index.field_id = double_id.get(); double_index.field_type = CDataType::Double; double_index.index_params["index_type"] = "sort"; auto double_data = dataset.get_col(double_id); double_index.index = std::move(GenScalarIndexing(N, double_data.data())); segment->LoadIndex(double_index); LoadIndexInfo nothing_index; nothing_index.field_id = nothing_id.get(); nothing_index.field_type = CDataType::Int32; nothing_index.index_params["index_type"] = "sort"; auto nothing_data = dataset.get_col(nothing_id); nothing_index.index = std::move(GenScalarIndexing(N, nothing_data.data())); segment->LoadIndex(nothing_index); auto sr = segment->Search(plan.get(), ph_group.get(), time); auto json = SearchResultToJson(*sr); std::cout << json.dump(1); } TEST(Sealed, Delete) { auto dim = 16; auto topK = 5; auto N = 10; auto metric_type = MetricType::METRIC_L2; auto schema = std::make_shared(); auto fakevec_id = schema->AddDebugField("fakevec", DataType::VECTOR_FLOAT, dim, metric_type); auto counter_id = schema->AddDebugField("counter", DataType::INT64); auto double_id = schema->AddDebugField("double", DataType::DOUBLE); auto nothing_id = schema->AddDebugField("nothing", DataType::INT32); schema->set_primary_field_id(counter_id); auto dataset = DataGen(schema, N); auto fakevec = dataset.get_col(fakevec_id); auto segment = CreateSealedSegment(schema); std::string dsl = R"({ "bool": { "must": [ { "range": { "double": { "GE": -1, "LT": 1 } } }, { "vector": { "fakevec": { "metric_type": "L2", "params": { "nprobe": 10 }, "query": "$0", "topk": 5, "round_decimal": 3 } } } ] } })"; Timestamp time = 1000000; auto plan = CreatePlan(*schema, dsl); auto num_queries = 5; auto ph_group_raw = CreatePlaceholderGroup(num_queries, 16, 1024); auto ph_group = ParsePlaceholderGroup(plan.get(), ph_group_raw.SerializeAsString()); ASSERT_ANY_THROW(segment->Search(plan.get(), ph_group.get(), time)); SealedLoadFieldData(dataset, *segment); int64_t row_count = 5; std::vector pks{1, 2, 3, 4, 5}; auto ids = std::make_unique(); ids->mutable_int_id()->mutable_data()->Add(pks.begin(), pks.end()); std::vector timestamps{10, 10, 10, 10, 10}; LoadDeletedRecordInfo info = {timestamps.data(), ids.get(), row_count}; segment->LoadDeletedRecord(info); std::vector tmp_block{0, 0}; BitsetType bitset(N, false); segment->mask_with_delete(bitset, 10, 11); ASSERT_EQ(bitset.count(), pks.size()); int64_t new_count = 3; std::vector new_pks{6, 7, 8}; auto new_ids = std::make_unique(); new_ids->mutable_int_id()->mutable_data()->Add(new_pks.begin(), new_pks.end()); std::vector new_timestamps{10, 10, 10}; auto reserved_offset = segment->PreDelete(new_count); ASSERT_EQ(reserved_offset, row_count); segment->Delete(reserved_offset, new_count, new_ids.get(), reinterpret_cast(new_timestamps.data())); }