2024-03-02 03:03:03 +00:00
|
|
|
import json
|
|
|
|
import sys
|
|
|
|
import pytest
|
|
|
|
import time
|
2024-06-24 02:36:25 +00:00
|
|
|
import uuid
|
2024-11-05 00:42:23 +00:00
|
|
|
from pymilvus import connections, db, MilvusClient
|
2024-03-02 03:03:03 +00:00
|
|
|
from utils.util_log import test_log as logger
|
|
|
|
from api.milvus import (VectorClient, CollectionClient, PartitionClient, IndexClient, AliasClient,
|
2024-12-07 06:42:45 +00:00
|
|
|
UserClient, RoleClient, ImportJobClient, StorageClient, Requests, DatabaseClient)
|
2024-03-02 03:03:03 +00:00
|
|
|
from utils.utils import get_data_by_payload
|
|
|
|
|
|
|
|
|
|
|
|
def get_config():
|
|
|
|
pass
|
|
|
|
|
|
|
|
|
|
|
|
class Base:
|
|
|
|
name = None
|
|
|
|
protocol = None
|
|
|
|
host = None
|
|
|
|
port = None
|
|
|
|
endpoint = None
|
|
|
|
api_key = None
|
|
|
|
username = None
|
|
|
|
password = None
|
|
|
|
invalid_api_key = None
|
|
|
|
vector_client = None
|
|
|
|
collection_client = None
|
|
|
|
partition_client = None
|
|
|
|
index_client = None
|
2024-03-21 07:31:09 +00:00
|
|
|
alias_client = None
|
2024-03-02 03:03:03 +00:00
|
|
|
user_client = None
|
|
|
|
role_client = None
|
|
|
|
import_job_client = None
|
|
|
|
storage_client = None
|
2024-11-05 00:42:23 +00:00
|
|
|
milvus_client = None
|
2024-12-07 06:42:45 +00:00
|
|
|
database_client = None
|
2024-03-02 03:03:03 +00:00
|
|
|
|
|
|
|
|
|
|
|
class TestBase(Base):
|
2024-06-24 02:36:25 +00:00
|
|
|
req = None
|
2024-10-19 11:13:25 +00:00
|
|
|
|
2024-03-02 03:03:03 +00:00
|
|
|
def teardown_method(self):
|
2024-12-07 06:42:45 +00:00
|
|
|
# Clean up collections
|
2024-03-02 03:03:03 +00:00
|
|
|
self.collection_client.api_key = self.api_key
|
|
|
|
all_collections = self.collection_client.collection_list()['data']
|
|
|
|
if self.name in all_collections:
|
|
|
|
logger.info(f"collection {self.name} exist, drop it")
|
|
|
|
payload = {
|
|
|
|
"collectionName": self.name,
|
|
|
|
}
|
|
|
|
try:
|
|
|
|
rsp = self.collection_client.collection_drop(payload)
|
|
|
|
except Exception as e:
|
2024-12-07 06:42:45 +00:00
|
|
|
logger.error(f"drop collection error: {e}")
|
|
|
|
|
2024-10-31 13:18:23 +00:00
|
|
|
for item in self.collection_client.name_list:
|
|
|
|
db_name = item[0]
|
|
|
|
c_name = item[1]
|
|
|
|
payload = {
|
|
|
|
"collectionName": c_name,
|
|
|
|
"dbName": db_name
|
|
|
|
}
|
|
|
|
try:
|
|
|
|
self.collection_client.collection_drop(payload)
|
|
|
|
except Exception as e:
|
2024-12-07 06:42:45 +00:00
|
|
|
logger.error(f"drop collection error: {e}")
|
|
|
|
|
|
|
|
|
|
|
|
# Clean up databases created by this client
|
|
|
|
self.database_client.api_key = self.api_key
|
|
|
|
for db_name in self.database_client.db_names[:]: # Create a copy of the list to iterate
|
|
|
|
logger.info(f"database {db_name} exist, drop it")
|
|
|
|
try:
|
|
|
|
rsp = self.database_client.database_drop({"dbName": db_name})
|
|
|
|
except Exception as e:
|
|
|
|
logger.error(f"drop database error: {e}")
|
2024-03-02 03:03:03 +00:00
|
|
|
|
|
|
|
@pytest.fixture(scope="function", autouse=True)
|
2024-03-21 07:31:09 +00:00
|
|
|
def init_client(self, endpoint, token, minio_host, bucket_name, root_path):
|
2024-06-24 02:36:25 +00:00
|
|
|
_uuid = str(uuid.uuid1())
|
|
|
|
self.req = Requests()
|
|
|
|
self.req.update_uuid(_uuid)
|
2024-03-02 03:03:03 +00:00
|
|
|
self.endpoint = f"{endpoint}"
|
|
|
|
self.api_key = f"{token}"
|
|
|
|
self.invalid_api_key = "invalid_token"
|
|
|
|
self.vector_client = VectorClient(self.endpoint, self.api_key)
|
2024-06-24 02:36:25 +00:00
|
|
|
self.vector_client.update_uuid(_uuid)
|
2024-03-02 03:03:03 +00:00
|
|
|
self.collection_client = CollectionClient(self.endpoint, self.api_key)
|
2024-06-24 02:36:25 +00:00
|
|
|
self.collection_client.update_uuid(_uuid)
|
2024-03-02 03:03:03 +00:00
|
|
|
self.partition_client = PartitionClient(self.endpoint, self.api_key)
|
2024-06-24 02:36:25 +00:00
|
|
|
self.partition_client.update_uuid(_uuid)
|
2024-03-02 03:03:03 +00:00
|
|
|
self.index_client = IndexClient(self.endpoint, self.api_key)
|
2024-06-24 02:36:25 +00:00
|
|
|
self.index_client.update_uuid(_uuid)
|
2024-03-02 03:03:03 +00:00
|
|
|
self.alias_client = AliasClient(self.endpoint, self.api_key)
|
2024-06-24 02:36:25 +00:00
|
|
|
self.alias_client.update_uuid(_uuid)
|
2024-03-02 03:03:03 +00:00
|
|
|
self.user_client = UserClient(self.endpoint, self.api_key)
|
2024-06-24 02:36:25 +00:00
|
|
|
self.user_client.update_uuid(_uuid)
|
2024-03-02 03:03:03 +00:00
|
|
|
self.role_client = RoleClient(self.endpoint, self.api_key)
|
2024-06-24 02:36:25 +00:00
|
|
|
self.role_client.update_uuid(_uuid)
|
2024-03-02 03:03:03 +00:00
|
|
|
self.import_job_client = ImportJobClient(self.endpoint, self.api_key)
|
2024-06-24 02:36:25 +00:00
|
|
|
self.import_job_client.update_uuid(_uuid)
|
2024-03-21 07:31:09 +00:00
|
|
|
self.storage_client = StorageClient(f"{minio_host}:9000", "minioadmin", "minioadmin", bucket_name, root_path)
|
2024-12-07 06:42:45 +00:00
|
|
|
self.database_client = DatabaseClient(self.endpoint, self.api_key)
|
|
|
|
self.database_client.update_uuid(_uuid)
|
2024-03-02 03:03:03 +00:00
|
|
|
if token is None:
|
|
|
|
self.vector_client.api_key = None
|
|
|
|
self.collection_client.api_key = None
|
|
|
|
self.partition_client.api_key = None
|
|
|
|
connections.connect(uri=endpoint, token=token)
|
|
|
|
|
2024-03-21 07:31:09 +00:00
|
|
|
def init_collection(self, collection_name, pk_field="id", metric_type="L2", dim=128, nb=100, batch_size=1000, return_insert_id=False):
|
2024-03-02 03:03:03 +00:00
|
|
|
# create collection
|
|
|
|
schema_payload = {
|
|
|
|
"collectionName": collection_name,
|
|
|
|
"dimension": dim,
|
|
|
|
"metricType": metric_type,
|
|
|
|
"description": "test collection",
|
|
|
|
"primaryField": pk_field,
|
|
|
|
"vectorField": "vector",
|
|
|
|
}
|
|
|
|
rsp = self.collection_client.collection_create(schema_payload)
|
2024-05-27 06:25:41 +00:00
|
|
|
assert rsp['code'] == 0
|
2024-03-02 03:03:03 +00:00
|
|
|
self.wait_collection_load_completed(collection_name)
|
|
|
|
batch_size = batch_size
|
|
|
|
batch = nb // batch_size
|
|
|
|
remainder = nb % batch_size
|
2024-09-24 02:21:12 +00:00
|
|
|
|
|
|
|
full_data = []
|
2024-03-21 07:31:09 +00:00
|
|
|
insert_ids = []
|
2024-03-02 03:03:03 +00:00
|
|
|
for i in range(batch):
|
|
|
|
nb = batch_size
|
|
|
|
data = get_data_by_payload(schema_payload, nb)
|
|
|
|
payload = {
|
|
|
|
"collectionName": collection_name,
|
|
|
|
"data": data
|
|
|
|
}
|
|
|
|
body_size = sys.getsizeof(json.dumps(payload))
|
|
|
|
logger.debug(f"body size: {body_size / 1024 / 1024} MB")
|
|
|
|
rsp = self.vector_client.vector_insert(payload)
|
2024-05-27 06:25:41 +00:00
|
|
|
assert rsp['code'] == 0
|
2024-03-21 07:31:09 +00:00
|
|
|
if return_insert_id:
|
|
|
|
insert_ids.extend(rsp['data']['insertIds'])
|
2024-09-24 02:21:12 +00:00
|
|
|
full_data.extend(data)
|
2024-03-02 03:03:03 +00:00
|
|
|
# insert remainder data
|
|
|
|
if remainder:
|
|
|
|
nb = remainder
|
|
|
|
data = get_data_by_payload(schema_payload, nb)
|
|
|
|
payload = {
|
|
|
|
"collectionName": collection_name,
|
|
|
|
"data": data
|
|
|
|
}
|
|
|
|
rsp = self.vector_client.vector_insert(payload)
|
2024-05-27 06:25:41 +00:00
|
|
|
assert rsp['code'] == 0
|
2024-03-21 07:31:09 +00:00
|
|
|
if return_insert_id:
|
|
|
|
insert_ids.extend(rsp['data']['insertIds'])
|
2024-09-24 02:21:12 +00:00
|
|
|
full_data.extend(data)
|
2024-03-21 07:31:09 +00:00
|
|
|
if return_insert_id:
|
2024-09-24 02:21:12 +00:00
|
|
|
return schema_payload, full_data, insert_ids
|
2024-03-02 03:03:03 +00:00
|
|
|
|
2024-09-24 02:21:12 +00:00
|
|
|
return schema_payload, full_data
|
2024-03-02 03:03:03 +00:00
|
|
|
|
|
|
|
def wait_collection_load_completed(self, name):
|
|
|
|
t0 = time.time()
|
|
|
|
timeout = 60
|
|
|
|
while True and time.time() - t0 < timeout:
|
|
|
|
rsp = self.collection_client.collection_describe(name)
|
|
|
|
if "data" in rsp and "load" in rsp["data"] and rsp["data"]["load"] == "LoadStateLoaded":
|
|
|
|
break
|
|
|
|
else:
|
|
|
|
time.sleep(5)
|
|
|
|
|
|
|
|
def create_database(self, db_name="default"):
|
|
|
|
all_db = db.list_database()
|
|
|
|
logger.info(f"all database: {all_db}")
|
|
|
|
if db_name not in all_db:
|
|
|
|
logger.info(f"create database: {db_name}")
|
|
|
|
try:
|
|
|
|
db.create_database(db_name=db_name)
|
|
|
|
except Exception as e:
|
|
|
|
logger.error(e)
|
|
|
|
|
|
|
|
def update_database(self, db_name="default"):
|
|
|
|
self.create_database(db_name=db_name)
|
2024-03-21 07:31:09 +00:00
|
|
|
db.using_database(db_name=db_name)
|
2024-03-02 03:03:03 +00:00
|
|
|
self.collection_client.db_name = db_name
|
|
|
|
self.vector_client.db_name = db_name
|
2024-03-21 07:31:09 +00:00
|
|
|
self.import_job_client.db_name = db_name
|
2024-10-19 11:13:25 +00:00
|
|
|
|
2024-11-16 08:00:31 +00:00
|
|
|
def wait_load_completed(self, collection_name, db_name="default", timeout=5):
|
2024-11-05 00:42:23 +00:00
|
|
|
t0 = time.time()
|
|
|
|
while True and time.time() - t0 < timeout:
|
|
|
|
rsp = self.collection_client.collection_describe(collection_name, db_name=db_name)
|
|
|
|
if "data" in rsp and "load" in rsp["data"] and rsp["data"]["load"] == "LoadStateLoaded":
|
2024-11-16 08:00:31 +00:00
|
|
|
logger.info(f"collection {collection_name} load completed in {time.time() - t0} seconds")
|
2024-11-05 00:42:23 +00:00
|
|
|
break
|
|
|
|
else:
|
2024-11-16 08:00:31 +00:00
|
|
|
time.sleep(1)
|