enhance: Add integration test for stats task (#37506)

issue: #33744

Signed-off-by: Cai Zhang <cai.zhang@zilliz.com>
pull/37545/head
cai.zhang 2024-11-08 15:48:26 +08:00 committed by GitHub
parent 0f514708ff
commit ae227e3934
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
2 changed files with 474 additions and 0 deletions

View File

@ -0,0 +1,444 @@
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
package statstask
import (
"context"
"fmt"
"math"
"strconv"
"testing"
"time"
"github.com/stretchr/testify/suite"
"go.uber.org/zap"
"google.golang.org/protobuf/proto"
"github.com/milvus-io/milvus-proto/go-api/v2/commonpb"
"github.com/milvus-io/milvus-proto/go-api/v2/milvuspb"
"github.com/milvus-io/milvus-proto/go-api/v2/schemapb"
"github.com/milvus-io/milvus/internal/metastore/kv/binlog"
"github.com/milvus-io/milvus/internal/proto/datapb"
"github.com/milvus-io/milvus/internal/storage"
"github.com/milvus-io/milvus/internal/util/hookutil"
"github.com/milvus-io/milvus/pkg/common"
"github.com/milvus-io/milvus/pkg/log"
"github.com/milvus-io/milvus/pkg/util/funcutil"
"github.com/milvus-io/milvus/pkg/util/merr"
"github.com/milvus-io/milvus/pkg/util/metric"
"github.com/milvus-io/milvus/pkg/util/paramtable"
"github.com/milvus-io/milvus/tests/integration"
)
type StatsTaskCheckerSuite struct {
integration.MiniClusterSuite
pkType schemapb.DataType
dbName string
dim int
batch int
batchCnt int
indexType string
metricType string
cm storage.ChunkManager
}
func TestStatsTask(t *testing.T) {
suite.Run(t, new(StatsTaskCheckerSuite))
}
func (s *StatsTaskCheckerSuite) initParams() {
s.dbName = "default"
s.dim = 128
s.batch = 2000
s.batchCnt = 5
s.indexType = integration.IndexFaissIvfFlat
s.metricType = metric.L2
cm, err := storage.NewChunkManagerFactoryWithParam(paramtable.Get()).NewPersistentStorageChunkManager(context.Background())
s.NoError(err)
s.cm = cm
}
func (s *StatsTaskCheckerSuite) TestStatsTaskChecker_Int64PK() {
s.initParams()
s.pkType = schemapb.DataType_Int64
s.run()
}
func (s *StatsTaskCheckerSuite) TestStatsTaskChecker_VarcharPK() {
s.initParams()
s.pkType = schemapb.DataType_VarChar
s.run()
}
func (s *StatsTaskCheckerSuite) run() {
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
c := s.Cluster
collectionName := "TestStatsTask" + funcutil.GenRandomStr()
var pkField *schemapb.FieldSchema
if s.pkType == schemapb.DataType_VarChar {
pkField = &schemapb.FieldSchema{
FieldID: 100,
Name: "pk",
IsPrimaryKey: true,
Description: "primary key",
DataType: schemapb.DataType_VarChar,
TypeParams: []*commonpb.KeyValuePair{{Key: "max_length", Value: "10000"}},
AutoID: false,
}
} else {
pkField = &schemapb.FieldSchema{
FieldID: 100,
Name: "pk",
IsPrimaryKey: true,
Description: "primary key",
DataType: schemapb.DataType_Int64,
TypeParams: []*commonpb.KeyValuePair{},
AutoID: false,
}
}
varcharField := &schemapb.FieldSchema{
FieldID: 101,
Name: "var",
IsPrimaryKey: false,
Description: "test enable match",
DataType: schemapb.DataType_VarChar,
TypeParams: []*commonpb.KeyValuePair{
{Key: "max_length", Value: "10000"},
},
}
vectorField := &schemapb.FieldSchema{
FieldID: 102,
Name: integration.FloatVecField,
DataType: schemapb.DataType_FloatVector,
TypeParams: []*commonpb.KeyValuePair{
{Key: "dim", Value: strconv.Itoa(s.dim)},
},
}
schema := integration.ConstructSchema(collectionName, s.dim, false, pkField, varcharField, vectorField)
marshaledSchema, err := proto.Marshal(schema)
s.NoError(err)
createCollectionStatus, err := c.Proxy.CreateCollection(ctx, &milvuspb.CreateCollectionRequest{
DbName: s.dbName,
CollectionName: collectionName,
Schema: marshaledSchema,
ShardsNum: common.DefaultShardsNum,
})
s.NoError(err)
if createCollectionStatus.GetErrorCode() != commonpb.ErrorCode_Success {
log.Warn("createCollectionStatus fail reason", zap.String("reason", createCollectionStatus.GetReason()))
}
s.Equal(createCollectionStatus.GetErrorCode(), commonpb.ErrorCode_Success)
log.Info("CreateCollection result", zap.Any("createCollectionStatus", createCollectionStatus))
showCollectionsResp, err := c.Proxy.ShowCollections(ctx, &milvuspb.ShowCollectionsRequest{})
s.NoError(err)
s.Equal(showCollectionsResp.GetStatus().GetErrorCode(), commonpb.ErrorCode_Success)
log.Info("ShowCollections result", zap.Any("showCollectionsResp", showCollectionsResp))
insertCheckReport := func() {
timeoutCtx, cancelFunc := context.WithTimeout(ctx, 5*time.Second)
defer cancelFunc()
for {
select {
case <-timeoutCtx.Done():
s.Fail("insert check timeout")
case report := <-c.Extension.GetReportChan():
reportInfo := report.(map[string]any)
log.Info("insert report info", zap.Any("reportInfo", reportInfo))
s.Equal(hookutil.OpTypeInsert, reportInfo[hookutil.OpTypeKey])
s.NotEqualValues(0, reportInfo[hookutil.RequestDataSizeKey])
return
}
}
}
go insertCheckReport()
// batch insert to generate some segments
for i := 0; i < s.batchCnt; i++ {
var pkColumn *schemapb.FieldData
if s.pkType == schemapb.DataType_VarChar {
stringData := make([]string, s.batch)
for j := 0; j < s.batch; j++ {
stringData[j] = fmt.Sprintf("%d", s.batch*(s.batchCnt-i)-j)
}
pkColumn = &schemapb.FieldData{
Type: schemapb.DataType_VarChar,
FieldName: "pk",
Field: &schemapb.FieldData_Scalars{
Scalars: &schemapb.ScalarField{
Data: &schemapb.ScalarField_StringData{
StringData: &schemapb.StringArray{
Data: stringData,
},
},
},
},
FieldId: 100,
}
} else {
intData := make([]int64, s.batch)
for j := 0; j < s.batch; j++ {
intData[j] = int64(s.batch*(s.batchCnt-i) - j)
}
pkColumn = &schemapb.FieldData{
Type: schemapb.DataType_VarChar,
FieldName: "pk",
Field: &schemapb.FieldData_Scalars{
Scalars: &schemapb.ScalarField{
Data: &schemapb.ScalarField_LongData{
LongData: &schemapb.LongArray{
Data: intData,
},
},
},
},
FieldId: 100,
}
}
stringData := make([]string, s.batch)
for j := 0; j < s.batch; j++ {
stringData[j] = fmt.Sprintf("hello milvus with %d", s.batch*(s.batchCnt-i)-j)
}
varcharColumn := &schemapb.FieldData{
Type: schemapb.DataType_VarChar,
FieldName: "var",
Field: &schemapb.FieldData_Scalars{
Scalars: &schemapb.ScalarField{
Data: &schemapb.ScalarField_StringData{
StringData: &schemapb.StringArray{
Data: stringData,
},
},
},
},
FieldId: 100,
}
fVecColumn := integration.NewFloatVectorFieldData(integration.FloatVecField, s.batch, s.dim)
hashKeys := integration.GenerateHashKeys(s.batch)
insertResult, err := c.Proxy.Insert(ctx, &milvuspb.InsertRequest{
DbName: s.dbName,
CollectionName: collectionName,
FieldsData: []*schemapb.FieldData{pkColumn, varcharColumn, fVecColumn},
HashKeys: hashKeys,
NumRows: uint32(s.batch),
})
s.NoError(err)
s.Equal(insertResult.GetStatus().GetErrorCode(), commonpb.ErrorCode_Success)
// flush
flushResp, err := c.Proxy.Flush(ctx, &milvuspb.FlushRequest{
DbName: s.dbName,
CollectionNames: []string{collectionName},
})
s.NoError(err)
segmentIDs, has := flushResp.GetCollSegIDs()[collectionName]
ids := segmentIDs.GetData()
s.Require().NotEmpty(segmentIDs)
s.Require().True(has)
flushTs, has := flushResp.GetCollFlushTs()[collectionName]
s.True(has)
segments, err := c.MetaWatcher.ShowSegments()
s.NoError(err)
s.NotEmpty(segments)
for _, segment := range segments {
log.Info("ShowSegments result", zap.String("segment", segment.String()))
}
s.WaitForFlush(ctx, ids, flushTs, s.dbName, collectionName)
}
// create index
createIndexStatus, err := c.Proxy.CreateIndex(ctx, &milvuspb.CreateIndexRequest{
CollectionName: collectionName,
FieldName: integration.FloatVecField,
IndexName: "_default",
ExtraParams: integration.ConstructIndexParam(s.dim, s.indexType, s.metricType),
})
if createIndexStatus.GetErrorCode() != commonpb.ErrorCode_Success {
log.Warn("createIndexStatus fail reason", zap.String("reason", createIndexStatus.GetReason()))
}
s.NoError(err)
s.Equal(commonpb.ErrorCode_Success, createIndexStatus.GetErrorCode())
s.WaitForIndexBuilt(ctx, collectionName, integration.FloatVecField)
// load
loadStatus, err := c.Proxy.LoadCollection(ctx, &milvuspb.LoadCollectionRequest{
DbName: s.dbName,
CollectionName: collectionName,
})
s.NoError(err)
if loadStatus.GetErrorCode() != commonpb.ErrorCode_Success {
log.Warn("loadStatus fail reason", zap.String("reason", loadStatus.GetReason()))
}
s.Equal(commonpb.ErrorCode_Success, loadStatus.GetErrorCode())
s.WaitForLoad(ctx, collectionName)
s.WaitForSortedSegmentLoaded(ctx, s.dbName, collectionName)
segments, err := c.MetaWatcher.ShowSegments()
s.NoError(err)
s.NotEmpty(segments)
for _, segment := range segments {
if segment.GetIsSorted() && segment.GetState() != commonpb.SegmentState_Dropped {
s.checkSegmentIsSorted(ctx, segment)
}
}
// search
nq := 10
topk := 10
roundDecimal := -1
params := integration.GetSearchParams(s.indexType, s.metricType)
searchReq := integration.ConstructSearchRequest("", collectionName, "",
integration.FloatVecField, schemapb.DataType_FloatVector, nil, s.metricType, params, nq, s.dim, topk, roundDecimal)
searchCheckReport := func() {
timeoutCtx, cancelFunc := context.WithTimeout(ctx, 5*time.Second)
defer cancelFunc()
for {
select {
case <-timeoutCtx.Done():
s.Fail("search check timeout")
case report := <-c.Extension.GetReportChan():
reportInfo := report.(map[string]any)
log.Info("search report info", zap.Any("reportInfo", reportInfo))
s.Equal(hookutil.OpTypeSearch, reportInfo[hookutil.OpTypeKey])
s.NotEqualValues(0, reportInfo[hookutil.ResultDataSizeKey])
s.NotEqualValues(0, reportInfo[hookutil.RelatedDataSizeKey])
s.EqualValues(s.batch*s.batchCnt, reportInfo[hookutil.RelatedCntKey])
return
}
}
}
go searchCheckReport()
searchResult, err := c.Proxy.Search(ctx, searchReq)
err = merr.CheckRPCCall(searchResult, err)
s.NoError(err)
queryCheckReport := func() {
timeoutCtx, cancelFunc := context.WithTimeout(ctx, 5*time.Second)
defer cancelFunc()
for {
select {
case <-timeoutCtx.Done():
s.Fail("query check timeout")
return
case report := <-c.Extension.GetReportChan():
reportInfo := report.(map[string]any)
log.Info("query report info", zap.Any("reportInfo", reportInfo))
s.Equal(hookutil.OpTypeQuery, reportInfo[hookutil.OpTypeKey])
s.NotEqualValues(0, reportInfo[hookutil.ResultDataSizeKey])
s.NotEqualValues(0, reportInfo[hookutil.RelatedDataSizeKey])
s.EqualValues(s.batch*s.batchCnt, reportInfo[hookutil.RelatedCntKey])
return
}
}
}
go queryCheckReport()
queryResult, err := c.Proxy.Query(ctx, &milvuspb.QueryRequest{
DbName: s.dbName,
CollectionName: collectionName,
Expr: "",
OutputFields: []string{"count(*)"},
})
if queryResult.GetStatus().GetErrorCode() != commonpb.ErrorCode_Success {
log.Warn("searchResult fail reason", zap.String("reason", queryResult.GetStatus().GetReason()))
}
s.NoError(err)
s.Equal(commonpb.ErrorCode_Success, queryResult.GetStatus().GetErrorCode())
status, err := c.Proxy.ReleaseCollection(ctx, &milvuspb.ReleaseCollectionRequest{
CollectionName: collectionName,
})
err = merr.CheckRPCCall(status, err)
s.NoError(err)
status, err = c.Proxy.DropCollection(ctx, &milvuspb.DropCollectionRequest{
CollectionName: collectionName,
})
err = merr.CheckRPCCall(status, err)
s.NoError(err)
log.Info("TestStatsTask succeed")
}
func (s *StatsTaskCheckerSuite) checkBinlogIsSorted(ctx context.Context, binlogPath string, lastValue interface{}) interface{} {
bs, err := s.cm.Read(ctx, binlogPath)
s.NoError(err)
reader, err := storage.NewBinlogReader(bs)
s.NoError(err)
defer reader.Close()
er, err := reader.NextEventReader()
s.NoError(err)
if s.pkType == schemapb.DataType_VarChar {
pks, _, err := er.GetStringFromPayload()
s.NoError(err)
for _, pk := range pks {
s.GreaterOrEqual(pk, lastValue)
lastValue = pk
}
return lastValue
}
pks, _, err := er.GetInt64FromPayload()
s.NoError(err)
for _, pk := range pks {
s.GreaterOrEqual(pk, lastValue)
lastValue = pk
}
return lastValue
}
func (s *StatsTaskCheckerSuite) checkSegmentIsSorted(ctx context.Context, segment *datapb.SegmentInfo) {
err := binlog.DecompressBinLogs(segment)
s.NoError(err)
var pkBinlogs *datapb.FieldBinlog
for _, fb := range segment.Binlogs {
if fb.FieldID == 100 {
pkBinlogs = fb
break
}
}
entitiesNum := int64(0)
if s.pkType == schemapb.DataType_VarChar {
lastValue := ""
for _, b := range pkBinlogs.Binlogs {
lastValue = s.checkBinlogIsSorted(ctx, b.GetLogPath(), lastValue).(string)
entitiesNum += b.GetEntriesNum()
}
} else {
lastValue := int64(math.MinInt64)
for _, b := range pkBinlogs.Binlogs {
lastValue = s.checkBinlogIsSorted(ctx, b.GetLogPath(), lastValue).(int64)
entitiesNum += b.GetEntriesNum()
}
}
s.Equal(segment.GetNumOfRows(), entitiesNum)
}

View File

@ -55,6 +55,36 @@ func (s *MiniClusterSuite) WaitForLoad(ctx context.Context, collection string) {
s.waitForLoadInternal(ctx, "", collection)
}
func (s *MiniClusterSuite) WaitForSortedSegmentLoaded(ctx context.Context, dbName, collection string) {
cluster := s.Cluster
getSegmentsSorted := func() bool {
querySegmentInfo, err := cluster.Proxy.GetQuerySegmentInfo(ctx, &milvuspb.GetQuerySegmentInfoRequest{
DbName: dbName,
CollectionName: collection,
})
if err != nil {
panic("GetQuerySegmentInfo fail")
}
for _, info := range querySegmentInfo.GetInfos() {
if !info.GetIsSorted() {
return false
}
}
return true
}
for !getSegmentsSorted() {
select {
case <-ctx.Done():
s.FailNow("failed to wait for get segments sorted")
return
default:
time.Sleep(500 * time.Millisecond)
}
}
}
func (s *MiniClusterSuite) waitForLoadInternal(ctx context.Context, dbName, collection string) {
cluster := s.Cluster
getLoadingProgress := func() *milvuspb.GetLoadingProgressResponse {