mirror of https://github.com/milvus-io/milvus.git
fix: enable milvus.yaml check (#34567)
See #32168 --------- Signed-off-by: Ted Xu <ted.xu@zilliz.com>pull/35001/head
parent
95c4600958
commit
63f0154dfb
2
Makefile
2
Makefile
|
@ -381,7 +381,7 @@ clean:
|
|||
|
||||
milvus-tools: print-build-info
|
||||
@echo "Building tools ..."
|
||||
@mkdir -p $(INSTALL_PATH)/tools && go env -w CGO_ENABLED="1" && GO111MODULE=on $(GO) build \
|
||||
@. $(PWD)/scripts/setenv.sh && mkdir -p $(INSTALL_PATH)/tools && go env -w CGO_ENABLED="1" && GO111MODULE=on $(GO) build \
|
||||
-pgo=$(PGO_PATH)/default.pgo -ldflags="-X 'main.BuildTags=$(BUILD_TAGS)' -X 'main.BuildTime=$(BUILD_TIME)' -X 'main.GitCommit=$(GIT_COMMIT)' -X 'main.GoVersion=$(GO_VERSION)'" \
|
||||
-o $(INSTALL_PATH)/tools $(PWD)/cmd/tools/* 1>/dev/null
|
||||
|
||||
|
|
|
@ -16,6 +16,7 @@ import (
|
|||
"bytes"
|
||||
"fmt"
|
||||
"os"
|
||||
"strings"
|
||||
"testing"
|
||||
|
||||
"github.com/stretchr/testify/assert"
|
||||
|
@ -30,20 +31,26 @@ import (
|
|||
// Please be noted that milvus.yaml is generated by code, so don't edit it directly, instead, change the code in paramtable
|
||||
// and run `make milvus-tools && ./bin/tools/config gen-yaml && mv milvus.yaml configs/milvus.yaml`.
|
||||
func TestYamlFile(t *testing.T) {
|
||||
log.SetLevel(zap.InfoLevel)
|
||||
w := bytes.Buffer{}
|
||||
WriteYaml(&w)
|
||||
|
||||
base := paramtable.NewBaseTable()
|
||||
f, err := os.Open(fmt.Sprintf("%s/%s", base.GetConfigDir(), "milvus.yaml"))
|
||||
assert.NoError(t, err, "expecting configs/milvus.yaml")
|
||||
log.Info("Verifying config", zap.String("file", f.Name()))
|
||||
defer f.Close()
|
||||
fileScanner := bufio.NewScanner(f)
|
||||
codeScanner := bufio.NewScanner(&w)
|
||||
|
||||
for fileScanner.Scan() && codeScanner.Scan() {
|
||||
if strings.Contains(codeScanner.Text(), "etcd:") || strings.Contains(codeScanner.Text(), "minio:") || strings.Contains(codeScanner.Text(), "pulsar:") {
|
||||
// Skip check of endpoints given by .env
|
||||
continue
|
||||
}
|
||||
if fileScanner.Text() != codeScanner.Text() {
|
||||
assert.FailNow(t, fmt.Sprintf("configs/milvus.yaml is not consistent with paramtable, file: [%s], code: [%s]. Do not edit milvus.yaml directly.",
|
||||
fileScanner.Text(), codeScanner.Text()))
|
||||
}
|
||||
log.Error("", zap.Any("file", fileScanner.Text()), zap.Any("code", codeScanner.Text()))
|
||||
}
|
||||
}
|
||||
|
|
|
@ -226,8 +226,6 @@ proxy:
|
|||
localPath: /tmp/milvus_access
|
||||
filename: # Log filename, leave empty to use stdout.
|
||||
maxSize: 64 # Max size for a single file, in MB.
|
||||
cacheSize: 0 # Size of log write cache, in B
|
||||
cacheFlushInterval: 3 # time interval of auto flush write cache, in Seconds. (Close auto flush if interval was 0)
|
||||
rotatedTime: 0 # Max time for single access log file in seconds
|
||||
remotePath: access_log/ # File path in minIO
|
||||
remoteMaxTime: 0 # Max time for log file in minIO, in hours
|
||||
|
@ -237,6 +235,8 @@ proxy:
|
|||
query:
|
||||
format: "[$time_now] [ACCESS] <$user_name: $user_addr> $method_name [status: $method_status] [code: $error_code] [sdk: $sdk_version] [msg: $error_msg] [traceID: $trace_id] [timeCost: $time_cost] [database: $database_name] [collection: $collection_name] [partitions: $partition_name] [expr: $method_expr]"
|
||||
methods: "Query,Search,Delete"
|
||||
cacheSize: 0 # Size of log of write cache, in byte. (Close write cache if size was 0)
|
||||
cacheFlushInterval: 3 # time interval of auto flush write cache, in seconds. (Close auto flush if interval was 0)
|
||||
connectionCheckIntervalSeconds: 120 # the interval time(in seconds) for connection manager to scan inactive client info
|
||||
connectionClientInfoTTLSeconds: 86400 # inactive client info TTL duration, in seconds
|
||||
maxConnectionNum: 10000 # the max client info numbers that proxy should manage, avoid too many client infos
|
||||
|
@ -278,6 +278,7 @@ queryCoord:
|
|||
rowCountMaxSteps: 50 # segment count based plan generator max steps
|
||||
randomMaxSteps: 10 # segment count based plan generator max steps
|
||||
growingRowCountWeight: 4 # the memory weight of growing segment row count
|
||||
delegatorMemoryOverloadFactor: 0.3 # the factor of delegator overloaded memory
|
||||
balanceCostThreshold: 0.001 # the threshold of balance cost, if the difference of cluster's cost after executing the balance plan is less than this value, the plan will not be executed
|
||||
checkSegmentInterval: 1000
|
||||
checkChannelInterval: 1000
|
||||
|
@ -286,8 +287,6 @@ queryCoord:
|
|||
channelTaskTimeout: 60000 # 1 minute
|
||||
segmentTaskTimeout: 120000 # 2 minute
|
||||
distPullInterval: 500
|
||||
collectionObserverInterval: 200
|
||||
checkExecutedFlagInterval: 100
|
||||
heartbeatAvailableInterval: 10000 # 10s, Only QueryNodes which fetched heartbeats within the duration are available
|
||||
loadTimeoutSeconds: 600
|
||||
distRequestTimeout: 5000 # the request timeout for querycoord fetching data distribution from querynodes, in milliseconds
|
||||
|
@ -305,6 +304,8 @@ queryCoord:
|
|||
gracefulStopTimeout: 5 # seconds. force stop node without graceful stop
|
||||
enableStoppingBalance: true # whether enable stopping balance
|
||||
channelExclusiveNodeFactor: 4 # the least node number for enable channel's exclusive mode
|
||||
collectionObserverInterval: 200 # the interval of collection observer
|
||||
checkExecutedFlagInterval: 100 # the interval of check executed flag to force to pull dist
|
||||
cleanExcludeSegmentInterval: 60 # the time duration of clean pipeline exclude segment which used for filter invalid data, in seconds
|
||||
ip: # if not specified, use the first unicastable address
|
||||
port: 19531
|
||||
|
@ -344,9 +345,9 @@ queryNode:
|
|||
warmup: disable
|
||||
mmap:
|
||||
mmapEnabled: false # Enable mmap for loading data
|
||||
growingMmapEnabled: false # Enable mmap for growing segment
|
||||
fixedFileSizeForMmapAlloc: 4 #MB, fixed file size for mmap chunk manager to store chunk data
|
||||
maxDiskUsagePercentageForMmapAlloc: 20 # max percentage of disk usage in memory mapping
|
||||
growingMmapEnabled: false # Enable mmap for using in growing raw data
|
||||
fixedFileSizeForMmapAlloc: 4 # tmp file size for mmap chunk manager
|
||||
maxDiskUsagePercentageForMmapAlloc: 20 # disk percentage used in mmap chunk manager
|
||||
lazyload:
|
||||
enabled: false # Enable lazyload for loading data
|
||||
waitTimeout: 30000 # max wait timeout duration in milliseconds before start to do lazyload search and retrieve
|
||||
|
@ -385,8 +386,9 @@ queryNode:
|
|||
flowGraph:
|
||||
maxQueueLength: 16 # Maximum length of task queue in flowgraph
|
||||
maxParallelism: 1024 # Maximum number of tasks executed in parallel in the flowgraph
|
||||
enableSegmentPrune: false # use partition prune function on shard delegator
|
||||
enableSegmentPrune: false # use partition stats to prune data in search/query on shard delegator
|
||||
queryStreamBatchSize: 4194304 # return batch size of stream query
|
||||
bloomFilterApplyParallelFactor: 4 # parallel factor when to apply pk to bloom filter, default to 4*CPU_CORE_NUM
|
||||
ip: # if not specified, use the first unicastable address
|
||||
port: 21123
|
||||
grpc:
|
||||
|
@ -429,9 +431,7 @@ dataCoord:
|
|||
maxSize: 1024 # Maximum size of a segment in MB
|
||||
diskSegmentMaxSize: 2048 # Maximun size of a segment in MB for collection which has Disk index
|
||||
sealProportion: 0.12
|
||||
# segment seal proportion jitter ratio, default value 0.1(10%),
|
||||
# if seal propertion is 12%, with jitter=0.1, the actuall applied ratio will be 10.8~12%
|
||||
sealProportionJitter: 0.1 #
|
||||
sealProportionJitter: 0.1 # segment seal proportion jitter ratio, default value 0.1(10%), if seal propertion is 12%, with jitter=0.1, the actuall applied ratio will be 10.8~12%
|
||||
assignmentExpiration: 2000 # The time of the assignment expiration in ms
|
||||
allocLatestExpireAttempt: 200 # The time attempting to alloc latest lastExpire from rootCoord after restart
|
||||
maxLife: 86400 # The max lifetime of segment in seconds, 24*60*60
|
||||
|
@ -451,13 +451,13 @@ dataCoord:
|
|||
# MUST BE GREATER THAN OR EQUAL TO <smallProportion>!!!
|
||||
# During compaction, the size of segment # of rows is able to exceed segment max # of rows by (expansionRate-1) * 100%.
|
||||
expansionRate: 1.25
|
||||
segmentFlushInterval: 2 # the minimal interval duration(unit: Seconds) between flusing operation on same segment
|
||||
sealPolicy:
|
||||
channel:
|
||||
# The size threshold in MB, if the total size of growing segments of each shard
|
||||
# The size threshold in MB, if the total size of growing segments of each shard
|
||||
# exceeds this threshold, the largest growing segment will be sealed.
|
||||
growingSegmentsMemSize: 4096
|
||||
autoUpgradeSegmentIndex: false # whether auto upgrade segment index to index engine's version
|
||||
segmentFlushInterval: 2 # the minimal interval duration(unit: Seconds) between flusing operation on same segment
|
||||
enableCompaction: true # Enable data segment compaction
|
||||
compaction:
|
||||
enableAutoCompaction: true
|
||||
|
@ -467,35 +467,29 @@ dataCoord:
|
|||
workerMaxParallelTaskNum: 2
|
||||
clustering:
|
||||
enable: true # Enable clustering compaction
|
||||
autoEnable: false # Enable auto background clustering compaction
|
||||
autoEnable: false # Enable auto clustering compaction
|
||||
triggerInterval: 600 # clustering compaction trigger interval in seconds
|
||||
stateCheckInterval: 10
|
||||
gcInterval: 600
|
||||
minInterval: 3600 # The minimum interval between clustering compaction executions of one collection, to avoid redundant compaction
|
||||
maxInterval: 259200 # If a collection haven't been clustering compacted for longer than maxInterval, force compact
|
||||
newDataSizeThreshold: 512m # If new data size is large than newDataSizeThreshold, execute clustering compaction
|
||||
timeout: 7200 # timeout in seconds for clustering compaction, the task will stop if timeout
|
||||
dropTolerance: 86400
|
||||
# clustering compaction will try best to distribute data into segments with size range in [preferSegmentSize, maxSegmentSize].
|
||||
# data will be clustered by preferSegmentSize, if a cluster is larger than maxSegmentSize, will spilt it into multi segment
|
||||
# buffer between (preferSegmentSize, maxSegmentSize) is left for new data in the same cluster(range), to avoid globally redistribute too often
|
||||
preferSegmentSize: 512m
|
||||
dropTolerance: 86400 # If clustering compaction job is finished for a long time, gc it
|
||||
preferSegmentSize: 512m
|
||||
maxSegmentSize: 1024m
|
||||
|
||||
# vector clustering related
|
||||
maxTrainSizeRatio: 0.8 # max data size ratio in Kmeans train, if larger than it, will down sampling to meet this limit
|
||||
maxCentroidsNum: 10240 # maximum centroids number in Kmeans train
|
||||
minCentroidsNum: 16 # minimum centroids number in Kmeans train
|
||||
minClusterSizeRatio: 0.01 # minimum cluster size / avg size in Kmeans train
|
||||
maxClusterSizeRatio: 10 #maximum cluster size / avg size in Kmeans train
|
||||
maxClusterSizeRatio: 10 # maximum cluster size / avg size in Kmeans train
|
||||
maxClusterSize: 5g # maximum cluster size in Kmeans train
|
||||
|
||||
levelzero:
|
||||
forceTrigger:
|
||||
minSize: 8388608 # The minmum size in bytes to force trigger a LevelZero Compaction, default as 8MB
|
||||
maxSize: 67108864 # The maxmum size in bytes to force trigger a LevelZero Compaction, default as 64MB
|
||||
deltalogMinNum: 10 # The minimum number of deltalog files to force trigger a LevelZero Compaction
|
||||
deltalogMaxNum: 30 # The maxmum number of deltalog files to force trigger a LevelZero Compaction, default as 30
|
||||
syncSegmentsInterval: 300 # The time interval for regularly syncing segments
|
||||
enableGarbageCollection: true
|
||||
gc:
|
||||
interval: 3600 # meta-based gc scanning interval in seconds
|
||||
|
@ -517,6 +511,10 @@ dataCoord:
|
|||
maxImportFileNumPerReq: 1024 # The maximum number of files allowed per single import request.
|
||||
waitForIndex: true # Indicates whether the import operation waits for the completion of index building.
|
||||
gracefulStopTimeout: 5 # seconds. force stop node without graceful stop
|
||||
slot:
|
||||
clusteringCompactionUsage: 16 # slot usage of clustering compaction job.
|
||||
mixCompactionUsage: 8 # slot usage of mix compaction job.
|
||||
l0DeleteCompactionUsage: 8 # slot usage of l0 compaction job.
|
||||
ip: # if not specified, use the first unicastable address
|
||||
port: 13333
|
||||
grpc:
|
||||
|
@ -524,11 +522,6 @@ dataCoord:
|
|||
serverMaxRecvSize: 268435456
|
||||
clientMaxSendSize: 268435456
|
||||
clientMaxRecvSize: 536870912
|
||||
syncSegmentsInterval: 300
|
||||
slot:
|
||||
clusteringCompactionUsage: 16
|
||||
mixCompactionUsage: 8
|
||||
l0DeleteCompactionUsage: 8
|
||||
|
||||
dataNode:
|
||||
dataSync:
|
||||
|
@ -556,6 +549,8 @@ dataNode:
|
|||
# if this parameter <= 0, will set it as the maximum number of CPUs that can be executing
|
||||
# suggest to set it bigger on large collection numbers to avoid blocking
|
||||
workPoolSize: -1
|
||||
# specify the size of global work pool for channel checkpoint updating
|
||||
# if this parameter <= 0, will set it as 10
|
||||
updateChannelCheckpointMaxParallel: 10
|
||||
updateChannelCheckpointInterval: 60 # the interval duration(in seconds) for datanode to update channel checkpoint of each channel
|
||||
updateChannelCheckpointRPCTimeout: 20 # timeout in seconds for UpdateChannelCheckpoint RPC call
|
||||
|
@ -569,6 +564,11 @@ dataNode:
|
|||
levelZeroBatchMemoryRatio: 0.05 # The minimal memory ratio of free memory for level zero compaction executing in batch mode
|
||||
levelZeroMaxBatchSize: -1 # Max batch size refers to the max number of L1/L2 segments in a batch when executing L0 compaction. Default to -1, any value that is less than 1 means no limit. Valid range: >= 1.
|
||||
gracefulStopTimeout: 1800 # seconds. force stop node without graceful stop
|
||||
slot:
|
||||
slotCap: 16 # The maximum number of tasks(e.g. compaction, importing) allowed to run concurrently on a datanode
|
||||
clusteringCompaction:
|
||||
memoryBufferRatio: 0.1 # The ratio of memory buffer of clustering compaction. Data larger than threshold will be flushed to storage.
|
||||
workPoolSize: 8 # worker pool size for one clustering compaction job.
|
||||
ip: # if not specified, use the first unicastable address
|
||||
port: 21124
|
||||
grpc:
|
||||
|
@ -576,23 +576,6 @@ dataNode:
|
|||
serverMaxRecvSize: 268435456
|
||||
clientMaxSendSize: 268435456
|
||||
clientMaxRecvSize: 536870912
|
||||
slot:
|
||||
slotCap: 16 # The maximum number of tasks(e.g. compaction, importing) allowed to run concurrently on a datanode.
|
||||
|
||||
clusteringCompaction:
|
||||
memoryBufferRatio: 0.1 # The ratio of memory buffer of clustering compaction. Data larger than threshold will be flushed to storage.
|
||||
workPoolSize: 8 # worker pool size for one clustering compaction job
|
||||
|
||||
streamingNode:
|
||||
# can specify ip for example
|
||||
# ip: 127.0.0.1
|
||||
ip: # if not specify address, will use the first unicastable address as local ip
|
||||
port: 19532
|
||||
grpc:
|
||||
serverMaxSendSize: 536870912
|
||||
serverMaxRecvSize: 536870912
|
||||
clientMaxSendSize: 268435456
|
||||
clientMaxRecvSize: 268435456
|
||||
|
||||
# Configures the system log output.
|
||||
log:
|
||||
|
@ -617,6 +600,7 @@ grpc:
|
|||
maxMaxAttempts: 10
|
||||
initialBackoff: 0.2
|
||||
maxBackoff: 10
|
||||
backoffMultiplier: 2
|
||||
minResetInterval: 1000
|
||||
maxCancelError: 32
|
||||
minSessionCheckInterval: 200
|
||||
|
@ -647,6 +631,7 @@ common:
|
|||
BeamWidthRatio: 4
|
||||
gracefulTime: 5000 # milliseconds. it represents the interval (in ms) by which the request arrival time needs to be subtracted in the case of Bounded Consistency.
|
||||
gracefulStopTimeout: 1800 # seconds. it will force quit the server if the graceful stop process is not completed during this time.
|
||||
bitmapIndexCardinalityBound: 500
|
||||
storageType: remote # please adjust in embedded Milvus: local, available values are [local, remote, opendal], value minio is deprecated, use remote instead
|
||||
# Default value: auto
|
||||
# Valid values: [auto, avx512, avx2, avx, sse4_2]
|
||||
|
@ -657,8 +642,8 @@ common:
|
|||
# The superusers will ignore some system check processes,
|
||||
# like the old password verification when updating the credential
|
||||
superUsers:
|
||||
defaultRootPassword: Milvus # default password for root user
|
||||
tlsMode: 0
|
||||
defaultRootPassword: Milvus
|
||||
session:
|
||||
ttl: 30 # ttl value when session granting a lease to register service
|
||||
retryTimes: 30 # retry times when session sending etcd requests
|
||||
|
@ -674,11 +659,12 @@ common:
|
|||
ttMsgEnabled: true # Whether the instance disable sending ts messages
|
||||
traceLogMode: 0 # trace request info
|
||||
bloomFilterSize: 100000 # bloom filter initial size
|
||||
bloomFilterType: BlockedBloomFilter # bloom filter type, support BasicBloomFilter and BlockedBloomFilter
|
||||
maxBloomFalsePositive: 0.001 # max false positive rate for bloom filter
|
||||
# clustering key/compaction related
|
||||
usePartitionKeyAsClusteringKey: false
|
||||
useVectorAsClusteringKey: false
|
||||
enableVectorClusteringKey: false
|
||||
bloomFilterApplyBatchSize: 1000 # batch size when to apply pk to bloom filter
|
||||
usePartitionKeyAsClusteringKey: false # if true, do clustering compaction and segment prune on partition key field
|
||||
useVectorAsClusteringKey: false # if true, do clustering compaction and segment prune on vector field
|
||||
enableVectorClusteringKey: false # if true, enable vector clustering key and vector clustering compaction
|
||||
|
||||
# QuotaConfig, configurations of Milvus quota and limits.
|
||||
# By default, we enable:
|
||||
|
@ -819,6 +805,10 @@ quotaAndLimits:
|
|||
diskQuotaPerDB: -1 # MB, (0, +inf), default no limit
|
||||
diskQuotaPerCollection: -1 # MB, (0, +inf), default no limit
|
||||
diskQuotaPerPartition: -1 # MB, (0, +inf), default no limit
|
||||
l0SegmentsRowCountProtection:
|
||||
enabled: false # switch to enable l0 segment row count quota
|
||||
lowWaterLevel: 32768 # l0 segment row count quota, low water level
|
||||
highWaterLevel: 65536 # l0 segment row count quota, low water level
|
||||
limitReading:
|
||||
# forceDeny false means dql requests are allowed (except for some
|
||||
# specific conditions, such as collection has been dropped), true means always reject all dql requests.
|
||||
|
@ -862,6 +852,7 @@ trace:
|
|||
otlp:
|
||||
endpoint: # example: "127.0.0.1:4318"
|
||||
secure: true
|
||||
initTimeoutSeconds: 10 # segcore initialization timeout in seconds, preventing otlp grpc hangs forever
|
||||
|
||||
#when using GPU indexing, Milvus will utilize a memory pool to avoid frequent memory allocation and deallocation.
|
||||
#here, you can set the size of the memory occupied by the memory pool, with the unit being MB.
|
||||
|
|
|
@ -807,6 +807,7 @@ like the old password verification when updating the credential`,
|
|||
Version: "2.4.6",
|
||||
Doc: "if true, do clustering compaction and segment prune on partition key field",
|
||||
DefaultValue: "false",
|
||||
Export: true,
|
||||
}
|
||||
p.UsePartitionKeyAsClusteringKey.Init(base.mgr)
|
||||
|
||||
|
@ -815,6 +816,7 @@ like the old password verification when updating the credential`,
|
|||
Version: "2.4.6",
|
||||
Doc: "if true, do clustering compaction and segment prune on vector field",
|
||||
DefaultValue: "false",
|
||||
Export: true,
|
||||
}
|
||||
p.UseVectorAsClusteringKey.Init(base.mgr)
|
||||
|
||||
|
@ -823,6 +825,7 @@ like the old password verification when updating the credential`,
|
|||
Version: "2.4.6",
|
||||
Doc: "if true, enable vector clustering key and vector clustering compaction",
|
||||
DefaultValue: "false",
|
||||
Export: true,
|
||||
}
|
||||
p.EnableVectorClusteringKey.Init(base.mgr)
|
||||
}
|
||||
|
@ -1352,7 +1355,7 @@ please adjust in embedded Milvus: false`,
|
|||
Key: "proxy.accessLog.cacheSize",
|
||||
Version: "2.3.2",
|
||||
DefaultValue: "0",
|
||||
Doc: "Size of log of write cache, in B. (Close write cache if size was 0",
|
||||
Doc: "Size of log of write cache, in byte. (Close write cache if size was 0)",
|
||||
Export: true,
|
||||
}
|
||||
p.AccessLog.CacheSize.Init(base.mgr)
|
||||
|
@ -1361,7 +1364,8 @@ please adjust in embedded Milvus: false`,
|
|||
Key: "proxy.accessLog.cacheFlushInterval",
|
||||
Version: "2.4.0",
|
||||
DefaultValue: "3",
|
||||
Doc: "time interval of auto flush write cache, in Seconds. (Close auto flush if interval was 0)",
|
||||
Doc: "time interval of auto flush write cache, in seconds. (Close auto flush if interval was 0)",
|
||||
Export: true,
|
||||
}
|
||||
p.AccessLog.CacheFlushInterval.Init(base.mgr)
|
||||
|
||||
|
@ -2120,7 +2124,7 @@ func (p *queryCoordConfig) init(base *BaseTable) {
|
|||
Version: "2.4.4",
|
||||
DefaultValue: "200",
|
||||
Doc: "the interval of collection observer",
|
||||
Export: false,
|
||||
Export: true,
|
||||
}
|
||||
p.CollectionObserverInterval.Init(base.mgr)
|
||||
|
||||
|
@ -2129,7 +2133,7 @@ func (p *queryCoordConfig) init(base *BaseTable) {
|
|||
Version: "2.4.4",
|
||||
DefaultValue: "100",
|
||||
Doc: "the interval of check executed flag to force to pull dist",
|
||||
Export: false,
|
||||
Export: true,
|
||||
}
|
||||
p.CheckExecutedFlagInterval.Init(base.mgr)
|
||||
|
||||
|
@ -2459,7 +2463,7 @@ func (p *queryNodeConfig) init(base *BaseTable) {
|
|||
p.FixedFileSizeForMmapManager.Init(base.mgr)
|
||||
|
||||
p.MaxMmapDiskPercentageForMmapManager = ParamItem{
|
||||
Key: "querynode.mmap.maxDiskUsagePercentageForMmapAlloc",
|
||||
Key: "queryNode.mmap.maxDiskUsagePercentageForMmapAlloc",
|
||||
Version: "2.4.6",
|
||||
DefaultValue: "20",
|
||||
Doc: "disk percentage used in mmap chunk manager",
|
||||
|
@ -3312,6 +3316,7 @@ During compaction, the size of segment # of rows is able to exceed segment max #
|
|||
Version: "2.4.6",
|
||||
Doc: "The time interval for regularly syncing segments",
|
||||
DefaultValue: "300", // 5 * 60 seconds
|
||||
Export: true,
|
||||
}
|
||||
p.SyncSegmentsInterval.Init(base.mgr)
|
||||
|
||||
|
@ -3383,6 +3388,7 @@ During compaction, the size of segment # of rows is able to exceed segment max #
|
|||
Version: "2.4.6",
|
||||
DefaultValue: "600",
|
||||
Doc: "clustering compaction trigger interval in seconds",
|
||||
Export: true,
|
||||
}
|
||||
p.ClusteringCompactionTriggerInterval.Init(base.mgr)
|
||||
|
||||
|
@ -3390,6 +3396,7 @@ During compaction, the size of segment # of rows is able to exceed segment max #
|
|||
Key: "dataCoord.compaction.clustering.stateCheckInterval",
|
||||
Version: "2.4.6",
|
||||
DefaultValue: "10",
|
||||
Export: true,
|
||||
}
|
||||
p.ClusteringCompactionStateCheckInterval.Init(base.mgr)
|
||||
|
||||
|
@ -3397,6 +3404,7 @@ During compaction, the size of segment # of rows is able to exceed segment max #
|
|||
Key: "dataCoord.compaction.clustering.gcInterval",
|
||||
Version: "2.4.6",
|
||||
DefaultValue: "600",
|
||||
Export: true,
|
||||
}
|
||||
p.ClusteringCompactionGCInterval.Init(base.mgr)
|
||||
|
||||
|
@ -3405,6 +3413,7 @@ During compaction, the size of segment # of rows is able to exceed segment max #
|
|||
Version: "2.4.6",
|
||||
Doc: "The minimum interval between clustering compaction executions of one collection, to avoid redundant compaction",
|
||||
DefaultValue: "3600",
|
||||
Export: true,
|
||||
}
|
||||
p.ClusteringCompactionMinInterval.Init(base.mgr)
|
||||
|
||||
|
@ -3413,6 +3422,7 @@ During compaction, the size of segment # of rows is able to exceed segment max #
|
|||
Version: "2.4.6",
|
||||
Doc: "If a collection haven't been clustering compacted for longer than maxInterval, force compact",
|
||||
DefaultValue: "86400",
|
||||
Export: true,
|
||||
}
|
||||
p.ClusteringCompactionMaxInterval.Init(base.mgr)
|
||||
|
||||
|
@ -3421,6 +3431,7 @@ During compaction, the size of segment # of rows is able to exceed segment max #
|
|||
Version: "2.4.6",
|
||||
Doc: "If new data size is large than newDataSizeThreshold, execute clustering compaction",
|
||||
DefaultValue: "512m",
|
||||
Export: true,
|
||||
}
|
||||
p.ClusteringCompactionNewDataSizeThreshold.Init(base.mgr)
|
||||
|
||||
|
@ -3428,7 +3439,6 @@ During compaction, the size of segment # of rows is able to exceed segment max #
|
|||
Key: "dataCoord.compaction.clustering.timeout",
|
||||
Version: "2.4.6",
|
||||
DefaultValue: "3600",
|
||||
Doc: "timeout in seconds for clustering compaction, the task will stop if timeout",
|
||||
}
|
||||
p.ClusteringCompactionTimeoutInSeconds.Init(base.mgr)
|
||||
|
||||
|
@ -3437,6 +3447,7 @@ During compaction, the size of segment # of rows is able to exceed segment max #
|
|||
Version: "2.4.6",
|
||||
Doc: "If clustering compaction job is finished for a long time, gc it",
|
||||
DefaultValue: "259200",
|
||||
Export: true,
|
||||
}
|
||||
p.ClusteringCompactionDropTolerance.Init(base.mgr)
|
||||
|
||||
|
|
|
@ -35,6 +35,13 @@ fi
|
|||
|
||||
# starting the timer
|
||||
beginTime=`date +%s`
|
||||
pushd cmd/tools
|
||||
$TEST_CMD -race -tags dynamic,test -v -coverpkg=./... -coverprofile=profile.out -covermode=atomic ./...
|
||||
if [ -f profile.out ]; then
|
||||
grep -v kafka profile.out | grep -v planparserv2/generated | grep -v mocks | sed '1d' >> ../${FILE_COVERAGE_INFO}
|
||||
rm profile.out
|
||||
fi
|
||||
popd
|
||||
for d in $(go list ./internal/... | grep -v -e vendor -e kafka -e planparserv2/generated -e mocks); do
|
||||
$TEST_CMD -race -tags dynamic,test -v -coverpkg=./... -coverprofile=profile.out -covermode=atomic "$d"
|
||||
if [ -f profile.out ]; then
|
||||
|
@ -42,13 +49,6 @@ for d in $(go list ./internal/... | grep -v -e vendor -e kafka -e planparserv2/g
|
|||
rm profile.out
|
||||
fi
|
||||
done
|
||||
for d in $(go list ./cmd/tools/... | grep -v -e vendor -e kafka -e planparserv2/generated -e mocks); do
|
||||
$TEST_CMD -race -tags dynamic,test -v -coverpkg=./... -coverprofile=profile.out -covermode=atomic "$d"
|
||||
if [ -f profile.out ]; then
|
||||
grep -v kafka profile.out | grep -v planparserv2/generated | grep -v mocks | sed '1d' >> ../${FILE_COVERAGE_INFO}
|
||||
rm profile.out
|
||||
fi
|
||||
done
|
||||
pushd pkg
|
||||
for d in $(go list ./... | grep -v -e vendor -e kafka -e planparserv2/generated -e mocks); do
|
||||
$TEST_CMD -race -tags dynamic,test -v -coverpkg=./... -coverprofile=profile.out -covermode=atomic "$d"
|
||||
|
|
Loading…
Reference in New Issue