Add log for singlenode

Signed-off-by: Xiangyu Wang <xiangyu.wang@zilliz.com>
pull/4973/head^2
Xiangyu Wang 2021-03-05 18:16:50 +08:00 committed by yefu.chen
parent 33d8e13f96
commit e5d595564a
15 changed files with 61 additions and 2 deletions

View File

@ -4,6 +4,8 @@ import (
"os" "os"
"github.com/zilliztech/milvus-distributed/cmd/distributed/roles" "github.com/zilliztech/milvus-distributed/cmd/distributed/roles"
"github.com/zilliztech/milvus-distributed/internal/log"
"github.com/zilliztech/milvus-distributed/internal/logutil"
) )
func initRoles(roles *roles.MilvusRoles) { func initRoles(roles *roles.MilvusRoles) {
@ -19,9 +21,24 @@ func initRoles(roles *roles.MilvusRoles) {
roles.EnableMsgStreamService = true roles.EnableMsgStreamService = true
} }
func initLogCfg() log.Config {
logCfg := log.Config{}
logCfg.Format = "text"
logCfg.Level = "debug"
logCfg.Development = true
logCfg.File.MaxSize = 300
logCfg.File.MaxBackups = 20
logCfg.File.MaxDays = 10
logCfg.File.Filename = "/tmp/milvus/singlenode.log"
return logCfg
}
func main() { func main() {
var roles roles.MilvusRoles var roles roles.MilvusRoles
initRoles(&roles) initRoles(&roles)
os.Setenv("QUERY_NODE_ID", "1") os.Setenv("QUERY_NODE_ID", "1")
logCfg := initLogCfg()
logutil.SetupLogger(&logCfg)
roles.Run(true) roles.Run(true)
} }

1
go.sum
View File

@ -288,6 +288,7 @@ github.com/prometheus/procfs v0.0.2/go.mod h1:TjEm7ze935MbeOT/UhFTIMYKhuLP4wbCsT
github.com/prometheus/tsdb v0.7.1/go.mod h1:qhTCs0VvXwvX/y3TZrWD7rabWM+ijKTux40TwIPHuXU= github.com/prometheus/tsdb v0.7.1/go.mod h1:qhTCs0VvXwvX/y3TZrWD7rabWM+ijKTux40TwIPHuXU=
github.com/protocolbuffers/protobuf v3.15.3+incompatible h1:5WExaSYHEGvU73sVHvqe+3/APOOyCVg/pDCeAlfpCrw= github.com/protocolbuffers/protobuf v3.15.3+incompatible h1:5WExaSYHEGvU73sVHvqe+3/APOOyCVg/pDCeAlfpCrw=
github.com/protocolbuffers/protobuf v3.15.4+incompatible h1:Blv4dGFGqHXX+r5Tqoc1ziXPMDElqZ+/ryYcE4bddN4= github.com/protocolbuffers/protobuf v3.15.4+incompatible h1:Blv4dGFGqHXX+r5Tqoc1ziXPMDElqZ+/ryYcE4bddN4=
github.com/protocolbuffers/protobuf v3.15.5+incompatible h1:NsnktN0DZ4i7hXZ6HPFH395SptFlMVhSc8XuhkiOwzI=
github.com/rogpeppe/fastuuid v0.0.0-20150106093220-6724a57986af/go.mod h1:XWv6SoW27p1b0cqNHllgS5HIMJraePCO15w5zCzIWYg= github.com/rogpeppe/fastuuid v0.0.0-20150106093220-6724a57986af/go.mod h1:XWv6SoW27p1b0cqNHllgS5HIMJraePCO15w5zCzIWYg=
github.com/rogpeppe/go-internal v1.3.0/go.mod h1:M8bDsm7K2OlrFYOpmOWEs/qY81heoFRclV5y23lUDJ4= github.com/rogpeppe/go-internal v1.3.0/go.mod h1:M8bDsm7K2OlrFYOpmOWEs/qY81heoFRclV5y23lUDJ4=
github.com/rs/xid v1.2.1 h1:mhH9Nq+C1fY2l1XIpgxIiUOfNpRBYH1kKcr+qfKgjRc= github.com/rs/xid v1.2.1 h1:mhH9Nq+C1fY2l1XIpgxIiUOfNpRBYH1kKcr+qfKgjRc=

View File

@ -655,18 +655,21 @@ func newInsertBufferNode(ctx context.Context, flushMeta *metaTable,
//input stream, data node time tick //input stream, data node time tick
wTt, _ := factory.NewMsgStream(ctx) wTt, _ := factory.NewMsgStream(ctx)
wTt.AsProducer([]string{Params.TimeTickChannelName}) wTt.AsProducer([]string{Params.TimeTickChannelName})
log.Debug("datanode AsProducer: " + Params.TimeTickChannelName)
var wTtMsgStream msgstream.MsgStream = wTt var wTtMsgStream msgstream.MsgStream = wTt
wTtMsgStream.Start() wTtMsgStream.Start()
// update statistics channel // update statistics channel
segS, _ := factory.NewMsgStream(ctx) segS, _ := factory.NewMsgStream(ctx)
segS.AsProducer([]string{Params.SegmentStatisticsChannelName}) segS.AsProducer([]string{Params.SegmentStatisticsChannelName})
log.Debug("datanode AsProducer: " + Params.SegmentStatisticsChannelName)
var segStatisticsMsgStream msgstream.MsgStream = segS var segStatisticsMsgStream msgstream.MsgStream = segS
segStatisticsMsgStream.Start() segStatisticsMsgStream.Start()
// segment flush completed channel // segment flush completed channel
cf, _ := factory.NewMsgStream(ctx) cf, _ := factory.NewMsgStream(ctx)
cf.AsProducer([]string{Params.CompleteFlushChannelName}) cf.AsProducer([]string{Params.CompleteFlushChannelName})
log.Debug("datanode AsProducer: " + Params.CompleteFlushChannelName)
var completeFlushStream msgstream.MsgStream = cf var completeFlushStream msgstream.MsgStream = cf
completeFlushStream.Start() completeFlushStream.Start()

View File

@ -2,7 +2,9 @@ package datanode
import ( import (
"context" "context"
"strings"
"github.com/zilliztech/milvus-distributed/internal/log"
"github.com/zilliztech/milvus-distributed/internal/msgstream" "github.com/zilliztech/milvus-distributed/internal/msgstream"
"github.com/zilliztech/milvus-distributed/internal/util/flowgraph" "github.com/zilliztech/milvus-distributed/internal/util/flowgraph"
) )
@ -16,6 +18,7 @@ func newDmInputNode(ctx context.Context, factory msgstream.Factory) *flowgraph.I
insertStream, _ := factory.NewTtMsgStream(ctx) insertStream, _ := factory.NewTtMsgStream(ctx)
insertStream.AsConsumer(consumeChannels, consumeSubName) insertStream.AsConsumer(consumeChannels, consumeSubName)
log.Debug("datanode AsConsumer: " + strings.Join(consumeChannels, ", ") + " : " + consumeSubName)
var stream msgstream.MsgStream = insertStream var stream msgstream.MsgStream = insertStream
node := flowgraph.NewInputNode(&stream, "dmInputNode", maxQueueLength, maxParallelism) node := flowgraph.NewInputNode(&stream, "dmInputNode", maxQueueLength, maxParallelism)
@ -30,6 +33,7 @@ func newDDInputNode(ctx context.Context, factory msgstream.Factory) *flowgraph.I
tmpStream, _ := factory.NewTtMsgStream(ctx) tmpStream, _ := factory.NewTtMsgStream(ctx)
tmpStream.AsConsumer(Params.DDChannelNames, consumeSubName) tmpStream.AsConsumer(Params.DDChannelNames, consumeSubName)
log.Debug("datanode AsConsumer: " + strings.Join(Params.DDChannelNames, ", ") + " : " + consumeSubName)
var stream msgstream.MsgStream = tmpStream var stream msgstream.MsgStream = tmpStream
node := flowgraph.NewInputNode(&stream, "ddInputNode", maxQueueLength, maxParallelism) node := flowgraph.NewInputNode(&stream, "ddInputNode", maxQueueLength, maxParallelism)

View File

@ -6,6 +6,7 @@ import (
"fmt" "fmt"
"path" "path"
"strconv" "strconv"
"strings"
"sync" "sync"
"sync/atomic" "sync/atomic"
"time" "time"
@ -190,6 +191,7 @@ func (s *Server) initMeta() error {
func (s *Server) initSegmentInfoChannel() { func (s *Server) initSegmentInfoChannel() {
segmentInfoStream, _ := s.msFactory.NewMsgStream(s.ctx) segmentInfoStream, _ := s.msFactory.NewMsgStream(s.ctx)
segmentInfoStream.AsProducer([]string{Params.SegmentInfoChannelName}) segmentInfoStream.AsProducer([]string{Params.SegmentInfoChannelName})
log.Debug("dataservice AsProducer: " + Params.SegmentInfoChannelName)
s.segmentInfoStream = segmentInfoStream s.segmentInfoStream = segmentInfoStream
s.segmentInfoStream.Start() s.segmentInfoStream.Start()
} }
@ -199,6 +201,7 @@ func (s *Server) initMsgProducer() error {
return err return err
} }
s.ttMsgStream.AsConsumer([]string{Params.TimeTickChannelName}, Params.DataServiceSubscriptionName) s.ttMsgStream.AsConsumer([]string{Params.TimeTickChannelName}, Params.DataServiceSubscriptionName)
log.Debug("dataservice AsConsumer: " + Params.TimeTickChannelName + " : " + Params.DataServiceSubscriptionName)
s.ttMsgStream.Start() s.ttMsgStream.Start()
s.ttBarrier = timesync.NewHardTimeTickBarrier(s.ctx, s.ttMsgStream, s.cluster.GetNodeIDs()) s.ttBarrier = timesync.NewHardTimeTickBarrier(s.ctx, s.ttMsgStream, s.cluster.GetNodeIDs())
s.ttBarrier.Start() s.ttBarrier.Start()
@ -206,6 +209,7 @@ func (s *Server) initMsgProducer() error {
return err return err
} }
s.k2sMsgStream.AsProducer(Params.K2SChannelNames) s.k2sMsgStream.AsProducer(Params.K2SChannelNames)
log.Debug("dataservice AsProducer: " + strings.Join(Params.K2SChannelNames, ", "))
s.k2sMsgStream.Start() s.k2sMsgStream.Start()
dataNodeTTWatcher := newDataNodeTimeTickWatcher(s.meta, s.segAllocator, s.cluster) dataNodeTTWatcher := newDataNodeTimeTickWatcher(s.meta, s.segAllocator, s.cluster)
k2sMsgWatcher := timesync.NewMsgTimeTickWatcher(s.k2sMsgStream) k2sMsgWatcher := timesync.NewMsgTimeTickWatcher(s.k2sMsgStream)
@ -324,6 +328,7 @@ func (s *Server) startStatsChannel(ctx context.Context) {
defer s.serverLoopWg.Done() defer s.serverLoopWg.Done()
statsStream, _ := s.msFactory.NewMsgStream(ctx) statsStream, _ := s.msFactory.NewMsgStream(ctx)
statsStream.AsConsumer([]string{Params.StatisticsChannelName}, Params.DataServiceSubscriptionName) statsStream.AsConsumer([]string{Params.StatisticsChannelName}, Params.DataServiceSubscriptionName)
log.Debug("dataservice AsConsumer: " + Params.StatisticsChannelName + " : " + Params.DataServiceSubscriptionName)
statsStream.Start() statsStream.Start()
defer statsStream.Close() defer statsStream.Close()
for { for {
@ -353,6 +358,7 @@ func (s *Server) startSegmentFlushChannel(ctx context.Context) {
defer s.serverLoopWg.Done() defer s.serverLoopWg.Done()
flushStream, _ := s.msFactory.NewMsgStream(ctx) flushStream, _ := s.msFactory.NewMsgStream(ctx)
flushStream.AsConsumer([]string{Params.SegmentInfoChannelName}, Params.DataServiceSubscriptionName) flushStream.AsConsumer([]string{Params.SegmentInfoChannelName}, Params.DataServiceSubscriptionName)
log.Debug("dataservice AsConsumer: " + Params.SegmentInfoChannelName + " : " + Params.DataServiceSubscriptionName)
flushStream.Start() flushStream.Start()
defer flushStream.Close() defer flushStream.Close()
for { for {
@ -388,6 +394,7 @@ func (s *Server) startDDChannel(ctx context.Context) {
defer s.serverLoopWg.Done() defer s.serverLoopWg.Done()
ddStream, _ := s.msFactory.NewMsgStream(ctx) ddStream, _ := s.msFactory.NewMsgStream(ctx)
ddStream.AsConsumer([]string{s.ddChannelName}, Params.DataServiceSubscriptionName) ddStream.AsConsumer([]string{s.ddChannelName}, Params.DataServiceSubscriptionName)
log.Debug("dataservice AsConsumer: " + s.ddChannelName + " : " + Params.DataServiceSubscriptionName)
ddStream.Start() ddStream.Start()
defer ddStream.Close() defer ddStream.Close()
for { for {

View File

@ -454,6 +454,7 @@ func (c *Core) setMsgStreams() error {
proxyTimeTickStream, _ := c.msFactory.NewMsgStream(c.ctx) proxyTimeTickStream, _ := c.msFactory.NewMsgStream(c.ctx)
proxyTimeTickStream.AsConsumer([]string{Params.ProxyTimeTickChannel}, Params.MsgChannelSubName) proxyTimeTickStream.AsConsumer([]string{Params.ProxyTimeTickChannel}, Params.MsgChannelSubName)
log.Debug("master AsConsumer: " + Params.ProxyTimeTickChannel + " : " + Params.MsgChannelSubName)
proxyTimeTickStream.Start() proxyTimeTickStream.Start()
// master time tick channel // master time tick channel
@ -462,6 +463,7 @@ func (c *Core) setMsgStreams() error {
} }
timeTickStream, _ := c.msFactory.NewMsgStream(c.ctx) timeTickStream, _ := c.msFactory.NewMsgStream(c.ctx)
timeTickStream.AsProducer([]string{Params.TimeTickChannel}) timeTickStream.AsProducer([]string{Params.TimeTickChannel})
log.Debug("masterservice AsProducer: " + Params.TimeTickChannel)
// master dd channel // master dd channel
if Params.DdChannel == "" { if Params.DdChannel == "" {
@ -469,6 +471,7 @@ func (c *Core) setMsgStreams() error {
} }
ddStream, _ := c.msFactory.NewMsgStream(c.ctx) ddStream, _ := c.msFactory.NewMsgStream(c.ctx)
ddStream.AsProducer([]string{Params.DdChannel}) ddStream.AsProducer([]string{Params.DdChannel})
log.Debug("masterservice AsProducer: " + Params.DdChannel)
c.SendTimeTick = func(t typeutil.Timestamp) error { c.SendTimeTick = func(t typeutil.Timestamp) error {
msgPack := ms.MsgPack{} msgPack := ms.MsgPack{}
@ -602,6 +605,7 @@ func (c *Core) setMsgStreams() error {
} }
dataServiceStream, _ := c.msFactory.NewMsgStream(c.ctx) dataServiceStream, _ := c.msFactory.NewMsgStream(c.ctx)
dataServiceStream.AsConsumer([]string{Params.DataServiceSegmentChannel}, Params.MsgChannelSubName) dataServiceStream.AsConsumer([]string{Params.DataServiceSegmentChannel}, Params.MsgChannelSubName)
log.Debug("master AsConsumer: " + Params.DataServiceSegmentChannel + " : " + Params.MsgChannelSubName)
dataServiceStream.Start() dataServiceStream.Start()
c.DataServiceSegmentChan = make(chan *datapb.SegmentInfo, 1024) c.DataServiceSegmentChan = make(chan *datapb.SegmentInfo, 1024)
c.DataNodeSegmentFlushCompletedChan = make(chan typeutil.UniqueID, 1024) c.DataNodeSegmentFlushCompletedChan = make(chan typeutil.UniqueID, 1024)

View File

@ -102,6 +102,8 @@ func (m *InsertChannelsMap) createInsertMsgStream(collID UniqueID, channels []st
stream, _ := m.msFactory.NewMsgStream(context.Background()) stream, _ := m.msFactory.NewMsgStream(context.Background())
stream.AsProducer(channels) stream.AsProducer(channels)
// FIXME(wxyu): use log.Debug instead
log.Println("proxynode AsProducer: ", channels)
repack := func(tsMsgs []msgstream.TsMsg, hashKeys [][]int32) (map[int32]*msgstream.MsgPack, error) { repack := func(tsMsgs []msgstream.TsMsg, hashKeys [][]int32) (map[int32]*msgstream.MsgPack, error) {
return insertRepackFunc(tsMsgs, hashKeys, m.nodeInstance.segAssigner, true) return insertRepackFunc(tsMsgs, hashKeys, m.nodeInstance.segAssigner, true)
} }

View File

@ -184,6 +184,8 @@ func (node *ProxyNode) Init() error {
node.queryMsgStream, _ = node.msFactory.NewMsgStream(node.ctx) node.queryMsgStream, _ = node.msFactory.NewMsgStream(node.ctx)
node.queryMsgStream.AsProducer(Params.SearchChannelNames) node.queryMsgStream.AsProducer(Params.SearchChannelNames)
// FIXME(wxyu): use log.Debug instead
log.Println("proxynode AsProducer: ", Params.SearchChannelNames)
log.Println("create query message stream ...") log.Println("create query message stream ...")
masterAddr := Params.MasterAddress masterAddr := Params.MasterAddress
@ -211,6 +213,8 @@ func (node *ProxyNode) Init() error {
node.manipulationMsgStream, _ = node.msFactory.NewMsgStream(node.ctx) node.manipulationMsgStream, _ = node.msFactory.NewMsgStream(node.ctx)
node.manipulationMsgStream.AsProducer(Params.InsertChannelNames) node.manipulationMsgStream.AsProducer(Params.InsertChannelNames)
// FIXME(wxyu): use log.Debug instead
log.Println("proxynode AsProducer: ", Params.InsertChannelNames)
repackFunc := func(tsMsgs []msgstream.TsMsg, hashKeys [][]int32) (map[int32]*msgstream.MsgPack, error) { repackFunc := func(tsMsgs []msgstream.TsMsg, hashKeys [][]int32) (map[int32]*msgstream.MsgPack, error) {
return insertRepackFunc(tsMsgs, hashKeys, node.segAssigner, true) return insertRepackFunc(tsMsgs, hashKeys, node.segAssigner, true)
} }

View File

@ -390,8 +390,8 @@ func (sched *TaskScheduler) queryResultLoop() {
defer sched.wg.Done() defer sched.wg.Done()
queryResultMsgStream, _ := sched.msFactory.NewMsgStream(sched.ctx) queryResultMsgStream, _ := sched.msFactory.NewMsgStream(sched.ctx)
queryResultMsgStream.AsConsumer(Params.SearchResultChannelNames, queryResultMsgStream.AsConsumer(Params.SearchResultChannelNames, Params.ProxySubName)
Params.ProxySubName) log.Println("proxynode AsConsumer: ", Params.SearchResultChannelNames, " : ", Params.ProxySubName)
queryNodeNum := Params.QueryNodeNum queryNodeNum := Params.QueryNodeNum
queryResultMsgStream.Start() queryResultMsgStream.Start()

View File

@ -55,6 +55,8 @@ func newTimeTick(ctx context.Context,
t.tickMsgStream, _ = t.msFactory.NewMsgStream(t.ctx) t.tickMsgStream, _ = t.msFactory.NewMsgStream(t.ctx)
t.tickMsgStream.AsProducer(Params.ProxyTimeTickChannelNames) t.tickMsgStream.AsProducer(Params.ProxyTimeTickChannelNames)
// FIXME(wxyu): use log.Debug instead
log.Println("proxynode AsProducer: ", Params.ProxyTimeTickChannelNames)
return t return t
} }

View File

@ -110,6 +110,8 @@ func (s *ProxyService) Init() error {
serviceTimeTickMsgStream, _ := s.msFactory.NewTtMsgStream(s.ctx) serviceTimeTickMsgStream, _ := s.msFactory.NewTtMsgStream(s.ctx)
serviceTimeTickMsgStream.AsProducer([]string{Params.ServiceTimeTickChannel}) serviceTimeTickMsgStream.AsProducer([]string{Params.ServiceTimeTickChannel})
// FIXME(wxyu): use log.Debug instead
log.Println("proxyservice AsProducer: ", []string{Params.ServiceTimeTickChannel})
log.Println("create service time tick producer channel: ", []string{Params.ServiceTimeTickChannel}) log.Println("create service time tick producer channel: ", []string{Params.ServiceTimeTickChannel})
channels := make([]string, Params.InsertChannelNum) channels := make([]string, Params.InsertChannelNum)
@ -119,11 +121,14 @@ func (s *ProxyService) Init() error {
} }
insertTickMsgStream, _ := s.msFactory.NewMsgStream(s.ctx) insertTickMsgStream, _ := s.msFactory.NewMsgStream(s.ctx)
insertTickMsgStream.AsProducer(channels) insertTickMsgStream.AsProducer(channels)
// FIXME(wxyu): use log.Debug instead
log.Println("proxyservice AsProducer: ", channels)
log.Println("create insert time tick producer channel: ", channels) log.Println("create insert time tick producer channel: ", channels)
nodeTimeTickMsgStream, _ := s.msFactory.NewMsgStream(s.ctx) nodeTimeTickMsgStream, _ := s.msFactory.NewMsgStream(s.ctx)
nodeTimeTickMsgStream.AsConsumer(Params.NodeTimeTickChannel, nodeTimeTickMsgStream.AsConsumer(Params.NodeTimeTickChannel,
"proxyservicesub") // TODO: add config "proxyservicesub") // TODO: add config
log.Println("proxynode AsConsumer: ", Params.NodeTimeTickChannel, " : ", "proxyservicesub")
log.Println("create node time tick consumer channel: ", Params.NodeTimeTickChannel) log.Println("create node time tick consumer channel: ", Params.NodeTimeTickChannel)
ttBarrier := newSoftTimeTickBarrier(s.ctx, nodeTimeTickMsgStream, []UniqueID{1}, 10) ttBarrier := newSoftTimeTickBarrier(s.ctx, nodeTimeTickMsgStream, []UniqueID{1}, 10)

View File

@ -81,6 +81,7 @@ func newServiceTimeNode(ctx context.Context, replica ReplicaInterface, factory m
timeTimeMsgStream, _ := factory.NewMsgStream(ctx) timeTimeMsgStream, _ := factory.NewMsgStream(ctx)
timeTimeMsgStream.AsProducer([]string{Params.QueryTimeTickChannelName}) timeTimeMsgStream.AsProducer([]string{Params.QueryTimeTickChannelName})
log.Debug("querynode AsProducer: " + Params.QueryTimeTickChannelName)
return &serviceTimeNode{ return &serviceTimeNode{
baseNode: baseNode, baseNode: baseNode,

View File

@ -15,6 +15,7 @@ import "C"
import ( import (
"context" "context"
"fmt" "fmt"
"strings"
"sync/atomic" "sync/atomic"
"errors" "errors"
@ -301,10 +302,12 @@ func (node *QueryNode) AddQueryChannel(in *queryPb.AddQueryChannelsRequest) (*co
consumeChannels := []string{in.RequestChannelID} consumeChannels := []string{in.RequestChannelID}
consumeSubName := Params.MsgChannelSubName consumeSubName := Params.MsgChannelSubName
node.searchService.searchMsgStream.AsConsumer(consumeChannels, consumeSubName) node.searchService.searchMsgStream.AsConsumer(consumeChannels, consumeSubName)
log.Debug("querynode AsConsumer: " + strings.Join(consumeChannels, ", ") + " : " + consumeSubName)
// add result channel // add result channel
producerChannels := []string{in.ResultChannelID} producerChannels := []string{in.ResultChannelID}
node.searchService.searchResultMsgStream.AsProducer(producerChannels) node.searchService.searchResultMsgStream.AsProducer(producerChannels)
log.Debug("querynode AsProducer: " + strings.Join(producerChannels, ", "))
status := &commonpb.Status{ status := &commonpb.Status{
ErrorCode: commonpb.ErrorCode_SUCCESS, ErrorCode: commonpb.ErrorCode_SUCCESS,
@ -391,6 +394,7 @@ func (node *QueryNode) WatchDmChannels(in *queryPb.WatchDmChannelsRequest) (*com
consumeChannels := in.ChannelIDs consumeChannels := in.ChannelIDs
consumeSubName := Params.MsgChannelSubName consumeSubName := Params.MsgChannelSubName
node.dataSyncService.dmStream.AsConsumer(consumeChannels, consumeSubName) node.dataSyncService.dmStream.AsConsumer(consumeChannels, consumeSubName)
log.Debug("querynode AsConsumer: " + strings.Join(consumeChannels, ", ") + " : " + consumeSubName)
status := &commonpb.Status{ status := &commonpb.Status{
ErrorCode: commonpb.ErrorCode_SUCCESS, ErrorCode: commonpb.ErrorCode_SUCCESS,

View File

@ -6,6 +6,7 @@ import (
"errors" "errors"
"regexp" "regexp"
"strconv" "strconv"
"strings"
"sync" "sync"
"go.uber.org/zap" "go.uber.org/zap"
@ -48,8 +49,10 @@ func newSearchService(ctx context.Context, replica ReplicaInterface, factory msg
consumeChannels := Params.SearchChannelNames consumeChannels := Params.SearchChannelNames
consumeSubName := Params.MsgChannelSubName consumeSubName := Params.MsgChannelSubName
searchStream.AsConsumer(consumeChannels, consumeSubName) searchStream.AsConsumer(consumeChannels, consumeSubName)
log.Debug("querynode AsConsumer: " + strings.Join(consumeChannels, ", ") + " : " + consumeSubName)
producerChannels := Params.SearchResultChannelNames producerChannels := Params.SearchResultChannelNames
searchResultStream.AsProducer(producerChannels) searchResultStream.AsProducer(producerChannels)
log.Debug("querynode AsProducer: " + strings.Join(producerChannels, ", "))
searchServiceCtx, searchServiceCancel := context.WithCancel(ctx) searchServiceCtx, searchServiceCancel := context.WithCancel(ctx)
msgBuffer := make(chan msgstream.TsMsg, receiveBufSize) msgBuffer := make(chan msgstream.TsMsg, receiveBufSize)

View File

@ -2,6 +2,7 @@ package querynode
import ( import (
"context" "context"
"strings"
"time" "time"
"github.com/zilliztech/milvus-distributed/internal/log" "github.com/zilliztech/milvus-distributed/internal/log"
@ -42,6 +43,7 @@ func (sService *statsService) start() {
statsStream, _ := sService.msFactory.NewMsgStream(sService.ctx) statsStream, _ := sService.msFactory.NewMsgStream(sService.ctx)
statsStream.AsProducer(producerChannels) statsStream.AsProducer(producerChannels)
log.Debug("querynode AsProducer: " + strings.Join(producerChannels, ", "))
var statsMsgStream msgstream.MsgStream = statsStream var statsMsgStream msgstream.MsgStream = statsStream