2021-01-19 04:10:49 +00:00
|
|
|
package dataservice
|
|
|
|
|
2021-01-22 03:07:07 +00:00
|
|
|
import (
|
|
|
|
"context"
|
|
|
|
"fmt"
|
|
|
|
"log"
|
2021-01-23 12:22:59 +00:00
|
|
|
"sync"
|
2021-01-22 11:43:27 +00:00
|
|
|
|
2021-01-25 07:17:17 +00:00
|
|
|
"github.com/zilliztech/milvus-distributed/internal/msgstream/util"
|
|
|
|
|
2021-01-22 11:43:27 +00:00
|
|
|
"github.com/zilliztech/milvus-distributed/internal/msgstream"
|
|
|
|
"github.com/zilliztech/milvus-distributed/internal/msgstream/pulsarms"
|
|
|
|
|
|
|
|
"github.com/zilliztech/milvus-distributed/internal/distributed/masterservice"
|
2021-01-22 03:07:07 +00:00
|
|
|
|
|
|
|
"github.com/zilliztech/milvus-distributed/internal/proto/milvuspb"
|
|
|
|
|
|
|
|
"github.com/zilliztech/milvus-distributed/internal/timesync"
|
|
|
|
|
|
|
|
etcdkv "github.com/zilliztech/milvus-distributed/internal/kv/etcd"
|
|
|
|
"go.etcd.io/etcd/clientv3"
|
|
|
|
|
|
|
|
"github.com/zilliztech/milvus-distributed/internal/proto/commonpb"
|
|
|
|
"github.com/zilliztech/milvus-distributed/internal/proto/datapb"
|
|
|
|
"github.com/zilliztech/milvus-distributed/internal/proto/internalpb2"
|
|
|
|
"github.com/zilliztech/milvus-distributed/internal/util/typeutil"
|
|
|
|
)
|
|
|
|
|
2021-01-22 11:43:27 +00:00
|
|
|
const role = "dataservice"
|
|
|
|
|
2021-01-22 03:07:07 +00:00
|
|
|
type DataService interface {
|
|
|
|
typeutil.Service
|
2021-01-25 07:17:17 +00:00
|
|
|
typeutil.Component
|
2021-01-22 03:07:07 +00:00
|
|
|
RegisterNode(req *datapb.RegisterNodeRequest) (*datapb.RegisterNodeResponse, error)
|
|
|
|
Flush(req *datapb.FlushRequest) (*commonpb.Status, error)
|
|
|
|
|
|
|
|
AssignSegmentID(req *datapb.AssignSegIDRequest) (*datapb.AssignSegIDResponse, error)
|
|
|
|
ShowSegments(req *datapb.ShowSegmentRequest) (*datapb.ShowSegmentResponse, error)
|
|
|
|
GetSegmentStates(req *datapb.SegmentStatesRequest) (*datapb.SegmentStatesResponse, error)
|
|
|
|
GetInsertBinlogPaths(req *datapb.InsertBinlogPathRequest) (*datapb.InsertBinlogPathsResponse, error)
|
2021-01-25 07:17:17 +00:00
|
|
|
GetSegmentInfoChannel() (string, error)
|
|
|
|
GetInsertChannels(req *datapb.InsertChannelRequest) ([]string, error)
|
2021-01-22 03:07:07 +00:00
|
|
|
GetCollectionStatistics(req *datapb.CollectionStatsRequest) (*datapb.CollectionStatsResponse, error)
|
|
|
|
GetPartitionStatistics(req *datapb.PartitionStatsRequest) (*datapb.PartitionStatsResponse, error)
|
|
|
|
GetComponentStates() (*internalpb2.ComponentStates, error)
|
|
|
|
}
|
|
|
|
|
|
|
|
type (
|
2021-01-22 11:43:27 +00:00
|
|
|
UniqueID = typeutil.UniqueID
|
|
|
|
Timestamp = typeutil.Timestamp
|
|
|
|
Server struct {
|
2021-01-25 07:16:28 +00:00
|
|
|
ctx context.Context
|
|
|
|
serverLoopCtx context.Context
|
|
|
|
serverLoopCancel context.CancelFunc
|
|
|
|
serverLoopWg sync.WaitGroup
|
|
|
|
state internalpb2.StateCode
|
|
|
|
client *etcdkv.EtcdKV
|
|
|
|
meta *meta
|
|
|
|
segAllocator segmentAllocator
|
|
|
|
statsHandler *statsHandler
|
|
|
|
insertChannelMgr *insertChannelManager
|
|
|
|
allocator allocator
|
|
|
|
cluster *dataNodeCluster
|
|
|
|
msgProducer *timesync.MsgProducer
|
|
|
|
registerFinishCh chan struct{}
|
|
|
|
masterClient *masterservice.GrpcClient
|
|
|
|
ttMsgStream msgstream.MsgStream
|
|
|
|
k2sMsgStream msgstream.MsgStream
|
|
|
|
ddChannelName string
|
|
|
|
segmentInfoStream msgstream.MsgStream
|
|
|
|
segmentFlushStream msgstream.MsgStream
|
2021-01-22 03:07:07 +00:00
|
|
|
}
|
|
|
|
)
|
|
|
|
|
2021-01-23 12:22:59 +00:00
|
|
|
func CreateServer(ctx context.Context, client *masterservice.GrpcClient) (*Server, error) {
|
2021-01-25 07:16:28 +00:00
|
|
|
Params.Init()
|
2021-01-22 11:43:27 +00:00
|
|
|
ch := make(chan struct{})
|
2021-01-22 03:07:07 +00:00
|
|
|
return &Server{
|
|
|
|
ctx: ctx,
|
|
|
|
state: internalpb2.StateCode_INITIALIZING,
|
|
|
|
insertChannelMgr: newInsertChannelManager(),
|
2021-01-22 11:43:27 +00:00
|
|
|
registerFinishCh: ch,
|
|
|
|
cluster: newDataNodeCluster(ch),
|
2021-01-23 12:22:59 +00:00
|
|
|
masterClient: client,
|
2021-01-22 03:07:07 +00:00
|
|
|
}, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) Init() error {
|
2021-01-22 11:43:27 +00:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) Start() error {
|
2021-01-25 07:16:28 +00:00
|
|
|
var err error
|
2021-01-22 11:43:27 +00:00
|
|
|
s.allocator = newAllocatorImpl(s.masterClient)
|
2021-01-25 07:16:28 +00:00
|
|
|
if err = s.initMeta(); err != nil {
|
2021-01-22 03:07:07 +00:00
|
|
|
return err
|
|
|
|
}
|
|
|
|
s.statsHandler = newStatsHandler(s.meta)
|
2021-01-25 07:16:28 +00:00
|
|
|
s.segAllocator, err = newSegmentAllocator(s.meta, s.allocator)
|
2021-01-22 03:07:07 +00:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2021-01-25 07:16:28 +00:00
|
|
|
s.initSegmentInfoChannel()
|
|
|
|
if err = s.initMsgProducer(); err != nil {
|
2021-01-22 11:43:27 +00:00
|
|
|
return err
|
|
|
|
}
|
2021-01-25 07:16:28 +00:00
|
|
|
if err = s.loadMetaFromMaster(); err != nil {
|
2021-01-22 03:07:07 +00:00
|
|
|
return err
|
|
|
|
}
|
2021-01-23 12:22:59 +00:00
|
|
|
s.startServerLoop()
|
2021-01-25 07:16:28 +00:00
|
|
|
s.waitDataNodeRegister()
|
2021-01-22 11:43:27 +00:00
|
|
|
s.state = internalpb2.StateCode_HEALTHY
|
|
|
|
log.Println("start success")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2021-01-22 03:07:07 +00:00
|
|
|
func (s *Server) initMeta() error {
|
|
|
|
etcdClient, err := clientv3.New(clientv3.Config{Endpoints: []string{Params.EtcdAddress}})
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
etcdKV := etcdkv.NewEtcdKV(etcdClient, Params.MetaRootPath)
|
|
|
|
s.client = etcdKV
|
2021-01-25 07:17:17 +00:00
|
|
|
s.meta, err = newMeta(etcdKV)
|
2021-01-22 03:07:07 +00:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2021-01-25 07:16:28 +00:00
|
|
|
func (s *Server) initSegmentInfoChannel() {
|
|
|
|
segmentInfoStream := pulsarms.NewPulsarMsgStream(s.ctx, 1024)
|
|
|
|
segmentInfoStream.SetPulsarClient(Params.PulsarAddress)
|
|
|
|
segmentInfoStream.CreatePulsarProducers([]string{Params.SegmentInfoChannelName})
|
|
|
|
s.segmentInfoStream = segmentInfoStream
|
|
|
|
s.segmentInfoStream.Start()
|
2021-01-22 11:43:27 +00:00
|
|
|
}
|
2021-01-22 03:07:07 +00:00
|
|
|
func (s *Server) initMsgProducer() error {
|
2021-01-25 07:16:28 +00:00
|
|
|
ttMsgStream := pulsarms.NewPulsarMsgStream(s.ctx, 1024)
|
2021-01-25 07:17:17 +00:00
|
|
|
ttMsgStream.SetPulsarClient(Params.PulsarAddress)
|
|
|
|
ttMsgStream.CreatePulsarConsumers([]string{Params.TimeTickChannelName}, Params.DataServiceSubscriptionName, util.NewUnmarshalDispatcher(), 1024)
|
|
|
|
s.ttMsgStream = ttMsgStream
|
2021-01-22 11:43:27 +00:00
|
|
|
s.ttMsgStream.Start()
|
|
|
|
timeTickBarrier := timesync.NewHardTimeTickBarrier(s.ttMsgStream, s.cluster.GetNodeIDs())
|
|
|
|
dataNodeTTWatcher := newDataNodeTimeTickWatcher(s.meta, s.segAllocator, s.cluster)
|
2021-01-25 07:16:28 +00:00
|
|
|
k2sStream := pulsarms.NewPulsarMsgStream(s.ctx, 1024)
|
|
|
|
k2sStream.SetPulsarClient(Params.PulsarAddress)
|
|
|
|
k2sStream.CreatePulsarProducers(Params.K2SChannelNames)
|
|
|
|
s.k2sMsgStream = k2sStream
|
|
|
|
s.k2sMsgStream.Start()
|
|
|
|
k2sMsgWatcher := timesync.NewMsgTimeTickWatcher(s.k2sMsgStream)
|
|
|
|
producer, err := timesync.NewTimeSyncMsgProducer(timeTickBarrier, dataNodeTTWatcher, k2sMsgWatcher)
|
2021-01-22 03:07:07 +00:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
s.msgProducer = producer
|
|
|
|
s.msgProducer.Start(s.ctx)
|
|
|
|
return nil
|
|
|
|
}
|
2021-01-23 12:22:59 +00:00
|
|
|
|
2021-01-22 11:43:27 +00:00
|
|
|
func (s *Server) loadMetaFromMaster() error {
|
|
|
|
log.Println("loading collection meta from master")
|
|
|
|
collections, err := s.masterClient.ShowCollections(&milvuspb.ShowCollectionRequest{
|
|
|
|
Base: &commonpb.MsgBase{
|
|
|
|
MsgType: commonpb.MsgType_kShowCollections,
|
|
|
|
MsgID: -1, // todo add msg id
|
|
|
|
Timestamp: 0, // todo
|
|
|
|
SourceID: -1, // todo
|
|
|
|
},
|
|
|
|
DbName: "",
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
for _, collectionName := range collections.CollectionNames {
|
|
|
|
collection, err := s.masterClient.DescribeCollection(&milvuspb.DescribeCollectionRequest{
|
|
|
|
Base: &commonpb.MsgBase{
|
|
|
|
MsgType: commonpb.MsgType_kDescribeCollection,
|
|
|
|
MsgID: -1, // todo
|
|
|
|
Timestamp: 0, // todo
|
|
|
|
SourceID: -1, // todo
|
|
|
|
},
|
|
|
|
DbName: "",
|
|
|
|
CollectionName: collectionName,
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
log.Println(err.Error())
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
partitions, err := s.masterClient.ShowPartitions(&milvuspb.ShowPartitionRequest{
|
|
|
|
Base: &commonpb.MsgBase{
|
|
|
|
MsgType: commonpb.MsgType_kShowPartitions,
|
|
|
|
MsgID: -1, // todo
|
|
|
|
Timestamp: 0, // todo
|
|
|
|
SourceID: -1, // todo
|
|
|
|
},
|
|
|
|
DbName: "",
|
|
|
|
CollectionName: collectionName,
|
|
|
|
CollectionID: collection.CollectionID,
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
log.Println(err.Error())
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
err = s.meta.AddCollection(&collectionInfo{
|
|
|
|
ID: collection.CollectionID,
|
|
|
|
Schema: collection.Schema,
|
2021-01-23 06:41:29 +00:00
|
|
|
Partitions: partitions.PartitionIDs,
|
2021-01-22 11:43:27 +00:00
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
log.Println(err.Error())
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
}
|
|
|
|
log.Println("load collection meta from master complete")
|
|
|
|
return nil
|
2021-01-22 03:07:07 +00:00
|
|
|
}
|
2021-01-25 07:16:28 +00:00
|
|
|
func (s *Server) startServerLoop() {
|
|
|
|
s.serverLoopCtx, s.serverLoopCancel = context.WithCancel(s.ctx)
|
|
|
|
s.serverLoopWg.Add(2)
|
|
|
|
go s.startStatsChannel(s.serverLoopCtx)
|
|
|
|
go s.startSegmentFlushChannel(s.serverLoopCtx)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) startStatsChannel(ctx context.Context) {
|
|
|
|
defer s.serverLoopWg.Done()
|
|
|
|
statsStream := pulsarms.NewPulsarMsgStream(ctx, 1024)
|
|
|
|
statsStream.SetPulsarClient(Params.PulsarAddress)
|
|
|
|
statsStream.CreatePulsarConsumers([]string{Params.StatisticsChannelName}, Params.DataServiceSubscriptionName, util.NewUnmarshalDispatcher(), 1024)
|
|
|
|
statsStream.Start()
|
|
|
|
defer statsStream.Close()
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-ctx.Done():
|
|
|
|
return
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
msgPack := statsStream.Consume()
|
|
|
|
for _, msg := range msgPack.Msgs {
|
|
|
|
statistics := msg.(*msgstream.SegmentStatisticsMsg)
|
|
|
|
for _, stat := range statistics.SegStats {
|
|
|
|
if err := s.statsHandler.HandleSegmentStat(stat); err != nil {
|
|
|
|
log.Println(err.Error())
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) startSegmentFlushChannel(ctx context.Context) {
|
|
|
|
defer s.serverLoopWg.Done()
|
|
|
|
flushStream := pulsarms.NewPulsarMsgStream(ctx, 1024)
|
|
|
|
flushStream.SetPulsarClient(Params.PulsarAddress)
|
|
|
|
flushStream.CreatePulsarConsumers([]string{Params.SegmentInfoChannelName}, Params.DataServiceSubscriptionName, util.NewUnmarshalDispatcher(), 1024)
|
|
|
|
flushStream.Start()
|
|
|
|
defer flushStream.Close()
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-ctx.Done():
|
|
|
|
log.Println("segment flush channel shut down")
|
|
|
|
return
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
msgPack := flushStream.Consume()
|
|
|
|
for _, msg := range msgPack.Msgs {
|
|
|
|
if msg.Type() != commonpb.MsgType_kSegmentFlushDone {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
realMsg := msg.(*msgstream.FlushCompletedMsg)
|
|
|
|
|
|
|
|
segmentInfo, err := s.meta.GetSegment(realMsg.SegmentID)
|
|
|
|
if err != nil {
|
|
|
|
log.Println(err.Error())
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
segmentInfo.FlushedTime = realMsg.BeginTimestamp
|
|
|
|
if err = s.meta.UpdateSegment(segmentInfo); err != nil {
|
|
|
|
log.Println(err.Error())
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) waitDataNodeRegister() {
|
|
|
|
log.Println("waiting data node to register")
|
|
|
|
<-s.registerFinishCh
|
|
|
|
log.Println("all data nodes register")
|
|
|
|
}
|
2021-01-22 03:07:07 +00:00
|
|
|
|
|
|
|
func (s *Server) Stop() error {
|
2021-01-22 11:43:27 +00:00
|
|
|
s.ttMsgStream.Close()
|
2021-01-25 07:16:28 +00:00
|
|
|
s.k2sMsgStream.Close()
|
2021-01-22 03:07:07 +00:00
|
|
|
s.msgProducer.Close()
|
2021-01-25 07:17:17 +00:00
|
|
|
s.segmentInfoStream.Close()
|
2021-01-23 12:22:59 +00:00
|
|
|
s.stopServerLoop()
|
2021-01-22 03:07:07 +00:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2021-01-23 12:22:59 +00:00
|
|
|
func (s *Server) stopServerLoop() {
|
|
|
|
s.serverLoopCancel()
|
|
|
|
s.serverLoopWg.Wait()
|
|
|
|
}
|
|
|
|
|
2021-01-22 03:07:07 +00:00
|
|
|
func (s *Server) GetComponentStates() (*internalpb2.ComponentStates, error) {
|
2021-01-22 11:43:27 +00:00
|
|
|
resp := &internalpb2.ComponentStates{
|
|
|
|
State: &internalpb2.ComponentInfo{
|
|
|
|
NodeID: Params.NodeID,
|
|
|
|
Role: role,
|
|
|
|
StateCode: s.state,
|
|
|
|
},
|
|
|
|
Status: &commonpb.Status{
|
|
|
|
ErrorCode: commonpb.ErrorCode_UNEXPECTED_ERROR,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
dataNodeStates, err := s.cluster.GetDataNodeStates()
|
|
|
|
if err != nil {
|
|
|
|
resp.Status.Reason = err.Error()
|
|
|
|
return resp, nil
|
|
|
|
}
|
|
|
|
resp.SubcomponentStates = dataNodeStates
|
|
|
|
resp.Status.ErrorCode = commonpb.ErrorCode_SUCCESS
|
|
|
|
return resp, nil
|
2021-01-22 03:07:07 +00:00
|
|
|
}
|
|
|
|
|
2021-01-25 07:17:17 +00:00
|
|
|
func (s *Server) GetTimeTickChannel() (string, error) {
|
|
|
|
return Params.TimeTickChannelName, nil
|
2021-01-22 03:07:07 +00:00
|
|
|
}
|
|
|
|
|
2021-01-25 07:17:17 +00:00
|
|
|
func (s *Server) GetStatisticsChannel() (string, error) {
|
|
|
|
return Params.StatisticsChannelName, nil
|
2021-01-22 03:07:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) RegisterNode(req *datapb.RegisterNodeRequest) (*datapb.RegisterNodeResponse, error) {
|
2021-01-23 12:22:59 +00:00
|
|
|
ret := &datapb.RegisterNodeResponse{
|
2021-01-22 03:07:07 +00:00
|
|
|
Status: &commonpb.Status{
|
2021-01-23 12:22:59 +00:00
|
|
|
ErrorCode: commonpb.ErrorCode_UNEXPECTED_ERROR,
|
2021-01-22 03:07:07 +00:00
|
|
|
},
|
2021-01-23 12:22:59 +00:00
|
|
|
}
|
|
|
|
s.cluster.Register(req.Address.Ip, req.Address.Port, req.Base.SourceID)
|
2021-01-25 07:17:17 +00:00
|
|
|
if s.ddChannelName == "" {
|
2021-01-24 12:26:35 +00:00
|
|
|
resp, err := s.masterClient.GetDdChannel()
|
2021-01-23 12:22:59 +00:00
|
|
|
if err != nil {
|
|
|
|
ret.Status.Reason = err.Error()
|
|
|
|
return ret, err
|
|
|
|
}
|
2021-01-24 12:26:35 +00:00
|
|
|
s.ddChannelName = resp
|
2021-01-23 12:22:59 +00:00
|
|
|
}
|
|
|
|
ret.Status.ErrorCode = commonpb.ErrorCode_SUCCESS
|
|
|
|
ret.InitParams = &internalpb2.InitParams{
|
|
|
|
NodeID: Params.NodeID,
|
|
|
|
StartParams: []*commonpb.KeyValuePair{
|
|
|
|
{Key: "DDChannelName", Value: s.ddChannelName},
|
|
|
|
{Key: "SegmentStatisticsChannelName", Value: Params.StatisticsChannelName},
|
|
|
|
{Key: "TimeTickChannelName", Value: Params.TimeTickChannelName},
|
2021-01-25 07:17:17 +00:00
|
|
|
{Key: "CompleteFlushChannelName", Value: Params.SegmentInfoChannelName},
|
2021-01-23 12:22:59 +00:00
|
|
|
},
|
|
|
|
}
|
|
|
|
return ret, nil
|
2021-01-22 03:07:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) Flush(req *datapb.FlushRequest) (*commonpb.Status, error) {
|
2021-01-25 07:17:17 +00:00
|
|
|
s.segAllocator.SealAllSegments(req.CollectionID)
|
2021-01-22 11:43:27 +00:00
|
|
|
return &commonpb.Status{
|
|
|
|
ErrorCode: commonpb.ErrorCode_SUCCESS,
|
|
|
|
}, nil
|
2021-01-22 03:07:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) AssignSegmentID(req *datapb.AssignSegIDRequest) (*datapb.AssignSegIDResponse, error) {
|
|
|
|
resp := &datapb.AssignSegIDResponse{
|
|
|
|
Status: &commonpb.Status{
|
|
|
|
ErrorCode: commonpb.ErrorCode_SUCCESS,
|
|
|
|
},
|
|
|
|
SegIDAssignments: make([]*datapb.SegIDAssignment, 0),
|
|
|
|
}
|
|
|
|
for _, r := range req.SegIDRequests {
|
|
|
|
result := &datapb.SegIDAssignment{
|
|
|
|
Status: &commonpb.Status{
|
|
|
|
ErrorCode: commonpb.ErrorCode_UNEXPECTED_ERROR,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
segmentID, retCount, expireTs, err := s.segAllocator.AllocSegment(r.CollectionID, r.PartitionID, r.ChannelName, int(r.Count))
|
|
|
|
if err != nil {
|
|
|
|
if _, ok := err.(errRemainInSufficient); !ok {
|
|
|
|
result.Status.Reason = fmt.Sprintf("allocation of Collection %d, Partition %d, Channel %s, Count %d error: %s",
|
|
|
|
r.CollectionID, r.PartitionID, r.ChannelName, r.Count, err.Error())
|
|
|
|
resp.SegIDAssignments = append(resp.SegIDAssignments, result)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Printf("no enough space for allocation of Collection %d, Partition %d, Channel %s, Count %d",
|
|
|
|
r.CollectionID, r.PartitionID, r.ChannelName, r.Count)
|
|
|
|
if err = s.openNewSegment(r.CollectionID, r.PartitionID, r.ChannelName); err != nil {
|
|
|
|
result.Status.Reason = fmt.Sprintf("open new segment of Collection %d, Partition %d, Channel %s, Count %d error: %s",
|
|
|
|
r.CollectionID, r.PartitionID, r.ChannelName, r.Count, err.Error())
|
|
|
|
resp.SegIDAssignments = append(resp.SegIDAssignments, result)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
segmentID, retCount, expireTs, err = s.segAllocator.AllocSegment(r.CollectionID, r.PartitionID, r.ChannelName, int(r.Count))
|
|
|
|
if err != nil {
|
|
|
|
result.Status.Reason = fmt.Sprintf("retry allocation of Collection %d, Partition %d, Channel %s, Count %d error: %s",
|
|
|
|
r.CollectionID, r.PartitionID, r.ChannelName, r.Count, err.Error())
|
|
|
|
resp.SegIDAssignments = append(resp.SegIDAssignments, result)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
result.Status.ErrorCode = commonpb.ErrorCode_SUCCESS
|
|
|
|
result.CollectionID = r.CollectionID
|
|
|
|
result.SegID = segmentID
|
|
|
|
result.PartitionID = r.PartitionID
|
|
|
|
result.Count = uint32(retCount)
|
|
|
|
result.ExpireTime = expireTs
|
|
|
|
result.ChannelName = r.ChannelName
|
|
|
|
resp.SegIDAssignments = append(resp.SegIDAssignments, result)
|
|
|
|
}
|
|
|
|
return resp, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) openNewSegment(collectionID UniqueID, partitionID UniqueID, channelName string) error {
|
|
|
|
group, err := s.insertChannelMgr.GetChannelGroup(collectionID, channelName)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2021-01-25 07:17:17 +00:00
|
|
|
|
|
|
|
id, err := s.allocator.allocID()
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
segmentInfo, err := BuildSegment(collectionID, partitionID, id, group)
|
2021-01-22 03:07:07 +00:00
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
if err = s.meta.AddSegment(segmentInfo); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2021-01-22 11:43:27 +00:00
|
|
|
if err = s.segAllocator.OpenSegment(segmentInfo); err != nil {
|
2021-01-22 03:07:07 +00:00
|
|
|
return err
|
|
|
|
}
|
2021-01-25 07:16:28 +00:00
|
|
|
infoMsg := &msgstream.SegmentInfoMsg{
|
|
|
|
SegmentMsg: datapb.SegmentMsg{
|
|
|
|
Base: &commonpb.MsgBase{
|
|
|
|
MsgType: commonpb.MsgType_kSegmentInfo,
|
|
|
|
MsgID: 0,
|
|
|
|
Timestamp: 0, // todo
|
|
|
|
SourceID: 0,
|
|
|
|
},
|
|
|
|
Segment: segmentInfo,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
msgPack := &pulsarms.MsgPack{
|
|
|
|
Msgs: []msgstream.TsMsg{infoMsg},
|
|
|
|
}
|
|
|
|
if err = s.segmentInfoStream.Produce(msgPack); err != nil {
|
|
|
|
return err
|
|
|
|
}
|
2021-01-22 03:07:07 +00:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) ShowSegments(req *datapb.ShowSegmentRequest) (*datapb.ShowSegmentResponse, error) {
|
|
|
|
ids := s.meta.GetSegmentsByCollectionAndPartitionID(req.CollectionID, req.PartitionID)
|
|
|
|
return &datapb.ShowSegmentResponse{SegmentIDs: ids}, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) GetSegmentStates(req *datapb.SegmentStatesRequest) (*datapb.SegmentStatesResponse, error) {
|
|
|
|
resp := &datapb.SegmentStatesResponse{
|
|
|
|
Status: &commonpb.Status{
|
|
|
|
ErrorCode: commonpb.ErrorCode_UNEXPECTED_ERROR,
|
|
|
|
},
|
|
|
|
}
|
|
|
|
|
|
|
|
segmentInfo, err := s.meta.GetSegment(req.SegmentID)
|
|
|
|
if err != nil {
|
|
|
|
resp.Status.Reason = "get segment states error: " + err.Error()
|
|
|
|
return resp, nil
|
|
|
|
}
|
|
|
|
resp.State = segmentInfo.State
|
|
|
|
resp.CreateTime = segmentInfo.OpenTime
|
|
|
|
resp.SealedTime = segmentInfo.SealedTime
|
|
|
|
resp.FlushedTime = segmentInfo.FlushedTime
|
2021-01-25 07:16:28 +00:00
|
|
|
resp.StartPositions = segmentInfo.StartPosition
|
|
|
|
resp.EndPositions = segmentInfo.EndPosition
|
2021-01-22 03:07:07 +00:00
|
|
|
return resp, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) GetInsertBinlogPaths(req *datapb.InsertBinlogPathRequest) (*datapb.InsertBinlogPathsResponse, error) {
|
|
|
|
panic("implement me")
|
|
|
|
}
|
|
|
|
|
2021-01-25 07:17:17 +00:00
|
|
|
func (s *Server) GetInsertChannels(req *datapb.InsertChannelRequest) ([]string, error) {
|
2021-01-22 03:07:07 +00:00
|
|
|
contains, ret := s.insertChannelMgr.ContainsCollection(req.CollectionID)
|
|
|
|
if contains {
|
2021-01-25 07:17:17 +00:00
|
|
|
return ret, nil
|
2021-01-22 03:07:07 +00:00
|
|
|
}
|
2021-01-22 11:43:27 +00:00
|
|
|
channelGroups, err := s.insertChannelMgr.AllocChannels(req.CollectionID, s.cluster.GetNumOfNodes())
|
2021-01-22 03:07:07 +00:00
|
|
|
if err != nil {
|
2021-01-25 07:17:17 +00:00
|
|
|
return nil, err
|
2021-01-22 03:07:07 +00:00
|
|
|
}
|
2021-01-22 11:43:27 +00:00
|
|
|
|
2021-01-22 03:07:07 +00:00
|
|
|
channels := make([]string, Params.InsertChannelNumPerCollection)
|
|
|
|
for _, group := range channelGroups {
|
2021-01-22 11:43:27 +00:00
|
|
|
channels = append(channels, group...)
|
2021-01-22 03:07:07 +00:00
|
|
|
}
|
2021-01-22 11:43:27 +00:00
|
|
|
s.cluster.WatchInsertChannels(channelGroups)
|
|
|
|
|
2021-01-25 07:17:17 +00:00
|
|
|
return channels, nil
|
2021-01-22 03:07:07 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) GetCollectionStatistics(req *datapb.CollectionStatsRequest) (*datapb.CollectionStatsResponse, error) {
|
|
|
|
// todo implement
|
|
|
|
return nil, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *Server) GetPartitionStatistics(req *datapb.PartitionStatsRequest) (*datapb.PartitionStatsResponse, error) {
|
|
|
|
// todo implement
|
|
|
|
return nil, nil
|
2021-01-19 04:10:49 +00:00
|
|
|
}
|
2021-01-25 07:17:17 +00:00
|
|
|
|
|
|
|
func (s *Server) GetSegmentInfoChannel() (string, error) {
|
|
|
|
return Params.SegmentInfoChannelName, nil
|
|
|
|
}
|