2022-04-12 11:47:33 +00:00
|
|
|
package kafka
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"context"
|
|
|
|
"encoding/binary"
|
|
|
|
"fmt"
|
|
|
|
"math/rand"
|
|
|
|
"os"
|
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
2022-08-03 11:06:35 +00:00
|
|
|
"github.com/confluentinc/confluent-kafka-go/kafka"
|
|
|
|
|
2022-04-12 11:47:33 +00:00
|
|
|
"github.com/milvus-io/milvus/internal/common"
|
|
|
|
"github.com/milvus-io/milvus/internal/log"
|
|
|
|
"github.com/milvus-io/milvus/internal/mq/msgstream/mqwrapper"
|
|
|
|
"github.com/milvus-io/milvus/internal/util/paramtable"
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
|
|
"go.uber.org/zap"
|
|
|
|
)
|
|
|
|
|
|
|
|
var Params paramtable.BaseTable
|
|
|
|
|
|
|
|
func TestMain(m *testing.M) {
|
|
|
|
Params.Init()
|
2022-08-03 11:06:35 +00:00
|
|
|
mockCluster, err := kafka.NewMockCluster(1)
|
|
|
|
defer mockCluster.Close()
|
|
|
|
if err != nil {
|
|
|
|
fmt.Printf("Failed to create MockCluster: %s\n", err)
|
|
|
|
os.Exit(1)
|
|
|
|
}
|
|
|
|
|
|
|
|
broker := mockCluster.BootstrapServers()
|
|
|
|
Params.Save("kafka.brokerList", broker)
|
|
|
|
|
2022-04-12 11:47:33 +00:00
|
|
|
exitCode := m.Run()
|
|
|
|
os.Exit(exitCode)
|
|
|
|
}
|
2022-08-03 11:06:35 +00:00
|
|
|
|
|
|
|
func getKafkaBrokerList() string {
|
|
|
|
brokerList := Params.Get("kafka.brokerList")
|
|
|
|
log.Info("get kafka broker list.", zap.String("address", brokerList))
|
|
|
|
return brokerList
|
|
|
|
}
|
|
|
|
|
2022-04-12 11:47:33 +00:00
|
|
|
func IntToBytes(n int) []byte {
|
|
|
|
tmp := int32(n)
|
|
|
|
bytesBuffer := bytes.NewBuffer([]byte{})
|
|
|
|
binary.Write(bytesBuffer, common.Endian, tmp)
|
|
|
|
return bytesBuffer.Bytes()
|
|
|
|
}
|
|
|
|
func BytesToInt(b []byte) int {
|
|
|
|
bytesBuffer := bytes.NewBuffer(b)
|
|
|
|
var tmp int32
|
|
|
|
binary.Read(bytesBuffer, common.Endian, &tmp)
|
|
|
|
return int(tmp)
|
|
|
|
}
|
|
|
|
|
|
|
|
// Consume1 will consume random messages and record the last MessageID it received
|
|
|
|
func Consume1(ctx context.Context, t *testing.T, kc *kafkaClient, topic string, subName string, c chan mqwrapper.MessageID, total *int) {
|
|
|
|
consumer, err := kc.Subscribe(mqwrapper.ConsumerOptions{
|
|
|
|
Topic: topic,
|
|
|
|
SubscriptionName: subName,
|
|
|
|
BufSize: 1024,
|
|
|
|
SubscriptionInitialPosition: mqwrapper.SubscriptionPositionEarliest,
|
|
|
|
})
|
|
|
|
assert.Nil(t, err)
|
|
|
|
assert.NotNil(t, consumer)
|
|
|
|
defer consumer.Close()
|
|
|
|
|
|
|
|
// get random number between 1 ~ 5
|
|
|
|
rand.Seed(time.Now().UnixNano())
|
|
|
|
cnt := 1 + rand.Int()%5
|
|
|
|
|
|
|
|
log.Info("Consume1 start")
|
|
|
|
var msg mqwrapper.Message
|
|
|
|
for i := 0; i < cnt; i++ {
|
|
|
|
select {
|
|
|
|
case <-ctx.Done():
|
|
|
|
log.Info("Consume1 channel closed")
|
|
|
|
return
|
|
|
|
case msg = <-consumer.Chan():
|
|
|
|
if msg == nil {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Info("Consume1 RECV", zap.Any("v", BytesToInt(msg.Payload())))
|
|
|
|
consumer.Ack(msg)
|
|
|
|
(*total)++
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
c <- msg.ID()
|
|
|
|
log.Info("Consume1 randomly RECV", zap.Any("number", cnt))
|
|
|
|
log.Info("Consume1 done")
|
|
|
|
}
|
|
|
|
|
|
|
|
// Consume2 will consume messages from specified MessageID
|
|
|
|
func Consume2(ctx context.Context, t *testing.T, kc *kafkaClient, topic string, subName string, msgID mqwrapper.MessageID, total *int) {
|
|
|
|
consumer, err := kc.Subscribe(mqwrapper.ConsumerOptions{
|
|
|
|
Topic: topic,
|
|
|
|
SubscriptionName: subName,
|
|
|
|
BufSize: 1024,
|
|
|
|
SubscriptionInitialPosition: mqwrapper.SubscriptionPositionEarliest,
|
|
|
|
})
|
|
|
|
assert.Nil(t, err)
|
|
|
|
assert.NotNil(t, consumer)
|
|
|
|
defer consumer.Close()
|
|
|
|
|
|
|
|
err = consumer.Seek(msgID, true)
|
|
|
|
assert.Nil(t, err)
|
|
|
|
|
|
|
|
mm := <-consumer.Chan()
|
|
|
|
consumer.Ack(mm)
|
|
|
|
log.Info("skip the last received message", zap.Any("skip msg", mm.ID()))
|
|
|
|
|
|
|
|
log.Info("Consume2 start")
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-ctx.Done():
|
|
|
|
log.Info("Consume2 channel closed")
|
|
|
|
return
|
|
|
|
case msg, ok := <-consumer.Chan():
|
|
|
|
if msg == nil || !ok {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
log.Info("Consume2 RECV", zap.Any("v", BytesToInt(msg.Payload())))
|
|
|
|
consumer.Ack(msg)
|
|
|
|
(*total)++
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func Consume3(ctx context.Context, t *testing.T, kc *kafkaClient, topic string, subName string, total *int) {
|
|
|
|
consumer, err := kc.Subscribe(mqwrapper.ConsumerOptions{
|
|
|
|
Topic: topic,
|
|
|
|
SubscriptionName: subName,
|
|
|
|
BufSize: 1024,
|
|
|
|
SubscriptionInitialPosition: mqwrapper.SubscriptionPositionEarliest,
|
|
|
|
})
|
|
|
|
assert.Nil(t, err)
|
|
|
|
assert.NotNil(t, consumer)
|
|
|
|
defer consumer.Close()
|
|
|
|
|
|
|
|
log.Info("Consume3 start")
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-ctx.Done():
|
|
|
|
log.Info("Consume3 channel closed")
|
|
|
|
return
|
|
|
|
case msg, ok := <-consumer.Chan():
|
|
|
|
if msg == nil || !ok {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
consumer.Ack(msg)
|
|
|
|
(*total)++
|
2022-08-04 08:02:35 +00:00
|
|
|
log.Info("Consume3 RECV", zap.Any("v", BytesToInt(msg.Payload())), zap.Int("total", *total))
|
2022-04-12 11:47:33 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestKafkaClient_ConsumeWithAck(t *testing.T) {
|
|
|
|
kc := createKafkaClient(t)
|
|
|
|
defer kc.Close()
|
|
|
|
assert.NotNil(t, kc)
|
|
|
|
|
|
|
|
rand.Seed(time.Now().UnixNano())
|
|
|
|
topic := fmt.Sprintf("test-topic-%d", rand.Int())
|
|
|
|
subName := fmt.Sprintf("test-subname-%d", rand.Int())
|
|
|
|
arr := []int{111, 222, 333, 444, 555, 666, 777}
|
|
|
|
c := make(chan mqwrapper.MessageID, 1)
|
|
|
|
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
|
|
|
|
var total1 int
|
|
|
|
var total2 int
|
|
|
|
var total3 int
|
|
|
|
|
|
|
|
producer := createProducer(t, kc, topic)
|
|
|
|
defer producer.Close()
|
|
|
|
produceData(ctx, t, producer, arr)
|
|
|
|
time.Sleep(100 * time.Millisecond)
|
|
|
|
|
|
|
|
ctx1, cancel1 := context.WithTimeout(ctx, 5*time.Second)
|
|
|
|
defer cancel1()
|
|
|
|
Consume1(ctx1, t, kc, topic, subName, c, &total1)
|
|
|
|
|
|
|
|
lastMsgID := <-c
|
|
|
|
log.Info("lastMsgID", zap.Any("lastMsgID", lastMsgID.(*kafkaID).messageID))
|
|
|
|
|
|
|
|
ctx2, cancel2 := context.WithTimeout(ctx, 3*time.Second)
|
|
|
|
Consume2(ctx2, t, kc, topic, subName, lastMsgID, &total2)
|
|
|
|
cancel2()
|
|
|
|
|
|
|
|
time.Sleep(5 * time.Second)
|
|
|
|
ctx3, cancel3 := context.WithTimeout(ctx, 3*time.Second)
|
|
|
|
Consume3(ctx3, t, kc, topic, subName, &total3)
|
|
|
|
cancel3()
|
|
|
|
|
|
|
|
cancel()
|
|
|
|
assert.Equal(t, len(arr), total1+total2)
|
|
|
|
|
|
|
|
assert.Equal(t, len(arr), total3)
|
|
|
|
}
|
|
|
|
|
2022-05-23 04:39:58 +00:00
|
|
|
func TestKafkaClient_SeekPosition(t *testing.T) {
|
2022-04-12 11:47:33 +00:00
|
|
|
kc := createKafkaClient(t)
|
|
|
|
defer kc.Close()
|
|
|
|
|
|
|
|
rand.Seed(time.Now().UnixNano())
|
2022-05-23 04:39:58 +00:00
|
|
|
ctx := context.Background()
|
2022-04-12 11:47:33 +00:00
|
|
|
topic := fmt.Sprintf("test-topic-%d", rand.Int())
|
|
|
|
subName := fmt.Sprintf("test-subname-%d", rand.Int())
|
|
|
|
|
|
|
|
producer := createProducer(t, kc, topic)
|
|
|
|
defer producer.Close()
|
|
|
|
|
2022-05-23 04:39:58 +00:00
|
|
|
data := []int{1, 2, 3}
|
|
|
|
ids := produceData(ctx, t, producer, data)
|
2022-04-12 11:47:33 +00:00
|
|
|
|
2022-05-23 04:39:58 +00:00
|
|
|
consumer := createConsumer(t, kc, topic, subName, mqwrapper.SubscriptionPositionLatest)
|
|
|
|
defer consumer.Close()
|
2022-04-12 11:47:33 +00:00
|
|
|
|
2022-05-23 04:39:58 +00:00
|
|
|
err := consumer.Seek(ids[2], true)
|
|
|
|
assert.Nil(t, err)
|
2022-04-12 11:47:33 +00:00
|
|
|
|
2022-05-23 04:39:58 +00:00
|
|
|
select {
|
|
|
|
case msg := <-consumer.Chan():
|
|
|
|
consumer.Ack(msg)
|
|
|
|
assert.Equal(t, 3, BytesToInt(msg.Payload()))
|
|
|
|
case <-time.After(10 * time.Second):
|
|
|
|
assert.FailNow(t, "should not wait")
|
|
|
|
}
|
2022-04-12 11:47:33 +00:00
|
|
|
}
|
|
|
|
|
2022-05-23 04:39:58 +00:00
|
|
|
func TestKafkaClient_ConsumeFromLatest(t *testing.T) {
|
2022-04-12 11:47:33 +00:00
|
|
|
kc := createKafkaClient(t)
|
|
|
|
defer kc.Close()
|
|
|
|
|
|
|
|
rand.Seed(time.Now().UnixNano())
|
|
|
|
ctx := context.Background()
|
|
|
|
topic := fmt.Sprintf("test-topic-%d", rand.Int())
|
|
|
|
subName := fmt.Sprintf("test-subname-%d", rand.Int())
|
|
|
|
|
|
|
|
producer := createProducer(t, kc, topic)
|
|
|
|
defer producer.Close()
|
|
|
|
|
2022-05-23 04:39:58 +00:00
|
|
|
data := []int{1, 2}
|
|
|
|
produceData(ctx, t, producer, data)
|
2022-04-12 11:47:33 +00:00
|
|
|
|
|
|
|
consumer := createConsumer(t, kc, topic, subName, mqwrapper.SubscriptionPositionLatest)
|
|
|
|
defer consumer.Close()
|
|
|
|
|
2022-05-23 04:39:58 +00:00
|
|
|
go func() {
|
|
|
|
time.Sleep(time.Second * 2)
|
|
|
|
data := []int{3}
|
|
|
|
produceData(ctx, t, producer, data)
|
|
|
|
}()
|
2022-04-12 11:47:33 +00:00
|
|
|
|
|
|
|
select {
|
|
|
|
case msg := <-consumer.Chan():
|
|
|
|
consumer.Ack(msg)
|
|
|
|
assert.Equal(t, 3, BytesToInt(msg.Payload()))
|
2022-05-23 04:39:58 +00:00
|
|
|
case <-time.After(5 * time.Second):
|
2022-04-12 11:47:33 +00:00
|
|
|
assert.FailNow(t, "should not wait")
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestKafkaClient_EarliestMessageID(t *testing.T) {
|
2022-08-03 11:06:35 +00:00
|
|
|
kafkaAddress := getKafkaBrokerList()
|
2022-04-12 11:47:33 +00:00
|
|
|
kc := NewKafkaClientInstance(kafkaAddress)
|
|
|
|
defer kc.Close()
|
|
|
|
|
|
|
|
mid := kc.EarliestMessageID()
|
|
|
|
assert.NotNil(t, mid)
|
|
|
|
}
|
|
|
|
|
2022-05-23 04:39:58 +00:00
|
|
|
func TestKafkaClient_MsgSerializAndDeserialize(t *testing.T) {
|
2022-08-03 11:06:35 +00:00
|
|
|
kafkaAddress := getKafkaBrokerList()
|
2022-05-23 04:39:58 +00:00
|
|
|
kc := NewKafkaClientInstance(kafkaAddress)
|
|
|
|
defer kc.Close()
|
|
|
|
|
|
|
|
mid := kc.EarliestMessageID()
|
|
|
|
msgID, err := kc.BytesToMsgID(mid.Serialize())
|
|
|
|
assert.NoError(t, err)
|
|
|
|
assert.True(t, msgID.AtEarliestPosition())
|
|
|
|
|
|
|
|
msgID, err = kc.StringToMsgID("1")
|
|
|
|
assert.NoError(t, err)
|
|
|
|
assert.NotNil(t, msgID)
|
|
|
|
|
|
|
|
msgID, err = kc.StringToMsgID("1.0")
|
|
|
|
assert.Error(t, err)
|
|
|
|
assert.Nil(t, msgID)
|
|
|
|
}
|
|
|
|
|
2022-06-02 04:12:03 +00:00
|
|
|
func TestKafkaClient_NewKafkaClientInstanceWithConfig(t *testing.T) {
|
|
|
|
config1 := ¶mtable.KafkaConfig{Address: "addr", SaslPassword: "password"}
|
|
|
|
assert.Panics(t, func() { NewKafkaClientInstanceWithConfig(config1) })
|
|
|
|
|
|
|
|
config2 := ¶mtable.KafkaConfig{Address: "addr", SaslUsername: "username"}
|
|
|
|
assert.Panics(t, func() { NewKafkaClientInstanceWithConfig(config2) })
|
|
|
|
|
|
|
|
config3 := ¶mtable.KafkaConfig{Address: "addr", SaslUsername: "username", SaslPassword: "password"}
|
|
|
|
client := NewKafkaClientInstanceWithConfig(config3)
|
|
|
|
assert.NotNil(t, client)
|
|
|
|
assert.NotNil(t, client.basicConfig)
|
2022-08-25 03:02:53 +00:00
|
|
|
|
|
|
|
consumerConfig := make(map[string]string)
|
|
|
|
consumerConfig["client.id"] = "dc"
|
|
|
|
config4 := ¶mtable.KafkaConfig{Address: "addr", SaslUsername: "username", SaslPassword: "password", ConsumerExtraConfig: consumerConfig}
|
|
|
|
client4 := NewKafkaClientInstanceWithConfig(config4)
|
|
|
|
assert.Equal(t, "dc", client4.consumerConfig["client.id"])
|
|
|
|
|
|
|
|
newConsumerConfig := client4.newConsumerConfig("test", 0)
|
|
|
|
clientID, err := newConsumerConfig.Get("client.id", "")
|
|
|
|
assert.Nil(t, err)
|
|
|
|
assert.Equal(t, "dc", clientID)
|
|
|
|
|
|
|
|
producerConfig := make(map[string]string)
|
|
|
|
producerConfig["client.id"] = "dc1"
|
|
|
|
config5 := ¶mtable.KafkaConfig{Address: "addr", SaslUsername: "username", SaslPassword: "password", ProducerExtraConfig: producerConfig}
|
|
|
|
client5 := NewKafkaClientInstanceWithConfig(config5)
|
|
|
|
assert.Equal(t, "dc1", client5.producerConfig["client.id"])
|
|
|
|
|
|
|
|
newProducerConfig := client5.newProducerConfig()
|
|
|
|
pClientID, err := newProducerConfig.Get("client.id", "")
|
|
|
|
assert.Nil(t, err)
|
|
|
|
assert.Equal(t, pClientID, "dc1")
|
|
|
|
|
2022-06-02 04:12:03 +00:00
|
|
|
}
|
|
|
|
|
2022-04-12 11:47:33 +00:00
|
|
|
func createKafkaClient(t *testing.T) *kafkaClient {
|
2022-08-03 11:06:35 +00:00
|
|
|
kafkaAddress := getKafkaBrokerList()
|
2022-04-12 11:47:33 +00:00
|
|
|
kc := NewKafkaClientInstance(kafkaAddress)
|
|
|
|
assert.NotNil(t, kc)
|
|
|
|
return kc
|
|
|
|
}
|
|
|
|
|
|
|
|
func createConsumer(t *testing.T,
|
|
|
|
kc *kafkaClient,
|
|
|
|
topic string,
|
|
|
|
groupID string,
|
|
|
|
initPosition mqwrapper.SubscriptionInitialPosition) mqwrapper.Consumer {
|
|
|
|
consumer, err := kc.Subscribe(mqwrapper.ConsumerOptions{
|
|
|
|
Topic: topic,
|
|
|
|
SubscriptionName: groupID,
|
|
|
|
BufSize: 1024,
|
|
|
|
SubscriptionInitialPosition: initPosition,
|
|
|
|
})
|
|
|
|
assert.Nil(t, err)
|
|
|
|
return consumer
|
|
|
|
}
|
|
|
|
|
|
|
|
func createProducer(t *testing.T, kc *kafkaClient, topic string) mqwrapper.Producer {
|
|
|
|
producer, err := kc.CreateProducer(mqwrapper.ProducerOptions{Topic: topic})
|
|
|
|
assert.Nil(t, err)
|
|
|
|
assert.NotNil(t, producer)
|
|
|
|
return producer
|
|
|
|
}
|
|
|
|
|
|
|
|
func produceData(ctx context.Context, t *testing.T, producer mqwrapper.Producer, arr []int) []mqwrapper.MessageID {
|
|
|
|
var msgIDs []mqwrapper.MessageID
|
|
|
|
for _, v := range arr {
|
|
|
|
msg := &mqwrapper.ProducerMessage{
|
|
|
|
Payload: IntToBytes(v),
|
|
|
|
Properties: map[string]string{},
|
|
|
|
}
|
|
|
|
msgID, err := producer.Send(ctx, msg)
|
|
|
|
msgIDs = append(msgIDs, msgID)
|
|
|
|
assert.Nil(t, err)
|
|
|
|
}
|
2022-08-04 08:02:35 +00:00
|
|
|
|
|
|
|
producer.(*kafkaProducer).p.Flush(500)
|
2022-04-12 11:47:33 +00:00
|
|
|
return msgIDs
|
|
|
|
}
|