func NewProducer(kafkaBrokerURL string, kafkaSchemaRegistryUrl string) {
producerConfig := getKafkaProducerConfig(config.EnvConfig)
producer, err := confluent_kafka.NewProducer(producerConfig)
if err != nil {
log.WithFields(log.Fields{"err": err}).Error("Failed to create Kafka Producer")
log.Panicf("Unable to create Kafka Producer")
}
client, err := schemaregistry.NewClient(schemaregistry.NewConfig(kafkaSchemaRegistryUrl))
if err != nil {
log.WithFields(log.Fields{"err": err}).Error("Failed to create Kafka Client")
log.Panicf("Unable to create Kafka Client")
}
serializer, err := protobuf.NewSerializer(client, serde.ValueSerde, protobuf.NewSerializerConfig())
if err != nil {
log.WithFields(log.Fields{"err": err}).Error("Failed to create Kafka Serializer")
log.Panicf("Unable to create Kafka Serializer")
}
KafkaProducerInstance = &KafkaProducer{
producer: producer,
serializer: serializer,
}
log.Info("Created Kafka Producer and Serializer")
}
func producerHelper[kdt KafkaMesageDataTypes](message kdt, topicName string) {
deliveryChan := make(chan confluent_kafka.Event)
payload, err := KafkaProducerInstance.serializer.Serialize(topicName, &message)
if err != nil {
log.Errorf("Failed to serialize payload: %v\n", err)
close(deliveryChan)
return
}
err = KafkaProducerInstance.producer.Produce(&confluent_kafka.Message{
TopicPartition: confluent_kafka.TopicPartition{Topic: &topicName, Partition: confluent_kafka.PartitionAny},
Value: payload,
}, deliveryChan)
if err != nil {
log.Errorf("Failed to Produce: %v\n", err)
close(deliveryChan)
return
}
e := <-deliveryChan
m := e.(*confluent_kafka.Message)
if m.TopicPartition.Error != nil {
log.Errorf("Delivery failed: %v\n", m.TopicPartition.Error)
close(deliveryChan)
return
} else {
log.Infof("Delivered message to topic %s [%d] at offset %v\n",
*m.TopicPartition.Topic, m.TopicPartition.Partition, m.TopicPartition.Offset)
}
close(deliveryChan)
}
from confluent_kafka import Consumer, KafkaError
import KafkaDiagnoseResult_pb2 # replace with your generated module name
from google.protobuf.message import DecodeError
# Kafka consumer configuration
conf = {
'bootstrap.servers': "localhost:9092/v3/", # Replace with your Kafka server address
'group.id': "myGroup",
'auto.offset.reset': 'earliest'
}
# Create a consumer instance
consumer = Consumer(conf)
# Subscribe to a topic
from confluent_kafka import Consumer, KafkaError
import KafkaDiagnoseResult_pb2
from google.protobuf.message import DecodeError
# Kafka consumer configuration
conf = {
'bootstrap.servers': "localhost:9092/v3/",
'group.id': "myGroup",
'auto.offset.reset': 'earliest'
}
# Create a consumer instance
consumer = Consumer(conf)
# Subscribe to a topic
consumer.subscribe(['diagnosis'])
try:
while True:
msg = consumer.poll(1.0)
if msg is None:
continue
if msg.error():
if msg.error().code() == KafkaError._PARTITION_EOF:
# End of partition event
continue
else:
print(msg.error())
break
# Deserialize the message
try:
data = KafkaDiagnoseResult_pb2.KafkaDiagnoseRequest()
data.ParseFromString(msg.value())
except DecodeError as e:
print(f"Error parsing message: {e}")
print(f"Raw message data: {msg.value()}")
print("Received message: ", data)
except KeyboardInterrupt:
pass
finally:
consumer.close()
Error parsing message
I am trying to debug it but unable to.
proton
to generate the pb2 file.Your help is appreciated.
Thank you
I can get the message in raw format:
Raw message data: b'\x00\x00\x00\x00\x02\x02\x08\n$1775100a-1a47-48b2-93b7-b7a331be59b4\x12\tcompleted'
print(" Decode 1: ", dict_str)
print("Decode 2: ", ast.literal_eval(dict_str))
Output from above code:
Unparsed Message: b'\x00\x00\x00\x00\x02\x02\x08\n$ccb0ad7e-abb2-4af6-90d1-187381f9d47e\x12\tcompleted'
Decode 1:
$ccb0ad7e-abb2-4af6-90d1-187381f9d47e completed
Inner Exception Here source code string cannot contain null bytes
Your Go client is serializing with the Schema Registry meaning your Python code must do the same. The records are not "just Protobuf" since there's a schema ID also encoded in the bytes, therefore a regular Protobuf parser will fail.
There's example code in the repo for consuming Protobuf with the Registry integration
https://github.com/confluentinc/confluent-kafka-python/blob/master/examples/protobuf_consumer.py