Hi Olga,

Sorry for the late reply.
I think that Gordon (cc’ed) could be able to answer your question.

Cheers,
Kostas

> On Oct 13, 2018, at 3:10 PM, Olga Luganska <trebl...@hotmail.com> wrote:
> 
> Any suggestions?
> 
> Thank you
> 
> Sent from my iPhone
> 
> On Oct 9, 2018, at 9:28 PM, Olga Luganska <trebl...@hotmail.com 
> <mailto:trebl...@hotmail.com>> wrote:
> 
>> Hello,
>> 
>> I would like to use Confluent Schema Registry in my streaming job.
>> I was able to make it work with the help of generic Kafka producer and 
>> FlinkKafkaConsumer which is using 
>> ConfluentRegistryAvroDeserializationSchema.  
>> 
>> FlinkKafkaConsumer011<GenericRecord> consumer = new 
>> FlinkKafkaConsumer011<>(MY_TOPIC,
>> ConfluentRegistryAvroDeserializationSchema.forGeneric(schema, SCHEMA_URI), 
>> kafkaProperties);
>> 
>> My question: is it possible to implement producer logic in the 
>> FlinkKafkaProducer to serialize message and store schema id in the Confluent 
>> Schema registry?
>> 
>> I don't think this is going to work with the current interface because 
>> creation and caching of the schema id in the Confluent Schema Registry is 
>> done with the help of 
>> io.confluent.kafka.serializers.KafkaAvroSerializer.class  and all 
>> FlinkKafkaProducer constructors have either SerializationSchema or 
>> KeyedSerializationSchema (part of Flink's own serialization stack) as one of 
>> the parameters.
>> If my assumption is wrong, could you please provide details of 
>> implementation?
>> 
>> Thank you very much,
>> Olga
>> 
>> 
>> 
>> 
>> 
>> ​
>> 
>> 
>> 
>> 

Reply via email to