Hi, kumar

Sorry for missed the original question, I think we can not create topic 
dynamically current, creating topic should belong to control flow rather a data 
flow, and user may has some custom configurations of the topic from my 
understanding. Maybe you need implement the logic of check/create/manage topic 
in your custom SinkFunction so that the topic can create dynamically in 
runtime. 

Best,
Leonard Xu

> 在 2020年6月1日,17:02,Prasanna kumar <prasannakumarram...@gmail.com> 写道:
> 
> Leaonard, 
> 
> Thanks for the reply and would look into those options.  
> But as for the original question, could we create a topic dynamically when 
> required . 
> 
> Prasanna.
> 
> On Mon, Jun 1, 2020 at 2:18 PM Leonard Xu <xbjt...@gmail.com 
> <mailto:xbjt...@gmail.com>> wrote:
> Hi, kumar
> 
> Flink support consume/produce from/to multiple kafka topics[1], in your case 
> you can implement KeyedSerializationSchema(legacy interface) or 
> KafkaSerializationSchema[2] to make one producer instance support send data 
> to multiple topics. There is an ITCase you can reference[3].
> 
> 
> Best,
> Leonard Xu
> 
> [1] 
> https://ci.apache.org/projects/flink/flink-docs-master/dev/connectors/kafka.html#kafka-producer
>  
> <https://ci.apache.org/projects/flink/flink-docs-master/dev/connectors/kafka.html#kafka-producer>
> [2]https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-kafka-base/src/main/java/org/apache/flink/streaming/connectors/kafka/KafkaSerializationSchema.java
>  
> <https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-kafka-base/src/main/java/org/apache/flink/streaming/connectors/kafka/KafkaSerializationSchema.java>
>  
> [3]https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-kafka-0.11/src/test/java/org/apache/flink/streaming/connectors/kafka/Kafka011ITCase.java#L126
>  
> <https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-kafka-0.11/src/test/java/org/apache/flink/streaming/connectors/kafka/Kafka011ITCase.java#L126>
>  
> 
>> 在 2020年6月1日,15:35,Prasanna kumar <prasannakumarram...@gmail.com 
>> <mailto:prasannakumarram...@gmail.com>> 写道:
>> 
>> Hi,
>> 
>> I have Use Case where i read events from a Single kafka Stream comprising of 
>> JSON messages.
>> 
>> Requirement is to split the stream into multiple output streams based on 
>> some criteria say based on Type of Event or Based on Type and Customer 
>> associated with the event. 
>> 
>> We could achieve the splitting of stream using Side outputs as i have seen 
>> in the documentation.
>> 
>> Our business environment is such that there could be new event types flowing 
>> in and would the Flink Kafka producer create the topics dynamically based on 
>> the inflowing events. I did not see any documentation saying that it could 
>> create.  
>> 
>> Or should it be always pre created by running a script separately. (Not a 
>> good scalable practice in our case)
>> 
>> Thanks,
>> Prasanna.
> 

Reply via email to