I think "auto.create.topics.enable" is enabled by default [1]?

Best,
Jark

[1]: https://kafka.apache.org/documentation/#auto.create.topics.enable

On Mon, 1 Jun 2020 at 19:55, Leonard Xu <xbjt...@gmail.com> wrote:

> I think @brat is right, I didn’t know the Kafka property
>  'auto.create.topics.enable’ , you can pass the property to Kafka Producer,
> that should work.
> Best,
> Leonard Xu
>
> 在 2020年6月1日,18:33,satya brat <bratsatya...@gmail.com> 写道:
>
> Prasanna,
> You might want to check the kafka broker configs where
> 'auto.create.topics.enable' helps with creating a new topic whenever a new
> message with non existent topic is published.
> https://kafka.apache.org/documentation/#brokerconfigs
>
> I am not too sure about pitfalls if any.
>
> On Mon, Jun 1, 2020 at 3:20 PM Leonard Xu <xbjt...@gmail.com> wrote:
>
>> Hi, kumar
>>
>> Sorry for missed the original question, I think we can not create topic
>> dynamically current, creating topic should belong to control flow rather a
>> data flow, and user may has some custom configurations of the topic from my
>> understanding. Maybe you need implement the logic of check/create/manage
>> topic in your custom SinkFunction so that the topic can create dynamically
>> in runtime.
>>
>> Best,
>> Leonard Xu
>>
>> 在 2020年6月1日,17:02,Prasanna kumar <prasannakumarram...@gmail.com> 写道:
>>
>> Leaonard,
>>
>> Thanks for the reply and would look into those options.
>> But as for the original question, could we create a topic dynamically
>> when required .
>>
>> Prasanna.
>>
>> On Mon, Jun 1, 2020 at 2:18 PM Leonard Xu <xbjt...@gmail.com> wrote:
>>
>>> Hi, kumar
>>>
>>> Flink support consume/produce from/to multiple kafka topics[1], in your
>>> case you can implement KeyedSerializationSchema(legacy interface) or
>>> KafkaSerializationSchema[2] to make one producer instance support send data
>>> to multiple topics. There is an ITCase you can reference[3].
>>>
>>>
>>> Best,
>>> Leonard Xu
>>>
>>> [1]
>>> https://ci.apache.org/projects/flink/flink-docs-master/dev/connectors/kafka.html#kafka-producer
>>> [2]
>>> https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-kafka-base/src/main/java/org/apache/flink/streaming/connectors/kafka/KafkaSerializationSchema.java
>>>
>>> [3]
>>> https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-kafka-0.11/src/test/java/org/apache/flink/streaming/connectors/kafka/Kafka011ITCase.java#L126
>>>
>>>
>>> 在 2020年6月1日,15:35,Prasanna kumar <prasannakumarram...@gmail.com> 写道:
>>>
>>> Hi,
>>>
>>> I have Use Case where i read events from a Single kafka Stream
>>> comprising of JSON messages.
>>>
>>> Requirement is to split the stream into multiple output streams based on
>>> some criteria say based on Type of Event or Based on Type and Customer
>>> associated with the event.
>>>
>>> We could achieve the splitting of stream using Side outputs as i have
>>> seen in the documentation.
>>>
>>> Our business environment is such that there could be new event types
>>> flowing in and would the Flink Kafka producer create the topics dynamically
>>> based on the inflowing events. I did not see any documentation saying
>>> that it could create.
>>>
>>> Or should it be always pre created by running a script separately. (Not
>>> a good scalable practice in our case)
>>>
>>> Thanks,
>>> Prasanna.
>>>
>>>
>>>
>>
>

Reply via email to