Hi Danny,

Thanks for the response.

In short without restarting we cannot add new sinks or sources.

For better understanding I will explain my problem more clearly.

My scenario is I have two topics, one is configuration topic and second one
is event activities.

* In the configuration topic I will get details of the kafka cluster
details and elasticsearch cluster details.
* In the event activities i will get events and each event will have a
tenantId
* Suppose now we are getting a new tenantId data I need to send data to the
respective elastic search cluster which I will come to know in the runtime
from the configuration topic.
* Is there a way to add a new elastic search sink in the same job without
restarting.

Before starting a job I can create two elastic search sinks and with a
condition I can route the data to the respective elastic search cluster. Is
there a way to do it in runtime?


Thanks and Regards,
Doinesh.


On Wed, Jul 1, 2020 at 5:24 PM Danny Chan <yuzhao....@gmail.com> wrote:

> Sorry, a job graph is solid while we compile it before submitting to the
> cluster, not dynamic as what you want.
>
> You did can write some wrapper operators which response to your own PRCs
> to run the appended operators you want,
> But the you should keep the consistency semantics by yourself.
>
> Best,
> Danny Chan
> 在 2020年6月28日 +0800 PM3:30,C DINESH <dinesh.kitt...@gmail.com>,写道:
>
> Hi All,
>
> In a flink job I have a pipeline. It is consuming data from one kafka
> topic and storing data to Elastic search cluster.
>
> without restarting the job can we add another kafka cluster and another
> elastic search sink to the job. Which means i will supply the new kafka
> cluster and elastic search details in the topic.  After consuming the data
> can our flink job add the new source and sink to the same job.
>
>
> Thanks & Regards,
> Dinesh.
>
>

Reply via email to