Running Flink job as a rest

2020-12-02 Thread dhurandar S
Can Flink job be running as Rest Server, Where Apache Flink job is listening on a port (443). When a user calls this URL with payload, data directly goes to the Apache Flink windowing function. Right now Flink can ingest data from Kafka or Kinesis, but we have a use case where we would like to pus

adding s3 object metadata while using StreamFileSink

2020-06-19 Thread dhurandar S
We are creating files in S3 and we want to update the S3 object metadata with some security-related information for governance purposes. Right now Apache Flink totally abstracts how and when S3 object gets created in the system. Is there a way that we can pass the S3 object metadata and update it

Re: changing the output files names in Streamfilesink from part-00 to something else

2020-05-18 Thread dhurandar S
ribe your needs? Why do you need to modify file names > flexibly? What kind of name do you want? > > Best, > Jingsong Lee > > On Thu, May 14, 2020 at 2:05 AM dhurandar S > wrote: > >> Yes we looked at it , >> The problem is the file name gets generated in a dy

Re: changing the output files names in Streamfilesink from part-00 to something else

2020-05-13 Thread dhurandar S
ojects/flink/flink-docs-stable/dev/connectors/streamfile_sink.html#part-file-configuration > > > > ------ > 发件人:dhurandar S > 日 期:2020年05月13日 05:13:04 > 收件人:user; > 主 题:changing the output files names in Streamfilesink from part-00

changing the output files names in Streamfilesink from part-00 to something else

2020-05-12 Thread dhurandar S
We want to change the name of the file being generated as the output of our StreamFileSink. , when files are generated they are named part-00*, is there a way that we can change the name. In Hadoop, we can change RecordWriters and MultipleOutputs. May I please some help in this regard. This is cau

Re: doing demultiplexing using Apache flink

2020-05-02 Thread dhurandar S
many > > -- > > Ververica GmbH > Registered at Amtsgericht Charlottenburg: HRB 158244 B > Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, Ji > (Tony) Cheng > > > > On Thu, Apr 30, 2020 at 5:21 AM dhurandar S > wrote: > >> Thank yo

Re: doing demultiplexing using Apache flink

2020-04-29 Thread dhurandar S
Thank you Alexander for the response. This is very helpful. Can i apply the same pattern to S3 as well , as in read from Kafka or Kinesis and write multiple files in S3 or multiple topics in Kinesis ? regards, Rahul On Wed, Apr 29, 2020 at 2:32 PM Alexander Fedulov wrote: > Hi Dhurandar, > > it

doing demultiplexing using Apache flink

2020-04-29 Thread dhurandar S
Hi , We have a use case where we have to demultiplex the incoming stream to multiple output streams. We read from 1 Kafka topic and as an output we generate multiple Kafka topics. The logic of generating each new Kafka topic is different and not known beforehand. Users of the system keep adding n

doing demultiplexing using Apache flink

2020-04-29 Thread dhurandar S
> > Hi , > > We have a use case where we have to demultiplex the incoming stream to > multiple output streams. > > We read from 1 Kafka topic and as an output we generate multiple Kafka > topics. The logic of generating each new Kafka topic is different and not > known beforehand. Users of the syst

doing demultiplexing using Apache flink

2020-04-29 Thread dhurandar S
Hi , We have a use case where we have to demultiplex the incoming stream to multiple output streams. We read from 1 Kafka topic and as an output we generate multiple Kafka topics. The logic of generating each new Kafka topic is different and not known beforehand. Users of the system keep adding n

doing demultiplexing using Apache flink

2020-04-29 Thread dhurandar S
Hi , We have a use case where we have to demultiplex the incoming stream to multiple output streams. We read from 1 Kafka topic and as an output we generate multiple Kafka topics. The logic of generating each new Kafka topic is different and not known beforehand. Users of the system keep adding n