Flink JDBC JDBCOutputFormat Open

2016-09-11 Thread Swapnil Chougule
Hi Team, I want to know how tasknumber & numtasks help in opening db connection in Flink JDBC JDBCOutputFormat Open. I checked with docs where it says: taskNumber - The number of the parallel instance.numTasks - The number of parallel tasks.But couldn't get clear idea among parallel instance & par

Re: RawSchema as deserialization schema

2016-09-11 Thread Swapnil Chougule
Thanks Maximilian. I implemented same & it worked for me. I was under impression that RawSchema is available from flink. Regards, Swapnil On Mon, Sep 5, 2016 at 8:48 PM, Maximilian Michels wrote: > Just implement DeserializationSchema and return the byte array from > Kafka. Byte array serializa

Re: Distributed Cache support for StreamExecutionEnvironment

2016-09-11 Thread Swapnil Chougule
Thanks Robert. It worked for me. I have used RichFunction's open() method. Regards, Swapnil On Fri, Sep 9, 2016 at 3:40 PM, Robert Metzger wrote: > Hi Swapnil, > > there's no support for something like DistributedCache in the DataStream > API. > However, as a workaround, you can rely on the Ric

Re: Enriching a tuple mapped from a datastream with data coming from a JDBC source

2016-09-11 Thread Konstantin Knauf
Hi Philipp, the easist way is a RichMap. In the open()-Method you can load the relevant database table into memory (e.g. a HashMap). In the map()-method you than just look up the entry in the HashMap. Of course, this only works if the dataset is small enough to fit in memory. Is it? Cheers, Kon

Re: Firing windows multiple times

2016-09-11 Thread aj.h
In the way that FLIP-2 would solve this problem, secondAggregate would ignore the early firing updates from firstAggregate to prevent double-counting, correct? If that's the case, I am trying to understand why we'd want to trigger early-fires every 30 seconds for the secondAggregate if it's only ac

Enriching a tuple mapped from a datastream with data coming from a JDBC source

2016-09-11 Thread Philipp Bussche
Hi there, I have a data stream (coming from Kafka) that contains information which I want to enrich with information that sits in a database before I handover the enriched tuple to a sink. How would I do that ? I was thinking of somehow combining my streaming job with a JDBC input but wasn't very s