Re: Kafka Producer - Null Pointer Exception when processing by element

2017-07-14 Thread Tzu-Li (Gordon) Tai
Hi, It seems like you’ve misunderstood how to use the FlinkKafkaProducer, or is there any specific reason why you want to emit elements to Kafka in a map function? The correct way to use it is to add it as a sink function to your pipeline, i.e. DataStream someStream = … someStream.addSink(new

Re: Joining two aggregated streams

2017-07-14 Thread Udhay
Hi Thanks for your suggestion. I ll try this one.:) -Udhay. -- View this message in context: http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Joining-two-aggregated-streams-tp14123p14289.html Sent from the Apache Flink User Mailing List archive. mailing list archive at Na

Kafka Producer - Null Pointer Exception when processing by element

2017-07-14 Thread earellano
I'm getting a NullPointerException when calling KakfaProducer010.processElement(StreamRecord). Specifically, this comes from its helper function invokeInternally(), and the function's internalProducer not being configured properly, resulting in passing a null value to one its helper functions. We'

Re: Reading static data

2017-07-14 Thread Mohit Anchlia
Is there a way to accomplish this for the batch operations? On Thu, Jul 13, 2017 at 4:59 AM, Timo Walther wrote: > Hi Mohit, > > do you plan to implement a batch or streaming job? If it is a streaming > job: You can use a connected stream (see [1], Slide 34). The static data is > one side of the

Question regarding check memory and CPU used for job

2017-07-14 Thread Claire Yuan
Hi all,  I am currently running a flink session on YARN and try to access the counter to get the infomation for CPU and memory used for my job. However, flink job does not look like other mapReduce job shown on YARN thus I cannot find the counter infos for it.  Is there any way I can find those

Re: High back-pressure after recovering from a save point

2017-07-14 Thread Gyula Fóra
It will work if you assign a new uid to the Kafka source. Gyula On Fri, Jul 14, 2017, 18:42 Tzu-Li (Gordon) Tai wrote: > One thing: do note that `FlinkKafkaConsumer#setStartFromLatest()` does not > have any effect when starting from savepoints. > i.e., the consumer will still start from whateve

Re: High back-pressure after recovering from a save point

2017-07-14 Thread Tzu-Li (Gordon) Tai
One thing: do note that `FlinkKafkaConsumer#setStartFromLatest()` does not have any effect when starting from savepoints. i.e., the consumer will still start from whatever offset is written in the savepoint. On 15 July 2017 at 12:38:10 AM, Tzu-Li (Gordon) Tai (tzuli...@apache.org) wrote: Can yo

Re: High back-pressure after recovering from a save point

2017-07-14 Thread Tzu-Li (Gordon) Tai
Can you try starting from the savepoint, but telling Kafka to start from the latest offset? (@gordon: Is that possible in Flink 1.3.1 or only in 1.4-SNAPSHOT ?) This is already possible in Flink 1.3.x. `FlinkKafkaConsumer#setStartFromLatest()` would be it. On 15 July 2017 at 12:33:53 AM, Steph

Re: High back-pressure after recovering from a save point

2017-07-14 Thread Stephan Ewen
Can you try starting from the savepoint, but telling Kafka to start from the latest offset? (@gordon: Is that possible in Flink 1.3.1 or only in 1.4-SNAPSHOT ?) On Fri, Jul 14, 2017 at 11:18 AM, Kien Truong wrote: > Hi, > > Sorry for the version typo, I'm running 1.3.1. I did not test with 1.2.

Re: S3 recovery and checkpoint directories exhibit explosive growth

2017-07-14 Thread Stephan Ewen
Hi! I am looping in Stefan and Xiaogang who worked a lot in incremental checkpointing. Some background on incremental checkpoints: Incremental checkpoints store "pieces" of the state (RocksDB ssTables) that are shared between checkpoints. Hence it naturally uses more files than no-incremental che

Re: global window trigger

2017-07-14 Thread jad mad
Hi Aljoscha thanks for the comment. is wrapping by a PurgingTrigger.of() the same as doing "return TriggerResult.FIRE_AND_PURGE;" inside of a custom trigger? gave it a test and the result seems the opposite of what I meant... instead of throwing away previous windows' contents, I wanna keep them

Custom Kryo serializer

2017-07-14 Thread Boris Lublinsky
Hi I have several implementations of my Model trait, trait Model { def score(input : AnyVal) : AnyVal def cleanup() : Unit def toBytes() : Array[Byte] def getType : Long } neither one of them are serializable, but are used in the state definition. So I implemented custom serializer imp

Executing Flink server From IntelliJ

2017-07-14 Thread Boris Lublinsky
Hi, I am trying to upgrade my project from Flink 1.2 to 1.3 and getting problems while trying to run Flink server from my Intellij project. The code // Execute on the local Flink server - to test queariable state def executeServer() : Unit = { // We use a mini cluster here for sake of simplic

Re: How to send local files to a flink job on YARN

2017-07-14 Thread Aljoscha Krettek
There’s a bit of a misconception here: in Flink there is no “driver” as there is in spark and the entry point of your program (“main()”) is not executed on the cluster but in the “client”. The main method is only responsible for constructing a program graph, this is then shipped to the cluster a

Re: High back-pressure after recovering from a save point

2017-07-14 Thread Kien Truong
Hi, Sorry for the version typo, I'm running 1.3.1. I did not test with 1.2.x. The jobs runs fine with almost 0 back-pressure if it's started from scratch or if I reuse the kafka consumers group id without specifying the safe point. Best regards, Kien On Jul 14, 2017, 15:59, at 15:59, Stephan E

Re: Flink Elasticsearch Connector: Lucene Error message

2017-07-14 Thread Aljoscha Krettek
This kind of error almost always hints at a dependency clash, i.e. there is some version of this code in the class path that clashed with the version that the Flink program uses. That’s why it works in local mode, where there are probably not many other dependencies and not in cluster mode. How

Re: High back-pressure after recovering from a save point

2017-07-14 Thread Stephan Ewen
Hi! Flink 1.3.2 does not yet exist. Do you mean 1.3.1 or latest master? Can you tell us whether this occurs only in 1.3.x and worked well in 1.2.x? If you just keep the job running without savepoint/restore, you do not get into backpressure situations? Thanks, Stephan On Fri, Jul 14, 2017 at 1

Re: Why would a kafka source checkpoint take so long?

2017-07-14 Thread Gyula Fóra
Hi, I have seen this again yesterday, now with some logging it looks like acquiring the lock took all the time. In this case it was pretty clear that the job started falling behind a few minutes before starting the checkpoint so backpressure seems to be the culprit. Thanks, Gyula Stephan Ewen e