checkpointing when yarn session crashed

2019-04-07 Thread Shengnan YU
Hi everyone: Recently I am doing a bucketingsink (to hdfs) job by flink on yarn. However I found that if the yarn session crashed, or I manually killed the yarn session, the file on hdfs does not rename to .pending state and the latest checkpoint does not have _metadata. Therefore I cannot resu

Re: Apache Flink - Question about broadcast state pattern usage

2019-04-07 Thread Guowei Ma
Hi 1. I think you could use "Using Managed Operator State"[1] (context.getOperatorStateStore().getBroadcastState()) to use the BroadCastState. But you must use it very carefully and guarantee the semantics of broadcast state yourself. I think "The Broadcast State Pattern"[2] is some best practice

Exceptions when launching counts on a Flink DataSet concurrently

2019-04-07 Thread Juan Rodríguez Hortalá
Hi, I have a very simple program using the local execution environment, that throws NPE and other exceptions related to concurrent access when launching a count for a DataSet from different threads. The program is https://gist.github.com/juanrh/685a89039e866c1067a6efbfc22c753e which is basically t

Re: Flink - Type Erasure Exception trying to use Tuple6 instead of Tuple

2019-04-07 Thread abhishek sharma
I agree with Timothy, POJO would be a much better approach. However, If you are trying to build some generic framework and for different streams, there would be different fields, you can follow the Map approach. For the latter approach, you need to write extra mapper class which will convert all t

How to see user configuration in UI dashboard

2019-04-07 Thread Avi Levi
Is it possible to (or how to) set custom parameters programmatically that can be viewed in configuration tab (UI) either executions config / user configuration ?

Re: Flink - Type Erasure Exception trying to use Tuple6 instead of Tuple

2019-04-07 Thread Timothy Victor
Could this just be solved by creating a POJO model class for your problem? That is, instead of using Tuple6 - create a class that encapsulates your data. This, I think, would solve your problem. But beyond that I think the code will be more understandable. It's hard to have a Tuple6 of all Str