Get UnknownTaskExecutorException when I add a new configuration in flink-conf.yaml

2019-01-31 Thread yinhua.dai
Hi Community,I added below item in flink-conf.yaml, and I saw UnknownTaskExecutorException each time when I start flink in Windows via start-cluster.bat.*fs.s3a.aws.credentials.provider: com.tr.apt.sqlengine.tables.aws.HadoopAWSCredentialsProviderChain*I'm sure that this new configuration is succes

Videos and slides on Flink Forward Beijing

2019-01-31 Thread Paul Lam
Hi, It’s been a while since Flink Forward Beijing, would the videos and slides be available on the website? Thanks! Best, Paul Lam

Re: No resource available error while testing HA

2019-01-31 Thread Averell
Hi Gary, I faced a similar problem yesterday, but don't know what was the cause yet. The situation that I observed is as follow: - At about 2:57, one of my EMR execution node (IP ...99) got disconnected from YARN resource manager (on RM I could not see that node anymore), despite that the node wa

Re: Writing a custom Rocksdb statistics collector

2019-01-31 Thread Harshvardhan Agrawal
It looks like the DBOptions that are created by the OptionsFactory class are used for opening RocksDB. And yes I missed the fact that DBOptions is not serializable. Thanks for pointing that out. I will go through the metrics exposed via Flink. But does this mean that there no good way of getting n

how to use Hadoop Inputformats with flink shaded s3?

2019-01-31 Thread Cliff Resnick
I need to process some Parquet data from S3 as a unioned input in my DataStream pipeline. From what I know, this requires using the hadoop AvroParquetInputFormat. The problem I'm running into is that also requires using un-shaded hadoop classes that conflict with the Flink shaded hadoop3 FileSyste

Continuous Monitoring of back-pressure

2019-01-31 Thread David Corley
I'm currently writing some code to convert the back-pressure REST API data into Prometheus-compatible output. I was just curious why back-pressure wasn't already exposed as a metric in the in-built Prometheus exporter? Is it because the thread-sampling is too intensive? Or too slow (particularly if

Re: Forking a stream with Flink

2019-01-31 Thread Dawid Wysakowicz
Hi Daniel, The answer to you original question is you can just keyBy[1] by e.g. the machineId and then computations on KeyedStream are applied independently for each key. Best, Dawid [1] https://ci.apache.org/projects/flink/flink-docs-release-1.7/dev/stream/operators/#datastream-transformations

Re: Forking a stream with Flink

2019-01-31 Thread Daniel Krenn
I don't get what happened here. Did Selvaraj just hijack this question? Or what is going on? Am Di., 29. Jan. 2019 um 17:01 Uhr schrieb Selvaraj chennappan < selvarajchennap...@gmail.com>: > I think there is misunderstanding . I want to compare raw json and > transformed record . > Hence I need t

Re: Data loss when restoring from savepoint

2019-01-31 Thread Juho Autio
Hello, is there anyone that could help with this? On Fri, Jan 11, 2019 at 8:14 AM Juho Autio wrote: > Stefan, would you have time to comment? > > On Wednesday, January 2, 2019, Juho Autio wrote: > >> Bump – does anyone know if Stefan will be available to comment the latest >> findings? Thanks.

Re: Why OnDataStream in flink accepts partial function but DataStream do not

2019-01-31 Thread Dawid Wysakowicz
Hi, AFAIK it clashes with def map[R: TypeInformation](mapper: MapFunction[T, R]): DataStream[R] if you use partially applied function. You can check the discussion behind the OnDataStream implementation in this jira[1]. [1] https://issues.apache.org/jira/browse/FLINK-1159 On 31/01/2019 04:10,