Re: Rocksdb Serialization issue

2020-03-05 Thread David Morin
buckets / e.g. how do you avoid reading the same file > on multiple parallel sinks? > Is your sink running in a keyed context? Maybe even provide the general > pipeline. > > On Thu, Mar 5, 2020 at 2:29 PM David Morin > wrote: > >> Hello Arvid, >> >> After some in

Re: Rocksdb Serialization issue

2020-03-05 Thread David Morin
> > Did you restart from checkpoint or savepoint? > > On Thu, Mar 5, 2020 at 1:14 AM David Morin > wrote: > >> Hello, >> >> I have this Exception in my datastream app and I can't find the root >> cause. >> I consume data from Kafka and it fails

Rocksdb Serialization issue

2020-03-04 Thread David Morin
Hello, I have this Exception in my datastream app and I can't find the root cause. I consume data from Kafka and it fails when I try to get a value from my MapState in RocksDB. It was working in previous release of my app but I can't find the cause of this error. java.lang.ArrayIndexOutOfBounds

Re: yarn session: one JVM per task

2020-02-25 Thread David Morin
ry mentioned > specifying "-s 1" at starting the session would be enough, and it should > work with all existing versions above (including) 1.8. > > Thank you~ > > Xintong Song > > > > On Tue, Feb 25, 2020 at 7:41 PM David Morin > wrote: > >> Hi Gary, >>

Re: yarn session: one JVM per task

2020-02-25 Thread David Morin
ink/flink-docs-stable/concepts/runtime.html#task-slots-and-resources > [3] > https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/stream/operators/#task-chaining-and-resource-groups > > On Tue, Feb 25, 2020 at 10:28 AM David Morin > wrote: > >> Hi Xintong, >> >

Re: yarn session: one JVM per task

2020-02-25 Thread David Morin
same JVM. Not that I'm aware > of. > > > Thank you~ > > Xintong Song > > > > On Mon, Feb 24, 2020 at 8:43 PM David Morin > wrote: > >> Hi, >> >> Thanks Xintong. >> I've noticed than when I use yarn-session.sh with --slots (-s) para

Re: yarn session: one JVM per task

2020-02-24 Thread David Morin
eline and of course to prevent from thread-safe issue Do you think I cannot be confident on that behaviour ? Regards, David On 2020/02/22 17:11:25, David Morin wrote: > Hi, > My app is based on a lib that is not thread safe (yet...). > In waiting of the patch has been pushed, how can I b

yarn session: one JVM per task

2020-02-22 Thread David Morin
Hi, My app is based on a lib that is not thread safe (yet...). In waiting of the patch has been pushed, how can I be sure that my Sink that uses this lib is in one JVM ? Context: I use one Yarn session and send my Flink jobs to this session Regards, David

Re: kafka: how to stop consumption temporarily

2020-01-08 Thread David Morin
e5a135c5ccfd6630634dd . > > [1] > https://cwiki.apache.org/confluence/display/FLINK/FLIP-27%3A+Refactor+Source+Interface > > On Mon, Jan 6, 2020 at 1:16 PM David Morin > wrote: > >> My naive solution can't work because a dump can be quite long. >> So, yes I have to find a

Re: kafka: how to stop consumption temporarily

2020-01-06 Thread David Morin
tions/59201503/flink-kafka-gracefully-close-flink-consuming-messages-from-kafka-source-after-a Any suggestions are welcomed thx. David On 2020/01/06 09:35:37, David Morin wrote: > Hi, > > Thanks for your replies. > Yes Terry. You are right. I can try to create a custom sourc

Re: kafka: how to stop consumption temporarily

2020-01-06 Thread David Morin
at you need adjust the Kafka > > consumer to have communication each other. When your begin the dump > > process, you need to notify another CDC-topic consumer to wait idle. > > > > > > Best, > > Terry Wang > > > > > > > >> 2020年1月2日 1

kafka: how to stop consumption temporarily

2020-01-02 Thread David Morin
Hi, Is there a way to stop temporarily to consume one kafka source in streaming mode ? Use case: I have to consume 2 topics but in fact one of them is more prioritized. One of this topic is dedicated to ingest data from db (change data capture) and one of them is dedicated to make a synchroniza

Re: How to prevent from launching 2 jobs at the same time

2019-09-23 Thread David Morin
is enabled or it picks up the configured JobManager address from the > configuration). > > Of course, you could also provide the JobManager address as a parameter. > > Cheers, > Till > > On Mon, Sep 23, 2019 at 9:08 AM David Morin > wrote: > > > Hi, > > &g

Re: How to prevent from launching 2 jobs at the same time

2019-09-23 Thread David Morin
> > yourself. > > > > Regards, > > Dian > > > > [1] https://ci.apache.org/projects/flink/flink-docs-master/ops/cli.html > > [2] > > https://ci.apache.org/projects/flink/flink-docs-master/monitoring/rest_api.html#jars-jarid-run > > > > 在 201

How to prevent from launching 2 jobs at the same time

2019-09-22 Thread David Morin
Hi, What is the best way to prevent from launching 2 jobs with the same name concurrently ? Instead of doing a check in the script that starts the Flink job, I would prefer to stop a job if another one with the same name is in progress (Exception or something like that). David

Re: Flink and kerberos

2019-08-29 Thread David Morin
cc_c', partition=7}, > KafkaTopicPartition{topic='gbl_auth_raw_occ_c', partition=14}, > KafkaTopicPartition{topic='gbl_auth_raw_occ_c', partition=13}, > KafkaTopicPartition{topic='gbl_auth_raw_occ_c', partition=12}, > KafkaTopicPartition{topic='

Re: Flink and kerberos

2019-08-29 Thread David Morin
Hello Vishwas, You can use a keytab if you prefer. You generate a keytab for your user and then you can reference it in the Flink configuration. Then this keytab will be handled by Flink in a secure way and TGT will be created based on this keytab. However, that seems to be working. Did you check

Re: Status of the Integration of Flink with Hive

2019-08-12 Thread David Morin
d, official > 1.9 should be out soon > - Hive features are in beta in 1.9 > > Feel free to open tickets if you have feature requests. > > > On Fri, Aug 9, 2019 at 8:00 AM David Morin > wrote: > >> Hi, >> >> I want to connect my Flink streaming job to Hive.

Status of the Integration of Flink with Hive

2019-08-09 Thread David Morin
Hi, I want to connect my Flink streaming job to Hive. At the moment, what is the best way to connect to Hive. Some features seems to be in development. Some really cool features have been described here: https://fr.slideshare.net/BowenLi9/integrating-flink-with-hive-xuefu-zhang-and-bowen-li-seatt

Re: dynamic metric

2019-06-20 Thread David Morin
if you implement a RichFunction: > > getRuntimeContext().getMetricGroup().gauge("MyGauge", new Gauge() > { > @Override > public Integer getValue() { > return valueToExpose; > } > }); > > Cheers, > Till > > On Fri, Jun 21, 2019 at 12:36 A

dynamic metric

2019-06-20 Thread David Morin
Hi, I want to create one metric related to the number of errors but in fact I would like to add some contextual labels ? What is the best way to do that ? gauge ? How to create this kind of metric dynamically during the run of the task (because open method is not possible because too early) ? Than