What does enableObjectReuse exactly do?

2021-05-07 Thread
I wrote a streaming job with scala, using only immutable case class. Is it safe to enable object reuse? Will it get benefits from enabling object reuse? I reached to documents but they cover neither streaming cases nor immutable data structures.

Re: Watermark won't advance in ProcessFunction

2019-10-28 Thread
, for the first program, you have > not registered the event time timer(though context.timerService. > registerEventTimeTimer) and also there is also no onTimer logic defined > to process it. > > 在 2019年10月28日,下午4:01,杨力 写道: > > It seems to be the case. But when I use timeWindow or

Re: Watermark won't advance in ProcessFunction

2019-10-28 Thread
ermark is still not generated yet and I think that's the reason why > it's Long.MinValue. > > Regards, > Dian > > 在 2019年10月28日,上午11:59,杨力 写道: > > I'm going to sort elements in a PriorityQueue and set up timers at > (currentWatermark + 1), following the ins

Watermark won't advance in ProcessFunction

2019-10-27 Thread
I'm going to sort elements in a PriorityQueue and set up timers at (currentWatermark + 1), following the instructions in https://ci.apache.org/projects/flink/flink-docs-release-1.9/dev/stream/operators/process_function.html#timer-coalescing . However, it seems that context.timerService().currentWa

Is it possible to register a custom TypeInfoFactory without using an annotation?

2019-08-26 Thread
I'd like to provide a custom serializer for a POJO class. But that class cannot be modified so it's not possible to add a @TypeInfo annotation to it. Are there any other ways to register one?

Re: Job xxx not found exception when starting Flink program in Local

2018-11-17 Thread
The web ui keeps job ids in web browsers. 徐涛 于2018年11月18日周日 上午8:12写道: > > Hi Chesnay, > Yes, it is. Later I notice before the later started cluster I have use > ./bin/start-cluster.sh to start a local cluster, and after I > ./bin/stop-cluster.sh to stop the cluster, the problem is gone. > But ho

Re: JMX Configuration: Missing Job Related Beans

2018-09-24 Thread
They are provided in taskmanagers. Sayat Satybaldiyev 于 2018年9月24日周一 下午6:38写道: > Dear all, > > While configuring JMX with Flink, I don't see some bean metrics that > belongs to the job, in particular, the number in/out records per operator. > I've checked REST API and those numbers provided ther

Re: CombinableGroupReducer says The Iterable can be iterated over only once

2018-09-14 Thread
A java.util.Iterable is expected to provide iterators again and again. On Fri, Sep 14, 2018 at 10:53 PM Alejandro Alcalde wrote: > Hello all, > > I am trying to replicate the code in the Docs ( > https://ci.apache.org/projects/flink/flink-docs-release-1.6/dev/batch/dataset_transformations.html#c

Re: Can rocksDBBackend handle rescaling?

2018-09-13 Thread
A checkpoint cannot be used for rescaling. You should take a savepoint, stop the application, and resume from the savepoint with a new parallelism. https://ci.apache.org/projects/flink/flink-docs-release-1.6/ops/state/savepoints.html On Fri, Sep 14, 2018 at 1:50 AM 廖嘉逸 wrote: > Hi, all > >

Re: Flink 1.6 Job fails with IllegalStateException: Buffer pool is destroyed.

2018-09-07 Thread
to config at least INFO level for logger in flink, and > currently the messages are so limited for debugging the problem. > > Best, > Zhijiang > > ------ > 发件人:杨力 > 发送时间:2018年9月7日(星期五) 17:21 > 收件人:Zhijiang(wangz

Flink 1.6 Job fails with IllegalStateException: Buffer pool is destroyed.

2018-09-06 Thread
Hi all, I am encountering a weird problem when running flink 1.6 in yarn per-job clusters. The job fails in about half an hour after it starts. Related logs is attached as an imange. This piece of log comes from one of the taskmanagers. There are not any other related log lines. No ERROR-level log

Re: Flink 1.6 ExecutionJobVertex.getTaskInformationOrBlobKey OutOfMemoryError

2018-08-13 Thread
equn > > [1] > https://ci.apache.org/projects/flink/flink-docs-master/ops/deployment/yarn_setup.html#run-a-flink-job-on-yarn > > On Mon, Aug 13, 2018 at 10:25 PM, 杨力 wrote: > >> I used to runFlink SQL in streaming mode with more than 70 sqls in >> version 1.4. With

Flink 1.6 ExecutionJobVertex.getTaskInformationOrBlobKey OutOfMemoryError

2018-08-13 Thread
I used to runFlink SQL in streaming mode with more than 70 sqls in version 1.4. With so many sqls loaded, akka.framesize has to be set to 200 MB to submit the job. When I am trying to run the job with flink 1.6.0, the HTTP-based job submission works perfectly but an OutOfMemoryError is thrown when

Jobs running on a yarn per-job cluster fail to restart when a task manager is lost

2018-05-22 Thread
Hi, I am running a streaming job without checkpointing enabled. A failute rate restart strategy have been set with StreamExecutionEvironment.setRestartStrategy. When a task manager is lost because of memory problems, the job manager try to restart the job without launching a new task manager, and

User-defined aggregation function and parallelism

2018-04-15 Thread
I am running flink SQL in streaming mode and implemented a UDAGG, which is used in keyed HOP windows. But I found that the throughput decreases dramatically when the function is used. Does UDAGG run in parallell? Or does it run only in one thread? Regards, Bill

Re: Access logs for a running Flink app in YARN cluster

2018-04-12 Thread
Maybe you can get them from yarn with rest API. Tao Xia 于 2018年4月13日周五 上午8:09写道: > Any good way to get access container logs from a running Flink app in YARN > cluster in EMR? > You can view the logs through YARN UI. But cannot programmatically access > it and send to other services. > The log a

Re: subuquery about flink sql

2018-04-03 Thread
You should add a column TUMBLE_ROWTIME(t, INTERVAL '60' SECOND) AS t to the select part of your subquery. 韩宁宁 <453673...@qq.com> 于 2018年4月3日周二 下午3:34写道: > Thank you for your reply. > > I think the table registration no problem。I guess it's a question of > subquery。 > It's no problem to execut

Re: bad data output

2018-03-29 Thread
You can use a split operator, generating 2 streams. Darshan Singh 于 2018年3月30日周五 上午2:53写道: > Hi > > I have a dataset which has almost 99% of correct data. As of now if say > some data is bad I just ignore it and log it and return only correct data. > I do this inside a map function. > > The part

Re: Extremely large job serialization produced by union operator

2018-03-14 Thread
n the > worker nodes. The size of the code depends on the number of fields in the > schema. > Operators and code are not shared across queries. > > Best, Fabian > > 2018-03-09 23:36 GMT+01:00 杨力 : > >> Thank you for your response. It occurs both in a standalone clus

Re: Extremely large job serialization produced by union operator

2018-03-09 Thread
tup? Logs could be > helpful as well. > > Piotrek > > > On 9 Mar 2018, at 11:00, 杨力 wrote: > > > > I wrote a flink-sql app with following topography. > > > > KafkaJsonTableSource -> SQL -> toAppendStream -> Map -> > JDBCAppend

Re: Flink 1.4 SQL API Streaming TableException

2018-03-09 Thread
To use a field in a table as timestamp, it must be declared as a rowtime attribute for the table. 1) Call env.setStreamTimeCharacteristic(TimeCharacteristic.EventTime). 2) Call withRowtimeAttribute on KafkaJsonTableSourceBuilder. Reference: 1. https://ci.apache.org/projects/flink/flink-docs-relea

Extremely large job serialization produced by union operator

2018-03-09 Thread
I wrote a flink-sql app with following topography. KafkaJsonTableSource -> SQL -> toAppendStream -> Map -> JDBCAppendTableSink KafkaJsonTableSource -> SQL -> toAppendStream -> Map -> JDBCAppendTableSink ... KafkaJsonTableSource -> SQL -> toAppendStream -> Map -> JDBCAppendTableSink I have a dozen

Re: Using time window with SQL nested query

2018-03-05 Thread
https://ci.apache.org/projects/flink/flink-docs-release-1.4/dev/table/sql.html#selecting-group-window-start-and-end-timestamps > > > > Am 3/5/18 um 11:26 AM schrieb 杨力: > > >

Using time window with SQL nested query

2018-03-05 Thread
I tried to use write a nested query with HOP window in a streaming environment. Table `source` consists of 3 column, a, b, and timestamp. SELECT a FROM (SELECT a, COUNT(*) FROM source GROUP BY HOP(timestamp, ..., ...), a, b) GROUP BY HOP(timestamp, ..., ...), a HAVING ... And flink throws an exc