Hi everybody,
I am using Flink 1.4.2 and periodically my job goes down with the following
exception in logs. Relaunching the job does not help, only restarting the
whole cluster.
Is there a JIRA problem for that? will upgrade to 1.5 help?
java.io.FileNotFoundException:
/tmp/flink-io-20a15b29-183
Hi all,
I am getting similar exception while upgrading from Flink 1.4 to 1.6:
```
06 Feb 2019 14:37:34,080 ERROR
org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Fatal error
occurred in the cluster entrypoint.
java.lang.RuntimeException: org.apache.flink.util.FlinkException: Could not
retr
my guess is that tmp directory got cleaned on your host and Flink couldn't
restore memory state from it upon startup.
Take a look at
https://ci.apache.org/projects/flink/flink-docs-stable/ops/config.html#configuring-temporary-io-directories
article, I think it is relevant
On Thu, Nov 1, 2018 at 8
that's what are You looking for:
> https://issues.apache.org/jira/browse/KAFKA-6221
>
> This issue is connected with Kafka itself rather than Flink.
>
> Best Regards,
> Dom.
>
> wt., 23 paź 2018 o 15:04 Alexander Smirnov
> napisał(a):
>
>> Hi,
>>
>
Hi,
I stumbled upon an exception in the "Exceptions" tab which I could not
explain. Do you know what could cause it? Unfortunately I don't know how to
reproduce it. Do you know if there is a respective JIRA issue for it?
Here's the exception's stack trace:
org.apache.flink.streaming.connectors.k
I think that's because you declared it as transient field.
Move the declaration inside of "open" function to resolve that
On Mon, Oct 22, 2018 at 3:48 PM Ahmad Hassan wrote:
> 2018-10-22 13:46:31,944 INFO org.apache.flink.runtime.taskmanager.Task
> - Window(SlidingProcessingT
You need to us an
> older point to restart.
>
> Best,
> Stefan
>
>
> Am 25.09.2018 um 16:53 schrieb Alexander Smirnov <
> alexander.smirn...@gmail.com>:
>
> Thanks Stefan.
>
> is it only Flink runtime should be updated, or the job should be
> recompiled
ra/browse/FLINK-8836 which
> would also match to your Flink version. I suggest to update to 1.4.3 or
> higher to avoid the issue in the future.
>
> Best,
> Stefan
>
>
> Am 25.09.2018 um 16:37 schrieb Alexander Smirnov <
> alexander.smirn...@gmail.com>:
>
>
I'm getting an exception on job starting from a savepoint. Why that could
happen?
Flink 1.4.2
java.lang.IllegalStateException: Could not initialize operator state
backend.
at
org.apache.flink.streaming.api.operators.AbstractStreamOperator.initOperatorState(AbstractStreamOperator.java:301
Thanks Hequn!
On Thu, 30 Aug 2018 at 04:49, Hequn Cheng wrote:
> Hi Alex,
>
> It seems a bug. There is a discussion here
> <http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Kryo-Exception-td20324.html>
> .
> Best, Hequn
>
> On Wed, Aug 29, 2018
Hi,
A job fell into a restart loop with the following exception. Is it
something known?
What could cause it?
Flink 1.4.2
16 Aug 2018 13:43:00,835 INFO org.apache.flink.runtime.taskmanager.Task -
Source: Custom Source -> (Filter -> Timestamps/Watermarks -> Map, Filter ->
Timestamps/Watermarks ->
Hello,
I have a cluster with multiple jobs running on it. One of the jobs has
checkpoints constantly failing
[image: image.png]
How do I investigate it?
Thank you,
Alex
Hello,
I noticed CPU utilization went high and took a thread dump on the task
manager node. Why would RocksDBMapState.entries() / seek0 call consumes CPU?
It is Flink 1.4.2
"Co-Flat Map (3/4)" #16129 prio=5 os_prio=0 tid=0x7fefac029000
nid=0x338f runnable [0x7feed2002000]
java.lang.Th
d .out are different. Usually, the .log file stores the log
> information output by the log framework. Flink uses slf4j as the log
> interface and supports log4j and logback configurations. The .out file
> stores the STDOUT information. This information is usually output by you
>
Hi,
could you please explain the difference between *.log and *.out files in
Flink?
What information is supposed to be in each of them?
Is "log" a subset of "out"?
How do I setup rotation with gzipping?
Thank you,
Alex
to denote development and stable releases?
Hi Alexey,
I know that Kibana(https://en.wikipedia.org/wiki/Kibana) can show logs from
different servers at one screen. May be this is what you are looking for
Alex
On Mon, May 14, 2018 at 5:17 PM NEKRASSOV, ALEXEI wrote:
> Is there a way to see logs from multiple Task Managers **all in one pl
1.0.2.
>
> As a side note, as far as we know our FlinkKafkaProducer011 works fine
> with Kafka 1.0.x.
>
> Piotrek
>
> On 7 May 2018, at 12:12, Alexander Smirnov
> wrote:
>
> Hi Piotr, using 0.11.0 Kafka version
>
> On Sat, May 5, 2018 at 10:19 AM Piotr Nowoj
rek
>
>
> On 4 May 2018, at 17:55, Alexander Smirnov
> wrote:
>
> Thanks for quick turnaround Stefan, Piotr
>
> This is a rare reproducible issue and I will keep an eye on it
>
> searching on the Stack Overflow I found
> https://stackoverflow.com/questions/433786
the correct Broker (the leader) but to a broker that is just a follower
> and the follower can not execute that request. However, I am not sure what
> causes this in the context of the FlinkKafkaProducer. Maybe Piotr (in CC)
> has an idea?
>
> Best,
> Stefan
>
> Am 04.05.
Hi,
what could cause the following exception?
org.apache.flink.streaming.connectors.kafka.FlinkKafka011Exception: Failed
to send data to Kafka: This server is not the leader for that
topic-partition.
at
org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer011.checkErroneous(FlinkKafkaPro
Hi,
I'm creating kafka producer with timestamps enabled following
instructions at
https://ci.apache.org/projects/flink/flink-docs-release-1.4/dev/connectors/kafka.html#kafka-producer
Optional> customPartitioner = Optional.empty();
FlinkKafkaProducer011 result = new
FlinkKafkaPr
ok, I got it. Barrier-n is an indicator or n-th checkpoint.
My first impression was that barriers are carrying offset information, but
it was wrong.
Thanks for unblocking ;-)
Alex
Hi,
I'm reading documentation about checkpointing:
https://ci.apache.org/projects/flink/flink-docs-master/internals/stream_checkpointing.html
It describes a case, when an operator receives data from all its incoming
streams alongs with barriers. There's also an illustration on that page for
the ca
Hi Fabian,
please share the workarounds, that must be helpful for my case as well
Thank you,
Alex
On Mon, Apr 23, 2018 at 2:14 PM Fabian Hueske wrote:
> Hi Miki,
>
> Sorry for the late response.
> There are basically two ways to implement an enrichment join as in your
> use case.
>
> 1) Keep t
That's absolutely no problem Tzu-Li. Either of them would work. Thank you!
On Thu, Apr 19, 2018 at 4:56 PM Tzu-Li (Gordon) Tai
wrote:
> @Alexander
> Sorry about that, that would be my mistake. I’ll close FLINK-9204 as a
> duplicate and leave my thoughts on FLINK-9155. Thanks for pointing out!
>
Hi,
I have a co-flatmap function which reads data from external DB on specific
events.
The API for the DB layer is homegrown and it uses multiple threads to speed
up reading.
Can it cause any problems if I use the multithreading API in the flatmap1
function? Is it allowed in Flink?
Or, maybe I s
a better approach.
>
> I'm not sure to what extend the source functions (Kafka, Kinesis) support
> such error tracking.
> Adding Gordon to the thread who knows the internals of the connectors.
>
> Best, Fabian
>
> 2018-04-08 17:53 GMT+02:00 Alexander Smirnov >:
>
>
this feature has been implemented in 1.4.0, take a look at
https://issues.apache.org/jira/browse/FLINK-4022
https://ci.apache.org/projects/flink/flink-docs-release-1.4/dev/connectors/kafka.html#kafka-consumers-topic-and-partition-discovery
On Wed, Apr 11, 2018 at 3:33 PM chandresh pancholi <
cha
I've seen similar problem, but it was not a heap size, but Metaspace.
It was caused by a job restarting in a loop. Looks like for each restart,
Flink loads new instance of classes and very soon in runs out of metaspace.
I've created a JIRA issue for this problem, but got no response from the
devel
I have the same question. In case of kafka source, it would be good to know
topic name and offset of the corrupted message for further investigation.
Looks like the only option is to write messages into a log file
On Fri, Apr 6, 2018 at 9:12 PM Elias Levy
wrote:
> I was wondering how are folks t
being completely ignored (regardless of it’s
> value) when user enables checkpointing:
>
> env.enableCheckpointing(5000, CheckpointingMode.EXACTLY_ONCE);
>
> I suspect this is a bug, but I have to confirm it.
>
> Thanks, Piotrek
>
> On 5 Apr 2018, at 12:40, Alexander Smirnov
=2147483647,
delayBetweenRestartAttempts=1) for 43ecfe9cb258b7f624aad9868d306edb.*
2018-04-05 22:38:29,656 INFO
org.apache.flink.runtime.executiongraph.ExecutionGraph- Job
recovers via failover strategy: full graph restart
On Thu, Apr 5, 2018 at 10:35 PM Alexander Smirnov <
alexander.sm
nk
> version. Otherwise ClassNotFoundException usually indicates that
> something is wrong with your dependencies. Maybe you can share your
> pom.xml with us.
>
> Regards,
> Timo
>
> Am 02.04.18 um 13:32 schrieb Alexander Smirnov:
> > I see a lot of messages in flink log like b
Hello,
I've defined restart strategy in flink-conf.yaml as none. WebUI / Job
Manager section confirms that.
But looks like this setting is disregarded.
When I go into job's configuration in the WebUI, in the Execution
Configuration section I can see:
Max. number of execution retries
I see a lot of messages in flink log like below. What's the cause?
02 Apr 2018 04:09:13,554 ERROR
org.apache.kafka.clients.producer.internals.Sender - Uncaught error in
kafka producer I/O thread:
org.apache.kafka.common.KafkaException: Error registering mbean
kafka.producer:type=producer-node-met
Thanks for confirming, Nico 👍🏻
On Wed, Mar 28, 2018 at 3:57 PM Nico Kruber wrote:
> If you refer to the files under the conf folder, these are only used by
> the standalone cluster startup scripts, i.e. bin/start-cluster.sh and
> bin/stop-cluster.sh
>
>
> Nico
>
> On 2
Hi,
are the files needed only on cluster startup stage?
are they only used by bash scripts?
Alex
y happen after 48 hours?
> Is there some indication of a failure in TaskManager log?
>
> If you will be still unable to solve the problem, please provide full
> TaskManager and JobManager logs.
>
> Piotrek
>
> On 21 Mar 2018, at 16:00, Alexander Smirnov
> wrote:
>
&g
. (akka.remote.Remoting)
The host is available, but I don't understand where port number comes from.
Task Manager uses another port (which is printed in logs on startup)
Could you please help to understand why it happens?
Thank you,
Alex
On Wed, Mar 21, 2018 at 4:19 PM Alexander Sm
Hello,
I've assembled a standalone cluster of 3 task managers and 3 job
managers(and 3 ZK) following the instructions at
https://ci.apache.org/projects/flink/flink-docs-release-1.4/ops/deployment/cluster_setup.html
and
https://ci.apache.org/projects/flink/flink-docs-release-1.4/ops/jobmanager_hig
perQuorumPeer does not add any functionality except creating
> ZooKeeper's myid file.
>
> Best,
> Gary
>
> On Wed, Mar 21, 2018 at 12:02 PM, Alexander Smirnov <
> alexander.smirn...@gmail.com> wrote:
>
>> Hi,
>>
>> For standalone cluster config
Hi,
For standalone cluster configuration, is it possible to use vanilla Apache
Zookeeper?
I saw there's a wrapper around it in Flink - FlinkZooKeeperQuorumPeer. Is
it mandatory to use it?
Thank you,
Alex
Thank you so much, it helped!
From: Piotr Nowojski mailto:pi...@data-artisans.com>>
Date: Thursday, October 12, 2017 at 6:00 PM
To: Alexander Smirnov mailto:asmir...@five9.com>>
Cc: "user@flink.apache.org<mailto:user@flink.apache.org>"
mailto:user@flink.apache.org>
Hello All,
I got the following error while attempting to execute a job via command line:
[root@flink01 bin]# ./flink run -c com.five9.stream.PrecomputeJob
/vagrant/flink-precompute-1.0-SNAPSHOT.jar -Xmx2048m -Xms2048m
Cluster configuration: Standalone cluster with JobManager at
flink01.pb.lx-dr
Hi Biplob,
Yes unix timestamp is what I¹m using now.
But the problem is that a time window like '1 day' is defined using
different start-end timestamps for users in different time zones
Let me try to draw it
|1--2-3-4---|
1 and 3 - time frames for European users
2 and 4 - t
Hello everybody,
I’m exploring Flink options to build statistics engine for call center solution.
One thing I’m not sure how to implement.
Currently have the following jobs in the architecture.
Job #1 – is for matching start and end events and calculate durations. Like
having call started and c
Hi,
source data, read from MQ, contains tenant Id.
Is there a way to route messages from particular tenant to particular Flink
node? Is it what can be configured?
Thank you,
Alex
idn't try the approach myself, though.
>
> On Mon, Apr 25, 2016 at 6:50 PM, Alexander Smirnov <
> alexander.smirn...@gmail.com> wrote:
>
>> Hello everybody!
>>
>> my RMQSource function receives string with JSONs in it.
>> Because many operations in Fli
Hello everybody!
my RMQSource function receives string with JSONs in it.
Because many operations in Flink rely on Tuple operations, I think it is a
good idea to convert JSON to Tuple.
I believe this task has been solved already :)
what's the common approach for this conversion?
Thank you,
Alex
fying Flink programs by config files (or graphically) would require
>> a data model, a DataStream/DataSet program generator and probably a code
>> generation component.
>>
>> Best, Fabian
>>
>> 2016-04-22 18:41 GMT+02:00 Alexander Smirnov <
>> alexander.s
Hi guys!
I’m new to Flink, and actually to this mailing list as well :) this is my first
message.
I’m still reading the documentation and I would say Flink is an amazing
system!! Thanks everybody who participated in the development!
The information I didn’t find in the documentation - if it is
52 matches
Mail list logo