Sure, it is already merged as FLINK-10816. Best, Stefan
> On 8. Nov 2018, at 11:53, Shailesh Jain <shailesh.j...@stellapps.com> wrote: > > Thanks a lot for looking into this issue Stefan. > > Could you please let me know the issue ID once you open it? It'll help me > understand the problem better, and also I could do a quick test in our > environment once the issue is resolved. > > Thanks, > Shailesh > > On Wed, Nov 7, 2018, 10:46 PM Till Rohrmann <trohrm...@apache.org > <mailto:trohrm...@apache.org> wrote: > Really good finding Stefan! > > On Wed, Nov 7, 2018 at 5:28 PM Stefan Richter <s.rich...@data-artisans.com > <mailto:s.rich...@data-artisans.com>> wrote: > Hi, > > I think I can already spot the problem: LockableTypeSerializer.duplicate() is > not properly implemented because it also has to call duplicate() on the > element serialiser that is passed into the constructor of the new instance. I > will open an issue and fix the problem. > > Best, > Stefan > >> On 7. Nov 2018, at 17:17, Till Rohrmann <trohrm...@apache.org >> <mailto:trohrm...@apache.org>> wrote: >> >> Hi Shailesh, >> >> could you maybe provide us with an example program which is able to >> reproduce this problem? This would help the community to better debug the >> problem. It looks not right and might point towards a bug in Flink. Thanks a >> lot! >> >> Cheers, >> Till >> >> On Tue, Oct 30, 2018 at 9:10 AM Dawid Wysakowicz <dwysakow...@apache.org >> <mailto:dwysakow...@apache.org>> wrote: >> This is some problem with serializing your events using Kryo. I'm adding >> Gordon to cc, as he was recently working with serializers. He might give you >> more insights what is going wrong. >> >> Best, >> >> Dawid >> >> On 25/10/2018 05:41, Shailesh Jain wrote: >>> Hi Dawid, >>> >>> I've upgraded to flink 1.6.1 and rebased by changes against the tag 1.6.1, >>> the only commit on top of 1.6 is this: >>> https://github.com/jainshailesh/flink/commit/797e3c4af5b28263fd98fb79daaba97cabf3392c >>> >>> <https://github.com/jainshailesh/flink/commit/797e3c4af5b28263fd98fb79daaba97cabf3392c> >>> >>> I ran two separate identical jobs (with and without checkpointing enabled), >>> I'm hitting a ArrayIndexOutOfBoundsException (and sometimes NPE) only when >>> checkpointing (HDFS backend) is enabled, with the below stack trace. >>> >>> I did see a similar problem with different operators here >>> (http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/IndexOutOfBoundsException-on-deserialization-after-updating-to-1-6-1-td23933.html >>> >>> <http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/IndexOutOfBoundsException-on-deserialization-after-updating-to-1-6-1-td23933.html>). >>> Is this a known issue which is getting addressed? >>> >>> Any ideas on what could be causing this? >>> >>> Thanks, >>> Shailesh >>> >>> >>> 2018-10-24 17:04:13,365 INFO org.apache.flink.runtime.taskmanager.Task >>> - SelectCepOperatorMixedTime (1/1) - >>> SelectCepOperatorMixedTime (1/1) (3d984b7919342a3886593401088ca2cd) >>> switched from RUNNING to FAILED. >>> org.apache.flink.util.FlinkRuntimeException: Failure happened in filter >>> function. >>> at org.apache.flink.cep.nfa.NFA.createDecisionGraph(NFA.java:731) >>> at org.apache.flink.cep.nfa.NFA.computeNextStates(NFA.java:541) >>> at org.apache.flink.cep.nfa.NFA.doProcess(NFA.java:284) >>> at org.apache.flink.cep.nfa.NFA.process(NFA.java:220) >>> at >>> org.apache.flink.cep.operator.AbstractKeyedCEPPatternOperator.processEvent(AbstractKeyedCEPPatternOperator.java:379) >>> at >>> org.apache.flink.cep.operator.AbstractKeyedCEPPatternMixedTimeApproachOperator.processElement(AbstractKeyedCEPPatternMixedTimeApproachOperator.java:45) >>> at >>> org.apache.flink.streaming.runtime.io.StreamInputProcessor.processInput(StreamInputProcessor.java:202) >>> at >>> org.apache.flink.streaming.runtime.tasks.OneInputStreamTask.run(OneInputStreamTask.java:105) >>> at >>> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:300) >>> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:711) >>> at java.lang.Thread.run(Thread.java:748) >>> Caused by: org.apache.flink.util.WrappingRuntimeException: >>> java.lang.ArrayIndexOutOfBoundsException: -1 >>> at >>> org.apache.flink.cep.nfa.sharedbuffer.SharedBuffer.lambda$materializeMatch$1(SharedBuffer.java:305) >>> at java.util.HashMap.computeIfAbsent(HashMap.java:1127) >>> at >>> org.apache.flink.cep.nfa.sharedbuffer.SharedBuffer.materializeMatch(SharedBuffer.java:301) >>> at >>> org.apache.flink.cep.nfa.sharedbuffer.SharedBuffer.materializeMatch(SharedBuffer.java:291) >>> at >>> org.apache.flink.cep.nfa.NFA$ConditionContext.getEventsForPattern(NFA.java:811) >>> at >>> com.stellapps.contrakcep.flink.patterns.critical.AgitatorMalfunctionAfterChillingPattern$1.filter(AgitatorMalfunctionAfterChillingPattern.java:70) >>> at >>> com.stellapps.contrakcep.flink.patterns.critical.AgitatorMalfunctionAfterChillingPattern$1.filter(AgitatorMalfunctionAfterChillingPattern.java:62) >>> at org.apache.flink.cep.nfa.NFA.checkFilterCondition(NFA.java:742) >>> at org.apache.flink.cep.nfa.NFA.createDecisionGraph(NFA.java:716) >>> ... 10 more >>> Caused by: java.lang.ArrayIndexOutOfBoundsException: -1 >>> at com.esotericsoftware.kryo.util.IntArray.add(IntArray.java:61) >>> at com.esotericsoftware.kryo.Kryo.readReferenceOrNull(Kryo.java:800) >>> at com.esotericsoftware.kryo.Kryo.readObject(Kryo.java:655) >>> at >>> org.apache.flink.api.java.typeutils.runtime.kryo.KryoSerializer.copy(KryoSerializer.java:231) >>> at >>> org.apache.flink.cep.nfa.sharedbuffer.Lockable$LockableTypeSerializer.copy(Lockable.java:120) >>> at >>> org.apache.flink.cep.nfa.sharedbuffer.Lockable$LockableTypeSerializer.copy(Lockable.java:95) >>> at >>> org.apache.flink.api.common.typeutils.base.MapSerializer.copy(MapSerializer.java:113) >>> at >>> org.apache.flink.api.common.typeutils.base.MapSerializer.copy(MapSerializer.java:49) >>> at >>> org.apache.flink.runtime.state.heap.CopyOnWriteStateTable.get(CopyOnWriteStateTable.java:287) >>> at >>> org.apache.flink.runtime.state.heap.CopyOnWriteStateTable.get(CopyOnWriteStateTable.java:311) >>> at >>> org.apache.flink.runtime.state.heap.HeapMapState.get(HeapMapState.java:85) >>> at >>> org.apache.flink.runtime.state.UserFacingMapState.get(UserFacingMapState.java:47) >>> at >>> org.apache.flink.cep.nfa.sharedbuffer.SharedBuffer.lambda$materializeMatch$1(SharedBuffer.java:303) >>> ... 18 more >>> >>> On Fri, Sep 28, 2018 at 11:00 AM Shailesh Jain <shailesh.j...@stellapps.com >>> <mailto:shailesh.j...@stellapps.com>> wrote: >>> Hi Dawid, >>> >>> Thanks for your time on this. The diff should have pointed out only the top >>> 3 commits, but since it did not, it is possible I did not rebase my branch >>> against 1.4.2 correctly. I'll check this out and get back to you if I hit >>> the same issue again. >>> >>> Thanks again, >>> Shailesh >>> >>> On Thu, Sep 27, 2018 at 1:00 PM Dawid Wysakowicz <dwysakow...@apache.org >>> <mailto:dwysakow...@apache.org>> wrote: >>> Hi Shailesh, >>> >>> I am afraid it is gonna be hard to help you, as this branch differs >>> significantly from 1.4.2 release (I've done diff across your branch and >>> tag/release-1.4.2). Moreover the code in the branch you've provided still >>> does not correspond to the lines in the exception you've posted previously. >>> Could you check if the problem occurs on vanilla flink as well? >>> >>> Best, >>> >>> Dawid >>> >>> >>> On 27/09/18 08:22, Shailesh Jain wrote: >>>> Hi Dawid, >>>> >>>> Yes, it is version 1.4.2. We are running vanilla flink, but have added a >>>> couple of changes in the CEP operator specifically (top 3 commits here: >>>> https://github.com/jainshailesh/flink/commits/poc_on_1.4.2 >>>> <https://github.com/jainshailesh/flink/commits/poc_on_1.4.2>). Changes >>>> I've made to CEP operators do not touch the checkpointing path, just >>>> overloading the operator for a specific way of handling event time. >>>> >>>> We are hitting this in production, so I'm not sure it'll be feasible to >>>> move to 1.6.0 immediately, but eventually yes. >>>> >>>> Thanks, >>>> Shailesh >>>> >>>> On Wed, Sep 26, 2018 at 5:44 PM Dawid Wysakowicz <dwysakow...@apache.org >>>> <mailto:dwysakow...@apache.org>> wrote: >>>> Hi Shailesh, >>>> >>>> Are you sure you are using version 1.4.2? Do you run a vanilla flink, or >>>> have you introduced some changes? I am asking cause the lines in >>>> stacktrace does not align with the source code for 1.4.2. >>>> >>>> Also it is a different exception than the one in the issue you've linked, >>>> so if it is a problem than it is definitely a different one. Last thing I >>>> would recommend upgrading to the newest version, as we rewritten the >>>> SharedBuffer implementation in 1.6.0. >>>> >>>> Best, >>>> >>>> Dawid >>>> >>>> >>>> On 26/09/18 13:50, Shailesh Jain wrote: >>>>> Hi, >>>>> >>>>> I think I've hit this same issue on a 3 node standalone cluster (1.4.2) >>>>> using HDFS (2.8.4) as state backend. >>>>> >>>>> 2018-09-26 17:07:39,370 INFO org.apache.flink.runtime.taskmanager.Task >>>>> - Attempting to fail task externally SelectCepOperator >>>>> (1/1) (3bec4aa1ef2226c4e0c5ff7b3860d340). >>>>> 2018-09-26 17:07:39,370 INFO org.apache.flink.runtime.taskmanager.Task >>>>> - SelectCepOperator (1/1) >>>>> (3bec4aa1ef2226c4e0c5ff7b3860d340) switched from RUNNING to FAILED. >>>>> AsynchronousException{java.lang.Exception: Could not materialize >>>>> checkpoint 6 for operator SelectCepOperator (1/1).} >>>>> at >>>>> org.apache.flink.streaming.runtime.tasks.StreamTask$AsyncCheckpointRunnable.run(StreamTask.java:948) >>>>> at >>>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) >>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >>>>> at >>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) >>>>> at >>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) >>>>> at java.lang.Thread.run(Thread.java:748) >>>>> Caused by: java.lang.Exception: Could not materialize checkpoint 6 for >>>>> operator SelectCepOperator (1/1). >>>>> ... 6 more >>>>> Caused by: java.util.concurrent.ExecutionException: >>>>> java.lang.NullPointerException >>>>> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >>>>> at java.util.concurrent.FutureTask.get(FutureTask.java:192) >>>>> at >>>>> org.apache.flink.util.FutureUtil.runIfNotDoneAndGet(FutureUtil.java:43) >>>>> at >>>>> org.apache.flink.streaming.runtime.tasks.StreamTask$AsyncCheckpointRunnable.run(StreamTask.java:894) >>>>> ... 5 more >>>>> Suppressed: java.lang.Exception: Could not properly cancel managed >>>>> keyed state future. >>>>> at >>>>> org.apache.flink.streaming.api.operators.OperatorSnapshotResult.cancel(OperatorSnapshotResult.java:91) >>>>> at >>>>> org.apache.flink.streaming.runtime.tasks.StreamTask$AsyncCheckpointRunnable.cleanup(StreamTask.java:976) >>>>> at >>>>> org.apache.flink.streaming.runtime.tasks.StreamTask$AsyncCheckpointRunnable.run(StreamTask.java:939) >>>>> ... 5 more >>>>> Caused by: java.util.concurrent.ExecutionException: >>>>> java.lang.NullPointerException >>>>> at java.util.concurrent.FutureTask.report(FutureTask.java:122) >>>>> at java.util.concurrent.FutureTask.get(FutureTask.java:192) >>>>> at >>>>> org.apache.flink.util.FutureUtil.runIfNotDoneAndGet(FutureUtil.java:43) >>>>> at >>>>> org.apache.flink.runtime.state.StateUtil.discardStateFuture(StateUtil.java:66) >>>>> at >>>>> org.apache.flink.streaming.api.operators.OperatorSnapshotResult.cancel(OperatorSnapshotResult.java:89) >>>>> ... 7 more >>>>> Caused by: java.lang.NullPointerException >>>>> at >>>>> org.apache.flink.cep.nfa.SharedBuffer$SharedBufferSerializer.serialize(SharedBuffer.java:954) >>>>> at >>>>> org.apache.flink.cep.nfa.SharedBuffer$SharedBufferSerializer.serialize(SharedBuffer.java:825) >>>>> at >>>>> org.apache.flink.cep.nfa.NFA$NFASerializer.serialize(NFA.java:888) >>>>> at >>>>> org.apache.flink.cep.nfa.NFA$NFASerializer.serialize(NFA.java:820) >>>>> at >>>>> org.apache.flink.runtime.state.heap.CopyOnWriteStateTableSnapshot.writeMappingsInKeyGroup(CopyOnWriteStateTableSnapshot.java:196) >>>>> at >>>>> org.apache.flink.runtime.state.heap.HeapKeyedStateBackend$1.performOperation(HeapKeyedStateBackend.java:390) >>>>> at >>>>> org.apache.flink.runtime.state.heap.HeapKeyedStateBackend$1.performOperation(HeapKeyedStateBackend.java:339) >>>>> at >>>>> org.apache.flink.runtime.io.async.AbstractAsyncCallableWithResources.call(AbstractAsyncCallableWithResources.java:75) >>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >>>>> at >>>>> org.apache.flink.util.FutureUtil.runIfNotDoneAndGet(FutureUtil.java:40) >>>>> at >>>>> org.apache.flink.streaming.runtime.tasks.StreamTask$AsyncCheckpointRunnable.run(StreamTask.java:894) >>>>> ... 5 more >>>>> [CIRCULAR REFERENCE:java.lang.NullPointerException] >>>>> >>>>> Any ideas on why I'm hitting this especially when this >>>>> (https://issues.apache.org/jira/browse/FLINK-7756 >>>>> <https://issues.apache.org/jira/browse/FLINK-7756>) says it has been >>>>> fixed in 1.4.2 ? >>>>> >>>>> On Sat, Nov 4, 2017 at 12:34 AM Federico D'Ambrosio >>>>> <federico.dambro...@smartlab.ws <mailto:federico.dambro...@smartlab.ws>> >>>>> wrote: >>>>> Thank you very much for your steady response, Kostas! >>>>> >>>>> Cheers, >>>>> Federico >>>>> >>>>> 2017-11-03 16:26 GMT+01:00 Kostas Kloudas <k.klou...@data-artisans.com >>>>> <mailto:k.klou...@data-artisans.com>>: >>>>> Hi Federico, >>>>> >>>>> Thanks for trying it out! >>>>> Great to hear that your problem was fixed! >>>>> >>>>> The feature freeze for the release is going to be next week, and I would >>>>> expect 1 or 2 more weeks testing. >>>>> So I would say in 2.5 weeks. But this is of course subject to potential >>>>> issues we may find during testing. >>>>> >>>>> Cheers, >>>>> Kostas >>>>> >>>>>> On Nov 3, 2017, at 4:22 PM, Federico D'Ambrosio >>>>>> <federico.dambro...@smartlab.ws <mailto:federico.dambro...@smartlab.ws>> >>>>>> wrote: >>>>>> >>>>>> Hi Kostas, >>>>>> >>>>>> I just tried running the same job with 1.4-SNAPSHOT for 10 minutes and >>>>>> it didn't crash, so that was the same underlying issue of the JIRA you >>>>>> linked. >>>>>> >>>>>> Do you happen to know when it's expected the 1.4 stable release? >>>>>> >>>>>> Thank you very much, >>>>>> Federico >>>>>> >>>>>> 2017-11-03 15:25 GMT+01:00 Kostas Kloudas <k.klou...@data-artisans.com >>>>>> <mailto:k.klou...@data-artisans.com>>: >>>>>> Perfect! thanks a lot! >>>>>> >>>>>> Kostas >>>>>> >>>>>>> On Nov 3, 2017, at 3:23 PM, Federico D'Ambrosio >>>>>>> <federico.dambro...@smartlab.ws >>>>>>> <mailto:federico.dambro...@smartlab.ws>> wrote: >>>>>>> >>>>>>> Hi Kostas, >>>>>>> >>>>>>> yes, I'm using 1.3.2. I'll try the current master and I'll get back to >>>>>>> you. >>>>>>> >>>>>>> 2017-11-03 15:21 GMT+01:00 Kostas Kloudas <k.klou...@data-artisans.com >>>>>>> <mailto:k.klou...@data-artisans.com>>: >>>>>>> Hi Federico, >>>>>>> >>>>>>> I assume that you are using Flink 1.3, right? >>>>>>> >>>>>>> In this case, in 1.4 we have fixed a bug that seems similar to your >>>>>>> case: >>>>>>> https://issues.apache.org/jira/browse/FLINK-7756 >>>>>>> <https://issues.apache.org/jira/browse/FLINK-7756> >>>>>>> >>>>>>> Could you try the current master to see if it fixes your problem? >>>>>>> >>>>>>> Thanks, >>>>>>> Kostas >>>>>>> >>>>>>>> On Nov 3, 2017, at 3:12 PM, Federico D'Ambrosio >>>>>>>> <federico.dambro...@smartlab.ws >>>>>>>> <mailto:federico.dambro...@smartlab.ws>> wrote: >>>>>>>> >>>>>>>> Could not find id for entry: >>>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Federico D'Ambrosio >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Federico D'Ambrosio >>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> Federico D'Ambrosio >>>> >>> >