Hi Dawid, Thank you for your reply. I'm out for the next few days, so I hope you don't mind me cc'ing my team in here. We all really appreciate you and the rest of the people monitoring the mailing list.
We've only seen this SharedBuffer problem in production, after sending around 20 GB of data through. In the Flink UI, we see the checkpoint status as: *Checkpoint failed: Checkpoint Coordinator is suspending.* It then tries to restore the last previously succeeded checkpoint, but cannot and throws the SharedBuffer exception. Our state size is around 200MB when it fails. Unfortunately, the platform we are running our cluster on only supports up to Flink 1.5. We will continue trying to find a reproducible example for you. I have found a few other people with a similar problem (attached at the bottom), but none seem to have been resolved. Our CEP setup looks generally like this: Pattern<AlertEvent, ?> pattern = Pattern.<AlertEvent>begin("alertOne") .where(new SimpleCondition<Event>() { @Override public boolean filter(AlertEvent event) { return event.level > 0; } }) .next("alertTwo").subtype(AlertEvent.class) .where(new IterativeCondition<SubEvent>() { @Override public boolean filter(AlertEvent subEvent, Context<AlertEvent> ctx) throws Exception { double surprisePercent = subEvent.surprise(); Iterable<AlertEvent> previousEvents = ctx.getEventsForPattern("alertOne"); for (AlertEvent prevEvent : previousEvents) { double prevSurprisePercent = prevEvent.surprise(); if (prevSurprisePercent > surprisePercent) { return false; } } return true; } }) .next("alertThree").subtype(AlertEvent.class) .where(new IterativeCondition<SubEvent>() { @Override public boolean filter(AlertEvent subEvent, Context<AlertEvent> ctx) throws Exception { double surprisePercent = subEvent.surprise(); Iterable<AlertEvent> previousEvents = ctx.getEventsForPattern("alertTwo"); for (AlertEvent prevEvent : previousEvents) { double prevSurprisePercent = prevEvent.surprise(); if (prevSurprisePercent > surprisePercent) { return false; } } return true; }}); PatternStream<AlertEvent> alertPatternStream = CEP.pattern(alertEvents, pattern); DataStream<Alert> confirmedAlerts = alertPatternStream .select(new PatternSelectFunction<AlertEvent, Alert> { private static final long serialVersionUID = 1L; @Override public Alert select(Map<String, List<AlertEvent>> patternIn) { List<AlertEvent> alertEventList = new ArrayList<>(); // Create an alert composed of three escalating events alertEventList.add(patternIn.get("alertOne").get(0)); alertEventList.add(patternIn.get("alertTwo").get(0)); alertEventList.add(patternIn.get("alertThree").get(0)); Alert confirmedAlert = new Alert(alertEventList); return confirmedAlert; } }) .uid("confirmedAlerts") .returns(Alert.class) .keyBy("id"); Once again thank you, Austin - https://stackoverflow.com/questions/36917569/flink-streaming-cep-could-not-find-previous-shared-buffer-entry-with-key - https://mail-archives.apache.org/mod_mbox/flink-user/201604.mbox/%3CCAC27z=nd0sx_ogbh0vuvzwetcke2t40tbxht5huh8uyzdte...@mail.gmail.com%3E - https://mail-archives.apache.org/mod_mbox/flink-user/201702.mbox/%3CCAGr9p8CzT3=cr+cOas=3k0becmiybfv+kt1fxv_rf8xabtu...@mail.gmail.com%3E On Thu, Nov 1, 2018 at 4:44 AM Dawid Wysakowicz <dwysakow...@apache.org> wrote: > Hi Austin, > > Could you elaborate a bit more what do you mean by "after a checkpoint > fails", what is the reason why checkpoint fails? Would it be possible for > you to prepare some reproducible example for that problem? Finally, I would > also recommend trying out Flink 1.6.x, as we reworked the underlying > structure for CEP - SharedBuffer. > > Best, > > Dawid > On 30/10/2018 20:59, Austin Cawley-Edwards wrote: > > Following up, we are using Flink 1.5.0 and Flink-CEP 2.11. > > Thanks, > Austin > > On Tue, Oct 30, 2018 at 3:58 PM Austin Cawley-Edwards < > austin.caw...@gmail.com> wrote: > >> Hi there, >> >> We have a streaming application that uses CEP processing but are getting >> this error fairly frequently after a checkpoint fails, though not sure if it >> is related. We have implemented both `hashCode` and `equals()` using >> `Objects.hash(...properties)` and basic equality, respectively. Has anyone >> seen this before using CEP? >> >> Here is the full exception: >> >> java.lang.RuntimeException: Exception occurred while processing valve output >> watermark: >> at >> org.apache.flink.streaming.runtime.io.StreamInputProcessor$ForwardingValveOutputHandler.handleWatermark(StreamInputProcessor.java:265) >> at >> org.apache.flink.streaming.runtime.streamstatus.StatusWatermarkValve.findAndOutputNewMinWatermarkAcrossAlignedChannels(StatusWatermarkValve.java:189) >> at >> org.apache.flink.streaming.runtime.streamstatus.StatusWatermarkValve.inputWatermark(StatusWatermarkValve.java:111) >> at >> org.apache.flink.streaming.runtime.io.StreamInputProcessor.processInput(StreamInputProcessor.java:184) >> at >> org.apache.flink.streaming.runtime.tasks.OneInputStreamTask.run(OneInputStreamTask.java:103) >> at >> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:306) >> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:703) >> at java.lang.Thread.run(Thread.java:748) >> Caused by: java.lang.IllegalStateException: Could not find previous entry >> with key: alertOne, value: >> {"totalDocumentCount":12,"alertLevel":3,"id":"entity:medium:13790","predicted":4.220480902392199,"timestamp":1539700396000} >> and timestamp: 1539700799999. This can indicate that either you did not >> implement the equals() and hashCode() methods of your input elements >> properly or that the element belonging to that entry has been already pruned. >> at org.apache.flink.cep.nfa.SharedBuffer.put(SharedBuffer.java:107) >> at org.apache.flink.cep.nfa.NFA.computeNextStates(NFA.java:566) >> at org.apache.flink.cep.nfa.NFA.process(NFA.java:252) >> at >> org.apache.flink.cep.operator.AbstractKeyedCEPPatternOperator.processEvent(AbstractKeyedCEPPatternOperator.java:332) >> at >> org.apache.flink.cep.operator.AbstractKeyedCEPPatternOperator.lambda$onEventTime$0(AbstractKeyedCEPPatternOperator.java:235) >> at >> java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1382) >> at >> java.util.stream.ReferencePipeline$Head.forEachOrdered(ReferencePipeline.java:590) >> at >> org.apache.flink.cep.operator.AbstractKeyedCEPPatternOperator.onEventTime(AbstractKeyedCEPPatternOperator.java:234) >> at >> org.apache.flink.streaming.api.operators.HeapInternalTimerService.advanceWatermark(HeapInternalTimerService.java:288) >> at >> org.apache.flink.streaming.api.operators.InternalTimeServiceManager.advanceWatermark(InternalTimeServiceManager.java:108) >> at >> org.apache.flink.streaming.api.operators.AbstractStreamOperator.processWatermark(AbstractStreamOperator.java:734) >> at >> org.apache.flink.streaming.runtime.io.StreamInputProcessor$ForwardingValveOutputHandler.handleWatermark(StreamInputProcessor.java:262) >> >> Best, >> >> Austin >> >>