We did look into fixing it ourselves, but decided that migrating to the datastream api, not using CEP, was more fruitful overall for us. Unfortunately, I don't have a good answer for you. The bug from a non-contributors stand point appears pretty deep in the codebase, but the authors are best ones to speak to this.
On Tue, Jul 6, 2021 at 7:01 PM Li Jim <lishijun121...@126.com> wrote: > > Hi, Jameson > Thanks very much for replying , I am really struggling on this. > I am using flowId as my keys, which means they will be matched and never use > again. > This seems like the scenario 2. I didn't know it is not fixed yet. > thank you again and do you have any solutions ? > > On 2021/07/07 01:47:00, Aeden Jameson <aeden.jame...@gmail.com> wrote: > > Hi Li, > > > > How big is your keyspace? Had a similar problem which turns out to > > be scenario 2 in this issue > > https://issues.apache.org/jira/browse/FLINK-19970. Looks like the bug > > in scenario 1 got fixed by scenario 2 did not. There's more detail in > > this thread, > > http://deprecated-apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/Flink-cep-checkpoint-size-td44141.html#a44168 > > . Hope that helps > > > > On Tue, Jul 6, 2021 at 12:44 AM Li Jim <lishijun121...@126.com> wrote: > > > > > > I am using Flink CEP to do some performance tests. > > > > > > Flink version 1.13.1. > > > > > > below is the sql: > > > > > > INSERT INTO to_kafka > > > SELECT bizName, wdName, wdValue , zbValue , flowId FROM kafka_source > > > MATCH_RECOGNIZE > > > ( > > > PARTITION BY flow_id > > > ORDER BY proctime > > > MEASURES A.biz_name as bizName, A.wd_name as wdName, A.wd_value as > > > wdValue, MAP[ > > > A.zb_name, A.zb_value, > > > B.zb_name, B.zb_value > > > ] as zbValue, A.flow_id as flowId > > > ONE ROW PER MATCH > > > AFTER MATCH SKIP PAST LAST ROW > > > PATTERN ( A B ) WITHIN INTERVAL '10' SECOND > > > DEFINE > > > B AS B.flow_id = A.flow_id > > > ); > > > > > > I add the 'within clause' to avoid state growing, but it does not work at > > > all. > > > > > > the checkpoint size is growing fast. > > > > > > I am using rocksdb incremental mode. > > > > > > > > > > > > -- > > Cheers, > > Aeden > >