[ANNOUNCE] Apache Flink Stateful Functions 2.2.1 released

2020-11-11 Thread Tzu-Li (Gordon) Tai
The Apache Flink community released the first bugfix release of the
Stateful Functions (StateFun) 2.2 series, version 2.2.1.

This release fixes a critical bug that causes restoring a Stateful
Functions cluster from snapshots (checkpoints or savepoints) to fail under
certain conditions.

*We strongly recommend all users to upgrade to this version.*

*Please check out the release announcement for details on upgrading to
2.2.1:*https://flink.apache.org/news/2020/11/11/release-statefun-2.2.1.html

The release is available for download at:
https://flink.apache.org/downloads.html

Maven artifacts for Stateful Functions can be found at:
https://search.maven.org/search?q=g:org.apache.flink%20statefun

Python SDK for Stateful Functions published to the PyPI index can be found
at:
https://pypi.org/project/apache-flink-statefun/

Official Dockerfiles for building Stateful Functions Docker images can be
found at:
https://github.com/apache/flink-statefun-docker

Alternatively, Ververica has volunteered to make Stateful Function's images
available for the community via their public Docker Hub registry:
https://hub.docker.com/r/ververica/flink-statefun

The full release notes are available in Jira:
https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12349291

We would like to thank all contributors of the Apache Flink community who
made this release possible!

Cheers,
Gordon


[jira] [Created] (FLINK-20086) Add documentation for the open method in UserDefinedFunction

2020-11-11 Thread Dian Fu (Jira)
Dian Fu created FLINK-20086:
---

 Summary: Add documentation for the open method in 
UserDefinedFunction
 Key: FLINK-20086
 URL: https://issues.apache.org/jira/browse/FLINK-20086
 Project: Flink
  Issue Type: Improvement
  Components: API / Python, Documentation
Reporter: Dian Fu
 Fix For: 1.12.0, 1.11.3


According to the questions asked by PyFlink users so far, many users are not 
aware that there is a open method in UserDefinedFunction where they could 
perform initialization work. This method is especially useful for ML users 
where they could do ML mode initialization.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [DISCUSS] Releasing Apache Flink 1.11.3

2020-11-11 Thread Khachatryan Roman
Hi,

I'd like FLINK-20079 [1] to be merged into 1.11 and included in 1.11.3.

[1] https://issues.apache.org/jira/browse/FLINK-20079

Regards,
Roman


On Tue, Nov 10, 2020 at 8:21 AM Xintong Song  wrote:

> Thanks for the notice, Dian.
>
> Thank you~
>
> Xintong Song
>
>
>
> On Tue, Nov 10, 2020 at 10:19 AM Dian Fu  wrote:
>
> > Hi Xintong,
> >
> > I want to bring one more issue to your attention [1]. The test case
> > UnalignedCheckpointCompatibilityITCase.test failed several times in the
> > last nightly test of release-1.11. We need to figure out if it's just an
> > instable test or caused by recent changes.
> >
> > [1] https://issues.apache.org/jira/browse/FLINK-20065
> >
> > > 在 2020年11月10日,上午9:24,Xintong Song  写道:
> > >
> > > Thanks for the replies.
> > >
> > > Thank you~
> > >
> > > Xintong Song
> > >
> > >
> > >
> > > On Tue, Nov 10, 2020 at 1:09 AM Becket Qin 
> wrote:
> > >
> > >> Hi Xintong,
> > >>
> > >> Thanks for driving the release. Just want to sync up on the FLIP-27
> > >> backporting. Stephan and I are still trying to backport a bunch of
> > patches
> > >> of Source to 1.11.3. Including:
> > >>
> > >> [FLINK-19698][connector/common] Add a close() method to the
> SplitReader.
> > >> [FLINK-19717] SourceReaderBase.pollNext may return END_OF_INPUT if
> > >> SplitReader.fetch throws
> > >> [FLINK-19535] [connector/common] Avoid failing a job multiple times in
> > >> SourceCoordinator.
> > >> [FLINK-19265] [FLINK-20049][core] Source API final adjustments.
> > >>
> > >> and a few more fixes.
> > >>
> > >> We are currently trying to fix them in 1.12 first so it might take a
> > little
> > >> longer to backport them to 1.11.3. I think it will probably take us a
> > few
> > >> more days to finish the backport. So that would roughly be the end of
> > this
> > >> week.
> > >>
> > >> Thanks,
> > >>
> > >> Jiangjie (Becket) Qin
> > >>
> > >>
> > >>
> > >>
> > >> On Mon, Nov 9, 2020 at 9:57 PM Till Rohrmann 
> > wrote:
> > >>
> > >>> Yes, I've downgraded FLINK-19816 to critical.
> > >>>
> > >>> Cheers,
> > >>> Till
> > >>>
> > >>> On Mon, Nov 9, 2020 at 10:19 AM Xintong Song 
> > >>> wrote:
> > >>>
> >  Thanks for the notice, Till.
> > 
> >  I just checked and found FLINK-20033 is already fixed. Shall we also
> >  downgrade FLINK-19816 to `Critical`?
> > 
> >  Thank you~
> > 
> >  Xintong Song
> > 
> > 
> > 
> >  On Mon, Nov 9, 2020 at 4:42 PM Till Rohrmann 
> > >>> wrote:
> > 
> > > I would like to bring one more critical issue to your attention
> which
> > >>> is
> > > FLINK-20033 [1]. I believe that this issue is actually causing what
> > >> has
> > > been reported in FLINK-19816 [2]. I hope to have it fixed by the
> end
> > >> of
> > > today. Once FLINK-20033 is fixed, I think that we don't have to
> block
> > >>> the
> > > release on FLINK-19816.
> > >
> > > [1] https://issues.apache.org/jira/browse/FLINK-20033
> > > [2] https://issues.apache.org/jira/browse/FLINK-19816
> > >
> > > Cheers,
> > > Till
> > >
> > > On Mon, Nov 9, 2020 at 4:05 AM Xintong Song  >
> >  wrote:
> > >
> > >> Hi devs,
> > >>
> > >> I'd like to provide an update on the progress of preparing release
> > > 1.11.3.
> > >>
> > >> *Blockers*
> > >> We currently have 3 remaining blockers. (3 resolved and 1 emerged
> > > compared
> > >> to last week)
> > >>
> > >> - [FLINK-19698] Add close() method and onCheckpointComplete() to
> > >> the
> > >> Source.
> > >> The issue has been fixed on the master branch. It's currently
> > >> blocked
> >  on
> > >> the FLIP-27 backportings to backport it to the 1.11 branch.
> > >>
> > >> - [FLINK-19717] SourceReaderBase.pollNext may return END_OF_INPUT
> > >> if
> > >> SplitReader.fetch throws
> > >> A PR has been opened and reviewed. From the discussions on the PR,
> > >> it
> > > looks
> > >> close to mergeable.
> > >>
> > >> - [FLINK-19816] Flink restored from a wrong checkpoint (a very old
> > >>> one
> > > and
> > >> not the last completed one)
> > >> This is a newly emerged blocker and Matthias is working on it.
> > >>
> > >> *Test Instabilities*
> > >> We currently have 27 test instabilities[1].
> > >> AFAIK, none of them are as serious as to block the 1.11.3 release.
> > >>
> > >> *FLIP-27 Backprotings*
> > >>
> > >> I noticed that there's no jira issues opened on the FLIP-27
> > >>> backporting
> > >> efforts, which is part of the major efforts planned for the 1.11.3
> > > release,
> > >> making it hard to track the progress.
> > >>
> > >>
> > >> @Stephan and @Becket, could you please share the updates on the
> > > backporting
> > >> efforts? How is the progress and when are the efforts expected to
> > >> be
> > >> finished? It would be appreciated and helpful if we can have a
> jira
> > > ticket
> > >

Re: [VOTE] FLIP-145: Support SQL windowing table-valued function (2nd)

2020-11-11 Thread Timo Walther

+1 (binding)

Thanks,
Timo

On 11.11.20 07:14, Pengcheng Liu wrote:

+1 (binding)

Jark Wu  于2020年11月11日周三 上午10:13写道:


+1 (binding)

On Tue, 10 Nov 2020 at 14:59, Jark Wu  wrote:


Hi all,

There is new feedback on the FLIP-145. So I would like to start a new

vote

for FLIP-145 [1],
which has been discussed and reached consensus in the discussion thread
[2].

The vote will be open until 15:00 (UTC+8) 13th Nov. (72h), unless there

is

an objection or not enough votes.

Best,
Jark

[1]:


https://cwiki.apache.org/confluence/display/FLINK/FLIP-145%3A+Support+SQL+windowing+table-valued+function

[2]:


http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-FLIP-145-Support-SQL-windowing-table-valued-function-td45269.html










[jira] [Created] (FLINK-20087) CheckpointCoordinator waits until all tasks finish initialization of states to trigger checkpoint

2020-11-11 Thread Jiayi Liao (Jira)
Jiayi Liao created FLINK-20087:
--

 Summary: CheckpointCoordinator waits until all tasks finish 
initialization of states to trigger checkpoint
 Key: FLINK-20087
 URL: https://issues.apache.org/jira/browse/FLINK-20087
 Project: Flink
  Issue Type: Improvement
  Components: Runtime / Checkpointing
Affects Versions: 1.9.0
Reporter: Jiayi Liao


{{CheckpointCoordinator}} will start triggering checkpoint after all tasks send 
RUNNING status to JobMaster. However, the {{initializeState}} could be 
time-expensive(a few minutes), for example the task needs to rescale with 
multiple old tasks' rocksdb data. During this process, the triggering of 
checkpoints are meaningless. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: Re: [VOTE] FLIP-145: Support SQL windowing table-valued function (2nd)

2020-11-11 Thread 刘大龙

+1

> -原始邮件-
> 发件人: "Timo Walther" 
> 发送时间: 2020-11-11 18:55:06 (星期三)
> 收件人: dev@flink.apache.org
> 抄送: 
> 主题: Re: [VOTE] FLIP-145: Support SQL windowing table-valued function (2nd)
> 
> +1 (binding)
> 
> Thanks,
> Timo
> 
> On 11.11.20 07:14, Pengcheng Liu wrote:
> > +1 (binding)
> > 
> > Jark Wu  于2020年11月11日周三 上午10:13写道:
> > 
> >> +1 (binding)
> >>
> >> On Tue, 10 Nov 2020 at 14:59, Jark Wu  wrote:
> >>
> >>> Hi all,
> >>>
> >>> There is new feedback on the FLIP-145. So I would like to start a new
> >> vote
> >>> for FLIP-145 [1],
> >>> which has been discussed and reached consensus in the discussion thread
> >>> [2].
> >>>
> >>> The vote will be open until 15:00 (UTC+8) 13th Nov. (72h), unless there
> >> is
> >>> an objection or not enough votes.
> >>>
> >>> Best,
> >>> Jark
> >>>
> >>> [1]:
> >>>
> >> https://cwiki.apache.org/confluence/display/FLINK/FLIP-145%3A+Support+SQL+windowing+table-valued+function
> >>> [2]:
> >>>
> >> http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-FLIP-145-Support-SQL-windowing-table-valued-function-td45269.html
> >>>
> >>
> >


[jira] [Created] (FLINK-20088) [Kinesis][Polling] Issue using Polling consumer at timestamp with empty shard

2020-11-11 Thread Danny Cranmer (Jira)
Danny Cranmer created FLINK-20088:
-

 Summary: [Kinesis][Polling] Issue using Polling consumer at 
timestamp with empty shard
 Key: FLINK-20088
 URL: https://issues.apache.org/jira/browse/FLINK-20088
 Project: Flink
  Issue Type: Improvement
  Components: Connectors / Kinesis
Reporter: Danny Cranmer
 Fix For: 1.12.0


*Background*

The consumer fails when a Polling record publisher uses a timestamp sentinel 
starting position and the first record batch is empty. This is because the 
consumer tries to recalculate the start position from the timestamp sentinel, 
this operation is not supported.

*Reproduction Steps*

Setup an application consuming from Kinesis with following properties and 
consume from an empty shard:
{code:java}
String format = "-MM-dd'T'HH:mm:ss";
String date = new SimpleDateFormat(format).format(new Date());

consumerConfig.setProperty(ConsumerConfigConstants.STREAM_INITIAL_TIMESTAMP, 
date);
consumerConfig.setProperty(ConsumerConfigConstants.STREAM_TIMESTAMP_DATE_FORMAT,
 format);
consumerConfig.setProperty(ConsumerConfigConstants.STREAM_INITIAL_POSITION, 
"AT_TIMESTAMP"); {code}
*Error*
{code:java}
Exception in thread "main" 
org.apache.flink.runtime.client.JobExecutionException: Job execution 
failed.Exception in thread "main" 
org.apache.flink.runtime.client.JobExecutionException: Job execution failed. at 
org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:147)
 at 
org.apache.flink.runtime.minicluster.MiniClusterJobClient.lambda$getJobExecutionResult$2(MiniClusterJobClient.java:119)
 at java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:616) 
at 
java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:591)
 at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488) 
at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975) 
at 
org.apache.flink.runtime.rpc.akka.AkkaInvocationHandler.lambda$invokeRpc$0(AkkaInvocationHandler.java:229)
 at 
java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:774)
 at 
java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:750)
 at 
java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:488) 
at java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:1975) 
at 
org.apache.flink.runtime.concurrent.FutureUtils$1.onComplete(FutureUtils.java:996)
 at akka.dispatch.OnComplete.internal(Future.scala:264) at 
akka.dispatch.OnComplete.internal(Future.scala:261) at 
akka.dispatch.japi$CallbackBridge.apply(Future.scala:191) at 
akka.dispatch.japi$CallbackBridge.apply(Future.scala:188) at 
scala.concurrent.impl.CallbackRunnable.run$$$capture(Promise.scala:36) at 
scala.concurrent.impl.CallbackRunnable.run(Promise.scala) at 
org.apache.flink.runtime.concurrent.Executors$DirectExecutionContext.execute(Executors.java:74)
 at scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44) 
at scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252) 
at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:572) at 
akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:22)
 at 
akka.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:21)
 at scala.concurrent.Future$$anonfun$andThen$1.apply(Future.scala:436) at 
scala.concurrent.Future$$anonfun$andThen$1.apply(Future.scala:435) at 
scala.concurrent.impl.CallbackRunnable.run$$$capture(Promise.scala:36) at 
scala.concurrent.impl.CallbackRunnable.run(Promise.scala) at 
akka.dispatch.BatchingExecutor$AbstractBatch.processBatch(BatchingExecutor.scala:55)
 at 
akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply$mcV$sp(BatchingExecutor.scala:91)
 at 
akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply(BatchingExecutor.scala:91)
 at 
akka.dispatch.BatchingExecutor$BlockableBatch$$anonfun$run$1.apply(BatchingExecutor.scala:91)
 at scala.concurrent.BlockContext$.withBlockContext(BlockContext.scala:72) at 
akka.dispatch.BatchingExecutor$BlockableBatch.run(BatchingExecutor.scala:90) at 
akka.dispatch.TaskInvocation.run(AbstractDispatcher.scala:40) at 
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(ForkJoinExecutorConfigurator.scala:44)
 at akka.dispatch.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at 
akka.dispatch.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) 
at akka.dispatch.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) at 
akka.dispatch.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)Caused
 by: org.apache.flink.runtime.JobException: Recovery is suppressed by 
NoRestartBackoffTimeStrategy at 
org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(Executio

[ANNOUNCE] Weekly Community Update 2020/44-45

2020-11-11 Thread Konstantin Knauf
Dear community,

two weeks have passed again and I am happy two share another update with
news on Flink 1.12, Flink 1.11.3 and the release of Stateful Functions
2.2.1. As everyone has been finishing the last bit and pieces of Flink
1.12, there are only a handful of new initiatives to cover this time
including a so-called hybrid source and incremental checkpointing for the
heap-based statebackends.

Flink Development
==

* [releases] The feature freeze for Flink 1.12 happened on Monday and a
first non-voting/testing release candidate has been published. [1] The
community is collecting (manual) testing tasks in the wiki [2].

* [releases] There are still a few blockers to resolve before a first
release candidate for Flink 1.11.3 is published. [3]

* [releases] Stateful Functions 2.2.0 experiences a critical bug that
causes restore from checkpoints or savepoints to fail in certain situations
(FLINK-19692). The proper fix will be included in Flink 1.11.3.  Since
Flink 1.11.3 still takes a few days, Gordon proposed to release Stateful
Functions 2.2.1 right away, that already fixes the issues when the
framework version across snapshot creation and restore is the same. The
release has already been approved and will be announced shortly. [4,5]

* [sql] Jark has updated FLIP-145 after a round of offline discussions. The
new windowing syntax will now also support session windows, propagate the
window time as a time attribute and the FLIP proposes to deprecate the
current GROUP BY window aggregation syntax. A new vote has been started
based on the recent changes to the FLIP. [6,7]

* [connectors] Nicholas Jiang has published a FLIP to support "Hybrid
Sources". A Hybrid Source consists of multiple regular sources that are
read from one after the other. Hybrid sources aim to make
reprocessing/backfilling of data easier if the data is already distributed
over multiple systems (e.g. last 14 days in Kafka, history in S3). [8]

* [statebackends] Roman has published FLIP-151 to support incremental
snapshotting for the heap-based state backend. Currently, incremental
snapshotting is only supported by the RocksDBStatebackend. The
HeapStatebackend is still preferable in a few situations and support for
incremental checkpointing would overcome its largest limitation (besides
limiting the state size to memory). [9]

* [docker] In contrast to what I wrote would become the outcome of the
discussion to make jemalloc the default memory allocator in the Apache
Flink docker image, jemalloc will indeed become the default. [10]

[1]
http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/ANNOUNCE-Feature-Freeze-of-Flink-1-12-tp46418.html
[2]
https://cwiki.apache.org/confluence/display/FLINK/1.12+Release+-+Community+Testing
[3]
http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-Releasing-Apache-Flink-1-11-3-tp45989.html
[4]
http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-Releasing-StateFun-hotfix-version-2-2-1-tp46239.html
[5]
http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/VOTE-Apache-Flink-Stateful-Functions-2-2-1-release-candidate-1-tp46303.html
[6]
http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-FLIP-145-Support-SQL-windowing-table-valued-function-tp45269.html
[7]
http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/VOTE-FLIP-145-Support-SQL-windowing-table-valued-function-2nd-tp46452.html
[8]
http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-FLIP-151-Incremental-snapshots-for-heap-based-state-backend-tp46284.html
[9]
http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/VOTE-Adopt-jemalloc-as-default-memory-allocator-in-docker-image-tp46382.html

Notable Bugs
==

* [FLINK-19970][1.11.2] There might be a state leak in the CEP library that
leads to an ever growing state size. I don't think this has been reproduced
yet, but for anyone using the CEP library this is an interesting one to
watch. [10]
* [FLINK-20033] [1.11.2] [1.10.2] When a Job Master is stopped (which
happens if the Dispatcher loses leadership) the current execution of its
Job is failed, which can lead to data loss if the number of restarts are
depleted. Fixed for 1.11.3 & 1.10.3. [11]

[10] https://issues.apache.org/jira/browse/FLINK-19970
[11] https://issues.apache.org/jira/browse/FLINK-20033

Events, Blog Posts, Misc
===

* Congxian Qiu is now an Apache Flink Committer. Congratulations! [12]

* Xianghu Wang has published a blog post outlining Apache Hudi's transition
away from a Spark-only and towards a Flink-first architecture. [13]

* Fred Teunissen & Erik de Nooij describe their solution to deal with
event-time skew when ingesting data from heterogeneous Kafka partitions
within one Flink Job on the Ververica Blog. [14]

[12]
http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/ANNOUNCE-New-Apache-Flink-Committer-Congxian-Qiu-tp46123p46208.html
[13] http://hudi.apache.org/blog/apache-hudi-meets-

[jira] [Created] (FLINK-20089) Avro format fails to serialize map value with null keys

2020-11-11 Thread hailong wang (Jira)
hailong wang created FLINK-20089:


 Summary: Avro format fails to serialize map value with null keys
 Key: FLINK-20089
 URL: https://issues.apache.org/jira/browse/FLINK-20089
 Project: Flink
  Issue Type: Bug
  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
Affects Versions: 1.12.0
Reporter: hailong wang


When AvroRowDataSerializationSchema serializing map data with null keys, it 
will throw NPE.
This is the same as 
[FLINK-19912|https://issues.apache.org/jira/projects/FLINK/issues/FLINK-19912]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-20090) Expose SlotId / SlotSharingGroup in Rest API

2020-11-11 Thread Maximilian Michels (Jira)
Maximilian Michels created FLINK-20090:
--

 Summary: Expose SlotId / SlotSharingGroup in Rest API 
 Key: FLINK-20090
 URL: https://issues.apache.org/jira/browse/FLINK-20090
 Project: Flink
  Issue Type: New Feature
  Components: Runtime / REST
Reporter: Maximilian Michels


There is no information on slot sharing exposed via the Rest API which would be 
useful to monitor how tasks are assigned to task slots.

We could include the SlotId in {{SubtaskExecutionAttemptDetailsInfo}} and 
provide a list of slots in {{TaskManagersInfo}}.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-20091) Introduce avro.ignore-parse-errors for AvroRowDataDeserializationSchema

2020-11-11 Thread hailong wang (Jira)
hailong wang created FLINK-20091:


 Summary: Introduce avro.ignore-parse-errors  for 
AvroRowDataDeserializationSchema
 Key: FLINK-20091
 URL: https://issues.apache.org/jira/browse/FLINK-20091
 Project: Flink
  Issue Type: Improvement
  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
Affects Versions: 1.12.0
Reporter: hailong wang


Introduce avro.ignore-parse-errors to allow users to skip rows with parsing 
errors instead of failing when deserializing avro format data.
This is useful when there are dirty data, for without this option, users can 
not skip the dirty row.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-20092) Multi-thread Flink compilation

2020-11-11 Thread Jira
Maciej Bryński created FLINK-20092:
--

 Summary: Multi-thread Flink compilation
 Key: FLINK-20092
 URL: https://issues.apache.org/jira/browse/FLINK-20092
 Project: Flink
  Issue Type: Bug
Reporter: Maciej Bryński


I'd like to use maven -T option when compiling flink.
{code:java}
 mvn -T 2C clean install -D"scala-2.12" -DskipTests{code}
Unfortunately my build is stuck on:
{code:java}
[INFO] --- maven-shade-plugin:3.2.1:shade (shade-flink) @ 
flink-fs-hadoop-shaded ---
[INFO] Including org.apache.hadoop:hadoop-common:jar:3.1.0 in the shaded jar.
[INFO] Including org.apache.hadoop:hadoop-annotations:jar:3.1.0 in the shaded 
jar.
[INFO] Including com.google.guava:guava:jar:11.0.2 in the shaded jar.
[INFO] Including commons-io:commons-io:jar:2.7 in the shaded jar.
[INFO] Including commons-collections:commons-collections:jar:3.2.2 in the 
shaded jar.
[INFO] Including commons-logging:commons-logging:jar:1.1.3 in the shaded jar.
[INFO] Including commons-lang:commons-lang:jar:2.6 in the shaded jar.
[INFO] Including commons-beanutils:commons-beanutils:jar:1.9.3 in the shaded 
jar.
[INFO] Including org.apache.commons:commons-configuration2:jar:2.1.1 in the 
shaded jar.
[INFO] Including org.apache.commons:commons-lang3:jar:3.3.2 in the shaded jar.
[INFO] Including com.google.re2j:re2j:jar:1.1 in the shaded jar.
[INFO] Including org.apache.hadoop:hadoop-auth:jar:3.1.0 in the shaded jar.
[INFO] Including org.apache.htrace:htrace-core4:jar:4.1.0-incubating in the 
shaded jar.
[INFO] Including com.fasterxml.jackson.core:jackson-databind:jar:2.10.1 in the 
shaded jar.
[INFO] Including com.fasterxml.jackson.core:jackson-annotations:jar:2.10.1 in 
the shaded jar.
[INFO] Including com.fasterxml.jackson.core:jackson-core:jar:2.10.1 in the 
shaded jar.
[INFO] Including org.codehaus.woodstox:stax2-api:jar:3.1.4 in the shaded jar.
[INFO] Including com.fasterxml.woodstox:woodstox-core:jar:5.0.3 in the shaded 
jar.
[INFO] Including org.apache.flink:force-shading:jar:1.12-SNAPSHOT in the shaded 
jar.
[INFO] No artifact matching filter io.netty:netty
[WARNING] Discovered module-info.class. Shading will break its strong 
encapsulation.
[WARNING] Discovered module-info.class. Shading will break its strong 
encapsulation.
[WARNING] Discovered module-info.class. Shading will break its strong 
encapsulation.
[INFO] Replacing original artifact with shaded artifact.
[INFO] Replacing 
/home/maverick/flink/flink-filesystems/flink-fs-hadoop-shaded/target/flink-fs-hadoop-shaded-1.12-SNAPSHOT.jar
 with 
/home/maverick/flink/flink-filesystems/flink-fs-hadoop-shaded/target/flink-fs-hadoop-shaded-1.12-SNAPSHOT-shaded.jar
[INFO] Dependency-reduced POM written at: 
/home/maverick/flink/flink-filesystems/flink-fs-hadoop-shaded/target/dependency-reduced-pom.xml
{code}
Can we make flink compilation working with multiple maven threads ?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: Register processing time timers when Operator.close() is called

2020-11-11 Thread Aljoscha Krettek

Hi!

This is an interesting topic and we recently created a Jira issue about 
this: https://issues.apache.org/jira/browse/FLINK-18647.


In Beam we even have a workaround for this: 
https://github.com/apache/beam/blob/0c01636fc8610414859d946cb93eabc904123fc8/runners/flink/src/main/java/org/apache/beam/runners/flink/translation/wrappers/streaming/DoFnOperator.java#L581


Maybe it's time to finally address this in Flink as well.

Best,
Aljoscha


On 11.11.20 01:02, Boyuan Zhang wrote:

Hi team,

I'm writing my custom Operator as a high fan-out operation and I use
processing time timers to defer processing some inputs When timers are
firing, the operator will continue to process the deferred elements. One
typical use case for my Operator is like:
ImpulseOperator -> my Operator -> downstream where the watermark of
ImpulseOperator advances to MAX_TIMESTAMP immediately.

One problem I have is that after my operator.close() is called, it's still
possible for me to set processing time timers and wait for these timers to
be fired. But it seems like Flink pauses invoking processing timers once
one operator.close() is called in the new version. I'm curious why Flink
decides to do so and any workaround I can do for my operator?

Thanks for your help!





[jira] [Created] (FLINK-20093) Create a download page for all optional sql client components

2020-11-11 Thread Dawid Wysakowicz (Jira)
Dawid Wysakowicz created FLINK-20093:


 Summary: Create a download page for all optional sql client 
components
 Key: FLINK-20093
 URL: https://issues.apache.org/jira/browse/FLINK-20093
 Project: Flink
  Issue Type: Improvement
  Components: Documentation, Table SQL / Ecosystem
Reporter: Dawid Wysakowicz
Assignee: Dawid Wysakowicz
 Fix For: 1.12.0


It would be nice to have a single page that lists all optional binaries for sql 
client. We could link such a page from the "Optional components" section in the 
https://flink.apache.org/downloads.html



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-20094) Make JobMaster.jobStatusChanged directly called by JobManagerJobStatusListener

2020-11-11 Thread Till Rohrmann (Jira)
Till Rohrmann created FLINK-20094:
-

 Summary: Make JobMaster.jobStatusChanged directly called by 
JobManagerJobStatusListener
 Key: FLINK-20094
 URL: https://issues.apache.org/jira/browse/FLINK-20094
 Project: Flink
  Issue Type: Improvement
  Components: Runtime / Coordination
Affects Versions: 1.12.0
Reporter: Till Rohrmann
 Fix For: 1.13.0


Since the {{ExecutionGraph}} is only modified by the main thread, we no longer 
need to splice the {{JobManagerJobStatusListener}} callbacks back into the 
{{JobMaster's}} main thread using {{runAsync}}. Instead it should be possible 
to directly call {{JobMaster.jobStatusChanged}}.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-20095) SchedulerBase.stopWithSavepoint schedules operation after ExecutionGraph terminates

2020-11-11 Thread Till Rohrmann (Jira)
Till Rohrmann created FLINK-20095:
-

 Summary: SchedulerBase.stopWithSavepoint schedules operation after 
ExecutionGraph terminates
 Key: FLINK-20095
 URL: https://issues.apache.org/jira/browse/FLINK-20095
 Project: Flink
  Issue Type: Improvement
  Components: Runtime / Coordination
Affects Versions: 1.12.0
Reporter: Till Rohrmann
 Fix For: 1.13.0


While looking into FLINK-20065, we realized that we cannot implement the change 
described in FLINK-20094 because the {{SchedulerBase}} schedules an 
asynchronous operation after the termination of the {{ExecutionGraph}}. This 
asynchronous operation needs to complete before the savepoint path is returned.

In order to solve FLINK-20094, we either need to keep track of these operations 
and suspend the {{SchedulerBase}} termination until all operations have 
completed or we might be to return the savepoint path without depending it on 
the completion of the {{handleAsync}} here: 
https://github.com/apache/flink/blob/master/flink-runtime/src/main/java/org/apache/flink/runtime/scheduler/SchedulerBase.java#L921.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-20096) Clean up PyFlink documentation

2020-11-11 Thread Seth Wiesman (Jira)
Seth Wiesman created FLINK-20096:


 Summary: Clean up PyFlink documentation
 Key: FLINK-20096
 URL: https://issues.apache.org/jira/browse/FLINK-20096
 Project: Flink
  Issue Type: Bug
  Components: Documentation
Reporter: Seth Wiesman
Assignee: Seth Wiesman
 Fix For: 1.12.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-20097) Race conditions in InputChannel.ChannelStatePersister

2020-11-11 Thread Roman Khachatryan (Jira)
Roman Khachatryan created FLINK-20097:
-

 Summary: Race conditions in InputChannel.ChannelStatePersister
 Key: FLINK-20097
 URL: https://issues.apache.org/jira/browse/FLINK-20097
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Checkpointing, Runtime / Network
Affects Versions: 1.12.0
Reporter: Roman Khachatryan
Assignee: Roman Khachatryan
 Fix For: 1.12.0


In InputChannel.ChannelStatePersister, stopPersisting() and checkForBarrier() 
always update pendingCheckpointBarrierId, potentially overwriting newer id (or 
BARRIER_RECEIVED value) with an old one.


For stopPersisting(), consider a case:
 # Two consecutive UC barriers arrive at the same channel (1st being stale at 
some point)
 # In RemoteInputChannel.onBuffer, netty thread updates 
pendingCheckpointBarrierId to BARRIER_RECEIVED
 # Task thread processes the 1st barrier and triggers a checkpoint
Task thread processes the 2nd barrier and aborts 1st checkpoint, calling 
stopPersisting() from UC controller and setting pendingCheckpointBarrierId to 
CHECKPOINT_COMPLETED
 # Task thread starts 2nd checkpoint and calls startPersisting() setting 
pendingCheckpointBarrierId to 2
 # now new buffers have a chance to be included in the 2nd checkpoint (though 
they belong to the next one)

 

For pendingCheckpointBarrierId(), consider an input gate with two channels A 
and B and two barriers 1 and 2:
 # Channel A receives both barriers, channel B receives nothing yet
 # Task thread processes both barriers on A, eventually triggering 2nd 
checkpoint
 # Channel A state is now BARRIER_RECEIVED, channel B - pending (with id=2)
 # Channel B receives the 1st barrier and becomes BARRIER_RECEIVED
 # No buffers in B between barriers 1 and 2 will be included in the checkpoint 
 # Channel B receives the 2nd barrier which will eventually conclude the 
checkpoint

 

I see a solution in doing an action only if passed checkpointId >= 
pendingCheckpointId. For that, a separate field will be needed to hold the 
status (RECEIVED/COMPLETED/PENDING). The class isn't thread-safe so it 
shouldn't be a problem.
 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-20098) Don't add flink-connector-files to flink-dist

2020-11-11 Thread Aljoscha Krettek (Jira)
Aljoscha Krettek created FLINK-20098:


 Summary: Don't add flink-connector-files to flink-dist
 Key: FLINK-20098
 URL: https://issues.apache.org/jira/browse/FLINK-20098
 Project: Flink
  Issue Type: Bug
  Components: API / DataStream
Reporter: Aljoscha Krettek
Assignee: Aljoscha Krettek
 Fix For: 1.12.0


We currently add both {{flink-connector-files}} and {{flink-connector-base}} to 
{{flink-dist}}. 

This implies, that users should use the dependency like this:
{code}

org.apache.flink
flink-connector-files
${project.version}
provided

{code}
which differs from other connectors where users don't need to specify 
{{provided}}.

Also, {{flink-connector-files}} has {{flink-connector-base}} as a provided 
dependency, which means that examples that use this dependency will not run 
out-of-box in IntelliJ because transitive provided dependencies will not be 
considered.

I propose to just remove the dependencies from {{flink-dist}} and let users use 
the File Connector like any other connector.

I believe the initial motivation for "providing" the File Connector in 
{{flink-dist}} was to allow us to use the File Connector under the hood in 
methods such as {{StreamExecutionEnvironment.readFile(...)}}. We could decide 
to deprecate and remove those methods or re-add the File Connector as an 
explicit (non-provided) dependency again in the future.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: Register processing time timers when Operator.close() is called

2020-11-11 Thread Boyuan Zhang
Thanks, Aljoscha!

Manually draining processing time timers during operator.close() is my
current workaround as well. It's just not efficient for me since I may set
the processing time timer for the callback after 5 mins but now I need to
fire them immediately.

https://issues.apache.org/jira/browse/FLINK-18647 is really helpful and
looking forward to the solution.

Thanks for your help!


On Wed, Nov 11, 2020 at 8:13 AM Aljoscha Krettek 
wrote:

> Hi!
>
> This is an interesting topic and we recently created a Jira issue about
> this: https://issues.apache.org/jira/browse/FLINK-18647.
>
> In Beam we even have a workaround for this:
>
> https://github.com/apache/beam/blob/0c01636fc8610414859d946cb93eabc904123fc8/runners/flink/src/main/java/org/apache/beam/runners/flink/translation/wrappers/streaming/DoFnOperator.java#L581
>
> Maybe it's time to finally address this in Flink as well.
>
> Best,
> Aljoscha
>
>
> On 11.11.20 01:02, Boyuan Zhang wrote:
> > Hi team,
> >
> > I'm writing my custom Operator as a high fan-out operation and I use
> > processing time timers to defer processing some inputs When timers are
> > firing, the operator will continue to process the deferred elements. One
> > typical use case for my Operator is like:
> > ImpulseOperator -> my Operator -> downstream where the watermark of
> > ImpulseOperator advances to MAX_TIMESTAMP immediately.
> >
> > One problem I have is that after my operator.close() is called, it's
> still
> > possible for me to set processing time timers and wait for these timers
> to
> > be fired. But it seems like Flink pauses invoking processing timers once
> > one operator.close() is called in the new version. I'm curious why Flink
> > decides to do so and any workaround I can do for my operator?
> >
> > Thanks for your help!
> >
>
>


[jira] [Created] (FLINK-20099) HeapStateBackend checkpoint error hidden under cryptic message

2020-11-11 Thread Nico Kruber (Jira)
Nico Kruber created FLINK-20099:
---

 Summary: HeapStateBackend checkpoint error hidden under cryptic 
message
 Key: FLINK-20099
 URL: https://issues.apache.org/jira/browse/FLINK-20099
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Checkpointing, Runtime / State Backends
Affects Versions: 1.11.2
Reporter: Nico Kruber
 Attachments: Screenshot_20201112_001331.png

When the memory state back-end hits a certain size, it fails to permit 
checkpoints. Even though a very detailed exception is thrown at its source, 
this is neither logged nor shown in the UI:
 * Logs just contain:

{code:java}
00:06:41.462 [jobmanager-future-thread-14] INFO  
org.apache.flink.runtime.checkpoint.CheckpointCoordinator - Decline checkpoint 
2 by task 8eb303cd3196310cb2671212f4ed013c of job 
c9b7a410bd3143864ca23ba89595d878 at 6a73bcf2-46b6-4735-a616-fdf09ff1471c @ 
localhost (dataPort=-1).
{code}
 * UI: (also see the attached Screenshot_20201112_001331.png)

{code:java}
Failure Message: The job has failed.
{code}
-> this isn't even true: the job is still running fine!

 

Debugging into {{PendingCheckpoint#abort()}} reveals that the causing exception 
is actually still in there but the detailed information from it is just never 
used.
 For reference, this is what is available there and should be logged or shown:
{code:java}
java.lang.Exception: Could not materialize checkpoint 2 for operator aggregates 
-> (Sink: sink-agg-365, Sink: sink-agg-180, Sink: sink-agg-45, Sink: 
sink-agg-30) (4/4).
at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.handleExecutionException(AsyncCheckpointRunnable.java:191)
at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.run(AsyncCheckpointRunnable.java:138)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.util.concurrent.ExecutionException: java.io.IOException: Size 
of the state is larger than the maximum permitted memory-backed state. 
Size=6122737 , maxSize=5242880 . Consider using a different state backend, like 
the File System State backend.
at java.util.concurrent.FutureTask.report(FutureTask.java:122)
at java.util.concurrent.FutureTask.get(FutureTask.java:192)
at 
org.apache.flink.runtime.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:479)
at 
org.apache.flink.streaming.api.operators.OperatorSnapshotFinalizer.(OperatorSnapshotFinalizer.java:50)
at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.run(AsyncCheckpointRunnable.java:102)
... 3 more
Caused by: java.io.IOException: Size of the state is larger than the maximum 
permitted memory-backed state. Size=6122737 , maxSize=5242880 . Consider using 
a different state backend, like the File System State backend.
at 
org.apache.flink.runtime.state.memory.MemCheckpointStreamFactory.checkSize(MemCheckpointStreamFactory.java:64)
at 
org.apache.flink.runtime.state.memory.MemCheckpointStreamFactory$MemoryCheckpointOutputStream.closeAndGetBytes(MemCheckpointStreamFactory.java:145)
at 
org.apache.flink.runtime.state.memory.MemCheckpointStreamFactory$MemoryCheckpointOutputStream.closeAndGetHandle(MemCheckpointStreamFactory.java:126)
at 
org.apache.flink.runtime.state.CheckpointStreamWithResultProvider$PrimaryStreamOnly.closeAndFinalizeCheckpointStreamResult(CheckpointStreamWithResultProvider.java:77)
at 
org.apache.flink.runtime.state.heap.HeapSnapshotStrategy$1.callInternal(HeapSnapshotStrategy.java:199)
at 
org.apache.flink.runtime.state.heap.HeapSnapshotStrategy$1.callInternal(HeapSnapshotStrategy.java:158)
at 
org.apache.flink.runtime.state.AsyncSnapshotCallable.call(AsyncSnapshotCallable.java:75)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at 
org.apache.flink.runtime.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:476)
... 5 more
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [DISCUSS] Releasing Apache Flink 1.11.3

2020-11-11 Thread Xintong Song
Thanks for the notice and fix, Roman.

Thank you~

Xintong Song



On Wed, Nov 11, 2020 at 5:53 PM Khachatryan Roman <
khachatryan.ro...@gmail.com> wrote:

> Hi,
>
> I'd like FLINK-20079 [1] to be merged into 1.11 and included in 1.11.3.
>
> [1] https://issues.apache.org/jira/browse/FLINK-20079
>
> Regards,
> Roman
>
>
> On Tue, Nov 10, 2020 at 8:21 AM Xintong Song 
> wrote:
>
> > Thanks for the notice, Dian.
> >
> > Thank you~
> >
> > Xintong Song
> >
> >
> >
> > On Tue, Nov 10, 2020 at 10:19 AM Dian Fu  wrote:
> >
> > > Hi Xintong,
> > >
> > > I want to bring one more issue to your attention [1]. The test case
> > > UnalignedCheckpointCompatibilityITCase.test failed several times in the
> > > last nightly test of release-1.11. We need to figure out if it's just
> an
> > > instable test or caused by recent changes.
> > >
> > > [1] https://issues.apache.org/jira/browse/FLINK-20065
> > >
> > > > 在 2020年11月10日,上午9:24,Xintong Song  写道:
> > > >
> > > > Thanks for the replies.
> > > >
> > > > Thank you~
> > > >
> > > > Xintong Song
> > > >
> > > >
> > > >
> > > > On Tue, Nov 10, 2020 at 1:09 AM Becket Qin 
> > wrote:
> > > >
> > > >> Hi Xintong,
> > > >>
> > > >> Thanks for driving the release. Just want to sync up on the FLIP-27
> > > >> backporting. Stephan and I are still trying to backport a bunch of
> > > patches
> > > >> of Source to 1.11.3. Including:
> > > >>
> > > >> [FLINK-19698][connector/common] Add a close() method to the
> > SplitReader.
> > > >> [FLINK-19717] SourceReaderBase.pollNext may return END_OF_INPUT if
> > > >> SplitReader.fetch throws
> > > >> [FLINK-19535] [connector/common] Avoid failing a job multiple times
> in
> > > >> SourceCoordinator.
> > > >> [FLINK-19265] [FLINK-20049][core] Source API final adjustments.
> > > >>
> > > >> and a few more fixes.
> > > >>
> > > >> We are currently trying to fix them in 1.12 first so it might take a
> > > little
> > > >> longer to backport them to 1.11.3. I think it will probably take us
> a
> > > few
> > > >> more days to finish the backport. So that would roughly be the end
> of
> > > this
> > > >> week.
> > > >>
> > > >> Thanks,
> > > >>
> > > >> Jiangjie (Becket) Qin
> > > >>
> > > >>
> > > >>
> > > >>
> > > >> On Mon, Nov 9, 2020 at 9:57 PM Till Rohrmann 
> > > wrote:
> > > >>
> > > >>> Yes, I've downgraded FLINK-19816 to critical.
> > > >>>
> > > >>> Cheers,
> > > >>> Till
> > > >>>
> > > >>> On Mon, Nov 9, 2020 at 10:19 AM Xintong Song <
> tonysong...@gmail.com>
> > > >>> wrote:
> > > >>>
> > >  Thanks for the notice, Till.
> > > 
> > >  I just checked and found FLINK-20033 is already fixed. Shall we
> also
> > >  downgrade FLINK-19816 to `Critical`?
> > > 
> > >  Thank you~
> > > 
> > >  Xintong Song
> > > 
> > > 
> > > 
> > >  On Mon, Nov 9, 2020 at 4:42 PM Till Rohrmann <
> trohrm...@apache.org>
> > > >>> wrote:
> > > 
> > > > I would like to bring one more critical issue to your attention
> > which
> > > >>> is
> > > > FLINK-20033 [1]. I believe that this issue is actually causing
> what
> > > >> has
> > > > been reported in FLINK-19816 [2]. I hope to have it fixed by the
> > end
> > > >> of
> > > > today. Once FLINK-20033 is fixed, I think that we don't have to
> > block
> > > >>> the
> > > > release on FLINK-19816.
> > > >
> > > > [1] https://issues.apache.org/jira/browse/FLINK-20033
> > > > [2] https://issues.apache.org/jira/browse/FLINK-19816
> > > >
> > > > Cheers,
> > > > Till
> > > >
> > > > On Mon, Nov 9, 2020 at 4:05 AM Xintong Song <
> tonysong...@gmail.com
> > >
> > >  wrote:
> > > >
> > > >> Hi devs,
> > > >>
> > > >> I'd like to provide an update on the progress of preparing
> release
> > > > 1.11.3.
> > > >>
> > > >> *Blockers*
> > > >> We currently have 3 remaining blockers. (3 resolved and 1
> emerged
> > > > compared
> > > >> to last week)
> > > >>
> > > >> - [FLINK-19698] Add close() method and onCheckpointComplete() to
> > > >> the
> > > >> Source.
> > > >> The issue has been fixed on the master branch. It's currently
> > > >> blocked
> > >  on
> > > >> the FLIP-27 backportings to backport it to the 1.11 branch.
> > > >>
> > > >> - [FLINK-19717] SourceReaderBase.pollNext may return
> END_OF_INPUT
> > > >> if
> > > >> SplitReader.fetch throws
> > > >> A PR has been opened and reviewed. From the discussions on the
> PR,
> > > >> it
> > > > looks
> > > >> close to mergeable.
> > > >>
> > > >> - [FLINK-19816] Flink restored from a wrong checkpoint (a very
> old
> > > >>> one
> > > > and
> > > >> not the last completed one)
> > > >> This is a newly emerged blocker and Matthias is working on it.
> > > >>
> > > >> *Test Instabilities*
> > > >> We currently have 27 test instabilities[1].
> > > >> AFAIK, none of them are as serious as to block the 1.11.3
> release.
> > > >>
> > > >> *FLIP-27

[jira] [Created] (FLINK-20100) Lag aggregate function does not return lag, but current row

2020-11-11 Thread Thilo Schneider (Jira)
Thilo Schneider created FLINK-20100:
---

 Summary: Lag aggregate function does not return lag, but current 
row
 Key: FLINK-20100
 URL: https://issues.apache.org/jira/browse/FLINK-20100
 Project: Flink
  Issue Type: Bug
  Components: API / Python
Affects Versions: 1.11.2
Reporter: Thilo Schneider


The lag aggregate function seems to always return the current row and not the 
row one lagged behind:

 

 
{code:java}
from pyflink.table import EnvironmentSettings, StreamTableEnvironment

env_settings = 
EnvironmentSettings.new_instance().in_streaming_mode().use_blink_planner().build()
t_env = StreamTableEnvironment.create( environment_settings=env_settings)

t_env.execute_sql("""
CREATE TABLE datagen (
 foo INT,
 message_time AS to_timestamp(from_unixtime(foo)),
 WATERMARK FOR message_time AS message_time
) WITH (
 'connector' = 'datagen',
 'rows-per-second'='3',
 'fields.foo.kind'='sequence',
 'fields.foo.start'='1',
 'fields.foo.end'='10'
)""")

t = t_env.sql_query("SELECT foo, lag(foo) OVER w AS lagfoo FROM datagen WINDOW 
w AS (ORDER BY message_time)")

t_env.execute_sql("CREATE TABLE output (foo INT, lagfoo INT) WITH ('connector' 
= 'print')")
t.execute_insert("output")
{code}
 

This results in

 
{code:java}
+I(1,1)  // Expected (1, null)
+I(2,2)  // Expected (2, 1)
+I(3,3)  // Expected (3, 2)
+I(4,4)  // and so on
+I(5,5)
+I(6,6)
+I(7,7)
+I(8,8)
+I(9,9)
+I(10,10)
{code}
 

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re:Re: Re: [VOTE] FLIP-145: Support SQL windowing table-valued function (2nd)

2020-11-11 Thread hailongwang


+1(no-binding)




在 2020-11-11 19:44:40,"刘大龙"  写道:
>
>+1
>
>> -原始邮件-
>> 发件人: "Timo Walther" 
>> 发送时间: 2020-11-11 18:55:06 (星期三)
>> 收件人: dev@flink.apache.org
>> 抄送: 
>> 主题: Re: [VOTE] FLIP-145: Support SQL windowing table-valued function (2nd)
>> 
>> +1 (binding)
>> 
>> Thanks,
>> Timo
>> 
>> On 11.11.20 07:14, Pengcheng Liu wrote:
>> > +1 (binding)
>> > 
>> > Jark Wu  于2020年11月11日周三 上午10:13写道:
>> > 
>> >> +1 (binding)
>> >>
>> >> On Tue, 10 Nov 2020 at 14:59, Jark Wu  wrote:
>> >>
>> >>> Hi all,
>> >>>
>> >>> There is new feedback on the FLIP-145. So I would like to start a new
>> >> vote
>> >>> for FLIP-145 [1],
>> >>> which has been discussed and reached consensus in the discussion thread
>> >>> [2].
>> >>>
>> >>> The vote will be open until 15:00 (UTC+8) 13th Nov. (72h), unless there
>> >> is
>> >>> an objection or not enough votes.
>> >>>
>> >>> Best,
>> >>> Jark
>> >>>
>> >>> [1]:
>> >>>
>> >> https://cwiki.apache.org/confluence/display/FLINK/FLIP-145%3A+Support+SQL+windowing+table-valued+function
>> >>> [2]:
>> >>>
>> >> http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-FLIP-145-Support-SQL-windowing-table-valued-function-td45269.html
>> >>>
>> >>
>> >


Re: Re: Re: [VOTE] FLIP-145: Support SQL windowing table-valued function (2nd)

2020-11-11 Thread Jingsong Li
+1

On Thu, Nov 12, 2020 at 2:07 PM hailongwang <18868816...@163.com> wrote:

>
>
> +1(no-binding)
>
>
>
>
> 在 2020-11-11 19:44:40,"刘大龙"  写道:
> >
> >+1
> >
> >> -原始邮件-
> >> 发件人: "Timo Walther" 
> >> 发送时间: 2020-11-11 18:55:06 (星期三)
> >> 收件人: dev@flink.apache.org
> >> 抄送:
> >> 主题: Re: [VOTE] FLIP-145: Support SQL windowing table-valued function
> (2nd)
> >>
> >> +1 (binding)
> >>
> >> Thanks,
> >> Timo
> >>
> >> On 11.11.20 07:14, Pengcheng Liu wrote:
> >> > +1 (binding)
> >> >
> >> > Jark Wu  于2020年11月11日周三 上午10:13写道:
> >> >
> >> >> +1 (binding)
> >> >>
> >> >> On Tue, 10 Nov 2020 at 14:59, Jark Wu  wrote:
> >> >>
> >> >>> Hi all,
> >> >>>
> >> >>> There is new feedback on the FLIP-145. So I would like to start a
> new
> >> >> vote
> >> >>> for FLIP-145 [1],
> >> >>> which has been discussed and reached consensus in the discussion
> thread
> >> >>> [2].
> >> >>>
> >> >>> The vote will be open until 15:00 (UTC+8) 13th Nov. (72h), unless
> there
> >> >> is
> >> >>> an objection or not enough votes.
> >> >>>
> >> >>> Best,
> >> >>> Jark
> >> >>>
> >> >>> [1]:
> >> >>>
> >> >>
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-145%3A+Support+SQL+windowing+table-valued+function
> >> >>> [2]:
> >> >>>
> >> >>
> http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-FLIP-145-Support-SQL-windowing-table-valued-function-td45269.html
> >> >>>
> >> >>
> >> >
>


-- 
Best, Jingsong Lee


Re: Re: Re: [VOTE] FLIP-145: Support SQL windowing table-valued function (2nd)

2020-11-11 Thread Danny Chan
+1

Jingsong Li  于2020年11月12日周四 下午2:18写道:

> +1
>
> On Thu, Nov 12, 2020 at 2:07 PM hailongwang <18868816...@163.com> wrote:
>
> >
> >
> > +1(no-binding)
> >
> >
> >
> >
> > 在 2020-11-11 19:44:40,"刘大龙"  写道:
> > >
> > >+1
> > >
> > >> -原始邮件-
> > >> 发件人: "Timo Walther" 
> > >> 发送时间: 2020-11-11 18:55:06 (星期三)
> > >> 收件人: dev@flink.apache.org
> > >> 抄送:
> > >> 主题: Re: [VOTE] FLIP-145: Support SQL windowing table-valued function
> > (2nd)
> > >>
> > >> +1 (binding)
> > >>
> > >> Thanks,
> > >> Timo
> > >>
> > >> On 11.11.20 07:14, Pengcheng Liu wrote:
> > >> > +1 (binding)
> > >> >
> > >> > Jark Wu  于2020年11月11日周三 上午10:13写道:
> > >> >
> > >> >> +1 (binding)
> > >> >>
> > >> >> On Tue, 10 Nov 2020 at 14:59, Jark Wu  wrote:
> > >> >>
> > >> >>> Hi all,
> > >> >>>
> > >> >>> There is new feedback on the FLIP-145. So I would like to start a
> > new
> > >> >> vote
> > >> >>> for FLIP-145 [1],
> > >> >>> which has been discussed and reached consensus in the discussion
> > thread
> > >> >>> [2].
> > >> >>>
> > >> >>> The vote will be open until 15:00 (UTC+8) 13th Nov. (72h), unless
> > there
> > >> >> is
> > >> >>> an objection or not enough votes.
> > >> >>>
> > >> >>> Best,
> > >> >>> Jark
> > >> >>>
> > >> >>> [1]:
> > >> >>>
> > >> >>
> >
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-145%3A+Support+SQL+windowing+table-valued+function
> > >> >>> [2]:
> > >> >>>
> > >> >>
> >
> http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-FLIP-145-Support-SQL-windowing-table-valued-function-td45269.html
> > >> >>>
> > >> >>
> > >> >
> >
>
>
> --
> Best, Jingsong Lee
>


Re: Re: Re: [VOTE] FLIP-145: Support SQL windowing table-valued function (2nd)

2020-11-11 Thread Benchao Li
+1

Danny Chan  于2020年11月12日周四 下午2:19写道:

> +1
>
> Jingsong Li  于2020年11月12日周四 下午2:18写道:
>
> > +1
> >
> > On Thu, Nov 12, 2020 at 2:07 PM hailongwang <18868816...@163.com> wrote:
> >
> > >
> > >
> > > +1(no-binding)
> > >
> > >
> > >
> > >
> > > 在 2020-11-11 19:44:40,"刘大龙"  写道:
> > > >
> > > >+1
> > > >
> > > >> -原始邮件-
> > > >> 发件人: "Timo Walther" 
> > > >> 发送时间: 2020-11-11 18:55:06 (星期三)
> > > >> 收件人: dev@flink.apache.org
> > > >> 抄送:
> > > >> 主题: Re: [VOTE] FLIP-145: Support SQL windowing table-valued function
> > > (2nd)
> > > >>
> > > >> +1 (binding)
> > > >>
> > > >> Thanks,
> > > >> Timo
> > > >>
> > > >> On 11.11.20 07:14, Pengcheng Liu wrote:
> > > >> > +1 (binding)
> > > >> >
> > > >> > Jark Wu  于2020年11月11日周三 上午10:13写道:
> > > >> >
> > > >> >> +1 (binding)
> > > >> >>
> > > >> >> On Tue, 10 Nov 2020 at 14:59, Jark Wu  wrote:
> > > >> >>
> > > >> >>> Hi all,
> > > >> >>>
> > > >> >>> There is new feedback on the FLIP-145. So I would like to start
> a
> > > new
> > > >> >> vote
> > > >> >>> for FLIP-145 [1],
> > > >> >>> which has been discussed and reached consensus in the discussion
> > > thread
> > > >> >>> [2].
> > > >> >>>
> > > >> >>> The vote will be open until 15:00 (UTC+8) 13th Nov. (72h),
> unless
> > > there
> > > >> >> is
> > > >> >>> an objection or not enough votes.
> > > >> >>>
> > > >> >>> Best,
> > > >> >>> Jark
> > > >> >>>
> > > >> >>> [1]:
> > > >> >>>
> > > >> >>
> > >
> >
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-145%3A+Support+SQL+windowing+table-valued+function
> > > >> >>> [2]:
> > > >> >>>
> > > >> >>
> > >
> >
> http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-FLIP-145-Support-SQL-windowing-table-valued-function-td45269.html
> > > >> >>>
> > > >> >>
> > > >> >
> > >
> >
> >
> > --
> > Best, Jingsong Lee
> >
>


-- 

Best,
Benchao Li


[jira] [Created] (FLINK-20101) Fix the wrong documentation of FROM_UNIXTIME function

2020-11-11 Thread Jark Wu (Jira)
Jark Wu created FLINK-20101:
---

 Summary: Fix the wrong documentation of FROM_UNIXTIME function
 Key: FLINK-20101
 URL: https://issues.apache.org/jira/browse/FLINK-20101
 Project: Flink
  Issue Type: Bug
  Components: Documentation, Table SQL / API
Reporter: Jark Wu
 Fix For: 1.12.0, 1.11.3
 Attachments: image-2020-11-12-15-43-16-755.png

The return value should be '1970-01-01 00:00:44' in UTC time zone. 

 !image-2020-11-12-15-43-16-755.png|thumbnail! 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)