Re: [ANNOUNCE] New Apache Flink Committer - Xintong Song

2020-06-05 Thread Jark Wu
Congratulations Xintong!

Best,
Jark

On Fri, 5 Jun 2020 at 14:32, Danny Chan  wrote:

> Congratulations Xintong !
>
> Best,
> Danny Chan
> 在 2020年6月5日 +0800 PM2:20,dev@flink.apache.org,写道:
> >
> > Congratulations Xintong
>


[jira] [Created] (FLINK-18146) The note of  StreamingJobGraphGenerator#createJobGraph has a misspelling

2020-06-05 Thread ZhuShang (Jira)
ZhuShang created FLINK-18146:


 Summary: The note of  StreamingJobGraphGenerator#createJobGraph 
has a misspelling
 Key: FLINK-18146
 URL: https://issues.apache.org/jira/browse/FLINK-18146
 Project: Flink
  Issue Type: Bug
  Components: Documentation
Affects Versions: 1.10.0
Reporter: ZhuShang


The note of 

StreamingJobGraphGenerator#createJobGraph in line 160 has a misspelling.
{code:java}
private JobGraph createJobGraph() {
...
// Generate deterministic hashes for the nodes in order to identify them across
// submission iff they didn't change.
Map hashes = 
defaultStreamGraphHasher.traverseStreamGraphAndGenerateHashes(streamGraph);
...
}
{code}
Is 'iff' should be 'if'?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18147) Orc document display is disordered

2020-06-05 Thread Shuai Xia (Jira)
Shuai Xia created FLINK-18147:
-

 Summary: Orc document display is disordered
 Key: FLINK-18147
 URL: https://issues.apache.org/jira/browse/FLINK-18147
 Project: Flink
  Issue Type: Bug
  Components: Documentation, FileSystems
Affects Versions: 1.11.0
Reporter: Shuai Xia
 Fix For: 1.11.0


Official documents show that there is a problem.

link: 
[https://ci.apache.org/projects/flink/flink-docs-master/dev/connectors/streamfile_sink.html#tab_java_2]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [ANNOUNCE] New Apache Flink Committer - Xintong Song

2020-06-05 Thread tison
Congrats, Xintong!

Best,
tison.


Jark Wu  于2020年6月5日周五 下午3:00写道:

> Congratulations Xintong!
>
> Best,
> Jark
>
> On Fri, 5 Jun 2020 at 14:32, Danny Chan  wrote:
>
> > Congratulations Xintong !
> >
> > Best,
> > Danny Chan
> > 在 2020年6月5日 +0800 PM2:20,dev@flink.apache.org,写道:
> > >
> > > Congratulations Xintong
> >
>


Re: [ANNOUNCE] New Apache Flink Committer - Xintong Song

2020-06-05 Thread Congxian Qiu
Congratulations! Xintong!
Best,
Congxian


Jark Wu  于2020年6月5日周五 下午3:00写道:

> Congratulations Xintong!
>
> Best,
> Jark
>
> On Fri, 5 Jun 2020 at 14:32, Danny Chan  wrote:
>
> > Congratulations Xintong !
> >
> > Best,
> > Danny Chan
> > 在 2020年6月5日 +0800 PM2:20,dev@flink.apache.org,写道:
> > >
> > > Congratulations Xintong
> >
>


[jira] [Created] (FLINK-18148) "Resuming Savepoint" e2e fails with TimeoutException in CliFrontend.stop()

2020-06-05 Thread Robert Metzger (Jira)
Robert Metzger created FLINK-18148:
--

 Summary: "Resuming Savepoint" e2e fails with TimeoutException in 
CliFrontend.stop() 
 Key: FLINK-18148
 URL: https://issues.apache.org/jira/browse/FLINK-18148
 Project: Flink
  Issue Type: Bug
  Components: Command Line Client
Affects Versions: 1.11.0
Reporter: Robert Metzger


https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=2759&view=logs&j=c88eea3b-64a0-564d-0031-9fdcd7b8abee&t=1e2bbe5b-4657-50be-1f07-d84bfce5b1f5

{code}

 The program finished with the following exception:

org.apache.flink.util.FlinkException: Could not stop with a savepoint job 
"081bda854bc250e01055ed1ba9d43178".
at 
org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java:495)
at 
org.apache.flink.client.cli.CliFrontend.runClusterAction(CliFrontend.java:864)
at org.apache.flink.client.cli.CliFrontend.stop(CliFrontend.java:487)
at 
org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:931)
at 
org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:992)
at 
org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30)
at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:992)
Caused by: java.util.concurrent.TimeoutException
at 
java.util.concurrent.CompletableFuture.timedGet(CompletableFuture.java:1784)
at 
java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1928)
at 
org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java:493)
... 6 more
Waiting for job (081bda854bc250e01055ed1ba9d43178) to reach terminal state 
FINISHED ...
Job (081bda854bc250e01055ed1ba9d43178) reached terminal state FINISHED
Savepoint location was empty. This may mean that the stop-with-savepoint failed.
[FAIL] Test script contains errors.
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [Discussion] flink elasticsearch connector supports

2020-06-05 Thread Forward Xu
hi Jacky Lau :

I agree with jark's point of view, the use of es is not just to read data,
more use is to group query, aggregate these.

Best,
Forward

Jacky Lau  于2020年6月5日周五 下午2:47写道:

> hi Etienne Chauchot:
> you can read here https://www.jianshu.com/p/d32e17dab90c, which is
> chinese.But you can konw that slice api has poor performance in es-hadoop
> project .
>
> And i found that es-hadoop has removed this and disable sliced scrolls by
> default. you can see below, which i found in the lastest es-hadoop release
> version
>  Configuration Changes
> `es.input.use.sliced.partitions` is deprecated in 6.5.0, and will be
> removed
> in 7.0.0. The default value for `es.input.max.docs.per.partition` (10)
> will also be removed in 7.0.0, thus disabling sliced scrolls by default,
> and
> switching them to be an explicitly opt-in feature.
>
> added[5.0.0]
> `es.input.max.docs.per.partition` ::
> When reading from an {es} cluster that supports scroll slicing ({es} v5.0.0
> and above), this parameter advises the
> connector on what the maximum number of documents per input partition
> should
> be. The connector will sample and estimate
> the number of documents on each shard to be read and divides each shard
> into
> input slices using the value supplied by
> this property. This property is a suggestion, not a guarantee. The final
> number of documents per partition is not
> guaranteed to be below this number, but rather, they will be close to this
> number. This property is ignored if you are
> reading from an {es} cluster that does not support scroll slicing ({es} any
> version below v5.0.0). By default, this
> value is unset, and the input partitions are calculated based on the number
> of shards in the indices being read.
>
>
>
> Jacky Lau wrote
> > hi Etienne Chauchot:
> > thanks for your discussion.
> > for 1) we do not supprt es  unbouded source currently
> >
> > for 2) RichParallelSourceFunction is used for streaming ,InputFormat is
> > for
> > batch
> >
> > for 3)  i downloaded beam just now. and the beam es connector is also
> > using
> > es-hadoop. i have read the code of es-hadoop(inputsplit contains shard
> and
> > slice. And i think it is better when diffirent shard has diffirent number
> > of
> > docs), which you can seed here
> > .https://github.com/elastic/elasticsearch-hadoop. But the code is not
> > good.
> > so we do not want to reference . and you can see presto, there is also
> > just
> > using inputsplit with shard not contains slice
> >
> > for 4) because flink es connectro has alreay using diffrent client (es 5
> > for
> > tranport client, es 6,7 for highlevelrest), we just  reuse it,which will
> > not
> > change too much code
> >
> >
> >
> > --
> > Sent from:
> http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/
>
>
>
>
>
> --
> Sent from: http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/
>


Re: [DISCUSS] Releasing "fat" and "slim" Flink distributions

2020-06-05 Thread Leonard Xu
+1 for Jingsong’s proposal to put flink-csv, flink-json and flink-avro under 
lib/ directory.
I have heard many SQL users(most of newbies) complaint the out-of-box 
experience in mail list.

Best,
Leonard Xu


> 在 2020年6月5日,14:39,Benchao Li  写道:
> 
> +1 to include them for sql-client by default;
> +0 to put into lib and exposed to all kinds of jobs, including DataStream.
> 
> Danny Chan  于2020年6月5日周五 下午2:31写道:
> 
>> +1, at least, we should keep an out of the box SQL-CLI, it’s very poor
>> experience to add such required format jars for SQL users.
>> 
>> Best,
>> Danny Chan
>> 在 2020年6月5日 +0800 AM11:14,Jingsong Li ,写道:
>>> Hi all,
>>> 
>>> Considering that 1.11 will be released soon, what about my previous
>>> proposal? Put flink-csv, flink-json and flink-avro under lib.
>>> These three formats are very small and no third party dependence, and
>> they
>>> are widely used by table users.
>>> 
>>> Best,
>>> Jingsong Lee
>>> 
>>> On Tue, May 12, 2020 at 4:19 PM Jingsong Li 
>> wrote:
>>> 
 Thanks for your discussion.
 
 Sorry to start discussing another thing:
 
 The biggest problem I see is the variety of problems caused by users'
>> lack
 of format dependency.
 As Aljoscha said, these three formats are very small and no third party
 dependence, and they are widely used by table users.
 Actually, we don't have any other built-in table formats now... In
>> total
 151K...
 
 73K flink-avro-1.10.0.jar
 36K flink-csv-1.10.0.jar
 42K flink-json-1.10.0.jar
 
 So, Can we just put them into "lib/" or flink-table-uber?
 It not solve all problems and maybe it is independent of "fat" and
>> "slim".
 But also improve usability.
 What do you think? Any objections?
 
 Best,
 Jingsong Lee
 
 On Mon, May 11, 2020 at 5:48 PM Chesnay Schepler 
 wrote:
 
> One downside would be that we're shipping more stuff when running on
> YARN for example, since the entire plugins directory is shiped by
>> default.
> 
> On 17/04/2020 16:38, Stephan Ewen wrote:
>> @Aljoscha I think that is an interesting line of thinking. the
>> swift-fs
> may
>> be rarely enough used to move it to an optional download.
>> 
>> I would still drop two more thoughts:
>> 
>> (1) Now that we have plugins support, is there a reason to have a
> metrics
>> reporter or file system in /opt instead of /plugins? They don't
>> spoil
> the
>> class path any more.
>> 
>> (2) I can imagine there still being a desire to have a "minimal"
>> docker
>> file, for users that want to keep the container images as small as
>> possible, to speed up deployment. It is fine if that would not be
>> the
>> default, though.
>> 
>> 
>> On Fri, Apr 17, 2020 at 12:16 PM Aljoscha Krettek <
>> aljos...@apache.org>
>> wrote:
>> 
>>> I think having such tools and/or tailor-made distributions can
>> be nice
>>> but I also think the discussion is missing the main point: The
>> initial
>>> observation/motivation is that apparently a lot of users (Kurt
>> and I
>>> talked about this) on the chinese DingTalk support groups, and
>> other
>>> support channels have problems when first using the SQL client
>> because
>>> of these missing connectors/formats. For these, having
>> additional tools
>>> would not solve anything because they would also not take that
>> extra
>>> step. I think that even tiny friction should be avoided because
>> the
>>> annoyance from it accumulates of the (hopefully) many users that
>> we
> want
>>> to have.
>>> 
>>> Maybe we should take a step back from discussing the
>> "fat"/"slim" idea
>>> and instead think about the composition of the current dist. As
>>> mentioned we have these jars in opt/:
>>> 
>>> 17M flink-azure-fs-hadoop-1.10.0.jar
>>> 52K flink-cep-scala_2.11-1.10.0.jar
>>> 180K flink-cep_2.11-1.10.0.jar
>>> 746K flink-gelly-scala_2.11-1.10.0.jar
>>> 626K flink-gelly_2.11-1.10.0.jar
>>> 512K flink-metrics-datadog-1.10.0.jar
>>> 159K flink-metrics-graphite-1.10.0.jar
>>> 1.0M flink-metrics-influxdb-1.10.0.jar
>>> 102K flink-metrics-prometheus-1.10.0.jar
>>> 10K flink-metrics-slf4j-1.10.0.jar
>>> 12K flink-metrics-statsd-1.10.0.jar
>>> 36M flink-oss-fs-hadoop-1.10.0.jar
>>> 28M flink-python_2.11-1.10.0.jar
>>> 22K flink-queryable-state-runtime_2.11-1.10.0.jar
>>> 18M flink-s3-fs-hadoop-1.10.0.jar
>>> 31M flink-s3-fs-presto-1.10.0.jar
>>> 196K flink-shaded-netty-tcnative-dynamic-2.0.25.Final-9.0.jar
>>> 518K flink-sql-client_2.11-1.10.0.jar
>>> 99K flink-state-processor-api_2.11-1.10.0.jar
>>> 25M flink-swift-fs-hadoop-1.10.0.jar
>>> 160M opt
>>> 
>>> The "filesystem" connectors ar ethe heavy hitters, there.
>>> 
>>> I downloaded most of the SQL connectors/formats and this is what
>> I got:
>>> 
>>> 

Re: [ANNOUNCE] New Apache Flink Committer - Xintong Song

2020-06-05 Thread Yu Li
Congratulations Xintong, well deserved!

Best Regards,
Yu


On Fri, 5 Jun 2020 at 15:12, Congxian Qiu  wrote:

> Congratulations! Xintong!
> Best,
> Congxian
>
>
> Jark Wu  于2020年6月5日周五 下午3:00写道:
>
> > Congratulations Xintong!
> >
> > Best,
> > Jark
> >
> > On Fri, 5 Jun 2020 at 14:32, Danny Chan  wrote:
> >
> > > Congratulations Xintong !
> > >
> > > Best,
> > > Danny Chan
> > > 在 2020年6月5日 +0800 PM2:20,dev@flink.apache.org,写道:
> > > >
> > > > Congratulations Xintong
> > >
> >
>


[jira] [Created] (FLINK-18149) Taskmanager logs could not show up in native K8s deployment

2020-06-05 Thread Yang Wang (Jira)
Yang Wang created FLINK-18149:
-

 Summary: Taskmanager logs could not show up in native K8s 
deployment
 Key: FLINK-18149
 URL: https://issues.apache.org/jira/browse/FLINK-18149
 Project: Flink
  Issue Type: Bug
  Components: Deployment / Kubernetes
Affects Versions: 1.11.0, 1.12.0
Reporter: Yang Wang
 Fix For: 1.11.0, 1.12.0


In FLINK-17935, we use {{flinkConfig.get(DeploymentOptionsInternal.CONF_DIR)}} 
to replace {{CliFrontend.getConfigurationDirectoryFromEnv}}. It will cause 
problem in native K8s integration. The root cause we set the 
{{DeploymentOptionsInternal.CONF_DIR}} in flink-conf.yaml to a local path. 
However, it does not exist in JobManager pod.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [ANNOUNCE] New Apache Flink Committer - Xintong Song

2020-06-05 Thread Rui Li
Congrats Xintong!

On Fri, Jun 5, 2020 at 3:12 PM Congxian Qiu  wrote:

> Congratulations! Xintong!
> Best,
> Congxian
>
>
> Jark Wu  于2020年6月5日周五 下午3:00写道:
>
> > Congratulations Xintong!
> >
> > Best,
> > Jark
> >
> > On Fri, 5 Jun 2020 at 14:32, Danny Chan  wrote:
> >
> > > Congratulations Xintong !
> > >
> > > Best,
> > > Danny Chan
> > > 在 2020年6月5日 +0800 PM2:20,dev@flink.apache.org,写道:
> > > >
> > > > Congratulations Xintong
> > >
> >
>


-- 
Best regards!
Rui Li


Re: [DISCUSS] Releasing "fat" and "slim" Flink distributions

2020-06-05 Thread Rui Li
+1 to add light-weighted formats into the lib

On Fri, Jun 5, 2020 at 3:28 PM Leonard Xu  wrote:

> +1 for Jingsong’s proposal to put flink-csv, flink-json and flink-avro
> under lib/ directory.
> I have heard many SQL users(most of newbies) complaint the out-of-box
> experience in mail list.
>
> Best,
> Leonard Xu
>
>
> > 在 2020年6月5日,14:39,Benchao Li  写道:
> >
> > +1 to include them for sql-client by default;
> > +0 to put into lib and exposed to all kinds of jobs, including
> DataStream.
> >
> > Danny Chan  于2020年6月5日周五 下午2:31写道:
> >
> >> +1, at least, we should keep an out of the box SQL-CLI, it’s very poor
> >> experience to add such required format jars for SQL users.
> >>
> >> Best,
> >> Danny Chan
> >> 在 2020年6月5日 +0800 AM11:14,Jingsong Li ,写道:
> >>> Hi all,
> >>>
> >>> Considering that 1.11 will be released soon, what about my previous
> >>> proposal? Put flink-csv, flink-json and flink-avro under lib.
> >>> These three formats are very small and no third party dependence, and
> >> they
> >>> are widely used by table users.
> >>>
> >>> Best,
> >>> Jingsong Lee
> >>>
> >>> On Tue, May 12, 2020 at 4:19 PM Jingsong Li 
> >> wrote:
> >>>
>  Thanks for your discussion.
> 
>  Sorry to start discussing another thing:
> 
>  The biggest problem I see is the variety of problems caused by users'
> >> lack
>  of format dependency.
>  As Aljoscha said, these three formats are very small and no third
> party
>  dependence, and they are widely used by table users.
>  Actually, we don't have any other built-in table formats now... In
> >> total
>  151K...
> 
>  73K flink-avro-1.10.0.jar
>  36K flink-csv-1.10.0.jar
>  42K flink-json-1.10.0.jar
> 
>  So, Can we just put them into "lib/" or flink-table-uber?
>  It not solve all problems and maybe it is independent of "fat" and
> >> "slim".
>  But also improve usability.
>  What do you think? Any objections?
> 
>  Best,
>  Jingsong Lee
> 
>  On Mon, May 11, 2020 at 5:48 PM Chesnay Schepler 
>  wrote:
> 
> > One downside would be that we're shipping more stuff when running on
> > YARN for example, since the entire plugins directory is shiped by
> >> default.
> >
> > On 17/04/2020 16:38, Stephan Ewen wrote:
> >> @Aljoscha I think that is an interesting line of thinking. the
> >> swift-fs
> > may
> >> be rarely enough used to move it to an optional download.
> >>
> >> I would still drop two more thoughts:
> >>
> >> (1) Now that we have plugins support, is there a reason to have a
> > metrics
> >> reporter or file system in /opt instead of /plugins? They don't
> >> spoil
> > the
> >> class path any more.
> >>
> >> (2) I can imagine there still being a desire to have a "minimal"
> >> docker
> >> file, for users that want to keep the container images as small as
> >> possible, to speed up deployment. It is fine if that would not be
> >> the
> >> default, though.
> >>
> >>
> >> On Fri, Apr 17, 2020 at 12:16 PM Aljoscha Krettek <
> >> aljos...@apache.org>
> >> wrote:
> >>
> >>> I think having such tools and/or tailor-made distributions can
> >> be nice
> >>> but I also think the discussion is missing the main point: The
> >> initial
> >>> observation/motivation is that apparently a lot of users (Kurt
> >> and I
> >>> talked about this) on the chinese DingTalk support groups, and
> >> other
> >>> support channels have problems when first using the SQL client
> >> because
> >>> of these missing connectors/formats. For these, having
> >> additional tools
> >>> would not solve anything because they would also not take that
> >> extra
> >>> step. I think that even tiny friction should be avoided because
> >> the
> >>> annoyance from it accumulates of the (hopefully) many users that
> >> we
> > want
> >>> to have.
> >>>
> >>> Maybe we should take a step back from discussing the
> >> "fat"/"slim" idea
> >>> and instead think about the composition of the current dist. As
> >>> mentioned we have these jars in opt/:
> >>>
> >>> 17M flink-azure-fs-hadoop-1.10.0.jar
> >>> 52K flink-cep-scala_2.11-1.10.0.jar
> >>> 180K flink-cep_2.11-1.10.0.jar
> >>> 746K flink-gelly-scala_2.11-1.10.0.jar
> >>> 626K flink-gelly_2.11-1.10.0.jar
> >>> 512K flink-metrics-datadog-1.10.0.jar
> >>> 159K flink-metrics-graphite-1.10.0.jar
> >>> 1.0M flink-metrics-influxdb-1.10.0.jar
> >>> 102K flink-metrics-prometheus-1.10.0.jar
> >>> 10K flink-metrics-slf4j-1.10.0.jar
> >>> 12K flink-metrics-statsd-1.10.0.jar
> >>> 36M flink-oss-fs-hadoop-1.10.0.jar
> >>> 28M flink-python_2.11-1.10.0.jar
> >>> 22K flink-queryable-state-runtime_2.11-1.10.0.jar
> >>> 18M flink-s3-fs-hadoop-1.10.0.jar
> >>> 31M flink-s3-fs-presto-1.10.0.jar
> >>> 196K flink-shaded-netty-tcnative-dynamic-2.0.25.Final-9.0

Re: [ANNOUNCE] New Apache Flink Committer - Xintong Song

2020-06-05 Thread Dawid Wysakowicz
Congratulations!

Best,

Dawid

On 05/06/2020 09:10, tison wrote:
> Congrats, Xintong!
>
> Best,
> tison.
>
>
> Jark Wu  于2020年6月5日周五 下午3:00写道:
>
>> Congratulations Xintong!
>>
>> Best,
>> Jark
>>
>> On Fri, 5 Jun 2020 at 14:32, Danny Chan  wrote:
>>
>>> Congratulations Xintong !
>>>
>>> Best,
>>> Danny Chan
>>> 在 2020年6月5日 +0800 PM2:20,dev@flink.apache.org,写道:
 Congratulations Xintong


signature.asc
Description: OpenPGP digital signature


[jira] [Created] (FLINK-18150) A single failing Kafka broker may cause jobs to fail indefinitely with TimeoutException: Timeout expired while fetching topic metadata

2020-06-05 Thread Julius Michaelis (Jira)
Julius Michaelis created FLINK-18150:


 Summary: A single failing Kafka broker may cause jobs to fail 
indefinitely with TimeoutException: Timeout expired while fetching topic 
metadata
 Key: FLINK-18150
 URL: https://issues.apache.org/jira/browse/FLINK-18150
 Project: Flink
  Issue Type: Bug
  Components: Connectors / Kafka
Affects Versions: 1.10.1
 Environment: It is a bit unclear to me under what circumstances this 
can be reproduced. I created a "minimum" non-working example at 
https://github.com/jcaesar/flink-kafka-ha-failure. Note that this deploys the 
minimum number of Kafka brokers, but it works just as well with replication 
factor 3 and 8 brokers, e.g.

I run this with
{code:bash}
docker-compose kill; and docker-compose rm -vf; and docker-compose up 
--abort-on-container-exit --build
{code}
The exception should appear on the webui after 5~6 minutes.

You verify that the Kafka cluster is running "normally" e.g. with:
{code:bash}
kafkacat -b localhost,localhost:9093 -L
{code}

So far, I only know that
* {{flink.partition-discovery.interval-millis}} must be set.
* The broker that failed must be part of the {{bootstrap.servers}}
* There needs to be a certain amount of topics or producers, but I'm unsure 
which is crucial
* Changing the values of {{metadata.request.timeout.ms}} or 
{{flink.partition-discovery.interval-millis}} does not seem to have any effect.
Reporter: Julius Michaelis


When a Kafka broker fails that is listed among the bootstrap servers and 
partition discovery is active, the Flink job reading from that Kafka may enter 
a failing loop.

At first, the job seems to react normally without failure with only a short 
latency spike when switching Kafka leaders.
Then, it fails with a
{code:plain}
org.apache.flink.streaming.connectors.kafka.internal.Handover$ClosedException
at 
org.apache.flink.streaming.connectors.kafka.internal.Handover.close(Handover.java:182)
at 
org.apache.flink.streaming.connectors.kafka.internal.KafkaFetcher.cancel(KafkaFetcher.java:175)
at 
org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumerBase.cancel(FlinkKafkaConsumerBase.java:821)
at 
org.apache.flink.streaming.api.operators.StreamSource.cancel(StreamSource.java:147)
at 
org.apache.flink.streaming.runtime.tasks.SourceStreamTask.cancelTask(SourceStreamTask.java:136)
at 
org.apache.flink.streaming.runtime.tasks.StreamTask.cancel(StreamTask.java:602)
at 
org.apache.flink.runtime.taskmanager.Task$TaskCanceler.run(Task.java:1355)
at java.lang.Thread.run(Thread.java:748)
{code}
It recovers, but processes fewer than the expected amount of records.

Finally,  the job fails with
{code:plain}
2020-06-05 13:59:37
org.apache.kafka.common.errors.TimeoutException: Timeout expired while fetching 
topic metadata
{code}
and repeats doing so while not processing any records.


I have also observed this behavior with partition-discovery turned off, but 
only when the Flink job failed (after a broker failure) and had to run 
checkpoint recovery for some other reason.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [ANNOUNCE] New Apache Flink Committer - Xintong Song

2020-06-05 Thread Yun Tang
Congratulations, Xintong

Best
Yun Tang



From: Dawid Wysakowicz
Sent: Friday, June 05, 2020 16:00
To: dev@flink.apache.org
Subject: Re: [ANNOUNCE] New Apache Flink Committer - Xintong Song

Congratulations!

Best,

Dawid

On 05/06/2020 09:10, tison wrote:
> Congrats, Xintong!
>
> Best,
> tison.
>
>
> Jark Wu  于2020年6月5日周五 下午3:00写道:
>
>> Congratulations Xintong!
>>
>> Best,
>> Jark
>>
>> On Fri, 5 Jun 2020 at 14:32, Danny Chan  wrote:
>>
>>> Congratulations Xintong !
>>>
>>> Best,
>>> Danny Chan
>>> 在 2020年6月5日 +0800 PM2:20,dev@flink.apache.org,写道:
 Congratulations Xintong


Re: [VOTE] Apache Flink Stateful Functions 2.1.0, release candidate #1

2020-06-05 Thread Robert Metzger
Thanks a lot for creating this release Gordon!

+1 (binding)

- maven staging repo looks fine (version tags, license files)
- source archive looks good (no binaries, no unexpected files, pom has
right version)
- quickly checked diff:
https://github.com/apache/flink-statefun/compare/release-2.0.0-rc6...release-2.1.0-rc1


On Fri, Jun 5, 2020 at 5:05 AM Congxian Qiu  wrote:

> @Tzu-Li (Gordon) Tai   Thanks for the info. `mvn
> clean
> install -Prun=e2e-tests` works for me. before verified demo on a clean
> source directory.
>
> Best,
> Congxian
>
>
> Tzu-Li (Gordon) Tai  于2020年6月4日周四 下午6:35写道:
>
> > +1 (binding)
> >
> > Legal
> >
> > - Verified signatures and hashes of staged Maven artifacts, source
> > distribution and Python SDK distribution
> > - Checked NOTICE file of statefun-flink-distribution and
> > statefun-ridesharing-example-simulator
> >
> > Functional
> >
> > - Full build with end-to-end-tests, JDK 8: mvn clean install
> > -Prun-e2e-tests
> > - Manually verified state TTL for remote functions
> > - Manually verified checkpointing with failure recovery
> > - Manually verified savepointing + manual restore
> > - Generated quickstart project from archetype works
> >
> >
> > On Thu, Jun 4, 2020 at 3:10 PM Hequn Cheng  wrote:
> >
> > > +1 (binding)
> > >
> > > - Signatures and hash are correct.
> > > - All artifacts to be released to Maven in the staging Nexus
> repository.
> > > - Verify that the source archives do not contain any binaries.
> > > - Go through all commits from the last release. No license problem
> > spotted.
> > > - Check end-to-end tests. All tests have been passed on Travis(both for
> > JDK
> > > 1.8 and 1.11).
> > >
> > > Best,
> > > Hequn
> > >
> > > On Thu, Jun 4, 2020 at 12:50 PM Tzu-Li (Gordon) Tai <
> tzuli...@apache.org
> > >
> > > wrote:
> > >
> > > > Hi Hequn,
> > > >
> > > > Sorry, I mis-tagged the wrong commit.
> > > > Just fixed this, the tag [1] [2] should now be pointing to the
> correct
> > > > commit that contains the updated version.
> > > >
> > > > Gordon
> > > >
> > > > [1]
> > > >
> > > >
> > >
> >
> https://gitbox.apache.org/repos/asf?p=flink-statefun.git;a=tag;h=c08c9850147d818fc8fed877a01ff87021f3cf21
> > > > [2] https://github.com/apache/flink-statefun/tree/release-2.1.0-rc1
> > > >
> > > > On Thu, Jun 4, 2020 at 12:10 PM Hequn Cheng 
> wrote:
> > > >
> > > > > It seems the release tag is not correct? The version in the poms
> > should
> > > > > be 2.1.0 instead of 2.1-SNAPSHOT.
> > > > >
> > > > > Best,
> > > > > Hequn
> > > > >
> > > > >
> > > > > On Thu, Jun 4, 2020 at 10:33 AM Congxian Qiu <
> qcx978132...@gmail.com
> > >
> > > > > wrote:
> > > > >
> > > > > > +1  (non-binding)
> > > > > >
> > > > > > maybe there is something that needs to be updated in
> > > > README.md(currently
> > > > > > the official docs link points to the master instead of 2.1)
> > > > > >
> > > > > > and have another question: do we need to add the command used to
> > > build
> > > > > the
> > > > > > base image locally(which was on the README.md in release-2.0.0)?
> > > > > >
> > > > > > checked
> > > > > > - sha & gpg, ok
> > > > > > - mvn clean install -Prun-e2e-test on 1.8.0_252, ok
> > > > > > - source archives do not contains any binaries
> > > > > > - maven clean install -Papache-release, ok (this step need a gpg
> > > secret
> > > > > > key)
> > > > > > - check all pom files, dockerfiles, examples point to the same
> > > version,
> > > > > ok
> > > > > > - check READM.md, nothing unexpected.
> > > > > > - but the official docs link points to the master instead of
> > 2.1
> > > > > > - run greeter&ride-share demo, ok
> > > > > >
> > > > > > Best,
> > > > > > Congxian
> > > > > >
> > > > > >
> > > > > > Tzu-Li (Gordon) Tai  于2020年6月1日周一 下午3:25写道:
> > > > > >
> > > > > > > Hi everyone,
> > > > > > >
> > > > > > > Please review and vote on the *release candidate #1* for the
> > > version
> > > > > > 2.1.0
> > > > > > > of
> > > > > > > Apache Flink Stateful Functions,
> > > > > > > as follows:
> > > > > > > [ ] +1, Approve the release
> > > > > > > [ ] -1, Do not approve the release (please provide specific
> > > comments)
> > > > > > >
> > > > > > > ***Testing Guideline***
> > > > > > >
> > > > > > > You can find here [1] a page in the project wiki on
> instructions
> > > for
> > > > > > > testing.
> > > > > > > To cast a vote, it is not necessary to perform all listed
> checks,
> > > > > > > but please mention which checks you have performed when voting.
> > > > > > >
> > > > > > > ***Release Overview***
> > > > > > >
> > > > > > > As an overview, the release consists of the following:
> > > > > > > a) Stateful Functions canonical source distribution, to be
> > deployed
> > > > to
> > > > > > the
> > > > > > > release repository at dist.apache.org
> > > > > > > b) Stateful Functions Python SDK distributions to be deployed
> to
> > > PyPI
> > > > > > > c) Maven artifacts to be deployed to the Maven Central
> Repository
> > > > > > >
> > > > > > > ***Staging Areas to Rev

Re: [Discussion] flink elasticsearch connector supports

2020-06-05 Thread Etienne Chauchot

Hi Jacky Lau,

1) yes I saw that

2) I saw sources like IntegerSource which are bounded and which extend 
RichParallelSourceFunction. This is why I mentioned it.


3) True, there is an hadoop ES connector in Beam but it is more of a 
side connector, the main one is ElasticsearchIO here: 
https://github.com/apache/beam/blob/095589c28f5c427bf99fc0330af91c859bb2ad6b/sdks/java/io/elasticsearch/src/main/java/org/apache/beam/sdk/io/elasticsearch/ElasticsearchIO.java#L156 
and it does not use hadoop.


4) Yes but using the same client could simplify the code in the end, but 
I agree it needs more change in the current code.


Etienne

On 05/06/2020 05:50, Jacky Lau wrote:

hi Etienne Chauchot:
thanks for your discussion.
for 1) we do not supprt es  unbouded source currently

for 2) RichParallelSourceFunction is used for streaming ,InputFormat is for
batch

for 3)  i downloaded beam just now. and the beam es connector is also using
es-hadoop. i have read the code of es-hadoop(inputsplit contains shard and
slice. And i think it is better when diffirent shard has diffirent number of
docs), which you can seed here
.https://github.com/elastic/elasticsearch-hadoop. But the code is not good.
so we do not want to reference . and you can see presto, there is also just
using inputsplit with shard not contains slice

for 4) because flink es connectro has alreay using diffrent client (es 5 for
tranport client, es 6,7 for highlevelrest), we just  reuse it,which will not
change too much code



--
Sent from: http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/


[jira] [Created] (FLINK-18151) Resolve CWE22 problems in pyflink_gateway_server.py

2020-06-05 Thread Hequn Cheng (Jira)
Hequn Cheng created FLINK-18151:
---

 Summary: Resolve CWE22 problems in pyflink_gateway_server.py 
 Key: FLINK-18151
 URL: https://issues.apache.org/jira/browse/FLINK-18151
 Project: Flink
  Issue Type: Bug
  Components: API / Python
Affects Versions: 1.10.1, 1.11.0, 1.12.0
Reporter: Hequn Cheng


For example, the code `if os.path.isfile(flink_conf_path):` contains CWE22 
problem that calling "os.path.isfile" with the tainted value in argument 1. 
This constructs a path or URI using the tainted value and may thus allow an 
attacker to access, modify, or test the existence of critical or sensitive 
files.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18152) Master starts despite IllegalConfigurationException

2020-06-05 Thread Chesnay Schepler (Jira)
Chesnay Schepler created FLINK-18152:


 Summary: Master starts despite IllegalConfigurationException
 Key: FLINK-18152
 URL: https://issues.apache.org/jira/browse/FLINK-18152
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Coordination
Affects Versions: 1.11.0
Reporter: Chesnay Schepler


When no memory configuration parameters are set for the Master process then 
this exception is logged by the BashJavaUtils:
{code}
Exception in thread "main" 
org.apache.flink.configuration.IllegalConfigurationException: Either required 
fine-grained memory (jobmanager.memory.heap.size), or Total Flink Memory size 
(Key: 'jobmanager.memory.flink.size' , default: null $at 
org.apache.flink.runtime.util.config.memory.ProcessMemoryUtils.failBecauseRequiredOptionsNotConfigured(ProcessMemoryUtils.java:111)

   at 
org.apache.flink.runtime.util.config.memory.ProcessMemoryUtils.memoryProcessSpecFromConfig(ProcessMemoryUtils.java:81)

at 
org.apache.flink.runtime.jobmanager.JobManagerProcessUtils.processSpecFromConfig(JobManagerProcessUtils.java:76)

  at 
org.apache.flink.runtime.jobmanager.JobManagerProcessUtils.processSpecFromConfigWithNewOptionToInterpretLegacyHeap(JobManagerProcessUtils.java:71)

at 
org.apache.flink.runtime.util.bash.BashJavaUtils.getJmResourceParams(BashJavaUtils.java:92)

   at 
org.apache.flink.runtime.util.bash.BashJavaUtils.runCommand(BashJavaUtils.java:66)

at 
org.apache.flink.runtime.util.bash.BashJavaUtils.main(BashJavaUtils.java:54)
{code}

Afterwards, the Master process however continues to start just fine.

TaskManagers will not be started given the same situation, which is 
inconsistent behavior, and it seems dangerous to start a process despite an 
illegal configuration.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18153) flink-daemon does not detect errors if JVM crashes immediately

2020-06-05 Thread Chesnay Schepler (Jira)
Chesnay Schepler created FLINK-18153:


 Summary: flink-daemon does not detect errors if JVM crashes 
immediately
 Key: FLINK-18153
 URL: https://issues.apache.org/jira/browse/FLINK-18153
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Coordination
Affects Versions: 1.11.0
Reporter: Chesnay Schepler


Configuring obscenely large amounts of memory that exceed the machine, which 
prevents the JMV startup, does not result in any error to be printed to the 
console. The logs are empty, and the only hint for an error is found in the 
.out file.

{code}
OpenJDK 64-Bit Server VM warning: INFO: os::commit_memory(0x7aaf7000, 
5241403604992, 0) failed; error='Not enough space' (errno=12)
#
# There is insufficient memory for the Java Runtime Environment to continue.
# Native memory allocation (mmap) failed to map 5241403604992 bytes for 
committing reserved memory.
{code}

This behavior is just a bit too subtle.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18155) Disable Log4j's JMX for (integration)tests

2020-06-05 Thread Robert Metzger (Jira)
Robert Metzger created FLINK-18155:
--

 Summary:  Disable Log4j's JMX for (integration)tests
 Key: FLINK-18155
 URL: https://issues.apache.org/jira/browse/FLINK-18155
 Project: Flink
  Issue Type: Improvement
  Components: Build System
Reporter: Robert Metzger


There were reports about Log4j2 initializing slowly when executing tests 
locally.
The root cause of this has not been understood yet, but we know that Log4j is 
starting a JXM server, which could be responsible for the slow startup.




--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18154) Unhelpful error message if heap.size takes up signficant portion of process.size

2020-06-05 Thread Chesnay Schepler (Jira)
Chesnay Schepler created FLINK-18154:


 Summary: Unhelpful error message if heap.size takes up signficant 
portion of process.size
 Key: FLINK-18154
 URL: https://issues.apache.org/jira/browse/FLINK-18154
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Coordination
Affects Versions: 1.11.0
Reporter: Chesnay Schepler


The process still starts fine though, similar to FLINK-18152.

{code}
[] - Loading configuration property: jobmanager.memory.process.size, 2000m
[] - Loading configuration property: jobmanager.memory.heap.size, 1999m
Exception in thread "main" java.lang.IllegalArgumentException: bytes must be >= 0
at 
org.apache.flink.util.Preconditions.checkArgument(Preconditions.java:139)
at org.apache.flink.configuration.MemorySize.(MemorySize.java:82)
at 
org.apache.flink.configuration.MemorySize.subtract(MemorySize.java:216)
at 
org.apache.flink.runtime.util.config.memory.ProcessMemoryUtils.deriveJvmMetaspaceAndOverheadFromTotalFlinkMemory(ProcessMemoryUtils.java:147)
at 
org.apache.flink.runtime.util.config.memory.ProcessMemoryUtils.deriveProcessSpecWithExplicitInternalMemory(ProcessMemoryUtils.java:86)
at 
org.apache.flink.runtime.util.config.memory.ProcessMemoryUtils.memoryProcessSpecFromConfig(ProcessMemoryUtils.java:71)
at 
org.apache.flink.runtime.jobmanager.JobManagerProcessUtils.processSpecFromConfig(JobManagerProcessUtils.java:76)
at 
org.apache.flink.runtime.jobmanager.JobManagerProcessUtils.processSpecFromConfigWithNewOptionToInterpretLegacyHeap(JobManagerProcessUtils.java:71)
at 
org.apache.flink.runtime.util.bash.BashJavaUtils.getJmResourceParams(BashJavaUtils.java:92)
at 
org.apache.flink.runtime.util.bash.BashJavaUtils.runCommand(BashJavaUtils.java:66)
at 
org.apache.flink.runtime.util.bash.BashJavaUtils.main(BashJavaUtils.java:54)
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18156) Misleading error message if derived JVM overhead is too small

2020-06-05 Thread Chesnay Schepler (Jira)
Chesnay Schepler created FLINK-18156:


 Summary: Misleading error message if derived JVM overhead is too 
small
 Key: FLINK-18156
 URL: https://issues.apache.org/jira/browse/FLINK-18156
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Coordination
Affects Versions: 1.11.0
Reporter: Chesnay Schepler


The error message says that the derived size is lower then the configured 
overhead range, but now such property was configured by the user.

{code}
[] - Loading configuration property: jobmanager.memory.process.size, 2000m
[] - Loading configuration property: jobmanager.memory.heap.size, 1500m
Exception in thread "main" 
org.apache.flink.configuration.IllegalConfigurationException: Derived JVM 
Overhead size (116.000mb (121634816 bytes)) is not in configured JVM Overhead 
range [192.000mb (201326592 bytes), 1024.000mb (107374182$at 
org.apache.flink.runtime.util.config.memory.ProcessMemoryUtils.sanityCheckJvmOverhead(ProcessMemoryUtils.java:171)
at 
org.apache.flink.runtime.util.config.memory.ProcessMemoryUtils.deriveJvmMetaspaceAndOverheadFromTotalFlinkMemory(ProcessMemoryUtils.java:148)
at 
org.apache.flink.runtime.util.config.memory.ProcessMemoryUtils.deriveProcessSpecWithExplicitInternalMemory(ProcessMemoryUtils.java:86)
at 
org.apache.flink.runtime.util.config.memory.ProcessMemoryUtils.memoryProcessSpecFromConfig(ProcessMemoryUtils.java:71)
at 
org.apache.flink.runtime.jobmanager.JobManagerProcessUtils.processSpecFromConfig(JobManagerProcessUtils.java:76)
at 
org.apache.flink.runtime.jobmanager.JobManagerProcessUtils.processSpecFromConfigWithNewOptionToInterpretLegacyHeap(JobManagerProcessUtils.java:71)
at 
org.apache.flink.runtime.util.bash.BashJavaUtils.getJmResourceParams(BashJavaUtils.java:92)
at 
org.apache.flink.runtime.util.bash.BashJavaUtils.runCommand(BashJavaUtils.java:66)
at 
org.apache.flink.runtime.util.bash.BashJavaUtils.main(BashJavaUtils.java:54)
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18157) Jobstore size check compares against offHeapMemory

2020-06-05 Thread Chesnay Schepler (Jira)
Chesnay Schepler created FLINK-18157:


 Summary: Jobstore size check compares against offHeapMemory
 Key: FLINK-18157
 URL: https://issues.apache.org/jira/browse/FLINK-18157
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Coordination
Affects Versions: 1.11.0
Reporter: Chesnay Schepler


Setting {{jobmanager.memory.off-heap.size}} to 0 results in this confusing 
error:
{code}
[] - Loading configuration property: jobmanager.memory.process.size, 2000m
[] - Loading configuration property: jobmanager.memory.heap.size, 1500m
[] - Loading configuration property: jobmanager.memory.jvm-overhead.min, 100m
[] - Loading configuration property: jobmanager.memory.jvm-overhead.max, 350m
[] - Loading configuration property: jobmanager.memory.off-heap.size, 0m
[] - The configured or derived JVM heap memory size (jobstore.cache-size: 0 
bytes) is less than the configured or default size of the job store cache 
(jobmanager.memory.heap.size: 50.000mb (52428800 bytes))
{code}

According to the documentation the jobstore uses the heap though.

{code}
private static JobManagerFlinkMemory createJobManagerFlinkMemory(
Configuration config,
MemorySize jvmHeap,
MemorySize offHeapMemory) {
verifyJvmHeapSize(jvmHeap);
verifyJobStoreCacheSize(config, offHeapMemory);
return new JobManagerFlinkMemory(jvmHeap, offHeapMemory);
}

private static void verifyJvmHeapSize(MemorySize jvmHeapSize) {
if (jvmHeapSize.compareTo(JobManagerOptions.MIN_JVM_HEAP_SIZE) < 1) {
LOG.warn(
"The configured or derived JVM heap memory size ({}) is 
less than its recommended minimum value ({})",
jvmHeapSize.toHumanReadableString(),
JobManagerOptions.MIN_JVM_HEAP_SIZE);
}
}
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18158) Add a utility to create a DDL statement from avro schema

2020-06-05 Thread Dawid Wysakowicz (Jira)
Dawid Wysakowicz created FLINK-18158:


 Summary: Add a utility to create a DDL statement from avro schema
 Key: FLINK-18158
 URL: https://issues.apache.org/jira/browse/FLINK-18158
 Project: Flink
  Issue Type: Improvement
  Components: Table SQL / API
Reporter: Dawid Wysakowicz


User asked if there is a way to create a TableSchema/Table originating from 
avro schema. 
https://lists.apache.org/thread.html/r9bd43449314230fad0b627a170db05284c9727371092fc275fc05b74%40%3Cuser.flink.apache.org%3E



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18159) Add support for trimming stack traces in user-facing components

2020-06-05 Thread Chesnay Schepler (Jira)
Chesnay Schepler created FLINK-18159:


 Summary: Add support for trimming stack traces in user-facing 
components
 Key: FLINK-18159
 URL: https://issues.apache.org/jira/browse/FLINK-18159
 Project: Flink
  Issue Type: New Feature
  Components: Command Line Client, Runtime / REST
Reporter: Chesnay Schepler
 Fix For: 1.12.0


Add a verbosity query parameter / flag(==config option) into the REST API / CLI 
to trim the exception stack traces, which means excluding the location 
information.

The result would be something like this:
{code}
org.apache.flink.runtime.client.JobSubmissionException: Failed to submit job.
Caused by: java.lang.RuntimeException: 
org.apache.flink.runtime.client.JobExecutionException: Could not set up 
JobManager
Caused by: org.apache.flink.runtime.client.JobExecutionException: Could not set 
up JobManager
Caused by: java.io.FileNotFoundException: Cannot find checkpoint or savepoint 
file/directory 'ashudasd' on file system 'file'.
{code}

This approach renders even the biggest stack traces fairly readable, and is 
rather convenient since it only requires changes in the actual user-facing 
components.

Logging would not be impacted by this.

The trimming was already implemented in this 
[PR|https://github.com/apache/flink/pull/12392], but the flags are missing.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [ANNOUNCE] New Apache Flink Committer - Xintong Song

2020-06-05 Thread Till Rohrmann
Congratulations!

Cheers,
Till

On Fri, Jun 5, 2020 at 10:00 AM Dawid Wysakowicz 
wrote:

> Congratulations!
>
> Best,
>
> Dawid
>
> On 05/06/2020 09:10, tison wrote:
> > Congrats, Xintong!
> >
> > Best,
> > tison.
> >
> >
> > Jark Wu  于2020年6月5日周五 下午3:00写道:
> >
> >> Congratulations Xintong!
> >>
> >> Best,
> >> Jark
> >>
> >> On Fri, 5 Jun 2020 at 14:32, Danny Chan  wrote:
> >>
> >>> Congratulations Xintong !
> >>>
> >>> Best,
> >>> Danny Chan
> >>> 在 2020年6月5日 +0800 PM2:20,dev@flink.apache.org,写道:
>  Congratulations Xintong
>


[jira] [Created] (FLINK-18160) YARN session logs about HADOOP_CONF_DIR even though HADOOP_CLASSPATH containing a config is set

2020-06-05 Thread Robert Metzger (Jira)
Robert Metzger created FLINK-18160:
--

 Summary: YARN session logs about HADOOP_CONF_DIR even though 
HADOOP_CLASSPATH containing a config is set
 Key: FLINK-18160
 URL: https://issues.apache.org/jira/browse/FLINK-18160
 Project: Flink
  Issue Type: Improvement
  Components: Deployment / YARN
Affects Versions: 1.11.0
Reporter: Robert Metzger


Flink prints 
{code}
Setting HADOOP_CONF_DIR=/etc/hadoop/conf because no HADOOP_CONF_DIR was set.
{code}

When running Flink on YARN with the HADOOP_CLASSPATH set. "hadoop classpath" 
also returns the configuration directory, so HADOOP_CONF_DIR is not needed 
anymore.

I suggest to revisit this log message / behavior as it is misleading



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18161) Changing parallelism is not possible in sql-client.sh

2020-06-05 Thread Robert Metzger (Jira)
Robert Metzger created FLINK-18161:
--

 Summary: Changing parallelism is not possible in sql-client.sh
 Key: FLINK-18161
 URL: https://issues.apache.org/jira/browse/FLINK-18161
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / Client
Affects Versions: 1.11.0
Reporter: Robert Metzger


I tried using 
{code}
SET execution.parallelism=12
{code}

and changing the parallelism in the configuration file.

My SQL queries were always running with p=1 for all operators.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [ANNOUNCE] New Apache Flink Committer - Xintong Song

2020-06-05 Thread Andrey Zagrebin
Welcome to committers and congrats, Xintong!

Cheers,
Andrey

On Fri, Jun 5, 2020 at 4:22 PM Till Rohrmann  wrote:

> Congratulations!
>
> Cheers,
> Till
>
> On Fri, Jun 5, 2020 at 10:00 AM Dawid Wysakowicz 
> wrote:
>
> > Congratulations!
> >
> > Best,
> >
> > Dawid
> >
> > On 05/06/2020 09:10, tison wrote:
> > > Congrats, Xintong!
> > >
> > > Best,
> > > tison.
> > >
> > >
> > > Jark Wu  于2020年6月5日周五 下午3:00写道:
> > >
> > >> Congratulations Xintong!
> > >>
> > >> Best,
> > >> Jark
> > >>
> > >> On Fri, 5 Jun 2020 at 14:32, Danny Chan  wrote:
> > >>
> > >>> Congratulations Xintong !
> > >>>
> > >>> Best,
> > >>> Danny Chan
> > >>> 在 2020年6月5日 +0800 PM2:20,dev@flink.apache.org,写道:
> >  Congratulations Xintong
> >
>


[jira] [Created] (FLINK-18162) AddSplitEvents should serialize the splits into bytes.

2020-06-05 Thread Jiangjie Qin (Jira)
Jiangjie Qin created FLINK-18162:


 Summary: AddSplitEvents should serialize the splits into bytes.
 Key: FLINK-18162
 URL: https://issues.apache.org/jira/browse/FLINK-18162
 Project: Flink
  Issue Type: Bug
  Components: Connectors / Common
Reporter: Jiangjie Qin


{{AddSplitsEvent}} is a serializable at the moment which contains a list of 
splits. However, the {{SourceSplit}} is not a subclass of {{Serializable}}. We 
need to serialize the splits in the {{AddSplitsEvent}} using SplitSerializer 
before sending it over the wire and deserialize the splits on the reader side.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [ANNOUNCE] New Apache Flink Committer - Xintong Song

2020-06-05 Thread Zhijiang
Congratulations! 
Best,
Zhijiang 
--
From:Andrey Zagrebin 
Send Time:2020年6月5日(星期五) 22:34
To:dev 
Subject:Re: [ANNOUNCE] New Apache Flink Committer - Xintong Song
Welcome to committers and congrats, Xintong!

Cheers,
Andrey

On Fri, Jun 5, 2020 at 4:22 PM Till Rohrmann  wrote:

> Congratulations!
>
> Cheers,
> Till
>
> On Fri, Jun 5, 2020 at 10:00 AM Dawid Wysakowicz 
> wrote:
>
> > Congratulations!
> >
> > Best,
> >
> > Dawid
> >
> > On 05/06/2020 09:10, tison wrote:
> > > Congrats, Xintong!
> > >
> > > Best,
> > > tison.
> > >
> > >
> > > Jark Wu  于2020年6月5日周五 下午3:00写道:
> > >
> > >> Congratulations Xintong!
> > >>
> > >> Best,
> > >> Jark
> > >>
> > >> On Fri, 5 Jun 2020 at 14:32, Danny Chan  wrote:
> > >>
> > >>> Congratulations Xintong !
> > >>>
> > >>> Best,
> > >>> Danny Chan
> > >>> 在 2020年6月5日 +0800 PM2:20,dev@flink.apache.org,写道:
> >  Congratulations Xintong
> >
>



[jira] [Created] (FLINK-18163) Should be volatile: network.api.writer.RecordWriter.flusherException

2020-06-05 Thread Roman Khachatryan (Jira)
Roman Khachatryan created FLINK-18163:
-

 Summary: Should be volatile: 
network.api.writer.RecordWriter.flusherException 
 Key: FLINK-18163
 URL: https://issues.apache.org/jira/browse/FLINK-18163
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Task
Affects Versions: 1.11.0
Reporter: Roman Khachatryan
Assignee: Roman Khachatryan
 Fix For: 1.12.0






--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-18164) null <> 'str' should be true

2020-06-05 Thread Benchao Li (Jira)
Benchao Li created FLINK-18164:
--

 Summary: null <> 'str' should be true
 Key: FLINK-18164
 URL: https://issues.apache.org/jira/browse/FLINK-18164
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / API
Reporter: Benchao Li


Currently, if we compare null with other literals, the result will always be 
false.
It's because the code gen always gives a default value (false) for the result. 
And I think it's a bug if `null <> 'str'` is false.

It's reported from user-zh: 
http://apache-flink.147419.n8.nabble.com/flink-sql-null-false-td3640.html

CC [~jark] [~ykt836]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


Re: [ANNOUNCE] New Apache Flink Committer - Xintong Song

2020-06-05 Thread Matt Wang
Congratulations!


---
Best,
Matt Wang


On 06/5/2020 22:34,Andrey Zagrebin wrote:
Welcome to committers and congrats, Xintong!

Cheers,
Andrey

On Fri, Jun 5, 2020 at 4:22 PM Till Rohrmann  wrote:

Congratulations!

Cheers,
Till

On Fri, Jun 5, 2020 at 10:00 AM Dawid Wysakowicz 
wrote:

Congratulations!

Best,

Dawid

On 05/06/2020 09:10, tison wrote:
Congrats, Xintong!

Best,
tison.


Jark Wu  于2020年6月5日周五 下午3:00写道:

Congratulations Xintong!

Best,
Jark

On Fri, 5 Jun 2020 at 14:32, Danny Chan  wrote:

Congratulations Xintong !

Best,
Danny Chan
在 2020年6月5日 +0800 PM2:20,dev@flink.apache.org,写道:
Congratulations Xintong




Re: [DISCUSS] (Document) Backwards Compatibility of Savepoints

2020-06-05 Thread Congxian Qiu
Sorry for jumping in late.

Currently, we only have a forward-compatible guarantee and do not have the
backward-compatible guarantee. And as this may take a large effort to
support the backward-compatible guarantee. so I agree that we should write
this down explicitly.

For the given scenario, I have a little question: Why do we want to restore
from the savepoint taken the new Flink version instead of the previous
savepoint, is that we want to minimize the source rewind?

Best,
Congxian


Steven Wu  于2020年6月3日周三 上午9:08写道:

> Current Flink documentation is actually pretty clear about no guarantees
> for backward compatibility.
>
> https://ci.apache.org/projects/flink/flink-docs-stable/ops/upgrading.html#compatibility-table
>
> On Tue, Jun 2, 2020 at 3:20 AM Yun Tang  wrote:
>
> > Since Flink lacks of such kind of experiments to ensure the backwards
> > compatibility of savepoints before, especially those built-in operators
> > with their own operator state.
> > I am afraid we need huge energy to cover all cases to give the most
> > correct result.
> >
> > I prefer to just point out this in documentation to say explicitly Flink
> > does not guarantee such kind of backwards compatibility.
> >
> > Best
> > Yun Tang
> > 
> > From: Ufuk Celebi 
> > Sent: Wednesday, May 27, 2020 16:42
> > To: dev@flink.apache.org 
> > Subject: Re: [DISCUSS] (Document) Backwards Compatibility of Savepoints
> >
> > I agree with Konstantin and Steven that it makes sense to point this out
> > explicitly.
> >
> > I think that the following would be helpful:
> >
> > 1/ Mention breaking compatibility in release notes
> >
> > 2/ Update the linked table to reflect compatibilities while pointing out
> > what the community commits to maintain going forward (e.g. "happens to
> > work" vs. "guaranteed to work")
> >
> > In general, the table is quite large. Would it make sense to order the
> > releases in reverse order (assuming that the table is more relevant for
> > recent releases)?
> >
> > – Ufuk
> >
> > On Tue, May 26, 2020 at 8:36 PM Steven Wu  wrote:
> >
> > > > A use case for this might be when you want to rollback a framework
> > > upgrade (after some time) due to e.g. a performance
> > > or stability issue.
> > >
> > > Downgrade (that Konstantin called out) is an important and realistic
> > > scenario. It will be great to support backward compatibility for
> > savepoint
> > > or at least document any breaking change.
> > >
> > > On Tue, May 26, 2020 at 4:39 AM Piotr Nowojski 
> > > wrote:
> > >
> > > > Hi,
> > > >
> > > > It might have been implicit choice, but so far we were not supporting
> > the
> > > > scenario that you are asking for. It has never been tested and we
> have
> > > > lot’s of state migration code sprinkled among our code base (for
> > example
> > > > upgrading state fields of the operators like [1]), that only supports
> > > > upgrades, not downgrades.
> > > >
> > > > Also we do not have testing infrastructure for checking the
> downgrades.
> > > We
> > > > would need to check if save points taken from master branch, are
> > readable
> > > > by previous releases (not release branch!).
> > > >
> > > > So all in all, I don’t think it can be easily done. It would require
> > some
> > > > effort to start maintaining backward compatibility.
> > > >
> > > > Piotrek
> > > >
> > > > [1]
> > > >
> > >
> >
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer011#migrateNextTransactionalIdHindState
> > > >
> > > > > On 26 May 2020, at 13:18, Konstantin Knauf 
> > wrote:
> > > > >
> > > > > Hi everyone,
> > > > >
> > > > > I recently stumbled across the fact that Savepoints created with
> > Flink
> > > > 1.11
> > > > > can not be read by Flink 1.10. A use case for this might be when
> you
> > > want
> > > > > to rollback a framework upgrade (after some time) due to e.g. a
> > > > performance
> > > > > or stability issue.
> > > > >
> > > > > From the documentation [1] it seems as if the Savepoint format is
> > > > generally
> > > > > only forward-compatible although in many cases it is actually also
> > > > > backwards compatible (e.g. Savepoint taken in Flink 1.10, restored
> > with
> > > > > Flink 1.9).
> > > > >
> > > > > Was it a deliberate choice not to document any backwards
> > compatibility?
> > > > If
> > > > > not, should we add the missing entries in the compatibility table?
> > > > >
> > > > > Thanks,
> > > > >
> > > > > Konstantin
> > > > >
> > > > > [1]
> > > > >
> > > >
> > >
> >
> https://ci.apache.org/projects/flink/flink-docs-master/ops/upgrading.html#compatibility-table
> > > > >
> > > > > --
> > > > >
> > > > > Konstantin Knauf
> > > > >
> > > > > https://twitter.com/snntrable
> > > > >
> > > > > https://github.com/knaufk
> > > >
> > > >
> > >
> >
>