[jira] [Updated] (FLINK-15983) add native reader for Hive parquet files

2020-03-23 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15983?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-15983:
---
Fix Version/s: (was: 1.11.0)

> add native reader for Hive parquet files
> 
>
> Key: FLINK-15983
> URL: https://issues.apache.org/jira/browse/FLINK-15983
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / Hive
>Reporter: Bowen Li
>Assignee: Jingsong Lee
>Priority: Critical
>
> add native reader for Hive parquet files, as well as benchmark that show how 
> faster the native reader is compared to generic hive reader, and publish a 
> blog about it
>  
> cc [~ykt836] [~lirui]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-15984) support hive stream table sink

2020-03-23 Thread Kurt Young (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15984?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kurt Young updated FLINK-15984:
---
Fix Version/s: (was: 1.11.0)

> support hive stream table sink
> --
>
> Key: FLINK-15984
> URL: https://issues.apache.org/jira/browse/FLINK-15984
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / Hive
>Reporter: Bowen Li
>Priority: Major
>
> support hive stream table sink for stream processing



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Resolved] (FLINK-13890) HiveCatalogUseBlinkITCase failed to get metastore connection

2020-03-23 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13890?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee resolved FLINK-13890.
--
Resolution: Duplicate

Feel free to re-open if the problem is re-produced.

> HiveCatalogUseBlinkITCase failed to get metastore connection
> 
>
> Key: FLINK-13890
> URL: https://issues.apache.org/jira/browse/FLINK-13890
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive, Tests
>Affects Versions: 1.10.0
>Reporter: Piotr Nowojski
>Priority: Critical
>
> Failed on master with
> {code:java}
> Could not connect to meta store using any of the URIs provided. Most recent 
> failure: org.apache.thrift.transport.TTransportException: 
> java.net.ConnectException: Connection refused (Connection refused)
> {code}
> https://api.travis-ci.org/v3/job/578208116/log.txt
> https://travis-ci.org/apache/flink/builds/578208104
> CC [~lzljs3620320] [~ykt836]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] hequn8128 commented on a change in pull request #11444: [FLINK-16608][python] Support primitive data types for vectorized Python UDF

2020-03-23 Thread GitBox
hequn8128 commented on a change in pull request #11444: [FLINK-16608][python] 
Support primitive data types for vectorized Python UDF
URL: https://github.com/apache/flink/pull/11444#discussion_r396243233
 
 

 ##
 File path: 
flink-python/src/main/java/org/apache/flink/table/runtime/arrow/writers/DateWriter.java
 ##
 @@ -0,0 +1,48 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.runtime.arrow.writers;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.table.runtime.functions.SqlDateTimeUtils;
+import org.apache.flink.types.Row;
+
+import org.apache.arrow.vector.DateDayVector;
+
+import java.sql.Date;
+
+/**
+ * {@link ArrowFieldWriter} for Date.
+ */
+@Internal
+public final class DateWriter extends ArrowFieldWriter {
+
+   public DateWriter(DateDayVector dateDayVector) {
+   super(dateDayVector);
+   }
+
+   @Override
+   public void doWrite(Row value, int ordinal) {
+   if (value.getField(ordinal) == null) {
+   ((DateDayVector) getValueVector()).setNull(getCount());
+   } else {
+   ((DateDayVector) getValueVector()).setSafe(
+   getCount(), 
SqlDateTimeUtils.dateToInternal(((Date) value.getField(ordinal;
 
 Review comment:
   `DateWriter` is used for the old planner. It seems we can't use the blink 
planner class directly here, otherwise, Pyflink will depend on blink planner 
even it is configured to use the old planner.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] hequn8128 commented on a change in pull request #11444: [FLINK-16608][python] Support primitive data types for vectorized Python UDF

2020-03-23 Thread GitBox
hequn8128 commented on a change in pull request #11444: [FLINK-16608][python] 
Support primitive data types for vectorized Python UDF
URL: https://github.com/apache/flink/pull/11444#discussion_r396237402
 
 

 ##
 File path: 
flink-python/src/main/java/org/apache/flink/table/runtime/arrow/writers/BaseRowDecimalWriter.java
 ##
 @@ -0,0 +1,50 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.runtime.arrow.writers;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.table.dataformat.BaseRow;
+
+import org.apache.arrow.vector.DecimalVector;
+
+/**
+ * {@link ArrowFieldWriter} for Decimal.
+ */
+@Internal
+public final class BaseRowDecimalWriter extends ArrowFieldWriter {
+
+   private final int precision;
+   private final int scale;
+
+   public BaseRowDecimalWriter(DecimalVector decimalVector, int precision, 
int scale) {
+   super(decimalVector);
+   this.precision = precision;
+   this.scale = scale;
+   }
+
+   @Override
+   public void doWrite(BaseRow row, int ordinal) {
+   if (row.isNullAt(ordinal)) {
+   ((DecimalVector) getValueVector()).setNull(getCount());
+   } else {
+   ((DecimalVector) getValueVector()).setSafe(
 
 Review comment:
   Should we also adjust the precision and scale here, like it in 
`DecimalWriter`? If we pass a decimal with different precision and scale, an 
exception will be thrown. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] hequn8128 commented on a change in pull request #11444: [FLINK-16608][python] Support primitive data types for vectorized Python UDF

2020-03-23 Thread GitBox
hequn8128 commented on a change in pull request #11444: [FLINK-16608][python] 
Support primitive data types for vectorized Python UDF
URL: https://github.com/apache/flink/pull/11444#discussion_r396237588
 
 

 ##
 File path: 
flink-python/src/main/java/org/apache/flink/table/runtime/arrow/writers/BaseRowDecimalWriter.java
 ##
 @@ -0,0 +1,50 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.table.runtime.arrow.writers;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.table.dataformat.BaseRow;
+
+import org.apache.arrow.vector.DecimalVector;
+
+/**
+ * {@link ArrowFieldWriter} for Decimal.
+ */
+@Internal
+public final class BaseRowDecimalWriter extends ArrowFieldWriter {
+
+   private final int precision;
+   private final int scale;
+
+   public BaseRowDecimalWriter(DecimalVector decimalVector, int precision, 
int scale) {
+   super(decimalVector);
+   this.precision = precision;
+   this.scale = scale;
+   }
+
+   @Override
+   public void doWrite(BaseRow row, int ordinal) {
+   if (row.isNullAt(ordinal)) {
+   ((DecimalVector) getValueVector()).setNull(getCount());
+   } else {
+   ((DecimalVector) getValueVector()).setSafe(
 
 Review comment:
   Should we also adjust the precision and scale here, like it in 
`DecimalWriter`? If we pass a decimal with different precision and scale for 
the BaseRowDecimalWriter, an exception will be thrown. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-16708) When a JDBC connection has been closed, the retry policy of the JDBCUpsertOutputFormat cannot take effect and may result in data loss

2020-03-23 Thread Weike Dong (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16708?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064580#comment-17064580
 ] 

Weike Dong commented on FLINK-16708:


Hi [~lzljs3620320] , do you have time to look into this issue? Thanks 

> When a JDBC connection has been closed, the retry policy of the 
> JDBCUpsertOutputFormat cannot take effect and may result in data loss
> -
>
> Key: FLINK-16708
> URL: https://issues.apache.org/jira/browse/FLINK-16708
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / JDBC
>Affects Versions: 1.10.0
>Reporter: tangshangwen
>Assignee: tangshangwen
>Priority: Major
>
> In our test environment, I used the tcpkill command to simulate a scenario 
> where the postgresql connection was closed. I found that the retry strategy 
> of the flush method did not take effect, and when it retried the second time, 
> it could not recognize that the connection had been closed because Before the 
> first check whether the connection is closed, the batchStatements of 
> PgStatement have been cleared, which causes the second execution to check 
> that the batchStatements are empty and return normally.
> {code:java}
> 2020-03-20 21:16:18.246 [jdbc-upsert-output-format-thread-1] ERROR 
> org.apache.flink.api.java.io.jdbc.JDBCUpsertOutputFormat  - JDBC executeBatch 
> error, retry times = 1
>  org.postgresql.util.PSQLException: This connection has been closed.
>  at 
> org.postgresql.jdbc.PgConnection.checkClosed(PgConnection.java:857)
>  at 
> org.postgresql.jdbc.PgConnection.getAutoCommit(PgConnection.java:817)
>  at 
> org.postgresql.jdbc.PgStatement.internalExecuteBatch(PgStatement.java:813)
>  at org.postgresql.jdbc.PgStatement.executeBatch(PgStatement.java:873)
>  at 
> org.postgresql.jdbc.PgPreparedStatement.executeBatch(PgPreparedStatement.java:1569)
>  at 
> org.apache.flink.api.java.io.jdbc.writer.AppendOnlyWriter.executeBatch(AppendOnlyWriter.java:62)
>  at 
> org.apache.flink.api.java.io.jdbc.JDBCUpsertOutputFormat.flush(JDBCUpsertOutputFormat.java:159)
>  at 
> org.apache.flink.api.java.io.jdbc.JDBCUpsertOutputFormat.lambda$open$0(JDBCUpsertOutputFormat.java:124)
>  at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>  at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
>  at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
>  at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  at java.lang.Thread.run(Thread.java:748)
>  2020-03-20 21:16:21.247 [jdbc-upsert-output-format-thread-1] ERROR 
> org.apache.flink.api.java.io.jdbc.JDBCUpsertOutputFormat  - JDBC executeBatch 
> error, retry times = 1
>  org.postgresql.util.PSQLException: This connection has been closed.
>  at 
> org.postgresql.jdbc.PgConnection.checkClosed(PgConnection.java:857)
>  at 
> org.postgresql.jdbc.PgConnection.getAutoCommit(PgConnection.java:817)
>  at 
> org.postgresql.jdbc.PgStatement.internalExecuteBatch(PgStatement.java:813)
>  at org.postgresql.jdbc.PgStatement.executeBatch(PgStatement.java:873)
>  at 
> org.postgresql.jdbc.PgPreparedStatement.executeBatch(PgPreparedStatement.java:1569)
>  at 
> org.apache.flink.api.java.io.jdbc.writer.AppendOnlyWriter.executeBatch(AppendOnlyWriter.java:62)
>  at 
> org.apache.flink.api.java.io.jdbc.JDBCUpsertOutputFormat.flush(JDBCUpsertOutputFormat.java:159)
>  at 
> org.apache.flink.api.java.io.jdbc.JDBCUpsertOutputFormat.lambda$open$0(JDBCUpsertOutputFormat.java:124)
>  at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>  at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
>  at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
>  at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  at java.lang.Thread.run(Thread.java:748)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-13420) Supported hive versions are hard coded

2020-03-23 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee closed FLINK-13420.

Resolution: Invalid

> Supported hive versions are hard coded
> --
>
> Key: FLINK-13420
> URL: https://issues.apache.org/jira/browse/FLINK-13420
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Hive
>Affects Versions: 1.9.0
>Reporter: Jeff Zhang
>Priority: Major
>  Labels: pull-request-available
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Currently, the supported hive versions are hardcoded in 
> [HiveShimLoader|[https://github.com/apache/flink/blob/master/flink-connectors/flink-connector-hive/src/main/java/org/apache/flink/table/catalog/hive/client/HiveShimLoader.java]]
>  That makes me unable to run hive connector for 1.2.2, I believe we only need 
> to check the major.minor version number, but ignore the bugfix version number.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] TisonKun closed pull request #11415: [FLINK-15667][k8s] Support to mount custom Hadoop configurations

2020-03-23 Thread GitBox
TisonKun closed pull request #11415: [FLINK-15667][k8s] Support to mount custom 
Hadoop configurations
URL: https://github.com/apache/flink/pull/11415
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-15667) Support to mount Hadoop Configurations

2020-03-23 Thread Zili Chen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15667?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064581#comment-17064581
 ] 

Zili Chen commented on FLINK-15667:
---

master via f56a075aa029832b9cbacab51649c900498e18e9

> Support to mount Hadoop Configurations
> --
>
> Key: FLINK-15667
> URL: https://issues.apache.org/jira/browse/FLINK-15667
> Project: Flink
>  Issue Type: Sub-task
>  Components: Deployment / Kubernetes
>Reporter: Canbin Zheng
>Assignee: Canbin Zheng
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Mounts the Hadoop configurations, either as a pre-defined config map, or a 
> local configuration directory on the Pod.
> We plan to provide two options for the users to mount custom Hadoop 
> configurations:
>  * option 1: specify an existing ConfigMap that contains custom Hadoop 
> Configurations, one can share a single ConfigMap for more than one Flink 
> clusters.
>  * option 2: create a dedicated ConfigMap containing Hadoop Configurations 
> loaded from the local directory specified by the *HADOOP_CONF_DIR* or 
> *HADOOP_HOME* environment, then bind that ConfigMap to the lifecycle of the 
> new Flink cluster.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-15667) Support to mount Hadoop Configurations

2020-03-23 Thread Zili Chen (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15667?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zili Chen closed FLINK-15667.
-
Resolution: Fixed

> Support to mount Hadoop Configurations
> --
>
> Key: FLINK-15667
> URL: https://issues.apache.org/jira/browse/FLINK-15667
> Project: Flink
>  Issue Type: Sub-task
>  Components: Deployment / Kubernetes
>Reporter: Canbin Zheng
>Assignee: Canbin Zheng
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Mounts the Hadoop configurations, either as a pre-defined config map, or a 
> local configuration directory on the Pod.
> We plan to provide two options for the users to mount custom Hadoop 
> configurations:
>  * option 1: specify an existing ConfigMap that contains custom Hadoop 
> Configurations, one can share a single ConfigMap for more than one Flink 
> clusters.
>  * option 2: create a dedicated ConfigMap containing Hadoop Configurations 
> loaded from the local directory specified by the *HADOOP_CONF_DIR* or 
> *HADOOP_HOME* environment, then bind that ConfigMap to the lifecycle of the 
> new Flink cluster.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #10973: [FLINK-15647][k8s] Support user-specified annotations for JM/TM Pods

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #10973: [FLINK-15647][k8s] Support 
user-specified annotations for JM/TM Pods
URL: https://github.com/apache/flink/pull/10973#issuecomment-580223229
 
 
   
   ## CI report:
   
   * 33776e6aa098656ddb2820bbac39b4b08f9fa3f8 Travis: 
[CANCELED](https://travis-ci.com/github/flink-ci/flink/builds/154537476) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6505)
 
   * 3bfbfa984bf69d1abf17594aabd4b0e3ec1f782b UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-15932) Add download url to hive dependencies

2020-03-23 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-15932?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee updated FLINK-15932:
-
Fix Version/s: (was: 1.11.0)

> Add download url to hive dependencies
> -
>
> Key: FLINK-15932
> URL: https://issues.apache.org/jira/browse/FLINK-15932
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Hive, Documentation
>Reporter: Jingsong Lee
>Priority: Major
>
> Now in 
> [https://ci.apache.org/projects/flink/flink-docs-master/dev/table/hive/#dependencies]
> We list all dependencies for all supported hive versions. But no download 
> url, it is not so handy.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16405) Fix hive-metastore dependency description in connecting to hive docs

2020-03-23 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16405?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee updated FLINK-16405:
-
Fix Version/s: (was: 1.10.1)
   (was: 1.11.0)

> Fix hive-metastore dependency description in connecting to hive docs
> 
>
> Key: FLINK-16405
> URL: https://issues.apache.org/jira/browse/FLINK-16405
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive, Documentation
>Reporter: Leonard Xu
>Priority: Major
>
>  we need hive-metastore dependency when connect to hive(version 3.1.x) if 
> using hive-metastore in embedded mode. but looks like we missed it: 
> [https://ci.apache.org/projects/flink/flink-docs-master/dev/table/hive/#connecting-to-hive]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-13278) add Chinese documentation for Hive source/sink

2020-03-23 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13278?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee updated FLINK-13278:
-
Parent: (was: FLINK-10729)
Issue Type: Task  (was: Sub-task)

> add Chinese documentation for Hive source/sink
> --
>
> Key: FLINK-13278
> URL: https://issues.apache.org/jira/browse/FLINK-13278
> Project: Flink
>  Issue Type: Task
>  Components: Connectors / Hive, Documentation
>Reporter: Bowen Li
>Assignee: Zijie Lu
>Priority: Major
> Fix For: 1.9.3
>
>
> add Chinese documentation of Hive source/sink in {{batch/connector_zh.md}}
> it's corresponding English one is FLINK-13277
> cc [~xuefuz] [~lirui] [~Terry1897]



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16475) PulsarConsumerSource messageReceiveTimeoutMs parameter should be configurable

2020-03-23 Thread Robert Metzger (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064583#comment-17064583
 ] 

Robert Metzger commented on FLINK-16475:


No worries, and thanks for referring to the new ticket!

> PulsarConsumerSource messageReceiveTimeoutMs parameter should be configurable
> -
>
> Key: FLINK-16475
> URL: https://issues.apache.org/jira/browse/FLINK-16475
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.9.2
> Environment: Not releveant
>Reporter: Jason Kania
>Priority: Major
>
> The messageReceiveTimeoutMs value in the PulsarConsumerSource class is 
> hardcoded to 100ms but should be configurable to accommodate different 
> hardware setups such as developer labs where the performance level may not be 
> as critical.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-16713) Flink can not support the type of source table to query data from es

2020-03-23 Thread jackray wang (Jira)
jackray wang created FLINK-16713:


 Summary: Flink can not support the type of source table to query 
data from es
 Key: FLINK-16713
 URL: https://issues.apache.org/jira/browse/FLINK-16713
 Project: Flink
  Issue Type: Improvement
  Components: Table SQL / Client
Affects Versions: 1.10.0
Reporter: jackray wang


[https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/table/connect.html#elasticsearch-connector]

For append-only queries, the connector can also operate in [append 
mode|https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/table/connect.html#update-modes]
 for exchanging only INSERT messages with the external system. If no key is 
defined by the query, a key is automatically generated by Elasticsearch.

I want to know ,why the connector of flink with ES just support sink but  
doesn't support source .Which version could add this feature to ?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-4407) Add DSL for specifying Window Triggers

2020-03-23 Thread Kostas Kloudas (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-4407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kostas Kloudas closed FLINK-4407.
-
Resolution: Won't Fix

> Add DSL for specifying Window Triggers
> --
>
> Key: FLINK-4407
> URL: https://issues.apache.org/jira/browse/FLINK-4407
> Project: Flink
>  Issue Type: New Feature
>  Components: API / DataStream
>Reporter: Kostas Kloudas
>Assignee: Kostas Kloudas
>Priority: Major
>  Labels: pull-request-available
>
> This issue refers to the implementation of the trigger DSL.
> The specification of the DSL has an open FLIP here:
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-9%3A+Trigger+DSL
> And is currently under discussion in the dev@ mailing list here:
> http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-FLIP-9-Trigger-DSL-td13065.html



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #11284: [FLINK-15911][runtime] Make Flink work with NAT.

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11284: [FLINK-15911][runtime] Make Flink 
work with NAT.
URL: https://github.com/apache/flink/pull/11284#issuecomment-593407935
 
 
   
   ## CI report:
   
   * 0be3fd11e11d708b254036e9e91fc71f1fd02181 Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/154513428) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6492)
 
   * ff670f742762f45214ca32df966e02f4dbe6f758 Travis: 
[PENDING](https://travis-ci.com/github/flink-ci/flink/builds/154538899) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6507)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Closed] (FLINK-10729) Create a Hive connector for Hive data access in Flink

2020-03-23 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-10729?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee closed FLINK-10729.

Resolution: Fixed

> Create a Hive connector for Hive data access in Flink
> -
>
> Key: FLINK-10729
> URL: https://issues.apache.org/jira/browse/FLINK-10729
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / Hive, Table SQL / Ecosystem
>Affects Versions: 1.6.2
>Reporter: Xuefu Zhang
>Assignee: Xuefu Zhang
>Priority: Major
>
> As part of Flink-Hive integration effort, it's important for Flink to access 
> (read/write) Hive data, which is the responsibility of Hive connector. While 
> there is a HCatalog data connector in the code base, it's not complete (i.e. 
> missing all connector related classes such as validators, etc.). Further, 
> HCatalog interface has many limitations such as accessing a subset of Hive 
> data, supporting a subset of Hive data types, etc. In addition, it's not 
> actively maintained. In fact, it's now only a sub-project in Hive.
> Therefore, here we propose a complete connector set for Hive tables, not via 
> HCatalog, but via direct Hive interface. HCatalog connector will be 
> deprecated.
> Please note that connector on Hive metadata is already covered in other 
> JIRAs, as {{HiveExternalCatalog}}.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-6215) Make the StatefulSequenceSource scalable.

2020-03-23 Thread Kostas Kloudas (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-6215?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kostas Kloudas reassigned FLINK-6215:
-

Assignee: (was: Kostas Kloudas)

> Make the StatefulSequenceSource scalable.
> -
>
> Key: FLINK-6215
> URL: https://issues.apache.org/jira/browse/FLINK-6215
> Project: Flink
>  Issue Type: Bug
>  Components: API / DataStream
>Affects Versions: 1.3.0
>Reporter: Kostas Kloudas
>Priority: Major
>  Labels: pull-request-available
>
> Currently the {{StatefulSequenceSource}} instantiates all the elements to 
> emit first and keeps them in memory. This is not scalable as for large 
> sequences of elements this can lead to out of memory exceptions.
> To solve this, we can pre-partition the sequence of elements based on the 
> {{maxParallelism}} parameter, and just keep state (to checkpoint) per such 
> partition.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-10729) Create a Hive connector for Hive data access in Flink

2020-03-23 Thread Jingsong Lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-10729?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064584#comment-17064584
 ] 

Jingsong Lee commented on FLINK-10729:
--

Close, Thanks for the great work [~xuefuz]

> Create a Hive connector for Hive data access in Flink
> -
>
> Key: FLINK-10729
> URL: https://issues.apache.org/jira/browse/FLINK-10729
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / Hive, Table SQL / Ecosystem
>Affects Versions: 1.6.2
>Reporter: Xuefu Zhang
>Assignee: Xuefu Zhang
>Priority: Major
>
> As part of Flink-Hive integration effort, it's important for Flink to access 
> (read/write) Hive data, which is the responsibility of Hive connector. While 
> there is a HCatalog data connector in the code base, it's not complete (i.e. 
> missing all connector related classes such as validators, etc.). Further, 
> HCatalog interface has many limitations such as accessing a subset of Hive 
> data, supporting a subset of Hive data types, etc. In addition, it's not 
> actively maintained. In fact, it's now only a sub-project in Hive.
> Therefore, here we propose a complete connector set for Hive tables, not via 
> HCatalog, but via direct Hive interface. HCatalog connector will be 
> deprecated.
> Please note that connector on Hive metadata is already covered in other 
> JIRAs, as {{HiveExternalCatalog}}.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Assigned] (FLINK-7771) Make the operator state queryable

2020-03-23 Thread Kostas Kloudas (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-7771?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kostas Kloudas reassigned FLINK-7771:
-

Assignee: (was: Kostas Kloudas)

> Make the operator state queryable
> -
>
> Key: FLINK-7771
> URL: https://issues.apache.org/jira/browse/FLINK-7771
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Queryable State
>Affects Versions: 1.4.0
>Reporter: Kostas Kloudas
>Priority: Major
>
> There seem to be some requests for making the operator (non-keyed) state 
> queryable. This means that the user will specify the *uuid* of the operator 
> and the *taskId*, and he will be able to access the state that corresponds to 
> that operator and for that specific task.
> This issue will serve to document the discussion on the topic, so that 
> everybody can participate.
> I also link [~till.rohrmann] and [~skonto] as he also mentioned that this 
> feature could be helpful.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #11481: [FLINK-13691][Table]Remove deprecated query config

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11481: [FLINK-13691][Table]Remove 
deprecated query config
URL: https://github.com/apache/flink/pull/11481#issuecomment-602174011
 
 
   
   ## CI report:
   
   * cb9eb2458536cf858d110b30c6481b3d6597c9bc Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/154532502) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6502)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11471: [FLINK-16692] flink joblistener can register from config

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11471: [FLINK-16692] flink joblistener can 
register from config
URL: https://github.com/apache/flink/pull/11471#issuecomment-602012274
 
 
   
   ## CI report:
   
   * ad9049971efa640a2b22ee8f258b0a0d7f64cc47 Travis: 
[CANCELED](https://travis-ci.com/github/flink-ci/flink/builds/154513464) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6493)
 
   * 7f0c7e868fce22437c1b4b7a02fece52afb70f4a UNKNOWN
   * 57f1b99ac6db6626e6c78cb7b39e2026b59d1282 UNKNOWN
   * 84606e23725fc6c688ce72da42d06b8f30ac4910 Travis: 
[PENDING](https://travis-ci.com/github/flink-ci/flink/builds/154537643) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6506)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11448: [FLINK-16666][python][table] Support new Python dependency configuration options in flink-java, flink-streaming-java and flink-table.

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11448: [FLINK-1][python][table] Support 
new Python dependency configuration options in flink-java, flink-streaming-java 
and flink-table.
URL: https://github.com/apache/flink/pull/11448#issuecomment-601042255
 
 
   
   ## CI report:
   
   * d595c9fddd3ac097d6861e53305ea7f7a558 Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/154041174) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6386)
 
   * 2821e4a2542bffceeef1b742f43a09cae359e8ba UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-16692) flink joblistener can register from config

2020-03-23 Thread Kostas Kloudas (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16692?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064587#comment-17064587
 ] 

Kostas Kloudas commented on FLINK-16692:


Hi, I think this is a nice feature to have. I have 2 comments on the design 
here:
1) I believe that the option should go in the {{DeploymentOptions}} rather than 
the {{ExecutionOptions}}
2) the name has to be aligned with the names of the other parameters, so it has 
to be something like: {{execution.job-listeners}}.

> flink joblistener can register from config
> --
>
> Key: FLINK-16692
> URL: https://issues.apache.org/jira/browse/FLINK-16692
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Core
>Affects Versions: 1.10.0
>Reporter: jackylau
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> we should do as spark does ,which can register listener from conf such as 
> "spark.extraListeners"。 And it will be convinient for users when users just 
> want to set hook



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] azagrebin commented on issue #11445: [FLINK-16615] Introduce data structures and utilities to calculate Job Manager memory components

2020-03-23 Thread GitBox
azagrebin commented on issue #11445: [FLINK-16615] Introduce data structures 
and utilities to calculate Job Manager memory components
URL: https://github.com/apache/flink/pull/11445#issuecomment-602426159
 
 
   Thanks for the review @xintongsong 
   I refactored more the class structure. The final composition is a bit 
different but goes into a similar direction. This should also allow better test 
separation (probably later) comparing to the inheritance approach which is also 
in one of commits.
   I also addressed smaller comments.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-16708) When a JDBC connection has been closed, the retry policy of the JDBCUpsertOutputFormat cannot take effect and may result in data loss

2020-03-23 Thread Jingsong Lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16708?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064588#comment-17064588
 ] 

Jingsong Lee commented on FLINK-16708:
--

Hi, Thanks for reporting. Can you try FLINK-16281 ?

> When a JDBC connection has been closed, the retry policy of the 
> JDBCUpsertOutputFormat cannot take effect and may result in data loss
> -
>
> Key: FLINK-16708
> URL: https://issues.apache.org/jira/browse/FLINK-16708
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / JDBC
>Affects Versions: 1.10.0
>Reporter: tangshangwen
>Assignee: tangshangwen
>Priority: Major
>
> In our test environment, I used the tcpkill command to simulate a scenario 
> where the postgresql connection was closed. I found that the retry strategy 
> of the flush method did not take effect, and when it retried the second time, 
> it could not recognize that the connection had been closed because Before the 
> first check whether the connection is closed, the batchStatements of 
> PgStatement have been cleared, which causes the second execution to check 
> that the batchStatements are empty and return normally.
> {code:java}
> 2020-03-20 21:16:18.246 [jdbc-upsert-output-format-thread-1] ERROR 
> org.apache.flink.api.java.io.jdbc.JDBCUpsertOutputFormat  - JDBC executeBatch 
> error, retry times = 1
>  org.postgresql.util.PSQLException: This connection has been closed.
>  at 
> org.postgresql.jdbc.PgConnection.checkClosed(PgConnection.java:857)
>  at 
> org.postgresql.jdbc.PgConnection.getAutoCommit(PgConnection.java:817)
>  at 
> org.postgresql.jdbc.PgStatement.internalExecuteBatch(PgStatement.java:813)
>  at org.postgresql.jdbc.PgStatement.executeBatch(PgStatement.java:873)
>  at 
> org.postgresql.jdbc.PgPreparedStatement.executeBatch(PgPreparedStatement.java:1569)
>  at 
> org.apache.flink.api.java.io.jdbc.writer.AppendOnlyWriter.executeBatch(AppendOnlyWriter.java:62)
>  at 
> org.apache.flink.api.java.io.jdbc.JDBCUpsertOutputFormat.flush(JDBCUpsertOutputFormat.java:159)
>  at 
> org.apache.flink.api.java.io.jdbc.JDBCUpsertOutputFormat.lambda$open$0(JDBCUpsertOutputFormat.java:124)
>  at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>  at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
>  at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
>  at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  at java.lang.Thread.run(Thread.java:748)
>  2020-03-20 21:16:21.247 [jdbc-upsert-output-format-thread-1] ERROR 
> org.apache.flink.api.java.io.jdbc.JDBCUpsertOutputFormat  - JDBC executeBatch 
> error, retry times = 1
>  org.postgresql.util.PSQLException: This connection has been closed.
>  at 
> org.postgresql.jdbc.PgConnection.checkClosed(PgConnection.java:857)
>  at 
> org.postgresql.jdbc.PgConnection.getAutoCommit(PgConnection.java:817)
>  at 
> org.postgresql.jdbc.PgStatement.internalExecuteBatch(PgStatement.java:813)
>  at org.postgresql.jdbc.PgStatement.executeBatch(PgStatement.java:873)
>  at 
> org.postgresql.jdbc.PgPreparedStatement.executeBatch(PgPreparedStatement.java:1569)
>  at 
> org.apache.flink.api.java.io.jdbc.writer.AppendOnlyWriter.executeBatch(AppendOnlyWriter.java:62)
>  at 
> org.apache.flink.api.java.io.jdbc.JDBCUpsertOutputFormat.flush(JDBCUpsertOutputFormat.java:159)
>  at 
> org.apache.flink.api.java.io.jdbc.JDBCUpsertOutputFormat.lambda$open$0(JDBCUpsertOutputFormat.java:124)
>  at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>  at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
>  at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
>  at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
>  at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  at java.lang.Thread.run(Thread.java:748)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-13483) PrestoS3FileSystemITCase.testDirectoryListing fails on Travis

2020-03-23 Thread Piotr Nowojski (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Piotr Nowojski updated FLINK-13483:
---
Affects Version/s: 1.10.0

> PrestoS3FileSystemITCase.testDirectoryListing fails on Travis
> -
>
> Key: FLINK-13483
> URL: https://issues.apache.org/jira/browse/FLINK-13483
> Project: Flink
>  Issue Type: Bug
>  Components: FileSystems
>Affects Versions: 1.10.0
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Lu Niu
>Priority: Critical
>
> https://api.travis-ci.org/v3/job/564894421/log.txt
> {code}
> PrestoS3FileSystemITCase>AbstractHadoopFileSystemITTest.testDirectoryListing:144->AbstractHadoopFileSystemITTest.checkPathExistence:61
>  expected: but was:
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-13483) PrestoS3FileSystemITCase.testDirectoryListing fails on Travis

2020-03-23 Thread Piotr Nowojski (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Piotr Nowojski updated FLINK-13483:
---
Fix Version/s: 1.11.0

> PrestoS3FileSystemITCase.testDirectoryListing fails on Travis
> -
>
> Key: FLINK-13483
> URL: https://issues.apache.org/jira/browse/FLINK-13483
> Project: Flink
>  Issue Type: Bug
>  Components: FileSystems
>Affects Versions: 1.10.0
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Lu Niu
>Priority: Critical
> Fix For: 1.11.0
>
>
> https://api.travis-ci.org/v3/job/564894421/log.txt
> {code}
> PrestoS3FileSystemITCase>AbstractHadoopFileSystemITTest.testDirectoryListing:144->AbstractHadoopFileSystemITTest.checkPathExistence:61
>  expected: but was:
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-13483) PrestoS3FileSystemITCase.testDirectoryListing fails on Travis

2020-03-23 Thread Piotr Nowojski (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-13483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Piotr Nowojski updated FLINK-13483:
---
Labels: test-stability  (was: )

> PrestoS3FileSystemITCase.testDirectoryListing fails on Travis
> -
>
> Key: FLINK-13483
> URL: https://issues.apache.org/jira/browse/FLINK-13483
> Project: Flink
>  Issue Type: Bug
>  Components: FileSystems
>Affects Versions: 1.10.0
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Lu Niu
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.11.0
>
>
> https://api.travis-ci.org/v3/job/564894421/log.txt
> {code}
> PrestoS3FileSystemITCase>AbstractHadoopFileSystemITTest.testDirectoryListing:144->AbstractHadoopFileSystemITTest.checkPathExistence:61
>  expected: but was:
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16713) Support source mode of elasticsearch connector

2020-03-23 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16713?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee updated FLINK-16713:
-
Component/s: (was: Table SQL / Client)
 Connectors / ElasticSearch

> Support source mode of elasticsearch connector
> --
>
> Key: FLINK-16713
> URL: https://issues.apache.org/jira/browse/FLINK-16713
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / ElasticSearch
>Affects Versions: 1.10.0
>Reporter: jackray wang
>Priority: Major
>
> [https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/table/connect.html#elasticsearch-connector]
> For append-only queries, the connector can also operate in [append 
> mode|https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/table/connect.html#update-modes]
>  for exchanging only INSERT messages with the external system. If no key is 
> defined by the query, a key is automatically generated by Elasticsearch.
> I want to know ,why the connector of flink with ES just support sink but  
> doesn't support source .Which version could add this feature to ?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16713) Support source mode of elasticsearch connector

2020-03-23 Thread Jingsong Lee (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16713?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jingsong Lee updated FLINK-16713:
-
Summary: Support source mode of elasticsearch connector  (was: Flink can 
not support the type of source table to query data from es)

> Support source mode of elasticsearch connector
> --
>
> Key: FLINK-16713
> URL: https://issues.apache.org/jira/browse/FLINK-16713
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Client
>Affects Versions: 1.10.0
>Reporter: jackray wang
>Priority: Major
>
> [https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/table/connect.html#elasticsearch-connector]
> For append-only queries, the connector can also operate in [append 
> mode|https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/table/connect.html#update-modes]
>  for exchanging only INSERT messages with the external system. If no key is 
> defined by the query, a key is automatically generated by Elasticsearch.
> I want to know ,why the connector of flink with ES just support sink but  
> doesn't support source .Which version could add this feature to ?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16713) Support source mode of elasticsearch connector

2020-03-23 Thread Jingsong Lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16713?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064590#comment-17064590
 ] 

Jingsong Lee commented on FLINK-16713:
--

Hi [~jackray], unfortunately, as far as I know, there is no plan... Welcome to 
contribute.

> Support source mode of elasticsearch connector
> --
>
> Key: FLINK-16713
> URL: https://issues.apache.org/jira/browse/FLINK-16713
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / ElasticSearch
>Affects Versions: 1.10.0
>Reporter: jackray wang
>Priority: Major
>
> [https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/table/connect.html#elasticsearch-connector]
> For append-only queries, the connector can also operate in [append 
> mode|https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/table/connect.html#update-modes]
>  for exchanging only INSERT messages with the external system. If no key is 
> defined by the query, a key is automatically generated by Elasticsearch.
> I want to know ,why the connector of flink with ES just support sink but  
> doesn't support source .Which version could add this feature to ?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #11482: [FLINK-16581][table] Minibatch deduplication lack state TTL bug fix

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11482: [FLINK-16581][table] Minibatch 
deduplication lack state TTL bug fix
URL: https://github.com/apache/flink/pull/11482#issuecomment-60726
 
 
   
   ## CI report:
   
   * 95c74ce398cdc779ac895edb7cf3433579f1 Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/154529184) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6499)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] TisonKun commented on issue #11320: [FLINK-16437][runtime] Make SlotManager allocate resource from ResourceManager at the worker granularity.

2020-03-23 Thread GitBox
TisonKun commented on issue #11320: [FLINK-16437][runtime] Make SlotManager 
allocate resource from ResourceManager at the worker granularity.
URL: https://github.com/apache/flink/pull/11320#issuecomment-602429506
 
 
   @flinkbot run azure


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11448: [FLINK-16666][python][table] Support new Python dependency configuration options in flink-java, flink-streaming-java and flink-table.

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11448: [FLINK-1][python][table] Support 
new Python dependency configuration options in flink-java, flink-streaming-java 
and flink-table.
URL: https://github.com/apache/flink/pull/11448#issuecomment-601042255
 
 
   
   ## CI report:
   
   * d595c9fddd3ac097d6861e53305ea7f7a558 Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/154041174) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6386)
 
   * 2821e4a2542bffceeef1b742f43a09cae359e8ba Travis: 
[PENDING](https://travis-ci.com/github/flink-ci/flink/builds/154540657) Azure: 
[PENDING](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6510)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] WeiZhong94 commented on a change in pull request #11448: [FLINK-16666][python][table] Support new Python dependency configuration options in flink-java, flink-streaming-java and flink

2020-03-23 Thread GitBox
WeiZhong94 commented on a change in pull request #11448: 
[FLINK-1][python][table] Support new Python dependency configuration 
options in flink-java, flink-streaming-java and flink-table.
URL: https://github.com/apache/flink/pull/11448#discussion_r396254822
 
 

 ##
 File path: 
flink-java/src/main/java/org/apache/flink/api/java/utils/PythonDependencyManager.java
 ##
 @@ -0,0 +1,282 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.api.java.utils;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.api.common.cache.DistributedCache;
+import org.apache.flink.api.java.tuple.Tuple2;
+import org.apache.flink.configuration.ConfigOption;
+import org.apache.flink.configuration.ConfigOptions;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.configuration.PythonOptions;
+import org.apache.flink.configuration.ReadableConfig;
+import org.apache.flink.util.Preconditions;
+
+import java.io.File;
+import java.util.ArrayList;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+import java.util.UUID;
+
+import static 
org.apache.flink.configuration.PythonOptions.PYTHON_CLIENT_EXECUTABLE;
+import static org.apache.flink.configuration.PythonOptions.PYTHON_EXECUTABLE;
+
+/**
+ * Utility class for Python dependency management. The dependencies will be 
registered at the distributed
+ * cache.
+ */
+@Internal
+public class PythonDependencyManager {
+
+   public static final String PYTHON_FILE_PREFIX = "python_file";
+   public static final String PYTHON_REQUIREMENTS_FILE_PREFIX = 
"python_requirements_file";
+   public static final String PYTHON_REQUIREMENTS_CACHE_PREFIX = 
"python_requirements_cache";
+   public static final String PYTHON_ARCHIVE_PREFIX = "python_archive";
+   public static final String CLUSTER = "cluster";
+   public static final String FILE = "file";
+   public static final String CACHE = "cache";
+   public static final String FILE_DELIMITER = ",";
+   public static final String PARAM_DELIMITER = "#";
+
+   // Metadata of Python dependencies.
+
+   public static final ConfigOption> PYTHON_FILES =
+   
ConfigOptions.key("python.internal.files-key-map").mapType().noDefaultValue();
+   public static final ConfigOption> 
PYTHON_REQUIREMENTS_FILE =
+   
ConfigOptions.key("python.internal.requirements-file-key").mapType().noDefaultValue();
+   public static final ConfigOption> PYTHON_ARCHIVES =
+   
ConfigOptions.key("python.internal.archives-key-map").mapType().noDefaultValue();
+
+   private final Configuration config;
+   private final List> cachedFiles;
+   private int counter = -1;
+
+   public PythonDependencyManager(
+   Configuration config,
+   List> 
cachedFiles) {
+   this.config = Preconditions.checkNotNull(config);
+   this.cachedFiles = cachedFiles;
+   }
+
+   /**
+* Adds a Python dependency which could be .py files, Python 
packages(.zip, .egg etc.) or
+* local directories. The dependencies will be added to the PYTHONPATH 
of the Python UDF worker
+* and the local Py4J python client.
+*
+* @param filePath The path of the Python dependency.
+*/
+   public void addPythonFile(String filePath) {
 
 Review comment:
   In this PR this method is only public for test purpose. But in following PR 
the "add_python_file" method of Python TableEnvironment will point to this 
method so I think the annotation is unnecessary as it will be removed soon.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] kl0u closed pull request #11472: [FLINK-16547][yarn] Respect the config option of FileJobGraphRetriever#JOB_GRAPH_FILE_PATH

2020-03-23 Thread GitBox
kl0u closed pull request #11472: [FLINK-16547][yarn] Respect the config option 
of FileJobGraphRetriever#JOB_GRAPH_FILE_PATH
URL: https://github.com/apache/flink/pull/11472
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-16547) Respect the config option of FileJobGraphRetriever#JOB_GRAPH_FILE_PATH

2020-03-23 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16547?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-16547:
---
Labels: pull-request-available  (was: )

> Respect the config option of FileJobGraphRetriever#JOB_GRAPH_FILE_PATH
> --
>
> Key: FLINK-16547
> URL: https://issues.apache.org/jira/browse/FLINK-16547
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / YARN
>Affects Versions: 1.10.0
>Reporter: Canbin Zheng
>Assignee: Canbin Zheng
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> Currently, in {{YarnClusterDescriptor#startAppMaster}}, we first write out 
> and upload the Flink Configuration file, then start to write out the JobGraph 
> file and set its name into the Flink Configuration object, the afterward 
> setting is not written into the Flink Configuration file so that it does not 
> take effect in the cluster side.
> Since in the client-side we name the JobGraph file with the default value of 
> FileJobGraphRetriever.JOB_GRAPH_FILE_PATH option, the cluster side could 
> succeed in retrieving that file. 
> This ticket proposes to write out the JobGraph file before the Configuration 
> file to ensure that the setting of FileJobGraphRetriever.JOB_GRAPH_FILE_PATH 
> is delivered to the cluster side.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] WeiZhong94 commented on a change in pull request #11448: [FLINK-16666][python][table] Support new Python dependency configuration options in flink-java, flink-streaming-java and flink

2020-03-23 Thread GitBox
WeiZhong94 commented on a change in pull request #11448: 
[FLINK-1][python][table] Support new Python dependency configuration 
options in flink-java, flink-streaming-java and flink-table.
URL: https://github.com/apache/flink/pull/11448#discussion_r396257250
 
 

 ##
 File path: 
flink-java/src/main/java/org/apache/flink/api/java/utils/PythonDependencyManager.java
 ##
 @@ -0,0 +1,282 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.api.java.utils;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.api.common.cache.DistributedCache;
+import org.apache.flink.api.java.tuple.Tuple2;
+import org.apache.flink.configuration.ConfigOption;
+import org.apache.flink.configuration.ConfigOptions;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.configuration.PythonOptions;
+import org.apache.flink.configuration.ReadableConfig;
+import org.apache.flink.util.Preconditions;
+
+import java.io.File;
+import java.util.ArrayList;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+import java.util.UUID;
+
+import static 
org.apache.flink.configuration.PythonOptions.PYTHON_CLIENT_EXECUTABLE;
+import static org.apache.flink.configuration.PythonOptions.PYTHON_EXECUTABLE;
+
+/**
+ * Utility class for Python dependency management. The dependencies will be 
registered at the distributed
+ * cache.
+ */
+@Internal
+public class PythonDependencyManager {
+
+   public static final String PYTHON_FILE_PREFIX = "python_file";
+   public static final String PYTHON_REQUIREMENTS_FILE_PREFIX = 
"python_requirements_file";
+   public static final String PYTHON_REQUIREMENTS_CACHE_PREFIX = 
"python_requirements_cache";
+   public static final String PYTHON_ARCHIVE_PREFIX = "python_archive";
+   public static final String CLUSTER = "cluster";
+   public static final String FILE = "file";
+   public static final String CACHE = "cache";
+   public static final String FILE_DELIMITER = ",";
+   public static final String PARAM_DELIMITER = "#";
+
+   // Metadata of Python dependencies.
+
+   public static final ConfigOption> PYTHON_FILES =
+   
ConfigOptions.key("python.internal.files-key-map").mapType().noDefaultValue();
+   public static final ConfigOption> 
PYTHON_REQUIREMENTS_FILE =
+   
ConfigOptions.key("python.internal.requirements-file-key").mapType().noDefaultValue();
+   public static final ConfigOption> PYTHON_ARCHIVES =
+   
ConfigOptions.key("python.internal.archives-key-map").mapType().noDefaultValue();
+
+   private final Configuration config;
+   private final List> cachedFiles;
+   private int counter = -1;
+
+   public PythonDependencyManager(
+   Configuration config,
+   List> 
cachedFiles) {
+   this.config = Preconditions.checkNotNull(config);
+   this.cachedFiles = cachedFiles;
+   }
+
+   /**
+* Adds a Python dependency which could be .py files, Python 
packages(.zip, .egg etc.) or
+* local directories. The dependencies will be added to the PYTHONPATH 
of the Python UDF worker
+* and the local Py4J python client.
+*
+* @param filePath The path of the Python dependency.
+*/
+   public void addPythonFile(String filePath) {
+   Preconditions.checkNotNull(filePath);
+   if (fileAlreadyRegistered(filePath)) {
 
 Review comment:
   This case is expected because the `configure()` method will be called when 
the 
   `translate()` method of the Table API planners is called. As the 
`translate()` can be called many times in interactive programing, a python file 
will be registered multiple times.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] WeiZhong94 commented on a change in pull request #11448: [FLINK-16666][python][table] Support new Python dependency configuration options in flink-java, flink-streaming-java and flink

2020-03-23 Thread GitBox
WeiZhong94 commented on a change in pull request #11448: 
[FLINK-1][python][table] Support new Python dependency configuration 
options in flink-java, flink-streaming-java and flink-table.
URL: https://github.com/apache/flink/pull/11448#discussion_r396257250
 
 

 ##
 File path: 
flink-java/src/main/java/org/apache/flink/api/java/utils/PythonDependencyManager.java
 ##
 @@ -0,0 +1,282 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.api.java.utils;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.api.common.cache.DistributedCache;
+import org.apache.flink.api.java.tuple.Tuple2;
+import org.apache.flink.configuration.ConfigOption;
+import org.apache.flink.configuration.ConfigOptions;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.configuration.PythonOptions;
+import org.apache.flink.configuration.ReadableConfig;
+import org.apache.flink.util.Preconditions;
+
+import java.io.File;
+import java.util.ArrayList;
+import java.util.HashMap;
+import java.util.List;
+import java.util.Map;
+import java.util.UUID;
+
+import static 
org.apache.flink.configuration.PythonOptions.PYTHON_CLIENT_EXECUTABLE;
+import static org.apache.flink.configuration.PythonOptions.PYTHON_EXECUTABLE;
+
+/**
+ * Utility class for Python dependency management. The dependencies will be 
registered at the distributed
+ * cache.
+ */
+@Internal
+public class PythonDependencyManager {
+
+   public static final String PYTHON_FILE_PREFIX = "python_file";
+   public static final String PYTHON_REQUIREMENTS_FILE_PREFIX = 
"python_requirements_file";
+   public static final String PYTHON_REQUIREMENTS_CACHE_PREFIX = 
"python_requirements_cache";
+   public static final String PYTHON_ARCHIVE_PREFIX = "python_archive";
+   public static final String CLUSTER = "cluster";
+   public static final String FILE = "file";
+   public static final String CACHE = "cache";
+   public static final String FILE_DELIMITER = ",";
+   public static final String PARAM_DELIMITER = "#";
+
+   // Metadata of Python dependencies.
+
+   public static final ConfigOption> PYTHON_FILES =
+   
ConfigOptions.key("python.internal.files-key-map").mapType().noDefaultValue();
+   public static final ConfigOption> 
PYTHON_REQUIREMENTS_FILE =
+   
ConfigOptions.key("python.internal.requirements-file-key").mapType().noDefaultValue();
+   public static final ConfigOption> PYTHON_ARCHIVES =
+   
ConfigOptions.key("python.internal.archives-key-map").mapType().noDefaultValue();
+
+   private final Configuration config;
+   private final List> cachedFiles;
+   private int counter = -1;
+
+   public PythonDependencyManager(
+   Configuration config,
+   List> 
cachedFiles) {
+   this.config = Preconditions.checkNotNull(config);
+   this.cachedFiles = cachedFiles;
+   }
+
+   /**
+* Adds a Python dependency which could be .py files, Python 
packages(.zip, .egg etc.) or
+* local directories. The dependencies will be added to the PYTHONPATH 
of the Python UDF worker
+* and the local Py4J python client.
+*
+* @param filePath The path of the Python dependency.
+*/
+   public void addPythonFile(String filePath) {
+   Preconditions.checkNotNull(filePath);
+   if (fileAlreadyRegistered(filePath)) {
 
 Review comment:
   This case is expected because the `configure()` method will be called when 
the 
   `translate()` method of the Table API planners is called. As the 
`translate()` can be called many times in interactive programing, a python file 
may be registered multiple times.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (FLINK-16714) Kafka committedOffsets value <0

2020-03-23 Thread pingle wang (Jira)
pingle wang created FLINK-16714:
---

 Summary: Kafka committedOffsets value <0
 Key: FLINK-16714
 URL: https://issues.apache.org/jira/browse/FLINK-16714
 Project: Flink
  Issue Type: Improvement
  Components: Runtime / Metrics
Affects Versions: 1.6.3
Reporter: pingle wang
 Attachments: image-2020-03-23-15-36-16-487.png, 
image-2020-03-23-15-36-27-910.png

dear all:
       use rest api get kafka offsets sometimes produces errors value like < 0, 
so use this value monitoring Kafka's lag size will produce false positives. 
This should be the problem of flink metrics collection?

thanks!

!image-2020-03-23-15-36-27-910.png!



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-16547) Respect the config option of FileJobGraphRetriever#JOB_GRAPH_FILE_PATH

2020-03-23 Thread Kostas Kloudas (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16547?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kostas Kloudas closed FLINK-16547.
--
Resolution: Fixed

Merged on master with 6049b8307482a1d2d67d9642675345fc5093c197

> Respect the config option of FileJobGraphRetriever#JOB_GRAPH_FILE_PATH
> --
>
> Key: FLINK-16547
> URL: https://issues.apache.org/jira/browse/FLINK-16547
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / YARN
>Affects Versions: 1.10.0
>Reporter: Canbin Zheng
>Assignee: Canbin Zheng
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> Currently, in {{YarnClusterDescriptor#startAppMaster}}, we first write out 
> and upload the Flink Configuration file, then start to write out the JobGraph 
> file and set its name into the Flink Configuration object, the afterward 
> setting is not written into the Flink Configuration file so that it does not 
> take effect in the cluster side.
> Since in the client-side we name the JobGraph file with the default value of 
> FileJobGraphRetriever.JOB_GRAPH_FILE_PATH option, the cluster side could 
> succeed in retrieving that file. 
> This ticket proposes to write out the JobGraph file before the Configuration 
> file to ensure that the setting of FileJobGraphRetriever.JOB_GRAPH_FILE_PATH 
> is delivered to the cluster side.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] wangyang0918 commented on a change in pull request #11284: [FLINK-15911][runtime] Make Flink work with NAT.

2020-03-23 Thread GitBox
wangyang0918 commented on a change in pull request #11284: 
[FLINK-15911][runtime] Make Flink work with NAT.
URL: https://github.com/apache/flink/pull/11284#discussion_r396259803
 
 

 ##
 File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/taskexecutor/TaskManagerRunner.java
 ##
 @@ -351,21 +351,21 @@ public static TaskExecutor startTaskManager(
 
LOG.info("Starting TaskManager with ResourceID: {}", 
resourceID);
 
-   InetAddress remoteAddress = 
InetAddress.getByName(rpcService.getAddress());
+   InetAddress externalAddress = 
InetAddress.getByName(rpcService.getAddress());
 
 Review comment:
   @tillrohrmann @xintongsong 
   FYI
   I think we do not have the requirement that other `TMs` need to be able to 
resolve a `TM's ` address. For current standalone Flink cluster on K8s, all the 
`TMs` could not resolve other's hostname. However, it works pretty well. This 
is the case for current master code without external address.
   
   When  introducing the external address, i think it is same that we do not 
require other `TMs` need to be able to resolve a `TM's ` address.
   
   Correct me if i am wrong somewhere.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zhengcanbin commented on issue #11472: [FLINK-16547][yarn] Respect the config option of FileJobGraphRetriever#JOB_GRAPH_FILE_PATH

2020-03-23 Thread GitBox
zhengcanbin commented on issue #11472: [FLINK-16547][yarn] Respect the config 
option of FileJobGraphRetriever#JOB_GRAPH_FILE_PATH
URL: https://github.com/apache/flink/pull/11472#issuecomment-602436628
 
 
   > Thanks for the work @zhengcanbin and for the review @TisonKun. I will 
merge this.
   > 
   > Side note for a potential future JIRA: in the `startAppMaster()` method we 
are using sometime the `configuration` argument to the method to get/set config 
options, and sometime the `flinkConfiguration` which is a class member. 
Shouldn't we always use the `configuration` argument? This will make the method 
more self-contained.
   
   Thanks for the review @kl0u @TisonKun! I am in favor of always using the 
`configuration` argument to make the method more self-contained. I would like 
to create a tracking JIRA.
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] kl0u commented on issue #11471: [FLINK-16692] flink joblistener can register from config

2020-03-23 Thread GitBox
kl0u commented on issue #11471: [FLINK-16692] flink joblistener can register 
from config
URL: https://github.com/apache/flink/pull/11471#issuecomment-602438160
 
 
   Hi @liuyongvs . Thanks for the work, I had some comments on the JIRA and I 
will have a better look at the PR now. In the meantime, could you make your 
changes as a single commit? Although not forced yet, we do not allow merge 
commits in the repo. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-16407) Remove "FAQ" section from Flink website

2020-03-23 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-16407:
---
Labels: pull-request-available  (was: )

> Remove "FAQ" section from Flink website
> ---
>
> Key: FLINK-16407
> URL: https://issues.apache.org/jira/browse/FLINK-16407
> Project: Flink
>  Issue Type: Task
>  Components: Project Website
>Reporter: Robert Metzger
>Priority: Major
>  Labels: pull-request-available
>
> I propose to remove the [FAQ|https://flink.apache.org/faq.html] page from the 
> Flink website for the following reasons:
> - the information on there is not very up to date, nor helpful or extensive 
> (its a small selection of what somebody a few years ago considered frequent 
> questions)
> - StackOverflow lists a different set of questions as most frequent: 
> https://stackoverflow.com/questions/tagged/apache-flink?tab=Votes
> - The page is only on position 39 in Google Analytics (in the last ~10 months)
> I'm happy to hear opinions on this!



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] kl0u edited a comment on issue #11471: [FLINK-16692] flink joblistener can register from config

2020-03-23 Thread GitBox
kl0u edited a comment on issue #11471: [FLINK-16692] flink joblistener can 
register from config
URL: https://github.com/apache/flink/pull/11471#issuecomment-602438160
 
 
   Hi @liuyongvs . Thanks for the work, I had some comments on the JIRA and I 
will have a better look at the PR now. 
   
   In the meantime, could you make your changes as a single commit and rename 
your branch? Although not forced yet, we do not allow merge commits in the 
repo. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zhijiangW commented on a change in pull request #11351: [FLINK-16404][runtime] Solve the potential deadlock problem when reducing exclusive buffers to zero

2020-03-23 Thread GitBox
zhijiangW commented on a change in pull request #11351: [FLINK-16404][runtime] 
Solve the potential deadlock problem when reducing exclusive buffers to zero
URL: https://github.com/apache/flink/pull/11351#discussion_r396264764
 
 

 ##
 File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/consumer/SingleInputGate.java
 ##
 @@ -140,6 +141,8 @@
/** Channels, which notified this input gate about available data. */
private final ArrayDeque inputChannelsWithData = new 
ArrayDeque<>();
 
+   private final HashMap 
channelsBlockedByCheckpoint = new HashMap<>();
 
 Review comment:
   It is better to reuse the blocked states from `CheckpointBarrierAligner` if 
possible to avoid managing it duplicated in another place and causing potential 
inconsistency.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11284: [FLINK-15911][runtime] Make Flink work with NAT.

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11284: [FLINK-15911][runtime] Make Flink 
work with NAT.
URL: https://github.com/apache/flink/pull/11284#issuecomment-593407935
 
 
   
   ## CI report:
   
   * ff670f742762f45214ca32df966e02f4dbe6f758 Travis: 
[PENDING](https://travis-ci.com/github/flink-ci/flink/builds/154538899) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6507)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-16421) Changing default catalog to hive without changing default database fails

2020-03-23 Thread Gyula Fora (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16421?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064611#comment-17064611
 ] 

Gyula Fora commented on FLINK-16421:


my point is that defualt_database should switch to the catalogs default

> Changing default catalog to hive without changing default database fails
> 
>
> Key: FLINK-16421
> URL: https://issues.apache.org/jira/browse/FLINK-16421
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive, Table SQL / Client
>Affects Versions: 1.10.0
>Reporter: Gyula Fora
>Priority: Major
>
> The default database in Hive is caled "default" not "default_database". This 
> causes an error when starting the SQL CLI with hive set as default catalog:
> {code:java}
> Caused by: org.apache.flink.table.catalog.exceptions.CatalogException: A 
> database with name [default_database] does not exist in the catalog: 
> [hive].Caused by: org.apache.flink.table.catalog.exceptions.CatalogException: 
> A database with name [default_database] does not exist in the catalog: 
> [hive]. at 
> org.apache.flink.table.catalog.CatalogManager.setCurrentDatabase(CatalogManager.java:174)
>  at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.useDatabase(TableEnvironmentImpl.java:631)
>  at java.util.Optional.ifPresent(Optional.java:159) at 
> org.apache.flink.table.client.gateway.local.ExecutionContext.initializeCatalogs(ExecutionContext.java:561)
>  at 
> org.apache.flink.table.client.gateway.local.ExecutionContext.initializeTableEnvironment(ExecutionContext.java:494)
>  at 
> org.apache.flink.table.client.gateway.local.ExecutionContext.(ExecutionContext.java:159)
>  at 
> org.apache.flink.table.client.gateway.local.ExecutionContext.(ExecutionContext.java:118)
>  at 
> org.apache.flink.table.client.gateway.local.ExecutionContext$Builder.build(ExecutionContext.java:744)
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-16715) Always use the configuration argument in YarnClusterDescriptor#startAppMaster to make it more self-contained

2020-03-23 Thread Canbin Zheng (Jira)
Canbin Zheng created FLINK-16715:


 Summary: Always use the configuration argument in 
YarnClusterDescriptor#startAppMaster to make it more self-contained
 Key: FLINK-16715
 URL: https://issues.apache.org/jira/browse/FLINK-16715
 Project: Flink
  Issue Type: Improvement
  Components: Deployment / YARN
Reporter: Canbin Zheng
 Fix For: 1.11.0


In the YarnClusterDescriptor#{{startAppMaster()}} we are using some time the 
configuration argument to the method to get/set config options, and 
sometimes the flinkConfiguration which is a class member. This ticket 
proposes to always use the configuration argument to make the method 
more self-contained.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16715) Always use the configuration argument in YarnClusterDescriptor#startAppMaster to make it more self-contained

2020-03-23 Thread Canbin Zheng (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16715?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Canbin Zheng updated FLINK-16715:
-
Description: In the YarnClusterDescriptor#{{startAppMaster()}} we are using 
some time the {{configuration}} argument to the method to get/set config 
options, and sometimes the {{flinkConfiguration}} which is a class member. This 
ticket proposes to always use the {{configuration}} argument to make the method 
more self-contained.  (was: In the YarnClusterDescriptor#{{startAppMaster()}} 
we are using some time the configuration argument to the method to 
get/set config options, and sometimes the flinkConfiguration which is a 
class member. This ticket proposes to always use the configuration 
argument to make the method more self-contained.)

> Always use the configuration argument in YarnClusterDescriptor#startAppMaster 
> to make it more self-contained
> 
>
> Key: FLINK-16715
> URL: https://issues.apache.org/jira/browse/FLINK-16715
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / YARN
>Reporter: Canbin Zheng
>Priority: Trivial
> Fix For: 1.11.0
>
>
> In the YarnClusterDescriptor#{{startAppMaster()}} we are using some time the 
> {{configuration}} argument to the method to get/set config options, and 
> sometimes the {{flinkConfiguration}} which is a class member. This ticket 
> proposes to always use the {{configuration}} argument to make the method more 
> self-contained.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16715) Always use the configuration argument in YarnClusterDescriptor#startAppMaster to make it more self-contained

2020-03-23 Thread Canbin Zheng (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16715?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Canbin Zheng updated FLINK-16715:
-
Priority: Trivial  (was: Minor)

> Always use the configuration argument in YarnClusterDescriptor#startAppMaster 
> to make it more self-contained
> 
>
> Key: FLINK-16715
> URL: https://issues.apache.org/jira/browse/FLINK-16715
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / YARN
>Reporter: Canbin Zheng
>Priority: Trivial
> Fix For: 1.11.0
>
>
> In the YarnClusterDescriptor#{{startAppMaster()}} we are using some time the 
> configuration argument to the method to get/set config options, and 
> sometimes the flinkConfiguration which is a class member. This ticket 
> proposes to always use the configuration argument to make the method 
> more self-contained.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] liuzhixing1006 commented on issue #11455: [FLINK-16098] [chinese-translation, Documentation] Translate "Overvie…

2020-03-23 Thread GitBox
liuzhixing1006 commented on issue #11455: [FLINK-16098] [chinese-translation, 
Documentation] Translate "Overvie…
URL: https://github.com/apache/flink/pull/11455#issuecomment-602444612
 
 
   Hi @JingsongLi @lirui-apache I have finished the modification, Is there 
anything else I need to do? Thanks~


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] zhijiangW commented on a change in pull request #11351: [FLINK-16404][runtime] Solve the potential deadlock problem when reducing exclusive buffers to zero

2020-03-23 Thread GitBox
zhijiangW commented on a change in pull request #11351: [FLINK-16404][runtime] 
Solve the potential deadlock problem when reducing exclusive buffers to zero
URL: https://github.com/apache/flink/pull/11351#discussion_r396269643
 
 

 ##
 File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/io/network/partition/consumer/RemoteInputChannel.java
 ##
 @@ -416,7 +452,18 @@ public int getUnannouncedCredit() {
 * @return Credit which was not announced to the sender yet.
 */
public int getAndResetUnannouncedCredit() {
-   return unannouncedCredit.getAndSet(0);
+   synchronized (bufferQueue) {
+   int credit = unannouncedCredit;
+   unannouncedCredit = 0;
+   if (credit == 0) {
 
 Review comment:
   It is tricky to make zero credit has a special meaning, and we also break 
the previous assumption without negative credits. 
   
   I think it might be better to define a separate message to describe the 
semantic of barrier alignment to unblock upstream side. Then the credit is 
always positive to be consistent as before, and we only define another blocked 
state on upstream side to control the buffer transport besides with positive 
credits. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-16713) Support source mode of elasticsearch connector

2020-03-23 Thread jackray wang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16713?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064613#comment-17064613
 ] 

jackray wang commented on FLINK-16713:
--

Hi [~lzljs3620320] ,this feature is essential for our company,so if nobody to 
develop this feature  ,i will take it.

 

 

> Support source mode of elasticsearch connector
> --
>
> Key: FLINK-16713
> URL: https://issues.apache.org/jira/browse/FLINK-16713
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / ElasticSearch
>Affects Versions: 1.10.0
>Reporter: jackray wang
>Priority: Major
>
> [https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/table/connect.html#elasticsearch-connector]
> For append-only queries, the connector can also operate in [append 
> mode|https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/table/connect.html#update-modes]
>  for exchanging only INSERT messages with the external system. If no key is 
> defined by the query, a key is automatically generated by Elasticsearch.
> I want to know ,why the connector of flink with ES just support sink but  
> doesn't support source .Which version could add this feature to ?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-16540) Fully specify bugfix version of Docker containers in Flink Playground docker-compose.yaml files

2020-03-23 Thread Fabian Hueske (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16540?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fabian Hueske closed FLINK-16540.
-
Fix Version/s: 1.9.0
   1.10.0
   1.11.0
   Resolution: Fixed

> Fully specify bugfix version of Docker containers in Flink Playground 
> docker-compose.yaml files
> ---
>
> Key: FLINK-16540
> URL: https://issues.apache.org/jira/browse/FLINK-16540
> Project: Flink
>  Issue Type: Improvement
>  Components: Examples
>Reporter: Fabian Hueske
>Assignee: Fabian Hueske
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0, 1.10.0, 1.9.0
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> I recently noticed that we do not guarantee client-jobmanager compatibility 
> among different bugfix versions of the same minor version (see 
> https://github.com/ververica/sql-training/issues/8 for details).
> In this case, a job submitted via a Flink 1.9.0 client could not be executed 
> on a Flink 1.9.2 cluster.
> For the playgrounds this can easily happen, because we build a custom Docker 
> image (with a fixed bugfix version) and load the latest Flink images for the 
> latest bugfix version of the same minor version.
> We can avoid such problems by fixing the bugfix version of the Flink images 
> that we run in the playground.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-16713) Support source mode of elasticsearch connector

2020-03-23 Thread jackray wang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16713?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064613#comment-17064613
 ] 

jackray wang edited comment on FLINK-16713 at 3/23/20, 8:15 AM:


Hi [~lzljs3620320] ,this feature is essential for our company,so if nobody to 
develop this feature  ,I will take it.

 

 


was (Author: jackray):
Hi [~lzljs3620320] ,this feature is essential for our company,so if nobody to 
develop this feature  ,i will take it.

 

 

> Support source mode of elasticsearch connector
> --
>
> Key: FLINK-16713
> URL: https://issues.apache.org/jira/browse/FLINK-16713
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / ElasticSearch
>Affects Versions: 1.10.0
>Reporter: jackray wang
>Priority: Major
>
> [https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/table/connect.html#elasticsearch-connector]
> For append-only queries, the connector can also operate in [append 
> mode|https://ci.apache.org/projects/flink/flink-docs-release-1.10/dev/table/connect.html#update-modes]
>  for exchanging only INSERT messages with the external system. If no key is 
> defined by the query, a key is automatically generated by Elasticsearch.
> I want to know ,why the connector of flink with ES just support sink but  
> doesn't support source .Which version could add this feature to ?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #11481: [FLINK-13691][Table]Remove deprecated query config

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11481: [FLINK-13691][Table]Remove 
deprecated query config
URL: https://github.com/apache/flink/pull/11481#issuecomment-602174011
 
 
   
   ## CI report:
   
   * cb9eb2458536cf858d110b30c6481b3d6597c9bc Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/154532502) Azure: 
[SUCCESS](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6502)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11448: [FLINK-16666][python][table] Support new Python dependency configuration options in flink-java, flink-streaming-java and flink-table.

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11448: [FLINK-1][python][table] Support 
new Python dependency configuration options in flink-java, flink-streaming-java 
and flink-table.
URL: https://github.com/apache/flink/pull/11448#issuecomment-601042255
 
 
   
   ## CI report:
   
   * 2821e4a2542bffceeef1b742f43a09cae359e8ba Travis: 
[PENDING](https://travis-ci.com/github/flink-ci/flink/builds/154540657) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6510)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11471: [FLINK-16692] flink joblistener can register from config

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11471: [FLINK-16692] flink joblistener can 
register from config
URL: https://github.com/apache/flink/pull/11471#issuecomment-602012274
 
 
   
   ## CI report:
   
   * 7f0c7e868fce22437c1b4b7a02fece52afb70f4a UNKNOWN
   * 57f1b99ac6db6626e6c78cb7b39e2026b59d1282 UNKNOWN
   * 84606e23725fc6c688ce72da42d06b8f30ac4910 Travis: 
[PENDING](https://travis-ci.com/github/flink-ci/flink/builds/154537643) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6506)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-16156) TableAggregateITCase.testGroupByFlatAggregate failed on Travis

2020-03-23 Thread Kurt Young (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16156?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064619#comment-17064619
 ] 

Kurt Young commented on FLINK-16156:


FunctionITCase.testDynamicTableFunction 

has been fixed in FLINK-16118

 

> TableAggregateITCase.testGroupByFlatAggregate failed on Travis
> --
>
> Key: FLINK-16156
> URL: https://issues.apache.org/jira/browse/FLINK-16156
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.10.0
>Reporter: Till Rohrmann
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.10.1, 1.11.0
>
>
> The {{TableAggregateITCase.testGroupByFlatAggregate}} failed because the 
> order of the result did not match:
> {code}
> 11:50:11.239 [ERROR] Failures: 
> 11:50:11.239 [ERROR]   FunctionITCase.testDynamicTableFunction:611 
> Expected: <[Test is a string, 42, null]>
>  but: was <[42, null, Test is a string]>
> 11:50:11.239 [ERROR] Errors: 
> 11:50:11.239 [ERROR]   TableAggregateITCase.testGroupByFlatAggregate:59 » 
> JobExecution Job execution ...
> 11:50:11.239 [INFO] 
> {code}
> https://api.travis-ci.com/v3/job/288392903/log.txt



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16156) TableAggregateITCase.testGroupByFlatAggregate failed on Travis

2020-03-23 Thread Kurt Young (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16156?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064620#comment-17064620
 ] 

Kurt Young commented on FLINK-16156:


TableAggregateITCase.testGroupByFlatAggregate seems to be a random submission 
failure

> TableAggregateITCase.testGroupByFlatAggregate failed on Travis
> --
>
> Key: FLINK-16156
> URL: https://issues.apache.org/jira/browse/FLINK-16156
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Planner
>Affects Versions: 1.10.0
>Reporter: Till Rohrmann
>Priority: Critical
>  Labels: test-stability
> Fix For: 1.10.1, 1.11.0
>
>
> The {{TableAggregateITCase.testGroupByFlatAggregate}} failed because the 
> order of the result did not match:
> {code}
> 11:50:11.239 [ERROR] Failures: 
> 11:50:11.239 [ERROR]   FunctionITCase.testDynamicTableFunction:611 
> Expected: <[Test is a string, 42, null]>
>  but: was <[42, null, Test is a string]>
> 11:50:11.239 [ERROR] Errors: 
> 11:50:11.239 [ERROR]   TableAggregateITCase.testGroupByFlatAggregate:59 » 
> JobExecution Job execution ...
> 11:50:11.239 [INFO] 
> {code}
> https://api.travis-ci.com/v3/job/288392903/log.txt



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16692) flink joblistener can register from config

2020-03-23 Thread jackylau (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16692?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064621#comment-17064621
 ] 

jackylau commented on FLINK-16692:
--

HI [~kkl0u], 

1) i will move the arguments to DeploymentOptions

2) you can review my code, i use "execution.job-listeners" ,which is same with 
you.

> flink joblistener can register from config
> --
>
> Key: FLINK-16692
> URL: https://issues.apache.org/jira/browse/FLINK-16692
> Project: Flink
>  Issue Type: Improvement
>  Components: API / Core
>Affects Versions: 1.10.0
>Reporter: jackylau
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> we should do as spark does ,which can register listener from conf such as 
> "spark.extraListeners"。 And it will be convinient for users when users just 
> want to set hook



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16175) Add config option to switch case sensitive for column names in SQL

2020-03-23 Thread Dawid Wysakowicz (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064624#comment-17064624
 ] 

Dawid Wysakowicz commented on FLINK-16175:
--

I agree with the principle of the ticket to support case sensitivity. I think 
though we should solve it in a uniform way for all aspects at once. Otherwise I 
think the behaviour would be inconsistent, e.g. case sensitive for column names 
but case insensitive for table names, which I am sure will surprise many users.

> Add config option to switch case sensitive for column names in SQL
> --
>
> Key: FLINK-16175
> URL: https://issues.apache.org/jira/browse/FLINK-16175
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / API, Table SQL / Planner
>Affects Versions: 1.11.0
>Reporter: Leonard Xu
>Assignee: Leonard Xu
>Priority: Major
>  Labels: usability
> Fix For: 1.11.0
>
>
> Flink SQL is default CaseSensitive and have no option to config. This issue 
> aims to support
> a configOption so that user can set CaseSensitive for their SQL.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] TisonKun commented on issue #11458: [FLINK-16625][k8s] Extract BootstrapTools#getEnvironmentVariables to a general utility in ConfigurationUtils

2020-03-23 Thread GitBox
TisonKun commented on issue #11458: [FLINK-16625][k8s] Extract 
BootstrapTools#getEnvironmentVariables to a general utility in 
ConfigurationUtils
URL: https://github.com/apache/flink/pull/11458#issuecomment-602454643
 
 
   @zhengcanbin please rebase on master. I think this PR itself is good to go.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink-statefun] tzulitai closed pull request #62: [FLINK-16685] Add a k8s example for the Python SDK

2020-03-23 Thread GitBox
tzulitai closed pull request #62: [FLINK-16685] Add a k8s example for the 
Python SDK
URL: https://github.com/apache/flink-statefun/pull/62
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11471: [FLINK-16692] flink joblistener can register from config

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11471: [FLINK-16692] flink joblistener can 
register from config
URL: https://github.com/apache/flink/pull/11471#issuecomment-602012274
 
 
   
   ## CI report:
   
   * 7f0c7e868fce22437c1b4b7a02fece52afb70f4a UNKNOWN
   * 57f1b99ac6db6626e6c78cb7b39e2026b59d1282 UNKNOWN
   * 84606e23725fc6c688ce72da42d06b8f30ac4910 Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/154537643) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6506)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11486: [FLINK-16712][task] Refactor StreamTask to construct final fields

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11486: [FLINK-16712][task] Refactor 
StreamTask to construct final fields
URL: https://github.com/apache/flink/pull/11486#issuecomment-602377929
 
 
   
   ## CI report:
   
   * 53b97eb753986785848b55a6626e01f20e070626 Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/154529257) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6500)
 
   * 9535112ff0782fd584600cc3fdbefe0457b1bcfb UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Closed] (FLINK-16685) Add an example of how to deploy a multi-language function to Kubernetes.

2020-03-23 Thread Tzu-Li (Gordon) Tai (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16685?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tzu-Li (Gordon) Tai closed FLINK-16685.
---
Fix Version/s: statefun-1.1
   Resolution: Fixed

Merged to master via b9752c895487c188b5e687b6445d06b388d02239

> Add an example of how to deploy a multi-language function to Kubernetes.
> 
>
> Key: FLINK-16685
> URL: https://issues.apache.org/jira/browse/FLINK-16685
> Project: Flink
>  Issue Type: Improvement
>  Components: Stateful Functions
>Reporter: Igal Shilman
>Assignee: Igal Shilman
>Priority: Major
>  Labels: pull-request-available
> Fix For: statefun-1.1
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> Add an example of how to deploy a multi-language function to Kubernetes.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Closed] (FLINK-16706) Update Stateful Functions master branch version to 2.0-SNAPSHOT

2020-03-23 Thread Tzu-Li (Gordon) Tai (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16706?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tzu-Li (Gordon) Tai closed FLINK-16706.
---
Fix Version/s: statefun-1.1
   Resolution: Fixed

Merged via 2c9074d443866048c49ec07784ec384cd3054602

> Update Stateful Functions master branch version to 2.0-SNAPSHOT
> ---
>
> Key: FLINK-16706
> URL: https://issues.apache.org/jira/browse/FLINK-16706
> Project: Flink
>  Issue Type: Task
>Reporter: Tzu-Li (Gordon) Tai
>Assignee: Tzu-Li (Gordon) Tai
>Priority: Blocker
> Fix For: statefun-1.1
>
>
> As discussed in 
> http://apache-flink-mailing-list-archive.1008284.n3.nabble.com/DISCUSS-Update-on-Flink-Stateful-Functions-amp-what-are-the-next-steps-td38646.html,
> the community wants the next (and first Apache-endorsed) Stateful Functions 
> release number to be 2.0.
> We should update the master branch snapshot version, as well as the {{Affects 
> Version}} and {{Fixed Version}} fields of all Stateful Functions JIRAs.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-16716) Update Roadmap after Flink 1.10 release

2020-03-23 Thread Fabian Hueske (Jira)
Fabian Hueske created FLINK-16716:
-

 Summary: Update Roadmap after Flink 1.10 release
 Key: FLINK-16716
 URL: https://issues.apache.org/jira/browse/FLINK-16716
 Project: Flink
  Issue Type: Bug
  Components: Project Website
Reporter: Fabian Hueske


The roadmap on the Flink website needs to be updated to reflect the new 
features of Flink 1.10 and the planned features and improvements of future 
releases.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] TisonKun commented on issue #11472: [FLINK-16547][yarn] Respect the config option of FileJobGraphRetriever#JOB_GRAPH_FILE_PATH

2020-03-23 Thread GitBox
TisonKun commented on issue #11472: [FLINK-16547][yarn] Respect the config 
option of FileJobGraphRetriever#JOB_GRAPH_FILE_PATH
URL: https://github.com/apache/flink/pull/11472#issuecomment-602458510
 
 
   > Thanks for the work @zhengcanbin and for the review @TisonKun. I will 
merge this.
   > 
   > Side note for a potential future JIRA: in the `startAppMaster()` method we 
are using sometime the `configuration` argument to the method to get/set config 
options, and sometime the `flinkConfiguration` which is a class member. 
Shouldn't we always use the `configuration` argument? This will make the method 
more self-contained.
   
   @kl0u yes I think I get some cycle of that problem already. However, I'd 
prefer refer to the field `flinkConfiguration`; otherwise user(previous me) 
might ask what's the difference between *this configuration object* and *the 
one I've seen as field*? It is one of Java best practices that we don't pass 
field as parameter in the same class.
   
   Besides, I've heard developers complain with the tight couplings between 
`YarnClusterDescriptor` and `Utils`. There are several huge & context-aware 
function. If YARN deployment is still under active development, we might think 
of refactor these codes gradually. cc @wangyang0918 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] TisonKun edited a comment on issue #11472: [FLINK-16547][yarn] Respect the config option of FileJobGraphRetriever#JOB_GRAPH_FILE_PATH

2020-03-23 Thread GitBox
TisonKun edited a comment on issue #11472: [FLINK-16547][yarn] Respect the 
config option of FileJobGraphRetriever#JOB_GRAPH_FILE_PATH
URL: https://github.com/apache/flink/pull/11472#issuecomment-602458510
 
 
   > Thanks for the work @zhengcanbin and for the review @TisonKun. I will 
merge this.
   > 
   > Side note for a potential future JIRA: in the `startAppMaster()` method we 
are using sometime the `configuration` argument to the method to get/set config 
options, and sometime the `flinkConfiguration` which is a class member. 
Shouldn't we always use the `configuration` argument? This will make the method 
more self-contained.
   
   @kl0u yes I think I get some cycle of that problem already. However, I'd 
prefer refer to the field `flinkConfiguration`; otherwise user(previous me) 
might ask what's the difference between *this configuration object* and *the 
one I've seen as field*? It is one of Java best practices that we don't pass 
field as parameter of this class's method.
   
   Besides, I've heard developers complain with the tight couplings between 
`YarnClusterDescriptor` and `Utils`. There are several huge & context-aware 
function. If YARN deployment is still under active development, we might think 
of refactor these codes gradually. cc @wangyang0918 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] kl0u commented on a change in pull request #11471: [FLINK-16692] flink joblistener can register from config

2020-03-23 Thread GitBox
kl0u commented on a change in pull request #11471: [FLINK-16692] flink 
joblistener can register from config
URL: https://github.com/apache/flink/pull/11471#discussion_r396289410
 
 

 ##
 File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/environment/StreamExecutionEnvironment.java
 ##
 @@ -761,6 +768,20 @@ public void configure(ReadableConfig configuration, 
ClassLoader classLoader) {
});
config.configure(configuration, classLoader);
checkpointCfg.configure(configuration);
+   configuration.getOptional(ExecutionOptions.JOB_LISTENERS)
+   .ifPresent(l -> {
+   for (String listener : l) {
+   try {
 
 Review comment:
   You can use  the method `InstantiationUtil#instantiate(final String 
className, final Class targetType, final ClassLoader classLoader)`.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] kl0u commented on a change in pull request #11471: [FLINK-16692] flink joblistener can register from config

2020-03-23 Thread GitBox
kl0u commented on a change in pull request #11471: [FLINK-16692] flink 
joblistener can register from config
URL: https://github.com/apache/flink/pull/11471#discussion_r396281963
 
 

 ##
 File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/environment/StreamExecutionEnvironment.java
 ##
 @@ -237,6 +237,13 @@ public ExecutionConfig getConfig() {
return cacheFile;
}
 
+   /**
+* Gets the config JobListeners.
+*/
 
 Review comment:
   Annotate as `@PublicEvolving`.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-16620) Add attempt information in logging

2020-03-23 Thread Roman Khachatryan (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16620?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064631#comment-17064631
 ] 

Roman Khachatryan commented on FLINK-16620:
---

I think this way we'll have in too many places, but probably it's fine.

And something shorter would be better IMO: "$taskName (4/5, attempt 1)".

Should we increment attempt number by 1 too?

> Add attempt information in logging
> --
>
> Key: FLINK-16620
> URL: https://issues.apache.org/jira/browse/FLINK-16620
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Task
>Affects Versions: 1.10.0
>Reporter: Jiayi Liao
>Assignee: Jiayi Liao
>Priority: Minor
>
> Currently  logging in places such as {{Task}} and {{StreamTask}} , is using 
> {{taskNameWithSubtasks}} in {{TaskInfo}} to represent an execution. I think 
> it'll be more user-friendly if we can add attempt information into the 
> logging.
> The format can be consitent with logging information in {{Execution}} :
> {code:java}
> MySink (3/10) - execution #0
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] kl0u commented on a change in pull request #11471: [FLINK-16692] flink joblistener can register from config

2020-03-23 Thread GitBox
kl0u commented on a change in pull request #11471: [FLINK-16692] flink 
joblistener can register from config
URL: https://github.com/apache/flink/pull/11471#discussion_r396291783
 
 

 ##
 File path: 
flink-core/src/main/java/org/apache/flink/configuration/ExecutionOptions.java
 ##
 @@ -55,4 +56,11 @@
"throughput")
)
.build());
+
+   public static final ConfigOption> JOB_LISTENERS =
 
 Review comment:
   I would suggest to move it to the `DeploymentOptions`.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] kl0u commented on a change in pull request #11471: [FLINK-16692] flink joblistener can register from config

2020-03-23 Thread GitBox
kl0u commented on a change in pull request #11471: [FLINK-16692] flink 
joblistener can register from config
URL: https://github.com/apache/flink/pull/11471#discussion_r396289877
 
 

 ##
 File path: 
flink-streaming-java/src/test/java/org/apache/flink/streaming/api/environment/StreamExecutionEnvironmentComplexConfigurationTest.java
 ##
 @@ -97,4 +105,47 @@ public void 
testNotOverridingCachedFilesFromConfiguration() {
Tuple2.of("file3", new 
DistributedCache.DistributedCacheEntry("/tmp3", true))
)));
}
+
+   @Test
+   public void testLoadingListenersFromConfiguration() {
+   StreamExecutionEnvironment envFromConfiguration = 
StreamExecutionEnvironment.getExecutionEnvironment();
+   List listenersClass = 
Arrays.asList(BasicJobSubmittedCounter.class, BasicJobExecutedCounter.class);
+
+   Configuration configuration = new Configuration();
+   configuration.setString("execution.job-listeners", 
listenersClass.stream().map(l -> l.getName()).collect(Collectors.joining(";")));
+
 
 Review comment:
   You can use the `ConfigUtils#encodeCollectionToConfig()`.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] kl0u commented on a change in pull request #11471: [FLINK-16692] flink joblistener can register from config

2020-03-23 Thread GitBox
kl0u commented on a change in pull request #11471: [FLINK-16692] flink 
joblistener can register from config
URL: https://github.com/apache/flink/pull/11471#discussion_r396281963
 
 

 ##
 File path: 
flink-streaming-java/src/main/java/org/apache/flink/streaming/api/environment/StreamExecutionEnvironment.java
 ##
 @@ -237,6 +237,13 @@ public ExecutionConfig getConfig() {
return cacheFile;
}
 
+   /**
+* Gets the config JobListeners.
+*/
 
 Review comment:
   Annotate as `@PublicEvolving`. And I would prefer if we did not even add a 
new method in the `env`, if possible.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-16715) Always use the configuration argument in YarnClusterDescriptor#startAppMaster to make it more self-contained

2020-03-23 Thread Xintong Song (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16715?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064634#comment-17064634
 ] 

Xintong Song commented on FLINK-16715:
--

I tend to think this differently.
Instead of always using {{configuration}} argument, I would suggest to always 
use the class field {{flinkConfiguration}}, for the following reasons.
* As a private non-static method, we don't really need {{startAppMaster}} to be 
self-contained.
* Even we change it to always use the argument, it does guarantee that in 
future, because the class field is visible to the method anyway. On the other 
hand, if we change it to always use the class filed, we can remove the method 
argument to make sure there is only one configuration visible to the method.
* It helps reducing the number of arguments needed for the {{startAppMaster}} 
method.
Actually, I think we should do the same thing for {{yarnClient}}, which is also 
a class field.

> Always use the configuration argument in YarnClusterDescriptor#startAppMaster 
> to make it more self-contained
> 
>
> Key: FLINK-16715
> URL: https://issues.apache.org/jira/browse/FLINK-16715
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / YARN
>Reporter: Canbin Zheng
>Priority: Trivial
> Fix For: 1.11.0
>
>
> In the YarnClusterDescriptor#{{startAppMaster()}} we are using some time the 
> {{configuration}} argument to the method to get/set config options, and 
> sometimes the {{flinkConfiguration}} which is a class member. This ticket 
> proposes to always use the {{configuration}} argument to make the method more 
> self-contained.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Created] (FLINK-16717) Use headless service for rpc and blob port when flink on K8S

2020-03-23 Thread vanderliang (Jira)
vanderliang created FLINK-16717:
---

 Summary: Use headless service for rpc and blob port when flink on 
K8S
 Key: FLINK-16717
 URL: https://issues.apache.org/jira/browse/FLINK-16717
 Project: Flink
  Issue Type: Improvement
  Components: Deployment / Kubernetes
Affects Versions: 1.10.0
 Environment: flink 1.10
Reporter: vanderliang


Current, when submitting a flink job cluster on kubernetes, the 
[https://github.com/apache/flink/blob/release-1.10/flink-container/kubernetes/job-cluster-service.yaml]
 will create Node Port type service for all ports. 

First, the RPC and blob only need a headless service which could avoid iptables 
or ipvs forwarding. 

Second,serverless K8S(like AWS EKS) in public cloud does not support Node 
Port,which cause some problem when running flink On serverless K8S.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #11448: [FLINK-16666][python][table] Support new Python dependency configuration options in flink-java, flink-streaming-java and flink-table.

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11448: [FLINK-1][python][table] Support 
new Python dependency configuration options in flink-java, flink-streaming-java 
and flink-table.
URL: https://github.com/apache/flink/pull/11448#issuecomment-601042255
 
 
   
   ## CI report:
   
   * 2821e4a2542bffceeef1b742f43a09cae359e8ba Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/154540657) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6510)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] flinkbot edited a comment on issue #11284: [FLINK-15911][runtime] Make Flink work with NAT.

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11284: [FLINK-15911][runtime] Make Flink 
work with NAT.
URL: https://github.com/apache/flink/pull/11284#issuecomment-593407935
 
 
   
   ## CI report:
   
   * ff670f742762f45214ca32df966e02f4dbe6f758 Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/154538899) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6507)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Created] (FLINK-16718) KvStateServerHandlerTest leaks Netty ByteBufs

2020-03-23 Thread Gary Yao (Jira)
Gary Yao created FLINK-16718:


 Summary: KvStateServerHandlerTest leaks Netty ByteBufs
 Key: FLINK-16718
 URL: https://issues.apache.org/jira/browse/FLINK-16718
 Project: Flink
  Issue Type: Bug
  Components: Runtime / Queryable State, Tests
Affects Versions: 1.10.0, 1.11.0
Reporter: Gary Yao
Assignee: Gary Yao
 Fix For: 1.11.0


The {{KvStateServerHandlerTest}} leaks Netty {{ByteBuf}}s.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #11486: [FLINK-16712][task] Refactor StreamTask to construct final fields

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11486: [FLINK-16712][task] Refactor 
StreamTask to construct final fields
URL: https://github.com/apache/flink/pull/11486#issuecomment-602377929
 
 
   
   ## CI report:
   
   * 9535112ff0782fd584600cc3fdbefe0457b1bcfb Travis: 
[FAILURE](https://travis-ci.com/github/flink-ci/flink/builds/154549644) Azure: 
[FAILURE](https://dev.azure.com/rmetzger/5bd3ef0a-4359-41af-abca-811b04098d2e/_build/results?buildId=6512)
 
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-16718) KvStateServerHandlerTest leaks Netty ByteBufs

2020-03-23 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16718?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-16718:
-
Description: The {{KvStateServerHandlerTest}} leaks Netty {{ByteBuf}} 
instances.  (was: The {{KvStateServerHandlerTest}} leaks Netty {{ByteBuf}}s.)

> KvStateServerHandlerTest leaks Netty ByteBufs
> -
>
> Key: FLINK-16718
> URL: https://issues.apache.org/jira/browse/FLINK-16718
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Queryable State, Tests
>Affects Versions: 1.10.0, 1.11.0
>Reporter: Gary Yao
>Assignee: Gary Yao
>Priority: Major
> Fix For: 1.11.0
>
>
> The {{KvStateServerHandlerTest}} leaks Netty {{ByteBuf}} instances.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16717) Use headless service for rpc and blob port when flink on K8S

2020-03-23 Thread Canbin Zheng (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16717?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064649#comment-17064649
 ] 

Canbin Zheng commented on FLINK-16717:
--

Is it a duplicate of FLINK-16602?

> Use headless service for rpc and blob port when flink on K8S
> 
>
> Key: FLINK-16717
> URL: https://issues.apache.org/jira/browse/FLINK-16717
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / Kubernetes
>Affects Versions: 1.10.0
> Environment: flink 1.10
>Reporter: vanderliang
>Priority: Major
>
> Current, when submitting a flink job cluster on kubernetes, the 
> [https://github.com/apache/flink/blob/release-1.10/flink-container/kubernetes/job-cluster-service.yaml]
>  will create Node Port type service for all ports. 
> First, the RPC and blob only need a headless service which could avoid 
> iptables or ipvs forwarding. 
> Second,serverless K8S(like AWS EKS) in public cloud does not support Node 
> Port,which cause some problem when running flink On serverless K8S.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Commented] (FLINK-16638) Flink checkStateMappingCompleteness doesn't include UserDefinedOperatorIDs

2020-03-23 Thread Roman Khachatryan (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16638?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064651#comment-17064651
 ] 

Roman Khachatryan commented on FLINK-16638:
---

Unfortunately, I can't follow the stacktrace (e.g. LegacyScheduler).

Are you sure your version is 1.10?

> Flink checkStateMappingCompleteness doesn't include UserDefinedOperatorIDs
> --
>
> Key: FLINK-16638
> URL: https://issues.apache.org/jira/browse/FLINK-16638
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.10.0
>Reporter: Bashar Abdul Jawad
>Priority: Critical
>
> [StateAssignmentOperation.checkStateMappingCompleteness|https://github.com/apache/flink/blob/master/flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/StateAssignmentOperation.java#L555]
>  doesn't check for UserDefinedOperatorIDs (specified using setUidHash), 
> causing the exception:
> {code}
>  java.lang.IllegalStateException: There is no operator for the state {}
> {code}
> to be thrown when a savepoint can't be mapped to an ExecutionJobVertex, even 
> when the operator hash is explicitly specified.
> I believe this logic should be extended to also include 
> UserDefinedOperatorIDs as so:
> {code:java}
> for (ExecutionJobVertex executionJobVertex : tasks) {
>   allOperatorIDs.addAll(executionJobVertex.getOperatorIDs());
>   allOperatorIDs.addAll(executionJobVertex.getUserDefinedOperatorIDs());
> }
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] liuyongvs commented on issue #11471: [FLINK-16692] flink joblistener can register from config

2020-03-23 Thread GitBox
liuyongvs commented on issue #11471: [FLINK-16692] flink joblistener can 
register from config
URL: https://github.com/apache/flink/pull/11471#issuecomment-602476909
 
 
   Thanks you so much. Becase my personal PC doen't run fast. so i just not run 
on my PC and commits directly. i will commit with one commits next time.
   > Hi @liuyongvs . Thanks for the work, I had some comments on the JIRA and I 
will have a better look at the PR now.
   > 
   > In the meantime, could you make your changes as a single commit and rename 
your branch? Although not forced yet, we do not allow merge commits in the repo.
   
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] GJL closed pull request #11453: [FLINK-16718][tests] Fix ByteBuf leak in KvStateServerHandlerTest

2020-03-23 Thread GitBox
GJL closed pull request #11453: [FLINK-16718][tests] Fix ByteBuf leak in 
KvStateServerHandlerTest
URL: https://github.com/apache/flink/pull/11453
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Updated] (FLINK-16718) KvStateServerHandlerTest leaks Netty ByteBufs

2020-03-23 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16718?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-16718:
---
Labels: pull-request-available  (was: )

> KvStateServerHandlerTest leaks Netty ByteBufs
> -
>
> Key: FLINK-16718
> URL: https://issues.apache.org/jira/browse/FLINK-16718
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Queryable State, Tests
>Affects Versions: 1.10.0, 1.11.0
>Reporter: Gary Yao
>Assignee: Gary Yao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.11.0
>
>
> The {{KvStateServerHandlerTest}} leaks Netty {{ByteBuf}} instances.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Updated] (FLINK-16718) KvStateServerHandlerTest leaks Netty ByteBufs

2020-03-23 Thread Gary Yao (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-16718?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gary Yao updated FLINK-16718:
-
Fix Version/s: 1.10.1

> KvStateServerHandlerTest leaks Netty ByteBufs
> -
>
> Key: FLINK-16718
> URL: https://issues.apache.org/jira/browse/FLINK-16718
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Queryable State, Tests
>Affects Versions: 1.10.0, 1.11.0
>Reporter: Gary Yao
>Assignee: Gary Yao
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.10.1, 1.11.0
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> The {{KvStateServerHandlerTest}} leaks Netty {{ByteBuf}} instances.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[GitHub] [flink] flinkbot edited a comment on issue #11347: [FLINK-14971][checkpointing] Make all the non-IO operations in CheckpointCoordinator single-threaded

2020-03-23 Thread GitBox
flinkbot edited a comment on issue #11347: [FLINK-14971][checkpointing] Make 
all the non-IO operations in CheckpointCoordinator single-threaded
URL: https://github.com/apache/flink/pull/11347#issuecomment-596323413
 
 
   
   ## CI report:
   
   * 655adff5f2d9298b34b167b5a355f4b536613246 Travis: 
[SUCCESS](https://travis-ci.com/github/flink-ci/flink/builds/153558707) 
   * 9f59db6be464318410e6ce165cb599233980b534 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run travis` re-run the last Travis build
- `@flinkbot run azure` re-run the last Azure build
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[GitHub] [flink] ifndef-SleePy commented on a change in pull request #11347: [FLINK-14971][checkpointing] Make all the non-IO operations in CheckpointCoordinator single-threaded

2020-03-23 Thread GitBox
ifndef-SleePy commented on a change in pull request #11347: 
[FLINK-14971][checkpointing] Make all the non-IO operations in 
CheckpointCoordinator single-threaded
URL: https://github.com/apache/flink/pull/11347#discussion_r396315368
 
 

 ##
 File path: 
flink-runtime/src/main/java/org/apache/flink/runtime/checkpoint/PendingCheckpoint.java
 ##
 @@ -311,25 +315,32 @@ public CheckpointException getFailureCause() {
try (CheckpointMetadataOutputStream out = 
targetLocation.createMetadataOutputStream()) {

Checkpoints.storeCheckpointMetadata(savepoint, out);
finalizedLocation = 
out.closeAndFinalizeCheckpoint();
+   }
 
+   CompletedCheckpoint completed = new 
CompletedCheckpoint(
+   jobId,
+   checkpointId,
+   checkpointTimestamp,
+   System.currentTimeMillis(),
+   operatorStates,
+   masterStates,
+   props,
+   finalizedLocation);
+
+   try {
+   
completedCheckpointStore.addCheckpoint(completed);
+   } catch (Throwable t) {
+   completed.discardOnFailedStoring();
 
 Review comment:
   It should be.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services


[jira] [Commented] (FLINK-16715) Always use the configuration argument in YarnClusterDescriptor#startAppMaster to make it more self-contained

2020-03-23 Thread Kostas Kloudas (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16715?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064658#comment-17064658
 ] 

Kostas Kloudas commented on FLINK-16715:


[~xintongsong] I see your point. In general the whole {{YarnClusterDescriptor}} 
could use some refactoring (e.g. the {{startAppMaster()}} has ~400 lines of 
code). I do not have strong feelings about using the one or the other 
({{configuration}} or {{flinkConfiguration}) and maybe using the class member 
makes it clearer that these changes will be visible to all methods in the 
class.  

> Always use the configuration argument in YarnClusterDescriptor#startAppMaster 
> to make it more self-contained
> 
>
> Key: FLINK-16715
> URL: https://issues.apache.org/jira/browse/FLINK-16715
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / YARN
>Reporter: Canbin Zheng
>Priority: Trivial
> Fix For: 1.11.0
>
>
> In the YarnClusterDescriptor#{{startAppMaster()}} we are using some time the 
> {{configuration}} argument to the method to get/set config options, and 
> sometimes the {{flinkConfiguration}} which is a class member. This ticket 
> proposes to always use the {{configuration}} argument to make the method more 
> self-contained.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


[jira] [Comment Edited] (FLINK-16715) Always use the configuration argument in YarnClusterDescriptor#startAppMaster to make it more self-contained

2020-03-23 Thread Kostas Kloudas (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-16715?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17064658#comment-17064658
 ] 

Kostas Kloudas edited comment on FLINK-16715 at 3/23/20, 9:33 AM:
--

[~xintongsong] I see your point. In general the whole {{YarnClusterDescriptor}} 
could use some refactoring (e.g. the {{startAppMaster()}} has ~400 lines of 
code). I do not have strong feelings about using the one or the other 
({{configuration}} or {{flinkConfiguration}}) and maybe using the class member 
makes it clearer that these changes will be visible to all methods in the 
class.  


was (Author: kkl0u):
[~xintongsong] I see your point. In general the whole {{YarnClusterDescriptor}} 
could use some refactoring (e.g. the {{startAppMaster()}} has ~400 lines of 
code). I do not have strong feelings about using the one or the other 
({{configuration}} or {{flinkConfiguration}) and maybe using the class member 
makes it clearer that these changes will be visible to all methods in the 
class.  

> Always use the configuration argument in YarnClusterDescriptor#startAppMaster 
> to make it more self-contained
> 
>
> Key: FLINK-16715
> URL: https://issues.apache.org/jira/browse/FLINK-16715
> Project: Flink
>  Issue Type: Improvement
>  Components: Deployment / YARN
>Reporter: Canbin Zheng
>Priority: Trivial
> Fix For: 1.11.0
>
>
> In the YarnClusterDescriptor#{{startAppMaster()}} we are using some time the 
> {{configuration}} argument to the method to get/set config options, and 
> sometimes the {{flinkConfiguration}} which is a class member. This ticket 
> proposes to always use the {{configuration}} argument to make the method more 
> self-contained.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)


  1   2   3   4   5   6   >