[jira] [Updated] (FLINK-31059) Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by native implementation

2023-02-21 Thread miamiaoxyz (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31059?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

miamiaoxyz updated FLINK-31059:
---
Attachment: image-2023-02-21-15-59-44-470.png

> Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by 
> native implementation
> -
>
> Key: FLINK-31059
> URL: https://issues.apache.org/jira/browse/FLINK-31059
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Hive
>Affects Versions: 1.17.0
>Reporter: dalongliu
>Assignee: miamiaoxyz
>Priority: Blocker
> Fix For: 1.17.0
>
> Attachments: image-2023-02-21-15-45-48-226.png, 
> image-2023-02-21-15-46-13-966.png, image-2023-02-21-15-47-54-043.png, 
> image-2023-02-21-15-49-58-854.png, image-2023-02-21-15-59-44-470.png
>
>
> This task aims to verify 
> [FLINK-29717|https://issues.apache.org/jira/browse/FLINK-29717] which 
> improves the hive udaf performance.
> As the document [PR|https://github.com/apache/flink/pull/21789] description, 
> please veriy:
> 1. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count/avg/min/max functions separately in the query to verify if the 
> hash-agg strategy is chosen via plan, and verify if the data results are the 
> same as when the option `table.exec.hive.native-agg-function.enabled` is 
> disabled.
> 2. Enabling the option `table.exec.hive.native-agg-function.enabled`, combine 
> sum/count/avg/min/max functions in query, verify if the hash-agg strategy is 
> chosen via plan, and verify if the data results are the same as when option 
> `table.exec.hive.native-agg-function.enabled` is disabled.
> 3. Enabling the option `table.exec.hive.native-agg-function.enabled`, count 
> or max array&struct and other complex types in query, verify whether the 
> sort-agg strategy is chosen via plan, verify whether the data result is the 
> same as when option `table.exec.hive.native-agg-function.enabled` is disabled.
> 4. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count and first_value/last_value functions in the query simultaneously, 
> verify that the sort-agg strategy is chosen via plan, verify that the data is 
> the same as when option `table.exec.hive.native-agg-function.enabled` is 
> disabled.
> 5. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count/avg/min/max functions in the query and open sort-agg strategy 
> forcibly, verify that the data results are the same as when option 
> `table.exec.hive.native-agg-function.enabled` is disabled.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] Samrat002 commented on pull request #21770: [FLINK-30277][python]Allow PYTHONPATH of Python Worker configurable

2023-02-21 Thread via GitHub


Samrat002 commented on PR #21770:
URL: https://github.com/apache/flink/pull/21770#issuecomment-1438005348

   Test failure associated ticket 
https://issues.apache.org/jira/browse/FLINK-31120


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Assigned] (FLINK-31120) ConcurrentModificationException occurred in StringFunctionsITCase.test

2023-02-21 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31120?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser reassigned FLINK-31120:
--

Assignee: Shuiqiang Chen

> ConcurrentModificationException occurred in StringFunctionsITCase.test
> --
>
> Key: FLINK-31120
> URL: https://issues.apache.org/jira/browse/FLINK-31120
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Runtime
>Affects Versions: 1.17.0
>Reporter: Matthias Pohl
>Assignee: Shuiqiang Chen
>Priority: Blocker
>  Labels: pull-request-available, test-stability
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=46255&view=logs&j=0c940707-2659-5648-cbe6-a1ad63045f0a&t=075c2716-8010-5565-fe08-3c4bb45824a4&l=12334
> {code}
> Feb 17 04:51:25 [ERROR] Tests run: 4, Failures: 0, Errors: 1, Skipped: 0, 
> Time elapsed: 10.725 s <<< FAILURE! - in 
> org.apache.flink.table.planner.functions.StringFunctionsITCase
> Feb 17 04:51:25 [ERROR] 
> org.apache.flink.table.planner.functions.StringFunctionsITCase.test(TestCase)[4]
>  Time elapsed: 4.367 s <<< ERROR!
> Feb 17 04:51:25 org.apache.flink.table.api.TableException: Failed to execute 
> sql
> Feb 17 04:51:25 at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.executeQueryOperation(TableEnvironmentImpl.java:974)
> Feb 17 04:51:25 at 
> org.apache.flink.table.api.internal.TableEnvironmentImpl.executeInternal(TableEnvironmentImpl.java:1422)
> Feb 17 04:51:25 at 
> org.apache.flink.table.api.internal.TableImpl.execute(TableImpl.java:476)
> Feb 17 04:51:25 at 
> org.apache.flink.table.planner.functions.BuiltInFunctionTestBase$ResultTestItem.test(BuiltInFunctionTestBase.java:354)
> Feb 17 04:51:25 at 
> org.apache.flink.table.planner.functions.BuiltInFunctionTestBase$TestSetSpec.lambda$getTestCase$4(BuiltInFunctionTestBase.java:320)
> Feb 17 04:51:25 at 
> org.apache.flink.table.planner.functions.BuiltInFunctionTestBase$TestCase.execute(BuiltInFunctionTestBase.java:113)
> Feb 17 04:51:25 at 
> org.apache.flink.table.planner.functions.BuiltInFunctionTestBase.test(BuiltInFunctionTestBase.java:93)
> Feb 17 04:51:25 at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> [...]
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-31109) Fails with proxy user not supported even when security.kerberos.fetch.delegation-token is set to false

2023-02-21 Thread Martijn Visser (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31109?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691447#comment-17691447
 ] 

Martijn Visser commented on FLINK-31109:


[~vsowrirajan] Is there any eta on this? Since its a blocker for 1.17, we would 
like to understand how long this is expecting to take before being resolved.

> Fails with proxy user not supported even when 
> security.kerberos.fetch.delegation-token is set to false
> --
>
> Key: FLINK-31109
> URL: https://issues.apache.org/jira/browse/FLINK-31109
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.17.0
>Reporter: Venkata krishnan Sowrirajan
>Priority: Blocker
>
> With
> {code:java}
> security.kerberos.fetch.delegation-token: false
> {code}
> and delegation tokens obtained through our internal service which sets both 
> HADOOP_TOKEN_FILE_LOCATION to pick up the DTs and also sets the 
> HADOOP_PROXY_USER which fails with the below error
> {code:java}
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/export/home/vsowrira/flink-1.18-SNAPSHOT/lib/log4j-slf4j-impl-2.17.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/export/apps/hadoop/hadoop-bin_2100503/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]
> org.apache.flink.runtime.security.modules.SecurityModule$SecurityInstallException:
>  Unable to set the Hadoop login user
>   at 
> org.apache.flink.runtime.security.modules.HadoopModule.install(HadoopModule.java:106)
>   at 
> org.apache.flink.runtime.security.SecurityUtils.installModules(SecurityUtils.java:76)
>   at 
> org.apache.flink.runtime.security.SecurityUtils.install(SecurityUtils.java:57)
>   at 
> org.apache.flink.client.cli.CliFrontend.mainInternal(CliFrontend.java:1188)
>   at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1157)
> Caused by: java.lang.UnsupportedOperationException: Proxy user is not 
> supported
>   at 
> org.apache.flink.runtime.security.token.hadoop.KerberosLoginProvider.throwProxyUserNotSupported(KerberosLoginProvider.java:137)
>   at 
> org.apache.flink.runtime.security.token.hadoop.KerberosLoginProvider.isLoginPossible(KerberosLoginProvider.java:81)
>   at 
> org.apache.flink.runtime.security.modules.HadoopModule.install(HadoopModule.java:73)
>   ... 4 more
> {code}
> This seems to have gotten changed after 
> [480e6edf|https://github.com/apache/flink/commit/480e6edf9732f8334ef7576080fdbfc98051cb28]
>  ([FLINK-28330][runtime][security] Remove old delegation token framework code)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-31109) Fails with proxy user not supported even when security.kerberos.fetch.delegation-token is set to false

2023-02-21 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser reassigned FLINK-31109:
--

Assignee: Venkata krishnan Sowrirajan

> Fails with proxy user not supported even when 
> security.kerberos.fetch.delegation-token is set to false
> --
>
> Key: FLINK-31109
> URL: https://issues.apache.org/jira/browse/FLINK-31109
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.17.0
>Reporter: Venkata krishnan Sowrirajan
>Assignee: Venkata krishnan Sowrirajan
>Priority: Blocker
>
> With
> {code:java}
> security.kerberos.fetch.delegation-token: false
> {code}
> and delegation tokens obtained through our internal service which sets both 
> HADOOP_TOKEN_FILE_LOCATION to pick up the DTs and also sets the 
> HADOOP_PROXY_USER which fails with the below error
> {code:java}
> SLF4J: Class path contains multiple SLF4J bindings.
> SLF4J: Found binding in 
> [jar:file:/export/home/vsowrira/flink-1.18-SNAPSHOT/lib/log4j-slf4j-impl-2.17.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: Found binding in 
> [jar:file:/export/apps/hadoop/hadoop-bin_2100503/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class]
> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an 
> explanation.
> SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]
> org.apache.flink.runtime.security.modules.SecurityModule$SecurityInstallException:
>  Unable to set the Hadoop login user
>   at 
> org.apache.flink.runtime.security.modules.HadoopModule.install(HadoopModule.java:106)
>   at 
> org.apache.flink.runtime.security.SecurityUtils.installModules(SecurityUtils.java:76)
>   at 
> org.apache.flink.runtime.security.SecurityUtils.install(SecurityUtils.java:57)
>   at 
> org.apache.flink.client.cli.CliFrontend.mainInternal(CliFrontend.java:1188)
>   at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1157)
> Caused by: java.lang.UnsupportedOperationException: Proxy user is not 
> supported
>   at 
> org.apache.flink.runtime.security.token.hadoop.KerberosLoginProvider.throwProxyUserNotSupported(KerberosLoginProvider.java:137)
>   at 
> org.apache.flink.runtime.security.token.hadoop.KerberosLoginProvider.isLoginPossible(KerberosLoginProvider.java:81)
>   at 
> org.apache.flink.runtime.security.modules.HadoopModule.install(HadoopModule.java:73)
>   ... 4 more
> {code}
> This seems to have gotten changed after 
> [480e6edf|https://github.com/apache/flink/commit/480e6edf9732f8334ef7576080fdbfc98051cb28]
>  ([FLINK-28330][runtime][security] Remove old delegation token framework code)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-31092) Hive ITCases fail with OutOfMemoryError

2023-02-21 Thread Martijn Visser (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31092?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691453#comment-17691453
 ] 

Martijn Visser commented on FLINK-31092:


[~luoyuxia] Can you take a look?

> Hive ITCases fail with OutOfMemoryError
> ---
>
> Key: FLINK-31092
> URL: https://issues.apache.org/jira/browse/FLINK-31092
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.17.0
>Reporter: Matthias Pohl
>Priority: Critical
>  Labels: test-stability
> Attachments: VisualVM-FLINK-31092.png
>
>
> We're experiencing a OutOfMemoryError where the heap space reaches the upper 
> limit:
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=46161&view=logs&j=fc5181b0-e452-5c8f-68de-1097947f6483&t=995c650b-6573-581c-9ce6-7ad4cc038461&l=23142
> {code}
> Feb 15 05:05:14 [INFO] Running 
> org.apache.flink.table.catalog.hive.HiveCatalogITCase
> Feb 15 05:05:17 [INFO] java.lang.OutOfMemoryError: Java heap space
> Feb 15 05:05:17 [INFO] Dumping heap to java_pid9669.hprof ...
> Feb 15 05:05:28 [INFO] Heap dump file created [1957090051 bytes in 11.718 
> secs]
> java.lang.OutOfMemoryError: Java heap space
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.cancelPingScheduler(ForkedBooter.java:209)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.acknowledgedExit(ForkedBooter.java:419)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:186)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548)
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-ml] zhipeng93 commented on a diff in pull request #213: [FLINK-31126] Move classes not depending on Flink runtime from flink-ml-core to flink-ml-servable-core

2023-02-21 Thread via GitHub


zhipeng93 commented on code in PR #213:
URL: https://github.com/apache/flink-ml/pull/213#discussion_r1112704530


##
flink-ml-servable-core/pom.xml:
##
@@ -0,0 +1,127 @@
+
+
+http://maven.apache.org/POM/4.0.0";
+ xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance";
+ xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/xsd/maven-4.0.0.xsd";>
+  4.0.0
+
+  
+org.apache.flink
+flink-ml-parent
+2.2-SNAPSHOT
+  
+
+  flink-ml-servable-core
+  Flink ML : Servable : Core
+
+  
+
+
+  org.apache.flink

Review Comment:
   It seems that `flink-ml-servable-core` still relies on `flink-core` and is 
not consistent with the goal of FLIP-289.
   Can you explain why do we move XXTypeInfo into `flink-ml-servable-core`? 
   
   Moreover, there are some classes that do not depend on flink runtime, but 
not used in `flink-ml-servable-core` (e.g., `MLMetrics`). Can we move it back 
where it is used?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-ml] zhipeng93 commented on a diff in pull request #213: [FLINK-31126] Move classes not depending on Flink runtime from flink-ml-core to flink-ml-servable-core

2023-02-21 Thread via GitHub


zhipeng93 commented on code in PR #213:
URL: https://github.com/apache/flink-ml/pull/213#discussion_r1112704530


##
flink-ml-servable-core/pom.xml:
##
@@ -0,0 +1,127 @@
+
+
+http://maven.apache.org/POM/4.0.0";
+ xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance";
+ xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/xsd/maven-4.0.0.xsd";>
+  4.0.0
+
+  
+org.apache.flink
+flink-ml-parent
+2.2-SNAPSHOT
+  
+
+  flink-ml-servable-core
+  Flink ML : Servable : Core
+
+  
+
+
+  org.apache.flink

Review Comment:
   It seems that `flink-ml-servable-core` still relies on `flink-core` and is 
not consistent with the goal of FLIP-289. Can you explain why do we move 
XXTypeInfo into `flink-ml-servable-core`? 
   
   Moreover, there are some classes that do not depend on flink runtime, but 
not used in `flink-ml-servable-core` (e.g., `MLMetrics`). Can we move it back 
where it is used?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Assigned] (FLINK-31159) Release Testing: Verify FLINK-30583 Provide the flame graph to the subtask level

2023-02-21 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31159?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser reassigned FLINK-31159:
--

Assignee: Hong Liang Teoh  (was: Jane Chan)

> Release Testing: Verify FLINK-30583 Provide the flame graph to the subtask 
> level
> 
>
> Key: FLINK-31159
> URL: https://issues.apache.org/jira/browse/FLINK-31159
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / REST, Runtime / Web Frontend
>Affects Versions: 1.17.0
>Reporter: Hong Liang Teoh
>Assignee: Hong Liang Teoh
>Priority: Major
> Fix For: 1.17.0
>
> Attachments: Map_record_distribution.png, all_subtasks.png, 
> subtask0.png, subtask3.png
>
>
> The issue aims to verify FLINK-30583.
> Please verify:
>  # When below conditions are met:
>  ** Job has more than 1 parallel subtask
>  ** Some subtasks are busy, whilst others are not (e.g. some receive records, 
> some don't)
>  ** The FlameGraph accurately reflects the busy subtask



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Resolved] (FLINK-31159) Release Testing: Verify FLINK-30583 Provide the flame graph to the subtask level

2023-02-21 Thread Hong Liang Teoh (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31159?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hong Liang Teoh resolved FLINK-31159.
-
Resolution: Fixed

This issue has been completed! Sorry for the churn [~qingyue] + [~renqs] 

> Release Testing: Verify FLINK-30583 Provide the flame graph to the subtask 
> level
> 
>
> Key: FLINK-31159
> URL: https://issues.apache.org/jira/browse/FLINK-31159
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / REST, Runtime / Web Frontend
>Affects Versions: 1.17.0
>Reporter: Hong Liang Teoh
>Assignee: Hong Liang Teoh
>Priority: Major
> Fix For: 1.17.0
>
> Attachments: Map_record_distribution.png, all_subtasks.png, 
> subtask0.png, subtask3.png
>
>
> The issue aims to verify FLINK-30583.
> Please verify:
>  # When below conditions are met:
>  ** Job has more than 1 parallel subtask
>  ** Some subtasks are busy, whilst others are not (e.g. some receive records, 
> some don't)
>  ** The FlameGraph accurately reflects the busy subtask



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-31159) Release Testing: Verify FLINK-30583 Provide the flame graph to the subtask level

2023-02-21 Thread Hong Liang Teoh (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31159?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hong Liang Teoh closed FLINK-31159.
---

> Release Testing: Verify FLINK-30583 Provide the flame graph to the subtask 
> level
> 
>
> Key: FLINK-31159
> URL: https://issues.apache.org/jira/browse/FLINK-31159
> Project: Flink
>  Issue Type: Sub-task
>  Components: Runtime / REST, Runtime / Web Frontend
>Affects Versions: 1.17.0
>Reporter: Hong Liang Teoh
>Assignee: Hong Liang Teoh
>Priority: Major
> Fix For: 1.17.0
>
> Attachments: Map_record_distribution.png, all_subtasks.png, 
> subtask0.png, subtask3.png
>
>
> The issue aims to verify FLINK-30583.
> Please verify:
>  # When below conditions are met:
>  ** Job has more than 1 parallel subtask
>  ** Some subtasks are busy, whilst others are not (e.g. some receive records, 
> some don't)
>  ** The FlameGraph accurately reflects the busy subtask



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31059) Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by native implementation

2023-02-21 Thread miamiaoxyz (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31059?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

miamiaoxyz updated FLINK-31059:
---
Attachment: image-2023-02-21-16-28-22-038.png

> Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by 
> native implementation
> -
>
> Key: FLINK-31059
> URL: https://issues.apache.org/jira/browse/FLINK-31059
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Hive
>Affects Versions: 1.17.0
>Reporter: dalongliu
>Assignee: miamiaoxyz
>Priority: Blocker
> Fix For: 1.17.0
>
> Attachments: image-2023-02-21-15-45-48-226.png, 
> image-2023-02-21-15-46-13-966.png, image-2023-02-21-15-47-54-043.png, 
> image-2023-02-21-15-49-58-854.png, image-2023-02-21-15-59-44-470.png, 
> image-2023-02-21-16-28-22-038.png
>
>
> This task aims to verify 
> [FLINK-29717|https://issues.apache.org/jira/browse/FLINK-29717] which 
> improves the hive udaf performance.
> As the document [PR|https://github.com/apache/flink/pull/21789] description, 
> please veriy:
> 1. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count/avg/min/max functions separately in the query to verify if the 
> hash-agg strategy is chosen via plan, and verify if the data results are the 
> same as when the option `table.exec.hive.native-agg-function.enabled` is 
> disabled.
> 2. Enabling the option `table.exec.hive.native-agg-function.enabled`, combine 
> sum/count/avg/min/max functions in query, verify if the hash-agg strategy is 
> chosen via plan, and verify if the data results are the same as when option 
> `table.exec.hive.native-agg-function.enabled` is disabled.
> 3. Enabling the option `table.exec.hive.native-agg-function.enabled`, count 
> or max array&struct and other complex types in query, verify whether the 
> sort-agg strategy is chosen via plan, verify whether the data result is the 
> same as when option `table.exec.hive.native-agg-function.enabled` is disabled.
> 4. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count and first_value/last_value functions in the query simultaneously, 
> verify that the sort-agg strategy is chosen via plan, verify that the data is 
> the same as when option `table.exec.hive.native-agg-function.enabled` is 
> disabled.
> 5. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count/avg/min/max functions in the query and open sort-agg strategy 
> forcibly, verify that the data results are the same as when option 
> `table.exec.hive.native-agg-function.enabled` is disabled.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31059) Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by native implementation

2023-02-21 Thread miamiaoxyz (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31059?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

miamiaoxyz updated FLINK-31059:
---
Attachment: image-2023-02-21-16-29-42-983.png

> Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by 
> native implementation
> -
>
> Key: FLINK-31059
> URL: https://issues.apache.org/jira/browse/FLINK-31059
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Hive
>Affects Versions: 1.17.0
>Reporter: dalongliu
>Assignee: miamiaoxyz
>Priority: Blocker
> Fix For: 1.17.0
>
> Attachments: image-2023-02-21-15-45-48-226.png, 
> image-2023-02-21-15-46-13-966.png, image-2023-02-21-15-47-54-043.png, 
> image-2023-02-21-15-49-58-854.png, image-2023-02-21-15-59-44-470.png, 
> image-2023-02-21-16-28-22-038.png, image-2023-02-21-16-29-42-983.png
>
>
> This task aims to verify 
> [FLINK-29717|https://issues.apache.org/jira/browse/FLINK-29717] which 
> improves the hive udaf performance.
> As the document [PR|https://github.com/apache/flink/pull/21789] description, 
> please veriy:
> 1. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count/avg/min/max functions separately in the query to verify if the 
> hash-agg strategy is chosen via plan, and verify if the data results are the 
> same as when the option `table.exec.hive.native-agg-function.enabled` is 
> disabled.
> 2. Enabling the option `table.exec.hive.native-agg-function.enabled`, combine 
> sum/count/avg/min/max functions in query, verify if the hash-agg strategy is 
> chosen via plan, and verify if the data results are the same as when option 
> `table.exec.hive.native-agg-function.enabled` is disabled.
> 3. Enabling the option `table.exec.hive.native-agg-function.enabled`, count 
> or max array&struct and other complex types in query, verify whether the 
> sort-agg strategy is chosen via plan, verify whether the data result is the 
> same as when option `table.exec.hive.native-agg-function.enabled` is disabled.
> 4. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count and first_value/last_value functions in the query simultaneously, 
> verify that the sort-agg strategy is chosen via plan, verify that the data is 
> the same as when option `table.exec.hive.native-agg-function.enabled` is 
> disabled.
> 5. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count/avg/min/max functions in the query and open sort-agg strategy 
> forcibly, verify that the data results are the same as when option 
> `table.exec.hive.native-agg-function.enabled` is disabled.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] snuyanzin commented on pull request #21958: [FLINK-31118][table] Add ARRAY_UNION function.

2023-02-21 Thread via GitHub


snuyanzin commented on PR #21958:
URL: https://github.com/apache/flink/pull/21958#issuecomment-1438042696

   > so i ask you do you have a good implements
   
   unfortunately right now i do not have where i can just pick it up. You 
probably could do some research in this direction


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-31059) Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by native implementation

2023-02-21 Thread miamiaoxyz (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31059?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

miamiaoxyz updated FLINK-31059:
---
Attachment: image-2023-02-21-16-31-58-361.png

> Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by 
> native implementation
> -
>
> Key: FLINK-31059
> URL: https://issues.apache.org/jira/browse/FLINK-31059
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Hive
>Affects Versions: 1.17.0
>Reporter: dalongliu
>Assignee: miamiaoxyz
>Priority: Blocker
> Fix For: 1.17.0
>
> Attachments: image-2023-02-21-15-45-48-226.png, 
> image-2023-02-21-15-46-13-966.png, image-2023-02-21-15-47-54-043.png, 
> image-2023-02-21-15-49-58-854.png, image-2023-02-21-15-59-44-470.png, 
> image-2023-02-21-16-28-22-038.png, image-2023-02-21-16-29-42-983.png, 
> image-2023-02-21-16-31-58-361.png
>
>
> This task aims to verify 
> [FLINK-29717|https://issues.apache.org/jira/browse/FLINK-29717] which 
> improves the hive udaf performance.
> As the document [PR|https://github.com/apache/flink/pull/21789] description, 
> please veriy:
> 1. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count/avg/min/max functions separately in the query to verify if the 
> hash-agg strategy is chosen via plan, and verify if the data results are the 
> same as when the option `table.exec.hive.native-agg-function.enabled` is 
> disabled.
> 2. Enabling the option `table.exec.hive.native-agg-function.enabled`, combine 
> sum/count/avg/min/max functions in query, verify if the hash-agg strategy is 
> chosen via plan, and verify if the data results are the same as when option 
> `table.exec.hive.native-agg-function.enabled` is disabled.
> 3. Enabling the option `table.exec.hive.native-agg-function.enabled`, count 
> or max array&struct and other complex types in query, verify whether the 
> sort-agg strategy is chosen via plan, verify whether the data result is the 
> same as when option `table.exec.hive.native-agg-function.enabled` is disabled.
> 4. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count and first_value/last_value functions in the query simultaneously, 
> verify that the sort-agg strategy is chosen via plan, verify that the data is 
> the same as when option `table.exec.hive.native-agg-function.enabled` is 
> disabled.
> 5. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count/avg/min/max functions in the query and open sort-agg strategy 
> forcibly, verify that the data results are the same as when option 
> `table.exec.hive.native-agg-function.enabled` is disabled.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-31164) Unexpected correlate variable $cor0 in the plan error in where clause

2023-02-21 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31164?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser closed FLINK-31164.
--
Resolution: Duplicate

> Unexpected correlate variable $cor0 in the plan error in where clause
> -
>
> Key: FLINK-31164
> URL: https://issues.apache.org/jira/browse/FLINK-31164
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.16.0
>Reporter: P Rohan Kumar
>Priority: Major
>
> {code:java}
> val env: StreamExecutionEnvironment = 
> StreamExecutionEnvironment.getExecutionEnvironment
> val tableEnv = StreamTableEnvironment.create(env)
> val accountsTd = 
> TableDescriptor.forConnector("datagen").option("rows-per-second", "10")
>   .option("number-of-rows", "10")
>   .schema(Schema
> .newBuilder()
> .column("account_num", DataTypes.VARCHAR(2147483647))
> .column("acc_name", DataTypes.VARCHAR(2147483647))
> .column("acc_phone_num", DataTypes.VARCHAR(2147483647))
> .build())
>   .build()
> val accountsTable = tableEnv.from(accountsTd)
> tableEnv.createTemporaryView("accounts", accountsTable)
> val transactionsTd = 
> TableDescriptor.forConnector("datagen").option("rows-per-second", "10")
>   .option("number-of-rows", "10")
>   .schema(Schema
> .newBuilder()
> .column("account_num", DataTypes.VARCHAR(2147483647))
> .column("transaction_place", DataTypes.VARCHAR(2147483647))
> .column("transaction_time", DataTypes.BIGINT())
> .column("amount", DataTypes.INT())
> .build())
>   .build()
> val transactionsTable = tableEnv.from(transactionsTd)
> tableEnv.createTemporaryView("transaction_data", transactionsTable)
> val newTable = tableEnv.sqlQuery("select   acc.account_num,  (select count(*) 
> from transaction_data where transaction_place = trans.transaction_place and 
> account_num = acc.account_num)  from  accounts acc,transaction_data trans")
> tableEnv.toChangelogStream(newTable).print()
> env.execute() {code}
> I get the following error if I run the above code.
>  
> {code:java}
> Exception in thread "main" org.apache.flink.table.api.TableException: 
> unexpected correlate variable $cor0 in the plan
>     at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkDecorrelateProgram.checkCorrelVariableExists(FlinkDecorrelateProgram.scala:59)
>     at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkDecorrelateProgram.optimize(FlinkDecorrelateProgram.scala:42)
>     at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkGroupProgram.$anonfun$optimize$2(FlinkGroupProgram.scala:59)
>     at 
> scala.collection.TraversableOnce$folder$1$.apply(TraversableOnce.scala:187)
>     at 
> scala.collection.TraversableOnce$folder$1$.apply(TraversableOnce.scala:185)
>     at scala.collection.Iterator.foreach(Iterator.scala:943)
>     at scala.collection.Iterator.foreach$(Iterator.scala:943)
>     at scala.collection.AbstractIterator.foreach(Iterator.scala:1431)
>     at scala.collection.IterableLike.foreach(IterableLike.scala:74)
>     at scala.collection.IterableLike.foreach$(IterableLike.scala:73)
>     at scala.collection.AbstractIterable.foreach(Iterable.scala:56)
>     at scala.collection.TraversableOnce.foldLeft(TraversableOnce.scala:189)
>     at scala.collection.TraversableOnce.foldLeft$(TraversableOnce.scala:184)
>     at scala.collection.AbstractTraversable.foldLeft(Traversable.scala:108)
>     at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkGroupProgram.$anonfun$optimize$1(FlinkGroupProgram.scala:56)
>     at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkGroupProgram.$anonfun$optimize$1$adapted(FlinkGroupProgram.scala:51)
>     at 
> scala.collection.TraversableOnce$folder$1$.apply(TraversableOnce.scala:187)
>     at 
> scala.collection.TraversableOnce$folder$1$.apply(TraversableOnce.scala:185)
>     at scala.collection.immutable.Range.foreach(Range.scala:158)
>     at scala.collection.TraversableOnce.foldLeft(TraversableOnce.scala:189)
>     at scala.collection.TraversableOnce.foldLeft$(TraversableOnce.scala:184)
>     at scala.collection.AbstractTraversable.foldLeft(Traversable.scala:108)
>     at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkGroupProgram.optimize(FlinkGroupProgram.scala:51)
>     at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkChainedProgram.$anonfun$optimize$1(FlinkChainedProgram.scala:59)
>     at 
> scala.collection.TraversableOnce$folder$1$.apply(TraversableOnce.scala:187)
>     at 
> scala.collection.TraversableOnce$folder$1$.apply(TraversableOnce.scala:185)
>     at scala.collection.Iterator.foreach(Iterator.scala:943)
>     at scala.collection.Iterator.foreach$(Iterator.scala:943)
>     at scala.collection.AbstractIterator.foreach(Iterator.scala:1431)
>     at scala.colle

[jira] [Updated] (FLINK-31059) Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by native implementation

2023-02-21 Thread miamiaoxyz (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31059?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

miamiaoxyz updated FLINK-31059:
---
Attachment: image-2023-02-21-16-35-46-294.png

> Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by 
> native implementation
> -
>
> Key: FLINK-31059
> URL: https://issues.apache.org/jira/browse/FLINK-31059
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Hive
>Affects Versions: 1.17.0
>Reporter: dalongliu
>Assignee: miamiaoxyz
>Priority: Blocker
> Fix For: 1.17.0
>
> Attachments: image-2023-02-21-15-45-48-226.png, 
> image-2023-02-21-15-46-13-966.png, image-2023-02-21-15-47-54-043.png, 
> image-2023-02-21-15-49-58-854.png, image-2023-02-21-15-59-44-470.png, 
> image-2023-02-21-16-28-22-038.png, image-2023-02-21-16-29-42-983.png, 
> image-2023-02-21-16-31-58-361.png, image-2023-02-21-16-35-46-294.png
>
>
> This task aims to verify 
> [FLINK-29717|https://issues.apache.org/jira/browse/FLINK-29717] which 
> improves the hive udaf performance.
> As the document [PR|https://github.com/apache/flink/pull/21789] description, 
> please veriy:
> 1. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count/avg/min/max functions separately in the query to verify if the 
> hash-agg strategy is chosen via plan, and verify if the data results are the 
> same as when the option `table.exec.hive.native-agg-function.enabled` is 
> disabled.
> 2. Enabling the option `table.exec.hive.native-agg-function.enabled`, combine 
> sum/count/avg/min/max functions in query, verify if the hash-agg strategy is 
> chosen via plan, and verify if the data results are the same as when option 
> `table.exec.hive.native-agg-function.enabled` is disabled.
> 3. Enabling the option `table.exec.hive.native-agg-function.enabled`, count 
> or max array&struct and other complex types in query, verify whether the 
> sort-agg strategy is chosen via plan, verify whether the data result is the 
> same as when option `table.exec.hive.native-agg-function.enabled` is disabled.
> 4. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count and first_value/last_value functions in the query simultaneously, 
> verify that the sort-agg strategy is chosen via plan, verify that the data is 
> the same as when option `table.exec.hive.native-agg-function.enabled` is 
> disabled.
> 5. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count/avg/min/max functions in the query and open sort-agg strategy 
> forcibly, verify that the data results are the same as when option 
> `table.exec.hive.native-agg-function.enabled` is disabled.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31142) Some queries lead to abrupt sql client close

2023-02-21 Thread Sergey Nuyanzin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31142?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Nuyanzin updated FLINK-31142:

Description: 
Although the behavior has been changed in 1.17.0, I'm not sure whether it is a 
blocker or not, since in both cases it is invalid query.

The difference in the behavior is that before 1.17.0
a query like 
{code:sql}
select /* multiline comment;
{code}
fails to execute and sql client prompts to submit another query.

In 1.17.0 it fails with 
{noformat}
Exception in thread "main" org.apache.flink.table.client.SqlClientException: 
Could not read from command line.
at 
org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:205)
at 
org.apache.flink.table.client.cli.CliClient.executeInteractive(CliClient.java:168)
at 
org.apache.flink.table.client.cli.CliClient.executeInInteractiveMode(CliClient.java:113)
at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:169)
at org.apache.flink.table.client.SqlClient.start(SqlClient.java:118)
at 
org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:228)
at org.apache.flink.table.client.SqlClient.main(SqlClient.java:179)
Caused by: org.apache.flink.sql.parser.impl.TokenMgrError: Lexical error at 
line 1, column 29.  Encountered:  after : ""
at 
org.apache.flink.sql.parser.impl.FlinkSqlParserImplTokenManager.getNextToken(FlinkSqlParserImplTokenManager.java:26752)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.scan(SqlCommandParserImpl.java:89)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.next(SqlCommandParserImpl.java:81)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.checkIncompleteStatement(SqlCommandParserImpl.java:141)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.getCommand(SqlCommandParserImpl.java:111)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.parseStatement(SqlCommandParserImpl.java:52)
at 
org.apache.flink.table.client.cli.parser.SqlMultiLineParser.parse(SqlMultiLineParser.java:82)
at 
org.jline.reader.impl.LineReaderImpl.acceptLine(LineReaderImpl.java:2964)
at 
org.jline.reader.impl.LineReaderImpl$1.apply(LineReaderImpl.java:3778)
at 
org.jline.reader.impl.LineReaderImpl.readLine(LineReaderImpl.java:679)
at 
org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:183)
... 6 more

Shutting down the session...
done.

{noformat}

And it shuts down the sessoin

  was:
Although the behavior has been changed in 1.17.0, I'm not sure whether it is a 
blocker or not, since in both cases it is invalid query.

The difference in the behavior is that before 1.17.0
a query like 
{code:sql}
select /* multiline comment;
{code}
query fails to execute and sql client prompts to submit another query.

In 1.17.0 it fails with 
{noformat}
Exception in thread "main" org.apache.flink.table.client.SqlClientException: 
Could not read from command line.
at 
org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:205)
at 
org.apache.flink.table.client.cli.CliClient.executeInteractive(CliClient.java:168)
at 
org.apache.flink.table.client.cli.CliClient.executeInInteractiveMode(CliClient.java:113)
at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:169)
at org.apache.flink.table.client.SqlClient.start(SqlClient.java:118)
at 
org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:228)
at org.apache.flink.table.client.SqlClient.main(SqlClient.java:179)
Caused by: org.apache.flink.sql.parser.impl.TokenMgrError: Lexical error at 
line 1, column 29.  Encountered:  after : ""
at 
org.apache.flink.sql.parser.impl.FlinkSqlParserImplTokenManager.getNextToken(FlinkSqlParserImplTokenManager.java:26752)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.scan(SqlCommandParserImpl.java:89)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.next(SqlCommandParserImpl.java:81)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.checkIncompleteStatement(SqlCommandParserImpl.java:141)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.getCommand(SqlCommandParserImpl.java:111)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.parseStatement(SqlCommandParserImpl.java:52)
at 
org.apache.flink.table.client.cli.parser.SqlMultiLineParser.parse(SqlMultiLineParser.java:82)
at 
org.jline.reader.impl.LineReaderImpl.acceptLine(LineReaderImpl.java:2964)
at 
org.jline.reader.impl.LineReaderImpl$1.apply(LineReaderImpl.java:3778)
at 
org.jline.reader.impl.LineReaderIm

[jira] [Updated] (FLINK-31142) Some queries lead to abrupt sql client close

2023-02-21 Thread Sergey Nuyanzin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31142?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Nuyanzin updated FLINK-31142:

Description: 
Although the behavior has been changed in 1.17.0, I'm not sure whether it is a 
blocker or not, since in both cases it is invalid query.

The difference in the behavior is that before 1.17.0
a query like 
{code:sql}
select /* multiline comment;
{code}
fails to execute and sql client prompts to submit another query.

In 1.17.0 it  shuts down the session failing with 
{noformat}
Exception in thread "main" org.apache.flink.table.client.SqlClientException: 
Could not read from command line.
at 
org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:205)
at 
org.apache.flink.table.client.cli.CliClient.executeInteractive(CliClient.java:168)
at 
org.apache.flink.table.client.cli.CliClient.executeInInteractiveMode(CliClient.java:113)
at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:169)
at org.apache.flink.table.client.SqlClient.start(SqlClient.java:118)
at 
org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:228)
at org.apache.flink.table.client.SqlClient.main(SqlClient.java:179)
Caused by: org.apache.flink.sql.parser.impl.TokenMgrError: Lexical error at 
line 1, column 29.  Encountered:  after : ""
at 
org.apache.flink.sql.parser.impl.FlinkSqlParserImplTokenManager.getNextToken(FlinkSqlParserImplTokenManager.java:26752)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.scan(SqlCommandParserImpl.java:89)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.next(SqlCommandParserImpl.java:81)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.checkIncompleteStatement(SqlCommandParserImpl.java:141)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.getCommand(SqlCommandParserImpl.java:111)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.parseStatement(SqlCommandParserImpl.java:52)
at 
org.apache.flink.table.client.cli.parser.SqlMultiLineParser.parse(SqlMultiLineParser.java:82)
at 
org.jline.reader.impl.LineReaderImpl.acceptLine(LineReaderImpl.java:2964)
at 
org.jline.reader.impl.LineReaderImpl$1.apply(LineReaderImpl.java:3778)
at 
org.jline.reader.impl.LineReaderImpl.readLine(LineReaderImpl.java:679)
at 
org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:183)
... 6 more

Shutting down the session...
done.

{noformat}



  was:
Although the behavior has been changed in 1.17.0, I'm not sure whether it is a 
blocker or not, since in both cases it is invalid query.

The difference in the behavior is that before 1.17.0
a query like 
{code:sql}
select /* multiline comment;
{code}
fails to execute and sql client prompts to submit another query.

In 1.17.0 it fails with 
{noformat}
Exception in thread "main" org.apache.flink.table.client.SqlClientException: 
Could not read from command line.
at 
org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:205)
at 
org.apache.flink.table.client.cli.CliClient.executeInteractive(CliClient.java:168)
at 
org.apache.flink.table.client.cli.CliClient.executeInInteractiveMode(CliClient.java:113)
at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:169)
at org.apache.flink.table.client.SqlClient.start(SqlClient.java:118)
at 
org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:228)
at org.apache.flink.table.client.SqlClient.main(SqlClient.java:179)
Caused by: org.apache.flink.sql.parser.impl.TokenMgrError: Lexical error at 
line 1, column 29.  Encountered:  after : ""
at 
org.apache.flink.sql.parser.impl.FlinkSqlParserImplTokenManager.getNextToken(FlinkSqlParserImplTokenManager.java:26752)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.scan(SqlCommandParserImpl.java:89)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.next(SqlCommandParserImpl.java:81)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.checkIncompleteStatement(SqlCommandParserImpl.java:141)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.getCommand(SqlCommandParserImpl.java:111)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.parseStatement(SqlCommandParserImpl.java:52)
at 
org.apache.flink.table.client.cli.parser.SqlMultiLineParser.parse(SqlMultiLineParser.java:82)
at 
org.jline.reader.impl.LineReaderImpl.acceptLine(LineReaderImpl.java:2964)
at 
org.jline.reader.impl.LineReaderImpl$1.apply(LineReaderImpl.java:3778)
at 
org.jline.reader.impl.LineReaderImpl.readLi

[jira] [Updated] (FLINK-31142) Some queries lead to abrupt sql client close

2023-02-21 Thread Sergey Nuyanzin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31142?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Nuyanzin updated FLINK-31142:

Labels:   (was: release-testing)

> Some queries lead to abrupt sql client close
> 
>
> Key: FLINK-31142
> URL: https://issues.apache.org/jira/browse/FLINK-31142
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.17.0
>Reporter: Sergey Nuyanzin
>Priority: Major
>
> Although the behavior has been changed in 1.17.0, I'm not sure whether it is 
> a blocker or not, since in both cases it is invalid query.
> The difference in the behavior is that before 1.17.0
> a query like 
> {code:sql}
> select /* multiline comment;
> {code}
> fails to execute and sql client prompts to submit another query.
> In 1.17.0 it  shuts down the session failing with 
> {noformat}
> Exception in thread "main" org.apache.flink.table.client.SqlClientException: 
> Could not read from command line.
>   at 
> org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:205)
>   at 
> org.apache.flink.table.client.cli.CliClient.executeInteractive(CliClient.java:168)
>   at 
> org.apache.flink.table.client.cli.CliClient.executeInInteractiveMode(CliClient.java:113)
>   at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:169)
>   at org.apache.flink.table.client.SqlClient.start(SqlClient.java:118)
>   at 
> org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:228)
>   at org.apache.flink.table.client.SqlClient.main(SqlClient.java:179)
> Caused by: org.apache.flink.sql.parser.impl.TokenMgrError: Lexical error at 
> line 1, column 29.  Encountered:  after : ""
>   at 
> org.apache.flink.sql.parser.impl.FlinkSqlParserImplTokenManager.getNextToken(FlinkSqlParserImplTokenManager.java:26752)
>   at 
> org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.scan(SqlCommandParserImpl.java:89)
>   at 
> org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.next(SqlCommandParserImpl.java:81)
>   at 
> org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.checkIncompleteStatement(SqlCommandParserImpl.java:141)
>   at 
> org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.getCommand(SqlCommandParserImpl.java:111)
>   at 
> org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.parseStatement(SqlCommandParserImpl.java:52)
>   at 
> org.apache.flink.table.client.cli.parser.SqlMultiLineParser.parse(SqlMultiLineParser.java:82)
>   at 
> org.jline.reader.impl.LineReaderImpl.acceptLine(LineReaderImpl.java:2964)
>   at 
> org.jline.reader.impl.LineReaderImpl$1.apply(LineReaderImpl.java:3778)
>   at 
> org.jline.reader.impl.LineReaderImpl.readLine(LineReaderImpl.java:679)
>   at 
> org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:183)
>   ... 6 more
> Shutting down the session...
> done.
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] rkhachatryan opened a new pull request, #21981: [FLINK-21450][runtime] Support LocalRecovery by AdaptiveScheduler

2023-02-21 Thread via GitHub


rkhachatryan opened a new pull request, #21981:
URL: https://github.com/apache/flink/pull/21981

   ## What is the purpose of the change
   
   Adjust slot assignment by Adaptive Scheduler
   to try to re-use previous allocations
   so that TMs can use Local Recovery.
   
   Contributed mostly by @dmvk.
   The main defferences from the original contribution:
   1. Previous `ExecutionGraph` is passed from the previous state explicitly 
(currently, `WaitingForResources` stage, which triggers the computation, 
doesn't have the graph)
   2. In `SlotAssigner`, the split into two methods is removed mostly for 
consistency (two methods mostly duplicated each other). That results in higher 
asymptotical complexity of `StateLocalitySlotAssigner` (`O(mnlog*mnlog)` vs 
`O(mnlog`)
   3. DoP is computed according to FLINK-30895
   
   ## Brief change log
   
- Support LocalRecovery by AdaptiveScheduler
- Add previous ExecutionGraph to WaitingForResources AdaptiveScheduler state
- Make LocalRecoveryITCase fail when allocations don't match
   
   ## Verifying this change
   
   - Adjusted `LocalRecoveryITCase`
   - Added `SlotSharingSlotAllocatorTest.testStickyAllocation`
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / no)
 - The serializers: (yes / no / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / no / 
don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / no / don't know)
 - The S3 file system connector: (yes / no / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / no)
 - If yes, how is the feature documented? (not applicable / docs / JavaDocs 
/ not documented)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-21450) Add local recovery support to adaptive scheduler

2023-02-21 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-21450?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-21450:
---
Labels: auto-deprioritized-major auto-deprioritized-minor auto-unassigned 
pull-request-available  (was: auto-deprioritized-major auto-deprioritized-minor 
auto-unassigned)

> Add local recovery support to adaptive scheduler
> 
>
> Key: FLINK-21450
> URL: https://issues.apache.org/jira/browse/FLINK-21450
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Coordination
>Reporter: Robert Metzger
>Assignee: Roman Khachatryan
>Priority: Major
>  Labels: auto-deprioritized-major, auto-deprioritized-minor, 
> auto-unassigned, pull-request-available
> Fix For: 1.18.0
>
>
> local recovery means that, on a failure, we are able to re-use the state in a 
> taskmanager, instead of loading it again from distributed storage (which 
> means the scheduler needs to know where which state is located, and schedule 
> tasks accordingly).
> Adaptive Scheduler is currently not respecting the location of state, so 
> failures require the re-loading of state from the distributed storage.
> Adding this feature will allow us to enable the {{Local recovery and sticky 
> scheduling end-to-end test}} for adaptive scheduler again.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31153:
--
Description: 
If you are doing a new major release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Minor releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
releasing/update_branch_version.sh
{code}
In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 as the last entry:
{code:java}
// ...
v1_12("1.12"),
v1_13("1.13"),
v1_14("1.14"),
v1_15("1.15"),
v1_16("1.16");
{code}
The newly created branch and updated {{master}} branch need to be pushed to the 
official repository.
h4. Flink Docker Repository

Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
[apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
sure that 
[apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
 points to the correct snapshot version; for {{dev-x.y}} it should point to 
{{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).

After pushing the new major release branch, as the last step you should also 
update the documentation workflow to also build the documentation for the new 
release branch. Check [Managing 
Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
 on details on how to do that. You may also want to manually trigger a build to 
make the changes visible as soon as possible.

 

h3. Expectations (Minor Version only)
 * Release branch has been created and pushed
 * Cron job has been added on the release branch in 
([apache-flink:./tools/azure-pipelines/build-apache-repo.yml|https://github.com/apache/flink/blob/master/tools/azure-pipelines/build-apache-repo.yml])
 * Originating branch has the version information updated to the new version
 * New version is added to the 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum.
 * Make sure {{flink-docker}} has {{dev-x.y}} branch and docker e2e tests run 
against this branch
 * docs/config.toml has been updated appropriately.
 * The {{dev-x.y}} branch of ({{{}$CURRENT_SNAPSHOT_VERSION{}}}) have been 
created in the Flink Benchmark repo.
 * The {{flink.version}} property of Flink Benchmark repo has been updated to 
the latest snapshot version.

  was:
If you are doing a new major release, you need to update Flink version in two 
repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Minor releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
releasing/update_branch_version.sh
{code}
In the {{master}} branch, add a new value (e.g. {{v1_16("1.16")}}) to 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 as the last entry:
{code:java}
// ...
v1_12("1.12"),
v1_13("1.13"),
v1_14("1.14"),
v1_15("1.15"),
v1_16("1.16");
{code}
The newly created branch and updated {{master}} branch need to be pushed to the 
official repository.
h4. Flink Docker Repository

Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
[apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
sure that 
[apache/flink-docker:.gi

[GitHub] [flink] reswqa commented on a diff in pull request #21977: [FLINK-31132] Compact without setting parallelism does not follow the configured sink parallelism for HiveTableSink

2023-02-21 Thread via GitHub


reswqa commented on code in PR #21977:
URL: https://github.com/apache/flink/pull/21977#discussion_r1112733682


##
flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableCompactSinkTest.java:
##
@@ -0,0 +1,195 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.connectors.hive;
+
+import org.apache.flink.api.common.RuntimeExecutionMode;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.connector.file.table.FileSystemConnectorOptions;
+import org.apache.flink.connector.file.table.batch.BatchSink;
+import org.apache.flink.sql.parser.hive.ddl.SqlCreateHiveTable;
+import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
+import org.apache.flink.streaming.api.graph.StreamGraph;
+import org.apache.flink.streaming.api.graph.StreamNode;
+import org.apache.flink.table.api.DataTypes;
+import org.apache.flink.table.api.TableSchema;
+import org.apache.flink.table.catalog.CatalogDatabaseImpl;
+import org.apache.flink.table.catalog.CatalogTable;
+import org.apache.flink.table.catalog.CatalogTableImpl;
+import org.apache.flink.table.catalog.CatalogTest;
+import org.apache.flink.table.catalog.Column;
+import org.apache.flink.table.catalog.ObjectIdentifier;
+import org.apache.flink.table.catalog.ObjectPath;
+import org.apache.flink.table.catalog.ResolvedCatalogTable;
+import org.apache.flink.table.catalog.ResolvedSchema;
+import org.apache.flink.table.catalog.hive.HiveCatalog;
+import org.apache.flink.table.catalog.hive.HiveTestUtils;
+import org.apache.flink.table.connector.sink.DataStreamSinkProvider;
+import org.apache.flink.table.connector.sink.DynamicTableSink;
+import org.apache.flink.table.data.binary.BinaryRowData;
+import org.apache.flink.table.factories.DynamicTableSinkFactory;
+import org.apache.flink.table.factories.FactoryUtil;
+import 
org.apache.flink.table.runtime.connector.sink.SinkRuntimeProviderContext;
+import org.apache.flink.util.TestLoggerExtension;
+
+import org.junit.jupiter.api.AfterAll;
+import org.junit.jupiter.api.BeforeAll;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.extension.ExtendWith;
+
+import java.util.Collections;
+import java.util.HashMap;
+import java.util.Map;
+import java.util.Optional;
+
+import static org.assertj.core.api.Assertions.assertThat;
+
+/** Tests for {@link HiveTableSink} enable auto-compaction. */
+@ExtendWith(TestLoggerExtension.class)
+class HiveTableCompactSinkTest {
+
+private static HiveCatalog catalog;
+
+private Configuration tableConf;
+
+@BeforeAll
+static void before() {
+catalog = HiveTestUtils.createHiveCatalog();
+catalog.open();
+}
+
+@AfterAll
+static void after() {
+catalog.close();
+}
+
+@BeforeEach
+void resetTableConf() {
+tableConf = new Configuration();
+tableConf.set(FactoryUtil.CONNECTOR, SqlCreateHiveTable.IDENTIFIER);
+tableConf.set(FileSystemConnectorOptions.AUTO_COMPACTION, true);
+}
+
+/** If only sink parallelism is set, compact operator should follow this 
setting. */
+@Test
+void testOnlySetSinkParallelism() throws Exception {
+final int sinkParallelism = 4;
+
+tableConf.set(FileSystemConnectorOptions.SINK_PARALLELISM, 
sinkParallelism);
+
+assertSinkAndCompactOperatorParallelism(
+tableConf, true, true, sinkParallelism, sinkParallelism);
+}
+
+@Test
+void testOnlySetCompactParallelism() throws Exception {
+final int compactParallelism = 4;
+
+tableConf.set(FileSystemConnectorOptions.COMPACTION_PARALLELISM, 
compactParallelism);
+
+assertSinkAndCompactOperatorParallelism(tableConf, false, true, -1, 
compactParallelism);
+}
+
+@Test
+void testSetBothSinkAndCompactParallelism() throws Exception {
+final int sinkParallelism = 8;
+final int compactParallelism = 4;
+
+tableConf.set(FileSystemConnectorOptions.SINK_PARALLELISM, 
sinkParallelism);
+tableConf.set(FileSystemConnectorOptions.COMPACTION_PARALLELISM, 
compactParallelism);
+
+assertSinkAndCompactOperator

[GitHub] [flink] reswqa commented on a diff in pull request #21977: [FLINK-31132] Compact without setting parallelism does not follow the configured sink parallelism for HiveTableSink

2023-02-21 Thread via GitHub


reswqa commented on code in PR #21977:
URL: https://github.com/apache/flink/pull/21977#discussion_r1112739085


##
flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableCompactSinkTest.java:
##
@@ -0,0 +1,195 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.connectors.hive;
+
+import org.apache.flink.api.common.RuntimeExecutionMode;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.connector.file.table.FileSystemConnectorOptions;
+import org.apache.flink.connector.file.table.batch.BatchSink;
+import org.apache.flink.sql.parser.hive.ddl.SqlCreateHiveTable;
+import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
+import org.apache.flink.streaming.api.graph.StreamGraph;
+import org.apache.flink.streaming.api.graph.StreamNode;
+import org.apache.flink.table.api.DataTypes;
+import org.apache.flink.table.api.TableSchema;
+import org.apache.flink.table.catalog.CatalogDatabaseImpl;
+import org.apache.flink.table.catalog.CatalogTable;
+import org.apache.flink.table.catalog.CatalogTableImpl;
+import org.apache.flink.table.catalog.CatalogTest;
+import org.apache.flink.table.catalog.Column;
+import org.apache.flink.table.catalog.ObjectIdentifier;
+import org.apache.flink.table.catalog.ObjectPath;
+import org.apache.flink.table.catalog.ResolvedCatalogTable;
+import org.apache.flink.table.catalog.ResolvedSchema;
+import org.apache.flink.table.catalog.hive.HiveCatalog;
+import org.apache.flink.table.catalog.hive.HiveTestUtils;
+import org.apache.flink.table.connector.sink.DataStreamSinkProvider;
+import org.apache.flink.table.connector.sink.DynamicTableSink;
+import org.apache.flink.table.data.binary.BinaryRowData;
+import org.apache.flink.table.factories.DynamicTableSinkFactory;
+import org.apache.flink.table.factories.FactoryUtil;
+import 
org.apache.flink.table.runtime.connector.sink.SinkRuntimeProviderContext;
+import org.apache.flink.util.TestLoggerExtension;
+
+import org.junit.jupiter.api.AfterAll;
+import org.junit.jupiter.api.BeforeAll;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.extension.ExtendWith;
+
+import java.util.Collections;
+import java.util.HashMap;
+import java.util.Map;
+import java.util.Optional;
+
+import static org.assertj.core.api.Assertions.assertThat;
+
+/** Tests for {@link HiveTableSink} enable auto-compaction. */
+@ExtendWith(TestLoggerExtension.class)
+class HiveTableCompactSinkTest {
+
+private static HiveCatalog catalog;
+
+private Configuration tableConf;
+
+@BeforeAll
+static void before() {
+catalog = HiveTestUtils.createHiveCatalog();
+catalog.open();
+}
+
+@AfterAll
+static void after() {
+catalog.close();
+}
+
+@BeforeEach
+void resetTableConf() {
+tableConf = new Configuration();
+tableConf.set(FactoryUtil.CONNECTOR, SqlCreateHiveTable.IDENTIFIER);
+tableConf.set(FileSystemConnectorOptions.AUTO_COMPACTION, true);
+}
+
+/** If only sink parallelism is set, compact operator should follow this 
setting. */
+@Test
+void testOnlySetSinkParallelism() throws Exception {
+final int sinkParallelism = 4;
+
+tableConf.set(FileSystemConnectorOptions.SINK_PARALLELISM, 
sinkParallelism);
+
+assertSinkAndCompactOperatorParallelism(
+tableConf, true, true, sinkParallelism, sinkParallelism);
+}
+
+@Test
+void testOnlySetCompactParallelism() throws Exception {
+final int compactParallelism = 4;
+
+tableConf.set(FileSystemConnectorOptions.COMPACTION_PARALLELISM, 
compactParallelism);
+
+assertSinkAndCompactOperatorParallelism(tableConf, false, true, -1, 
compactParallelism);
+}
+
+@Test
+void testSetBothSinkAndCompactParallelism() throws Exception {
+final int sinkParallelism = 8;
+final int compactParallelism = 4;
+
+tableConf.set(FileSystemConnectorOptions.SINK_PARALLELISM, 
sinkParallelism);
+tableConf.set(FileSystemConnectorOptions.COMPACTION_PARALLELISM, 
compactParallelism);
+
+assertSinkAndCompactOperator

[GitHub] [flink] reswqa commented on a diff in pull request #21977: [FLINK-31132] Compact without setting parallelism does not follow the configured sink parallelism for HiveTableSink

2023-02-21 Thread via GitHub


reswqa commented on code in PR #21977:
URL: https://github.com/apache/flink/pull/21977#discussion_r1112739085


##
flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableCompactSinkTest.java:
##
@@ -0,0 +1,195 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.connectors.hive;
+
+import org.apache.flink.api.common.RuntimeExecutionMode;
+import org.apache.flink.configuration.Configuration;
+import org.apache.flink.connector.file.table.FileSystemConnectorOptions;
+import org.apache.flink.connector.file.table.batch.BatchSink;
+import org.apache.flink.sql.parser.hive.ddl.SqlCreateHiveTable;
+import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
+import org.apache.flink.streaming.api.graph.StreamGraph;
+import org.apache.flink.streaming.api.graph.StreamNode;
+import org.apache.flink.table.api.DataTypes;
+import org.apache.flink.table.api.TableSchema;
+import org.apache.flink.table.catalog.CatalogDatabaseImpl;
+import org.apache.flink.table.catalog.CatalogTable;
+import org.apache.flink.table.catalog.CatalogTableImpl;
+import org.apache.flink.table.catalog.CatalogTest;
+import org.apache.flink.table.catalog.Column;
+import org.apache.flink.table.catalog.ObjectIdentifier;
+import org.apache.flink.table.catalog.ObjectPath;
+import org.apache.flink.table.catalog.ResolvedCatalogTable;
+import org.apache.flink.table.catalog.ResolvedSchema;
+import org.apache.flink.table.catalog.hive.HiveCatalog;
+import org.apache.flink.table.catalog.hive.HiveTestUtils;
+import org.apache.flink.table.connector.sink.DataStreamSinkProvider;
+import org.apache.flink.table.connector.sink.DynamicTableSink;
+import org.apache.flink.table.data.binary.BinaryRowData;
+import org.apache.flink.table.factories.DynamicTableSinkFactory;
+import org.apache.flink.table.factories.FactoryUtil;
+import 
org.apache.flink.table.runtime.connector.sink.SinkRuntimeProviderContext;
+import org.apache.flink.util.TestLoggerExtension;
+
+import org.junit.jupiter.api.AfterAll;
+import org.junit.jupiter.api.BeforeAll;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.extension.ExtendWith;
+
+import java.util.Collections;
+import java.util.HashMap;
+import java.util.Map;
+import java.util.Optional;
+
+import static org.assertj.core.api.Assertions.assertThat;
+
+/** Tests for {@link HiveTableSink} enable auto-compaction. */
+@ExtendWith(TestLoggerExtension.class)
+class HiveTableCompactSinkTest {
+
+private static HiveCatalog catalog;
+
+private Configuration tableConf;
+
+@BeforeAll
+static void before() {
+catalog = HiveTestUtils.createHiveCatalog();
+catalog.open();
+}
+
+@AfterAll
+static void after() {
+catalog.close();
+}
+
+@BeforeEach
+void resetTableConf() {
+tableConf = new Configuration();
+tableConf.set(FactoryUtil.CONNECTOR, SqlCreateHiveTable.IDENTIFIER);
+tableConf.set(FileSystemConnectorOptions.AUTO_COMPACTION, true);
+}
+
+/** If only sink parallelism is set, compact operator should follow this 
setting. */
+@Test
+void testOnlySetSinkParallelism() throws Exception {
+final int sinkParallelism = 4;
+
+tableConf.set(FileSystemConnectorOptions.SINK_PARALLELISM, 
sinkParallelism);
+
+assertSinkAndCompactOperatorParallelism(
+tableConf, true, true, sinkParallelism, sinkParallelism);
+}
+
+@Test
+void testOnlySetCompactParallelism() throws Exception {
+final int compactParallelism = 4;
+
+tableConf.set(FileSystemConnectorOptions.COMPACTION_PARALLELISM, 
compactParallelism);
+
+assertSinkAndCompactOperatorParallelism(tableConf, false, true, -1, 
compactParallelism);
+}
+
+@Test
+void testSetBothSinkAndCompactParallelism() throws Exception {
+final int sinkParallelism = 8;
+final int compactParallelism = 4;
+
+tableConf.set(FileSystemConnectorOptions.SINK_PARALLELISM, 
sinkParallelism);
+tableConf.set(FileSystemConnectorOptions.COMPACTION_PARALLELISM, 
compactParallelism);
+
+assertSinkAndCompactOperator

[GitHub] [flink] reswqa commented on pull request #21977: [FLINK-31132] Compact without setting parallelism does not follow the configured sink parallelism for HiveTableSink

2023-02-21 Thread via GitHub


reswqa commented on PR #21977:
URL: https://github.com/apache/flink/pull/21977#issuecomment-1438069932

   Thanks @luoyuxia and @JunRuiLee for the review, I have updated this pr in 
two fixup commits, PTAL agian~


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691462#comment-17691462
 ] 

Matthias Pohl commented on FLINK-31153:
---

{quote}
BTW, you mentioned in the description: you need to update Flink version in 
"two" repositories followed with three repos. Did you mean "three"?
 {quote}
Thanks for the hint. The content in the description is coming from the [Flink 
docs on how to create a 
release|https://cwiki.apache.org/confluence/display/FLINK/Flink+Release+Management].
 I copied it and tried to align it but must have missed that bit. (y)

{quote}
It seems those tasks contain more or different information. Are you planning to 
build a template for the future release?
{quote}
We're experimenting a bit with improving the documentation. One idea is to move 
the documentation into Jira to enable release managers to document their 
efforts (see [the release sync protocol from February 14, 2023 
week|https://cwiki.apache.org/confluence/display/FLINK/1.17+Release#id-1.17Release-2023-02-14]).
 For now, this is meant as a test. If it works well, we will create an improved 
Jira issue collection which can serve as templates. AFAIU, the Apache Jira 
doesn't provide templates, though. But having a dedicated set of FLINK Jira 
issues that can be cloned might be good enough.

> Create a release branch
> ---
>
> Key: FLINK-31153
> URL: https://issues.apache.org/jira/browse/FLINK-31153
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Priority: Major
>
> If you are doing a new major release, you need to update Flink version in the 
> following repositories:
>  * [apache/flink|https://github.com/apache/flink]
>  * [apache/flink-docker|https://github.com/apache/flink-docker]
>  * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]
> Minor releases don't require the these repositories to be touched. Simply 
> checkout the already existing branch for that version:
> {code:java}
> $ cd ./tools
> $ git checkout release-$SHORT_RELEASE_VERSION
> {code}
> h4. Flink repository
> Create a branch for the new version that we want to release before updating 
> the master branch to the next development version:
> {code:bash}
> $ cd ./tools
> $ releasing/create_snapshot_branch.sh
> $ git checkout master
> $ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
> releasing/update_branch_version.sh
> {code}
> In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  as the last entry:
> {code:java}
> // ...
> v1_12("1.12"),
> v1_13("1.13"),
> v1_14("1.14"),
> v1_15("1.15"),
> v1_16("1.16");
> {code}
> The newly created branch and updated {{master}} branch need to be pushed to 
> the official repository.
> h4. Flink Docker Repository
> Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
> [apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
> sure that 
> [apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
>  points to the correct snapshot version; for {{dev-x.y}} it should point to 
> {{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
> recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).
> After pushing the new major release branch, as the last step you should also 
> update the documentation workflow to also build the documentation for the new 
> release branch. Check [Managing 
> Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
>  on details on how to do that. You may also want to manually trigger a build 
> to make the changes visible as soon as possible.
>  
> 
> h3. Expectations (Minor Version only)
>  * Release branch has been created and pushed
>  * Cron job has been added on the release branch in 
> ([apache-flink:./tools/azure-pipelines/build-apache-repo.yml|https://github.com/apache/flink/blob/master/tools/azure-pipelines/build-apache-repo.yml])
>  * Originating branch has the version information updated to the new version
>  * New version is added to the 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  enum.
>  * Make sure {{flink-docker}} has {{dev-x.y}} branch and docker e2e tests run 
> against this branch
>  * docs/config.toml has been updated appropriately.
>  * The {{dev-x.y}} branch of ({{{}$CURRENT_SNAPSHOT_VERSION{}}}) have been 
> created in the Flink Benchmark repo.
>  * The {{flink.version}} property of Fl

[GitHub] [flink] flinkbot commented on pull request #21981: [FLINK-21450][runtime] Support LocalRecovery by AdaptiveScheduler

2023-02-21 Thread via GitHub


flinkbot commented on PR #21981:
URL: https://github.com/apache/flink/pull/21981#issuecomment-1438071101

   
   ## CI report:
   
   * 90f669d3099333992edcd961682b0cde4fa06dc8 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] rkhachatryan commented on a diff in pull request #21981: [FLINK-21450][runtime] Support LocalRecovery by AdaptiveScheduler

2023-02-21 Thread via GitHub


rkhachatryan commented on code in PR #21981:
URL: https://github.com/apache/flink/pull/21981#discussion_r1112741455


##
flink-runtime/src/main/java/org/apache/flink/runtime/scheduler/adaptive/allocator/SlotSharingSlotAllocator.java:
##
@@ -121,16 +133,13 @@ public Optional 
determineParallelism(
 slotSharingGroupParallelism.get(
 slotSharingGroup.getSlotSharingGroupId()));
 
-final Iterable 
sharedSlotToVertexAssignment =
+final List sharedSlotToVertexAssignment 
=
 createExecutionSlotSharingGroups(vertexParallelism);
 
-for (ExecutionSlotSharingGroup executionSlotSharingGroup :
-sharedSlotToVertexAssignment) {
-final SlotInfo slotInfo = slotIterator.next();
-
-assignments.add(
-new 
ExecutionSlotSharingGroupAndSlot(executionSlotSharingGroup, slotInfo));
-}
+SlotAssigner.AssignmentResult result =
+slotAssigner.assignSlots(freeSlots, 
sharedSlotToVertexAssignment);
+assignments.addAll(result.assignments);
+freeSlots = result.remainingSlots;

Review Comment:
   Because `assignSlots` is called in a loop, and internally iterates over the 
`freeSlots`, the complexity here is suboptimal. Lower-priority slots are added 
over and over into the queue inside `assignSlots`, resulting in quadratic 
complexity.
   
   I was thinking about making free slots part of `SlotAssigner` state (passed 
into constructor).
   
   WDYT @dmvk ?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] pnowojski commented on a diff in pull request #21923: FLINK-13871: Consolidate volatile status fields in StreamTask

2023-02-21 Thread via GitHub


pnowojski commented on code in PR #21923:
URL: https://github.com/apache/flink/pull/21923#discussion_r1112741636


##
flink-streaming-java/src/main/java/org/apache/flink/streaming/runtime/tasks/StreamTask.java:
##
@@ -949,7 +952,9 @@ public final void cleanUp(Throwable throwable) throws 
Exception {
 // disabled the interruptions or not.
 getCompletionFuture().exceptionally(unused -> null).join();
 // clean up everything we initialized
-isRunning = false;
+if (!isCanceled() && !isFailing()) {

Review Comment:
   > On a side note, those volatile accesses aren't on the hot path so 
performance-wise it shouldn't change much
   
   Yes, this doesn't affect performance. This ticket is about cleaning up the 
code and improving readability.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Comment Edited] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691462#comment-17691462
 ] 

Matthias Pohl edited comment on FLINK-31153 at 2/21/23 8:48 AM:


{quote}
BTW, you mentioned in the description: you need to update Flink version in 
"two" repositories followed with three repos. Did you mean "three"?
 {quote}
Thanks for the hint. I updated the description accordingly. The content in the 
description is coming from the [Flink docs on how to create a 
release|https://cwiki.apache.org/confluence/display/FLINK/Flink+Release+Management].
 I copied it and tried to align it but must have missed that bit. (y)

{quote}
It seems those tasks contain more or different information. Are you planning to 
build a template for the future release?
{quote}
We're experimenting a bit with improving the documentation. One idea is to move 
the documentation into Jira to enable release managers to document their 
efforts (see [the release sync protocol from February 14, 2023 
week|https://cwiki.apache.org/confluence/display/FLINK/1.17+Release#id-1.17Release-2023-02-14]).
 For now, this is meant as a test. If it works well, we will create an improved 
Jira issue collection which can serve as templates. AFAIU, the Apache Jira 
doesn't provide templates, though. But having a dedicated set of FLINK Jira 
issues that can be cloned might be good enough.


was (Author: mapohl):
{quote}
BTW, you mentioned in the description: you need to update Flink version in 
"two" repositories followed with three repos. Did you mean "three"?
 {quote}
Thanks for the hint. The content in the description is coming from the [Flink 
docs on how to create a 
release|https://cwiki.apache.org/confluence/display/FLINK/Flink+Release+Management].
 I copied it and tried to align it but must have missed that bit. (y)

{quote}
It seems those tasks contain more or different information. Are you planning to 
build a template for the future release?
{quote}
We're experimenting a bit with improving the documentation. One idea is to move 
the documentation into Jira to enable release managers to document their 
efforts (see [the release sync protocol from February 14, 2023 
week|https://cwiki.apache.org/confluence/display/FLINK/1.17+Release#id-1.17Release-2023-02-14]).
 For now, this is meant as a test. If it works well, we will create an improved 
Jira issue collection which can serve as templates. AFAIU, the Apache Jira 
doesn't provide templates, though. But having a dedicated set of FLINK Jira 
issues that can be cloned might be good enough.

> Create a release branch
> ---
>
> Key: FLINK-31153
> URL: https://issues.apache.org/jira/browse/FLINK-31153
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Priority: Major
>
> If you are doing a new major release, you need to update Flink version in the 
> following repositories:
>  * [apache/flink|https://github.com/apache/flink]
>  * [apache/flink-docker|https://github.com/apache/flink-docker]
>  * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]
> Minor releases don't require the these repositories to be touched. Simply 
> checkout the already existing branch for that version:
> {code:java}
> $ cd ./tools
> $ git checkout release-$SHORT_RELEASE_VERSION
> {code}
> h4. Flink repository
> Create a branch for the new version that we want to release before updating 
> the master branch to the next development version:
> {code:bash}
> $ cd ./tools
> $ releasing/create_snapshot_branch.sh
> $ git checkout master
> $ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
> releasing/update_branch_version.sh
> {code}
> In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  as the last entry:
> {code:java}
> // ...
> v1_12("1.12"),
> v1_13("1.13"),
> v1_14("1.14"),
> v1_15("1.15"),
> v1_16("1.16");
> {code}
> The newly created branch and updated {{master}} branch need to be pushed to 
> the official repository.
> h4. Flink Docker Repository
> Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
> [apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
> sure that 
> [apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
>  points to the correct snapshot version; for {{dev-x.y}} it should point to 
> {{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
> recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).
> After pushing the new major release branch, as the last step you should also 
> 

[GitHub] [flink] MartijnVisser merged pull request #21964: [FLINK-30948][Formats/AWS] Remove GlueSchemaRegistry Avro and JSON fo…

2023-02-21 Thread via GitHub


MartijnVisser merged PR #21964:
URL: https://github.com/apache/flink/pull/21964


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-30948) Remove flink-avro-glue-schema-registry and flink-json-glue-schema-registry from Flink main repo

2023-02-21 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30948?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser updated FLINK-30948:
---
Fix Version/s: 1.17.0

> Remove flink-avro-glue-schema-registry and flink-json-glue-schema-registry 
> from Flink main repo
> ---
>
> Key: FLINK-30948
> URL: https://issues.apache.org/jira/browse/FLINK-30948
> Project: Flink
>  Issue Type: Sub-task
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
> Environment: Remove flink-avro-glue-schema-registry and 
> flink-json-glue-schema-registry from Flink main repo, along with associated 
> end-to-end tests
>Reporter: Hong Liang Teoh
>Assignee: Hong Liang Teoh
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31153:
--
Description: 
If you are doing a new minor release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Patch releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
releasing/update_branch_version.sh
{code}
In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 as the last entry:
{code:java}
// ...
v1_12("1.12"),
v1_13("1.13"),
v1_14("1.14"),
v1_15("1.15"),
v1_16("1.16");
{code}
The newly created branch and updated {{master}} branch need to be pushed to the 
official repository.
h4. Flink Docker Repository

Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
[apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
sure that 
[apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
 points to the correct snapshot version; for {{dev-x.y}} it should point to 
{{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).

After pushing the new minor release branch, as the last step you should also 
update the documentation workflow to also build the documentation for the new 
release branch. Check [Managing 
Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
 on details on how to do that. You may also want to manually trigger a build to 
make the changes visible as soon as possible.

 

h3. Expectations (Minor Version only)
 * Release branch has been created and pushed
 * Cron job has been added on the release branch in 
([apache-flink:./tools/azure-pipelines/build-apache-repo.yml|https://github.com/apache/flink/blob/master/tools/azure-pipelines/build-apache-repo.yml])
 * Originating branch has the version information updated to the new version
 * New version is added to the 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum.
 * Make sure {{flink-docker}} has {{dev-x.y}} branch and docker e2e tests run 
against this branch
 * docs/config.toml has been updated appropriately.
 * The {{dev-x.y}} branch of ({{{}$CURRENT_SNAPSHOT_VERSION{}}}) have been 
created in the Flink Benchmark repo.
 * The {{flink.version}} property of Flink Benchmark repo has been updated to 
the latest snapshot version.

  was:
If you are doing a new major release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Minor releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
releasing/update_branch_version.sh
{code}
In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 as the last entry:
{code:java}
// ...
v1_12("1.12"),
v1_13("1.13"),
v1_14("1.14"),
v1_15("1.15"),
v1_16("1.16");
{code}
The newly created branch and updated {{master}} branch need to be pushed to the 
official repository.
h4. Flink Docker Repository

Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
[apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
sure that 
[apache/fl

[jira] [Commented] (FLINK-31059) Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by native implementation

2023-02-21 Thread miamiaoxyz (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31059?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691464#comment-17691464
 ] 

miamiaoxyz commented on FLINK-31059:


I first use sql client to test the feature, but the 
`exec.hive.native-agg-function.enabled`  do not work. 


I then use ITCase to verify.
 #    I use IT case to turn on and off `exec.hive.native-agg-function.enabled` 
to verify the two results are the same, testSql function test whether the same 
sql get same result with on and off `exec.hive.native-agg-function.enabled`.

I verfied that the plan use Hashagg when turn on the 
`exec.hive.native-agg-function.enabled`, and the plan use SortAgg when turn off 
by IT case.
!image-2023-02-21-15-45-48-226.png|width=549,height=234!
It pass all the IT Case below. 
!image-2023-02-21-15-46-13-966.png|width=501,height=371!
2. I verified that data results are the same when combine sum/count/avg/min/max 
functions in query using `exec.hive.native-agg-function.enabled` on and off 
using the IT case below.

I verfied that the plan use Hashagg when turn on the 
`exec.hive.native-agg-function.enabled`, and the plan use SortAgg when turn off 
by IT case.

!image-2023-02-21-15-49-58-854.png|width=536,height=219!

3. For  `array` and `struct` do not support the max function.  For count 
function, it does not store `array` or `struct` in agg, so they use bigint 
instead, and hash-agg is chosen  .

!image-2023-02-21-15-59-44-470.png|width=1016,height=189!

4. For `first_value` and `last_value` are not implemented in hive,  
[https://cwiki.apache.org/confluence/display/Hive/LanguageManual+UDF#LanguageManualUDF-Built-inAggregateFunctions(UDAF)]
 I use `collect_set` to test  instead. All the plan use SortAgg, and get same 
result, which meet the expectations.

```

---  turn on `table.exec.hive.native-agg-function.enabled`

== Abstract Syntax Tree ==
LogicalProject(x=[$0], _o__c1=[$1])
+- LogicalAggregate(group=[\{0}], agg#0=[collect_set($1)])
   +- LogicalProject($f0=[$0], $f1=[$1])
      +- LogicalTableScan(table=[[test-catalog, default, foo]])

== Optimized Physical Plan ==
SortAggregate(isMerge=[true], groupBy=[x], select=[x, Final_collect_set($f1) AS 
$f1])
+- Sort(orderBy=[x ASC])
   +- Exchange(distribution=[hash[x]])
      +- LocalSortAggregate(groupBy=[x], select=[x, Partial_collect_set(y) AS 
$f1])
         +- Sort(orderBy=[x ASC])
            +- TableSourceScan(table=[[test-catalog, default, foo]], fields=[x, 
y])

== Optimized Execution Plan ==
SortAggregate(isMerge=[true], groupBy=[x], select=[x, Final_collect_set($f1) AS 
$f1])
+- Exchange(distribution=[forward])
   +- Sort(orderBy=[x ASC])
      +- Exchange(distribution=[hash[x]])
         +- LocalSortAggregate(groupBy=[x], select=[x, Partial_collect_set(y) 
AS $f1])
            +- Exchange(distribution=[forward])
               +- Sort(orderBy=[x ASC])
                  +- TableSourceScan(table=[[test-catalog, default, foo]], 
fields=[x, y])

---  turn off `table.exec.hive.native-agg-function.enabled`

== Abstract Syntax Tree ==
LogicalProject(x=[$0], _o__c1=[$1])
+- LogicalAggregate(group=[\{0}], agg#0=[collect_set($1)])
   +- LogicalProject($f0=[$0], $f1=[$1])
      +- LogicalTableScan(table=[[test-catalog, default, foo]])

== Optimized Physical Plan ==
SortAggregate(isMerge=[true], groupBy=[x], select=[x, Final_collect_set($f1) AS 
$f1])
+- Sort(orderBy=[x ASC])
   +- Exchange(distribution=[hash[x]])
      +- LocalSortAggregate(groupBy=[x], select=[x, Partial_collect_set(y) AS 
$f1])
         +- Sort(orderBy=[x ASC])
            +- TableSourceScan(table=[[test-catalog, default, foo]], fields=[x, 
y])

== Optimized Execution Plan ==
SortAggregate(isMerge=[true], groupBy=[x], select=[x, Final_collect_set($f1) AS 
$f1])
+- Exchange(distribution=[forward])
   +- Sort(orderBy=[x ASC])
      +- Exchange(distribution=[hash[x]])
         +- LocalSortAggregate(groupBy=[x], select=[x, Partial_collect_set(y) 
AS $f1])
            +- Exchange(distribution=[forward])
               +- Sort(orderBy=[x ASC])
                  +- TableSourceScan(table=[[test-catalog, default, foo]], 
fields=[x, y])

```

 

 

!image-2023-02-21-16-31-58-361.png|width=620,height=261!

 

5. I disable the hashagg to force use sortagg to process all of the test above, 
which  can see that the result of forcing to close hashagg is the same as the 
result of turn on and off`exec.hive.native-agg-function.enabled`, which meets 
the expectations

!image-2023-02-21-16-35-46-294.png|width=632,height=392!

 

Problems:

a.  The `exec.hive.native-agg-function.enabled`  do not work on sql client. the 
hashagg is not chosen on sql client.

!https://intranetproxy.alipay.com/skylark/lark/0/2023/png/83756403/1676952029939-182fa078-3a07-4e45-bdbb-832f7f74c838.png|width=703,height=383,id=u4fc84338!

b. Enable and disable `table.exec.hive.native-agg-function.enabled` get 
different result.

!image-2023-02-21-16-28-22-

[GitHub] [flink] FangYongs commented on a diff in pull request #21901: [FLINK-30968][sql-client] Sql client supports dynamic config to open session

2023-02-21 Thread via GitHub


FangYongs commented on code in PR #21901:
URL: https://github.com/apache/flink/pull/21901#discussion_r1112748189


##
flink-table/flink-sql-client/src/main/java/org/apache/flink/table/client/gateway/DefaultContextUtils.java:
##
@@ -51,12 +52,17 @@ public static DefaultContext 
buildDefaultContext(CliOptions.EmbeddedCliOptions o
 } else {
 libDirs = Collections.emptyList();
 }
-return DefaultContext.load(
-options.getPythonConfiguration(), discoverDependencies(jars, 
libDirs), true);
+Configuration pythonConfiguration = options.getPythonConfiguration();
+pythonConfiguration.addAll(
+
ConfigurationUtils.createConfiguration(options.getSessionConfig()));
+return DefaultContext.load(pythonConfiguration, 
discoverDependencies(jars, libDirs), true);

Review Comment:
   LGTM, DONE



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] MartijnVisser merged pull request #21969: [FLINK-30948][Formats/AWS] Remove GlueSchemaRegistry Avro and JSON fo…

2023-02-21 Thread via GitHub


MartijnVisser merged PR #21969:
URL: https://github.com/apache/flink/pull/21969


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Assigned] (FLINK-31150) Cross team testing

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl reassigned FLINK-31150:
-

Assignee: Qingsheng Ren

> Cross team testing
> --
>
> Key: FLINK-31150
> URL: https://issues.apache.org/jira/browse/FLINK-31150
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Assignee: Qingsheng Ren
>Priority: Major
>
> For user facing features that go into the release we'd like to ensure they 
> can actually _be used_ by Flink users. To achieve this the release managers 
> ensure that an issue for cross team testing is created in the Apache Flink 
> Jira. This can and should be picked up by other community members to verify 
> the functionality and usability of the feature.
> The issue should contain some entry points which enables other community 
> members to test it. It should not contain documentation on how to use the 
> feature as this should be part of the actual documentation. The cross team 
> tests are performed after the feature freeze. Documentation should be in 
> place before that. Those tests are manual tests, so do not confuse them with 
> automated tests.
> To sum that up:
>  * User facing features should be tested by other contributors
>  * The scope is usability and sanity of the feature
>  * The feature needs to be already documented
>  * The contributor creates an issue containing some pointers on how to get 
> started (e.g. link to the documentation, suggested targets of verification)
>  * Other community members pick those issues up and provide feedback
>  * Cross team testing happens right after the feature freeze



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-31165) Over Agg: The window rank function without order by error in top N query

2023-02-21 Thread P Rohan Kumar (Jira)
P Rohan Kumar created FLINK-31165:
-

 Summary: Over Agg: The window rank function without order by error 
in top N query
 Key: FLINK-31165
 URL: https://issues.apache.org/jira/browse/FLINK-31165
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / API
Affects Versions: 1.16.0
Reporter: P Rohan Kumar


 
{code:java}
val env: StreamExecutionEnvironment = 
StreamExecutionEnvironment.getExecutionEnvironment

val tableEnv = StreamTableEnvironment.create(env)


val td = TableDescriptor.forConnector("datagen").option("rows-per-second", "10")
  .option("number-of-rows", "10")
  .schema(Schema
.newBuilder()
.column("NAME", DataTypes.VARCHAR(2147483647))
.column("ROLLNO", DataTypes.DECIMAL(5, 0))
.column("DOB", DataTypes.DATE())
.column("CLASS", DataTypes.DECIMAL(2, 0))
.column("SUBJECT", DataTypes.VARCHAR(2147483647))
.build())
  .build()

val table = tableEnv.from(td)


tableEnv.createTemporaryView("temp_table", table)

val newTable = tableEnv.sqlQuery("select temp_table.*,cast('2022-01-01' as 
date) SRC_NO from temp_table")

tableEnv.createTemporaryView("temp_table2", newTable)


val newTable2 = tableEnv.sqlQuery("select * from (select 
NAME,ROLLNO,row_number() over (partition by NAME ORDER BY SRC_NO) AS rownum  
from temp_table2 a) where rownum <= 1")

tableEnv.toChangelogStream(newTable2).print()

env.execute()
 {code}
 

 

I am getting the below error if I run the above code.

I have already provided an order by column.

If I change the order by column to some other column, such as "SUBJECT", then 
the job runs fine.

 

 
{code:java}
Exception in thread "main" java.lang.RuntimeException: Error while applying 
rule FlinkLogicalOverAggregateConverter(in:NONE,out:LOGICAL), args 
[rel#245:LogicalWindow.NONE.any.None: 
0.[NONE].[NONE](input=RelSubset#244,window#0=window(partition {0} rows between 
UNBOUNDED PRECEDING and CURRENT ROW aggs [ROW_NUMBER()]))]
    at 
org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:256)
    at 
org.apache.calcite.plan.volcano.IterativeRuleDriver.drive(IterativeRuleDriver.java:58)
    at 
org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:510)
    at org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:312)
    at 
org.apache.flink.table.planner.plan.optimize.program.FlinkVolcanoProgram.optimize(FlinkVolcanoProgram.scala:62)
    at 
org.apache.flink.table.planner.plan.optimize.program.FlinkChainedProgram.$anonfun$optimize$1(FlinkChainedProgram.scala:59)
    at 
scala.collection.TraversableOnce$folder$1$.apply(TraversableOnce.scala:187)
    at 
scala.collection.TraversableOnce$folder$1$.apply(TraversableOnce.scala:185)
    at scala.collection.Iterator.foreach(Iterator.scala:943)
    at scala.collection.Iterator.foreach$(Iterator.scala:943)
    at scala.collection.AbstractIterator.foreach(Iterator.scala:1431)
    at scala.collection.IterableLike.foreach(IterableLike.scala:74)
    at scala.collection.IterableLike.foreach$(IterableLike.scala:73)
    at scala.collection.AbstractIterable.foreach(Iterable.scala:56)
    at scala.collection.TraversableOnce.foldLeft(TraversableOnce.scala:189)
    at scala.collection.TraversableOnce.foldLeft$(TraversableOnce.scala:184)
    at scala.collection.AbstractTraversable.foldLeft(Traversable.scala:108)
    at 
org.apache.flink.table.planner.plan.optimize.program.FlinkChainedProgram.optimize(FlinkChainedProgram.scala:55)
    at 
org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.optimizeTree(StreamCommonSubGraphBasedOptimizer.scala:176)
    at 
org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.doOptimize(StreamCommonSubGraphBasedOptimizer.scala:83)
    at 
org.apache.flink.table.planner.plan.optimize.CommonSubGraphBasedOptimizer.optimize(CommonSubGraphBasedOptimizer.scala:87)
    at 
org.apache.flink.table.planner.delegation.PlannerBase.optimize(PlannerBase.scala:315)
    at 
org.apache.flink.table.planner.delegation.PlannerBase.translate(PlannerBase.scala:195)
    at 
org.apache.flink.table.api.bridge.internal.AbstractStreamTableEnvironmentImpl.toStreamInternal(AbstractStreamTableEnvironmentImpl.java:224)
    at 
org.apache.flink.table.api.bridge.internal.AbstractStreamTableEnvironmentImpl.toStreamInternal(AbstractStreamTableEnvironmentImpl.java:219)
    at 
org.apache.flink.table.api.bridge.scala.internal.StreamTableEnvironmentImpl.toChangelogStream(StreamTableEnvironmentImpl.scala:160)
    at org.example.OverAggregateBug$.main(OverAggregateBug.scala:39)
    at org.example.OverAggregateBug.main(OverAggregateBug.scala)
Caused by: org.apache.flink.table.api.ValidationException: Over Agg: The window 
rank function without order by. please re-check the over window statement.
    at 
org.apache.flink.table.planner.plan.nodes.logical.FlinkLogicalOverAggregateConverter.$anonfu

[jira] [Assigned] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl reassigned FLINK-31153:
-

Assignee: Leonard Xu

> Create a release branch
> ---
>
> Key: FLINK-31153
> URL: https://issues.apache.org/jira/browse/FLINK-31153
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Assignee: Leonard Xu
>Priority: Major
>
> If you are doing a new minor release, you need to update Flink version in the 
> following repositories:
>  * [apache/flink|https://github.com/apache/flink]
>  * [apache/flink-docker|https://github.com/apache/flink-docker]
>  * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]
> Patch releases don't require the these repositories to be touched. Simply 
> checkout the already existing branch for that version:
> {code:java}
> $ cd ./tools
> $ git checkout release-$SHORT_RELEASE_VERSION
> {code}
> h4. Flink repository
> Create a branch for the new version that we want to release before updating 
> the master branch to the next development version:
> {code:bash}
> $ cd ./tools
> $ releasing/create_snapshot_branch.sh
> $ git checkout master
> $ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
> releasing/update_branch_version.sh
> {code}
> In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  as the last entry:
> {code:java}
> // ...
> v1_12("1.12"),
> v1_13("1.13"),
> v1_14("1.14"),
> v1_15("1.15"),
> v1_16("1.16");
> {code}
> The newly created branch and updated {{master}} branch need to be pushed to 
> the official repository.
> h4. Flink Docker Repository
> Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
> [apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
> sure that 
> [apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
>  points to the correct snapshot version; for {{dev-x.y}} it should point to 
> {{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
> recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).
> After pushing the new minor release branch, as the last step you should also 
> update the documentation workflow to also build the documentation for the new 
> release branch. Check [Managing 
> Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
>  on details on how to do that. You may also want to manually trigger a build 
> to make the changes visible as soon as possible.
>  
> 
> h3. Expectations (Minor Version only)
>  * Release branch has been created and pushed
>  * Cron job has been added on the release branch in 
> ([apache-flink:./tools/azure-pipelines/build-apache-repo.yml|https://github.com/apache/flink/blob/master/tools/azure-pipelines/build-apache-repo.yml])
>  * Originating branch has the version information updated to the new version
>  * New version is added to the 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  enum.
>  * Make sure {{flink-docker}} has {{dev-x.y}} branch and docker e2e tests run 
> against this branch
>  * docs/config.toml has been updated appropriately.
>  * The {{dev-x.y}} branch of ({{{}$CURRENT_SNAPSHOT_VERSION{}}}) have been 
> created in the Flink Benchmark repo.
>  * The {{flink.version}} property of Flink Benchmark repo has been updated to 
> the latest snapshot version.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-31150) Cross team testing

2023-02-21 Thread Matthias Pohl (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31150?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691465#comment-17691465
 ] 

Matthias Pohl commented on FLINK-31150:
---

[~renqs] created FLINK-30926 as an umbrella ticket for the 1.17 release-testing 
efforts.

> Cross team testing
> --
>
> Key: FLINK-31150
> URL: https://issues.apache.org/jira/browse/FLINK-31150
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Assignee: Qingsheng Ren
>Priority: Major
>
> For user facing features that go into the release we'd like to ensure they 
> can actually _be used_ by Flink users. To achieve this the release managers 
> ensure that an issue for cross team testing is created in the Apache Flink 
> Jira. This can and should be picked up by other community members to verify 
> the functionality and usability of the feature.
> The issue should contain some entry points which enables other community 
> members to test it. It should not contain documentation on how to use the 
> feature as this should be part of the actual documentation. The cross team 
> tests are performed after the feature freeze. Documentation should be in 
> place before that. Those tests are manual tests, so do not confuse them with 
> automated tests.
> To sum that up:
>  * User facing features should be tested by other contributors
>  * The scope is usability and sanity of the feature
>  * The feature needs to be already documented
>  * The contributor creates an issue containing some pointers on how to get 
> started (e.g. link to the documentation, suggested targets of verification)
>  * Other community members pick those issues up and provide feedback
>  * Cross team testing happens right after the feature freeze



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-30948) Remove flink-avro-glue-schema-registry and flink-json-glue-schema-registry from Flink main repo

2023-02-21 Thread Martijn Visser (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-30948?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691466#comment-17691466
 ] 

Martijn Visser commented on FLINK-30948:


Fixed in:

master: 29f009b7e8c714cd5af0626e9725eb8538a4bd0f
release-1.17: d58829335557dac6ce428df5a80d4244fccf4491

> Remove flink-avro-glue-schema-registry and flink-json-glue-schema-registry 
> from Flink main repo
> ---
>
> Key: FLINK-30948
> URL: https://issues.apache.org/jira/browse/FLINK-30948
> Project: Flink
>  Issue Type: Sub-task
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
> Environment: Remove flink-avro-glue-schema-registry and 
> flink-json-glue-schema-registry from Flink main repo, along with associated 
> end-to-end tests
>Reporter: Hong Liang Teoh
>Assignee: Hong Liang Teoh
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-30948) Remove flink-avro-glue-schema-registry and flink-json-glue-schema-registry from Flink main repo

2023-02-21 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30948?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser closed FLINK-30948.
--
Resolution: Fixed

> Remove flink-avro-glue-schema-registry and flink-json-glue-schema-registry 
> from Flink main repo
> ---
>
> Key: FLINK-30948
> URL: https://issues.apache.org/jira/browse/FLINK-30948
> Project: Flink
>  Issue Type: Sub-task
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
> Environment: Remove flink-avro-glue-schema-registry and 
> flink-json-glue-schema-registry from Flink main repo, along with associated 
> end-to-end tests
>Reporter: Hong Liang Teoh
>Assignee: Hong Liang Teoh
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.17.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] dmvk commented on a diff in pull request #21981: [FLINK-21450][runtime] Support LocalRecovery by AdaptiveScheduler

2023-02-21 Thread via GitHub


dmvk commented on code in PR #21981:
URL: https://github.com/apache/flink/pull/21981#discussion_r1112754334


##
flink-runtime/src/main/java/org/apache/flink/runtime/scheduler/adaptive/AdaptiveScheduler.java:
##
@@ -794,7 +802,23 @@ public void goToWaitingForResources() {
 LOG,
 desiredResources,
 this.initialResourceAllocationTimeout,
-this.resourceStabilizationTimeout));
+this.resourceStabilizationTimeout,
+null));
+}
+
+@Override
+public void goToWaitingForResources(ExecutionGraph executionGraph) {

Review Comment:
   I'm wondering, do we need an extra `goToWaitingForResources` method? 🤔 
   
   Can we do something along these lines instead?
   
   ```
   final ExecutionGraph previousExecutionGraph = 
state.as(StateWithExecutionGraph.class)
   .map(StateWithExecutionGraph::getExecutionGraph)
   .orElse(null);
   ```



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31153:
--
Description: 
If you are doing a new minor release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Patch releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
releasing/update_branch_version.sh
{code}
In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 as the last entry:
{code:java}
// ...
v1_12("1.12"),
v1_13("1.13"),
v1_14("1.14"),
v1_15("1.15"),
v1_16("1.16");
{code}
The newly created branch and updated {{master}} branch need to be pushed to the 
official repository.
h4. Flink Docker Repository

Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
[apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
sure that 
[apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
 points to the correct snapshot version; for {{dev-x.y}} it should point to 
{{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).

After pushing the new minor release branch, as the last step you should also 
update the documentation workflow to also build the documentation for the new 
release branch. Check [Managing 
Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
 on details on how to do that. You may also want to manually trigger a build to 
make the changes visible as soon as possible.

 

h3. Expectations (Minor Version only)
 * Release branch has been created and pushed
 * Cron job has been added on the release branch in 
([apache-flink:./tools/azure-pipelines/build-apache-repo.yml|https://github.com/apache/flink/blob/master/tools/azure-pipelines/build-apache-repo.yml])
 * Originating branch has the version information updated to the new version
 * New version is added to the 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum.
 * Make sure [flink-docker|https://github.com/apache/flink-docker/] has 
{{dev-x.y}} branch and docker e2e tests run against this branch
 * docs/config.toml has been updated appropriately.
 * The {{dev-x.y}} branch of ({{{}$CURRENT_SNAPSHOT_VERSION{}}}) have been 
created in the Flink Benchmark repo.
 * The {{flink.version}} property of Flink Benchmark repo has been updated to 
the latest snapshot version.

  was:
If you are doing a new minor release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Patch releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
releasing/update_branch_version.sh
{code}
In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 as the last entry:
{code:java}
// ...
v1_12("1.12"),
v1_13("1.13"),
v1_14("1.14"),
v1_15("1.15"),
v1_16("1.16");
{code}
The newly created branch and updated {{master}} branch need to be pushed to the 
official repository.
h4. Flink Docker Repository

Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
[apache/flink-docker|https://github.com/apache/flink-docker] r

[GitHub] [flink-ml] Fanoid commented on a diff in pull request #213: [FLINK-31126] Move classes not depending on Flink runtime from flink-ml-core to flink-ml-servable-core

2023-02-21 Thread via GitHub


Fanoid commented on code in PR #213:
URL: https://github.com/apache/flink-ml/pull/213#discussion_r1112756657


##
flink-ml-servable-core/pom.xml:
##
@@ -0,0 +1,127 @@
+
+
+http://maven.apache.org/POM/4.0.0";
+ xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance";
+ xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/xsd/maven-4.0.0.xsd";>
+  4.0.0
+
+  
+org.apache.flink
+flink-ml-parent
+2.2-SNAPSHOT
+  
+
+  flink-ml-servable-core
+  Flink ML : Servable : Core
+
+  
+
+
+  org.apache.flink

Review Comment:
   @lindong28 and I had a discussion about `flink-core` in a previous draft PR 
[1]. We agreed this change can be put in a follow-up PR.
   
   [1] https://github.com/apache/flink-ml/pull/199#discussion_r1090122715



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Assigned] (FLINK-31092) Hive ITCases fail with OutOfMemoryError

2023-02-21 Thread Leonard Xu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31092?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Leonard Xu reassigned FLINK-31092:
--

Assignee: luoyuxia

> Hive ITCases fail with OutOfMemoryError
> ---
>
> Key: FLINK-31092
> URL: https://issues.apache.org/jira/browse/FLINK-31092
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Hive
>Affects Versions: 1.17.0
>Reporter: Matthias Pohl
>Assignee: luoyuxia
>Priority: Critical
>  Labels: test-stability
> Attachments: VisualVM-FLINK-31092.png
>
>
> We're experiencing a OutOfMemoryError where the heap space reaches the upper 
> limit:
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=46161&view=logs&j=fc5181b0-e452-5c8f-68de-1097947f6483&t=995c650b-6573-581c-9ce6-7ad4cc038461&l=23142
> {code}
> Feb 15 05:05:14 [INFO] Running 
> org.apache.flink.table.catalog.hive.HiveCatalogITCase
> Feb 15 05:05:17 [INFO] java.lang.OutOfMemoryError: Java heap space
> Feb 15 05:05:17 [INFO] Dumping heap to java_pid9669.hprof ...
> Feb 15 05:05:28 [INFO] Heap dump file created [1957090051 bytes in 11.718 
> secs]
> java.lang.OutOfMemoryError: Java heap space
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.cancelPingScheduler(ForkedBooter.java:209)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.acknowledgedExit(ForkedBooter.java:419)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:186)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548)
> {code}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-30951) Release Testing: Verify FLINK-29635 Hive sink should support merge files in batch mode

2023-02-21 Thread Shengkai Fang (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-30951?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shengkai Fang closed FLINK-30951.
-
Resolution: Fixed

> Release Testing: Verify FLINK-29635 Hive sink should support merge files in 
> batch mode
> --
>
> Key: FLINK-30951
> URL: https://issues.apache.org/jira/browse/FLINK-30951
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Hive
>Reporter: luoyuxia
>Assignee: Shengkai Fang
>Priority: Blocker
> Fix For: 1.17.0
>
> Attachments: screenshot-1.png
>
>
> The issue aims to verfiy FLINK-29635.
> Please verify in batch mode, the document is in 
> [https://nightlies.apache.org/flink/flink-docs-master/docs/connectors/table/hive/hive_read_write/#file-compaction]:
>  
> 1: enable auto-compaction, write some data to a Hive table which results in 
> the average size of files is less than compaction.small-files.avg-size(16MB 
> by default), verfiy these files should be merged.
> 2:  enable auto-compaction, set compaction.small-files.avg-size to a smaller 
> values, then write some data to a Hive table which results in the average 
> size of files is greater thant the compaction.small-files.avg-size, verfiy 
> these files shouldn't be merged.
> 3. set sink.parallelism manually, check the parallelism of the compact 
> operator is equal to sink.parallelism.
> 4. set compaction.parallelism manually, check the parallelism of the compact 
> operator is equal to compaction.parallelism.
> 5. set compaction.file-size, check the size of the each target file merged is 
> about the `compaction.file-size`.
>  
> We shoud verify it with writing non-partitioned table, static partition 
> table, dynamic partition table.
> We can find the example sql for how to create & write hive table in the 
> codebase  
> [HiveTableCompactSinkITCase|[https://github.com/apache/flink/blob/0915c9850d861165e283acc0f60545cd836f0567/flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableCompactSinkITCase.java]].
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-ml] zhipeng93 commented on a diff in pull request #213: [FLINK-31126] Move classes not depending on Flink runtime from flink-ml-core to flink-ml-servable-core

2023-02-21 Thread via GitHub


zhipeng93 commented on code in PR #213:
URL: https://github.com/apache/flink-ml/pull/213#discussion_r1112772912


##
flink-ml-servable-core/pom.xml:
##
@@ -0,0 +1,127 @@
+
+
+http://maven.apache.org/POM/4.0.0";
+ xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance";
+ xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 
http://maven.apache.org/xsd/maven-4.0.0.xsd";>
+  4.0.0
+
+  
+org.apache.flink
+flink-ml-parent
+2.2-SNAPSHOT
+  
+
+  flink-ml-servable-core
+  Flink ML : Servable : Core
+
+  
+
+
+  org.apache.flink

Review Comment:
   Thanks for the explanation. I agree it is a reasonable solution in the long 
term.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-31059) Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by native implementation

2023-02-21 Thread dalongliu (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31059?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691472#comment-17691472
 ] 

dalongliu commented on FLINK-31059:
---

[~miamiaoxyz] Thanks for your verify, for the problem, I will see it.

> Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by 
> native implementation
> -
>
> Key: FLINK-31059
> URL: https://issues.apache.org/jira/browse/FLINK-31059
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Hive
>Affects Versions: 1.17.0
>Reporter: dalongliu
>Assignee: miamiaoxyz
>Priority: Blocker
> Fix For: 1.17.0
>
> Attachments: image-2023-02-21-15-45-48-226.png, 
> image-2023-02-21-15-46-13-966.png, image-2023-02-21-15-47-54-043.png, 
> image-2023-02-21-15-49-58-854.png, image-2023-02-21-15-59-44-470.png, 
> image-2023-02-21-16-28-22-038.png, image-2023-02-21-16-29-42-983.png, 
> image-2023-02-21-16-31-58-361.png, image-2023-02-21-16-35-46-294.png
>
>
> This task aims to verify 
> [FLINK-29717|https://issues.apache.org/jira/browse/FLINK-29717] which 
> improves the hive udaf performance.
> As the document [PR|https://github.com/apache/flink/pull/21789] description, 
> please veriy:
> 1. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count/avg/min/max functions separately in the query to verify if the 
> hash-agg strategy is chosen via plan, and verify if the data results are the 
> same as when the option `table.exec.hive.native-agg-function.enabled` is 
> disabled.
> 2. Enabling the option `table.exec.hive.native-agg-function.enabled`, combine 
> sum/count/avg/min/max functions in query, verify if the hash-agg strategy is 
> chosen via plan, and verify if the data results are the same as when option 
> `table.exec.hive.native-agg-function.enabled` is disabled.
> 3. Enabling the option `table.exec.hive.native-agg-function.enabled`, count 
> or max array&struct and other complex types in query, verify whether the 
> sort-agg strategy is chosen via plan, verify whether the data result is the 
> same as when option `table.exec.hive.native-agg-function.enabled` is disabled.
> 4. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count and first_value/last_value functions in the query simultaneously, 
> verify that the sort-agg strategy is chosen via plan, verify that the data is 
> the same as when option `table.exec.hive.native-agg-function.enabled` is 
> disabled.
> 5. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count/avg/min/max functions in the query and open sort-agg strategy 
> forcibly, verify that the data results are the same as when option 
> `table.exec.hive.native-agg-function.enabled` is disabled.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-ml] zhipeng93 commented on pull request #213: [FLINK-31126] Move classes not depending on Flink runtime from flink-ml-core to flink-ml-servable-core

2023-02-21 Thread via GitHub


zhipeng93 commented on PR #213:
URL: https://github.com/apache/flink-ml/pull/213#issuecomment-1438119732

   Thanks for the PR. LGTM.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-ml] zhipeng93 merged pull request #213: [FLINK-31126] Move classes not depending on Flink runtime from flink-ml-core to flink-ml-servable-core

2023-02-21 Thread via GitHub


zhipeng93 merged PR #213:
URL: https://github.com/apache/flink-ml/pull/213


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Closed] (FLINK-31059) Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by native implementation

2023-02-21 Thread dalongliu (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31059?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

dalongliu closed FLINK-31059.
-
Resolution: Fixed

> Release Testing: Verify FLINK-29717 Supports hive udaf such as sum/count by 
> native implementation
> -
>
> Key: FLINK-31059
> URL: https://issues.apache.org/jira/browse/FLINK-31059
> Project: Flink
>  Issue Type: Sub-task
>  Components: Connectors / Hive
>Affects Versions: 1.17.0
>Reporter: dalongliu
>Assignee: miamiaoxyz
>Priority: Blocker
> Fix For: 1.17.0
>
> Attachments: image-2023-02-21-15-45-48-226.png, 
> image-2023-02-21-15-46-13-966.png, image-2023-02-21-15-47-54-043.png, 
> image-2023-02-21-15-49-58-854.png, image-2023-02-21-15-59-44-470.png, 
> image-2023-02-21-16-28-22-038.png, image-2023-02-21-16-29-42-983.png, 
> image-2023-02-21-16-31-58-361.png, image-2023-02-21-16-35-46-294.png
>
>
> This task aims to verify 
> [FLINK-29717|https://issues.apache.org/jira/browse/FLINK-29717] which 
> improves the hive udaf performance.
> As the document [PR|https://github.com/apache/flink/pull/21789] description, 
> please veriy:
> 1. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count/avg/min/max functions separately in the query to verify if the 
> hash-agg strategy is chosen via plan, and verify if the data results are the 
> same as when the option `table.exec.hive.native-agg-function.enabled` is 
> disabled.
> 2. Enabling the option `table.exec.hive.native-agg-function.enabled`, combine 
> sum/count/avg/min/max functions in query, verify if the hash-agg strategy is 
> chosen via plan, and verify if the data results are the same as when option 
> `table.exec.hive.native-agg-function.enabled` is disabled.
> 3. Enabling the option `table.exec.hive.native-agg-function.enabled`, count 
> or max array&struct and other complex types in query, verify whether the 
> sort-agg strategy is chosen via plan, verify whether the data result is the 
> same as when option `table.exec.hive.native-agg-function.enabled` is disabled.
> 4. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count and first_value/last_value functions in the query simultaneously, 
> verify that the sort-agg strategy is chosen via plan, verify that the data is 
> the same as when option `table.exec.hive.native-agg-function.enabled` is 
> disabled.
> 5. Enabling the option `table.exec.hive.native-agg-function.enabled`, use the 
> sum/count/avg/min/max functions in the query and open sort-agg strategy 
> forcibly, verify that the data results are the same as when option 
> `table.exec.hive.native-agg-function.enabled` is disabled.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-ml] lindong28 commented on pull request #213: [FLINK-31126] Move classes not depending on Flink runtime from flink-ml-core to flink-ml-servable-core

2023-02-21 Thread via GitHub


lindong28 commented on PR #213:
URL: https://github.com/apache/flink-ml/pull/213#issuecomment-1438122258

   Thank you both for the review!


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31153:
--
Description: 
If you are doing a new minor release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Patch releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
releasing/update_branch_version.sh
{code}
In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 as the last entry:
{code:java}
// ...
v1_12("1.12"),
v1_13("1.13"),
v1_14("1.14"),
v1_15("1.15"),
v1_16("1.16");
{code}
The newly created branch and updated {{master}} branch need to be pushed to the 
official repository.
h4. Flink Docker Repository

Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
[apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
sure that 
[apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
 points to the correct snapshot version; for {{dev-x.y}} it should point to 
{{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).

After pushing the new minor release branch, as the last step you should also 
update the documentation workflow to also build the documentation for the new 
release branch. Check [Managing 
Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
 on details on how to do that. You may also want to manually trigger a build to 
make the changes visible as soon as possible.

 

h3. Expectations (Minor Version only)
 * Release branch has been created and pushed
 * Cron job has been added on the release branch in 
([apache-flink:./tools/azure-pipelines/build-apache-repo.yml|https://github.com/apache/flink/blob/master/tools/azure-pipelines/build-apache-repo.yml])
 * Changes on the new release branch are picked up by [Azure 
CI|https://dev.azure.com/apache-flink/apache-flink/_build?definitionId=1&_a=summary]
 * Originating branch has the version information updated to the new version
 * New version is added to the 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum.
 * Make sure [flink-docker|https://github.com/apache/flink-docker/] has 
{{dev-x.y}} branch and docker e2e tests run against this branch in the 
corresponding Apache Flink release branch (see 
[apache/flink:flink-end-to-end-tests/test-scripts/common_docker.sh:51|https://github.com/apache/flink/blob/master/flink-end-to-end-tests/test-scripts/common_docker.sh#L51])
 * 
[apache-flink:docs/config.toml|https://github.com/apache/flink/blob/release-1.17/docs/config.toml]
 has been updated appropriately in the new Apache Flink release branch.
 * The {{dev-x.y}} branch of ({{{}$CURRENT_SNAPSHOT_VERSION{}}}) have been 
created in the Flink Benchmark repo.
 * The {{flink.version}} property (see 
[apache/flink-benchmarks:pom.xml|https://github.com/apache/flink-benchmarks/blob/master/pom.xml#L48]
 of Flink Benchmark repo has been updated to the latest snapshot version.

  was:
If you are doing a new minor release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Patch releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
releasing/update_branch_version.sh
{code}
In the {{master}} branch, add a new valu

[jira] [Updated] (FLINK-31147) Create a new version in JIRA

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31147?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31147:
--
Description: 
When contributors resolve an issue in JIRA, they are tagging it with a release 
that will contain their changes. With the release currently underway, new 
issues should be resolved against a subsequent future release. Therefore, you 
should create a release item for this subsequent release, as follows:
 # In JIRA, navigate to the [Flink > Administration > 
Versions|https://issues.apache.org/jira/plugins/servlet/project-config/FLINK/versions].
 # Add a new release: choose the next minor version number compared to the one 
currently underway, select today’s date as the Start Date, and choose Add.
(Note: Only PMC members have access to the project administration. If you do 
not have access, ask on the mailing list for assistance.)

 

h3. Expectations
 * The new version should be listed in the dropdown menu of {{fixVersion}} or 
{{affectedVersion}} under "unreleased versions" when creating a new Jira issue.

  was:
When contributors resolve an issue in JIRA, they are tagging it with a release 
that will contain their changes. With the release currently underway, new 
issues should be resolved against a subsequent future release. Therefore, you 
should create a release item for this subsequent release, as follows:
 # In JIRA, navigate to the [Flink > Administration > 
Versions|https://issues.apache.org/jira/plugins/servlet/project-config/FLINK/versions].
 # Add a new release: choose the next minor version number compared to the one 
currently underway, select today’s date as the Start Date, and choose Add.
(Note: Only PMC members have access to the project administration. If you do 
not have access, ask on the mailing list for assistance.)


> Create a new version in JIRA
> 
>
> Key: FLINK-31147
> URL: https://issues.apache.org/jira/browse/FLINK-31147
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Priority: Major
>
> When contributors resolve an issue in JIRA, they are tagging it with a 
> release that will contain their changes. With the release currently underway, 
> new issues should be resolved against a subsequent future release. Therefore, 
> you should create a release item for this subsequent release, as follows:
>  # In JIRA, navigate to the [Flink > Administration > 
> Versions|https://issues.apache.org/jira/plugins/servlet/project-config/FLINK/versions].
>  # Add a new release: choose the next minor version number compared to the 
> one currently underway, select today’s date as the Start Date, and choose Add.
> (Note: Only PMC members have access to the project administration. If you do 
> not have access, ask on the mailing list for assistance.)
>  
> 
> h3. Expectations
>  * The new version should be listed in the dropdown menu of {{fixVersion}} or 
> {{affectedVersion}} under "unreleased versions" when creating a new Jira 
> issue.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31149) Review and update documentation

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31149?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31149:
--
Description: 
There are a few pages in the documentation that need to be reviewed and updated 
for each release.
 * Ensure that there exists a release notes page for each non-bugfix release 
(e.g., 1.5.0) in {{{}./docs/release-notes/{}}}, that it is up-to-date, and 
linked from the start page of the documentation.
 * Upgrading Applications and Flink Versions: 
[https://ci.apache.org/projects/flink/flink-docs-master/ops/upgrading.html]
 * ...

 

h3. Expectations
 * Update upgrade compatibility table 
([apache-flink:./docs/content/docs/ops/upgrading.md|https://github.com/apache/flink/blob/master/docs/content/docs/ops/upgrading.md#compatibility-table]
 and 
[apache-flink:./docs/content.zh/docs/ops/upgrading.md|https://github.com/apache/flink/blob/master/docs/content.zh/docs/ops/upgrading.md#compatibility-table]).
 * Update [Release Overview in 
Confluence|https://cwiki.apache.org/confluence/display/FLINK/Release+Management+and+Feature+Plan]
 * (minor only) The documentation for the new major release is visible under 
[https://nightlies.apache.org/flink/flink-docs-release-$SHORT_RELEASE_VERSION] 
(after at least one [doc 
build|https://github.com/apache/flink/actions/workflows/docs.yml] succeeded).
 * (minor only) The documentation for the new major release does not contain 
"-SNAPSHOT" in its version title, and all links refer to the corresponding 
version docs instead of {{{}master{}}}.

  was:
There are a few pages in the documentation that need to be reviewed and updated 
for each release.
 * Ensure that there exists a release notes page for each non-bugfix release 
(e.g., 1.5.0) in {{{}./docs/release-notes/{}}}, that it is up-to-date, and 
linked from the start page of the documentation.
 * Upgrading Applications and Flink Versions: 
[https://ci.apache.org/projects/flink/flink-docs-master/ops/upgrading.html]
 * ...

 

h3. Expectations
 * Update upgrade compatibility table 
([apache-flink:./docs/content/docs/ops/upgrading.md|https://github.com/apache/flink/blob/master/docs/content/docs/ops/upgrading.md#compatibility-table]
 and 
[apache-flink:./docs/content.zh/docs/ops/upgrading.md|https://github.com/apache/flink/blob/master/docs/content.zh/docs/ops/upgrading.md#compatibility-table]).
 * Update [Release Overview in 
Confluence|https://cwiki.apache.org/confluence/display/FLINK/Release+Management+and+Feature+Plan]
 * (major only) The documentation for the new major release is visible under 
https://nightlies.apache.org/flink/flink-docs-release-$SHORT_RELEASE_VERSION 
(after at least one [doc 
build|https://github.com/apache/flink/actions/workflows/docs.yml] succeeded).
 * (major only) The documentation for the new major release does not contain 
"-SNAPSHOT" in its version title, and all links refer to the corresponding 
version docs instead of {{master}}.


> Review and update documentation
> ---
>
> Key: FLINK-31149
> URL: https://issues.apache.org/jira/browse/FLINK-31149
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Priority: Major
>
> There are a few pages in the documentation that need to be reviewed and 
> updated for each release.
>  * Ensure that there exists a release notes page for each non-bugfix release 
> (e.g., 1.5.0) in {{{}./docs/release-notes/{}}}, that it is up-to-date, and 
> linked from the start page of the documentation.
>  * Upgrading Applications and Flink Versions: 
> [https://ci.apache.org/projects/flink/flink-docs-master/ops/upgrading.html]
>  * ...
>  
> 
> h3. Expectations
>  * Update upgrade compatibility table 
> ([apache-flink:./docs/content/docs/ops/upgrading.md|https://github.com/apache/flink/blob/master/docs/content/docs/ops/upgrading.md#compatibility-table]
>  and 
> [apache-flink:./docs/content.zh/docs/ops/upgrading.md|https://github.com/apache/flink/blob/master/docs/content.zh/docs/ops/upgrading.md#compatibility-table]).
>  * Update [Release Overview in 
> Confluence|https://cwiki.apache.org/confluence/display/FLINK/Release+Management+and+Feature+Plan]
>  * (minor only) The documentation for the new major release is visible under 
> [https://nightlies.apache.org/flink/flink-docs-release-$SHORT_RELEASE_VERSION]
>  (after at least one [doc 
> build|https://github.com/apache/flink/actions/workflows/docs.yml] succeeded).
>  * (minor only) The documentation for the new major release does not contain 
> "-SNAPSHOT" in its version title, and all links refer to the corresponding 
> version docs instead of {{{}master{}}}.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink] reswqa commented on a diff in pull request #21977: [FLINK-31132] Compact without setting parallelism does not follow the configured sink parallelism for HiveTableSink

2023-02-21 Thread via GitHub


reswqa commented on code in PR #21977:
URL: https://github.com/apache/flink/pull/21977#discussion_r1112641083


##
flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableCompactSinkITCase.java:
##
@@ -42,17 +44,18 @@
 import static org.assertj.core.api.Assertions.assertThat;
 
 /** IT case for Hive table compaction in batch mode. */
-public class HiveTableCompactSinkITCase {
+@ExtendWith(TestLoggerExtension.class)
+class HiveTableCompactSinkITCase {
 
 @RegisterExtension
-private static final MiniClusterExtension MINI_CLUSTER = new 
MiniClusterExtension();
+public static final MiniClusterExtension MINI_CLUSTER = new 
MiniClusterExtension();

Review Comment:
   TBH, The reason I changed it is only to follow the usage in the 
`MiniClusterExtension` class 
[javadoc](https://github.com/apache/flink/blob/bacdc326b58749924acbd8921d63eda06663a225/flink-test-utils-parent/flink-test-utils/src/main/java/org/apache/flink/test/junit5/MiniClusterExtension.java#L70).
   
   At the same time, in the `architecture check` of IT case, we can see similar 
[requirement](https://github.com/apache/flink/blob/bacdc326b58749924acbd8921d63eda06663a225/flink-architecture-tests/flink-architecture-tests-test/src/main/java/org/apache/flink/architecture/rules/ITCaseRules.java#L67).
 Unfortunately, I looked at the actual check logic and didn't seem to require 
it to be public, which really confused me.
   
   
   



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] FangYongs commented on pull request #21956: [FLINK-31113] Support AND filter push down in orc

2023-02-21 Thread via GitHub


FangYongs commented on PR #21956:
URL: https://github.com/apache/flink/pull/21956#issuecomment-1438133167

   Thanks @libenchao , I have updated


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-ml] lindong28 commented on pull request #214: [FLINK-31127] Add public API classes for FLIP-289

2023-02-21 Thread via GitHub


lindong28 commented on PR #214:
URL: https://github.com/apache/flink-ml/pull/214#issuecomment-1438134181

   @jiangxin369 Can you help review this PR?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-31152) Setup for the executing release manager

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31152?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31152:
--
Description: 
h4. GPG Key

You need to have a GPG key to sign the release artifacts. Please be aware of 
the ASF-wide [release signing 
guidelines|https://www.apache.org/dev/release-signing.html]. If you don’t have 
a GPG key associated with your Apache account, please create one according to 
the guidelines.

Determine your Apache GPG Key and Key ID, as follows:
{code:java}
$ gpg --list-keys
{code}
This will list your GPG keys. One of these should reflect your Apache account, 
for example:
{code:java}
--
pub   2048R/845E6689 2016-02-23
uid  Nomen Nescio 
sub   2048R/BA4D50BE 2016-02-23
{code}
In the example above, the key ID is the 8-digit hex string in the {{pub}} line: 
{{{}845E6689{}}}.

Now, add your Apache GPG key to the Flink’s {{KEYS}} file in the [Apache Flink 
release KEYS file|https://dist.apache.org/repos/dist/release/flink/KEYS] 
repository at [dist.apache.org|http://dist.apache.org/]. Follow the 
instructions listed at the top of these files. (Note: Only PMC members have 
write access to the release repository. If you end up getting 403 errors ask on 
the mailing list for assistance.)

Configure {{git}} to use this key when signing code by giving it your key ID, 
as follows:
{code:java}
$ git config --global user.signingkey 845E6689
{code}
You may drop the {{--global}} option if you’d prefer to use this key for the 
current repository only.

You may wish to start {{gpg-agent}} to unlock your GPG key only once using your 
passphrase. Otherwise, you may need to enter this passphrase hundreds of times. 
The setup for {{gpg-agent}} varies based on operating system, but may be 
something like this:
{code:bash}
$ eval $(gpg-agent --daemon --no-grab --write-env-file $HOME/.gpg-agent-info)
$ export GPG_TTY=$(tty)
$ export GPG_AGENT_INFO
{code}
h4. Access to Apache Nexus repository

Configure access to the [Apache Nexus 
repository|https://repository.apache.org/], which enables final deployment of 
releases to the Maven Central Repository.
 # You log in with your Apache account.
 # Confirm you have appropriate access by finding {{org.apache.flink}} under 
{{{}Staging Profiles{}}}.
 # Navigate to your {{Profile}} (top right drop-down menu of the page).
 # Choose {{User Token}} from the dropdown, then click {{{}Access User 
Token{}}}. Copy a snippet of the Maven XML configuration block.
 # Insert this snippet twice into your global Maven {{settings.xml}} file, 
typically {{{}${HOME}/.m2/settings.xml{}}}. The end result should look like 
this, where {{TOKEN_NAME}} and {{TOKEN_PASSWORD}} are your secret tokens:
{code:xml}

   
 
   apache.releases.https
   TOKEN_NAME
   TOKEN_PASSWORD
 
 
   apache.snapshots.https
   TOKEN_NAME
   TOKEN_PASSWORD
 
   
 
{code}

h4. Website development setup

Get ready for updating the Flink website by following the [website development 
instructions|https://flink.apache.org/contributing/improve-website.html].
h4. GNU Tar Setup for Mac (Skip this step if you are not using a Mac)

The default tar application on Mac does not support GNU archive format and 
defaults to Pax. This bloats the archive with unnecessary metadata that can 
result in additional files when decompressing (see [1.15.2-RC2 vote 
thread|https://lists.apache.org/thread/mzbgsb7y9vdp9bs00gsgscsjv2ygy58q]). 
Install gnu-tar and create a symbolic link to use in preference of the default 
tar program.
{code:bash}
$ brew install gnu-tar
$ ln -s /usr/local/bin/gtar /usr/local/bin/tar
$ which tar
{code}
 

h3. Expectations
 * Release Manager’s GPG key is published to 
[dist.apache.org|http://dist.apache.org/]
 * Release Manager’s GPG key is configured in git configuration
 * Release Manager's GPG key is configured as the default gpg key.
 * Release Manager has {{org.apache.flink}} listed under Staging Profiles in 
Nexus
 * Release Manager’s Nexus User Token is configured in settings.xml

  was:
h4. GPG Key

You need to have a GPG key to sign the release artifacts. Please be aware of 
the ASF-wide [release signing 
guidelines|https://www.apache.org/dev/release-signing.html]. If you don’t have 
a GPG key associated with your Apache account, please create one according to 
the guidelines.

Determine your Apache GPG Key and Key ID, as follows:
{code:java}
$ gpg --list-keys
{code}
This will list your GPG keys. One of these should reflect your Apache account, 
for example:
{code:java}
--
pub   2048R/845E6689 2016-02-23
uid  Nomen Nescio 
sub   2048R/BA4D50BE 2016-02-23
{code}
In the example above, the key ID is the 8-digit hex string in the {{pub}} line: 
{{{}845E6689{}}}.

Now, add your Apache GPG key to the Flink’s {{KEYS}} file in the [Apache Flink

[jira] [Updated] (FLINK-31144) Slow scheduling on large-scale batch jobs

2023-02-21 Thread Julien Tournay (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Julien Tournay updated FLINK-31144:
---
Attachment: image-2023-02-21-10-29-49-388.png

> Slow scheduling on large-scale batch jobs 
> --
>
> Key: FLINK-31144
> URL: https://issues.apache.org/jira/browse/FLINK-31144
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Reporter: Julien Tournay
>Priority: Major
> Attachments: flink-1.17-snapshot-1676473798013.nps, 
> image-2023-02-21-10-29-49-388.png
>
>
> When executing a complex job graph at high parallelism 
> `DefaultPreferredLocationsRetriever.getPreferredLocationsBasedOnInputs` can 
> get slow and cause long pauses where the JobManager becomes unresponsive and 
> all the taskmanagers just wait. I've attached a VisualVM snapshot to 
> illustrate the problem.[^flink-1.17-snapshot-1676473798013.nps]
> At Spotify we have complex jobs where this issue can cause batch "pause" of 
> 40+ minutes and make the overall execution 30% slower or more.
> More importantly this prevent us from running said jobs on larger cluster as 
> adding resources to the cluster worsen the issue.
> We have successfully tested a modified Flink version where 
> `DefaultPreferredLocationsRetriever.getPreferredLocationsBasedOnInputs` was 
> completely commented and simply returns an empty collection and confirmed it 
> solves the issue.
> In the same spirit as a recent change 
> ([https://github.com/apache/flink/blob/43f419d0eccba86ecc8040fa6f521148f1e358ff/flink-runtime/src/main/java/org/apache/flink/runtime/scheduler/DefaultPreferredLocationsRetriever.java#L98-L102)]
>  there could be a mechanism in place to detect when Flink run into this 
> specific issue and just skip the call to `getInputLocationFutures`  
> [https://github.com/apache/flink/blob/43f419d0eccba86ecc8040fa6f521148f1e358ff/flink-runtime/src/main/java/org/apache/flink/runtime/scheduler/DefaultPreferredLocationsRetriever.java#L105-L108.]
> I'm not familiar enough with the internals of Flink to propose a more 
> advanced fix, however it seems like a configurable threshold on the number of 
> consumer vertices above which the preferred location is not computed would 
> do. If this  solution is good enough, I'd be happy to submit a PR.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[GitHub] [flink-ml] zhipeng93 commented on a diff in pull request #212: [FLINK-31125] Flink ML benchmark framework should minimize the source operator overhead

2023-02-21 Thread via GitHub


zhipeng93 commented on code in PR #212:
URL: https://github.com/apache/flink-ml/pull/212#discussion_r1112795550


##
flink-ml-benchmark/src/main/java/org/apache/flink/ml/benchmark/datagenerator/common/DenseVectorArrayGenerator.java:
##
@@ -42,7 +42,7 @@ protected RowGenerator[] getRowGenerators() {
 return new RowGenerator[] {
 new RowGenerator(getNumValues(), getSeed()) {
 @Override
-protected Row nextRow() {
+protected Row getRow() {

Review Comment:
   Reusing one row to simulate the input may cause the benchmark result not 
stable, since processing two different rows may take different time.
   
   The two optimizations introduced in this PR give us 12% speedup. Is there a 
breakdown analysis of this two optimization?
   
   If the data generation cost is indeed high, can we sample a fixed number of 
rows and reuse it later?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-ml] zhipeng93 commented on a diff in pull request #212: [FLINK-31125] Flink ML benchmark framework should minimize the source operator overhead

2023-02-21 Thread via GitHub


zhipeng93 commented on code in PR #212:
URL: https://github.com/apache/flink-ml/pull/212#discussion_r1112795550


##
flink-ml-benchmark/src/main/java/org/apache/flink/ml/benchmark/datagenerator/common/DenseVectorArrayGenerator.java:
##
@@ -42,7 +42,7 @@ protected RowGenerator[] getRowGenerators() {
 return new RowGenerator[] {
 new RowGenerator(getNumValues(), getSeed()) {
 @Override
-protected Row nextRow() {
+protected Row getRow() {

Review Comment:
   Reusing one row to simulate the input may cause the benchmark result not 
stable, since processing two different rows may take different time.
   
   The two optimizations introduced in this PR give us 12% speedup. Is there a 
breakdown analysis of this two optimization?
   
   If the data generation cost is indeed high, can we sample a fixed number of 
rows and reuse it in the evalution?



##
flink-ml-benchmark/src/main/java/org/apache/flink/ml/benchmark/datagenerator/common/DenseVectorArrayGenerator.java:
##
@@ -42,7 +42,7 @@ protected RowGenerator[] getRowGenerators() {
 return new RowGenerator[] {
 new RowGenerator(getNumValues(), getSeed()) {
 @Override
-protected Row nextRow() {
+protected Row getRow() {

Review Comment:
   Reusing one row to simulate the input may cause the benchmark result not 
stable, since processing two different rows may take different time.
   
   The two optimizations introduced in this PR give us 12% speedup. Is there a 
breakdown analysis of this two optimization?
   
   If the data generation cost is indeed high, can we sample a fixed number of 
rows and reuse it in the evaluation?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-31144) Slow scheduling on large-scale batch jobs

2023-02-21 Thread Julien Tournay (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31144?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691474#comment-17691474
 ] 

Julien Tournay commented on FLINK-31144:


Hi [~huwh],


Thank you for the quick reply :)


{quote}when the topology is complex.
{quote}

Indeed. For the issue to be noticeable, the jobgraph has to be fairly complex, 
feature all-to-all distributions and execute with a high parallelism.

 
{quote}1. Is the slow scheduling or the scheduled result of location preferred 
make your job slow?
{quote}

Yes it very much does. We have a job that takes ~2h30 (after many many tweaks 
to get the best possible perf.). It's impossible to get it to run in less time 
because adding more taskmanagers make the scheduling slow and overall the 
execution gets longer. Removing preferred location makes it possible to run it 
in less that 2h (We're aiming at ~1h45min).

 
{quote}2. "we have complex jobs where this issue can cause batch "pause" of 40+ 
minutes"  What does "pause" meaning? Is the getPreferredLocationsBasedOnInputs 
take more than 40+ minutes?
{quote}

By "pause" I mean that at the beginning of the execution, the taskmanagers will 
wait for the JobManager for ~40min and then will start processing. With Flink 
1.17 and no preferred location, the "pause" is down to ~5min.

I should also mention the JM is very unresponsive and the web console struggles 
the show anything. 


{quote}Could you provide the topology of the complex job.
{quote}

I can but not sure what format to use. The graph is quite big and a simple 
screenshot is unreadable: !image-2023-02-21-10-29-49-388.png!

I can maybe share the archived execution json file (~500Mb) if that's helpful ?

 

> Slow scheduling on large-scale batch jobs 
> --
>
> Key: FLINK-31144
> URL: https://issues.apache.org/jira/browse/FLINK-31144
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Reporter: Julien Tournay
>Priority: Major
> Attachments: flink-1.17-snapshot-1676473798013.nps, 
> image-2023-02-21-10-29-49-388.png
>
>
> When executing a complex job graph at high parallelism 
> `DefaultPreferredLocationsRetriever.getPreferredLocationsBasedOnInputs` can 
> get slow and cause long pauses where the JobManager becomes unresponsive and 
> all the taskmanagers just wait. I've attached a VisualVM snapshot to 
> illustrate the problem.[^flink-1.17-snapshot-1676473798013.nps]
> At Spotify we have complex jobs where this issue can cause batch "pause" of 
> 40+ minutes and make the overall execution 30% slower or more.
> More importantly this prevent us from running said jobs on larger cluster as 
> adding resources to the cluster worsen the issue.
> We have successfully tested a modified Flink version where 
> `DefaultPreferredLocationsRetriever.getPreferredLocationsBasedOnInputs` was 
> completely commented and simply returns an empty collection and confirmed it 
> solves the issue.
> In the same spirit as a recent change 
> ([https://github.com/apache/flink/blob/43f419d0eccba86ecc8040fa6f521148f1e358ff/flink-runtime/src/main/java/org/apache/flink/runtime/scheduler/DefaultPreferredLocationsRetriever.java#L98-L102)]
>  there could be a mechanism in place to detect when Flink run into this 
> specific issue and just skip the call to `getInputLocationFutures`  
> [https://github.com/apache/flink/blob/43f419d0eccba86ecc8040fa6f521148f1e358ff/flink-runtime/src/main/java/org/apache/flink/runtime/scheduler/DefaultPreferredLocationsRetriever.java#L105-L108.]
> I'm not familiar enough with the internals of Flink to propose a more 
> advanced fix, however it seems like a configurable threshold on the number of 
> consumer vertices above which the preferred location is not computed would 
> do. If this  solution is good enough, I'd be happy to submit a PR.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31153:
--
Description: 
If you are doing a new minor release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Patch releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
releasing/update_branch_version.sh
{code}
In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 as the last entry:
{code:java}
// ...
v1_12("1.12"),
v1_13("1.13"),
v1_14("1.14"),
v1_15("1.15"),
v1_16("1.16");
{code}
The newly created branch and updated {{master}} branch need to be pushed to the 
official repository.
h4. Flink Docker Repository

Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
[apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
sure that 
[apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
 points to the correct snapshot version; for {{dev-x.y}} it should point to 
{{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).

After pushing the new minor release branch, as the last step you should also 
update the documentation workflow to also build the documentation for the new 
release branch. Check [Managing 
Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
 on details on how to do that. You may also want to manually trigger a build to 
make the changes visible as soon as possible.

 

h3. Expectations (Minor Version only)
 * Release branch has been created and pushed
 * Cron job has been added on the release branch in 
([apache-flink:./tools/azure-pipelines/build-apache-repo.yml|https://github.com/apache/flink/blob/master/tools/azure-pipelines/build-apache-repo.yml])
 * Changes on the new release branch are picked up by [Azure 
CI|https://dev.azure.com/apache-flink/apache-flink/_build?definitionId=1&_a=summary]
 * {{master}} branch has the version information updated to the new version 
(check pom.xml files and 
 * 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum)
 * New version is added to the 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum.
 * Make sure [flink-docker|https://github.com/apache/flink-docker/] has 
{{dev-x.y}} branch and docker e2e tests run against this branch in the 
corresponding Apache Flink release branch (see 
[apache/flink:flink-end-to-end-tests/test-scripts/common_docker.sh:51|https://github.com/apache/flink/blob/master/flink-end-to-end-tests/test-scripts/common_docker.sh#L51])
 * 
[apache-flink:docs/config.toml|https://github.com/apache/flink/blob/release-1.17/docs/config.toml]
 has been updated appropriately in the new Apache Flink release branch.
 * The {{flink.version}} property (see 
[apache/flink-benchmarks:pom.xml|https://github.com/apache/flink-benchmarks/blob/master/pom.xml#L48]
 of Flink Benchmark repo has been updated to the latest snapshot version.

  was:
If you are doing a new minor release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Patch releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VE

[GitHub] [flink-ml] zhipeng93 closed pull request #176: Bump jackson-databind from 2.12.6.1 to 2.12.7.1

2023-02-21 Thread via GitHub


zhipeng93 closed pull request #176: Bump jackson-databind from 2.12.6.1 to 
2.12.7.1
URL: https://github.com/apache/flink-ml/pull/176


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink-ml] dependabot[bot] commented on pull request #176: Bump jackson-databind from 2.12.6.1 to 2.12.7.1

2023-02-21 Thread via GitHub


dependabot[bot] commented on PR #176:
URL: https://github.com/apache/flink-ml/pull/176#issuecomment-1438157269

   OK, I won't notify you again about this release, but will get in touch when 
a new version is available. If you'd rather skip all updates until the next 
major or minor version, let me know by commenting `@dependabot ignore this 
major version` or `@dependabot ignore this minor version`.
   
   If you change your mind, just re-open this PR and I'll resolve any conflicts 
on it.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] dmvk commented on a diff in pull request #21981: [FLINK-21450][runtime] Support LocalRecovery by AdaptiveScheduler

2023-02-21 Thread via GitHub


dmvk commented on code in PR #21981:
URL: https://github.com/apache/flink/pull/21981#discussion_r1112810683


##
flink-runtime/src/main/java/org/apache/flink/runtime/scheduler/adaptive/allocator/StateLocalitySlotAssigner.java:
##
@@ -0,0 +1,209 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements.  See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License.  You may obtain a copy of the License at
+ *
+ *http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.runtime.scheduler.adaptive.allocator;
+
+import org.apache.flink.annotation.Internal;
+import org.apache.flink.runtime.clusterframework.types.AllocationID;
+import org.apache.flink.runtime.executiongraph.ExecutionGraph;
+import org.apache.flink.runtime.executiongraph.ExecutionJobVertex;
+import org.apache.flink.runtime.executiongraph.ExecutionVertex;
+import org.apache.flink.runtime.jobgraph.JobVertexID;
+import org.apache.flink.runtime.jobmaster.SlotInfo;
+import 
org.apache.flink.runtime.scheduler.adaptive.allocator.SlotSharingSlotAllocator.ExecutionSlotSharingGroup;
+import 
org.apache.flink.runtime.scheduler.adaptive.allocator.SlotSharingSlotAllocator.ExecutionSlotSharingGroupAndSlot;
+import org.apache.flink.runtime.scheduler.strategy.ExecutionVertexID;
+import org.apache.flink.runtime.state.KeyGroupRange;
+import org.apache.flink.runtime.state.KeyGroupRangeAssignment;
+import org.apache.flink.util.Preconditions;
+
+import java.util.ArrayList;
+import java.util.Collection;
+import java.util.Comparator;
+import java.util.HashMap;
+import java.util.Iterator;
+import java.util.List;
+import java.util.Map;
+import java.util.PriorityQueue;
+import java.util.stream.StreamSupport;
+
+import static java.util.function.Function.identity;
+import static java.util.stream.Collectors.toMap;
+
+/** A {@link SlotAssigner} that assigns slots based on the number of local key 
groups. */
+@Internal
+public class StateLocalitySlotAssigner implements SlotAssigner {
+
+private static class AllocationScore implements 
Comparable {
+
+private final String group;
+private final AllocationID allocationId;
+
+public AllocationScore(String group, AllocationID allocationId, int 
score) {
+this.group = group;
+this.allocationId = allocationId;
+this.score = score;
+}
+
+private final int score;
+
+public String getGroup() {
+return group;
+}
+
+public AllocationID getAllocationId() {
+return allocationId;
+}
+
+public int getScore() {
+return score;
+}
+
+@Override
+public int compareTo(StateLocalitySlotAssigner.AllocationScore other) {
+int result = Integer.compare(score, other.score);
+if (result != 0) {
+return result;
+}
+result = other.allocationId.compareTo(allocationId);
+if (result != 0) {
+return result;
+}
+return other.group.compareTo(group);
+}
+}
+
+private final Map> locality;
+private final Map maxParallelism;
+
+public StateLocalitySlotAssigner(ExecutionGraph archivedExecutionGraph) {
+this(
+calculateLocalKeyGroups(archivedExecutionGraph),
+StreamSupport.stream(
+
archivedExecutionGraph.getVerticesTopologically().spliterator(),
+false)
+.collect(
+toMap(
+ExecutionJobVertex::getJobVertexId,
+
ExecutionJobVertex::getMaxParallelism)));
+}
+
+public StateLocalitySlotAssigner(
+Map> locality,
+Map maxParallelism) {
+this.locality = locality;
+this.maxParallelism = maxParallelism;
+}
+
+@Override
+public AssignmentResult assignSlots(
+Collection slots, 
Collection groups) {
+
+final Map parallelism = new HashMap<>();
+groups.forEach(
+group ->
+group.getContainedExecutionVertices()
+.forEach(
+evi ->
+parallelism.merge(
+

[GitHub] [flink] dmvk commented on a diff in pull request #21981: [FLINK-21450][runtime] Support LocalRecovery by AdaptiveScheduler

2023-02-21 Thread via GitHub


dmvk commented on code in PR #21981:
URL: https://github.com/apache/flink/pull/21981#discussion_r1112815613


##
flink-runtime/src/main/java/org/apache/flink/runtime/scheduler/adaptive/allocator/SlotSharingSlotAllocator.java:
##
@@ -121,16 +133,13 @@ public Optional 
determineParallelism(
 slotSharingGroupParallelism.get(
 slotSharingGroup.getSlotSharingGroupId()));
 
-final Iterable 
sharedSlotToVertexAssignment =
+final List sharedSlotToVertexAssignment 
=
 createExecutionSlotSharingGroups(vertexParallelism);
 
-for (ExecutionSlotSharingGroup executionSlotSharingGroup :
-sharedSlotToVertexAssignment) {
-final SlotInfo slotInfo = slotIterator.next();
-
-assignments.add(
-new 
ExecutionSlotSharingGroupAndSlot(executionSlotSharingGroup, slotInfo));
-}
+SlotAssigner.AssignmentResult result =
+slotAssigner.assignSlots(freeSlots, 
sharedSlotToVertexAssignment);
+assignments.addAll(result.assignments);
+freeSlots = result.remainingSlots;

Review Comment:
   🤔 Is there a reason to run this within the loop at all? Can we simply move 
this to after the loop?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-31142) Some queries lead to abrupt sql client close

2023-02-21 Thread Sergey Nuyanzin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31142?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Nuyanzin updated FLINK-31142:

Priority: Blocker  (was: Major)

> Some queries lead to abrupt sql client close
> 
>
> Key: FLINK-31142
> URL: https://issues.apache.org/jira/browse/FLINK-31142
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / Client
>Affects Versions: 1.17.0
>Reporter: Sergey Nuyanzin
>Priority: Blocker
>
> Although the behavior has been changed in 1.17.0, I'm not sure whether it is 
> a blocker or not, since in both cases it is invalid query.
> The difference in the behavior is that before 1.17.0
> a query like 
> {code:sql}
> select /* multiline comment;
> {code}
> fails to execute and sql client prompts to submit another query.
> In 1.17.0 it  shuts down the session failing with 
> {noformat}
> Exception in thread "main" org.apache.flink.table.client.SqlClientException: 
> Could not read from command line.
>   at 
> org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:205)
>   at 
> org.apache.flink.table.client.cli.CliClient.executeInteractive(CliClient.java:168)
>   at 
> org.apache.flink.table.client.cli.CliClient.executeInInteractiveMode(CliClient.java:113)
>   at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:169)
>   at org.apache.flink.table.client.SqlClient.start(SqlClient.java:118)
>   at 
> org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:228)
>   at org.apache.flink.table.client.SqlClient.main(SqlClient.java:179)
> Caused by: org.apache.flink.sql.parser.impl.TokenMgrError: Lexical error at 
> line 1, column 29.  Encountered:  after : ""
>   at 
> org.apache.flink.sql.parser.impl.FlinkSqlParserImplTokenManager.getNextToken(FlinkSqlParserImplTokenManager.java:26752)
>   at 
> org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.scan(SqlCommandParserImpl.java:89)
>   at 
> org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.next(SqlCommandParserImpl.java:81)
>   at 
> org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.checkIncompleteStatement(SqlCommandParserImpl.java:141)
>   at 
> org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.getCommand(SqlCommandParserImpl.java:111)
>   at 
> org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.parseStatement(SqlCommandParserImpl.java:52)
>   at 
> org.apache.flink.table.client.cli.parser.SqlMultiLineParser.parse(SqlMultiLineParser.java:82)
>   at 
> org.jline.reader.impl.LineReaderImpl.acceptLine(LineReaderImpl.java:2964)
>   at 
> org.jline.reader.impl.LineReaderImpl$1.apply(LineReaderImpl.java:3778)
>   at 
> org.jline.reader.impl.LineReaderImpl.readLine(LineReaderImpl.java:679)
>   at 
> org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:183)
>   ... 6 more
> Shutting down the session...
> done.
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31142) Some queries lead to abrupt sql client close

2023-02-21 Thread Sergey Nuyanzin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31142?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Nuyanzin updated FLINK-31142:

Description: 
Although the behavior has been changed in 1.17.0, I'm not sure whether it is a 
blocker or not, since in both cases it is invalid query.
I put it to blocker just because of regression.

The difference in the behavior is that before 1.17.0
a query like 
{code:sql}
select /* multiline comment;
{code}
fails to execute and sql client prompts to submit another query.

In 1.17.0 it  shuts down the session failing with 
{noformat}
Exception in thread "main" org.apache.flink.table.client.SqlClientException: 
Could not read from command line.
at 
org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:205)
at 
org.apache.flink.table.client.cli.CliClient.executeInteractive(CliClient.java:168)
at 
org.apache.flink.table.client.cli.CliClient.executeInInteractiveMode(CliClient.java:113)
at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:169)
at org.apache.flink.table.client.SqlClient.start(SqlClient.java:118)
at 
org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:228)
at org.apache.flink.table.client.SqlClient.main(SqlClient.java:179)
Caused by: org.apache.flink.sql.parser.impl.TokenMgrError: Lexical error at 
line 1, column 29.  Encountered:  after : ""
at 
org.apache.flink.sql.parser.impl.FlinkSqlParserImplTokenManager.getNextToken(FlinkSqlParserImplTokenManager.java:26752)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.scan(SqlCommandParserImpl.java:89)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.next(SqlCommandParserImpl.java:81)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.checkIncompleteStatement(SqlCommandParserImpl.java:141)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.getCommand(SqlCommandParserImpl.java:111)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.parseStatement(SqlCommandParserImpl.java:52)
at 
org.apache.flink.table.client.cli.parser.SqlMultiLineParser.parse(SqlMultiLineParser.java:82)
at 
org.jline.reader.impl.LineReaderImpl.acceptLine(LineReaderImpl.java:2964)
at 
org.jline.reader.impl.LineReaderImpl$1.apply(LineReaderImpl.java:3778)
at 
org.jline.reader.impl.LineReaderImpl.readLine(LineReaderImpl.java:679)
at 
org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:183)
... 6 more

Shutting down the session...
done.

{noformat}



  was:
Although the behavior has been changed in 1.17.0, I'm not sure whether it is a 
blocker or not, since in both cases it is invalid query.

The difference in the behavior is that before 1.17.0
a query like 
{code:sql}
select /* multiline comment;
{code}
fails to execute and sql client prompts to submit another query.

In 1.17.0 it  shuts down the session failing with 
{noformat}
Exception in thread "main" org.apache.flink.table.client.SqlClientException: 
Could not read from command line.
at 
org.apache.flink.table.client.cli.CliClient.getAndExecuteStatements(CliClient.java:205)
at 
org.apache.flink.table.client.cli.CliClient.executeInteractive(CliClient.java:168)
at 
org.apache.flink.table.client.cli.CliClient.executeInInteractiveMode(CliClient.java:113)
at org.apache.flink.table.client.SqlClient.openCli(SqlClient.java:169)
at org.apache.flink.table.client.SqlClient.start(SqlClient.java:118)
at 
org.apache.flink.table.client.SqlClient.startClient(SqlClient.java:228)
at org.apache.flink.table.client.SqlClient.main(SqlClient.java:179)
Caused by: org.apache.flink.sql.parser.impl.TokenMgrError: Lexical error at 
line 1, column 29.  Encountered:  after : ""
at 
org.apache.flink.sql.parser.impl.FlinkSqlParserImplTokenManager.getNextToken(FlinkSqlParserImplTokenManager.java:26752)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.scan(SqlCommandParserImpl.java:89)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl$TokenIterator.next(SqlCommandParserImpl.java:81)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.checkIncompleteStatement(SqlCommandParserImpl.java:141)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.getCommand(SqlCommandParserImpl.java:111)
at 
org.apache.flink.table.client.cli.parser.SqlCommandParserImpl.parseStatement(SqlCommandParserImpl.java:52)
at 
org.apache.flink.table.client.cli.parser.SqlMultiLineParser.parse(SqlMultiLineParser.java:82)
at 
org.jline.reader.impl.LineReaderImpl.acceptLine(LineReaderImpl.java:2964)
at 
org.jline.reader.impl.LineReaderImpl$1.apply(LineRea

[jira] [Updated] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31153:
--
Description: 
If you are doing a new minor release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Patch releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
releasing/update_branch_version.sh
{code}
In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 as the last entry:
{code:java}
// ...
v1_12("1.12"),
v1_13("1.13"),
v1_14("1.14"),
v1_15("1.15"),
v1_16("1.16");
{code}
The newly created branch and updated {{master}} branch need to be pushed to the 
official repository.
h4. Flink Docker Repository

Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
[apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
sure that 
[apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
 points to the correct snapshot version; for {{dev-x.y}} it should point to 
{{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).

After pushing the new minor release branch, as the last step you should also 
update the documentation workflow to also build the documentation for the new 
release branch. Check [Managing 
Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
 on details on how to do that. You may also want to manually trigger a build to 
make the changes visible as soon as possible.

 

h3. Expectations (Minor Version only)
 * Release branch has been created and pushed
 * Changes on the new release branch are picked up by [Azure 
CI|https://dev.azure.com/apache-flink/apache-flink/_build?definitionId=1&_a=summary]
 * {{master}} branch has the version information updated to the new version 
(check pom.xml files and 
 * 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum)
 * New version is added to the 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum.
 * Make sure [flink-docker|https://github.com/apache/flink-docker/] has 
{{dev-x.y}} branch and docker e2e tests run against this branch in the 
corresponding Apache Flink release branch (see 
[apache/flink:flink-end-to-end-tests/test-scripts/common_docker.sh:51|https://github.com/apache/flink/blob/master/flink-end-to-end-tests/test-scripts/common_docker.sh#L51])
 * 
[apache-flink:docs/config.toml|https://github.com/apache/flink/blob/release-1.17/docs/config.toml]
 has been updated appropriately in the new Apache Flink release branch.
 * The {{flink.version}} property (see 
[apache/flink-benchmarks:pom.xml|https://github.com/apache/flink-benchmarks/blob/master/pom.xml#L48]
 of Flink Benchmark repo has been updated to the latest snapshot version.

  was:
If you are doing a new minor release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Patch releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
releasing/update_branch_version.sh
{code}
In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
[apache-flink:flink-annotations/src/main/java/org/apache/

[jira] [Commented] (FLINK-31153) Create a release branch

2023-02-21 Thread Jing Ge (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691491#comment-17691491
 ] 

Jing Ge commented on FLINK-31153:
-

Make sense. Thanks [~mapohl] for the clarification. Looking forward to the 
structure Flink Release Task Collection Template - FRACT ? ;)

> Create a release branch
> ---
>
> Key: FLINK-31153
> URL: https://issues.apache.org/jira/browse/FLINK-31153
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Assignee: Leonard Xu
>Priority: Major
>
> If you are doing a new minor release, you need to update Flink version in the 
> following repositories:
>  * [apache/flink|https://github.com/apache/flink]
>  * [apache/flink-docker|https://github.com/apache/flink-docker]
>  * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]
> Patch releases don't require the these repositories to be touched. Simply 
> checkout the already existing branch for that version:
> {code:java}
> $ cd ./tools
> $ git checkout release-$SHORT_RELEASE_VERSION
> {code}
> h4. Flink repository
> Create a branch for the new version that we want to release before updating 
> the master branch to the next development version:
> {code:bash}
> $ cd ./tools
> $ releasing/create_snapshot_branch.sh
> $ git checkout master
> $ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
> releasing/update_branch_version.sh
> {code}
> In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  as the last entry:
> {code:java}
> // ...
> v1_12("1.12"),
> v1_13("1.13"),
> v1_14("1.14"),
> v1_15("1.15"),
> v1_16("1.16");
> {code}
> The newly created branch and updated {{master}} branch need to be pushed to 
> the official repository.
> h4. Flink Docker Repository
> Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
> [apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
> sure that 
> [apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
>  points to the correct snapshot version; for {{dev-x.y}} it should point to 
> {{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
> recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).
> After pushing the new minor release branch, as the last step you should also 
> update the documentation workflow to also build the documentation for the new 
> release branch. Check [Managing 
> Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
>  on details on how to do that. You may also want to manually trigger a build 
> to make the changes visible as soon as possible.
>  
> 
> h3. Expectations (Minor Version only)
>  * Release branch has been created and pushed
>  * Changes on the new release branch are picked up by [Azure 
> CI|https://dev.azure.com/apache-flink/apache-flink/_build?definitionId=1&_a=summary]
>  * {{master}} branch has the version information updated to the new version 
> (check pom.xml files and 
>  * 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  enum)
>  * New version is added to the 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  enum.
>  * Make sure [flink-docker|https://github.com/apache/flink-docker/] has 
> {{dev-x.y}} branch and docker e2e tests run against this branch in the 
> corresponding Apache Flink release branch (see 
> [apache/flink:flink-end-to-end-tests/test-scripts/common_docker.sh:51|https://github.com/apache/flink/blob/master/flink-end-to-end-tests/test-scripts/common_docker.sh#L51])
>  * 
> [apache-flink:docs/config.toml|https://github.com/apache/flink/blob/release-1.17/docs/config.toml]
>  has been updated appropriately in the new Apache Flink release branch.
>  * The {{flink.version}} property (see 
> [apache/flink-benchmarks:pom.xml|https://github.com/apache/flink-benchmarks/blob/master/pom.xml#L48]
>  of Flink Benchmark repo has been updated to the latest snapshot version.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31153:
--
Description: 
If you are doing a new minor release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Patch releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
releasing/update_branch_version.sh
{code}
In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 as the last entry:
{code:java}
// ...
v1_12("1.12"),
v1_13("1.13"),
v1_14("1.14"),
v1_15("1.15"),
v1_16("1.16");
{code}
The newly created branch and updated {{master}} branch need to be pushed to the 
official repository.
h4. Flink Docker Repository

Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
[apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
sure that 
[apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
 points to the correct snapshot version; for {{dev-x.y}} it should point to 
{{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).

After pushing the new minor release branch, as the last step you should also 
update the documentation workflow to also build the documentation for the new 
release branch. Check [Managing 
Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
 on details on how to do that. You may also want to manually trigger a build to 
make the changes visible as soon as possible.

h4. Flink Benchmark Repository
First of all, checkout the {{master}} branch to {{dev-x.y}} branch in 
[apache/flink-benchmarks|https://github.com/apache/flink-benchmarks], so that 
we can have a branch named {{dev-x.y}} which could be built on top of 
(${{CURRENT_SNAPSHOT_VERSION}}).

Then, inside the repository you need to manually update the {{flink.version 
}}property inside the main *pom.xml* file. It should be pointing to the most 
recent snapshot version ({{{}$NEXT_SNAPSHOT_VERSION{}}}). For example:
{code:xml}
1.18-SNAPSHOT
{code}

 

h3. Expectations (Minor Version only)
 * Release branch has been created and pushed
 * Changes on the new release branch are picked up by [Azure 
CI|https://dev.azure.com/apache-flink/apache-flink/_build?definitionId=1&_a=summary]
 * {{master}} branch has the version information updated to the new version 
(check pom.xml files and 
 * 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum)
 * New version is added to the 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum.
 * Make sure [flink-docker|https://github.com/apache/flink-docker/] has 
{{dev-x.y}} branch and docker e2e tests run against this branch in the 
corresponding Apache Flink release branch (see 
[apache/flink:flink-end-to-end-tests/test-scripts/common_docker.sh:51|https://github.com/apache/flink/blob/master/flink-end-to-end-tests/test-scripts/common_docker.sh#L51])
 * 
[apache-flink:docs/config.toml|https://github.com/apache/flink/blob/release-1.17/docs/config.toml]
 has been updated appropriately in the new Apache Flink release branch.
 * The {{flink.version}} property (see 
[apache/flink-benchmarks:pom.xml|https://github.com/apache/flink-benchmarks/blob/master/pom.xml#L48]
 of Flink Benchmark repo has been updated to the latest snapshot version.

  was:
If you are doing a new minor release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Patch releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd .

[jira] [Updated] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31153:
--
Description: 
If you are doing a new minor release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Patch releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
releasing/update_branch_version.sh
{code}
In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 as the last entry:
{code:java}
// ...
v1_12("1.12"),
v1_13("1.13"),
v1_14("1.14"),
v1_15("1.15"),
v1_16("1.16");
{code}
The newly created branch and updated {{master}} branch need to be pushed to the 
official repository.
h4. Flink Docker Repository

Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
[apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
sure that 
[apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
 points to the correct snapshot version; for {{dev-x.y}} it should point to 
{{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).

After pushing the new minor release branch, as the last step you should also 
update the documentation workflow to also build the documentation for the new 
release branch. Check [Managing 
Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
 on details on how to do that. You may also want to manually trigger a build to 
make the changes visible as soon as possible.

h4. Flink Benchmark Repository
First of all, checkout the {{master}} branch to {{dev-x.y}} branch in 
[apache/flink-benchmarks|https://github.com/apache/flink-benchmarks], so that 
we can have a branch named {{dev-x.y}} which could be built on top of 
(${{CURRENT_SNAPSHOT_VERSION}}).

Then, inside the repository you need to manually update the {{flink.version}} 
property inside the parent *pom.xml* file. It should be pointing to the most 
recent snapshot version ($NEXT_SNAPSHOT_VERSION). For example:
{code:xml}
1.18-SNAPSHOT
{code}

 

h3. Expectations (Minor Version only)
 * Release branch has been created and pushed
 * Changes on the new release branch are picked up by [Azure 
CI|https://dev.azure.com/apache-flink/apache-flink/_build?definitionId=1&_a=summary]
 * {{master}} branch has the version information updated to the new version 
(check pom.xml files and 
 * 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum)
 * New version is added to the 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum.
 * Make sure [flink-docker|https://github.com/apache/flink-docker/] has 
{{dev-x.y}} branch and docker e2e tests run against this branch in the 
corresponding Apache Flink release branch (see 
[apache/flink:flink-end-to-end-tests/test-scripts/common_docker.sh:51|https://github.com/apache/flink/blob/master/flink-end-to-end-tests/test-scripts/common_docker.sh#L51])
 * 
[apache-flink:docs/config.toml|https://github.com/apache/flink/blob/release-1.17/docs/config.toml]
 has been updated appropriately in the new Apache Flink release branch.
 * The {{flink.version}} property (see 
[apache/flink-benchmarks:pom.xml|https://github.com/apache/flink-benchmarks/blob/master/pom.xml#L48]
 of Flink Benchmark repo has been updated to the latest snapshot version.

  was:
If you are doing a new minor release, you need to update Flink version in the 
following repositories:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Patch releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools

[jira] [Updated] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31153:
--
Description: 
If you are doing a new minor release, you need to update Flink version in the 
following repositories and the [AzureCI project 
configuration|https://dev.azure.com/apache-flink/apache-flink/]:
 * [apache/flink|https://github.com/apache/flink]
 * [apache/flink-docker|https://github.com/apache/flink-docker]
 * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]

Patch releases don't require the these repositories to be touched. Simply 
checkout the already existing branch for that version:
{code:java}
$ cd ./tools
$ git checkout release-$SHORT_RELEASE_VERSION
{code}
h4. Flink repository

Create a branch for the new version that we want to release before updating the 
master branch to the next development version:
{code:bash}
$ cd ./tools
$ releasing/create_snapshot_branch.sh
$ git checkout master
$ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
releasing/update_branch_version.sh
{code}
In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 as the last entry:
{code:java}
// ...
v1_12("1.12"),
v1_13("1.13"),
v1_14("1.14"),
v1_15("1.15"),
v1_16("1.16");
{code}
The newly created branch and updated {{master}} branch need to be pushed to the 
official repository.
h4. Flink Docker Repository

Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
[apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
sure that 
[apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
 points to the correct snapshot version; for {{dev-x.y}} it should point to 
{{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).

After pushing the new minor release branch, as the last step you should also 
update the documentation workflow to also build the documentation for the new 
release branch. Check [Managing 
Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
 on details on how to do that. You may also want to manually trigger a build to 
make the changes visible as soon as possible.

h4. Flink Benchmark Repository
First of all, checkout the {{master}} branch to {{dev-x.y}} branch in 
[apache/flink-benchmarks|https://github.com/apache/flink-benchmarks], so that 
we can have a branch named {{dev-x.y}} which could be built on top of 
(${{CURRENT_SNAPSHOT_VERSION}}).

Then, inside the repository you need to manually update the {{flink.version}} 
property inside the parent *pom.xml* file. It should be pointing to the most 
recent snapshot version ($NEXT_SNAPSHOT_VERSION). For example:
{code:xml}
1.18-SNAPSHOT
{code}

h4. AzureCI Project Configuration
The new release branch needs to be configured within AzureCI to make azure 
aware of the new release branch. This matter can only be handled by Ververica 
employees since they are owning the AzureCI setup.
 

h3. Expectations (Minor Version only)
 * Release branch has been created and pushed
 * Changes on the new release branch are picked up by [Azure 
CI|https://dev.azure.com/apache-flink/apache-flink/_build?definitionId=1&_a=summary]
 * {{master}} branch has the version information updated to the new version 
(check pom.xml files and 
 * 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum)
 * New version is added to the 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum.
 * Make sure [flink-docker|https://github.com/apache/flink-docker/] has 
{{dev-x.y}} branch and docker e2e tests run against this branch in the 
corresponding Apache Flink release branch (see 
[apache/flink:flink-end-to-end-tests/test-scripts/common_docker.sh:51|https://github.com/apache/flink/blob/master/flink-end-to-end-tests/test-scripts/common_docker.sh#L51])
 * 
[apache-flink:docs/config.toml|https://github.com/apache/flink/blob/release-1.17/docs/config.toml]
 has been updated appropriately in the new Apache Flink release branch.
 * The {{flink.version}} property (see 
[apache/flink-benchmarks:pom.xml|https://github.com/apache/flink-benchmarks/blob/master/pom.xml#L48]
 of Flink Benchmark repo has been updated to the latest snapshot version.

  was:
If you are doing a new minor release, you need to update Flink version in the 
following repositories:
 * [apache/

[jira] [Closed] (FLINK-31036) StateCheckpointedITCase timed out

2023-02-21 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31036?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan closed FLINK-31036.
---
Resolution: Fixed

> StateCheckpointedITCase timed out
> -
>
> Key: FLINK-31036
> URL: https://issues.apache.org/jira/browse/FLINK-31036
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing, Tests
>Affects Versions: 1.17.0
>Reporter: Matthias Pohl
>Assignee: Rui Fan
>Priority: Major
>  Labels: pull-request-available, test-stability
> Fix For: 1.17.0, 1.18.0
>
> Attachments: image-2023-02-16-20-29-52-050.png
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=46023&view=logs&j=39d5b1d5-3b41-54dc-6458-1e2ddd1cdcf3&t=0c010d0c-3dec-5bf1-d408-7b18988b1b2b&l=10608
> {code}
> "Legacy Source Thread - Source: Custom Source -> Filter (6/12)#69980" 
> #13718026 prio=5 os_prio=0 tid=0x7f05f44f0800 nid=0x128157 waiting on 
> condition [0x7f059feef000]
>java.lang.Thread.State: WAITING (parking)
>   at sun.misc.Unsafe.park(Native Method)
>   - parking to wait for  <0xf0a974e8> (a 
> java.util.concurrent.CompletableFuture$Signaller)
>   at java.util.concurrent.locks.LockSupport.park(LockSupport.java:175)
>   at 
> java.util.concurrent.CompletableFuture$Signaller.block(CompletableFuture.java:1707)
>   at 
> java.util.concurrent.ForkJoinPool.managedBlock(ForkJoinPool.java:3323)
>   at 
> java.util.concurrent.CompletableFuture.waitingGet(CompletableFuture.java:1742)
>   at 
> java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)
>   at 
> org.apache.flink.runtime.io.network.buffer.LocalBufferPool.requestMemorySegmentBlocking(LocalBufferPool.java:384)
>   at 
> org.apache.flink.runtime.io.network.buffer.LocalBufferPool.requestBufferBuilderBlocking(LocalBufferPool.java:356)
>   at 
> org.apache.flink.runtime.io.network.partition.BufferWritingResultPartition.requestNewBufferBuilderFromPool(BufferWritingResultPartition.java:414)
>   at 
> org.apache.flink.runtime.io.network.partition.BufferWritingResultPartition.requestNewUnicastBufferBuilder(BufferWritingResultPartition.java:390)
>   at 
> org.apache.flink.runtime.io.network.partition.BufferWritingResultPartition.appendUnicastDataForRecordContinuation(BufferWritingResultPartition.java:328)
>   at 
> org.apache.flink.runtime.io.network.partition.BufferWritingResultPartition.emitRecord(BufferWritingResultPartition.java:161)
>   at 
> org.apache.flink.runtime.io.network.api.writer.RecordWriter.emit(RecordWriter.java:107)
>   at 
> org.apache.flink.runtime.io.network.api.writer.ChannelSelectorRecordWriter.emit(ChannelSelectorRecordWriter.java:55)
>   at 
> org.apache.flink.streaming.runtime.io.RecordWriterOutput.pushToRecordWriter(RecordWriterOutput.java:105)
>   at 
> org.apache.flink.streaming.runtime.io.RecordWriterOutput.collect(RecordWriterOutput.java:91)
>   at 
> org.apache.flink.streaming.runtime.io.RecordWriterOutput.collect(RecordWriterOutput.java:45)
>   at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:59)
>   at 
> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:31)
>   at 
> org.apache.flink.streaming.api.operators.StreamFilter.processElement(StreamFilter.java:39)
>   at 
> org.apache.flink.streaming.runtime.io.RecordProcessorUtils$$Lambda$1311/1256184070.accept(Unknown
>  Source)
>   at 
> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.pushToOperator(CopyingChainingOutput.java:75)
>   at 
> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:50)
>   at 
> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:29)
>   at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$ManualWatermarkContext.processAndCollect(StreamSourceContexts.java:418)
>   at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$WatermarkContext.collect(StreamSourceContexts.java:513)
>   - locked <0xd55035c0> (a java.lang.Object)
>   at 
> org.apache.flink.streaming.api.operators.StreamSourceContexts$SwitchingOnClose.collect(StreamSourceContexts.java:103)
>   at 
> org.apache.flink.test.checkpointing.StateCheckpointedITCase$StringGeneratingSourceFunction.run(StateCheckpointedITCase.java:178)
>   - locked <0xd55035c0> (a java.lang.Object)
>   at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:110)
>   at 
> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:67)
>   at 
> org.apache.flink.streaming.runtime.tasks.SourceStreamTask$LegacySourceFunctionThread.run(S

[GitHub] [flink] luoyuxia commented on a diff in pull request #21977: [FLINK-31132] Compact without setting parallelism does not follow the configured sink parallelism for HiveTableSink

2023-02-21 Thread via GitHub


luoyuxia commented on code in PR #21977:
URL: https://github.com/apache/flink/pull/21977#discussion_r1112798182


##
flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableCompactSinkTest.java:
##
@@ -18,178 +18,174 @@
 
 package org.apache.flink.connectors.hive;
 
-import org.apache.flink.api.common.RuntimeExecutionMode;
-import org.apache.flink.configuration.Configuration;
-import org.apache.flink.connector.file.table.FileSystemConnectorOptions;
+import org.apache.flink.api.dag.Transformation;
 import org.apache.flink.connector.file.table.batch.BatchSink;
-import org.apache.flink.sql.parser.hive.ddl.SqlCreateHiveTable;
-import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
-import org.apache.flink.streaming.api.graph.StreamGraph;
-import org.apache.flink.streaming.api.graph.StreamNode;
-import org.apache.flink.table.api.DataTypes;
-import org.apache.flink.table.api.TableSchema;
-import org.apache.flink.table.catalog.CatalogDatabaseImpl;
-import org.apache.flink.table.catalog.CatalogTable;
-import org.apache.flink.table.catalog.CatalogTableImpl;
-import org.apache.flink.table.catalog.CatalogTest;
-import org.apache.flink.table.catalog.Column;
-import org.apache.flink.table.catalog.ObjectIdentifier;
-import org.apache.flink.table.catalog.ObjectPath;
-import org.apache.flink.table.catalog.ResolvedCatalogTable;
-import org.apache.flink.table.catalog.ResolvedSchema;
+import org.apache.flink.table.api.SqlDialect;
+import org.apache.flink.table.api.TableEnvironment;
+import org.apache.flink.table.api.internal.TableEnvironmentImpl;
 import org.apache.flink.table.catalog.hive.HiveCatalog;
 import org.apache.flink.table.catalog.hive.HiveTestUtils;
-import org.apache.flink.table.connector.sink.DataStreamSinkProvider;
-import org.apache.flink.table.connector.sink.DynamicTableSink;
-import org.apache.flink.table.data.binary.BinaryRowData;
-import org.apache.flink.table.factories.DynamicTableSinkFactory;
-import org.apache.flink.table.factories.FactoryUtil;
-import 
org.apache.flink.table.runtime.connector.sink.SinkRuntimeProviderContext;
+import org.apache.flink.table.operations.ModifyOperation;
+import org.apache.flink.table.operations.Operation;
+import org.apache.flink.table.planner.delegation.PlannerBase;
 import org.apache.flink.util.TestLoggerExtension;
 
-import org.junit.jupiter.api.AfterAll;
-import org.junit.jupiter.api.BeforeAll;
+import org.junit.jupiter.api.AfterEach;
 import org.junit.jupiter.api.BeforeEach;
 import org.junit.jupiter.api.Test;
 import org.junit.jupiter.api.extension.ExtendWith;
 
 import java.util.Collections;
-import java.util.HashMap;
-import java.util.Map;
-import java.util.Optional;
+import java.util.List;
 
+import static 
org.apache.flink.connector.file.table.FileSystemConnectorOptions.COMPACTION_PARALLELISM;
+import static 
org.apache.flink.connector.file.table.FileSystemConnectorOptions.SINK_PARALLELISM;
 import static org.assertj.core.api.Assertions.assertThat;
 
 /** Tests for {@link HiveTableSink} enable auto-compaction. */
 @ExtendWith(TestLoggerExtension.class)
 class HiveTableCompactSinkTest {

Review Comment:
   Since it's mainly for test parallelism, how about rename it to 
`HiveTableCompactSinkParallelismTest`?
   Also the java doc for this class should update.



##
flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableCompactSinkTest.java:
##
@@ -18,178 +18,174 @@
 
 package org.apache.flink.connectors.hive;
 
-import org.apache.flink.api.common.RuntimeExecutionMode;
-import org.apache.flink.configuration.Configuration;
-import org.apache.flink.connector.file.table.FileSystemConnectorOptions;
+import org.apache.flink.api.dag.Transformation;
 import org.apache.flink.connector.file.table.batch.BatchSink;
-import org.apache.flink.sql.parser.hive.ddl.SqlCreateHiveTable;
-import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
-import org.apache.flink.streaming.api.graph.StreamGraph;
-import org.apache.flink.streaming.api.graph.StreamNode;
-import org.apache.flink.table.api.DataTypes;
-import org.apache.flink.table.api.TableSchema;
-import org.apache.flink.table.catalog.CatalogDatabaseImpl;
-import org.apache.flink.table.catalog.CatalogTable;
-import org.apache.flink.table.catalog.CatalogTableImpl;
-import org.apache.flink.table.catalog.CatalogTest;
-import org.apache.flink.table.catalog.Column;
-import org.apache.flink.table.catalog.ObjectIdentifier;
-import org.apache.flink.table.catalog.ObjectPath;
-import org.apache.flink.table.catalog.ResolvedCatalogTable;
-import org.apache.flink.table.catalog.ResolvedSchema;
+import org.apache.flink.table.api.SqlDialect;
+import org.apache.flink.table.api.TableEnvironment;
+import org.apache.flink.table.api.internal.TableEnvironmentImpl;
 import org.apache.flink.table.catalog.hive.HiveCatalog;
 import org.apache.flink.table.catalog.hive.HiveTestUtils;
-import org.apache.flink.

[GitHub] [flink] luoyuxia commented on a diff in pull request #21977: [FLINK-31132] Compact without setting parallelism does not follow the configured sink parallelism for HiveTableSink

2023-02-21 Thread via GitHub


luoyuxia commented on code in PR #21977:
URL: https://github.com/apache/flink/pull/21977#discussion_r1112838547


##
flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableCompactSinkTest.java:
##
@@ -18,178 +18,174 @@
 
 package org.apache.flink.connectors.hive;
 
-import org.apache.flink.api.common.RuntimeExecutionMode;
-import org.apache.flink.configuration.Configuration;
-import org.apache.flink.connector.file.table.FileSystemConnectorOptions;
+import org.apache.flink.api.dag.Transformation;
 import org.apache.flink.connector.file.table.batch.BatchSink;
-import org.apache.flink.sql.parser.hive.ddl.SqlCreateHiveTable;
-import org.apache.flink.streaming.api.environment.StreamExecutionEnvironment;
-import org.apache.flink.streaming.api.graph.StreamGraph;
-import org.apache.flink.streaming.api.graph.StreamNode;
-import org.apache.flink.table.api.DataTypes;
-import org.apache.flink.table.api.TableSchema;
-import org.apache.flink.table.catalog.CatalogDatabaseImpl;
-import org.apache.flink.table.catalog.CatalogTable;
-import org.apache.flink.table.catalog.CatalogTableImpl;
-import org.apache.flink.table.catalog.CatalogTest;
-import org.apache.flink.table.catalog.Column;
-import org.apache.flink.table.catalog.ObjectIdentifier;
-import org.apache.flink.table.catalog.ObjectPath;
-import org.apache.flink.table.catalog.ResolvedCatalogTable;
-import org.apache.flink.table.catalog.ResolvedSchema;
+import org.apache.flink.table.api.SqlDialect;
+import org.apache.flink.table.api.TableEnvironment;
+import org.apache.flink.table.api.internal.TableEnvironmentImpl;
 import org.apache.flink.table.catalog.hive.HiveCatalog;
 import org.apache.flink.table.catalog.hive.HiveTestUtils;
-import org.apache.flink.table.connector.sink.DataStreamSinkProvider;
-import org.apache.flink.table.connector.sink.DynamicTableSink;
-import org.apache.flink.table.data.binary.BinaryRowData;
-import org.apache.flink.table.factories.DynamicTableSinkFactory;
-import org.apache.flink.table.factories.FactoryUtil;
-import 
org.apache.flink.table.runtime.connector.sink.SinkRuntimeProviderContext;
+import org.apache.flink.table.operations.ModifyOperation;
+import org.apache.flink.table.operations.Operation;
+import org.apache.flink.table.planner.delegation.PlannerBase;
 import org.apache.flink.util.TestLoggerExtension;
 
-import org.junit.jupiter.api.AfterAll;
-import org.junit.jupiter.api.BeforeAll;
+import org.junit.jupiter.api.AfterEach;
 import org.junit.jupiter.api.BeforeEach;
 import org.junit.jupiter.api.Test;
 import org.junit.jupiter.api.extension.ExtendWith;
 
 import java.util.Collections;
-import java.util.HashMap;
-import java.util.Map;
-import java.util.Optional;
+import java.util.List;
 
+import static 
org.apache.flink.connector.file.table.FileSystemConnectorOptions.COMPACTION_PARALLELISM;
+import static 
org.apache.flink.connector.file.table.FileSystemConnectorOptions.SINK_PARALLELISM;
 import static org.assertj.core.api.Assertions.assertThat;
 
 /** Tests for {@link HiveTableSink} enable auto-compaction. */
 @ExtendWith(TestLoggerExtension.class)
 class HiveTableCompactSinkTest {
+/**
+ * Represents the parallelism need not check, it should follow the setting 
of planer or auto

Review Comment:
   If my understand commented below is correct.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-31165) Over Agg: The window rank function without order by error in top N query

2023-02-21 Thread Martijn Visser (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31165?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691493#comment-17691493
 ] 

Martijn Visser commented on FLINK-31165:


[~godfrey] [~lincoln.86xy] Any thoughts on this one?

> Over Agg: The window rank function without order by error in top N query
> 
>
> Key: FLINK-31165
> URL: https://issues.apache.org/jira/browse/FLINK-31165
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Affects Versions: 1.16.0
>Reporter: P Rohan Kumar
>Priority: Major
>
>  
> {code:java}
> val env: StreamExecutionEnvironment = 
> StreamExecutionEnvironment.getExecutionEnvironment
> val tableEnv = StreamTableEnvironment.create(env)
> val td = TableDescriptor.forConnector("datagen").option("rows-per-second", 
> "10")
>   .option("number-of-rows", "10")
>   .schema(Schema
> .newBuilder()
> .column("NAME", DataTypes.VARCHAR(2147483647))
> .column("ROLLNO", DataTypes.DECIMAL(5, 0))
> .column("DOB", DataTypes.DATE())
> .column("CLASS", DataTypes.DECIMAL(2, 0))
> .column("SUBJECT", DataTypes.VARCHAR(2147483647))
> .build())
>   .build()
> val table = tableEnv.from(td)
> tableEnv.createTemporaryView("temp_table", table)
> val newTable = tableEnv.sqlQuery("select temp_table.*,cast('2022-01-01' as 
> date) SRC_NO from temp_table")
> tableEnv.createTemporaryView("temp_table2", newTable)
> val newTable2 = tableEnv.sqlQuery("select * from (select 
> NAME,ROLLNO,row_number() over (partition by NAME ORDER BY SRC_NO) AS rownum  
> from temp_table2 a) where rownum <= 1")
> tableEnv.toChangelogStream(newTable2).print()
> env.execute()
>  {code}
>  
>  
> I am getting the below error if I run the above code.
> I have already provided an order by column.
> If I change the order by column to some other column, such as "SUBJECT", then 
> the job runs fine.
>  
>  
> {code:java}
> Exception in thread "main" java.lang.RuntimeException: Error while applying 
> rule FlinkLogicalOverAggregateConverter(in:NONE,out:LOGICAL), args 
> [rel#245:LogicalWindow.NONE.any.None: 
> 0.[NONE].[NONE](input=RelSubset#244,window#0=window(partition {0} rows 
> between UNBOUNDED PRECEDING and CURRENT ROW aggs [ROW_NUMBER()]))]
>     at 
> org.apache.calcite.plan.volcano.VolcanoRuleCall.onMatch(VolcanoRuleCall.java:256)
>     at 
> org.apache.calcite.plan.volcano.IterativeRuleDriver.drive(IterativeRuleDriver.java:58)
>     at 
> org.apache.calcite.plan.volcano.VolcanoPlanner.findBestExp(VolcanoPlanner.java:510)
>     at org.apache.calcite.tools.Programs$RuleSetProgram.run(Programs.java:312)
>     at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkVolcanoProgram.optimize(FlinkVolcanoProgram.scala:62)
>     at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkChainedProgram.$anonfun$optimize$1(FlinkChainedProgram.scala:59)
>     at 
> scala.collection.TraversableOnce$folder$1$.apply(TraversableOnce.scala:187)
>     at 
> scala.collection.TraversableOnce$folder$1$.apply(TraversableOnce.scala:185)
>     at scala.collection.Iterator.foreach(Iterator.scala:943)
>     at scala.collection.Iterator.foreach$(Iterator.scala:943)
>     at scala.collection.AbstractIterator.foreach(Iterator.scala:1431)
>     at scala.collection.IterableLike.foreach(IterableLike.scala:74)
>     at scala.collection.IterableLike.foreach$(IterableLike.scala:73)
>     at scala.collection.AbstractIterable.foreach(Iterable.scala:56)
>     at scala.collection.TraversableOnce.foldLeft(TraversableOnce.scala:189)
>     at scala.collection.TraversableOnce.foldLeft$(TraversableOnce.scala:184)
>     at scala.collection.AbstractTraversable.foldLeft(Traversable.scala:108)
>     at 
> org.apache.flink.table.planner.plan.optimize.program.FlinkChainedProgram.optimize(FlinkChainedProgram.scala:55)
>     at 
> org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.optimizeTree(StreamCommonSubGraphBasedOptimizer.scala:176)
>     at 
> org.apache.flink.table.planner.plan.optimize.StreamCommonSubGraphBasedOptimizer.doOptimize(StreamCommonSubGraphBasedOptimizer.scala:83)
>     at 
> org.apache.flink.table.planner.plan.optimize.CommonSubGraphBasedOptimizer.optimize(CommonSubGraphBasedOptimizer.scala:87)
>     at 
> org.apache.flink.table.planner.delegation.PlannerBase.optimize(PlannerBase.scala:315)
>     at 
> org.apache.flink.table.planner.delegation.PlannerBase.translate(PlannerBase.scala:195)
>     at 
> org.apache.flink.table.api.bridge.internal.AbstractStreamTableEnvironmentImpl.toStreamInternal(AbstractStreamTableEnvironmentImpl.java:224)
>     at 
> org.apache.flink.table.api.bridge.internal.AbstractStreamTableEnvironmentImpl.toStreamInternal(AbstractStreamTableEnvironmentImpl.java:219)
>     at 
> org.apache.flink.table.api.bridge.scala

[GitHub] [flink] snuyanzin commented on a diff in pull request #21945: [FLINK-31091][Table SQL/Client] Add Ser/de for Interval types

2023-02-21 Thread via GitHub


snuyanzin commented on code in PR #21945:
URL: https://github.com/apache/flink/pull/21945#discussion_r1112846597


##
flink-table/flink-sql-gateway/src/main/java/org/apache/flink/table/gateway/rest/serde/LogicalTypeJsonSerializer.java:
##
@@ -87,6 +89,10 @@ public final class LogicalTypeJsonSerializer extends 
StdSerializer
 public static final String FIELD_NAME_CLASS = "class";
 public static final String FIELD_NAME_SERIALIZER = "serializer";
 
+// INTERVAL

Review Comment:
   updated usage of `precision` with usage of `yearPrecision` and 
`dayPrecision` based on offline discussion in slack



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[GitHub] [flink] reswqa commented on a diff in pull request #21977: [FLINK-31132] Compact without setting parallelism does not follow the configured sink parallelism for HiveTableSink

2023-02-21 Thread via GitHub


reswqa commented on code in PR #21977:
URL: https://github.com/apache/flink/pull/21977#discussion_r1112847462


##
flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableCompactSinkTest.java:
##
@@ -0,0 +1,191 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.connectors.hive;
+
+import org.apache.flink.api.dag.Transformation;
+import org.apache.flink.connector.file.table.batch.BatchSink;
+import org.apache.flink.table.api.SqlDialect;
+import org.apache.flink.table.api.TableEnvironment;
+import org.apache.flink.table.api.internal.TableEnvironmentImpl;
+import org.apache.flink.table.catalog.hive.HiveCatalog;
+import org.apache.flink.table.catalog.hive.HiveTestUtils;
+import org.apache.flink.table.operations.ModifyOperation;
+import org.apache.flink.table.operations.Operation;
+import org.apache.flink.table.planner.delegation.PlannerBase;
+import org.apache.flink.util.TestLoggerExtension;
+
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.extension.ExtendWith;
+
+import java.util.Collections;
+import java.util.List;
+
+import static 
org.apache.flink.connector.file.table.FileSystemConnectorOptions.COMPACTION_PARALLELISM;
+import static 
org.apache.flink.connector.file.table.FileSystemConnectorOptions.SINK_PARALLELISM;
+import static org.assertj.core.api.Assertions.assertThat;
+
+/** Tests for {@link HiveTableSink} enable auto-compaction. */
+@ExtendWith(TestLoggerExtension.class)
+class HiveTableCompactSinkTest {
+/**
+ * Represents the parallelism need not check, it should follow the setting 
of planer or auto
+ * inference.
+ */
+public static final int NEED_NOT_CHECK_PARALLELISM = -1;
+
+private HiveCatalog catalog;
+
+private TableEnvironment tableEnv;
+
+@BeforeEach
+void before() {
+catalog = HiveTestUtils.createHiveCatalog();
+catalog.open();
+tableEnv = HiveTestUtils.createTableEnvInBatchMode(SqlDialect.HIVE);
+tableEnv.registerCatalog(catalog.getName(), catalog);
+tableEnv.useCatalog(catalog.getName());
+}
+
+@AfterEach
+void after() {
+if (catalog != null) {
+catalog.close();
+}
+}
+
+/** If only sink parallelism is set, compact operator should follow this 
setting. */
+@Test
+void testOnlySetSinkParallelism() {
+final int sinkParallelism = 4;
+
+tableEnv.executeSql(
+String.format(
+"CREATE TABLE src ("
++ " key string,"
++ " value string"
++ ") TBLPROPERTIES ("
++ " 'auto-compaction' = 'true', "
++ " '%s' = '%s' )",
+SINK_PARALLELISM.key(), sinkParallelism));
+
+assertSinkAndCompactOperatorParallelism(true, true, sinkParallelism, 
sinkParallelism);
+}
+
+@Test
+void testOnlySetCompactParallelism() {
+final int compactParallelism = 4;
+
+tableEnv.executeSql(
+String.format(
+"CREATE TABLE src ("
++ " key string,"
++ " value string"
++ ") TBLPROPERTIES ("
++ " 'auto-compaction' = 'true', "
++ " '%s' = '%s' )",
+COMPACTION_PARALLELISM.key(), compactParallelism));
+
+assertSinkAndCompactOperatorParallelism(
+false, true, NEED_NOT_CHECK_PARALLELISM, compactParallelism);
+}
+
+@Test
+void testSetBothSinkAndCompactParallelism() {
+final int sinkParallelism = 8;
+final int compactParallelism = 4;
+
+tableEnv.executeSql(
+String.format(
+"CREATE TABLE src ("
++ " key string,"
++ " value string"
++ ") TBLPROPERTIES ("
++ 

[GitHub] [flink] MartijnVisser commented on pull request #21965: [FLINK-31121] Support discarding too large records in kafka sink

2023-02-21 Thread via GitHub


MartijnVisser commented on PR #21965:
URL: https://github.com/apache/flink/pull/21965#issuecomment-1438209724

   Can we hold off merging this PR since we're in the process of externalizing 
the Kafka connector, which means that this PR should then be moved to the 
external Kafka connector repository?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31153:
--
Fix Version/s: 1.17.0

> Create a release branch
> ---
>
> Key: FLINK-31153
> URL: https://issues.apache.org/jira/browse/FLINK-31153
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Assignee: Leonard Xu
>Priority: Major
> Fix For: 1.17.0
>
>
> If you are doing a new minor release, you need to update Flink version in the 
> following repositories and the [AzureCI project 
> configuration|https://dev.azure.com/apache-flink/apache-flink/]:
>  * [apache/flink|https://github.com/apache/flink]
>  * [apache/flink-docker|https://github.com/apache/flink-docker]
>  * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]
> Patch releases don't require the these repositories to be touched. Simply 
> checkout the already existing branch for that version:
> {code:java}
> $ cd ./tools
> $ git checkout release-$SHORT_RELEASE_VERSION
> {code}
> h4. Flink repository
> Create a branch for the new version that we want to release before updating 
> the master branch to the next development version:
> {code:bash}
> $ cd ./tools
> $ releasing/create_snapshot_branch.sh
> $ git checkout master
> $ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
> releasing/update_branch_version.sh
> {code}
> In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  as the last entry:
> {code:java}
> // ...
> v1_12("1.12"),
> v1_13("1.13"),
> v1_14("1.14"),
> v1_15("1.15"),
> v1_16("1.16");
> {code}
> The newly created branch and updated {{master}} branch need to be pushed to 
> the official repository.
> h4. Flink Docker Repository
> Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
> [apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
> sure that 
> [apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
>  points to the correct snapshot version; for {{dev-x.y}} it should point to 
> {{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
> recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).
> After pushing the new minor release branch, as the last step you should also 
> update the documentation workflow to also build the documentation for the new 
> release branch. Check [Managing 
> Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
>  on details on how to do that. You may also want to manually trigger a build 
> to make the changes visible as soon as possible.
> h4. Flink Benchmark Repository
> First of all, checkout the {{master}} branch to {{dev-x.y}} branch in 
> [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks], so that 
> we can have a branch named {{dev-x.y}} which could be built on top of 
> (${{CURRENT_SNAPSHOT_VERSION}}).
> Then, inside the repository you need to manually update the {{flink.version}} 
> property inside the parent *pom.xml* file. It should be pointing to the most 
> recent snapshot version ($NEXT_SNAPSHOT_VERSION). For example:
> {code:xml}
> 1.18-SNAPSHOT
> {code}
> h4. AzureCI Project Configuration
> The new release branch needs to be configured within AzureCI to make azure 
> aware of the new release branch. This matter can only be handled by Ververica 
> employees since they are owning the AzureCI setup.
>  
> 
> h3. Expectations (Minor Version only)
>  * Release branch has been created and pushed
>  * Changes on the new release branch are picked up by [Azure 
> CI|https://dev.azure.com/apache-flink/apache-flink/_build?definitionId=1&_a=summary]
>  * {{master}} branch has the version information updated to the new version 
> (check pom.xml files and 
>  * 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  enum)
>  * New version is added to the 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  enum.
>  * Make sure [flink-docker|https://github.com/apache/flink-docker/] has 
> {{dev-x.y}} branch and docker e2e tests run against this branch in the 
> corresponding Apache Flink release branch (see 
> [apache/flink:flink-end-to-end-tests/test-scripts/common_docker.sh:51|https://github.com/apache/flink/blob/master/flink-end-to-end-tests/test-scripts/common_docke

[jira] [Updated] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31153:
--
Affects Version/s: 1.17.0

> Create a release branch
> ---
>
> Key: FLINK-31153
> URL: https://issues.apache.org/jira/browse/FLINK-31153
> Project: Flink
>  Issue Type: Sub-task
>Affects Versions: 1.17.0
>Reporter: Matthias Pohl
>Assignee: Leonard Xu
>Priority: Major
> Fix For: 1.17.0
>
>
> If you are doing a new minor release, you need to update Flink version in the 
> following repositories and the [AzureCI project 
> configuration|https://dev.azure.com/apache-flink/apache-flink/]:
>  * [apache/flink|https://github.com/apache/flink]
>  * [apache/flink-docker|https://github.com/apache/flink-docker]
>  * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]
> Patch releases don't require the these repositories to be touched. Simply 
> checkout the already existing branch for that version:
> {code:java}
> $ cd ./tools
> $ git checkout release-$SHORT_RELEASE_VERSION
> {code}
> h4. Flink repository
> Create a branch for the new version that we want to release before updating 
> the master branch to the next development version:
> {code:bash}
> $ cd ./tools
> $ releasing/create_snapshot_branch.sh
> $ git checkout master
> $ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
> releasing/update_branch_version.sh
> {code}
> In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  as the last entry:
> {code:java}
> // ...
> v1_12("1.12"),
> v1_13("1.13"),
> v1_14("1.14"),
> v1_15("1.15"),
> v1_16("1.16");
> {code}
> The newly created branch and updated {{master}} branch need to be pushed to 
> the official repository.
> h4. Flink Docker Repository
> Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
> [apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
> sure that 
> [apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
>  points to the correct snapshot version; for {{dev-x.y}} it should point to 
> {{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
> recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).
> After pushing the new minor release branch, as the last step you should also 
> update the documentation workflow to also build the documentation for the new 
> release branch. Check [Managing 
> Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
>  on details on how to do that. You may also want to manually trigger a build 
> to make the changes visible as soon as possible.
> h4. Flink Benchmark Repository
> First of all, checkout the {{master}} branch to {{dev-x.y}} branch in 
> [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks], so that 
> we can have a branch named {{dev-x.y}} which could be built on top of 
> (${{CURRENT_SNAPSHOT_VERSION}}).
> Then, inside the repository you need to manually update the {{flink.version}} 
> property inside the parent *pom.xml* file. It should be pointing to the most 
> recent snapshot version ($NEXT_SNAPSHOT_VERSION). For example:
> {code:xml}
> 1.18-SNAPSHOT
> {code}
> h4. AzureCI Project Configuration
> The new release branch needs to be configured within AzureCI to make azure 
> aware of the new release branch. This matter can only be handled by Ververica 
> employees since they are owning the AzureCI setup.
>  
> 
> h3. Expectations (Minor Version only)
>  * Release branch has been created and pushed
>  * Changes on the new release branch are picked up by [Azure 
> CI|https://dev.azure.com/apache-flink/apache-flink/_build?definitionId=1&_a=summary]
>  * {{master}} branch has the version information updated to the new version 
> (check pom.xml files and 
>  * 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  enum)
>  * New version is added to the 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  enum.
>  * Make sure [flink-docker|https://github.com/apache/flink-docker/] has 
> {{dev-x.y}} branch and docker e2e tests run against this branch in the 
> corresponding Apache Flink release branch (see 
> [apache/flink:flink-end-to-end-tests/test-scripts/common_docker.sh:51|https://github.com/apache/flink/blob/master/flink-end-to-e

[jira] [Resolved] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl resolved FLINK-31153.
---
Resolution: Fixed

* (/) Release branch has been created and pushed
 * (/) Changes on the new release branch are picked up by [Azure 
CI|https://dev.azure.com/apache-flink/apache-flink/_build?definitionId=1&_a=summary]
 * {{(/) master}} branch has the version information updated to the new version 
(check pom.xml files)
 ** Related commit is 
[923361a3|https://github.com/apache/flink/commit/923361a3006c7c4ce8acbc74a9b756f93c93020a]
 * (/) New version is added to the 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum.
 ** Related commit is [d76c49b4|https://github.com/apache/flink/commit/d76c49b4]
 * (/) Make sure [flink-docker|https://github.com/apache/flink-docker/] has 
{{dev-x.y}} branch
 * (/) Docker e2e tests run against the new flink-docker branch in the 
corresponding Apache Flink release branch (see 
[apache/flink:flink-end-to-end-tests/test-scripts/common_docker.sh:51|https://github.com/apache/flink/blob/master/flink-end-to-end-tests/test-scripts/common_docker.sh#L51])
 ** Related commit is [df4d2ff7|https://github.com/apache/flink/commit/df4d2ff7]
 * [(/) 
apache-flink:docs/config.toml|https://github.com/apache/flink/blob/release-1.17/docs/config.toml]
 has been updated appropriately in the new Apache Flink release branch.
 ** Related commit is [df4d2ff7|https://github.com/apache/flink/commit/df4d2ff7]
 * The {{flink.version}} property (see 
[apache/flink-benchmarks:pom.xml|https://github.com/apache/flink-benchmarks/blob/master/pom.xml#L48]
 of Flink Benchmark repo has been updated to the latest snapshot version.
 ** Related commit is 
[79fdbaf6|https://github.com/apache/flink-benchmarks/commit/79fdbaf6f7ae2da63d08374b3052422afcec34ea]

> Create a release branch
> ---
>
> Key: FLINK-31153
> URL: https://issues.apache.org/jira/browse/FLINK-31153
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Assignee: Leonard Xu
>Priority: Major
>
> If you are doing a new minor release, you need to update Flink version in the 
> following repositories and the [AzureCI project 
> configuration|https://dev.azure.com/apache-flink/apache-flink/]:
>  * [apache/flink|https://github.com/apache/flink]
>  * [apache/flink-docker|https://github.com/apache/flink-docker]
>  * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]
> Patch releases don't require the these repositories to be touched. Simply 
> checkout the already existing branch for that version:
> {code:java}
> $ cd ./tools
> $ git checkout release-$SHORT_RELEASE_VERSION
> {code}
> h4. Flink repository
> Create a branch for the new version that we want to release before updating 
> the master branch to the next development version:
> {code:bash}
> $ cd ./tools
> $ releasing/create_snapshot_branch.sh
> $ git checkout master
> $ OLD_VERSION=$CURRENT_SNAPSHOT_VERSION NEW_VERSION=$NEXT_SNAPSHOT_VERSION 
> releasing/update_branch_version.sh
> {code}
> In the {{master}} branch, add a new value (e.g. {{{}v1_16("1.16"){}}}) to 
> [apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
>  as the last entry:
> {code:java}
> // ...
> v1_12("1.12"),
> v1_13("1.13"),
> v1_14("1.14"),
> v1_15("1.15"),
> v1_16("1.16");
> {code}
> The newly created branch and updated {{master}} branch need to be pushed to 
> the official repository.
> h4. Flink Docker Repository
> Afterwards fork off from {{dev-master}} a {{dev-x.y}} branch in the 
> [apache/flink-docker|https://github.com/apache/flink-docker] repository. Make 
> sure that 
> [apache/flink-docker:.github/workflows/ci.yml|https://github.com/apache/flink-docker/blob/dev-master/.github/workflows/ci.yml]
>  points to the correct snapshot version; for {{dev-x.y}} it should point to 
> {{{}x.y-SNAPSHOT{}}}, while for {{dev-master}} it should point to the most 
> recent snapshot version (\{[$NEXT_SNAPSHOT_VERSION}}).
> After pushing the new minor release branch, as the last step you should also 
> update the documentation workflow to also build the documentation for the new 
> release branch. Check [Managing 
> Documentation|https://cwiki.apache.org/confluence/display/FLINK/Managing+Documentation]
>  on details on how to do that. You may also want to manually trigger a build 
> to make the changes visible as soon as possible.
> h4. Flink Benchmark Repository
> First of all, checkout the {{master}} branch to {{dev-x.y}} branch in 
> [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks], so that 
> we can hav

[jira] [Comment Edited] (FLINK-31153) Create a release branch

2023-02-21 Thread Matthias Pohl (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31153?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691495#comment-17691495
 ] 

Matthias Pohl edited comment on FLINK-31153 at 2/21/23 10:14 AM:
-

Closing this issue. All the expected artifacts are created:
* (/) Release branch has been created and pushed
 * (/) Changes on the new release branch are picked up by [Azure 
CI|https://dev.azure.com/apache-flink/apache-flink/_build?definitionId=1&_a=summary]
 * {{(/) master}} branch has the version information updated to the new version 
(check pom.xml files)
 ** Related commit is 
[923361a3|https://github.com/apache/flink/commit/923361a3006c7c4ce8acbc74a9b756f93c93020a]
 * (/) New version is added to the 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum.
 ** Related commit is [d76c49b4|https://github.com/apache/flink/commit/d76c49b4]
 * (/) Make sure [flink-docker|https://github.com/apache/flink-docker/] has 
{{dev-x.y}} branch
 * (/) Docker e2e tests run against the new flink-docker branch in the 
corresponding Apache Flink release branch (see 
[apache/flink:flink-end-to-end-tests/test-scripts/common_docker.sh:51|https://github.com/apache/flink/blob/master/flink-end-to-end-tests/test-scripts/common_docker.sh#L51])
 ** Related commit is [df4d2ff7|https://github.com/apache/flink/commit/df4d2ff7]
 * [(/) 
apache-flink:docs/config.toml|https://github.com/apache/flink/blob/release-1.17/docs/config.toml]
 has been updated appropriately in the new Apache Flink release branch.
 ** Related commit is [df4d2ff7|https://github.com/apache/flink/commit/df4d2ff7]
 * The {{flink.version}} property (see 
[apache/flink-benchmarks:pom.xml|https://github.com/apache/flink-benchmarks/blob/master/pom.xml#L48]
 of Flink Benchmark repo has been updated to the latest snapshot version.
 ** Related commit is 
[79fdbaf6|https://github.com/apache/flink-benchmarks/commit/79fdbaf6f7ae2da63d08374b3052422afcec34ea]


was (Author: mapohl):
* (/) Release branch has been created and pushed
 * (/) Changes on the new release branch are picked up by [Azure 
CI|https://dev.azure.com/apache-flink/apache-flink/_build?definitionId=1&_a=summary]
 * {{(/) master}} branch has the version information updated to the new version 
(check pom.xml files)
 ** Related commit is 
[923361a3|https://github.com/apache/flink/commit/923361a3006c7c4ce8acbc74a9b756f93c93020a]
 * (/) New version is added to the 
[apache-flink:flink-annotations/src/main/java/org/apache/flink/FlinkVersion|https://github.com/apache/flink/blob/master/flink-annotations/src/main/java/org/apache/flink/FlinkVersion.java]
 enum.
 ** Related commit is [d76c49b4|https://github.com/apache/flink/commit/d76c49b4]
 * (/) Make sure [flink-docker|https://github.com/apache/flink-docker/] has 
{{dev-x.y}} branch
 * (/) Docker e2e tests run against the new flink-docker branch in the 
corresponding Apache Flink release branch (see 
[apache/flink:flink-end-to-end-tests/test-scripts/common_docker.sh:51|https://github.com/apache/flink/blob/master/flink-end-to-end-tests/test-scripts/common_docker.sh#L51])
 ** Related commit is [df4d2ff7|https://github.com/apache/flink/commit/df4d2ff7]
 * [(/) 
apache-flink:docs/config.toml|https://github.com/apache/flink/blob/release-1.17/docs/config.toml]
 has been updated appropriately in the new Apache Flink release branch.
 ** Related commit is [df4d2ff7|https://github.com/apache/flink/commit/df4d2ff7]
 * The {{flink.version}} property (see 
[apache/flink-benchmarks:pom.xml|https://github.com/apache/flink-benchmarks/blob/master/pom.xml#L48]
 of Flink Benchmark repo has been updated to the latest snapshot version.
 ** Related commit is 
[79fdbaf6|https://github.com/apache/flink-benchmarks/commit/79fdbaf6f7ae2da63d08374b3052422afcec34ea]

> Create a release branch
> ---
>
> Key: FLINK-31153
> URL: https://issues.apache.org/jira/browse/FLINK-31153
> Project: Flink
>  Issue Type: Sub-task
>Affects Versions: 1.17.0
>Reporter: Matthias Pohl
>Assignee: Leonard Xu
>Priority: Major
> Fix For: 1.17.0
>
>
> If you are doing a new minor release, you need to update Flink version in the 
> following repositories and the [AzureCI project 
> configuration|https://dev.azure.com/apache-flink/apache-flink/]:
>  * [apache/flink|https://github.com/apache/flink]
>  * [apache/flink-docker|https://github.com/apache/flink-docker]
>  * [apache/flink-benchmarks|https://github.com/apache/flink-benchmarks]
> Patch releases don't require the these repositories to be touched. Simply 
> checkout the already existing branch for that version:
> {code:java}
> $ cd ./tools
> $ git checkout release-$SHORT_RELEASE_VERSION
> {code}
> h4. Fli

[GitHub] [flink] reswqa commented on a diff in pull request #21977: [FLINK-31132] Compact without setting parallelism does not follow the configured sink parallelism for HiveTableSink

2023-02-21 Thread via GitHub


reswqa commented on code in PR #21977:
URL: https://github.com/apache/flink/pull/21977#discussion_r1112847462


##
flink-connectors/flink-connector-hive/src/test/java/org/apache/flink/connectors/hive/HiveTableCompactSinkTest.java:
##
@@ -0,0 +1,191 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.flink.connectors.hive;
+
+import org.apache.flink.api.dag.Transformation;
+import org.apache.flink.connector.file.table.batch.BatchSink;
+import org.apache.flink.table.api.SqlDialect;
+import org.apache.flink.table.api.TableEnvironment;
+import org.apache.flink.table.api.internal.TableEnvironmentImpl;
+import org.apache.flink.table.catalog.hive.HiveCatalog;
+import org.apache.flink.table.catalog.hive.HiveTestUtils;
+import org.apache.flink.table.operations.ModifyOperation;
+import org.apache.flink.table.operations.Operation;
+import org.apache.flink.table.planner.delegation.PlannerBase;
+import org.apache.flink.util.TestLoggerExtension;
+
+import org.junit.jupiter.api.AfterEach;
+import org.junit.jupiter.api.BeforeEach;
+import org.junit.jupiter.api.Test;
+import org.junit.jupiter.api.extension.ExtendWith;
+
+import java.util.Collections;
+import java.util.List;
+
+import static 
org.apache.flink.connector.file.table.FileSystemConnectorOptions.COMPACTION_PARALLELISM;
+import static 
org.apache.flink.connector.file.table.FileSystemConnectorOptions.SINK_PARALLELISM;
+import static org.assertj.core.api.Assertions.assertThat;
+
+/** Tests for {@link HiveTableSink} enable auto-compaction. */
+@ExtendWith(TestLoggerExtension.class)
+class HiveTableCompactSinkTest {
+/**
+ * Represents the parallelism need not check, it should follow the setting 
of planer or auto
+ * inference.
+ */
+public static final int NEED_NOT_CHECK_PARALLELISM = -1;
+
+private HiveCatalog catalog;
+
+private TableEnvironment tableEnv;
+
+@BeforeEach
+void before() {
+catalog = HiveTestUtils.createHiveCatalog();
+catalog.open();
+tableEnv = HiveTestUtils.createTableEnvInBatchMode(SqlDialect.HIVE);
+tableEnv.registerCatalog(catalog.getName(), catalog);
+tableEnv.useCatalog(catalog.getName());
+}
+
+@AfterEach
+void after() {
+if (catalog != null) {
+catalog.close();
+}
+}
+
+/** If only sink parallelism is set, compact operator should follow this 
setting. */
+@Test
+void testOnlySetSinkParallelism() {
+final int sinkParallelism = 4;
+
+tableEnv.executeSql(
+String.format(
+"CREATE TABLE src ("
++ " key string,"
++ " value string"
++ ") TBLPROPERTIES ("
++ " 'auto-compaction' = 'true', "
++ " '%s' = '%s' )",
+SINK_PARALLELISM.key(), sinkParallelism));
+
+assertSinkAndCompactOperatorParallelism(true, true, sinkParallelism, 
sinkParallelism);
+}
+
+@Test
+void testOnlySetCompactParallelism() {
+final int compactParallelism = 4;
+
+tableEnv.executeSql(
+String.format(
+"CREATE TABLE src ("
++ " key string,"
++ " value string"
++ ") TBLPROPERTIES ("
++ " 'auto-compaction' = 'true', "
++ " '%s' = '%s' )",
+COMPACTION_PARALLELISM.key(), compactParallelism));
+
+assertSinkAndCompactOperatorParallelism(
+false, true, NEED_NOT_CHECK_PARALLELISM, compactParallelism);
+}
+
+@Test
+void testSetBothSinkAndCompactParallelism() {
+final int sinkParallelism = 8;
+final int compactParallelism = 4;
+
+tableEnv.executeSql(
+String.format(
+"CREATE TABLE src ("
++ " key string,"
++ " value string"
++ ") TBLPROPERTIES ("
++ 

[jira] [Updated] (FLINK-31150) Cross team testing

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31150:
--
Description: 
For user facing features that go into the release we'd like to ensure they can 
actually _be used_ by Flink users. To achieve this the release managers ensure 
that an issue for cross team testing is created in the Apache Flink Jira. This 
can and should be picked up by other community members to verify the 
functionality and usability of the feature.
The issue should contain some entry points which enables other community 
members to test it. It should not contain documentation on how to use the 
feature as this should be part of the actual documentation. The cross team 
tests are performed after the feature freeze. Documentation should be in place 
before that. Those tests are manual tests, so do not confuse them with 
automated tests.
To sum that up:
 * User facing features should be tested by other contributors
 * The scope is usability and sanity of the feature
 * The feature needs to be already documented
 * The contributor creates an issue containing some pointers on how to get 
started (e.g. link to the documentation, suggested targets of verification)
 * Other community members pick those issues up and provide feedback
 * Cross team testing happens right after the feature freeze




h3. Expectations
* Jira issues for each expected release task according to the release plan is 
created and labeled as {{release-testing}}.
* All the created release-testing-related Jira issues are resolved and the 
corresponding blocker issues are fixed.

  was:
For user facing features that go into the release we'd like to ensure they can 
actually _be used_ by Flink users. To achieve this the release managers ensure 
that an issue for cross team testing is created in the Apache Flink Jira. This 
can and should be picked up by other community members to verify the 
functionality and usability of the feature.
The issue should contain some entry points which enables other community 
members to test it. It should not contain documentation on how to use the 
feature as this should be part of the actual documentation. The cross team 
tests are performed after the feature freeze. Documentation should be in place 
before that. Those tests are manual tests, so do not confuse them with 
automated tests.
To sum that up:
 * User facing features should be tested by other contributors
 * The scope is usability and sanity of the feature
 * The feature needs to be already documented
 * The contributor creates an issue containing some pointers on how to get 
started (e.g. link to the documentation, suggested targets of verification)
 * Other community members pick those issues up and provide feedback
 * Cross team testing happens right after the feature freeze

--
h3. Expectations
* Jira issues for each expected release task according to the release plan is 
created and labeled as {{release-testing}}.
* All the created release-testing-related Jira issues are resolved and the 
corresponding blocker issues are fixed.


> Cross team testing
> --
>
> Key: FLINK-31150
> URL: https://issues.apache.org/jira/browse/FLINK-31150
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Assignee: Qingsheng Ren
>Priority: Major
>
> For user facing features that go into the release we'd like to ensure they 
> can actually _be used_ by Flink users. To achieve this the release managers 
> ensure that an issue for cross team testing is created in the Apache Flink 
> Jira. This can and should be picked up by other community members to verify 
> the functionality and usability of the feature.
> The issue should contain some entry points which enables other community 
> members to test it. It should not contain documentation on how to use the 
> feature as this should be part of the actual documentation. The cross team 
> tests are performed after the feature freeze. Documentation should be in 
> place before that. Those tests are manual tests, so do not confuse them with 
> automated tests.
> To sum that up:
>  * User facing features should be tested by other contributors
>  * The scope is usability and sanity of the feature
>  * The feature needs to be already documented
>  * The contributor creates an issue containing some pointers on how to get 
> started (e.g. link to the documentation, suggested targets of verification)
>  * Other community members pick those issues up and provide feedback
>  * Cross team testing happens right after the feature freeze
> 
> h3. Expectations
> * Jira issues for each expected release task according to the release plan is 
> created and labeled as {{release-testing}}.
> * All the created release-testing-related Jira issues are resolved and the 
> corresponding blocker issues are 

[jira] [Commented] (FLINK-30337) Stateful Functions application throws an exception when stopping a job gracefully creating a final savepoint

2023-02-21 Thread Alexey Novakov (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-30337?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691499#comment-17691499
 ] 

Alexey Novakov commented on FLINK-30337:


I have been able to reproduce this issue as well. 

Flink CLI Stop command fails with above exception after timeout and there is 
one more exception in *statefun-worker* pod:

 
{code:java}
2023-02-21 10:06:36,818 INFO  
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable [] - 
feedback-union -> functions (1/1)#0 - asynchronous part of checkpoint 2 could 
not be completed.
java.util.concurrent.CancellationException: null
    at java.util.concurrent.FutureTask.report(Unknown Source) ~[?:?]
    at java.util.concurrent.FutureTask.get(Unknown Source) ~[?:?]
    at 
org.apache.flink.util.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:645)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.api.operators.OperatorSnapshotFinalizer.(OperatorSnapshotFinalizer.java:57)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.finalizeNonFinishedSnapshots(AsyncCheckpointRunnable.java:177)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.run(AsyncCheckpointRunnable.java:124)
 [flink-dist_2.12-1.14.3.jar:1.14.3]
    at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) [?:?]
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) [?:?]
    at java.lang.Thread.run(Unknown Source) [?:?] {code}
 

 

> Stateful Functions application throws an exception when stopping a job 
> gracefully creating a final savepoint
> 
>
> Key: FLINK-30337
> URL: https://issues.apache.org/jira/browse/FLINK-30337
> Project: Flink
>  Issue Type: Bug
>  Components: Stateful Functions
>Affects Versions: statefun-3.2.0
>Reporter: Ali Bahadir Zeybek
>Priority: Minor
>
> When running a Stateful Functions applications, if the stop[1] command is 
> executed, the client throws a FlinkException with the following stack trace 
> where *953498833da99ec437758b49b7d5befd* is the specific job id:
>  
> {code:java}
> The program finished with the following 
> exception:org.apache.flink.util.FlinkException: Could not stop with a 
> savepoint job "953498833da99ec437758b49b7d5befd".
>     at 
> org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java:581)
>     at 
> org.apache.flink.client.cli.CliFrontend.runClusterAction(CliFrontend.java:1002)
>     at org.apache.flink.client.cli.CliFrontend.stop(CliFrontend.java:569)
>     at 
> org.apache.flink.client.cli.CliFrontend.parseAndRun(CliFrontend.java:1069)
>     at 
> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:1132)
>     at 
> org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
>     at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1132)
> Caused by: java.util.concurrent.TimeoutException
>     at java.base/java.util.concurrent.CompletableFuture.timedGet(Unknown 
> Source)
>     at java.base/java.util.concurrent.CompletableFuture.get(Unknown Source)
>     at 
> org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java:579)
>     ... 6 more {code}
>  
> How to reproduce:
>  # Follow the README[2] of the k8s deployment example of the 
> *flink-statefun-playground* project to have a running application
>  # Open the Flink UI that is started to get the *JOB_ID*
>  # Detect the *STATEFUN_MASTER_POD_NAME* by running: *kubectl get pods 
> --namespace statefun*
>  # Start a shell into the *statefun-master* pod by issuing the: *kubectl exec 
> -it --namespace statefun $STATEFUN_MASTER_POD_NAME – /bin/bash*
>  # Run the stop command: *./bin/flink stop --savepointPath 
> /tmp/flink-savepoints $JOB_ID*
>  
> [1]: 
> [https://nightlies.apache.org/flink/flink-docs-release-1.15/docs/deployment/cli/#stopping-a-job-gracefully-creating-a-final-savepoint]
> [2]: 
> [https://github.com/apache/flink-statefun-playground/blob/main/deployments/k8s/README.md]
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31150) Cross team testing

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31150:
--
Description: 
For user facing features that go into the release we'd like to ensure they can 
actually _be used_ by Flink users. To achieve this the release managers ensure 
that an issue for cross team testing is created in the Apache Flink Jira. This 
can and should be picked up by other community members to verify the 
functionality and usability of the feature.
The issue should contain some entry points which enables other community 
members to test it. It should not contain documentation on how to use the 
feature as this should be part of the actual documentation. The cross team 
tests are performed after the feature freeze. Documentation should be in place 
before that. Those tests are manual tests, so do not confuse them with 
automated tests.
To sum that up:
 * User facing features should be tested by other contributors
 * The scope is usability and sanity of the feature
 * The feature needs to be already documented
 * The contributor creates an issue containing some pointers on how to get 
started (e.g. link to the documentation, suggested targets of verification)
 * Other community members pick those issues up and provide feedback
 * Cross team testing happens right after the feature freeze




h3. Expectations
* Jira issues for each expected release task according to the release plan is 
created and labeled as {{release-testing}}.
* All the created release-testing-related Jira issues are resolved and the 
corresponding blocker issues are fixed.

  was:
For user facing features that go into the release we'd like to ensure they can 
actually _be used_ by Flink users. To achieve this the release managers ensure 
that an issue for cross team testing is created in the Apache Flink Jira. This 
can and should be picked up by other community members to verify the 
functionality and usability of the feature.
The issue should contain some entry points which enables other community 
members to test it. It should not contain documentation on how to use the 
feature as this should be part of the actual documentation. The cross team 
tests are performed after the feature freeze. Documentation should be in place 
before that. Those tests are manual tests, so do not confuse them with 
automated tests.
To sum that up:
 * User facing features should be tested by other contributors
 * The scope is usability and sanity of the feature
 * The feature needs to be already documented
 * The contributor creates an issue containing some pointers on how to get 
started (e.g. link to the documentation, suggested targets of verification)
 * Other community members pick those issues up and provide feedback
 * Cross team testing happens right after the feature freeze



h3. Expectations
* Jira issues for each expected release task according to the release plan is 
created and labeled as {{release-testing}}.
* All the created release-testing-related Jira issues are resolved and the 
corresponding blocker issues are fixed.


> Cross team testing
> --
>
> Key: FLINK-31150
> URL: https://issues.apache.org/jira/browse/FLINK-31150
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Assignee: Qingsheng Ren
>Priority: Major
>
> For user facing features that go into the release we'd like to ensure they 
> can actually _be used_ by Flink users. To achieve this the release managers 
> ensure that an issue for cross team testing is created in the Apache Flink 
> Jira. This can and should be picked up by other community members to verify 
> the functionality and usability of the feature.
> The issue should contain some entry points which enables other community 
> members to test it. It should not contain documentation on how to use the 
> feature as this should be part of the actual documentation. The cross team 
> tests are performed after the feature freeze. Documentation should be in 
> place before that. Those tests are manual tests, so do not confuse them with 
> automated tests.
> To sum that up:
>  * User facing features should be tested by other contributors
>  * The scope is usability and sanity of the feature
>  * The feature needs to be already documented
>  * The contributor creates an issue containing some pointers on how to get 
> started (e.g. link to the documentation, suggested targets of verification)
>  * Other community members pick those issues up and provide feedback
>  * Cross team testing happens right after the feature freeze
> 
> h3. Expectations
> * Jira issues for each expected release task according to the release plan is 
> created and labeled as {{release-testing}}.
> * All the created release-testing-related Jira issues are resolved and the 
> corresponding blocker issues are f

[jira] [Updated] (FLINK-31150) Cross team testing

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31150:
--
Description: 
For user facing features that go into the release we'd like to ensure they can 
actually _be used_ by Flink users. To achieve this the release managers ensure 
that an issue for cross team testing is created in the Apache Flink Jira. This 
can and should be picked up by other community members to verify the 
functionality and usability of the feature.
The issue should contain some entry points which enables other community 
members to test it. It should not contain documentation on how to use the 
feature as this should be part of the actual documentation. The cross team 
tests are performed after the feature freeze. Documentation should be in place 
before that. Those tests are manual tests, so do not confuse them with 
automated tests.
To sum that up:
 * User facing features should be tested by other contributors
 * The scope is usability and sanity of the feature
 * The feature needs to be already documented
 * The contributor creates an issue containing some pointers on how to get 
started (e.g. link to the documentation, suggested targets of verification)
 * Other community members pick those issues up and provide feedback
 * Cross team testing happens right after the feature freeze



h3. Expectations
* Jira issues for each expected release task according to the release plan is 
created and labeled as {{release-testing}}.
* All the created release-testing-related Jira issues are resolved and the 
corresponding blocker issues are fixed.

  was:
For user facing features that go into the release we'd like to ensure they can 
actually _be used_ by Flink users. To achieve this the release managers ensure 
that an issue for cross team testing is created in the Apache Flink Jira. This 
can and should be picked up by other community members to verify the 
functionality and usability of the feature.
The issue should contain some entry points which enables other community 
members to test it. It should not contain documentation on how to use the 
feature as this should be part of the actual documentation. The cross team 
tests are performed after the feature freeze. Documentation should be in place 
before that. Those tests are manual tests, so do not confuse them with 
automated tests.
To sum that up:
 * User facing features should be tested by other contributors
 * The scope is usability and sanity of the feature
 * The feature needs to be already documented
 * The contributor creates an issue containing some pointers on how to get 
started (e.g. link to the documentation, suggested targets of verification)
 * Other community members pick those issues up and provide feedback
 * Cross team testing happens right after the feature freeze




h3. Expectations
* Jira issues for each expected release task according to the release plan is 
created and labeled as {{release-testing}}.
* All the created release-testing-related Jira issues are resolved and the 
corresponding blocker issues are fixed.


> Cross team testing
> --
>
> Key: FLINK-31150
> URL: https://issues.apache.org/jira/browse/FLINK-31150
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Assignee: Qingsheng Ren
>Priority: Major
>
> For user facing features that go into the release we'd like to ensure they 
> can actually _be used_ by Flink users. To achieve this the release managers 
> ensure that an issue for cross team testing is created in the Apache Flink 
> Jira. This can and should be picked up by other community members to verify 
> the functionality and usability of the feature.
> The issue should contain some entry points which enables other community 
> members to test it. It should not contain documentation on how to use the 
> feature as this should be part of the actual documentation. The cross team 
> tests are performed after the feature freeze. Documentation should be in 
> place before that. Those tests are manual tests, so do not confuse them with 
> automated tests.
> To sum that up:
>  * User facing features should be tested by other contributors
>  * The scope is usability and sanity of the feature
>  * The feature needs to be already documented
>  * The contributor creates an issue containing some pointers on how to get 
> started (e.g. link to the documentation, suggested targets of verification)
>  * Other community members pick those issues up and provide feedback
>  * Cross team testing happens right after the feature freeze
> 
> h3. Expectations
> * Jira issues for each expected release task according to the release plan is 
> created and labeled as {{release-testing}}.
> * All the created release-testing-related Jira issues are resolved and the 
> corresponding blocker issues are f

[jira] [Updated] (FLINK-31150) Cross team testing

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31150:
--
Description: 
For user facing features that go into the release we'd like to ensure they can 
actually _be used_ by Flink users. To achieve this the release managers ensure 
that an issue for cross team testing is created in the Apache Flink Jira. This 
can and should be picked up by other community members to verify the 
functionality and usability of the feature.
The issue should contain some entry points which enables other community 
members to test it. It should not contain documentation on how to use the 
feature as this should be part of the actual documentation. The cross team 
tests are performed after the feature freeze. Documentation should be in place 
before that. Those tests are manual tests, so do not confuse them with 
automated tests.
To sum that up:
 * User facing features should be tested by other contributors
 * The scope is usability and sanity of the feature
 * The feature needs to be already documented
 * The contributor creates an issue containing some pointers on how to get 
started (e.g. link to the documentation, suggested targets of verification)
 * Other community members pick those issues up and provide feedback
 * Cross team testing happens right after the feature freeze


h3. Expectations
* Jira issues for each expected release task according to the release plan is 
created and labeled as {{release-testing}}.
* All the created release-testing-related Jira issues are resolved and the 
corresponding blocker issues are fixed.

  was:
For user facing features that go into the release we'd like to ensure they can 
actually _be used_ by Flink users. To achieve this the release managers ensure 
that an issue for cross team testing is created in the Apache Flink Jira. This 
can and should be picked up by other community members to verify the 
functionality and usability of the feature.
The issue should contain some entry points which enables other community 
members to test it. It should not contain documentation on how to use the 
feature as this should be part of the actual documentation. The cross team 
tests are performed after the feature freeze. Documentation should be in place 
before that. Those tests are manual tests, so do not confuse them with 
automated tests.
To sum that up:
 * User facing features should be tested by other contributors
 * The scope is usability and sanity of the feature
 * The feature needs to be already documented
 * The contributor creates an issue containing some pointers on how to get 
started (e.g. link to the documentation, suggested targets of verification)
 * Other community members pick those issues up and provide feedback
 * Cross team testing happens right after the feature freeze



h3. Expectations
* Jira issues for each expected release task according to the release plan is 
created and labeled as {{release-testing}}.
* All the created release-testing-related Jira issues are resolved and the 
corresponding blocker issues are fixed.


> Cross team testing
> --
>
> Key: FLINK-31150
> URL: https://issues.apache.org/jira/browse/FLINK-31150
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Assignee: Qingsheng Ren
>Priority: Major
>
> For user facing features that go into the release we'd like to ensure they 
> can actually _be used_ by Flink users. To achieve this the release managers 
> ensure that an issue for cross team testing is created in the Apache Flink 
> Jira. This can and should be picked up by other community members to verify 
> the functionality and usability of the feature.
> The issue should contain some entry points which enables other community 
> members to test it. It should not contain documentation on how to use the 
> feature as this should be part of the actual documentation. The cross team 
> tests are performed after the feature freeze. Documentation should be in 
> place before that. Those tests are manual tests, so do not confuse them with 
> automated tests.
> To sum that up:
>  * User facing features should be tested by other contributors
>  * The scope is usability and sanity of the feature
>  * The feature needs to be already documented
>  * The contributor creates an issue containing some pointers on how to get 
> started (e.g. link to the documentation, suggested targets of verification)
>  * Other community members pick those issues up and provide feedback
>  * Cross team testing happens right after the feature freeze
> 
> h3. Expectations
> * Jira issues for each expected release task according to the release plan is 
> created and labeled as {{release-testing}}.
> * All the created release-testing-related Jira issues are resolved and the 
> corresponding blocker issues are fix

[jira] [Comment Edited] (FLINK-30337) Stateful Functions application throws an exception when stopping a job gracefully creating a final savepoint

2023-02-21 Thread Alexey Novakov (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-30337?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691499#comment-17691499
 ] 

Alexey Novakov edited comment on FLINK-30337 at 2/21/23 10:16 AM:
--

I have been able to reproduce this issue as well. 

Flink CLI Stop command fails with above exception after timeout and there is 
one more exception in *statefun-worker* pod:
{code:java}
2023-02-21 10:06:36,818 INFO  
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable [] - 
feedback-union -> functions (1/1)#0 - asynchronous part of checkpoint 2 could 
not be completed.
java.util.concurrent.CancellationException: null
    at java.util.concurrent.FutureTask.report(Unknown Source) ~[?:?]
    at java.util.concurrent.FutureTask.get(Unknown Source) ~[?:?]
    at 
org.apache.flink.util.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:645)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.api.operators.OperatorSnapshotFinalizer.(OperatorSnapshotFinalizer.java:57)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.finalizeNonFinishedSnapshots(AsyncCheckpointRunnable.java:177)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.run(AsyncCheckpointRunnable.java:124)
 [flink-dist_2.12-1.14.3.jar:1.14.3]
    at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) [?:?]
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) [?:?]
    at java.lang.Thread.run(Unknown Source) [?:?] {code}
 

 


was (Author: novakov.alex):
I have been able to reproduce this issue as well. 

Flink CLI Stop command fails with above exception after timeout and there is 
one more exception in *statefun-worker* pod:

 
{code:java}
2023-02-21 10:06:36,818 INFO  
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable [] - 
feedback-union -> functions (1/1)#0 - asynchronous part of checkpoint 2 could 
not be completed.
java.util.concurrent.CancellationException: null
    at java.util.concurrent.FutureTask.report(Unknown Source) ~[?:?]
    at java.util.concurrent.FutureTask.get(Unknown Source) ~[?:?]
    at 
org.apache.flink.util.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:645)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.api.operators.OperatorSnapshotFinalizer.(OperatorSnapshotFinalizer.java:57)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.finalizeNonFinishedSnapshots(AsyncCheckpointRunnable.java:177)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.run(AsyncCheckpointRunnable.java:124)
 [flink-dist_2.12-1.14.3.jar:1.14.3]
    at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) [?:?]
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) [?:?]
    at java.lang.Thread.run(Unknown Source) [?:?] {code}
 

 

> Stateful Functions application throws an exception when stopping a job 
> gracefully creating a final savepoint
> 
>
> Key: FLINK-30337
> URL: https://issues.apache.org/jira/browse/FLINK-30337
> Project: Flink
>  Issue Type: Bug
>  Components: Stateful Functions
>Affects Versions: statefun-3.2.0
>Reporter: Ali Bahadir Zeybek
>Priority: Minor
>
> When running a Stateful Functions applications, if the stop[1] command is 
> executed, the client throws a FlinkException with the following stack trace 
> where *953498833da99ec437758b49b7d5befd* is the specific job id:
>  
> {code:java}
> The program finished with the following 
> exception:org.apache.flink.util.FlinkException: Could not stop with a 
> savepoint job "953498833da99ec437758b49b7d5befd".
>     at 
> org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java:581)
>     at 
> org.apache.flink.client.cli.CliFrontend.runClusterAction(CliFrontend.java:1002)
>     at org.apache.flink.client.cli.CliFrontend.stop(CliFrontend.java:569)
>     at 
> org.apache.flink.client.cli.CliFrontend.parseAndRun(CliFrontend.java:1069)
>     at 
> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:1132)
>     at 
> org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
>     at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1132)
> Caused by: java.util.concurrent.TimeoutException
>     at java.base/java.util.concurrent.CompletableFuture.timedGet(Unknown 
> Source)
>     at java.base/java.util.concurrent.CompletableFuture.get(Unknown Source)
>     at 
> org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFron

[jira] [Updated] (FLINK-31150) Cross team testing

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31150:
--
Description: 
For user facing features that go into the release we'd like to ensure they can 
actually _be used_ by Flink users. To achieve this the release managers ensure 
that an issue for cross team testing is created in the Apache Flink Jira. This 
can and should be picked up by other community members to verify the 
functionality and usability of the feature.
The issue should contain some entry points which enables other community 
members to test it. It should not contain documentation on how to use the 
feature as this should be part of the actual documentation. The cross team 
tests are performed after the feature freeze. Documentation should be in place 
before that. Those tests are manual tests, so do not confuse them with 
automated tests.
To sum that up:
 * User facing features should be tested by other contributors
 * The scope is usability and sanity of the feature
 * The feature needs to be already documented
 * The contributor creates an issue containing some pointers on how to get 
started (e.g. link to the documentation, suggested targets of verification)
 * Other community members pick those issues up and provide feedback
 * Cross team testing happens right after the feature freeze



h3. Expectations
* Jira issues for each expected release task according to the release plan is 
created and labeled as {{release-testing}}.
* All the created release-testing-related Jira issues are resolved and the 
corresponding blocker issues are fixed.

  was:
For user facing features that go into the release we'd like to ensure they can 
actually _be used_ by Flink users. To achieve this the release managers ensure 
that an issue for cross team testing is created in the Apache Flink Jira. This 
can and should be picked up by other community members to verify the 
functionality and usability of the feature.
The issue should contain some entry points which enables other community 
members to test it. It should not contain documentation on how to use the 
feature as this should be part of the actual documentation. The cross team 
tests are performed after the feature freeze. Documentation should be in place 
before that. Those tests are manual tests, so do not confuse them with 
automated tests.
To sum that up:
 * User facing features should be tested by other contributors
 * The scope is usability and sanity of the feature
 * The feature needs to be already documented
 * The contributor creates an issue containing some pointers on how to get 
started (e.g. link to the documentation, suggested targets of verification)
 * Other community members pick those issues up and provide feedback
 * Cross team testing happens right after the feature freeze


h3. Expectations
* Jira issues for each expected release task according to the release plan is 
created and labeled as {{release-testing}}.
* All the created release-testing-related Jira issues are resolved and the 
corresponding blocker issues are fixed.


> Cross team testing
> --
>
> Key: FLINK-31150
> URL: https://issues.apache.org/jira/browse/FLINK-31150
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Assignee: Qingsheng Ren
>Priority: Major
>
> For user facing features that go into the release we'd like to ensure they 
> can actually _be used_ by Flink users. To achieve this the release managers 
> ensure that an issue for cross team testing is created in the Apache Flink 
> Jira. This can and should be picked up by other community members to verify 
> the functionality and usability of the feature.
> The issue should contain some entry points which enables other community 
> members to test it. It should not contain documentation on how to use the 
> feature as this should be part of the actual documentation. The cross team 
> tests are performed after the feature freeze. Documentation should be in 
> place before that. Those tests are manual tests, so do not confuse them with 
> automated tests.
> To sum that up:
>  * User facing features should be tested by other contributors
>  * The scope is usability and sanity of the feature
>  * The feature needs to be already documented
>  * The contributor creates an issue containing some pointers on how to get 
> started (e.g. link to the documentation, suggested targets of verification)
>  * Other community members pick those issues up and provide feedback
>  * Cross team testing happens right after the feature freeze
> 
> h3. Expectations
> * Jira issues for each expected release task according to the release plan is 
> created and labeled as {{release-testing}}.
> * All the created release-testing-related Jira issues are resolved and the 
> corresponding blocker issues are fix

[jira] [Updated] (FLINK-31150) Cross team testing

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31150:
--
Description: 
For user facing features that go into the release we'd like to ensure they can 
actually _be used_ by Flink users. To achieve this the release managers ensure 
that an issue for cross team testing is created in the Apache Flink Jira. This 
can and should be picked up by other community members to verify the 
functionality and usability of the feature.
The issue should contain some entry points which enables other community 
members to test it. It should not contain documentation on how to use the 
feature as this should be part of the actual documentation. The cross team 
tests are performed after the feature freeze. Documentation should be in place 
before that. Those tests are manual tests, so do not confuse them with 
automated tests.
To sum that up:
 * User facing features should be tested by other contributors
 * The scope is usability and sanity of the feature
 * The feature needs to be already documented
 * The contributor creates an issue containing some pointers on how to get 
started (e.g. link to the documentation, suggested targets of verification)
 * Other community members pick those issues up and provide feedback
 * Cross team testing happens right after the feature freeze

--
h3. Expectations
* Jira issues for each expected release task according to the release plan is 
created and labeled as {{release-testing}}.
* All the created release-testing-related Jira issues are resolved and the 
corresponding blocker issues are fixed.

  was:
For user facing features that go into the release we'd like to ensure they can 
actually _be used_ by Flink users. To achieve this the release managers ensure 
that an issue for cross team testing is created in the Apache Flink Jira. This 
can and should be picked up by other community members to verify the 
functionality and usability of the feature.
The issue should contain some entry points which enables other community 
members to test it. It should not contain documentation on how to use the 
feature as this should be part of the actual documentation. The cross team 
tests are performed after the feature freeze. Documentation should be in place 
before that. Those tests are manual tests, so do not confuse them with 
automated tests.
To sum that up:
 * User facing features should be tested by other contributors
 * The scope is usability and sanity of the feature
 * The feature needs to be already documented
 * The contributor creates an issue containing some pointers on how to get 
started (e.g. link to the documentation, suggested targets of verification)
 * Other community members pick those issues up and provide feedback
 * Cross team testing happens right after the feature freeze


> Cross team testing
> --
>
> Key: FLINK-31150
> URL: https://issues.apache.org/jira/browse/FLINK-31150
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Assignee: Qingsheng Ren
>Priority: Major
>
> For user facing features that go into the release we'd like to ensure they 
> can actually _be used_ by Flink users. To achieve this the release managers 
> ensure that an issue for cross team testing is created in the Apache Flink 
> Jira. This can and should be picked up by other community members to verify 
> the functionality and usability of the feature.
> The issue should contain some entry points which enables other community 
> members to test it. It should not contain documentation on how to use the 
> feature as this should be part of the actual documentation. The cross team 
> tests are performed after the feature freeze. Documentation should be in 
> place before that. Those tests are manual tests, so do not confuse them with 
> automated tests.
> To sum that up:
>  * User facing features should be tested by other contributors
>  * The scope is usability and sanity of the feature
>  * The feature needs to be already documented
>  * The contributor creates an issue containing some pointers on how to get 
> started (e.g. link to the documentation, suggested targets of verification)
>  * Other community members pick those issues up and provide feedback
>  * Cross team testing happens right after the feature freeze
> --
> h3. Expectations
> * Jira issues for each expected release task according to the release plan is 
> created and labeled as {{release-testing}}.
> * All the created release-testing-related Jira issues are resolved and the 
> corresponding blocker issues are fixed.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31150) Cross team testing

2023-02-21 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31150?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-31150:
--
Description: 
For user facing features that go into the release we'd like to ensure they can 
actually _be used_ by Flink users. To achieve this the release managers ensure 
that an issue for cross team testing is created in the Apache Flink Jira. This 
can and should be picked up by other community members to verify the 
functionality and usability of the feature.
The issue should contain some entry points which enables other community 
members to test it. It should not contain documentation on how to use the 
feature as this should be part of the actual documentation. The cross team 
tests are performed after the feature freeze. Documentation should be in place 
before that. Those tests are manual tests, so do not confuse them with 
automated tests.
To sum that up:
 * User facing features should be tested by other contributors
 * The scope is usability and sanity of the feature
 * The feature needs to be already documented
 * The contributor creates an issue containing some pointers on how to get 
started (e.g. link to the documentation, suggested targets of verification)
 * Other community members pick those issues up and provide feedback
 * Cross team testing happens right after the feature freeze

 

h3. Expectations
 * Jira issues for each expected release task according to the release plan is 
created and labeled as {{{}release-testing{}}}.
 * All the created release-testing-related Jira issues are resolved and the 
corresponding blocker issues are fixed.

  was:
For user facing features that go into the release we'd like to ensure they can 
actually _be used_ by Flink users. To achieve this the release managers ensure 
that an issue for cross team testing is created in the Apache Flink Jira. This 
can and should be picked up by other community members to verify the 
functionality and usability of the feature.
The issue should contain some entry points which enables other community 
members to test it. It should not contain documentation on how to use the 
feature as this should be part of the actual documentation. The cross team 
tests are performed after the feature freeze. Documentation should be in place 
before that. Those tests are manual tests, so do not confuse them with 
automated tests.
To sum that up:
 * User facing features should be tested by other contributors
 * The scope is usability and sanity of the feature
 * The feature needs to be already documented
 * The contributor creates an issue containing some pointers on how to get 
started (e.g. link to the documentation, suggested targets of verification)
 * Other community members pick those issues up and provide feedback
 * Cross team testing happens right after the feature freeze




h3. Expectations
* Jira issues for each expected release task according to the release plan is 
created and labeled as {{release-testing}}.
* All the created release-testing-related Jira issues are resolved and the 
corresponding blocker issues are fixed.


> Cross team testing
> --
>
> Key: FLINK-31150
> URL: https://issues.apache.org/jira/browse/FLINK-31150
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Matthias Pohl
>Assignee: Qingsheng Ren
>Priority: Major
>
> For user facing features that go into the release we'd like to ensure they 
> can actually _be used_ by Flink users. To achieve this the release managers 
> ensure that an issue for cross team testing is created in the Apache Flink 
> Jira. This can and should be picked up by other community members to verify 
> the functionality and usability of the feature.
> The issue should contain some entry points which enables other community 
> members to test it. It should not contain documentation on how to use the 
> feature as this should be part of the actual documentation. The cross team 
> tests are performed after the feature freeze. Documentation should be in 
> place before that. Those tests are manual tests, so do not confuse them with 
> automated tests.
> To sum that up:
>  * User facing features should be tested by other contributors
>  * The scope is usability and sanity of the feature
>  * The feature needs to be already documented
>  * The contributor creates an issue containing some pointers on how to get 
> started (e.g. link to the documentation, suggested targets of verification)
>  * Other community members pick those issues up and provide feedback
>  * Cross team testing happens right after the feature freeze
>  
> 
> h3. Expectations
>  * Jira issues for each expected release task according to the release plan 
> is created and labeled as {{{}release-testing{}}}.
>  * All the created release-testing-related Jira issues are resolved and the 
> corresponding blo

[jira] [Comment Edited] (FLINK-30337) Stateful Functions application throws an exception when stopping a job gracefully creating a final savepoint

2023-02-21 Thread Alexey Novakov (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-30337?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17691499#comment-17691499
 ] 

Alexey Novakov edited comment on FLINK-30337 at 2/21/23 10:17 AM:
--

I have been able to reproduce this issue as well. 

Flink CLI Stop command fails with above exception after timeout and there is 
one more exception in *statefun-worker* pod:
{code:java}
2023-02-21 10:06:36,818 INFO  
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable [] - 
feedback-union -> functions (1/1)#0 - asynchronous part of checkpoint 2 could 
not be completed.
java.util.concurrent.CancellationException: null
    at java.util.concurrent.FutureTask.report(Unknown Source) ~[?:?]
    at java.util.concurrent.FutureTask.get(Unknown Source) ~[?:?]
    at 
org.apache.flink.util.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:645)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.api.operators.OperatorSnapshotFinalizer.(OperatorSnapshotFinalizer.java:57)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.finalizeNonFinishedSnapshots(AsyncCheckpointRunnable.java:177)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.run(AsyncCheckpointRunnable.java:124)
 [flink-dist_2.12-1.14.3.jar:1.14.3]
    at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) [?:?]
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) [?:?]
    at java.lang.Thread.run(Unknown Source) [?:?]{code}


was (Author: novakov.alex):
I have been able to reproduce this issue as well. 

Flink CLI Stop command fails with above exception after timeout and there is 
one more exception in *statefun-worker* pod:
{code:java}
2023-02-21 10:06:36,818 INFO  
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable [] - 
feedback-union -> functions (1/1)#0 - asynchronous part of checkpoint 2 could 
not be completed.
java.util.concurrent.CancellationException: null
    at java.util.concurrent.FutureTask.report(Unknown Source) ~[?:?]
    at java.util.concurrent.FutureTask.get(Unknown Source) ~[?:?]
    at 
org.apache.flink.util.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:645)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.api.operators.OperatorSnapshotFinalizer.(OperatorSnapshotFinalizer.java:57)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.finalizeNonFinishedSnapshots(AsyncCheckpointRunnable.java:177)
 ~[flink-dist_2.12-1.14.3.jar:1.14.3]
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.run(AsyncCheckpointRunnable.java:124)
 [flink-dist_2.12-1.14.3.jar:1.14.3]
    at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) [?:?]
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) [?:?]
    at java.lang.Thread.run(Unknown Source) [?:?] {code}
 

 

> Stateful Functions application throws an exception when stopping a job 
> gracefully creating a final savepoint
> 
>
> Key: FLINK-30337
> URL: https://issues.apache.org/jira/browse/FLINK-30337
> Project: Flink
>  Issue Type: Bug
>  Components: Stateful Functions
>Affects Versions: statefun-3.2.0
>Reporter: Ali Bahadir Zeybek
>Priority: Minor
>
> When running a Stateful Functions applications, if the stop[1] command is 
> executed, the client throws a FlinkException with the following stack trace 
> where *953498833da99ec437758b49b7d5befd* is the specific job id:
>  
> {code:java}
> The program finished with the following 
> exception:org.apache.flink.util.FlinkException: Could not stop with a 
> savepoint job "953498833da99ec437758b49b7d5befd".
>     at 
> org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java:581)
>     at 
> org.apache.flink.client.cli.CliFrontend.runClusterAction(CliFrontend.java:1002)
>     at org.apache.flink.client.cli.CliFrontend.stop(CliFrontend.java:569)
>     at 
> org.apache.flink.client.cli.CliFrontend.parseAndRun(CliFrontend.java:1069)
>     at 
> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:1132)
>     at 
> org.apache.flink.runtime.security.contexts.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:28)
>     at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1132)
> Caused by: java.util.concurrent.TimeoutException
>     at java.base/java.util.concurrent.CompletableFuture.timedGet(Unknown 
> Source)
>     at java.base/java.util.concurrent.CompletableFuture.get(Unknown Source)
>     at 
> org.apache.flink.client.cli.CliFrontend.lambda$stop$5(CliFrontend.java

  1   2   3   4   >