Re: [PR] [FLINK-33704][BP 1.18][Filesytems] Update GCS filesystems to latest available versions [flink]

2023-12-18 Thread via GitHub


MartijnVisser commented on PR #23935:
URL: https://github.com/apache/flink/pull/23935#issuecomment-1859729348

   > Shall be #23837
   
   Indeed, thanks for that! 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33704][BP 1.18][Filesytems] Update GCS filesystems to latest available versions [flink]

2023-12-18 Thread via GitHub


MartijnVisser merged PR #23935:
URL: https://github.com/apache/flink/pull/23935


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-33704) Update GCS filesystems to latest available versions

2023-12-18 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33704?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser updated FLINK-33704:
---
Fix Version/s: 1.18.1

> Update GCS filesystems to latest available versions
> ---
>
> Key: FLINK-33704
> URL: https://issues.apache.org/jira/browse/FLINK-33704
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Connectors / FileSystem, FileSystems
>Reporter: Martijn Visser
>Assignee: Martijn Visser
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
>
> Update GS SDK from 2.15.0 to 2.29.1 and GS Hadoop Connector from 2.2.15 to 
> 2.2.18



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Reopened] (FLINK-33704) Update GCS filesystems to latest available versions

2023-12-18 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33704?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser reopened FLINK-33704:


> Update GCS filesystems to latest available versions
> ---
>
> Key: FLINK-33704
> URL: https://issues.apache.org/jira/browse/FLINK-33704
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Connectors / FileSystem, FileSystems
>Reporter: Martijn Visser
>Assignee: Martijn Visser
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0
>
>
> Update GS SDK from 2.15.0 to 2.29.1 and GS Hadoop Connector from 2.2.15 to 
> 2.2.18



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33603][1.18][Filesystem] Bundle Guava for GCS connector [flink]

2023-12-18 Thread via GitHub


MartijnVisser closed pull request #23920: [FLINK-33603][1.18][Filesystem] 
Bundle Guava for GCS connector
URL: https://github.com/apache/flink/pull/23920


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33603][1.18][Filesystem] Bundle Guava for GCS connector [flink]

2023-12-18 Thread via GitHub


MartijnVisser commented on PR #23920:
URL: https://github.com/apache/flink/pull/23920#issuecomment-1859733229

   Superseded by https://issues.apache.org/jira/browse/FLINK-33704


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [ FLINK-33603][FileSystems] shade guava in gs-fs filesystem [flink]

2023-12-18 Thread via GitHub


MartijnVisser commented on PR #23489:
URL: https://github.com/apache/flink/pull/23489#issuecomment-1859734539

   Superseded by https://issues.apache.org/jira/browse/FLINK-33704 - Thanks for 
the help/thinking on this @singhravidutt 
   
   @singhravidutt @cnauroth Do you have any bandwidth to add a GCS connector 
test to Flink?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [ FLINK-33603][FileSystems] shade guava in gs-fs filesystem [flink]

2023-12-18 Thread via GitHub


MartijnVisser closed pull request #23489: [ FLINK-33603][FileSystems] shade 
guava in gs-fs filesystem
URL: https://github.com/apache/flink/pull/23489


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-25509) FLIP-208: Add RecordEvaluator to dynamically stop source based on de-serialized records

2023-12-18 Thread Martijn Visser (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-25509?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798044#comment-17798044
 ] 

Martijn Visser commented on FLINK-25509:


[~ruanhang1993] [~lindong] Both the {{v3.0}} and the {{main}} branch of the 
Flink Kafka connector work for Flink 1.18 (see 
https://github.com/apache/flink-connector-kafka/actions/runs/7235099577 which 
shows the nightly run). I'm inclined to make it only compatible with {{main}} 
so that a new Flink Kafka connector v3.1 would be released. There are some 
other changes already in {{main}} as well (different Kafka client)

> FLIP-208: Add RecordEvaluator to dynamically stop source based on 
> de-serialized records
> ---
>
> Key: FLINK-25509
> URL: https://issues.apache.org/jira/browse/FLINK-25509
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / Common, Connectors / Kafka
>Reporter: Dong Lin
>Assignee: Hang Ruan
>Priority: Major
>  Labels: pull-request-available, stale-assigned
>
> This feature is needed to migrate applications which uses 
> KafkaDeserializationSchema::isEndOfStream() from using FlinkKafkaConsumer to 
> using KafkaSource.
> Please checkout 
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-208%3A+Add+RecordEvaluator+to+dynamically+stop+source+based+on+de-serialized+records
>  for the motivation and the proposed changes.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[PR] Reduce getTable call for hive client and optimize graph generation time [flink]

2023-12-18 Thread via GitHub


hehuiyuan opened a new pull request, #23945:
URL: https://github.com/apache/flink/pull/23945

   
   ## What is the purpose of the change
   
   HiveCatalog.getHiveTable method wastes a lot of time when generate graph, 
because the number of calls  is relatively high.
   
   
   I have an sql task with over 2000 rows,  the HiveCatalog.getHiveTable  
method is called 4879 times , but only six hive tables were used. 
   
   
![image](https://github.com/apache/flink/assets/18002496/d5f0daf3-f80a-4790-ae21-4e75dff9cfd7)
   
   The client.getTable method costs a lot of time.  One call takes 
approximately 50 milliseconds,
   
   4879 * 50 =243950ms  = 243.95s  = 4min
   
   ## Brief change log
   
   Reduce getTable call for hive client by using cache.
   
   
   ## Verifying this change
   
   
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / no)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / no)
 - The serializers: (yes / no / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / no / 
don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / no / don't know)
 - The S3 file system connector: (yes / no / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / no)
 - If yes, how is the feature documented? (not applicable / docs / JavaDocs 
/ not documented)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-21949][table] Support ARRAY_AGG aggregate function [flink]

2023-12-18 Thread via GitHub


Jiabao-Sun commented on PR #23411:
URL: https://github.com/apache/flink/pull/23411#issuecomment-1859780440

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-33871) Reduce getTable call for hive client and optimize graph generation time

2023-12-18 Thread hehuiyuan (Jira)
hehuiyuan created FLINK-33871:
-

 Summary: Reduce getTable call for hive client and optimize graph 
generation time
 Key: FLINK-33871
 URL: https://issues.apache.org/jira/browse/FLINK-33871
 Project: Flink
  Issue Type: Improvement
Reporter: hehuiyuan


HiveCatalog.getHiveTable method wastes a lot of time when generate graph, 
because the number of calls  is relatively high.


I have an sql task with over 2000 rows,  the HiveCatalog.getHiveTable  method 
is called 4879 times , but only six hive tables were used. 

![image](https://github.com/apache/flink/assets/18002496/d5f0daf3-f80a-4790-ae21-4e75dff9cfd7)

The client.getTable method costs a lot of time.  

![image](https://github.com/apache/flink/assets/18002496/be0d176f-3915-4b92-a177-f1cfaf6d2927)
There is a statistic that jobmanager interacts with hive when generate graph.

If One call takes approximately 50 milliseconds ,
How much time it spends  : 4879 * 50 =243950ms  = 243.95s  = 4min

We can cache and  client.getTable method  is only  called six times.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33871) Reduce getTable call for hive client and optimize graph generation time

2023-12-18 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33871?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-33871:
---
Labels: pull-request-available  (was: )

> Reduce getTable call for hive client and optimize graph generation time
> ---
>
> Key: FLINK-33871
> URL: https://issues.apache.org/jira/browse/FLINK-33871
> Project: Flink
>  Issue Type: Improvement
>Reporter: hehuiyuan
>Priority: Major
>  Labels: pull-request-available
>
> HiveCatalog.getHiveTable method wastes a lot of time when generate graph, 
> because the number of calls  is relatively high.
> I have an sql task with over 2000 rows,  the HiveCatalog.getHiveTable  method 
> is called 4879 times , but only six hive tables were used. 
> ![image](https://github.com/apache/flink/assets/18002496/d5f0daf3-f80a-4790-ae21-4e75dff9cfd7)
> The client.getTable method costs a lot of time.  
> ![image](https://github.com/apache/flink/assets/18002496/be0d176f-3915-4b92-a177-f1cfaf6d2927)
> There is a statistic that jobmanager interacts with hive when generate graph.
> If One call takes approximately 50 milliseconds ,
> How much time it spends  : 4879 * 50 =243950ms  = 243.95s  = 4min
> We can cache and  client.getTable method  is only  called six times.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33871][HIVE]Reduce getTable call for hive client and optimize graph generation time [flink]

2023-12-18 Thread via GitHub


flinkbot commented on PR #23945:
URL: https://github.com/apache/flink/pull/23945#issuecomment-1859784457

   
   ## CI report:
   
   * c6efa328604d58eb5ac27e39cef9f1c428b336b9 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-31650][metrics][rest] Remove transient metrics for subtasks in terminal state [flink]

2023-12-18 Thread via GitHub


JunRuiLee commented on code in PR #23447:
URL: https://github.com/apache/flink/pull/23447#discussion_r1429717030


##
flink-runtime/src/main/java/org/apache/flink/runtime/rest/handler/legacy/metrics/MetricStore.java:
##
@@ -126,9 +140,14 @@ public synchronized void 
updateCurrentExecutionAttempts(Collection j
 
taskMetricStore.getSubtaskMetricStore(
 
subtaskIndex))
 .ifPresent(
-subtaskMetricStore ->
-
subtaskMetricStore.retainAttempts(
-
attempts.getCurrentAttempts()));
+subtaskMetricStore -> {
+
subtaskMetricStore.retainAttempts(

Review Comment:
   Thanks for clarification @X-czh .
   
   I'm not quite certain I understand your concern.
   In my opinion, this issue is unrelated to the web interface and more related 
to the inconsistency in the MetricStore because the WebUI is also get data from 
MetricStore. Specifically, the metrics in the subtaskMetricsStore are being 
removed, while the metrics in the taskMetricsStore are not synchronously 
removed, which could be confusing for users.
   
   Based on your changes, you can perform the following test:
   For a jobVertex that has already finished, you can use the 
JobVertexMetricsHandler to retrieve subtask metrics like below:
   
http://localhost:8081/jobs//vertices//metrics?get=0.backPressuredTimeMsPerSecond,0.busyTimeMsPerSecond
   Then, compare the results with the SubtaskMetricsHandler:
   
http://localhost:8081/jobs//vertices//subtasks/0/metrics?get=backPressuredTimeMsPerSecond,busyTimeMsPerSecond
   
   The results from these two endpoints are different. In my local test, the 
results are as shown in the attached image. I prefer that cleaning up should be 
done simultaneously for both, WDYT?
   
   
![image](https://github.com/apache/flink/assets/107924572/7ecdf3a0-f9ca-4ddf-aadd-a5d6a684f7f9)
   



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-25509) FLIP-208: Add RecordEvaluator to dynamically stop source based on de-serialized records

2023-12-18 Thread Dong Lin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-25509?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798047#comment-17798047
 ] 

Dong Lin commented on FLINK-25509:
--

[~martijnvisser] Just to make sure that I understand you correctly, do you mean 
it is OK to put this feature in flink-kafka-connector's main branch? This would 
keep its main branch compatible with Flink 1.18. But it won't be compatible 
with Flink 1.17.

> FLIP-208: Add RecordEvaluator to dynamically stop source based on 
> de-serialized records
> ---
>
> Key: FLINK-25509
> URL: https://issues.apache.org/jira/browse/FLINK-25509
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / Common, Connectors / Kafka
>Reporter: Dong Lin
>Assignee: Hang Ruan
>Priority: Major
>  Labels: pull-request-available, stale-assigned
>
> This feature is needed to migrate applications which uses 
> KafkaDeserializationSchema::isEndOfStream() from using FlinkKafkaConsumer to 
> using KafkaSource.
> Please checkout 
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-208%3A+Add+RecordEvaluator+to+dynamically+stop+source+based+on+de-serialized+records
>  for the motivation and the proposed changes.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-25509) FLIP-208: Add RecordEvaluator to dynamically stop source based on de-serialized records

2023-12-18 Thread Martijn Visser (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-25509?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798066#comment-17798066
 ] 

Martijn Visser commented on FLINK-25509:


[~lindong] Yes, that's what I suggest. The one thing I would like to check with 
[~tzulitai] is if we want to cut {{main}} now first and create a Flink Kafka 
v3.1 branch, which would (at this point) have support for both Flink 1.17 and 
Flink 1.18. Then with the PR for this ticket opened, we would drop support for 
Flink 1.17 and therefore release it as part of a new Flink Kafka v4.0 
branch/release which only would support Flink 1.18 and newer. For users of 
Flink 1.17, they would stick with the v3.1 version (until support for Flink 
1.17 is stopped completely) 

> FLIP-208: Add RecordEvaluator to dynamically stop source based on 
> de-serialized records
> ---
>
> Key: FLINK-25509
> URL: https://issues.apache.org/jira/browse/FLINK-25509
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / Common, Connectors / Kafka
>Reporter: Dong Lin
>Assignee: Hang Ruan
>Priority: Major
>  Labels: pull-request-available, stale-assigned
>
> This feature is needed to migrate applications which uses 
> KafkaDeserializationSchema::isEndOfStream() from using FlinkKafkaConsumer to 
> using KafkaSource.
> Please checkout 
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-208%3A+Add+RecordEvaluator+to+dynamically+stop+source+based+on+de-serialized+records
>  for the motivation and the proposed changes.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-31650][metrics][rest] Remove transient metrics for subtasks in terminal state [flink]

2023-12-18 Thread via GitHub


X-czh commented on code in PR #23447:
URL: https://github.com/apache/flink/pull/23447#discussion_r1429759033


##
flink-runtime/src/main/java/org/apache/flink/runtime/rest/handler/legacy/metrics/MetricStore.java:
##
@@ -126,9 +140,14 @@ public synchronized void 
updateCurrentExecutionAttempts(Collection j
 
taskMetricStore.getSubtaskMetricStore(
 
subtaskIndex))
 .ifPresent(
-subtaskMetricStore ->
-
subtaskMetricStore.retainAttempts(
-
attempts.getCurrentAttempts()));
+subtaskMetricStore -> {
+
subtaskMetricStore.retainAttempts(

Review Comment:
   Hi @JunRuiLee, thanks for the clarification. I get your point and my concern 
is actual about maintainability of the code in the future. When there're both 
insertion and deletion operations, the duplication of task-level metrics 
actually makes it more difficult to maintain consistency (as is the case here), 
so I think it would be better to optimize the duplication issue here in the 
future.
   
   I'll clean up simultaneously for both cases first here, and create a new 
issue for the optimization later. WDYT?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] test request [flink]

2023-12-18 Thread via GitHub


singhravidutt closed pull request #23933: test request
URL: https://github.com/apache/flink/pull/23933


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33632] Adding custom flink mutator [flink-kubernetes-operator]

2023-12-18 Thread via GitHub


AncyRominus commented on code in PR #733:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/733#discussion_r1429764133


##
flink-kubernetes-webhook/pom.xml:
##
@@ -127,6 +132,11 @@ under the License.
 
 
 
+
+org.apache.maven.plugins
+maven-surefire-plugin
+${surefire-plugin.version}

Review Comment:
This plugin is required for the successful execution of the test 
AdmissionHandlerTest.testMutateHandler. Else it will throw <500 Internal Server 
Error> while mutating the SessionJob as it’s using the deployment from the 
informerManager which uses Kubernetes Client.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-15571) Create a Redis Streams Connector for Flink

2023-12-18 Thread Martijn Visser (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-15571?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798078#comment-17798078
 ] 

Martijn Visser commented on FLINK-15571:


[~jinsuichen] If there were volunteering maintainers for this, probably yes

> Create a Redis Streams Connector for Flink
> --
>
> Key: FLINK-15571
> URL: https://issues.apache.org/jira/browse/FLINK-15571
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / Redis Streams
>Reporter: Tugdual Grall
>Assignee: ZhuoYu Chen
>Priority: Minor
>  Labels: pull-request-available, stale-assigned
>
> Redis has a "log data structure" called Redis Streams, it would be nice to 
> integrate Redis Streams and Apache Flink as:
>  * Source
>  * Sink
> See Redis Streams introduction: [https://redis.io/topics/streams-intro]
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33632] Adding custom flink mutator [flink-kubernetes-operator]

2023-12-18 Thread via GitHub


AncyRominus commented on code in PR #733:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/733#discussion_r1429772463


##
flink-kubernetes-operator/pom.xml:
##
@@ -208,6 +208,18 @@ under the License.
 junit-jupiter-params
 test
 
+
+
+io.javaoperatorsdk
+kubernetes-webhooks-framework-core
+${operator.sdk.webhook-framework.version}
+
+
+*
+*
+
+
+

Review Comment:
   This  dependency is for the Kubernetes-webhooks-framework. This can be 
removed as we don't have any strong relation with this dependency.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-25509) FLIP-208: Add RecordEvaluator to dynamically stop source based on de-serialized records

2023-12-18 Thread Dong Lin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-25509?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798080#comment-17798080
 ] 

Dong Lin commented on FLINK-25509:
--

[~martijnvisser] Cool. I totally agree with your suggestion.

> FLIP-208: Add RecordEvaluator to dynamically stop source based on 
> de-serialized records
> ---
>
> Key: FLINK-25509
> URL: https://issues.apache.org/jira/browse/FLINK-25509
> Project: Flink
>  Issue Type: New Feature
>  Components: Connectors / Common, Connectors / Kafka
>Reporter: Dong Lin
>Assignee: Hang Ruan
>Priority: Major
>  Labels: pull-request-available, stale-assigned
>
> This feature is needed to migrate applications which uses 
> KafkaDeserializationSchema::isEndOfStream() from using FlinkKafkaConsumer to 
> using KafkaSource.
> Please checkout 
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-208%3A+Add+RecordEvaluator+to+dynamically+stop+source+based+on+de-serialized+records
>  for the motivation and the proposed changes.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33632] Adding custom flink mutator [flink-kubernetes-operator]

2023-12-18 Thread via GitHub


AncyRominus commented on PR #733:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/733#issuecomment-1859901718

   Thanks for the comments @tagarr and @gyfora !!! 
   The modify labels are set in FlinkMutator.java and the other mutator changes 
(if any) be added to the default or the custom mutators. The default/custom 
mutator sessionJob changes are iterated and the mutateSessionJobs are processed.
   
   Yes, we can add the sessionjob mutation for Update as well as for Create 
operations. 
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-31481][table] Support enhanced show databases syntax [flink]

2023-12-18 Thread via GitHub


jeyhunkarimov commented on PR #23612:
URL: https://github.com/apache/flink/pull/23612#issuecomment-1859901892

   Hi @jnh5y could you please check if your comments are addressed? Thanks!


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-31049) Add support for Kafka record headers to KafkaSink

2023-12-18 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31049?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser updated FLINK-31049:
---
Fix Version/s: kafka-3.1.0

> Add support for Kafka record headers to KafkaSink
> -
>
> Key: FLINK-31049
> URL: https://issues.apache.org/jira/browse/FLINK-31049
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka
>Reporter: Alex Gout
>Assignee: Alex Gout
>Priority: Minor
>  Labels: KafkaSink, pull-request-available, stale-assigned
> Fix For: kafka-3.1.0
>
>   Original Estimate: 48h
>  Remaining Estimate: 48h
>
> The default org.apache.flink.connector.kafka.sink.KafkaSink does not support 
> adding Kafka record headers. In some implementations, downstream consumers 
> might rely on Kafka record headers being set.
>  
> A way to add Headers would be to create a custom 
> KafkaRecordSerializationSchema and inject that into the KafkaSink.
> However, I'm assuming the KafkaRecordSerializationSchemaBuilder was added for 
> convenience and allows a more usable approach of creating a KafkaSink without 
> having to deal with details like the RecordProducer directly. This builder 
> does not support adding record headers.
> This is where I think it should be added.
> The code responsible for creating the Kafka record involves  
> org.apache.flink.connector.kafka.sink.KafkaRecordSerializationSchemaWrapper 
> where the RecordProducer is created. 
> It is relatively simple to add support for record headers by adding a 
> "HeaderProducer" to the KafkaRecordSerializationSchemaBuilder next to the key 
> and value serializers and using the appropriate RecordProducer constructor.
>  
> The issue was discussed 
> [here|https://lists.apache.org/thread/shlbbcqho0q9w5shjwdlscnsywjvbfro].
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-31049) Add support for Kafka record headers to KafkaSink

2023-12-18 Thread Martijn Visser (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31049?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798082#comment-17798082
 ] 

Martijn Visser commented on FLINK-31049:


Merged by [~tzulitai] in apache/flink-connector-kafka:main 
a7785630e714af303b224c38d9a6caa89a551265

> Add support for Kafka record headers to KafkaSink
> -
>
> Key: FLINK-31049
> URL: https://issues.apache.org/jira/browse/FLINK-31049
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka
>Reporter: Alex Gout
>Assignee: Alex Gout
>Priority: Minor
>  Labels: KafkaSink, pull-request-available, stale-assigned
> Fix For: kafka-3.1.0
>
>   Original Estimate: 48h
>  Remaining Estimate: 48h
>
> The default org.apache.flink.connector.kafka.sink.KafkaSink does not support 
> adding Kafka record headers. In some implementations, downstream consumers 
> might rely on Kafka record headers being set.
>  
> A way to add Headers would be to create a custom 
> KafkaRecordSerializationSchema and inject that into the KafkaSink.
> However, I'm assuming the KafkaRecordSerializationSchemaBuilder was added for 
> convenience and allows a more usable approach of creating a KafkaSink without 
> having to deal with details like the RecordProducer directly. This builder 
> does not support adding record headers.
> This is where I think it should be added.
> The code responsible for creating the Kafka record involves  
> org.apache.flink.connector.kafka.sink.KafkaRecordSerializationSchemaWrapper 
> where the RecordProducer is created. 
> It is relatively simple to add support for record headers by adding a 
> "HeaderProducer" to the KafkaRecordSerializationSchemaBuilder next to the key 
> and value serializers and using the appropriate RecordProducer constructor.
>  
> The issue was discussed 
> [here|https://lists.apache.org/thread/shlbbcqho0q9w5shjwdlscnsywjvbfro].
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33819) Support setting CompressType in RocksDBStateBackend

2023-12-18 Thread Yue Ma (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798083#comment-17798083
 ] 

Yue Ma commented on FLINK-33819:


[~masteryhx] Thanks for replying.
{quote}Linked FLINK-20684 as it has been discussed before.
{quote}
Sorry for missing the previous ticket and creating the duplicated one
{quote}Linked FLINK-11313 which talks about the LZ4 Compression which should be 
more usable than Snappy.
{quote}
We did not use LZ4 in the production environment, and for small state jobs, we 
directly turn off the compression. And for job with large states,  we adopted a 
compression algorithm based on Snappy optimization.
{quote}Do you have some test results on it ?
{quote}
Yes. we did some benchmark test for the *SnappyCompression* and *NoCompression*
And result show that and Read Performance. After turning off Compression, State 
Benchmark read performance can be improved by *80% to 100%*

We also conducted end-to-end online job testing, and after turning off 
Compression, {*}the CPU usage of the job decreased by 16%{*}, while the 
Checkpoint Total Size increased by *4-5 times.*

It is obvious that closing Compression is not only about benefits, but also 
brings some space amplification. What I want to express is that we may need to 
provide such a configuration for users to balance how to exchange space for time
 
{quote}BTW, If we'd like to introduce such a option, it's  better to guarantee 
the compalibility.
{quote}
Sorry, I didn't understand the compatibility issue here. I understand that it 
is compatible here. After switching the Compression Type, the newly generated 
file will be compressed using the new Compress Type, and the existing file can 
still be read and written with old Compress Type. 

> Support setting CompressType in RocksDBStateBackend
> ---
>
> Key: FLINK-33819
> URL: https://issues.apache.org/jira/browse/FLINK-33819
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / State Backends
>Affects Versions: 1.18.0
>Reporter: Yue Ma
>Priority: Major
> Fix For: 1.19.0
>
> Attachments: image-2023-12-14-11-32-32-968.png, 
> image-2023-12-14-11-35-22-306.png
>
>
> Currently, RocksDBStateBackend does not support setting the compression 
> level, and Snappy is used for compression by default. But we have some 
> scenarios where compression will use a lot of CPU resources. Turning off 
> compression can significantly reduce CPU overhead. So we may need to support 
> a parameter for users to set the CompressType of Rocksdb.
>   !image-2023-12-14-11-35-22-306.png!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-31049) Add support for Kafka record headers to KafkaSink

2023-12-18 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31049?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser closed FLINK-31049.
--
Resolution: Fixed

> Add support for Kafka record headers to KafkaSink
> -
>
> Key: FLINK-31049
> URL: https://issues.apache.org/jira/browse/FLINK-31049
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka
>Reporter: Alex Gout
>Assignee: Alex Gout
>Priority: Minor
>  Labels: KafkaSink, pull-request-available, stale-assigned
> Fix For: kafka-3.1.0
>
>   Original Estimate: 48h
>  Remaining Estimate: 48h
>
> The default org.apache.flink.connector.kafka.sink.KafkaSink does not support 
> adding Kafka record headers. In some implementations, downstream consumers 
> might rely on Kafka record headers being set.
>  
> A way to add Headers would be to create a custom 
> KafkaRecordSerializationSchema and inject that into the KafkaSink.
> However, I'm assuming the KafkaRecordSerializationSchemaBuilder was added for 
> convenience and allows a more usable approach of creating a KafkaSink without 
> having to deal with details like the RecordProducer directly. This builder 
> does not support adding record headers.
> This is where I think it should be added.
> The code responsible for creating the Kafka record involves  
> org.apache.flink.connector.kafka.sink.KafkaRecordSerializationSchemaWrapper 
> where the RecordProducer is created. 
> It is relatively simple to add support for record headers by adding a 
> "HeaderProducer" to the KafkaRecordSerializationSchemaBuilder next to the key 
> and value serializers and using the appropriate RecordProducer constructor.
>  
> The issue was discussed 
> [here|https://lists.apache.org/thread/shlbbcqho0q9w5shjwdlscnsywjvbfro].
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33153) Kafka using latest-offset maybe missing data

2023-12-18 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33153?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser updated FLINK-33153:
---
Affects Version/s: (was: kafka-4.1.0)

> Kafka using latest-offset maybe missing data
> 
>
> Key: FLINK-33153
> URL: https://issues.apache.org/jira/browse/FLINK-33153
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
>Reporter: tanjialiang
>Priority: Minor
>
> When Kafka start with the latest-offset strategy, it does not fetch the 
> latest snapshot offset and specify it for consumption. Instead, it sets the 
> startingOffset to -1 (KafkaPartitionSplit.LATEST_OFFSET, which makes 
> currentOffset = -1, and call the KafkaConsumer's  seekToEnd API). The 
> currentOffset is only set to the consumed offset + 1 when the task consumes 
> data, and this currentOffset is stored in the state during checkpointing. If 
> there are very few messages in Kafka and a partition has not consumed any 
> data, and I stop the task with a savepoint, then write data to that 
> partition, and start the task with the savepoint, the task will resume from 
> the saved state. Due to the startingOffset in the state being -1, it will 
> cause the task to miss the data that was written before the recovery point.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-31650][metrics][rest] Remove transient metrics for subtasks in terminal state [flink]

2023-12-18 Thread via GitHub


JunRuiLee commented on code in PR #23447:
URL: https://github.com/apache/flink/pull/23447#discussion_r1429785675


##
flink-runtime/src/main/java/org/apache/flink/runtime/rest/handler/legacy/metrics/MetricStore.java:
##
@@ -126,9 +140,14 @@ public synchronized void 
updateCurrentExecutionAttempts(Collection j
 
taskMetricStore.getSubtaskMetricStore(
 
subtaskIndex))
 .ifPresent(
-subtaskMetricStore ->
-
subtaskMetricStore.retainAttempts(
-
attempts.getCurrentAttempts()));
+subtaskMetricStore -> {
+
subtaskMetricStore.retainAttempts(

Review Comment:
   Clean up simultaneously for both cases first here sounds good to me, Please 
feel free and go ahead~



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-33365) Missing filter condition in execution plan containing lookup join with mysql jdbc connector

2023-12-18 Thread david radley (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33365?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798091#comment-17798091
 ] 

david radley commented on FLINK-33365:
--

[~libenchao] [~Sergey Nuyanzin] Bringing back a discussion onto the issue. I 
have many of the scenarios working including the Or scenario. There are 
scenarios

like the one [https://lists.apache.org/thread/7r49c5ffohz1oqm3oz2g60g5lnovq0l2] 
if the jdbc lookup source has rows that do not match the filter, then nulls 
come out in the lookup join resultset. The same behaviour occurs with predicate 
pushdown disabled and with my current patch (when I push it up).

On the discussion thread [~ruanhang1993]  agreed that the nulls are not correct.

 

The way I am reading this is that, unless we have a reason to see these nulls, 
there is an issue in lookup joins in certain scenarios. It does not appear to 
relate to JDBC. Are we ok if I make this fix work the same as when predicate 
pushdown is not enabled. And raise a separate issue for the lookup join which 
appears not to be JDBC related.  

 

 

 

> Missing filter condition in execution plan containing lookup join with mysql 
> jdbc connector
> ---
>
> Key: FLINK-33365
> URL: https://issues.apache.org/jira/browse/FLINK-33365
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / JDBC
>Affects Versions: 1.18.0, 1.17.1
> Environment: Flink 1.17.1 & Flink 1.18.0 with 
> flink-connector-jdbc-3.1.1-1.17.jar
>Reporter: macdoor615
>Assignee: david radley
>Priority: Critical
>  Labels: pull-request-available
> Attachments: flink-connector-jdbc-3.0.0-1.16.png, 
> flink-connector-jdbc-3.1.1-1.17.png
>
>
> create table in flink with sql-client.sh
> {code:java}
> CREATE TABLE default_catalog.default_database.a (
>   ip string, 
>   proctime as proctime()
> ) 
> WITH (
>   'connector' = 'datagen'
> );{code}
> create table in mysql
> {code:java}
> create table b (
>   ip varchar(20), 
>   type int
> );  {code}
>  
> Flink 1.17.1/ 1.18.0 and *flink-connector-jdbc-3.1.1-1.17.jar*
> excute in sql-client.sh 
> {code:java}
> explain SELECT * FROM default_catalog.default_database.a left join 
> bnpmp_mysql_test.gem_tmp.b FOR SYSTEM_TIME AS OF a.proctime on b.type = 0 and 
> a.ip = b.ip; {code}
> get the execution plan
> {code:java}
> ...
> == Optimized Execution Plan ==
> Calc(select=[ip, PROCTIME_MATERIALIZE(proctime) AS proctime, ip0, type])
> +- LookupJoin(table=[bnpmp_mysql_test.gem_tmp.b], joinType=[LeftOuterJoin], 
> lookup=[ip=ip], select=[ip, proctime, ip, CAST(0 AS INTEGER) AS type, CAST(ip 
> AS VARCHAR(2147483647)) AS ip0])
>    +- Calc(select=[ip, PROCTIME() AS proctime])
>       +- TableSourceScan(table=[[default_catalog, default_database, a]], 
> fields=[ip]){code}
>  
> excute same sql in sql-client with Flink 1.17.1/ 1.18.0 and 
> *flink-connector-jdbc-3.0.0-1.16.jar*
> get the execution plan
> {code:java}
> ...
> == Optimized Execution Plan ==
> Calc(select=[ip, PROCTIME_MATERIALIZE(proctime) AS proctime, ip0, type])
> +- LookupJoin(table=[bnpmp_mysql_test.gem_tmp.b], joinType=[LeftOuterJoin], 
> lookup=[type=0, ip=ip], where=[(type = 0)], select=[ip, proctime, ip, CAST(0 
> AS INTEGER) AS type, CAST(ip AS VARCHAR(2147483647)) AS ip0])
>    +- Calc(select=[ip, PROCTIME() AS proctime])
>       +- TableSourceScan(table=[[default_catalog, default_database, a]], 
> fields=[ip]) {code}
> with flink-connector-jdbc-3.1.1-1.17.jar,  the condition is 
> *lookup=[ip=ip]*
> with flink-connector-jdbc-3.0.0-1.16.jar ,  the condition is 
> *lookup=[type=0, ip=ip], where=[(type = 0)]*
>  
> In out real world production environment, this lead incorrect data output
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-22765) ExceptionUtilsITCase.testIsMetaspaceOutOfMemoryError is unstable

2023-12-18 Thread Sergey Nuyanzin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-22765?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798092#comment-17798092
 ] 

Sergey Nuyanzin commented on FLINK-22765:
-

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=55591&view=logs&j=a657ddbf-d986-5381-9649-342d9c92e7fb&t=dc085d4a-05c8-580e-06ab-21f5624dab16&l=8574

> ExceptionUtilsITCase.testIsMetaspaceOutOfMemoryError is unstable
> 
>
> Key: FLINK-22765
> URL: https://issues.apache.org/jira/browse/FLINK-22765
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.14.0, 1.13.5, 1.15.0, 1.17.2, 1.19.0
>Reporter: Robert Metzger
>Assignee: Robert Metzger
>Priority: Major
>  Labels: pull-request-available, stale-assigned, test-stability
> Fix For: 1.14.0, 1.16.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=18292&view=logs&j=39d5b1d5-3b41-54dc-6458-1e2ddd1cdcf3&t=a99e99c7-21cd-5a1f-7274-585e62b72f56
> {code}
> May 25 00:56:38 java.lang.AssertionError: 
> May 25 00:56:38 
> May 25 00:56:38 Expected: is ""
> May 25 00:56:38  but: was "The system is out of resources.\nConsult the 
> following stack trace for details."
> May 25 00:56:38   at 
> org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
> May 25 00:56:38   at org.junit.Assert.assertThat(Assert.java:956)
> May 25 00:56:38   at org.junit.Assert.assertThat(Assert.java:923)
> May 25 00:56:38   at 
> org.apache.flink.runtime.util.ExceptionUtilsITCase.run(ExceptionUtilsITCase.java:94)
> May 25 00:56:38   at 
> org.apache.flink.runtime.util.ExceptionUtilsITCase.testIsMetaspaceOutOfMemoryError(ExceptionUtilsITCase.java:70)
> May 25 00:56:38   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
> May 25 00:56:38   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> May 25 00:56:38   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> May 25 00:56:38   at java.lang.reflect.Method.invoke(Method.java:498)
> May 25 00:56:38   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> May 25 00:56:38   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> May 25 00:56:38   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> May 25 00:56:38   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> May 25 00:56:38   at 
> org.apache.flink.util.TestNameProvider$1.evaluate(TestNameProvider.java:45)
> May 25 00:56:38   at 
> org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
> May 25 00:56:38   at org.junit.rules.RunRules.evaluate(RunRules.java:20)
> May 25 00:56:38   at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> May 25 00:56:38   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> May 25 00:56:38   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> May 25 00:56:38   at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> May 25 00:56:38   at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> May 25 00:56:38   at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> May 25 00:56:38   at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> May 25 00:56:38   at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> May 25 00:56:38   at 
> org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
> May 25 00:56:38   at org.junit.rules.RunRules.evaluate(RunRules.java:20)
> May 25 00:56:38   at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> May 25 00:56:38   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
> May 25 00:56:38   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
> May 25 00:56:38   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
> May 25 00:56:38   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
> May 25 00:56:38   at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
> May 25 00:56:38   at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
> May 25 00:56:38   at 
> org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
> May 25 00:56:38   at 
> org.apache.maven.surefire.boote

Re: [PR] [FLINK-30593][autoscaler] Improve restart time tracking [flink-kubernetes-operator]

2023-12-18 Thread via GitHub


mxm commented on code in PR #735:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/735#discussion_r1429815765


##
flink-autoscaler/src/main/java/org/apache/flink/autoscaler/JobAutoScalerImpl.java:
##
@@ -182,22 +182,22 @@ private void runScalingLogic(Context ctx, 
AutoscalerFlinkMetrics autoscalerMetri
 jobTopology.getVerticesInTopologicalOrder(),
 () -> lastEvaluatedMetrics.get(ctx.getJobKey()));
 
-if (!collectedMetrics.isFullyCollected()) {
-// We have done an upfront evaluation, but we are not ready for 
scaling.
-resetRecommendedParallelism(evaluatedMetrics);
-return;
-}
-
 var scalingHistory = getTrimmedScalingHistory(stateStore, ctx, now);
 // A scaling tracking without an end time gets created whenever a 
scaling decision is
 // applied. Here, when the job transitions to RUNNING, we record the 
time for it.
 if (ctx.getJobStatus() == JobStatus.RUNNING) {
-if (scalingTracking.setEndTimeIfTrackedAndParallelismMatches(
+if 
(scalingTracking.recordRestartDurationIfTrackedAndParallelismMatches(
 now, jobTopology, scalingHistory)) {
 stateStore.storeScalingTracking(ctx, scalingTracking);
 }
 }

Review Comment:
   We don't need the RUNNING job state check. This block can be reduced to:
   
   ```java
 if 
(scalingTracking.recordRestartDurationIfTrackedAndParallelismMatches(
 now, jobTopology, scalingHistory)) {
 stateStore.storeScalingTracking(ctx, scalingTracking);
 }
   ```
   
   The reason is that this method only gets called when the job is in running 
state (see line 99). Enforcing a RUNNING state has always been a precondition 
for executing autoscaling.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-30593][autoscaler] Improve restart time tracking [flink-kubernetes-operator]

2023-12-18 Thread via GitHub


mxm commented on code in PR #735:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/735#discussion_r1429815765


##
flink-autoscaler/src/main/java/org/apache/flink/autoscaler/JobAutoScalerImpl.java:
##
@@ -182,22 +182,22 @@ private void runScalingLogic(Context ctx, 
AutoscalerFlinkMetrics autoscalerMetri
 jobTopology.getVerticesInTopologicalOrder(),
 () -> lastEvaluatedMetrics.get(ctx.getJobKey()));
 
-if (!collectedMetrics.isFullyCollected()) {
-// We have done an upfront evaluation, but we are not ready for 
scaling.
-resetRecommendedParallelism(evaluatedMetrics);
-return;
-}
-
 var scalingHistory = getTrimmedScalingHistory(stateStore, ctx, now);
 // A scaling tracking without an end time gets created whenever a 
scaling decision is
 // applied. Here, when the job transitions to RUNNING, we record the 
time for it.
 if (ctx.getJobStatus() == JobStatus.RUNNING) {
-if (scalingTracking.setEndTimeIfTrackedAndParallelismMatches(
+if 
(scalingTracking.recordRestartDurationIfTrackedAndParallelismMatches(
 now, jobTopology, scalingHistory)) {
 stateStore.storeScalingTracking(ctx, scalingTracking);
 }
 }

Review Comment:
   We don't need the RUNNING job state check. This block can be reduced to:
   
   ```java
 if 
(scalingTracking.recordRestartDurationIfTrackedAndParallelismMatches(
 now, jobTopology, scalingHistory)) {
 stateStore.storeScalingTracking(ctx, scalingTracking);
 }
   ```
   
   The reason is that this method only gets called when the job is in running 
state (see line 99). Enforcing a RUNNING state has always been a precondition 
for executing the autoscaling logic.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-30593][autoscaler] Improve restart time tracking [flink-kubernetes-operator]

2023-12-18 Thread via GitHub


mxm commented on PR #735:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/735#issuecomment-1859964401

   This needs a rebase. I'll run the tests afterwards.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-33706) Build_wheels_on_macos fails on AZP

2023-12-18 Thread Sergey Nuyanzin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33706?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798093#comment-17798093
 ] 

Sergey Nuyanzin commented on FLINK-33706:
-

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=55593&view=logs&j=f73b5736-8355-5390-ec71-4dfdec0ce6c5&t=90f7230e-bf5a-531b-8566-ad48d3e03bbb&l=815

> Build_wheels_on_macos fails on AZP
> --
>
> Key: FLINK-33706
> URL: https://issues.apache.org/jira/browse/FLINK-33706
> Project: Flink
>  Issue Type: Bug
>  Components: API / Python, Test Infrastructure
>Affects Versions: 1.19.0
>Reporter: Sergey Nuyanzin
>Priority: Major
>  Labels: test-stability
>
> This build 
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=55044&view=logs&j=f73b5736-8355-5390-ec71-4dfdec0ce6c5&t=90f7230e-bf5a-531b-8566-ad48d3e03bbb&l=427
> fails on AZP as 
> {noformat}
>note: This error originates from a subprocess, and is likely not a 
> problem with pip.
> ERROR: Failed cleaning build dir for crcmod
> Building wheel for dill (setup.py): started
> Building wheel for dill (setup.py): finished with status 'done'
> Created wheel for dill: filename=dill-0.0.0-py3-none-any.whl size=899 
> sha256=39d0b4b66ce11f42313482f4ad825029e861fd6dab87a743a95d75a44a1fedd6
> Stored in directory: 
> /Users/runner/Library/Caches/pip/wheels/07/35/78/e9004fa30578734db7f10e7a211605f3f0778d2bdde38a239d
> Building wheel for hdfs (setup.py): started
> Building wheel for hdfs (setup.py): finished with status 'done'
> Created wheel for hdfs: filename=UNKNOWN-0.0.0-py3-none-any.whl 
> size=928 
> sha256=cb3fd7d8c71b52bbc27cfb75842f9d4d9c6f3b847f3f4fe50323c945a0e38ccc
> Stored in directory: 
> /Users/runner/Library/Caches/pip/wheels/68/dd/29/c1a590238f9ebbe4f7ee9b3583f5185d0b9577e23f05c990eb
> WARNING: Built wheel for hdfs is invalid: Wheel has unexpected file 
> name: expected 'hdfs', got 'UNKNOWN'
> Building wheel for pymongo (pyproject.toml): started
> Building wheel for pymongo (pyproject.toml): finished with status 
> 'done'
> Created wheel for pymongo: 
> filename=pymongo-4.6.1-cp38-cp38-macosx_10_9_x86_64.whl size=478012 
> sha256=5dfc6fdb6a8a399f8f9da44e28bae19be244b15c8000cd3b2d7d6ff513cc6277
> Stored in directory: 
> /Users/runner/Library/Caches/pip/wheels/54/d8/0e/2a61e90bb3872d903b15eb3c94cb70f438fb8792a28fee7bb1
> Building wheel for docopt (setup.py): started
> Building wheel for docopt (setup.py): finished with status 'done'
> Created wheel for docopt: filename=UNKNOWN-0.0.0-py3-none-any.whl 
> size=920 
> sha256=612c56cd1a6344b8def6c4d3c3c1c8bb10e1f2b0d978fee0fc8b9281026e8288
> Stored in directory: 
> /Users/runner/Library/Caches/pip/wheels/56/ea/58/ead137b087d9e326852a851351d1debf4ada529b6ac0ec4e8c
> WARNING: Built wheel for docopt is invalid: Wheel has unexpected file 
> name: expected 'docopt', got 'UNKNOWN'
>   Successfully built dill pymongo
>   Failed to build fastavro crcmod hdfs docopt
>   ERROR: Could not build wheels for fastavro, which is required to 
> install pyproject.toml-based projects
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-31472) AsyncSinkWriterThrottlingTest failed with Illegal mailbox thread

2023-12-18 Thread Sergey Nuyanzin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31472?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798094#comment-17798094
 ] 

Sergey Nuyanzin commented on FLINK-31472:
-

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=55601&view=logs&j=1c002d28-a73d-5309-26ee-10036d8476b4&t=d1c117a6-8f13-5466-55f0-d48dbb767fcd&l=10380

> AsyncSinkWriterThrottlingTest failed with Illegal mailbox thread
> 
>
> Key: FLINK-31472
> URL: https://issues.apache.org/jira/browse/FLINK-31472
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Common
>Affects Versions: 1.17.0, 1.16.1, 1.19.0
>Reporter: Ran Tao
>Priority: Major
>  Labels: test-stability
>
> when run mvn clean test, this case failed occasionally.
> {noformat}
> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.955 
> s <<< FAILURE! - in 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest
> [ERROR] 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest.testSinkThroughputShouldThrottleToHalfBatchSize
>   Time elapsed: 0.492 s  <<< ERROR!
> java.lang.IllegalStateException: Illegal thread detected. This method must be 
> called from inside the mailbox thread!
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.TaskMailboxImpl.checkIsMailboxThread(TaskMailboxImpl.java:262)
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.TaskMailboxImpl.take(TaskMailboxImpl.java:137)
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxExecutorImpl.yield(MailboxExecutorImpl.java:84)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriter.flush(AsyncSinkWriter.java:367)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriter.lambda$registerCallback$3(AsyncSinkWriter.java:315)
>         at 
> org.apache.flink.streaming.runtime.tasks.TestProcessingTimeService$CallbackTask.onProcessingTime(TestProcessingTimeService.java:199)
>         at 
> org.apache.flink.streaming.runtime.tasks.TestProcessingTimeService.setCurrentTime(TestProcessingTimeService.java:76)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest.testSinkThroughputShouldThrottleToHalfBatchSize(AsyncSinkWriterThrottlingTest.java:64)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>         at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:498)
>         at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
>         at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>         at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
>         at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>         at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
>         at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
>         at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
>         at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
>         at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
>         at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
>         at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
>         at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
>         at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
>         at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
>         at org.junit.runner.JUnitCore.run(JUnitCore.java:115)
>         at 
> org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42)
>         at 
> org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80)
>         at 
> org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:72)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:147)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:127)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(Engi

[jira] [Updated] (FLINK-31472) AsyncSinkWriterThrottlingTest failed with Illegal mailbox thread

2023-12-18 Thread Sergey Nuyanzin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Nuyanzin updated FLINK-31472:

Description: 
when run mvn clean test, this case failed occasionally.
{noformat}
[ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.955 s 
<<< FAILURE! - in 
org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest
[ERROR] 
org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest.testSinkThroughputShouldThrottleToHalfBatchSize
  Time elapsed: 0.492 s  <<< ERROR!
java.lang.IllegalStateException: Illegal thread detected. This method must be 
called from inside the mailbox thread!
        at 
org.apache.flink.streaming.runtime.tasks.mailbox.TaskMailboxImpl.checkIsMailboxThread(TaskMailboxImpl.java:262)
        at 
org.apache.flink.streaming.runtime.tasks.mailbox.TaskMailboxImpl.take(TaskMailboxImpl.java:137)
        at 
org.apache.flink.streaming.runtime.tasks.mailbox.MailboxExecutorImpl.yield(MailboxExecutorImpl.java:84)
        at 
org.apache.flink.connector.base.sink.writer.AsyncSinkWriter.flush(AsyncSinkWriter.java:367)
        at 
org.apache.flink.connector.base.sink.writer.AsyncSinkWriter.lambda$registerCallback$3(AsyncSinkWriter.java:315)
        at 
org.apache.flink.streaming.runtime.tasks.TestProcessingTimeService$CallbackTask.onProcessingTime(TestProcessingTimeService.java:199)
        at 
org.apache.flink.streaming.runtime.tasks.TestProcessingTimeService.setCurrentTime(TestProcessingTimeService.java:76)
        at 
org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest.testSinkThroughputShouldThrottleToHalfBatchSize(AsyncSinkWriterThrottlingTest.java:64)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:498)
        at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
        at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
        at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
        at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
        at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
        at 
org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
        at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
        at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
        at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
        at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
        at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
        at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
        at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
        at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
        at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
        at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
        at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
        at org.junit.runner.JUnitCore.run(JUnitCore.java:115)
        at 
org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42)
        at 
org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80)
        at 
org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:72)
        at 
org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:147)
        at 
org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:127)
        at 
org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:90)
        at 
org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:55)
        at 
org.junit.platform.launcher.core.EngineExecutionOrchestrator.withInterceptedStreams(EngineExecutionOrchestrator.java:102)
        at 
org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:54)
        at 
org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:114)
        at 
org.junit.platform.launcher.core.DefaultLauncher.execute(DefaultLauncher.java:86)
        at 
org.junit.platform.launcher.core.DefaultLauncherSession$DelegatingLauncher.execute(DefaultLauncherSession.java:86)
        at 
org.junit.platform.launcher.core.SessionPerRequestLauncher.execute(SessionPerRequestLauncher.java:53)

[jira] [Commented] (FLINK-22765) ExceptionUtilsITCase.testIsMetaspaceOutOfMemoryError is unstable

2023-12-18 Thread Sergey Nuyanzin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-22765?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798095#comment-17798095
 ] 

Sergey Nuyanzin commented on FLINK-22765:
-

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=55601&view=logs&j=a657ddbf-d986-5381-9649-342d9c92e7fb&t=dc085d4a-05c8-580e-06ab-21f5624dab16&l=8573

> ExceptionUtilsITCase.testIsMetaspaceOutOfMemoryError is unstable
> 
>
> Key: FLINK-22765
> URL: https://issues.apache.org/jira/browse/FLINK-22765
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.14.0, 1.13.5, 1.15.0, 1.17.2, 1.19.0
>Reporter: Robert Metzger
>Assignee: Robert Metzger
>Priority: Major
>  Labels: pull-request-available, stale-assigned, test-stability
> Fix For: 1.14.0, 1.16.0
>
>
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=18292&view=logs&j=39d5b1d5-3b41-54dc-6458-1e2ddd1cdcf3&t=a99e99c7-21cd-5a1f-7274-585e62b72f56
> {code}
> May 25 00:56:38 java.lang.AssertionError: 
> May 25 00:56:38 
> May 25 00:56:38 Expected: is ""
> May 25 00:56:38  but: was "The system is out of resources.\nConsult the 
> following stack trace for details."
> May 25 00:56:38   at 
> org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
> May 25 00:56:38   at org.junit.Assert.assertThat(Assert.java:956)
> May 25 00:56:38   at org.junit.Assert.assertThat(Assert.java:923)
> May 25 00:56:38   at 
> org.apache.flink.runtime.util.ExceptionUtilsITCase.run(ExceptionUtilsITCase.java:94)
> May 25 00:56:38   at 
> org.apache.flink.runtime.util.ExceptionUtilsITCase.testIsMetaspaceOutOfMemoryError(ExceptionUtilsITCase.java:70)
> May 25 00:56:38   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
> May 25 00:56:38   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> May 25 00:56:38   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> May 25 00:56:38   at java.lang.reflect.Method.invoke(Method.java:498)
> May 25 00:56:38   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
> May 25 00:56:38   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> May 25 00:56:38   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
> May 25 00:56:38   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> May 25 00:56:38   at 
> org.apache.flink.util.TestNameProvider$1.evaluate(TestNameProvider.java:45)
> May 25 00:56:38   at 
> org.junit.rules.TestWatcher$1.evaluate(TestWatcher.java:55)
> May 25 00:56:38   at org.junit.rules.RunRules.evaluate(RunRules.java:20)
> May 25 00:56:38   at 
> org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
> May 25 00:56:38   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
> May 25 00:56:38   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
> May 25 00:56:38   at 
> org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
> May 25 00:56:38   at 
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
> May 25 00:56:38   at 
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
> May 25 00:56:38   at 
> org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
> May 25 00:56:38   at 
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
> May 25 00:56:38   at 
> org.junit.rules.ExternalResource$1.evaluate(ExternalResource.java:48)
> May 25 00:56:38   at org.junit.rules.RunRules.evaluate(RunRules.java:20)
> May 25 00:56:38   at 
> org.junit.runners.ParentRunner.run(ParentRunner.java:363)
> May 25 00:56:38   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
> May 25 00:56:38   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
> May 25 00:56:38   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
> May 25 00:56:38   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
> May 25 00:56:38   at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
> May 25 00:56:38   at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
> May 25 00:56:38   at 
> org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
> May 25 00:56:38   at 
> org.apache.maven.surefire.boote

[jira] [Commented] (FLINK-31472) AsyncSinkWriterThrottlingTest failed with Illegal mailbox thread

2023-12-18 Thread Sergey Nuyanzin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31472?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798096#comment-17798096
 ] 

Sergey Nuyanzin commented on FLINK-31472:
-

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=55602&view=logs&j=1c002d28-a73d-5309-26ee-10036d8476b4&t=d1c117a6-8f13-5466-55f0-d48dbb767fcd&l=10634

> AsyncSinkWriterThrottlingTest failed with Illegal mailbox thread
> 
>
> Key: FLINK-31472
> URL: https://issues.apache.org/jira/browse/FLINK-31472
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Common
>Affects Versions: 1.17.0, 1.16.1, 1.19.0
>Reporter: Ran Tao
>Priority: Major
>  Labels: test-stability
>
> when run mvn clean test, this case failed occasionally.
> {noformat}
> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.955 
> s <<< FAILURE! - in 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest
> [ERROR] 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest.testSinkThroughputShouldThrottleToHalfBatchSize
>   Time elapsed: 0.492 s  <<< ERROR!
> java.lang.IllegalStateException: Illegal thread detected. This method must be 
> called from inside the mailbox thread!
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.TaskMailboxImpl.checkIsMailboxThread(TaskMailboxImpl.java:262)
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.TaskMailboxImpl.take(TaskMailboxImpl.java:137)
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxExecutorImpl.yield(MailboxExecutorImpl.java:84)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriter.flush(AsyncSinkWriter.java:367)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriter.lambda$registerCallback$3(AsyncSinkWriter.java:315)
>         at 
> org.apache.flink.streaming.runtime.tasks.TestProcessingTimeService$CallbackTask.onProcessingTime(TestProcessingTimeService.java:199)
>         at 
> org.apache.flink.streaming.runtime.tasks.TestProcessingTimeService.setCurrentTime(TestProcessingTimeService.java:76)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest.testSinkThroughputShouldThrottleToHalfBatchSize(AsyncSinkWriterThrottlingTest.java:64)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>         at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:498)
>         at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
>         at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>         at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
>         at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>         at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
>         at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
>         at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
>         at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
>         at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
>         at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
>         at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
>         at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
>         at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
>         at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
>         at org.junit.runner.JUnitCore.run(JUnitCore.java:115)
>         at 
> org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42)
>         at 
> org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80)
>         at 
> org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:72)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:147)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:127)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(Engi

Re: [PR] [FLINK-33795] Add a new config to forbid autoscale execution in the configured excluded periods [flink-kubernetes-operator]

2023-12-18 Thread via GitHub


mxm commented on code in PR #728:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/728#discussion_r1426756716


##
flink-autoscaler/src/main/java/org/apache/flink/autoscaler/utils/AutoScalerUtils.java:
##
@@ -94,4 +99,106 @@ public static boolean excludeVerticesFromScaling(
 conf.set(AutoScalerOptions.VERTEX_EXCLUDE_IDS, new 
ArrayList<>(excludedIds));
 return anyAdded;
 }
+
+/** Quartz doesn't have the invertTimeRange flag so rewrite this method. */
+static boolean isTimeIncluded(CronCalendar cron, long timeInMillis) {
+if (cron.getBaseCalendar() != null
+&& !cron.getBaseCalendar().isTimeIncluded(timeInMillis)) {
+return false;
+} else {
+return cron.getCronExpression().isSatisfiedBy(new 
Date(timeInMillis));
+}
+}
+
+static Optional interpretAsDaily(String subExpression) {
+String[] splits = subExpression.split("-");
+if (splits.length != 2) {
+return Optional.empty();
+}
+try {
+DailyCalendar daily = new DailyCalendar(splits[0], splits[1]);
+daily.setInvertTimeRange(true);
+return Optional.of(daily);
+} catch (Exception e) {
+return Optional.empty();
+}
+}
+
+static Optional interpretAsCron(String subExpression) {
+try {
+return Optional.of(new CronCalendar(subExpression));
+} catch (Exception e) {
+return Optional.empty();

Review Comment:
   I think we should log this exception. Otherwise it is going to be hard to 
figure out what is wrong with the cron string. This also applies to all other 
methods in this class which have this pattern.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-32523) NotifyCheckpointAbortedITCase.testNotifyCheckpointAborted fails with timeout on AZP

2023-12-18 Thread Sergey Nuyanzin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-32523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798098#comment-17798098
 ] 

Sergey Nuyanzin commented on FLINK-32523:
-

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=55602&view=logs&j=8fd9202e-fd17-5b26-353c-ac1ff76c8f28&t=ea7cf968-e585-52cb-e0fc-f48de023a7ca&l=8330

> NotifyCheckpointAbortedITCase.testNotifyCheckpointAborted fails with timeout 
> on AZP
> ---
>
> Key: FLINK-32523
> URL: https://issues.apache.org/jira/browse/FLINK-32523
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.16.2, 1.18.0, 1.17.1, 1.19.0
>Reporter: Sergey Nuyanzin
>Assignee: Hangxiang Yu
>Priority: Critical
>  Labels: pull-request-available, stale-assigned, test-stability
> Attachments: failure.log
>
>
> This build
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=50795&view=logs&j=39d5b1d5-3b41-54dc-6458-1e2ddd1cdcf3&t=0c010d0c-3dec-5bf1-d408-7b18988b1b2b&l=8638
>  fails with timeout
> {noformat}
> Jul 03 01:26:35 org.junit.runners.model.TestTimedOutException: test timed out 
> after 10 milliseconds
> Jul 03 01:26:35   at java.lang.Object.wait(Native Method)
> Jul 03 01:26:35   at java.lang.Object.wait(Object.java:502)
> Jul 03 01:26:35   at 
> org.apache.flink.core.testutils.OneShotLatch.await(OneShotLatch.java:61)
> Jul 03 01:26:35   at 
> org.apache.flink.test.checkpointing.NotifyCheckpointAbortedITCase.verifyAllOperatorsNotifyAborted(NotifyCheckpointAbortedITCase.java:198)
> Jul 03 01:26:35   at 
> org.apache.flink.test.checkpointing.NotifyCheckpointAbortedITCase.testNotifyCheckpointAborted(NotifyCheckpointAbortedITCase.java:189)
> Jul 03 01:26:35   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
> Jul 03 01:26:35   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> Jul 03 01:26:35   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> Jul 03 01:26:35   at java.lang.reflect.Method.invoke(Method.java:498)
> Jul 03 01:26:35   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
> Jul 03 01:26:35   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> Jul 03 01:26:35   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
> Jul 03 01:26:35   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> Jul 03 01:26:35   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
> Jul 03 01:26:35   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
> Jul 03 01:26:35   at 
> java.util.concurrent.FutureTask.run(FutureTask.java:266)
> Jul 03 01:26:35   at java.lang.Thread.run(Thread.java:748)
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-32523) NotifyCheckpointAbortedITCase.testNotifyCheckpointAborted fails with timeout on AZP

2023-12-18 Thread Sergey Nuyanzin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-32523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798098#comment-17798098
 ] 

Sergey Nuyanzin edited comment on FLINK-32523 at 12/18/23 9:59 AM:
---

1.18: 
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=55602&view=logs&j=8fd9202e-fd17-5b26-353c-ac1ff76c8f28&t=ea7cf968-e585-52cb-e0fc-f48de023a7ca&l=8330


was (Author: sergey nuyanzin):
https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=55602&view=logs&j=8fd9202e-fd17-5b26-353c-ac1ff76c8f28&t=ea7cf968-e585-52cb-e0fc-f48de023a7ca&l=8330

> NotifyCheckpointAbortedITCase.testNotifyCheckpointAborted fails with timeout 
> on AZP
> ---
>
> Key: FLINK-32523
> URL: https://issues.apache.org/jira/browse/FLINK-32523
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.16.2, 1.18.0, 1.17.1, 1.19.0
>Reporter: Sergey Nuyanzin
>Assignee: Hangxiang Yu
>Priority: Critical
>  Labels: pull-request-available, stale-assigned, test-stability
> Attachments: failure.log
>
>
> This build
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=50795&view=logs&j=39d5b1d5-3b41-54dc-6458-1e2ddd1cdcf3&t=0c010d0c-3dec-5bf1-d408-7b18988b1b2b&l=8638
>  fails with timeout
> {noformat}
> Jul 03 01:26:35 org.junit.runners.model.TestTimedOutException: test timed out 
> after 10 milliseconds
> Jul 03 01:26:35   at java.lang.Object.wait(Native Method)
> Jul 03 01:26:35   at java.lang.Object.wait(Object.java:502)
> Jul 03 01:26:35   at 
> org.apache.flink.core.testutils.OneShotLatch.await(OneShotLatch.java:61)
> Jul 03 01:26:35   at 
> org.apache.flink.test.checkpointing.NotifyCheckpointAbortedITCase.verifyAllOperatorsNotifyAborted(NotifyCheckpointAbortedITCase.java:198)
> Jul 03 01:26:35   at 
> org.apache.flink.test.checkpointing.NotifyCheckpointAbortedITCase.testNotifyCheckpointAborted(NotifyCheckpointAbortedITCase.java:189)
> Jul 03 01:26:35   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)
> Jul 03 01:26:35   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> Jul 03 01:26:35   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> Jul 03 01:26:35   at java.lang.reflect.Method.invoke(Method.java:498)
> Jul 03 01:26:35   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
> Jul 03 01:26:35   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> Jul 03 01:26:35   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
> Jul 03 01:26:35   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> Jul 03 01:26:35   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:299)
> Jul 03 01:26:35   at 
> org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:293)
> Jul 03 01:26:35   at 
> java.util.concurrent.FutureTask.run(FutureTask.java:266)
> Jul 03 01:26:35   at java.lang.Thread.run(Thread.java:748)
> {noformat}



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33795] Add a new config to forbid autoscale execution in the configured excluded periods [flink-kubernetes-operator]

2023-12-18 Thread via GitHub


mxm commented on code in PR #728:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/728#discussion_r1429838068


##
flink-autoscaler/src/main/java/org/apache/flink/autoscaler/ScalingMetricCollector.java:
##
@@ -143,7 +145,11 @@ public CollectedMetricHistory updateMetrics(
 if (now.isBefore(windowFullTime)) {
 LOG.info("Metric window not full until {}", windowFullTime);
 } else {
-collectedMetrics.setFullyCollected(true);
+if (isExcluded) {
+LOG.info("autoscaling now in excluded period");

Review Comment:
   ```suggestion
   LOG.info("Autoscaling on halt based on exclusion rule {}",
   conf.get(AutoScalerOptions. EXCLUDED_PERIODS));
   ```
   
   I think this could be helpful for debugging.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33795] Add a new config to forbid autoscale execution in the configured excluded periods [flink-kubernetes-operator]

2023-12-18 Thread via GitHub


mxm commented on PR #728:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/728#issuecomment-1859986239

   Thanks @flashJd! Changes look great.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-32877][Filesystem]add HTTP options to gcs-cloud-storage client [flink]

2023-12-18 Thread via GitHub


singhravidutt commented on PR #23226:
URL: https://github.com/apache/flink/pull/23226#issuecomment-1859988955

   > Also @singhravidutt what's your Jira user name, so I can assign this 
ticket to you?
   
   I don't have a Jira account. @cnauroth can you take the ownership of jira 
ticket?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-33449) Support array_contains_seq function

2023-12-18 Thread leo cheng (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33449?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798100#comment-17798100
 ] 

leo cheng commented on FLINK-33449:
---

Hi [~xuyangzhong] Thank you

Pls do it.

> Support array_contains_seq function
> ---
>
> Key: FLINK-33449
> URL: https://issues.apache.org/jira/browse/FLINK-33449
> Project: Flink
>  Issue Type: New Feature
>Reporter: leo cheng
>Priority: Minor
>  Labels: pull-request-available
>
> support function array_contains_seq like trino contains_sequence
> trino: 
> https://trino.io/docs/current/functions/array.html?highlight=contains_sequence#contains_sequence



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-31472) AsyncSinkWriterThrottlingTest failed with Illegal mailbox thread

2023-12-18 Thread Ahmed Hamdy (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31472?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798101#comment-17798101
 ] 

Ahmed Hamdy commented on FLINK-31472:
-

hi [~Sergey Nuyanzin] could you assign me the ticket, I will take a look ASAP.

> AsyncSinkWriterThrottlingTest failed with Illegal mailbox thread
> 
>
> Key: FLINK-31472
> URL: https://issues.apache.org/jira/browse/FLINK-31472
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Common
>Affects Versions: 1.17.0, 1.16.1, 1.19.0
>Reporter: Ran Tao
>Priority: Major
>  Labels: test-stability
>
> when run mvn clean test, this case failed occasionally.
> {noformat}
> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.955 
> s <<< FAILURE! - in 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest
> [ERROR] 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest.testSinkThroughputShouldThrottleToHalfBatchSize
>   Time elapsed: 0.492 s  <<< ERROR!
> java.lang.IllegalStateException: Illegal thread detected. This method must be 
> called from inside the mailbox thread!
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.TaskMailboxImpl.checkIsMailboxThread(TaskMailboxImpl.java:262)
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.TaskMailboxImpl.take(TaskMailboxImpl.java:137)
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxExecutorImpl.yield(MailboxExecutorImpl.java:84)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriter.flush(AsyncSinkWriter.java:367)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriter.lambda$registerCallback$3(AsyncSinkWriter.java:315)
>         at 
> org.apache.flink.streaming.runtime.tasks.TestProcessingTimeService$CallbackTask.onProcessingTime(TestProcessingTimeService.java:199)
>         at 
> org.apache.flink.streaming.runtime.tasks.TestProcessingTimeService.setCurrentTime(TestProcessingTimeService.java:76)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest.testSinkThroughputShouldThrottleToHalfBatchSize(AsyncSinkWriterThrottlingTest.java:64)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>         at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:498)
>         at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
>         at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>         at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
>         at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>         at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
>         at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
>         at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
>         at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
>         at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
>         at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
>         at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
>         at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
>         at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
>         at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
>         at org.junit.runner.JUnitCore.run(JUnitCore.java:115)
>         at 
> org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42)
>         at 
> org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80)
>         at 
> org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:72)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:147)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:127)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:90)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrche

[jira] [Commented] (FLINK-33858) CI fails with No space left on device

2023-12-18 Thread Jing Ge (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33858?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798102#comment-17798102
 ] 

Jing Ge commented on FLINK-33858:
-

Does anyone know if we could just delete 
/home/agent03/myagent/_work/_temp/debug_files/mvn-1.log? Or we must archive 
i.e. zip it? Thanks!

> CI fails with No space left on device
> -
>
> Key: FLINK-33858
> URL: https://issues.apache.org/jira/browse/FLINK-33858
> Project: Flink
>  Issue Type: Bug
>  Components: Build System / CI
>Reporter: Sergey Nuyanzin
>Priority: Blocker
>
> AlibabaCI003-agent01
> AlibabaCI003-agent03
> AlibabaCI003-agent05
> [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=8&view=logs&j=0c940707-2659-5648-cbe6-a1ad63045f0a&t=075c2716-8010-5565-fe08-3c4bb45824a4&l=9765]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-33365) Missing filter condition in execution plan containing lookup join with mysql jdbc connector

2023-12-18 Thread Sergey Nuyanzin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33365?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798105#comment-17798105
 ] 

Sergey Nuyanzin edited comment on FLINK-33365 at 12/18/23 10:25 AM:


Since 
1. predicate push down from one side brings so many issues and probably there 
is something else which is not known yet.
2. IIRC this issue was a blocker before jdbc connector release

I tend to think 
1. to revert/disable push down commit 
((https://github.com/apache/flink-connector-jdbc/commit/3e3b40e8cfadfc16a8ab74d4ef6a3ab3ceafa57b)
2. make a release
3. continue working on pushdown

what do you think?




was (Author: sergey nuyanzin):
Since 
1. predicate push down from one side brings so many issues and probably there 
is something else which is not known yet.
2. IIRC this issue was a blocker before jdbc connector release

I tend to think 
1. to revert push down commit 
((https://github.com/apache/flink-connector-jdbc/commit/3e3b40e8cfadfc16a8ab74d4ef6a3ab3ceafa57b)
2. make a release
3. continue working on pushdown

what do you think?



> Missing filter condition in execution plan containing lookup join with mysql 
> jdbc connector
> ---
>
> Key: FLINK-33365
> URL: https://issues.apache.org/jira/browse/FLINK-33365
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / JDBC
>Affects Versions: 1.18.0, 1.17.1
> Environment: Flink 1.17.1 & Flink 1.18.0 with 
> flink-connector-jdbc-3.1.1-1.17.jar
>Reporter: macdoor615
>Assignee: david radley
>Priority: Critical
>  Labels: pull-request-available
> Attachments: flink-connector-jdbc-3.0.0-1.16.png, 
> flink-connector-jdbc-3.1.1-1.17.png
>
>
> create table in flink with sql-client.sh
> {code:java}
> CREATE TABLE default_catalog.default_database.a (
>   ip string, 
>   proctime as proctime()
> ) 
> WITH (
>   'connector' = 'datagen'
> );{code}
> create table in mysql
> {code:java}
> create table b (
>   ip varchar(20), 
>   type int
> );  {code}
>  
> Flink 1.17.1/ 1.18.0 and *flink-connector-jdbc-3.1.1-1.17.jar*
> excute in sql-client.sh 
> {code:java}
> explain SELECT * FROM default_catalog.default_database.a left join 
> bnpmp_mysql_test.gem_tmp.b FOR SYSTEM_TIME AS OF a.proctime on b.type = 0 and 
> a.ip = b.ip; {code}
> get the execution plan
> {code:java}
> ...
> == Optimized Execution Plan ==
> Calc(select=[ip, PROCTIME_MATERIALIZE(proctime) AS proctime, ip0, type])
> +- LookupJoin(table=[bnpmp_mysql_test.gem_tmp.b], joinType=[LeftOuterJoin], 
> lookup=[ip=ip], select=[ip, proctime, ip, CAST(0 AS INTEGER) AS type, CAST(ip 
> AS VARCHAR(2147483647)) AS ip0])
>    +- Calc(select=[ip, PROCTIME() AS proctime])
>       +- TableSourceScan(table=[[default_catalog, default_database, a]], 
> fields=[ip]){code}
>  
> excute same sql in sql-client with Flink 1.17.1/ 1.18.0 and 
> *flink-connector-jdbc-3.0.0-1.16.jar*
> get the execution plan
> {code:java}
> ...
> == Optimized Execution Plan ==
> Calc(select=[ip, PROCTIME_MATERIALIZE(proctime) AS proctime, ip0, type])
> +- LookupJoin(table=[bnpmp_mysql_test.gem_tmp.b], joinType=[LeftOuterJoin], 
> lookup=[type=0, ip=ip], where=[(type = 0)], select=[ip, proctime, ip, CAST(0 
> AS INTEGER) AS type, CAST(ip AS VARCHAR(2147483647)) AS ip0])
>    +- Calc(select=[ip, PROCTIME() AS proctime])
>       +- TableSourceScan(table=[[default_catalog, default_database, a]], 
> fields=[ip]) {code}
> with flink-connector-jdbc-3.1.1-1.17.jar,  the condition is 
> *lookup=[ip=ip]*
> with flink-connector-jdbc-3.0.0-1.16.jar ,  the condition is 
> *lookup=[type=0, ip=ip], where=[(type = 0)]*
>  
> In out real world production environment, this lead incorrect data output
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33365) Missing filter condition in execution plan containing lookup join with mysql jdbc connector

2023-12-18 Thread Sergey Nuyanzin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33365?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798105#comment-17798105
 ] 

Sergey Nuyanzin commented on FLINK-33365:
-

Since 
1. predicate push down from one side brings so many issues and probably there 
is something else which is not known yet.
2. IIRC this issue was a blocker before jdbc connector release

I tend to think 
1. to revert push down commit 
((https://github.com/apache/flink-connector-jdbc/commit/3e3b40e8cfadfc16a8ab74d4ef6a3ab3ceafa57b)
2. make a release
3. continue working on pushdown

what do you think?



> Missing filter condition in execution plan containing lookup join with mysql 
> jdbc connector
> ---
>
> Key: FLINK-33365
> URL: https://issues.apache.org/jira/browse/FLINK-33365
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / JDBC
>Affects Versions: 1.18.0, 1.17.1
> Environment: Flink 1.17.1 & Flink 1.18.0 with 
> flink-connector-jdbc-3.1.1-1.17.jar
>Reporter: macdoor615
>Assignee: david radley
>Priority: Critical
>  Labels: pull-request-available
> Attachments: flink-connector-jdbc-3.0.0-1.16.png, 
> flink-connector-jdbc-3.1.1-1.17.png
>
>
> create table in flink with sql-client.sh
> {code:java}
> CREATE TABLE default_catalog.default_database.a (
>   ip string, 
>   proctime as proctime()
> ) 
> WITH (
>   'connector' = 'datagen'
> );{code}
> create table in mysql
> {code:java}
> create table b (
>   ip varchar(20), 
>   type int
> );  {code}
>  
> Flink 1.17.1/ 1.18.0 and *flink-connector-jdbc-3.1.1-1.17.jar*
> excute in sql-client.sh 
> {code:java}
> explain SELECT * FROM default_catalog.default_database.a left join 
> bnpmp_mysql_test.gem_tmp.b FOR SYSTEM_TIME AS OF a.proctime on b.type = 0 and 
> a.ip = b.ip; {code}
> get the execution plan
> {code:java}
> ...
> == Optimized Execution Plan ==
> Calc(select=[ip, PROCTIME_MATERIALIZE(proctime) AS proctime, ip0, type])
> +- LookupJoin(table=[bnpmp_mysql_test.gem_tmp.b], joinType=[LeftOuterJoin], 
> lookup=[ip=ip], select=[ip, proctime, ip, CAST(0 AS INTEGER) AS type, CAST(ip 
> AS VARCHAR(2147483647)) AS ip0])
>    +- Calc(select=[ip, PROCTIME() AS proctime])
>       +- TableSourceScan(table=[[default_catalog, default_database, a]], 
> fields=[ip]){code}
>  
> excute same sql in sql-client with Flink 1.17.1/ 1.18.0 and 
> *flink-connector-jdbc-3.0.0-1.16.jar*
> get the execution plan
> {code:java}
> ...
> == Optimized Execution Plan ==
> Calc(select=[ip, PROCTIME_MATERIALIZE(proctime) AS proctime, ip0, type])
> +- LookupJoin(table=[bnpmp_mysql_test.gem_tmp.b], joinType=[LeftOuterJoin], 
> lookup=[type=0, ip=ip], where=[(type = 0)], select=[ip, proctime, ip, CAST(0 
> AS INTEGER) AS type, CAST(ip AS VARCHAR(2147483647)) AS ip0])
>    +- Calc(select=[ip, PROCTIME() AS proctime])
>       +- TableSourceScan(table=[[default_catalog, default_database, a]], 
> fields=[ip]) {code}
> with flink-connector-jdbc-3.1.1-1.17.jar,  the condition is 
> *lookup=[ip=ip]*
> with flink-connector-jdbc-3.0.0-1.16.jar ,  the condition is 
> *lookup=[type=0, ip=ip], where=[(type = 0)]*
>  
> In out real world production environment, this lead incorrect data output
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33858) CI fails with No space left on device

2023-12-18 Thread Sergey Nuyanzin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33858?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798108#comment-17798108
 ] 

Sergey Nuyanzin commented on FLINK-33858:
-

zip would be the safest

otherwise if it relates to successful build then it could be removed
if it relates to failure then probably would make sense probably to attach here 
zip version for all mvn-*, not only mvn-1.log in case the size is suitable 

> CI fails with No space left on device
> -
>
> Key: FLINK-33858
> URL: https://issues.apache.org/jira/browse/FLINK-33858
> Project: Flink
>  Issue Type: Bug
>  Components: Build System / CI
>Reporter: Sergey Nuyanzin
>Priority: Blocker
>
> AlibabaCI003-agent01
> AlibabaCI003-agent03
> AlibabaCI003-agent05
> [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=8&view=logs&j=0c940707-2659-5648-cbe6-a1ad63045f0a&t=075c2716-8010-5565-fe08-3c4bb45824a4&l=9765]



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33704) Update GCS filesystems to latest available versions

2023-12-18 Thread Jing Ge (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33704?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798109#comment-17798109
 ] 

Jing Ge commented on FLINK-33704:
-

Thanks [~martijnvisser] for fixing it! Can we resolve the ticket now?

> Update GCS filesystems to latest available versions
> ---
>
> Key: FLINK-33704
> URL: https://issues.apache.org/jira/browse/FLINK-33704
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Connectors / FileSystem, FileSystems
>Reporter: Martijn Visser
>Assignee: Martijn Visser
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
>
> Update GS SDK from 2.15.0 to 2.29.1 and GS Hadoop Connector from 2.2.15 to 
> 2.2.18



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-31472) AsyncSinkWriterThrottlingTest failed with Illegal mailbox thread

2023-12-18 Thread Sergey Nuyanzin (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sergey Nuyanzin reassigned FLINK-31472:
---

Assignee: Ahmed Hamdy

> AsyncSinkWriterThrottlingTest failed with Illegal mailbox thread
> 
>
> Key: FLINK-31472
> URL: https://issues.apache.org/jira/browse/FLINK-31472
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Common
>Affects Versions: 1.17.0, 1.16.1, 1.19.0
>Reporter: Ran Tao
>Assignee: Ahmed Hamdy
>Priority: Major
>  Labels: test-stability
>
> when run mvn clean test, this case failed occasionally.
> {noformat}
> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.955 
> s <<< FAILURE! - in 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest
> [ERROR] 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest.testSinkThroughputShouldThrottleToHalfBatchSize
>   Time elapsed: 0.492 s  <<< ERROR!
> java.lang.IllegalStateException: Illegal thread detected. This method must be 
> called from inside the mailbox thread!
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.TaskMailboxImpl.checkIsMailboxThread(TaskMailboxImpl.java:262)
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.TaskMailboxImpl.take(TaskMailboxImpl.java:137)
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxExecutorImpl.yield(MailboxExecutorImpl.java:84)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriter.flush(AsyncSinkWriter.java:367)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriter.lambda$registerCallback$3(AsyncSinkWriter.java:315)
>         at 
> org.apache.flink.streaming.runtime.tasks.TestProcessingTimeService$CallbackTask.onProcessingTime(TestProcessingTimeService.java:199)
>         at 
> org.apache.flink.streaming.runtime.tasks.TestProcessingTimeService.setCurrentTime(TestProcessingTimeService.java:76)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest.testSinkThroughputShouldThrottleToHalfBatchSize(AsyncSinkWriterThrottlingTest.java:64)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>         at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:498)
>         at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
>         at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>         at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
>         at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>         at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
>         at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
>         at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
>         at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
>         at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
>         at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
>         at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
>         at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
>         at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
>         at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
>         at org.junit.runner.JUnitCore.run(JUnitCore.java:115)
>         at 
> org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42)
>         at 
> org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80)
>         at 
> org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:72)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:147)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:127)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:90)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.lambda$execute$0(EngineExecutionOrchestrator.java:55)

[jira] [Commented] (FLINK-31472) AsyncSinkWriterThrottlingTest failed with Illegal mailbox thread

2023-12-18 Thread Sergey Nuyanzin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31472?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798112#comment-17798112
 ] 

Sergey Nuyanzin commented on FLINK-31472:
-

[~chalixar], thanks for volunteering, done 

> AsyncSinkWriterThrottlingTest failed with Illegal mailbox thread
> 
>
> Key: FLINK-31472
> URL: https://issues.apache.org/jira/browse/FLINK-31472
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Common
>Affects Versions: 1.17.0, 1.16.1, 1.19.0
>Reporter: Ran Tao
>Assignee: Ahmed Hamdy
>Priority: Major
>  Labels: test-stability
>
> when run mvn clean test, this case failed occasionally.
> {noformat}
> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.955 
> s <<< FAILURE! - in 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest
> [ERROR] 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest.testSinkThroughputShouldThrottleToHalfBatchSize
>   Time elapsed: 0.492 s  <<< ERROR!
> java.lang.IllegalStateException: Illegal thread detected. This method must be 
> called from inside the mailbox thread!
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.TaskMailboxImpl.checkIsMailboxThread(TaskMailboxImpl.java:262)
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.TaskMailboxImpl.take(TaskMailboxImpl.java:137)
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxExecutorImpl.yield(MailboxExecutorImpl.java:84)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriter.flush(AsyncSinkWriter.java:367)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriter.lambda$registerCallback$3(AsyncSinkWriter.java:315)
>         at 
> org.apache.flink.streaming.runtime.tasks.TestProcessingTimeService$CallbackTask.onProcessingTime(TestProcessingTimeService.java:199)
>         at 
> org.apache.flink.streaming.runtime.tasks.TestProcessingTimeService.setCurrentTime(TestProcessingTimeService.java:76)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest.testSinkThroughputShouldThrottleToHalfBatchSize(AsyncSinkWriterThrottlingTest.java:64)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>         at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:498)
>         at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
>         at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>         at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
>         at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>         at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
>         at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
>         at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
>         at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
>         at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
>         at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
>         at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
>         at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
>         at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
>         at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
>         at org.junit.runner.JUnitCore.run(JUnitCore.java:115)
>         at 
> org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42)
>         at 
> org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80)
>         at 
> org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:72)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:147)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:127)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:90)
>         at 
> org.junit.platform.launcher.core.EngineExecut

Re: [PR] [FLINK-33588][Flink-Runtime] Fix NullArgumentException of getQuantile method in DescriptiveStatisticsHistogramStatistics [flink]

2023-12-18 Thread via GitHub


JingGe merged PR #23931:
URL: https://github.com/apache/flink/pull/23931


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-33588) Fix Flink Checkpointing Statistics Bug

2023-12-18 Thread Jing Ge (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33588?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798114#comment-17798114
 ] 

Jing Ge commented on FLINK-33588:
-

master: 5e1deeea19c2e73f3bc6f1f9881a778d13abf6a4

> Fix Flink Checkpointing Statistics Bug
> --
>
> Key: FLINK-33588
> URL: https://issues.apache.org/jira/browse/FLINK-33588
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.14.5, 1.16.0, 1.17.0, 1.15.2, 1.14.6, 1.18.0, 1.17.1
>Reporter: Tongtong Zhu
>Assignee: Tongtong Zhu
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
> Attachments: FLINK-33588.patch, image-2023-12-11-17-35-23-391.png, 
> image-2023-12-13-11-35-43-780.png, image-2023-12-15-13-59-28-201.png, 
> newCommit-FLINK-33688.patch
>
>
> When the Flink task is first started, the checkpoint data is null due to the 
> lack of data, and Percentile throws a null pointer exception when calculating 
> the percentage. After multiple tests, I found that it is necessary to set an 
> initial value for the statistical data value of the checkpoint when the 
> checkpoint data is null (i.e. at the beginning of the task) to solve this 
> problem.
> The following is an abnormal description of the bug:
> 2023-09-13 15:02:54,608 ERROR 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler
>  [] - Unhandled exception.
> org.apache.commons.math3.exception.NullArgumentException: input array
>     at 
> org.apache.commons.math3.util.MathArrays.verifyValues(MathArrays.java:1650) 
> ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.commons.math3.stat.descriptive.AbstractUnivariateStatistic.test(AbstractUnivariateStatistic.java:158)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.commons.math3.stat.descriptive.rank.Percentile.evaluate(Percentile.java:272)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.commons.math3.stat.descriptive.rank.Percentile.evaluate(Percentile.java:241)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.metrics.DescriptiveStatisticsHistogramStatistics$CommonMetricsSnapshot.getPercentile(DescriptiveStatisticsHistogramStatistics.java:159)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.metrics.DescriptiveStatisticsHistogramStatistics.getQuantile(DescriptiveStatisticsHistogramStatistics.java:53)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.checkpoint.StatsSummarySnapshot.getQuantile(StatsSummarySnapshot.java:108)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.messages.checkpoints.StatsSummaryDto.valueOf(StatsSummaryDto.java:81)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.createCheckpointingStatistics(CheckpointingStatisticsHandler.java:129)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.handleRequest(CheckpointingStatisticsHandler.java:84)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.handleRequest(CheckpointingStatisticsHandler.java:58)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.AbstractAccessExecutionGraphHandler.handleRequest(AbstractAccessExecutionGraphHandler.java:68)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.AbstractExecutionGraphHandler.lambda$handleRequest$0(AbstractExecutionGraphHandler.java:87)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:602) 
> [?:1.8.0_151]
>     at 
> java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:577)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:442)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) 
> [?:1.8.0_151]
>     at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_151]
>     at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  [?:1.8.0_151]
>     at java.lan

[jira] [Updated] (FLINK-31472) AsyncSinkWriterThrottlingTest failed with Illegal mailbox thread

2023-12-18 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31472?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-31472:
---
Labels: pull-request-available test-stability  (was: test-stability)

> AsyncSinkWriterThrottlingTest failed with Illegal mailbox thread
> 
>
> Key: FLINK-31472
> URL: https://issues.apache.org/jira/browse/FLINK-31472
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Common
>Affects Versions: 1.17.0, 1.16.1, 1.19.0
>Reporter: Ran Tao
>Assignee: Ahmed Hamdy
>Priority: Major
>  Labels: pull-request-available, test-stability
>
> when run mvn clean test, this case failed occasionally.
> {noformat}
> [ERROR] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.955 
> s <<< FAILURE! - in 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest
> [ERROR] 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest.testSinkThroughputShouldThrottleToHalfBatchSize
>   Time elapsed: 0.492 s  <<< ERROR!
> java.lang.IllegalStateException: Illegal thread detected. This method must be 
> called from inside the mailbox thread!
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.TaskMailboxImpl.checkIsMailboxThread(TaskMailboxImpl.java:262)
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.TaskMailboxImpl.take(TaskMailboxImpl.java:137)
>         at 
> org.apache.flink.streaming.runtime.tasks.mailbox.MailboxExecutorImpl.yield(MailboxExecutorImpl.java:84)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriter.flush(AsyncSinkWriter.java:367)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriter.lambda$registerCallback$3(AsyncSinkWriter.java:315)
>         at 
> org.apache.flink.streaming.runtime.tasks.TestProcessingTimeService$CallbackTask.onProcessingTime(TestProcessingTimeService.java:199)
>         at 
> org.apache.flink.streaming.runtime.tasks.TestProcessingTimeService.setCurrentTime(TestProcessingTimeService.java:76)
>         at 
> org.apache.flink.connector.base.sink.writer.AsyncSinkWriterThrottlingTest.testSinkThroughputShouldThrottleToHalfBatchSize(AsyncSinkWriterThrottlingTest.java:64)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>         at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:498)
>         at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
>         at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>         at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
>         at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>         at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
>         at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
>         at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
>         at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331)
>         at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79)
>         at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329)
>         at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66)
>         at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293)
>         at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306)
>         at org.junit.runners.ParentRunner.run(ParentRunner.java:413)
>         at org.junit.runner.JUnitCore.run(JUnitCore.java:137)
>         at org.junit.runner.JUnitCore.run(JUnitCore.java:115)
>         at 
> org.junit.vintage.engine.execution.RunnerExecutor.execute(RunnerExecutor.java:42)
>         at 
> org.junit.vintage.engine.VintageTestEngine.executeAllChildren(VintageTestEngine.java:80)
>         at 
> org.junit.vintage.engine.VintageTestEngine.execute(VintageTestEngine.java:72)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:147)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:127)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrchestrator.execute(EngineExecutionOrchestrator.java:90)
>         at 
> org.junit.platform.launcher.core.EngineExecutionOrch

[PR] [FLINK-31472][Connectors/Base] Update AsyncSink Throttling test to use concurrent mailbox [flink]

2023-12-18 Thread via GitHub


vahmed-hamdy opened a new pull request, #23946:
URL: https://github.com/apache/flink/pull/23946

   
   
   
   ## What is the purpose of the change
   
   This pull request updates `AsyncSinkThrottlingTest` to use concurrent 
`TestSinkInitContextAnyThreadMailbox` due the concurrent nature of the test and 
to address CI stability issues.
   
   
   ## Brief change log
   - Updated  `AsyncSinkThrottlingTest` to use concurrent 
`TestSinkInitContextAnyThreadMailbox`
   
   
   ## Verifying this change
   
   
   
   This change is a trivial rework for test suite, changes are verified by 
rerunning tests in debug mode.
   
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): no
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: no
 - The serializers: no
 - The runtime per-record code paths (performance sensitive):no
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: no
 - The S3 file system connector: no
   
   ## Documentation
   
 - Does this pull request introduce a new feature? no
 - If yes, how is the feature documented? not applicable 
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-33588) Fix Flink Checkpointing Statistics Bug

2023-12-18 Thread Jing Ge (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33588?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798115#comment-17798115
 ] 

Jing Ge commented on FLINK-33588:
-

[~zhutong66] could you please cp the change to 1.18?

> Fix Flink Checkpointing Statistics Bug
> --
>
> Key: FLINK-33588
> URL: https://issues.apache.org/jira/browse/FLINK-33588
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.14.5, 1.16.0, 1.17.0, 1.15.2, 1.14.6, 1.18.0, 1.17.1
>Reporter: Tongtong Zhu
>Assignee: Tongtong Zhu
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
> Attachments: FLINK-33588.patch, image-2023-12-11-17-35-23-391.png, 
> image-2023-12-13-11-35-43-780.png, image-2023-12-15-13-59-28-201.png, 
> newCommit-FLINK-33688.patch
>
>
> When the Flink task is first started, the checkpoint data is null due to the 
> lack of data, and Percentile throws a null pointer exception when calculating 
> the percentage. After multiple tests, I found that it is necessary to set an 
> initial value for the statistical data value of the checkpoint when the 
> checkpoint data is null (i.e. at the beginning of the task) to solve this 
> problem.
> The following is an abnormal description of the bug:
> 2023-09-13 15:02:54,608 ERROR 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler
>  [] - Unhandled exception.
> org.apache.commons.math3.exception.NullArgumentException: input array
>     at 
> org.apache.commons.math3.util.MathArrays.verifyValues(MathArrays.java:1650) 
> ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.commons.math3.stat.descriptive.AbstractUnivariateStatistic.test(AbstractUnivariateStatistic.java:158)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.commons.math3.stat.descriptive.rank.Percentile.evaluate(Percentile.java:272)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.commons.math3.stat.descriptive.rank.Percentile.evaluate(Percentile.java:241)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.metrics.DescriptiveStatisticsHistogramStatistics$CommonMetricsSnapshot.getPercentile(DescriptiveStatisticsHistogramStatistics.java:159)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.metrics.DescriptiveStatisticsHistogramStatistics.getQuantile(DescriptiveStatisticsHistogramStatistics.java:53)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.checkpoint.StatsSummarySnapshot.getQuantile(StatsSummarySnapshot.java:108)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.messages.checkpoints.StatsSummaryDto.valueOf(StatsSummaryDto.java:81)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.createCheckpointingStatistics(CheckpointingStatisticsHandler.java:129)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.handleRequest(CheckpointingStatisticsHandler.java:84)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.handleRequest(CheckpointingStatisticsHandler.java:58)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.AbstractAccessExecutionGraphHandler.handleRequest(AbstractAccessExecutionGraphHandler.java:68)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.AbstractExecutionGraphHandler.lambda$handleRequest$0(AbstractExecutionGraphHandler.java:87)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:602) 
> [?:1.8.0_151]
>     at 
> java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:577)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:442)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) 
> [?:1.8.0_151]
>     at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_151]
>     at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>  [?:1.8.0_151]
>     at java

Re: [PR] [FLINK-31472][Connectors/Base] Update AsyncSink Throttling test to use concurrent mailbox [flink]

2023-12-18 Thread via GitHub


flinkbot commented on PR #23946:
URL: https://github.com/apache/flink/pull/23946#issuecomment-1860072502

   
   ## CI report:
   
   * 4f0d2bc42eef189b04854b445eea2e7c7470933d UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-33704) Update GCS filesystems to latest available versions

2023-12-18 Thread Martijn Visser (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33704?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798124#comment-17798124
 ] 

Martijn Visser commented on FLINK-33704:


[~jingge] I'm trying to confirm if this issue occurs in 1.17 or not. 

> Update GCS filesystems to latest available versions
> ---
>
> Key: FLINK-33704
> URL: https://issues.apache.org/jira/browse/FLINK-33704
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Connectors / FileSystem, FileSystems
>Reporter: Martijn Visser
>Assignee: Martijn Visser
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
>
> Update GS SDK from 2.15.0 to 2.29.1 and GS Hadoop Connector from 2.2.15 to 
> 2.2.18



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33704) Update GCS filesystems to latest available versions

2023-12-18 Thread Martijn Visser (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33704?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798126#comment-17798126
 ] 

Martijn Visser commented on FLINK-33704:


[~jingge] Closing this issue, it doesn't appear in 1.17.2

 !screenshot-1.png! 

> Update GCS filesystems to latest available versions
> ---
>
> Key: FLINK-33704
> URL: https://issues.apache.org/jira/browse/FLINK-33704
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Connectors / FileSystem, FileSystems
>Reporter: Martijn Visser
>Assignee: Martijn Visser
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
> Attachments: screenshot-1.png
>
>
> Update GS SDK from 2.15.0 to 2.29.1 and GS Hadoop Connector from 2.2.15 to 
> 2.2.18



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33704) Update GCS filesystems to latest available versions

2023-12-18 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33704?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser updated FLINK-33704:
---
Attachment: screenshot-1.png

> Update GCS filesystems to latest available versions
> ---
>
> Key: FLINK-33704
> URL: https://issues.apache.org/jira/browse/FLINK-33704
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Connectors / FileSystem, FileSystems
>Reporter: Martijn Visser
>Assignee: Martijn Visser
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
> Attachments: screenshot-1.png
>
>
> Update GS SDK from 2.15.0 to 2.29.1 and GS Hadoop Connector from 2.2.15 to 
> 2.2.18



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-33603) Fix guava shading for GCS connector

2023-12-18 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33603?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser closed FLINK-33603.
--
Resolution: Resolved

FLINK-33704

> Fix guava shading for GCS connector
> ---
>
> Key: FLINK-33603
> URL: https://issues.apache.org/jira/browse/FLINK-33603
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem
>Affects Versions: 1.17.1
>Reporter: Jayadeep Jayaraman
>Priority: Major
>  Labels: pull-request-available
>
> GCS connector has guava shading issue, This change introduced dependency on 
> guava version({{{}31.1-jre{}}}) required by {{{}google-cloud-storage{}}}. 
> Upgrade of {{google-cloud-storage}} lead to runtime failure because of new 
> functionalities added in {{{}31.1-jre{}}}.
> This change pins guava version to the one required by storage client 
> specifically in {{{}flink-gs-fs-hadoop{}}}, leaving all other filesystem 
> implementation untouched.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-33704) Update GCS filesystems to latest available versions

2023-12-18 Thread Martijn Visser (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33704?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798126#comment-17798126
 ] 

Martijn Visser edited comment on FLINK-33704 at 12/18/23 10:48 AM:
---

[~jingge] Closing this issue, it doesn't appear in 1.17.2

!screenshot-1.png|width=623,height=294!


was (Author: martijnvisser):
[~jingge] Closing this issue, it doesn't appear in 1.17.2

 !screenshot-1.png! 

> Update GCS filesystems to latest available versions
> ---
>
> Key: FLINK-33704
> URL: https://issues.apache.org/jira/browse/FLINK-33704
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Connectors / FileSystem, FileSystems
>Reporter: Martijn Visser
>Assignee: Martijn Visser
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
> Attachments: screenshot-1.png
>
>
> Update GS SDK from 2.15.0 to 2.29.1 and GS Hadoop Connector from 2.2.15 to 
> 2.2.18



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-33704) Update GCS filesystems to latest available versions

2023-12-18 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33704?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser closed FLINK-33704.
--
Resolution: Fixed

> Update GCS filesystems to latest available versions
> ---
>
> Key: FLINK-33704
> URL: https://issues.apache.org/jira/browse/FLINK-33704
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Connectors / FileSystem, FileSystems
>Reporter: Martijn Visser
>Assignee: Martijn Visser
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
> Attachments: screenshot-1.png
>
>
> Update GS SDK from 2.15.0 to 2.29.1 and GS Hadoop Connector from 2.2.15 to 
> 2.2.18



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33704) Update GCS filesystems to latest available versions

2023-12-18 Thread Jing Ge (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33704?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798130#comment-17798130
 ] 

Jing Ge commented on FLINK-33704:
-

[~martijnvisser] thanks for letting us know.

> Update GCS filesystems to latest available versions
> ---
>
> Key: FLINK-33704
> URL: https://issues.apache.org/jira/browse/FLINK-33704
> Project: Flink
>  Issue Type: Technical Debt
>  Components: Connectors / FileSystem, FileSystems
>Reporter: Martijn Visser
>Assignee: Martijn Visser
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
> Attachments: screenshot-1.png
>
>
> Update GS SDK from 2.15.0 to 2.29.1 and GS Hadoop Connector from 2.2.15 to 
> 2.2.18



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-33793) java.lang.NoSuchMethodError when checkpointing in Google Cloud Storage

2023-12-18 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33793?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser closed FLINK-33793.
--
Resolution: Fixed

Fixed via FLINK-33704

> java.lang.NoSuchMethodError when checkpointing in Google Cloud Storage
> --
>
> Key: FLINK-33793
> URL: https://issues.apache.org/jira/browse/FLINK-33793
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.18.0
> Environment: Flink 1.18
>Reporter: ChangZhuo Chen (陳昌倬)
>Priority: Critical
>
> We have the following exception when checkpointing in Flink 1.18 + Google 
> Cloud Storage. The same code works well in Flink 1.17:
>  
> {{2023-12-11 07:45:28,861 ERROR 
> org.apache.flink.util.FatalExitExceptionHandler              [] - FATAL: 
> Thread 'jobmanager-io-thread-5' produced an uncaught exception. Stopping the 
> process...}}
> {{java.lang.NoSuchMethodError: 'com.google.common.collect.ImmutableMap 
> com.google.common.collect.ImmutableMap$Builder.buildOrThrow()'}}
> {{        at 
> com.google.cloud.storage.UnifiedOpts$Opts.getRpcOptions(UnifiedOpts.java:2096)
>  ~[?:?]}}
> {{        at 
> com.google.cloud.storage.StorageImpl.writer(StorageImpl.java:624) ~[?:?]}}
> {{        at com.google.cloud.storage.StorageImpl.writer(StorageImpl.java:90) 
> ~[?:?]}}
> {{        at 
> org.apache.flink.fs.gs.storage.GSBlobStorageImpl.writeBlob(GSBlobStorageImpl.java:64)
>  ~[?:?]}}
> {{        at 
> org.apache.flink.fs.gs.writer.GSRecoverableFsDataOutputStream.createWriteChannel(GSRecoverableFsDataOutputStream.java:229)
>  ~[?:?]}}
> {{        at 
> org.apache.flink.fs.gs.writer.GSRecoverableFsDataOutputStream.write(GSRecoverableFsDataOutputStream.java:152)
>  ~[?:?]}}
> {{        at 
> org.apache.flink.fs.gs.writer.GSRecoverableFsDataOutputStream.write(GSRecoverableFsDataOutputStream.java:135)
>  ~[?:?]}}
> {{        at 
> org.apache.flink.fs.gs.writer.GSRecoverableFsDataOutputStream.write(GSRecoverableFsDataOutputStream.java:128)
>  ~[?:?]}}
> {{        at 
> org.apache.flink.runtime.state.filesystem.FsCheckpointMetadataOutputStream.write(FsCheckpointMetadataOutputStream.java:73)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at java.io.DataOutputStream.writeInt(Unknown Source) ~[?:?]}}
> {{        at 
> org.apache.flink.runtime.checkpoint.Checkpoints.storeCheckpointMetadata(Checkpoints.java:98)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.checkpoint.Checkpoints.storeCheckpointMetadata(Checkpoints.java:88)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.checkpoint.Checkpoints.storeCheckpointMetadata(Checkpoints.java:83)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.checkpoint.PendingCheckpoint.finalizeCheckpoint(PendingCheckpoint.java:335)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.finalizeCheckpoint(CheckpointCoordinator.java:1404)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.completePendingCheckpoint(CheckpointCoordinator.java:1303)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.receiveAcknowledgeMessage(CheckpointCoordinator.java:1195)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.scheduler.ExecutionGraphHandler.lambda$acknowledgeCheckpoint$1(ExecutionGraphHandler.java:89)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.scheduler.ExecutionGraphHandler.lambda$processCheckpointCoordinatorMessage$3(ExecutionGraphHandler.java:119)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown 
> Source) ~[?:?]}}
> {{        at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown 
> Source) ~[?:?]}}
> {{        at java.lang.Thread.run(Unknown Source) [?:?]}}
>  
> The issue has been reported in GitHub 
> [https://github.com/apache/flink/pull/22281#issuecomment-1728553794.] 
> However, it is still not fixed yet in 1.18.0.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33603) Fix guava shading for GCS connector

2023-12-18 Thread Martijn Visser (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33603?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Martijn Visser updated FLINK-33603:
---
Affects Version/s: 1.18.0
   (was: 1.17.1)

> Fix guava shading for GCS connector
> ---
>
> Key: FLINK-33603
> URL: https://issues.apache.org/jira/browse/FLINK-33603
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / FileSystem
>Affects Versions: 1.18.0
>Reporter: Jayadeep Jayaraman
>Priority: Major
>  Labels: pull-request-available
>
> GCS connector has guava shading issue, This change introduced dependency on 
> guava version({{{}31.1-jre{}}}) required by {{{}google-cloud-storage{}}}. 
> Upgrade of {{google-cloud-storage}} lead to runtime failure because of new 
> functionalities added in {{{}31.1-jre{}}}.
> This change pins guava version to the one required by storage client 
> specifically in {{{}flink-gs-fs-hadoop{}}}, leaving all other filesystem 
> implementation untouched.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-33588) Fix Flink Checkpointing Statistics Bug

2023-12-18 Thread Jing Ge (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33588?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798115#comment-17798115
 ] 

Jing Ge edited comment on FLINK-33588 at 12/18/23 10:54 AM:


[~zhutong66] could you please backport the change to 1.18?


was (Author: jingge):
[~zhutong66] could you please cp the change to 1.18?

> Fix Flink Checkpointing Statistics Bug
> --
>
> Key: FLINK-33588
> URL: https://issues.apache.org/jira/browse/FLINK-33588
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.14.5, 1.16.0, 1.17.0, 1.15.2, 1.14.6, 1.18.0, 1.17.1
>Reporter: Tongtong Zhu
>Assignee: Tongtong Zhu
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
> Attachments: FLINK-33588.patch, image-2023-12-11-17-35-23-391.png, 
> image-2023-12-13-11-35-43-780.png, image-2023-12-15-13-59-28-201.png, 
> newCommit-FLINK-33688.patch
>
>
> When the Flink task is first started, the checkpoint data is null due to the 
> lack of data, and Percentile throws a null pointer exception when calculating 
> the percentage. After multiple tests, I found that it is necessary to set an 
> initial value for the statistical data value of the checkpoint when the 
> checkpoint data is null (i.e. at the beginning of the task) to solve this 
> problem.
> The following is an abnormal description of the bug:
> 2023-09-13 15:02:54,608 ERROR 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler
>  [] - Unhandled exception.
> org.apache.commons.math3.exception.NullArgumentException: input array
>     at 
> org.apache.commons.math3.util.MathArrays.verifyValues(MathArrays.java:1650) 
> ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.commons.math3.stat.descriptive.AbstractUnivariateStatistic.test(AbstractUnivariateStatistic.java:158)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.commons.math3.stat.descriptive.rank.Percentile.evaluate(Percentile.java:272)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.commons.math3.stat.descriptive.rank.Percentile.evaluate(Percentile.java:241)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.metrics.DescriptiveStatisticsHistogramStatistics$CommonMetricsSnapshot.getPercentile(DescriptiveStatisticsHistogramStatistics.java:159)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.metrics.DescriptiveStatisticsHistogramStatistics.getQuantile(DescriptiveStatisticsHistogramStatistics.java:53)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.checkpoint.StatsSummarySnapshot.getQuantile(StatsSummarySnapshot.java:108)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.messages.checkpoints.StatsSummaryDto.valueOf(StatsSummaryDto.java:81)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.createCheckpointingStatistics(CheckpointingStatisticsHandler.java:129)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.handleRequest(CheckpointingStatisticsHandler.java:84)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.handleRequest(CheckpointingStatisticsHandler.java:58)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.AbstractAccessExecutionGraphHandler.handleRequest(AbstractAccessExecutionGraphHandler.java:68)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.AbstractExecutionGraphHandler.lambda$handleRequest$0(AbstractExecutionGraphHandler.java:87)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:602) 
> [?:1.8.0_151]
>     at 
> java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:577)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:442)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) 
> [?:1.8.0_151]
>     at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_151]
>     at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>  

[jira] [Commented] (FLINK-33365) Missing filter condition in execution plan containing lookup join with mysql jdbc connector

2023-12-18 Thread Benchao Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33365?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798131#comment-17798131
 ] 

Benchao Li commented on FLINK-33365:


[~davidradl] I didn't see a problem in the example you gave in this thread: 
https://lists.apache.org/thread/7r49c5ffohz1oqm3oz2g60g5lnovq0l2. 

The result is correct since you are using 'LEFT JOIN', for the records from 
left table which does not have matching rows, it should emit the line, and pad 
nulls for fields in right table. Am I missing something?

The behavior should be the same when pushdown is enabled or not as you stated 
in previously.

> Missing filter condition in execution plan containing lookup join with mysql 
> jdbc connector
> ---
>
> Key: FLINK-33365
> URL: https://issues.apache.org/jira/browse/FLINK-33365
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / JDBC
>Affects Versions: 1.18.0, 1.17.1
> Environment: Flink 1.17.1 & Flink 1.18.0 with 
> flink-connector-jdbc-3.1.1-1.17.jar
>Reporter: macdoor615
>Assignee: david radley
>Priority: Critical
>  Labels: pull-request-available
> Attachments: flink-connector-jdbc-3.0.0-1.16.png, 
> flink-connector-jdbc-3.1.1-1.17.png
>
>
> create table in flink with sql-client.sh
> {code:java}
> CREATE TABLE default_catalog.default_database.a (
>   ip string, 
>   proctime as proctime()
> ) 
> WITH (
>   'connector' = 'datagen'
> );{code}
> create table in mysql
> {code:java}
> create table b (
>   ip varchar(20), 
>   type int
> );  {code}
>  
> Flink 1.17.1/ 1.18.0 and *flink-connector-jdbc-3.1.1-1.17.jar*
> excute in sql-client.sh 
> {code:java}
> explain SELECT * FROM default_catalog.default_database.a left join 
> bnpmp_mysql_test.gem_tmp.b FOR SYSTEM_TIME AS OF a.proctime on b.type = 0 and 
> a.ip = b.ip; {code}
> get the execution plan
> {code:java}
> ...
> == Optimized Execution Plan ==
> Calc(select=[ip, PROCTIME_MATERIALIZE(proctime) AS proctime, ip0, type])
> +- LookupJoin(table=[bnpmp_mysql_test.gem_tmp.b], joinType=[LeftOuterJoin], 
> lookup=[ip=ip], select=[ip, proctime, ip, CAST(0 AS INTEGER) AS type, CAST(ip 
> AS VARCHAR(2147483647)) AS ip0])
>    +- Calc(select=[ip, PROCTIME() AS proctime])
>       +- TableSourceScan(table=[[default_catalog, default_database, a]], 
> fields=[ip]){code}
>  
> excute same sql in sql-client with Flink 1.17.1/ 1.18.0 and 
> *flink-connector-jdbc-3.0.0-1.16.jar*
> get the execution plan
> {code:java}
> ...
> == Optimized Execution Plan ==
> Calc(select=[ip, PROCTIME_MATERIALIZE(proctime) AS proctime, ip0, type])
> +- LookupJoin(table=[bnpmp_mysql_test.gem_tmp.b], joinType=[LeftOuterJoin], 
> lookup=[type=0, ip=ip], where=[(type = 0)], select=[ip, proctime, ip, CAST(0 
> AS INTEGER) AS type, CAST(ip AS VARCHAR(2147483647)) AS ip0])
>    +- Calc(select=[ip, PROCTIME() AS proctime])
>       +- TableSourceScan(table=[[default_catalog, default_database, a]], 
> fields=[ip]) {code}
> with flink-connector-jdbc-3.1.1-1.17.jar,  the condition is 
> *lookup=[ip=ip]*
> with flink-connector-jdbc-3.0.0-1.16.jar ,  the condition is 
> *lookup=[type=0, ip=ip], where=[(type = 0)]*
>  
> In out real world production environment, this lead incorrect data output
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-27876][table-planner] Choose the right side as build side when using default shuffle hash strategy if left size is equal with right [flink]

2023-12-18 Thread via GitHub


xuyangzhong commented on PR #19866:
URL: https://github.com/apache/flink/pull/19866#issuecomment-1860163668

   > @xuyangzhong Do you think we should pick this fix to release-1.18?
   
   I feel that cp is not needed because it changes the plan and will cause 
compatibility issues. WDYT?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-33560) Externalize Kinesis Python connector code

2023-12-18 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33560?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer updated FLINK-33560:
--
Fix Version/s: aws-connector-4.3.0

> Externalize Kinesis Python connector code
> -
>
> Key: FLINK-33560
> URL: https://issues.apache.org/jira/browse/FLINK-33560
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / Python, Connectors / Kinesis
>Affects Versions: 1.18.0
>Reporter: Márton Balassi
>Assignee: Danny Cranmer
>Priority: Major
> Fix For: 1.19.0, aws-connector-4.3.0
>
>
> See description of parent ticket for context.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33560) Externalize Kinesis Python connector code

2023-12-18 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33560?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer updated FLINK-33560:
--
Affects Version/s: aws-connector-4.2.0

> Externalize Kinesis Python connector code
> -
>
> Key: FLINK-33560
> URL: https://issues.apache.org/jira/browse/FLINK-33560
> Project: Flink
>  Issue Type: Sub-task
>  Components: API / Python, Connectors / Kinesis
>Affects Versions: 1.18.0, aws-connector-4.2.0
>Reporter: Márton Balassi
>Assignee: Danny Cranmer
>Priority: Major
> Fix For: 1.19.0, aws-connector-4.3.0
>
>
> See description of parent ticket for context.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33449) Support array_contains_seq function

2023-12-18 Thread Martijn Visser (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33449?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798135#comment-17798135
 ] 

Martijn Visser commented on FLINK-33449:


I think a Dev ML discussion makes sense, I would probably vote for 
ARRAY_CONTAINS_SEQUENCE 

> Support array_contains_seq function
> ---
>
> Key: FLINK-33449
> URL: https://issues.apache.org/jira/browse/FLINK-33449
> Project: Flink
>  Issue Type: New Feature
>Reporter: leo cheng
>Priority: Minor
>  Labels: pull-request-available
>
> support function array_contains_seq like trino contains_sequence
> trino: 
> https://trino.io/docs/current/functions/array.html?highlight=contains_sequence#contains_sequence



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-32877][Filesystem]add HTTP options to gcs-cloud-storage client [flink]

2023-12-18 Thread via GitHub


MartijnVisser commented on PR #23226:
URL: https://github.com/apache/flink/pull/23226#issuecomment-1860175028

   > I don't have a Jira account. @cnauroth can you take the ownership of jira 
ticket?
   
   @singhravidutt Please request one via 
https://selfserve.apache.org/jira-account.html
   
   Also, please rebase your PRs (merge comitts can't be merged)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[PR] Fix NullArgumentException of getQuantile method in DescriptiveStatisticsHistogramStatistics [flink]

2023-12-18 Thread via GitHub


JingGe opened a new pull request, #23947:
URL: https://github.com/apache/flink/pull/23947

   BP of #23931 with no change
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-33804) Add Option to disable showing metrics in JobMananger UI

2023-12-18 Thread Martijn Visser (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33804?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798136#comment-17798136
 ] 

Martijn Visser commented on FLINK-33804:


[~qqibrow]  Please open a discussion with a proposal on the Dev ML for this

> Add Option to disable showing metrics in JobMananger UI
> ---
>
> Key: FLINK-33804
> URL: https://issues.apache.org/jira/browse/FLINK-33804
> Project: Flink
>  Issue Type: Improvement
>  Components: Runtime / Metrics
>Reporter: Lu Niu
>Priority: Major
>
> Flink allows users to view metric in JobMananger UI. However there are 2 
> problems we found:
>  # The JobManager is required to aggregate metrics from all task managers. 
> When the metric cardinality is quite high, this process can trigger a 
> JobManager Full GC and slow response time.
>  # Flink user cases in prod usually have their own dashboard to view metrics. 
> so this feature sometimes is not useful.
> In light of this, we propose to add option to disable this feature.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-33365) Missing filter condition in execution plan containing lookup join with mysql jdbc connector

2023-12-18 Thread Benchao Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33365?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798131#comment-17798131
 ] 

Benchao Li edited comment on FLINK-33365 at 12/18/23 11:09 AM:
---

[~davidradl] I didn't see a problem in the example you gave in this thread: 
https://lists.apache.org/thread/7r49c5ffohz1oqm3oz2g60g5lnovq0l2. 

The result is correct since you are using 'LEFT JOIN', for the records from 
left table which does not have matching rows, it should emit the line, and pad 
nulls for fields in right table. Am I missing something?

The behavior should be the same when pushdown is enabled or not as you stated 
previously.


was (Author: libenchao):
[~davidradl] I didn't see a problem in the example you gave in this thread: 
https://lists.apache.org/thread/7r49c5ffohz1oqm3oz2g60g5lnovq0l2. 

The result is correct since you are using 'LEFT JOIN', for the records from 
left table which does not have matching rows, it should emit the line, and pad 
nulls for fields in right table. Am I missing something?

The behavior should be the same when pushdown is enabled or not as you stated 
in previously.

> Missing filter condition in execution plan containing lookup join with mysql 
> jdbc connector
> ---
>
> Key: FLINK-33365
> URL: https://issues.apache.org/jira/browse/FLINK-33365
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / JDBC
>Affects Versions: 1.18.0, 1.17.1
> Environment: Flink 1.17.1 & Flink 1.18.0 with 
> flink-connector-jdbc-3.1.1-1.17.jar
>Reporter: macdoor615
>Assignee: david radley
>Priority: Critical
>  Labels: pull-request-available
> Attachments: flink-connector-jdbc-3.0.0-1.16.png, 
> flink-connector-jdbc-3.1.1-1.17.png
>
>
> create table in flink with sql-client.sh
> {code:java}
> CREATE TABLE default_catalog.default_database.a (
>   ip string, 
>   proctime as proctime()
> ) 
> WITH (
>   'connector' = 'datagen'
> );{code}
> create table in mysql
> {code:java}
> create table b (
>   ip varchar(20), 
>   type int
> );  {code}
>  
> Flink 1.17.1/ 1.18.0 and *flink-connector-jdbc-3.1.1-1.17.jar*
> excute in sql-client.sh 
> {code:java}
> explain SELECT * FROM default_catalog.default_database.a left join 
> bnpmp_mysql_test.gem_tmp.b FOR SYSTEM_TIME AS OF a.proctime on b.type = 0 and 
> a.ip = b.ip; {code}
> get the execution plan
> {code:java}
> ...
> == Optimized Execution Plan ==
> Calc(select=[ip, PROCTIME_MATERIALIZE(proctime) AS proctime, ip0, type])
> +- LookupJoin(table=[bnpmp_mysql_test.gem_tmp.b], joinType=[LeftOuterJoin], 
> lookup=[ip=ip], select=[ip, proctime, ip, CAST(0 AS INTEGER) AS type, CAST(ip 
> AS VARCHAR(2147483647)) AS ip0])
>    +- Calc(select=[ip, PROCTIME() AS proctime])
>       +- TableSourceScan(table=[[default_catalog, default_database, a]], 
> fields=[ip]){code}
>  
> excute same sql in sql-client with Flink 1.17.1/ 1.18.0 and 
> *flink-connector-jdbc-3.0.0-1.16.jar*
> get the execution plan
> {code:java}
> ...
> == Optimized Execution Plan ==
> Calc(select=[ip, PROCTIME_MATERIALIZE(proctime) AS proctime, ip0, type])
> +- LookupJoin(table=[bnpmp_mysql_test.gem_tmp.b], joinType=[LeftOuterJoin], 
> lookup=[type=0, ip=ip], where=[(type = 0)], select=[ip, proctime, ip, CAST(0 
> AS INTEGER) AS type, CAST(ip AS VARCHAR(2147483647)) AS ip0])
>    +- Calc(select=[ip, PROCTIME() AS proctime])
>       +- TableSourceScan(table=[[default_catalog, default_database, a]], 
> fields=[ip]) {code}
> with flink-connector-jdbc-3.1.1-1.17.jar,  the condition is 
> *lookup=[ip=ip]*
> with flink-connector-jdbc-3.0.0-1.16.jar ,  the condition is 
> *lookup=[type=0, ip=ip], where=[(type = 0)]*
>  
> In out real world production environment, this lead incorrect data output
>  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32007) Implement Python Wrappers for DynamoDB Connector

2023-12-18 Thread Danny Cranmer (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32007?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Danny Cranmer updated FLINK-32007:
--
Fix Version/s: aws-connector-4.3.0
   (was: 1.19.0)

> Implement Python Wrappers for DynamoDB Connector
> 
>
> Key: FLINK-32007
> URL: https://issues.apache.org/jira/browse/FLINK-32007
> Project: Flink
>  Issue Type: New Feature
>  Components: API / Python, Connectors / DynamoDB
>Reporter: Ahmed Hamdy
>Assignee: Khanh Vu
>Priority: Minor
> Fix For: aws-connector-4.3.0
>
>
> Implement Python API Wrappers for DynamoDB Sink



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] Fix NullArgumentException of getQuantile method in DescriptiveStatisticsHistogramStatistics [flink]

2023-12-18 Thread via GitHub


JingGe commented on PR #23947:
URL: https://github.com/apache/flink/pull/23947#issuecomment-1860197970

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] Fix NullArgumentException of getQuantile method in DescriptiveStatisticsHistogramStatistics [flink]

2023-12-18 Thread via GitHub


flinkbot commented on PR #23947:
URL: https://github.com/apache/flink/pull/23947#issuecomment-1860198985

   
   ## CI report:
   
   * 28314d4fed2721dbabac426b41a416c79df772f0 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[PR] [FLINK-33823] Make PlannerQueryOperation SQL serializable [flink]

2023-12-18 Thread via GitHub


dawidwys opened a new pull request, #23948:
URL: https://github.com/apache/flink/pull/23948

   
   ## What is the purpose of the change
   
   Make PlannerQueryOperation SQL serializable
   
   
   ## Verifying this change
   
   Added tests in:
   * org.apache.flink.table.api.QueryOperationSqlSerializationTest
   * org.apache.flink.table.api.QueryOperationSqlExecutionTest
   
   ## Does this pull request potentially affect one of the following parts:
   
 - Dependencies (does it add or upgrade a dependency): (yes / **no**)
 - The public API, i.e., is any changed class annotated with 
`@Public(Evolving)`: (yes / **no**)
 - The serializers: (yes / **no** / don't know)
 - The runtime per-record code paths (performance sensitive): (yes / **no** 
/ don't know)
 - Anything that affects deployment or recovery: JobManager (and its 
components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / **no** / don't 
know)
 - The S3 file system connector: (yes / **no** / don't know)
   
   ## Documentation
   
 - Does this pull request introduce a new feature? (yes / **no**)
 - If yes, how is the feature documented? (**not applicable** / docs / 
JavaDocs / not documented)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-33823) Serialize PlannerQueryOperation into SQL

2023-12-18 Thread ASF GitHub Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33823?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

ASF GitHub Bot updated FLINK-33823:
---
Labels: pull-request-available  (was: )

> Serialize PlannerQueryOperation into SQL
> 
>
> Key: FLINK-33823
> URL: https://issues.apache.org/jira/browse/FLINK-33823
> Project: Flink
>  Issue Type: Sub-task
>Reporter: Dawid Wysakowicz
>Assignee: Dawid Wysakowicz
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-33588) Fix Flink Checkpointing Statistics Bug

2023-12-18 Thread Jing Ge (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33588?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798115#comment-17798115
 ] 

Jing Ge edited comment on FLINK-33588 at 12/18/23 11:22 AM:


[~zhutong66] bp to 1.18: https://github.com/apache/flink/pull/23947


was (Author: jingge):
[~zhutong66] could you please backport the change to 1.18?

> Fix Flink Checkpointing Statistics Bug
> --
>
> Key: FLINK-33588
> URL: https://issues.apache.org/jira/browse/FLINK-33588
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.14.5, 1.16.0, 1.17.0, 1.15.2, 1.14.6, 1.18.0, 1.17.1
>Reporter: Tongtong Zhu
>Assignee: Tongtong Zhu
>Priority: Critical
>  Labels: pull-request-available
> Fix For: 1.19.0, 1.18.1
>
> Attachments: FLINK-33588.patch, image-2023-12-11-17-35-23-391.png, 
> image-2023-12-13-11-35-43-780.png, image-2023-12-15-13-59-28-201.png, 
> newCommit-FLINK-33688.patch
>
>
> When the Flink task is first started, the checkpoint data is null due to the 
> lack of data, and Percentile throws a null pointer exception when calculating 
> the percentage. After multiple tests, I found that it is necessary to set an 
> initial value for the statistical data value of the checkpoint when the 
> checkpoint data is null (i.e. at the beginning of the task) to solve this 
> problem.
> The following is an abnormal description of the bug:
> 2023-09-13 15:02:54,608 ERROR 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler
>  [] - Unhandled exception.
> org.apache.commons.math3.exception.NullArgumentException: input array
>     at 
> org.apache.commons.math3.util.MathArrays.verifyValues(MathArrays.java:1650) 
> ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.commons.math3.stat.descriptive.AbstractUnivariateStatistic.test(AbstractUnivariateStatistic.java:158)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.commons.math3.stat.descriptive.rank.Percentile.evaluate(Percentile.java:272)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.commons.math3.stat.descriptive.rank.Percentile.evaluate(Percentile.java:241)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.metrics.DescriptiveStatisticsHistogramStatistics$CommonMetricsSnapshot.getPercentile(DescriptiveStatisticsHistogramStatistics.java:159)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.metrics.DescriptiveStatisticsHistogramStatistics.getQuantile(DescriptiveStatisticsHistogramStatistics.java:53)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.checkpoint.StatsSummarySnapshot.getQuantile(StatsSummarySnapshot.java:108)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.messages.checkpoints.StatsSummaryDto.valueOf(StatsSummaryDto.java:81)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.createCheckpointingStatistics(CheckpointingStatisticsHandler.java:129)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.handleRequest(CheckpointingStatisticsHandler.java:84)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.checkpoints.CheckpointingStatisticsHandler.handleRequest(CheckpointingStatisticsHandler.java:58)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.AbstractAccessExecutionGraphHandler.handleRequest(AbstractAccessExecutionGraphHandler.java:68)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> org.apache.flink.runtime.rest.handler.job.AbstractExecutionGraphHandler.lambda$handleRequest$0(AbstractExecutionGraphHandler.java:87)
>  ~[flink-dist_2.12-1.14.5.jar:1.14.5]
>     at 
> java.util.concurrent.CompletableFuture.uniApply(CompletableFuture.java:602) 
> [?:1.8.0_151]
>     at 
> java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:577)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.CompletableFuture$Completion.run(CompletableFuture.java:442)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) 
> [?:1.8.0_151]
>     at java.util.concurrent.FutureTask.run(FutureTask.java:266) [?:1.8.0_151]
>     at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>  [?:1.8.0_151]
>     at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor

[jira] [Commented] (FLINK-32346) JdbcNumericBetweenParametersProvider Sharding key boundaries large storage long integer overflow, use BigDecimal instead Long

2023-12-18 Thread Benchao Li (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-32346?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798139#comment-17798139
 ] 

Benchao Li commented on FLINK-32346:


[~sunnny] I agree the partition columns can/should consider more data types. 

Currently the documentation and the code already diverge. The document says 
"The scan.partition.column must be a numeric, date, or timestamp column from 
the table in question", however the code only supports numeric (not bigger than 
long). 

> JdbcNumericBetweenParametersProvider  Sharding key boundaries large storage 
> long integer overflow, use BigDecimal instead Long
> --
>
> Key: FLINK-32346
> URL: https://issues.apache.org/jira/browse/FLINK-32346
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / JDBC
>Reporter: zhilinli
>Priority: Major
> Attachments: image-2023-06-15-16-42-16-773.png, 
> image-2023-06-15-16-46-13-188.png
>
>
> *JdbcNumericBetweenParametersProvider.class*
> Sharding key boundaries large storage long integer overflow, use BigDecimal 
> instead Long, so that length types such as DecimalType(30,0) are compatible 
> and LONG cannot be stored Can be assigned to me and I want to complete it  
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33823] Make PlannerQueryOperation SQL serializable [flink]

2023-12-18 Thread via GitHub


flinkbot commented on PR #23948:
URL: https://github.com/apache/flink/pull/23948#issuecomment-1860215435

   
   ## CI report:
   
   * 675da5c03660ff590f1ceda33c14452181d0c2c6 UNKNOWN
   
   
   Bot commands
 The @flinkbot bot supports the following commands:
   
- `@flinkbot run azure` re-run the last Azure build
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Updated] (FLINK-33865) exponential-delay.attempts-before-reset-backoff doesn't work when it's set in Job Configuration

2023-12-18 Thread Rui Fan (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33865?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rui Fan updated FLINK-33865:

Description: 
exponential-delay.attempts-before-reset-backoff doesn't work when it's set in 
Job Configuration.
h2. Reason:

When exponential-delay.attempts-before-reset-backoff is set by job 
Configuration instead of cluster configuration. ExecutionConfig#configure will 
call RestartStrategies#parseConfiguration to create the 
ExponentialDelayRestartStrategyConfiguration. And then 
RestartBackoffTimeStrategyFactoryLoader#getJobRestartStrategyFactory will 
create the ExponentialDelayRestartBackoffTimeStrategyFactory by the 
ExponentialDelayRestartStrategyConfiguration.

Since 1.19, RestartStrategies and RestartStrategyConfiguration are depreated, 
so ExponentialDelayRestartStrategyConfiguration doesn't support 
exponential-delay.attempts-before-reset-backoff. So if we set 
exponential-delay.attempts-before-reset-backoff at job level, it won't be 
supported.
h2. Solution

If we use the ExponentialDelayRestartStrategyConfiguration to save 
restartStrategy related options in the ExecutionConfig, all new options are set 
at job level will be missed. 
So we can use the Configuration to save the restartStrategy options inside of 
ExecutionConfig.

!image-2023-12-17-17-56-59-138.png|width=1212,height=256!

  was:
exponential-delay.attempts-before-reset-backoff doesn't work when it's set in 
Job Configuration.

h2. Reason: 

when exponential-delay.attempts-before-reset-backoff is set by job 
Configuration instead of cluster configuration.

ExecutionConfig#configure will call RestartStrategies#parseConfiguration to 
create the ExponentialDelayRestartStrategyConfiguration. And then 
RestartBackoffTimeStrategyFactoryLoader#getJobRestartStrategyFactory will 
create the ExponentialDelayRestartBackoffTimeStrategyFactory by the 
ExponentialDelayRestartStrategyConfiguration.

Since 1.19, RestartStrategies and RestartStrategyConfiguration are depreated, 
so ExponentialDelayRestartStrategyConfiguration doesn't support 
exponential-delay.attempts-before-reset-backoff. So if we set 
exponential-delay.attempts-before-reset-backoff  at job level, it won't be 
supported.

h2. Solution

If we use the ExponentialDelayRestartStrategyConfiguration to save 
restartStrategy related options in the ExecutionConfig, all new options are set 
at job level will be missed. 
So we can use the Configuration to save the restartStrategy options inside of 
ExecutionConfig.


 !image-2023-12-17-17-56-59-138.png! 


> exponential-delay.attempts-before-reset-backoff doesn't work when it's set in 
> Job Configuration
> ---
>
> Key: FLINK-33865
> URL: https://issues.apache.org/jira/browse/FLINK-33865
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Reporter: Rui Fan
>Assignee: Rui Fan
>Priority: Major
>  Labels: pull-request-available
> Attachments: image-2023-12-17-17-56-59-138.png
>
>
> exponential-delay.attempts-before-reset-backoff doesn't work when it's set in 
> Job Configuration.
> h2. Reason:
> When exponential-delay.attempts-before-reset-backoff is set by job 
> Configuration instead of cluster configuration. ExecutionConfig#configure 
> will call RestartStrategies#parseConfiguration to create the 
> ExponentialDelayRestartStrategyConfiguration. And then 
> RestartBackoffTimeStrategyFactoryLoader#getJobRestartStrategyFactory will 
> create the ExponentialDelayRestartBackoffTimeStrategyFactory by the 
> ExponentialDelayRestartStrategyConfiguration.
> Since 1.19, RestartStrategies and RestartStrategyConfiguration are depreated, 
> so ExponentialDelayRestartStrategyConfiguration doesn't support 
> exponential-delay.attempts-before-reset-backoff. So if we set 
> exponential-delay.attempts-before-reset-backoff at job level, it won't be 
> supported.
> h2. Solution
> If we use the ExponentialDelayRestartStrategyConfiguration to save 
> restartStrategy related options in the ExecutionConfig, all new options are 
> set at job level will be missed. 
> So we can use the Configuration to save the restartStrategy options inside of 
> ExecutionConfig.
> !image-2023-12-17-17-56-59-138.png|width=1212,height=256!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33862) Flink Unit Test Failures on 1.18.0

2023-12-18 Thread Prabhu Joseph (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33862?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798143#comment-17798143
 ] 

Prabhu Joseph commented on FLINK-33862:
---

[FLINK-28203|https://issues.apache.org/jira/browse/FLINK-28203] has marked 
bundled dependecies as optional. Setting the flag flink.markBundledAsOptional 
to false has fixed the unit test failures. 

> Flink Unit Test Failures on 1.18.0
> --
>
> Key: FLINK-33862
> URL: https://issues.apache.org/jira/browse/FLINK-33862
> Project: Flink
>  Issue Type: Bug
>Affects Versions: 1.18.0, 1.19.0
>Reporter: Prabhu Joseph
>Priority: Major
>
> Flink Unit Test Failures on 1.18.0. There are 100+ unit test cases failing 
> due to below common issues.
> *Issue 1*
> {code:java}
> ./mvnw -DfailIfNoTests=false -Dmaven.test.failure.ignore=true 
> -Dtest=ExecutionPlanAfterExecutionTest test
> [INFO] Running org.apache.flink.client.program.ExecutionPlanAfterExecutionTest
> org.apache.flink.runtime.client.JobExecutionException: Job execution failed.
>   at 
> org.apache.flink.runtime.jobmaster.JobResult.toJobExecutionResult(JobResult.java:144)
>   at 
> org.apache.flink.runtime.minicluster.MiniClusterJobClient.lambda$getJobExecutionResult$3(MiniClusterJobClient.java:141)
>   at 
> java.base/java.util.concurrent.CompletableFuture$UniApply.tryFire(CompletableFuture.java:642)
>   at 
> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
>   at 
> java.base/java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:2073)
>   at 
> org.apache.flink.runtime.rpc.pekko.PekkoInvocationHandler.lambda$invokeRpc$1(PekkoInvocationHandler.java:268)
>   at 
> java.base/java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:859)
>   at 
> java.base/java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837)
>   at 
> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
>   at 
> java.base/java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:2073)
>   at 
> org.apache.flink.util.concurrent.FutureUtils.doForward(FutureUtils.java:1267)
>   at 
> org.apache.flink.runtime.concurrent.ClassLoadingUtils.lambda$guardCompletionWithContextClassLoader$1(ClassLoadingUtils.java:93)
>   at 
> org.apache.flink.runtime.concurrent.ClassLoadingUtils.runWithContextClassLoader(ClassLoadingUtils.java:68)
>   at 
> org.apache.flink.runtime.concurrent.ClassLoadingUtils.lambda$guardCompletionWithContextClassLoader$2(ClassLoadingUtils.java:92)
>   at 
> java.base/java.util.concurrent.CompletableFuture.uniWhenComplete(CompletableFuture.java:859)
>   at 
> java.base/java.util.concurrent.CompletableFuture$UniWhenComplete.tryFire(CompletableFuture.java:837)
>   at 
> java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506)
>   at 
> java.base/java.util.concurrent.CompletableFuture.complete(CompletableFuture.java:2073)
>   at 
> org.apache.flink.runtime.concurrent.pekko.ScalaFutureUtils$1.onComplete(ScalaFutureUtils.java:47)
>   at org.apache.pekko.dispatch.OnComplete.internal(Future.scala:310)
>   at org.apache.pekko.dispatch.OnComplete.internal(Future.scala:307)
>   at org.apache.pekko.dispatch.japi$CallbackBridge.apply(Future.scala:234)
>   at org.apache.pekko.dispatch.japi$CallbackBridge.apply(Future.scala:231)
>   at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:64)
>   at 
> org.apache.flink.runtime.concurrent.pekko.ScalaFutureUtils$DirectExecutionContext.execute(ScalaFutureUtils.java:65)
>   at 
> scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:72)
>   at 
> scala.concurrent.impl.Promise$DefaultPromise.$anonfun$tryComplete$1(Promise.scala:288)
>   at 
> scala.concurrent.impl.Promise$DefaultPromise.$anonfun$tryComplete$1$adapted(Promise.scala:288)
>   at 
> scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:288)
>   at org.apache.pekko.pattern.PromiseActorRef.$bang(AskSupport.scala:629)
>   at 
> org.apache.pekko.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:34)
>   at 
> org.apache.pekko.pattern.PipeToSupport$PipeableFuture$$anonfun$pipeTo$1.applyOrElse(PipeToSupport.scala:33)
>   at scala.concurrent.Future.$anonfun$andThen$1(Future.scala:536)
>   at scala.concurrent.impl.Promise.liftedTree1$1(Promise.scala:33)
>   at scala.concurrent.impl.Promise.$anonfun$transform$1(Promise.scala:33)
>   at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:64)
>   at 
> org.apache.pekko.dispatch.BatchingExecutor$AbstractBatch.processBatch(BatchingExecutor.scala:73)
>   at 
> org.apac

[jira] [Commented] (FLINK-33793) java.lang.NoSuchMethodError when checkpointing in Google Cloud Storage

2023-12-18 Thread Jun Qin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33793?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798144#comment-17798144
 ] 

Jun Qin commented on FLINK-33793:
-

Could we update the Fix versions field such that everyone knows? 

> java.lang.NoSuchMethodError when checkpointing in Google Cloud Storage
> --
>
> Key: FLINK-33793
> URL: https://issues.apache.org/jira/browse/FLINK-33793
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Checkpointing
>Affects Versions: 1.18.0
> Environment: Flink 1.18
>Reporter: ChangZhuo Chen (陳昌倬)
>Priority: Critical
>
> We have the following exception when checkpointing in Flink 1.18 + Google 
> Cloud Storage. The same code works well in Flink 1.17:
>  
> {{2023-12-11 07:45:28,861 ERROR 
> org.apache.flink.util.FatalExitExceptionHandler              [] - FATAL: 
> Thread 'jobmanager-io-thread-5' produced an uncaught exception. Stopping the 
> process...}}
> {{java.lang.NoSuchMethodError: 'com.google.common.collect.ImmutableMap 
> com.google.common.collect.ImmutableMap$Builder.buildOrThrow()'}}
> {{        at 
> com.google.cloud.storage.UnifiedOpts$Opts.getRpcOptions(UnifiedOpts.java:2096)
>  ~[?:?]}}
> {{        at 
> com.google.cloud.storage.StorageImpl.writer(StorageImpl.java:624) ~[?:?]}}
> {{        at com.google.cloud.storage.StorageImpl.writer(StorageImpl.java:90) 
> ~[?:?]}}
> {{        at 
> org.apache.flink.fs.gs.storage.GSBlobStorageImpl.writeBlob(GSBlobStorageImpl.java:64)
>  ~[?:?]}}
> {{        at 
> org.apache.flink.fs.gs.writer.GSRecoverableFsDataOutputStream.createWriteChannel(GSRecoverableFsDataOutputStream.java:229)
>  ~[?:?]}}
> {{        at 
> org.apache.flink.fs.gs.writer.GSRecoverableFsDataOutputStream.write(GSRecoverableFsDataOutputStream.java:152)
>  ~[?:?]}}
> {{        at 
> org.apache.flink.fs.gs.writer.GSRecoverableFsDataOutputStream.write(GSRecoverableFsDataOutputStream.java:135)
>  ~[?:?]}}
> {{        at 
> org.apache.flink.fs.gs.writer.GSRecoverableFsDataOutputStream.write(GSRecoverableFsDataOutputStream.java:128)
>  ~[?:?]}}
> {{        at 
> org.apache.flink.runtime.state.filesystem.FsCheckpointMetadataOutputStream.write(FsCheckpointMetadataOutputStream.java:73)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at java.io.DataOutputStream.writeInt(Unknown Source) ~[?:?]}}
> {{        at 
> org.apache.flink.runtime.checkpoint.Checkpoints.storeCheckpointMetadata(Checkpoints.java:98)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.checkpoint.Checkpoints.storeCheckpointMetadata(Checkpoints.java:88)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.checkpoint.Checkpoints.storeCheckpointMetadata(Checkpoints.java:83)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.checkpoint.PendingCheckpoint.finalizeCheckpoint(PendingCheckpoint.java:335)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.finalizeCheckpoint(CheckpointCoordinator.java:1404)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.completePendingCheckpoint(CheckpointCoordinator.java:1303)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.checkpoint.CheckpointCoordinator.receiveAcknowledgeMessage(CheckpointCoordinator.java:1195)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.scheduler.ExecutionGraphHandler.lambda$acknowledgeCheckpoint$1(ExecutionGraphHandler.java:89)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at 
> org.apache.flink.runtime.scheduler.ExecutionGraphHandler.lambda$processCheckpointCoordinatorMessage$3(ExecutionGraphHandler.java:119)
>  ~[flink-dist-1.18.0.jar:1.18.0]}}
> {{        at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown 
> Source) ~[?:?]}}
> {{        at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown 
> Source) ~[?:?]}}
> {{        at java.lang.Thread.run(Unknown Source) [?:?]}}
>  
> The issue has been reported in GitHub 
> [https://github.com/apache/flink/pull/22281#issuecomment-1728553794.] 
> However, it is still not fixed yet in 1.18.0.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33872) Checkpoint history does not display for completed jobs

2023-12-18 Thread Hong Liang Teoh (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33872?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hong Liang Teoh updated FLINK-33872:

Description: 
Prior to https://issues.apache.org/jira/browse/FLINK-32469, we see checkpoint 
history for completed jobs (CANCELED, FAILED, FINISHED).

After https://issues.apache.org/jira/browse/FLINK-32469, the checkpoint history 
does not show up for completed jobs. 

*Reproduction steps:*
 # Start a Flink cluster.
 # Submit a job with checkpointing enabled.
 # Wait until at least 1 checkpoint completes.
 # Cancel job.
 # Open the Flink dashboard > Job > Checkpoints > History.

We will see log line in JobManager saying "FlinkJobNotFoundException: Could not 
find Flink job (  )"

*Snapshot of failure:*

When job is running, we can see checkpoints.

!image-2023-12-18-11-37-11-914.png|width=1023,height=350!

When job has been CANCELLED, we no longer see checkpoints data.

!image-2023-12-18-11-37-29-596.png|width=1023,height=307!

  was:
Prior to https://issues.apache.org/jira/browse/FLINK-32469, we see checkpoint 
history for completed jobs (CANCELED, FAILED, FINISHED).

After https://issues.apache.org/jira/browse/FLINK-32469, the checkpoint history 
does not show up for completed jobs. 

*Reproduction steps:*
 # Start a Flink cluster.
 # Submit a job with checkpointing enabled.
 # Wait until at least 1 checkpoint completes.
 # Cancel job.
 # Open the Flink dashboard > Job > Checkpoints > History.

We will see log line in JobManager saying "FlinkJobNotFoundException: Could not 
find Flink job (  )"

*Snapshot of failure:*

When job is running, we can see checkpoints.

!image-2023-12-18-11-37-11-914.png!

When job has been CANCELLED, we no longer see checkpoints data.

!image-2023-12-18-11-37-29-596.png!


> Checkpoint history does not display for completed jobs
> --
>
> Key: FLINK-33872
> URL: https://issues.apache.org/jira/browse/FLINK-33872
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / REST
>Affects Versions: 1.18.0
>Reporter: Hong Liang Teoh
>Priority: Major
> Fix For: 1.19.0, 1.18.2
>
> Attachments: image-2023-12-18-11-37-11-914.png, 
> image-2023-12-18-11-37-29-596.png
>
>
> Prior to https://issues.apache.org/jira/browse/FLINK-32469, we see checkpoint 
> history for completed jobs (CANCELED, FAILED, FINISHED).
> After https://issues.apache.org/jira/browse/FLINK-32469, the checkpoint 
> history does not show up for completed jobs. 
> *Reproduction steps:*
>  # Start a Flink cluster.
>  # Submit a job with checkpointing enabled.
>  # Wait until at least 1 checkpoint completes.
>  # Cancel job.
>  # Open the Flink dashboard > Job > Checkpoints > History.
> We will see log line in JobManager saying "FlinkJobNotFoundException: Could 
> not find Flink job (  )"
> *Snapshot of failure:*
> When job is running, we can see checkpoints.
> !image-2023-12-18-11-37-11-914.png|width=1023,height=350!
> When job has been CANCELLED, we no longer see checkpoints data.
> !image-2023-12-18-11-37-29-596.png|width=1023,height=307!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33872) Checkpoint history does not display for completed jobs

2023-12-18 Thread Hong Liang Teoh (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33872?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hong Liang Teoh updated FLINK-33872:

Description: 
Prior to https://issues.apache.org/jira/browse/FLINK-32469, we see checkpoint 
history for completed jobs (CANCELED, FAILED, FINISHED).

After https://issues.apache.org/jira/browse/FLINK-32469, the checkpoint history 
does not show up for completed jobs. 

*Reproduction steps:*
 # Start a Flink cluster.
 # Submit a job with checkpointing enabled.
 # Wait until at least 1 checkpoint completes.
 # Cancel job.
 # Open the Flink dashboard > Job > Checkpoints > History.

We will see log line in JobManager saying "FlinkJobNotFoundException: Could not 
find Flink job (  )"

*Snapshot of failure:*

When job is running, we can see checkpoints.

!image-2023-12-18-11-37-11-914.png|width=862,height=295!

When job has been CANCELLED, we no longer see checkpoints data.

!image-2023-12-18-11-37-29-596.png|width=860,height=258!

  was:
Prior to https://issues.apache.org/jira/browse/FLINK-32469, we see checkpoint 
history for completed jobs (CANCELED, FAILED, FINISHED).

After https://issues.apache.org/jira/browse/FLINK-32469, the checkpoint history 
does not show up for completed jobs. 

*Reproduction steps:*
 # Start a Flink cluster.
 # Submit a job with checkpointing enabled.
 # Wait until at least 1 checkpoint completes.
 # Cancel job.
 # Open the Flink dashboard > Job > Checkpoints > History.

We will see log line in JobManager saying "FlinkJobNotFoundException: Could not 
find Flink job (  )"

*Snapshot of failure:*

When job is running, we can see checkpoints.

!image-2023-12-18-11-37-11-914.png|width=1023,height=350!

When job has been CANCELLED, we no longer see checkpoints data.

!image-2023-12-18-11-37-29-596.png|width=1023,height=307!


> Checkpoint history does not display for completed jobs
> --
>
> Key: FLINK-33872
> URL: https://issues.apache.org/jira/browse/FLINK-33872
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / REST
>Affects Versions: 1.18.0
>Reporter: Hong Liang Teoh
>Priority: Major
> Fix For: 1.19.0, 1.18.2
>
> Attachments: image-2023-12-18-11-37-11-914.png, 
> image-2023-12-18-11-37-29-596.png
>
>
> Prior to https://issues.apache.org/jira/browse/FLINK-32469, we see checkpoint 
> history for completed jobs (CANCELED, FAILED, FINISHED).
> After https://issues.apache.org/jira/browse/FLINK-32469, the checkpoint 
> history does not show up for completed jobs. 
> *Reproduction steps:*
>  # Start a Flink cluster.
>  # Submit a job with checkpointing enabled.
>  # Wait until at least 1 checkpoint completes.
>  # Cancel job.
>  # Open the Flink dashboard > Job > Checkpoints > History.
> We will see log line in JobManager saying "FlinkJobNotFoundException: Could 
> not find Flink job (  )"
> *Snapshot of failure:*
> When job is running, we can see checkpoints.
> !image-2023-12-18-11-37-11-914.png|width=862,height=295!
> When job has been CANCELLED, we no longer see checkpoints data.
> !image-2023-12-18-11-37-29-596.png|width=860,height=258!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-33872) Checkpoint history does not display for completed jobs

2023-12-18 Thread Hong Liang Teoh (Jira)
Hong Liang Teoh created FLINK-33872:
---

 Summary: Checkpoint history does not display for completed jobs
 Key: FLINK-33872
 URL: https://issues.apache.org/jira/browse/FLINK-33872
 Project: Flink
  Issue Type: Bug
  Components: Runtime / REST
Affects Versions: 1.18.0
Reporter: Hong Liang Teoh
 Fix For: 1.19.0, 1.18.2
 Attachments: image-2023-12-18-11-37-11-914.png, 
image-2023-12-18-11-37-29-596.png

Prior to https://issues.apache.org/jira/browse/FLINK-32469, we see checkpoint 
history for completed jobs (CANCELED, FAILED, FINISHED).

After https://issues.apache.org/jira/browse/FLINK-32469, the checkpoint history 
does not show up for completed jobs. 

*Reproduction steps:*
 # Start a Flink cluster.
 # Submit a job with checkpointing enabled.
 # Wait until at least 1 checkpoint completes.
 # Cancel job.
 # Open the Flink dashboard > Job > Checkpoints > History.

We will see log line in JobManager saying "FlinkJobNotFoundException: Could not 
find Flink job (  )"

*Snapshot of failure:*

When job is running, we can see checkpoints.

!image-2023-12-18-11-37-11-914.png!

When job has been CANCELLED, we no longer see checkpoints data.

!image-2023-12-18-11-37-29-596.png!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33778][table] Cleanup usage of deprecated TableConfig#setIdleS… [flink]

2023-12-18 Thread via GitHub


liuyongvs commented on PR #23894:
URL: https://github.com/apache/flink/pull/23894#issuecomment-1860241060

   > LTGM. BTW, should we also update this function in pyflink? cc @HuangXingBo 
Further more, please modify the description of this pr.
   
   the python code of  has no use, we do not need modify
   ```
   def set_idle_state_retention_time(self,
 min_time: datetime.timedelta,
 max_time: datetime.timedelta):
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Assigned] (FLINK-33872) Checkpoint history does not display for completed jobs

2023-12-18 Thread Hong Liang Teoh (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33872?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hong Liang Teoh reassigned FLINK-33872:
---

Assignee: Hong Liang Teoh

> Checkpoint history does not display for completed jobs
> --
>
> Key: FLINK-33872
> URL: https://issues.apache.org/jira/browse/FLINK-33872
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / REST
>Affects Versions: 1.18.0
>Reporter: Hong Liang Teoh
>Assignee: Hong Liang Teoh
>Priority: Major
> Fix For: 1.19.0, 1.18.2
>
> Attachments: image-2023-12-18-11-37-11-914.png, 
> image-2023-12-18-11-37-29-596.png
>
>
> Prior to https://issues.apache.org/jira/browse/FLINK-32469, we see checkpoint 
> history for completed jobs (CANCELED, FAILED, FINISHED).
> After https://issues.apache.org/jira/browse/FLINK-32469, the checkpoint 
> history does not show up for completed jobs. 
> *Reproduction steps:*
>  # Start a Flink cluster.
>  # Submit a job with checkpointing enabled.
>  # Wait until at least 1 checkpoint completes.
>  # Cancel job.
>  # Open the Flink dashboard > Job > Checkpoints > History.
> We will see log line in JobManager saying "FlinkJobNotFoundException: Could 
> not find Flink job (  )"
> *Snapshot of failure:*
> When job is running, we can see checkpoints.
> !image-2023-12-18-11-37-11-914.png|width=862,height=295!
> When job has been CANCELLED, we no longer see checkpoints data.
> !image-2023-12-18-11-37-29-596.png|width=860,height=258!



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33778][table] Cleanup usage of deprecated TableConfig#setIdleS… [flink]

2023-12-18 Thread via GitHub


HuangXingBo merged PR #23894:
URL: https://github.com/apache/flink/pull/23894


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-33778) Cleanup usage of deprecated TableConfig#setIdleStateRetentionTime(Time, Time)

2023-12-18 Thread Jacky Lau (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17798149#comment-17798149
 ] 

Jacky Lau commented on FLINK-33778:
---

merged 3c86dcadf5366fa0026125051d69b0a8913d5e61

> Cleanup usage of deprecated TableConfig#setIdleStateRetentionTime(Time, Time)
> -
>
> Key: FLINK-33778
> URL: https://issues.apache.org/jira/browse/FLINK-33778
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Affects Versions: 1.19.0
>Reporter: Jacky Lau
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Closed] (FLINK-33778) Cleanup usage of deprecated TableConfig#setIdleStateRetentionTime(Time, Time)

2023-12-18 Thread Jacky Lau (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33778?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jacky Lau closed FLINK-33778.
-
Resolution: Fixed

> Cleanup usage of deprecated TableConfig#setIdleStateRetentionTime(Time, Time)
> -
>
> Key: FLINK-33778
> URL: https://issues.apache.org/jira/browse/FLINK-33778
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Affects Versions: 1.19.0
>Reporter: Jacky Lau
>Priority: Major
>  Labels: pull-request-available
> Fix For: 1.19.0
>
>




--
This message was sent by Atlassian Jira
(v8.20.10#820010)


  1   2   3   >