Re: [PR] [FLINK-14102] Introduce DB2Dialect. [flink-connector-jdbc]

2023-10-06 Thread via GitHub


davidradl commented on PR #8:
URL: 
https://github.com/apache/flink-connector-jdbc/pull/8#issuecomment-1750191357

   @MartijnVisser This is something we also would really like to be merged. 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-18356) flink-table-planner Exit code 137 returned from process

2023-10-06 Thread Sergey Nuyanzin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18356?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17772493#comment-17772493
 ] 

Sergey Nuyanzin commented on FLINK-18356:
-

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=53575&view=logs&j=a9db68b9-a7e0-54b6-0f98-010e0aff39e2&t=cdd32e0b-6047-565b-c58f-14054472f1be&l=11720

> flink-table-planner Exit code 137 returned from process
> ---
>
> Key: FLINK-18356
> URL: https://issues.apache.org/jira/browse/FLINK-18356
> Project: Flink
>  Issue Type: Bug
>  Components: Build System / Azure Pipelines, Tests
>Affects Versions: 1.12.0, 1.13.0, 1.14.0, 1.15.0, 1.16.0, 1.17.0, 1.18.0, 
> 1.19.0
>Reporter: Piotr Nowojski
>Assignee: Yunhong Zheng
>Priority: Critical
>  Labels: pull-request-available, test-stability
> Attachments: 1234.jpg, app-profiling_4.gif, 
> image-2023-01-11-22-21-57-784.png, image-2023-01-11-22-22-32-124.png, 
> image-2023-02-16-20-18-09-431.png, image-2023-07-11-19-28-52-851.png, 
> image-2023-07-11-19-35-54-530.png, image-2023-07-11-19-41-18-626.png, 
> image-2023-07-11-19-41-37-105.png
>
>
> {noformat}
> = test session starts 
> ==
> platform linux -- Python 3.7.3, pytest-5.4.3, py-1.8.2, pluggy-0.13.1
> cachedir: .tox/py37-cython/.pytest_cache
> rootdir: /__w/3/s/flink-python
> collected 568 items
> pyflink/common/tests/test_configuration.py ..[  
> 1%]
> pyflink/common/tests/test_execution_config.py ...[  
> 5%]
> pyflink/dataset/tests/test_execution_environment.py .
> ##[error]Exit code 137 returned from process: file name '/bin/docker', 
> arguments 'exec -i -u 1002 
> 97fc4e22522d2ced1f4d23096b8929045d083dd0a99a4233a8b20d0489e9bddb 
> /__a/externals/node/bin/node /__w/_temp/containerHandlerInvoker.js'.
> Finishing: Test - python
> {noformat}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=3729&view=logs&j=9cada3cb-c1d3-5621-16da-0f718fb86602&t=8d78fe4f-d658-5c70-12f8-4921589024c3



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-18356) flink-table-planner Exit code 137 returned from process

2023-10-06 Thread Sergey Nuyanzin (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-18356?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17772496#comment-17772496
 ] 

Sergey Nuyanzin commented on FLINK-18356:
-

https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=53578&view=logs&j=a9db68b9-a7e0-54b6-0f98-010e0aff39e2&t=cdd32e0b-6047-565b-c58f-14054472f1be&l=1000

> flink-table-planner Exit code 137 returned from process
> ---
>
> Key: FLINK-18356
> URL: https://issues.apache.org/jira/browse/FLINK-18356
> Project: Flink
>  Issue Type: Bug
>  Components: Build System / Azure Pipelines, Tests
>Affects Versions: 1.12.0, 1.13.0, 1.14.0, 1.15.0, 1.16.0, 1.17.0, 1.18.0, 
> 1.19.0
>Reporter: Piotr Nowojski
>Assignee: Yunhong Zheng
>Priority: Critical
>  Labels: pull-request-available, test-stability
> Attachments: 1234.jpg, app-profiling_4.gif, 
> image-2023-01-11-22-21-57-784.png, image-2023-01-11-22-22-32-124.png, 
> image-2023-02-16-20-18-09-431.png, image-2023-07-11-19-28-52-851.png, 
> image-2023-07-11-19-35-54-530.png, image-2023-07-11-19-41-18-626.png, 
> image-2023-07-11-19-41-37-105.png
>
>
> {noformat}
> = test session starts 
> ==
> platform linux -- Python 3.7.3, pytest-5.4.3, py-1.8.2, pluggy-0.13.1
> cachedir: .tox/py37-cython/.pytest_cache
> rootdir: /__w/3/s/flink-python
> collected 568 items
> pyflink/common/tests/test_configuration.py ..[  
> 1%]
> pyflink/common/tests/test_execution_config.py ...[  
> 5%]
> pyflink/dataset/tests/test_execution_environment.py .
> ##[error]Exit code 137 returned from process: file name '/bin/docker', 
> arguments 'exec -i -u 1002 
> 97fc4e22522d2ced1f4d23096b8929045d083dd0a99a4233a8b20d0489e9bddb 
> /__a/externals/node/bin/node /__w/_temp/containerHandlerInvoker.js'.
> Finishing: Test - python
> {noformat}
> https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=3729&view=logs&j=9cada3cb-c1d3-5621-16da-0f718fb86602&t=8d78fe4f-d658-5c70-12f8-4921589024c3



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-27076) Setup runners

2023-10-06 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-27076:
--
Description: 
h2. GitHub Documentation
* About self-hosted runners
* Adding self-hosted runners
* Using self-hosted runners in a GitHub Actions workflow
* Security hardening for self-hosted runners
* Autoscaling with self-hosted runners
* Alternative: Approving workflow runs from public forks: Not suitable for 
projects with large amounts of contributors (see Slack thread)
h2. Apache Infra
* Infra ML discussion thread from Apache Pekko
* GitHub self-hosted-runners (most-recently updated wiki article)
* GitHub Actions status (last modified on Jun 20, 2021)
.h2 GitHub Runner
* actions/runner
* Closed issue #494
* Feature request
* Draft PR #783 with fix from @ashb
.h2 GitHub Runner Fork from @ashb
* ashb/runner (releases/pr-security-options branch)
* Automated release of custom GitHub runner - Apparently, we need to deploy the 
most-recent version of the GitHub runner to avoid having the runner being 
rejected by GitHub.
.h2 Apache Flink ML
* GHA migration roadmap (December 16, 2021)
* Notice thread on migration to GitHub Actions (April 4, 2022)
.h2 Miscellaneous
* Apache Pulsar restricts CI runs to reviewed PRs only. Contributors are asked 
to create a branch in their fork as well to use GitHub’s runners, instead. The 
project itself relies on Apache’s hosted runners. (see related PR)
There is a discussion about it in the follow-ups of this Infra Slack post
There are concerns shared about losing contributors due to the extra work 
that’s necessary to set this up.


> Setup runners
> -
>
> Key: FLINK-27076
> URL: https://issues.apache.org/jira/browse/FLINK-27076
> Project: Flink
>  Issue Type: Sub-task
>  Components: Build System / CI
>Reporter: Chesnay Schepler
>Assignee: Chesnay Schepler
>Priority: Major
>
> h2. GitHub Documentation
> * About self-hosted runners
> * Adding self-hosted runners
> * Using self-hosted runners in a GitHub Actions workflow
> * Security hardening for self-hosted runners
> * Autoscaling with self-hosted runners
> * Alternative: Approving workflow runs from public forks: Not suitable for 
> projects with large amounts of contributors (see Slack thread)
> h2. Apache Infra
> * Infra ML discussion thread from Apache Pekko
> * GitHub self-hosted-runners (most-recently updated wiki article)
> * GitHub Actions status (last modified on Jun 20, 2021)
> .h2 GitHub Runner
> * actions/runner
> * Closed issue #494
> * Feature request
> * Draft PR #783 with fix from @ashb
> .h2 GitHub Runner Fork from @ashb
> * ashb/runner (releases/pr-security-options branch)
> * Automated release of custom GitHub runner - Apparently, we need to deploy 
> the most-recent version of the GitHub runner to avoid having the runner being 
> rejected by GitHub.
> .h2 Apache Flink ML
> * GHA migration roadmap (December 16, 2021)
> * Notice thread on migration to GitHub Actions (April 4, 2022)
> .h2 Miscellaneous
> * Apache Pulsar restricts CI runs to reviewed PRs only. Contributors are 
> asked to create a branch in their fork as well to use GitHub’s runners, 
> instead. The project itself relies on Apache’s hosted runners. (see related 
> PR)
> There is a discussion about it in the follow-ups of this Infra Slack post
> There are concerns shared about losing contributors due to the extra work 
> that’s necessary to set this up.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-27076) Setup runners

2023-10-06 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-27076:
--
Description: 
h2. GitHub Documentation
 * [About self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/about-self-hosted-runners]
 * [Adding self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/adding-self-hosted-runners]
 * [Using self-hosted runners in a GitHub Actions 
workflow|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/using-self-hosted-runners-in-a-workflow]
 * [Security hardening for self-hosted 
runners|https://docs.github.com/en/actions/security-guides/security-hardening-for-github-actions#hardening-for-self-hosted-runners]
 * [Autoscaling with self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/autoscaling-with-self-hosted-runners]
 * Alternative: [Approving workflow runs from public 
forks|https://docs.github.com/en/actions/managing-workflow-runs/approving-workflow-runs-from-public-forks]
 (not suitable for projects with large amounts of contributors; see [Slack 
thread|https://the-asf.slack.com/archives/CBX4TSBQ8/p1640076905230600])

h2. Apache Infra
 * [Infra ML discussion thread from Apache 
Pekko|https://cwiki.apache.org/confluence/display/INFRA/GitHub+self-hosted+runners]
 * [GitHub 
self-hosted-runners|https://cwiki.apache.org/confluence/display/INFRA/GitHub+self-hosted+runners]
 (most-recently updated wiki article)
 * [GitHub Actions 
status|https://cwiki.apache.org/confluence/display/BUILDS/GitHub+Actions+status]
 (last modified on Jun 20, 2021)

h2. GitHub Runner
 * [actions/runner|https://github.com/actions/runner]
 * [Closed issue #494|https://github.com/actions/runner/issues/494]
 * [Feature request|https://github.com/orgs/community/discussions/53430]
 * [Draft PR #783|https://github.com/actions/runner/pull/783] with fix from 
[@ashb|https://github.com/ashb]

h2. GitHub Runner Fork from [@ashb|https://github.com/ashb]
 * [ashb/runner|https://github.com/ashb/runner] 
([releases/pr-security-options|https://github.com/actions/runner/compare/main...ashb:runner:releases/pr-security-options]
 branch)
 * [Automated release of custom GitHub 
runner|https://github.com/ashb/runner/blob/main/.github/workflows/sync-with-upstream.yml]
 - Apparently, we need to deploy the most-recent version of the GitHub runner 
to avoid having the runner being rejected by GitHub.

h2. Apache Flink ML
 * [GHA migration 
roadmap|https://lists.apache.org/thread/vcyx2nx0mhklqwm827vgykv8pc54gg3k] 
(December 16, 2021)
 * [Notice thread on migration to GitHub 
Actions|https://lists.apache.org/thread/438krxbs7y9lg0f009f1hlojqm54y57s] 
(April 4, 2022)

h2. Miscellaneous
 * Apache Pulsar restricts CI runs to reviewed PRs only. Contributors are asked 
to create a branch in their fork as well to use GitHub’s runners, instead. The 
project itself relies on Apache’s hosted runners. (see [related 
PR|https://github.com/apache/pulsar/pull/17693/files])
 ** There is a discussion about it in the follow-ups of [this Infra Slack 
post|https://the-asf.slack.com/archives/CBX4TSBQ8/p1663346680496919?thread_ts=1661512133.913279&cid=CBX4TSBQ8]
 ** There are concerns shared about losing contributors due to the extra work 
that’s necessary to set this up.

  was:
h2. GitHub Documentation
* About self-hosted runners
* Adding self-hosted runners
* Using self-hosted runners in a GitHub Actions workflow
* Security hardening for self-hosted runners
* Autoscaling with self-hosted runners
* Alternative: Approving workflow runs from public forks: Not suitable for 
projects with large amounts of contributors (see Slack thread)
h2. Apache Infra
* Infra ML discussion thread from Apache Pekko
* GitHub self-hosted-runners (most-recently updated wiki article)
* GitHub Actions status (last modified on Jun 20, 2021)
.h2 GitHub Runner
* actions/runner
* Closed issue #494
* Feature request
* Draft PR #783 with fix from @ashb
.h2 GitHub Runner Fork from @ashb
* ashb/runner (releases/pr-security-options branch)
* Automated release of custom GitHub runner - Apparently, we need to deploy the 
most-recent version of the GitHub runner to avoid having the runner being 
rejected by GitHub.
.h2 Apache Flink ML
* GHA migration roadmap (December 16, 2021)
* Notice thread on migration to GitHub Actions (April 4, 2022)
.h2 Miscellaneous
* Apache Pulsar restricts CI runs to reviewed PRs only. Contributors are asked 
to create a branch in their fork as well to use GitHub’s runners, instead. The 
project itself relies on Apache’s hosted runners. (see related PR)
There is a discussion about it in the follow-ups of this Infra Slack post
There are concerns shared about losing contributors due to the extra work 
that’s necessary to set this up.



> Setup runners

[jira] [Updated] (FLINK-27076) Setup runners

2023-10-06 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-27076:
--
Description: 
Below is a collection of links around the topic of self-hosted runners and the 
current state of the GitHub Actions migration efforts in Apache Flink.
h2. GitHub Documentation
 * [About self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/about-self-hosted-runners]
 * [Adding self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/adding-self-hosted-runners]
 * [Using self-hosted runners in a GitHub Actions 
workflow|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/using-self-hosted-runners-in-a-workflow]
 * [Security hardening for self-hosted 
runners|https://docs.github.com/en/actions/security-guides/security-hardening-for-github-actions#hardening-for-self-hosted-runners]
 * [Autoscaling with self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/autoscaling-with-self-hosted-runners]
 * Alternative: [Approving workflow runs from public 
forks|https://docs.github.com/en/actions/managing-workflow-runs/approving-workflow-runs-from-public-forks]
 (not suitable for projects with large amounts of contributors; see [Slack 
thread|https://the-asf.slack.com/archives/CBX4TSBQ8/p1640076905230600])

h2. Apache Infra
 * [Infra ML discussion thread from Apache 
Pekko|https://cwiki.apache.org/confluence/display/INFRA/GitHub+self-hosted+runners]
 * [GitHub 
self-hosted-runners|https://cwiki.apache.org/confluence/display/INFRA/GitHub+self-hosted+runners]
 (most-recently updated wiki article)
 * [GitHub Actions 
status|https://cwiki.apache.org/confluence/display/BUILDS/GitHub+Actions+status]
 (last modified on Jun 20, 2021)

h2. GitHub Runner
 * [actions/runner|https://github.com/actions/runner]
 * [Closed issue #494|https://github.com/actions/runner/issues/494]
 * [Feature request|https://github.com/orgs/community/discussions/53430]
 * [Draft PR #783|https://github.com/actions/runner/pull/783] with fix from 
[@ashb|https://github.com/ashb]

h2. GitHub Runner Fork from [@ashb|https://github.com/ashb]
 * [ashb/runner|https://github.com/ashb/runner] 
([releases/pr-security-options|https://github.com/actions/runner/compare/main...ashb:runner:releases/pr-security-options]
 branch)
 * [Automated release of custom GitHub 
runner|https://github.com/ashb/runner/blob/main/.github/workflows/sync-with-upstream.yml]
 - Apparently, we need to deploy the most-recent version of the GitHub runner 
to avoid having the runner being rejected by GitHub.

h2. Apache Flink ML
 * [GHA migration 
roadmap|https://lists.apache.org/thread/vcyx2nx0mhklqwm827vgykv8pc54gg3k] 
(December 16, 2021)
 * [Notice thread on migration to GitHub 
Actions|https://lists.apache.org/thread/438krxbs7y9lg0f009f1hlojqm54y57s] 
(April 4, 2022)

h2. Miscellaneous
 * Apache Pulsar restricts CI runs to reviewed PRs only. Contributors are asked 
to create a branch in their fork as well to use GitHub’s runners, instead. The 
project itself relies on Apache’s hosted runners. (see [related 
PR|https://github.com/apache/pulsar/pull/17693/files])
 ** There is a discussion about it in the follow-ups of [this Infra Slack 
post|https://the-asf.slack.com/archives/CBX4TSBQ8/p1663346680496919?thread_ts=1661512133.913279&cid=CBX4TSBQ8]
 ** There are concerns shared about losing contributors due to the extra work 
that’s necessary to set this up.

  was:
h2. GitHub Documentation
 * [About self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/about-self-hosted-runners]
 * [Adding self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/adding-self-hosted-runners]
 * [Using self-hosted runners in a GitHub Actions 
workflow|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/using-self-hosted-runners-in-a-workflow]
 * [Security hardening for self-hosted 
runners|https://docs.github.com/en/actions/security-guides/security-hardening-for-github-actions#hardening-for-self-hosted-runners]
 * [Autoscaling with self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/autoscaling-with-self-hosted-runners]
 * Alternative: [Approving workflow runs from public 
forks|https://docs.github.com/en/actions/managing-workflow-runs/approving-workflow-runs-from-public-forks]
 (not suitable for projects with large amounts of contributors; see [Slack 
thread|https://the-asf.slack.com/archives/CBX4TSBQ8/p1640076905230600])

h2. Apache Infra
 * [Infra ML discussion thread from Apache 
Pekko|https://cwiki.apache.org/confluence/display/INFRA/GitHub+self-hosted+runners]
 * [GitHub 
self-hosted-runners|ht

[jira] [Created] (FLINK-33197) PyFlink support for ByteArraySchema

2023-10-06 Thread Liu Chong (Jira)
Liu Chong created FLINK-33197:
-

 Summary: PyFlink support for ByteArraySchema
 Key: FLINK-33197
 URL: https://issues.apache.org/jira/browse/FLINK-33197
 Project: Flink
  Issue Type: New Feature
  Components: API / Python
Affects Versions: 1.17.0
Reporter: Liu Chong


Currently in Python Flink API, when reading messages from a Kafka source, only 
SimpleStringSchema is available.
If the data is in arbitary binary format(e.g. marshalled Protocol Buffer msg) 
it may not be decodable with the default 'utf-8' encoding. 
There's currently a workaround which is to manually set the encoding to 
'ISO-8859-1' which supports all possible byte combinations. 
However this is not an elegant solution.
We should support ByteArraySchema which outputs a raw byte array for subsequent 
unmarshalling.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-27076) Setup runners

2023-10-06 Thread Matthias Pohl (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-27076?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Matthias Pohl updated FLINK-27076:
--
Description: 
Below is a collection of links around the topic of self-hosted runners and the 
current state of the GitHub Actions migration efforts in Apache Flink.
h2. GitHub Documentation
 * [About self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/about-self-hosted-runners]
 * [Adding self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/adding-self-hosted-runners]
 * [Using self-hosted runners in a GitHub Actions 
workflow|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/using-self-hosted-runners-in-a-workflow]
 * [Security hardening for self-hosted 
runners|https://docs.github.com/en/actions/security-guides/security-hardening-for-github-actions#hardening-for-self-hosted-runners]
 * [Autoscaling with self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/autoscaling-with-self-hosted-runners]
 * Alternative: [Approving workflow runs from public 
forks|https://docs.github.com/en/actions/managing-workflow-runs/approving-workflow-runs-from-public-forks]
 (not suitable for projects with large amounts of contributors; see [Slack 
thread|https://the-asf.slack.com/archives/CBX4TSBQ8/p1640076905230600])

h2. Apache Infra
 * [Infra ML discussion thread from Apache 
Pekko|https://cwiki.apache.org/confluence/display/INFRA/GitHub+self-hosted+runners]
 * [GitHub 
self-hosted-runners|https://cwiki.apache.org/confluence/display/INFRA/GitHub+self-hosted+runners]
 (most-recently updated wiki article)
 * [GitHub Actions 
status|https://cwiki.apache.org/confluence/display/BUILDS/GitHub+Actions+status]
 (last modified on Jun 20, 2021)

h2. GitHub Runner
 * [actions/runner|https://github.com/actions/runner]
 * [Closed issue #494|https://github.com/actions/runner/issues/494]
 * [Feature request|https://github.com/orgs/community/discussions/53430]
 * [Draft PR #783|https://github.com/actions/runner/pull/783] with fix from 
[@ashb|https://github.com/ashb]

h2. GitHub Runner Fork from [@ashb|https://github.com/ashb]
 * [ashb/runner|https://github.com/ashb/runner] 
([releases/pr-security-options|https://github.com/actions/runner/compare/main...ashb:runner:releases/pr-security-options]
 branch)
 * [Automated release of custom GitHub 
runner|https://github.com/ashb/runner/blob/main/.github/workflows/sync-with-upstream.yml]
 - Apparently, we need to deploy the most-recent version of the GitHub runner 
to avoid having the runner being rejected by GitHub.

h2. Apache Flink
 * ML: [GHA migration 
roadmap|https://lists.apache.org/thread/vcyx2nx0mhklqwm827vgykv8pc54gg3k] 
(December 16, 2021)
 * ML: [Notice thread on migration to GitHub 
Actions|https://lists.apache.org/thread/438krxbs7y9lg0f009f1hlojqm54y57s] 
(April 4, 2022)
 * [exp_github_actions 
branch|https://github.com/apache/flink/tree/exp_github_actions] in Apache Flink 
repo

h2. Miscellaneous
 * Apache Pulsar restricts CI runs to reviewed PRs only. Contributors are asked 
to create a branch in their fork as well to use GitHub’s runners, instead. The 
project itself relies on Apache’s hosted runners. (see [related 
PR|https://github.com/apache/pulsar/pull/17693/files])
 ** There is a discussion about it in the follow-ups of [this Infra Slack 
post|https://the-asf.slack.com/archives/CBX4TSBQ8/p1663346680496919?thread_ts=1661512133.913279&cid=CBX4TSBQ8]
 ** There are concerns shared about losing contributors due to the extra work 
that’s necessary to set this up.

  was:
Below is a collection of links around the topic of self-hosted runners and the 
current state of the GitHub Actions migration efforts in Apache Flink.
h2. GitHub Documentation
 * [About self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/about-self-hosted-runners]
 * [Adding self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/adding-self-hosted-runners]
 * [Using self-hosted runners in a GitHub Actions 
workflow|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/using-self-hosted-runners-in-a-workflow]
 * [Security hardening for self-hosted 
runners|https://docs.github.com/en/actions/security-guides/security-hardening-for-github-actions#hardening-for-self-hosted-runners]
 * [Autoscaling with self-hosted 
runners|https://docs.github.com/en/actions/hosting-your-own-runners/managing-self-hosted-runners/autoscaling-with-self-hosted-runners]
 * Alternative: [Approving workflow runs from public 
forks|https://docs.github.com/en/actions/managing-workflow-runs/approving-workflow-runs-from-public-forks]
 (not suitable for projects with large amounts of contributors

[jira] [Comment Edited] (FLINK-33104) Nightly run for Flink Kafka connector fails due to architecture tests failing

2023-10-06 Thread Etienne Chauchot (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33104?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17770338#comment-17770338
 ] 

Etienne Chauchot edited comment on FLINK-33104 at 10/6/23 10:27 AM:


[~martijnvisser] actually it was never fixed in cassandra (see FLINK-32353): 
the problem is that when archunit rules change we need to update the archunit 
violation store. And there is only a single violation store. As the nightly 
tests the connector against several versions of flink, there will be failures 
if not all these versions have the same archunit rules. To fix this problem we 
need to skip archunit tests on Flink versions that are not the main one (the 
one the connector is built against) so that the single arunit violation store 
contains the violations for the main flink version. I proposed some change in 
the github action script 
[here|https://issues.apache.org/jira/browse/FLINK-32563]  for which I was 
waiting for your feedback.


was (Author: echauchot):
[~martijnvisser] actually it was never fixed in cassandra (see FLINK-32353): 
the problem is that when archunit rules change we need to update the archunit 
violation store. And there is only a single violation store. As the nightly 
tests the connector against several versions of flink, there will be failures 
if not all these versions have the same archunit rules. To fix this problem we 
need to skip archunit tests on Flink versions that are not the main one (the 
one the connector is built against) so that the single arunit violation store 
contains the violations for the main flink version. I proposed some change in 
the github action script here  for which I was waiting for your feedback.

> Nightly run for Flink Kafka connector fails due to architecture tests failing
> -
>
> Key: FLINK-33104
> URL: https://issues.apache.org/jira/browse/FLINK-33104
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
>Affects Versions: kafka-3.1.0
>Reporter: Martijn Visser
>Priority: Blocker
>
> {code:java}
> 2023-09-17T00:29:07.1675694Z [WARNING] Tests run: 18, Failures: 0, Errors: 0, 
> Skipped: 9, Time elapsed: 308.532 s - in 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducerMigrationTest
> 2023-09-17T00:29:07.5171608Z [INFO] 
> 2023-09-17T00:29:07.5172360Z [INFO] Results:
> 2023-09-17T00:29:07.5172773Z [INFO] 
> 2023-09-17T00:29:07.5173139Z [ERROR] Failures: 
> 2023-09-17T00:29:07.5174181Z [ERROR]   Architecture Violation [Priority: 
> MEDIUM] - Rule 'ITCASE tests should use a MiniCluster resource or extension' 
> was violated (13 times):
> 2023-09-17T00:29:07.5176050Z 
> org.apache.flink.connector.kafka.sink.FlinkKafkaInternalProducerITCase does 
> not satisfy: only one of the following predicates match:
> 2023-09-17T00:29:07.5177452Z * reside in a package 
> 'org.apache.flink.runtime.*' and contain any fields that are static, final, 
> and of type InternalMiniClusterExtension and annotated with @RegisterExtension
> 2023-09-17T00:29:07.5179831Z * reside outside of package 
> 'org.apache.flink.runtime.*' and contain any fields that are static, final, 
> and of type MiniClusterExtension and annotated with @RegisterExtension or are 
> , and of type MiniClusterTestEnvironment and annotated with @TestEnv
> 2023-09-17T00:29:07.5181277Z * reside in a package 
> 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class 
> InternalMiniClusterExtension
> 2023-09-17T00:29:07.5182154Z * reside outside of package 
> 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class 
> MiniClusterExtension
> 2023-09-17T00:29:07.5182951Z  or contain any fields that are public, static, 
> and of type MiniClusterWithClientResource and final and annotated with 
> @ClassRule or contain any fields that is of type 
> MiniClusterWithClientResource and public and final and not static and 
> annotated with @Rule
> 2023-09-17T00:29:07.5183906Z 
> org.apache.flink.connector.kafka.sink.KafkaSinkITCase does not satisfy: only 
> one of the following predicates match:
> 2023-09-17T00:29:07.5184769Z * reside in a package 
> 'org.apache.flink.runtime.*' and contain any fields that are static, final, 
> and of type InternalMiniClusterExtension and annotated with @RegisterExtension
> 2023-09-17T00:29:07.5185812Z * reside outside of package 
> 'org.apache.flink.runtime.*' and contain any fields that are static, final, 
> and of type MiniClusterExtension and annotated with @RegisterExtension or are 
> , and of type MiniClusterTestEnvironment and annotated with @TestEnv
> 2023-09-17T00:29:07.5186880Z * reside in a package 
> 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class 
> InternalMiniClusterExtension
> 2023-09-17T00:29:07.518792

[jira] [Commented] (FLINK-33104) Nightly run for Flink Kafka connector fails due to architecture tests failing

2023-10-06 Thread Etienne Chauchot (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33104?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17772516#comment-17772516
 ] 

Etienne Chauchot commented on FLINK-33104:
--

[~tzulitai] yes this is exactly what I wanted to do for Cassandra but it 
requires a change in the github actions scripts and/or on main build proposed 
[here|https://issues.apache.org/jira/browse/FLINK-32563] but I received no 
feedback. Among the 3 proposed changes in that ticket, I prefer the one that 
links the dep convergence tests and the archunit tests into a single maven 
profile (like architecture tests). I'll do this change in a PR.

> Nightly run for Flink Kafka connector fails due to architecture tests failing
> -
>
> Key: FLINK-33104
> URL: https://issues.apache.org/jira/browse/FLINK-33104
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Kafka
>Affects Versions: kafka-3.1.0
>Reporter: Martijn Visser
>Priority: Blocker
>
> {code:java}
> 2023-09-17T00:29:07.1675694Z [WARNING] Tests run: 18, Failures: 0, Errors: 0, 
> Skipped: 9, Time elapsed: 308.532 s - in 
> org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducerMigrationTest
> 2023-09-17T00:29:07.5171608Z [INFO] 
> 2023-09-17T00:29:07.5172360Z [INFO] Results:
> 2023-09-17T00:29:07.5172773Z [INFO] 
> 2023-09-17T00:29:07.5173139Z [ERROR] Failures: 
> 2023-09-17T00:29:07.5174181Z [ERROR]   Architecture Violation [Priority: 
> MEDIUM] - Rule 'ITCASE tests should use a MiniCluster resource or extension' 
> was violated (13 times):
> 2023-09-17T00:29:07.5176050Z 
> org.apache.flink.connector.kafka.sink.FlinkKafkaInternalProducerITCase does 
> not satisfy: only one of the following predicates match:
> 2023-09-17T00:29:07.5177452Z * reside in a package 
> 'org.apache.flink.runtime.*' and contain any fields that are static, final, 
> and of type InternalMiniClusterExtension and annotated with @RegisterExtension
> 2023-09-17T00:29:07.5179831Z * reside outside of package 
> 'org.apache.flink.runtime.*' and contain any fields that are static, final, 
> and of type MiniClusterExtension and annotated with @RegisterExtension or are 
> , and of type MiniClusterTestEnvironment and annotated with @TestEnv
> 2023-09-17T00:29:07.5181277Z * reside in a package 
> 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class 
> InternalMiniClusterExtension
> 2023-09-17T00:29:07.5182154Z * reside outside of package 
> 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class 
> MiniClusterExtension
> 2023-09-17T00:29:07.5182951Z  or contain any fields that are public, static, 
> and of type MiniClusterWithClientResource and final and annotated with 
> @ClassRule or contain any fields that is of type 
> MiniClusterWithClientResource and public and final and not static and 
> annotated with @Rule
> 2023-09-17T00:29:07.5183906Z 
> org.apache.flink.connector.kafka.sink.KafkaSinkITCase does not satisfy: only 
> one of the following predicates match:
> 2023-09-17T00:29:07.5184769Z * reside in a package 
> 'org.apache.flink.runtime.*' and contain any fields that are static, final, 
> and of type InternalMiniClusterExtension and annotated with @RegisterExtension
> 2023-09-17T00:29:07.5185812Z * reside outside of package 
> 'org.apache.flink.runtime.*' and contain any fields that are static, final, 
> and of type MiniClusterExtension and annotated with @RegisterExtension or are 
> , and of type MiniClusterTestEnvironment and annotated with @TestEnv
> 2023-09-17T00:29:07.5186880Z * reside in a package 
> 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class 
> InternalMiniClusterExtension
> 2023-09-17T00:29:07.5187929Z * reside outside of package 
> 'org.apache.flink.runtime.*' and is annotated with @ExtendWith with class 
> MiniClusterExtension
> 2023-09-17T00:29:07.5189073Z  or contain any fields that are public, static, 
> and of type MiniClusterWithClientResource and final and annotated with 
> @ClassRule or contain any fields that is of type 
> MiniClusterWithClientResource and public and final and not static and 
> annotated with @Rule
> 2023-09-17T00:29:07.5190076Z 
> org.apache.flink.connector.kafka.sink.KafkaTransactionLogITCase does not 
> satisfy: only one of the following predicates match:
> 2023-09-17T00:29:07.5190946Z * reside in a package 
> 'org.apache.flink.runtime.*' and contain any fields that are static, final, 
> and of type InternalMiniClusterExtension and annotated with @RegisterExtension
> 2023-09-17T00:29:07.5191983Z * reside outside of package 
> 'org.apache.flink.runtime.*' and contain any fields that are static, final, 
> and of type MiniClusterExtension and annotated with @RegisterExtension or are 
> , and of type MiniClusterTestEnvironment and annotated wit

Re: [PR] [FLINK-32353] [hotfix] build against last flink 1.16 and 1.17 and enable archunit only on main flink release [flink-connector-cassandra]

2023-10-06 Thread via GitHub


echauchot commented on PR #19:
URL: 
https://github.com/apache/flink-connector-cassandra/pull/19#issuecomment-1750384738

   Hi @tzulitai yes I fell back this PR's scope to just upgrading the versions 
hence the self merge. But, as I just wrote in [this 
ticket](https://issues.apache.org/jira/browse/FLINK-33104?focusedCommentId=17772516&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-17772516),
 I'll submit a new PR to change the build system and tackle this problem. 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33159] [build] use variables in pom file for Java and Maven versions [flink]

2023-10-06 Thread via GitHub


davidradl commented on code in PR #23469:
URL: https://github.com/apache/flink/pull/23469#discussion_r1348594028


##
flink-walkthroughs/flink-walkthrough-datastream-java/src/main/resources/archetype-resources/pom.xml:
##
@@ -31,7 +31,6 @@ under the License.


UTF-8
@project.version@
-   1.8

Review Comment:
   no worries -  sorry misunderstood - I will reinstate this.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33159] [build] use variables in pom file for Java and Maven versions [flink]

2023-10-06 Thread via GitHub


davidradl commented on code in PR #23469:
URL: https://github.com/apache/flink/pull/23469#discussion_r1348594028


##
flink-walkthroughs/flink-walkthrough-datastream-java/src/main/resources/archetype-resources/pom.xml:
##
@@ -31,7 +31,6 @@ under the License.


UTF-8
@project.version@
-   1.8

Review Comment:
   no worries -  sorry I misunderstood - I will reinstate this.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33046][FLIP-333] Redesign Apache Flink website [flink-web]

2023-10-06 Thread via GitHub


dannycranmer commented on PR #676:
URL: https://github.com/apache/flink-web/pull/676#issuecomment-1750463733

   > Not sure if it's intended, but when hovering over the Flink logo in top 
left, then "Apache Flink" becomes visible (in dark mode). I think we should 
always display it, because the sudden appearance after a hover feels wrong?
   
   This is not intentional. I cannot reproduce this, I am also using Chrome on 
Mac. Any other details you can provide to help me reproduce? Chrome version?
   
   > The menu isn't on the left side like in the FLIP, but on the right hand 
side
   
   We put it on the left initially and internal feedback was that users thought 
it was a nav, rather than page index. We moved to right as per this feedback, 
this also aligns with the previous website. What do you think? I can raise this 
on the discussion thread if you think it is worth it.
   
   > When a link has been visited, we're using a different color in the menu. 
Can we just keep the same color as it is initially? This looks messy
   
   Ack, I agree, will do. How about link within a page?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[PR] add announcement blog post for Flink 1.18 [flink-web]

2023-10-06 Thread via GitHub


knaufk opened a new pull request, #680:
URL: https://github.com/apache/flink-web/pull/680

   (no comment)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33159] [build] use variables in pom file for Java and Maven versions [flink]

2023-10-06 Thread via GitHub


davidradl commented on PR #23469:
URL: https://github.com/apache/flink/pull/23469#issuecomment-1750644172

   @ferenc-csaky thanks I have made the requested change.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-32380] Support Java Records with PojoTypeInfo/Serializer [flink]

2023-10-06 Thread via GitHub


gyfora commented on PR #23490:
URL: https://github.com/apache/flink/pull/23490#issuecomment-1750668456

   I just realised that the current implementation won't work with schema 
evolution as we have to "remap" the parameter indexes for the constructor 
afterwards, also defining new primitive fields can be problematic. 


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] add announcement blog post for Flink 1.18 [flink-web]

2023-10-06 Thread via GitHub


davidradl commented on code in PR #680:
URL: https://github.com/apache/flink-web/pull/680#discussion_r1348732092


##
docs/content/posts/2023-10-10-release-1.18.0.md:
##
@@ -0,0 +1,542 @@
+---
+authors:
+- JingGe:
+  name: "Jing Ge"
+  twitter: jingengineer
+- KonstantinKnauf:
+  name: "Konstantin Knauf"
+  twitter: snntrable
+- SergeyNuyanzin:
+  name: "Sergey Nuyanzin"
+  twitter: uckamello
+- QingshengRen:
+  name: "Qingsheng Ren"
+  twitter: renqstuite
+date: "2023-10-10T08:00:00Z"
+subtitle: ""
+title: Announcing the Release of Apache Flink 1.18
+aliases:
+- /news/2023/10/10/release-1.18.0.html
+---
+
+The Apache Flink PMC is pleased to announce the release of Apache Flink 
1.18.0. As usual, we are looking at a packed 
+release with a wide variety of improvements and new features. Overall, 176 
people contributed to this release completing 
+18 FLIPS and 700+ issues. Thank you!
+
+Let's dive into the highlights.
+
+# Towards a Streaming Lakehouse
+
+## Flink SQL Improvements
+
+### Introduce Flink JDBC Driver For Sql Gateway 
+
+Flink 1.18 comes with a JDBC Driver for the Flink SQL Gateway. So, you can now 
use any SQL Client that supports JDBC to 
+interact with your tables via Flink SQL. Here is an example using 
[SQLLine](https://julianhyde.github.io/sqlline/manual.html). 
+
+```shell
+sqlline> !connect jdbc:flink://localhost:8083
+```
+
+```shell
+sqlline version 1.12.0
+sqlline> !connect jdbc:flink://localhost:8083
+Enter username for jdbc:flink://localhost:8083:
+Enter password for jdbc:flink://localhost:8083:
+0: jdbc:flink://localhost:8083> CREATE TABLE T(
+. . . . . . . . . . . . . . .)>  a INT,
+. . . . . . . . . . . . . . .)>  b VARCHAR(10)
+. . . . . . . . . . . . . . .)>  ) WITH (
+. . . . . . . . . . . . . . .)>  'connector' = 'filesystem',
+. . . . . . . . . . . . . . .)>  'path' = 'file:///tmp/T.csv',
+. . . . . . . . . . . . . . .)>  'format' = 'csv'
+. . . . . . . . . . . . . . .)>  );
+No rows affected (0.122 seconds)
+0: jdbc:flink://localhost:8083> INSERT INTO T VALUES (1, 'Hi'), (2, 'Hello');
++--+
+|  job id  |
++--+
+| fbade1ab4450fc57ebd5269fdf60dcfd |
++--+
+1 row selected (1.282 seconds)
+0: jdbc:flink://localhost:8083> SELECT * FROM T;
++---+---+
+| a |   b   |
++---+---+
+| 1 | Hi|
+| 2 | Hello |
++---+---+
+2 rows selected (1.955 seconds)
+0: jdbc:flink://localhost:8083>
+```
+
+**More Information**
+* 
[Documentation](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/jdbcdriver/)
 
+* [FLIP-293: Introduce Flink Jdbc Driver For Sql 
Gateway](https://cwiki.apache.org/confluence/display/FLINK/FLIP-293%3A+Introduce+Flink+Jdbc+Driver+For+Sql+Gateway)
+
+
+### Stored Procedures
+
+Stored Procedures provide a convenient way to encapsulate complex logic to 
perform data manipulation or administrative 
+tasks in Apache Flink itself. Therefore, Flink introduces the support for 
calling stored procedures. 
+Flink now allows catalog developers to develop their own built-in stored 
procedures and then enables users to call these
+predefined stored procedures.
+
+**More Information**
+* 
[Documentation](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/procedures/)
+* [FLIP-311: Support Call Stored 
Procedure](https://cwiki.apache.org/confluence/display/FLINK/FLIP-311%3A+Support+Call+Stored+Procedure)
+
+### Extended DDL Support
+
+From this release onwards, Flink supports
+
+- `REPLACE TABLE AS SELECT`
+- `CREATE OR REPLACE TABLE AS SELECT`
+
+and both these commands and previously supported `CREATE TABLE AS` can now 
support atomicity provided the underlying 
+connector supports this.
+
+Moreover, Apache Flink now supports TRUNCATE TABLE in batch execution mode. As 
before, the underlying connector needs 
+to implement and provide this capability
+
+And, finally, we have also added support for adding, dropping and listing 
partitions via
+
+- `ALTER TABLE ADD PARTITION`
+- `ALTER TABLE DROP PARTITION`
+- `SHOW PARTITIONS`
+
+**More Information**
+- [Documentation on 
TRUNCATE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/truncate/)
+- [Documentation on CREATE OR 
REPLACE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/create/#create-or-replace-table)
+- [Documentation on ALTER 
TABLE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/alter/#alter-table)
+- [FLIP-302: Support TRUNCATE TABLE statement in batch 
mode](https://cwiki.apache.org/confluence/display/FLINK/FLIP-302%3A+Support+TRUNCATE+TABLE+statement+in+batch+mode)
+- [FLIP-303: Support REPLACE TABLE AS SELECT 
statement](https://cwiki.apache.org/confluence/display/FLINK/FLIP-303%3A+Support+REPLACE+TABLE+AS+SELECT+statement)
+- [FLIP-305: Support atomic for CREATE TABLE AS SELECT(CTAS) 
statement](https://cwiki.apache.org/confluence/display/FLINK

Re: [PR] add announcement blog post for Flink 1.18 [flink-web]

2023-10-06 Thread via GitHub


davidradl commented on code in PR #680:
URL: https://github.com/apache/flink-web/pull/680#discussion_r1348735360


##
docs/content/posts/2023-10-10-release-1.18.0.md:
##
@@ -0,0 +1,542 @@
+---
+authors:
+- JingGe:
+  name: "Jing Ge"
+  twitter: jingengineer
+- KonstantinKnauf:
+  name: "Konstantin Knauf"
+  twitter: snntrable
+- SergeyNuyanzin:
+  name: "Sergey Nuyanzin"
+  twitter: uckamello
+- QingshengRen:
+  name: "Qingsheng Ren"
+  twitter: renqstuite
+date: "2023-10-10T08:00:00Z"
+subtitle: ""
+title: Announcing the Release of Apache Flink 1.18
+aliases:
+- /news/2023/10/10/release-1.18.0.html
+---
+
+The Apache Flink PMC is pleased to announce the release of Apache Flink 
1.18.0. As usual, we are looking at a packed 
+release with a wide variety of improvements and new features. Overall, 176 
people contributed to this release completing 
+18 FLIPS and 700+ issues. Thank you!
+
+Let's dive into the highlights.
+
+# Towards a Streaming Lakehouse
+
+## Flink SQL Improvements
+
+### Introduce Flink JDBC Driver For Sql Gateway 
+
+Flink 1.18 comes with a JDBC Driver for the Flink SQL Gateway. So, you can now 
use any SQL Client that supports JDBC to 
+interact with your tables via Flink SQL. Here is an example using 
[SQLLine](https://julianhyde.github.io/sqlline/manual.html). 
+
+```shell
+sqlline> !connect jdbc:flink://localhost:8083
+```
+
+```shell
+sqlline version 1.12.0
+sqlline> !connect jdbc:flink://localhost:8083
+Enter username for jdbc:flink://localhost:8083:
+Enter password for jdbc:flink://localhost:8083:
+0: jdbc:flink://localhost:8083> CREATE TABLE T(
+. . . . . . . . . . . . . . .)>  a INT,
+. . . . . . . . . . . . . . .)>  b VARCHAR(10)
+. . . . . . . . . . . . . . .)>  ) WITH (
+. . . . . . . . . . . . . . .)>  'connector' = 'filesystem',
+. . . . . . . . . . . . . . .)>  'path' = 'file:///tmp/T.csv',
+. . . . . . . . . . . . . . .)>  'format' = 'csv'
+. . . . . . . . . . . . . . .)>  );
+No rows affected (0.122 seconds)
+0: jdbc:flink://localhost:8083> INSERT INTO T VALUES (1, 'Hi'), (2, 'Hello');
++--+
+|  job id  |
++--+
+| fbade1ab4450fc57ebd5269fdf60dcfd |
++--+
+1 row selected (1.282 seconds)
+0: jdbc:flink://localhost:8083> SELECT * FROM T;
++---+---+
+| a |   b   |
++---+---+
+| 1 | Hi|
+| 2 | Hello |
++---+---+
+2 rows selected (1.955 seconds)
+0: jdbc:flink://localhost:8083>
+```
+
+**More Information**
+* 
[Documentation](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/jdbcdriver/)
 
+* [FLIP-293: Introduce Flink Jdbc Driver For Sql 
Gateway](https://cwiki.apache.org/confluence/display/FLINK/FLIP-293%3A+Introduce+Flink+Jdbc+Driver+For+Sql+Gateway)
+
+
+### Stored Procedures
+
+Stored Procedures provide a convenient way to encapsulate complex logic to 
perform data manipulation or administrative 
+tasks in Apache Flink itself. Therefore, Flink introduces the support for 
calling stored procedures. 
+Flink now allows catalog developers to develop their own built-in stored 
procedures and then enables users to call these
+predefined stored procedures.
+
+**More Information**
+* 
[Documentation](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/procedures/)
+* [FLIP-311: Support Call Stored 
Procedure](https://cwiki.apache.org/confluence/display/FLINK/FLIP-311%3A+Support+Call+Stored+Procedure)
+
+### Extended DDL Support
+
+From this release onwards, Flink supports
+
+- `REPLACE TABLE AS SELECT`
+- `CREATE OR REPLACE TABLE AS SELECT`
+
+and both these commands and previously supported `CREATE TABLE AS` can now 
support atomicity provided the underlying 
+connector supports this.
+
+Moreover, Apache Flink now supports TRUNCATE TABLE in batch execution mode. As 
before, the underlying connector needs 
+to implement and provide this capability
+
+And, finally, we have also added support for adding, dropping and listing 
partitions via
+
+- `ALTER TABLE ADD PARTITION`
+- `ALTER TABLE DROP PARTITION`
+- `SHOW PARTITIONS`
+
+**More Information**
+- [Documentation on 
TRUNCATE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/truncate/)
+- [Documentation on CREATE OR 
REPLACE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/create/#create-or-replace-table)
+- [Documentation on ALTER 
TABLE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/alter/#alter-table)
+- [FLIP-302: Support TRUNCATE TABLE statement in batch 
mode](https://cwiki.apache.org/confluence/display/FLINK/FLIP-302%3A+Support+TRUNCATE+TABLE+statement+in+batch+mode)
+- [FLIP-303: Support REPLACE TABLE AS SELECT 
statement](https://cwiki.apache.org/confluence/display/FLINK/FLIP-303%3A+Support+REPLACE+TABLE+AS+SELECT+statement)
+- [FLIP-305: Support atomic for CREATE TABLE AS SELECT(CTAS) 
statement](https://cwiki.apache.org/confluence/display/FLINK

Re: [PR] add announcement blog post for Flink 1.18 [flink-web]

2023-10-06 Thread via GitHub


snuyanzin commented on code in PR #680:
URL: https://github.com/apache/flink-web/pull/680#discussion_r1348755495


##
docs/content/posts/2023-10-10-release-1.18.0.md:
##
@@ -0,0 +1,542 @@
+---
+authors:
+- JingGe:
+  name: "Jing Ge"
+  twitter: jingengineer
+- KonstantinKnauf:
+  name: "Konstantin Knauf"
+  twitter: snntrable
+- SergeyNuyanzin:
+  name: "Sergey Nuyanzin"
+  twitter: uckamello
+- QingshengRen:
+  name: "Qingsheng Ren"
+  twitter: renqstuite
+date: "2023-10-10T08:00:00Z"
+subtitle: ""
+title: Announcing the Release of Apache Flink 1.18
+aliases:
+- /news/2023/10/10/release-1.18.0.html
+---
+
+The Apache Flink PMC is pleased to announce the release of Apache Flink 
1.18.0. As usual, we are looking at a packed 
+release with a wide variety of improvements and new features. Overall, 176 
people contributed to this release completing 
+18 FLIPS and 700+ issues. Thank you!
+
+Let's dive into the highlights.
+
+# Towards a Streaming Lakehouse
+
+## Flink SQL Improvements
+
+### Introduce Flink JDBC Driver For Sql Gateway 
+
+Flink 1.18 comes with a JDBC Driver for the Flink SQL Gateway. So, you can now 
use any SQL Client that supports JDBC to 
+interact with your tables via Flink SQL. Here is an example using 
[SQLLine](https://julianhyde.github.io/sqlline/manual.html). 

Review Comment:
   based on license it should be ok
   this tool is used by  Apache Calcite, Apache Drill, APache Phoenix and other 
projects
   There is no new features and that's the reason it was not updated



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] add announcement blog post for Flink 1.18 [flink-web]

2023-10-06 Thread via GitHub


snuyanzin commented on code in PR #680:
URL: https://github.com/apache/flink-web/pull/680#discussion_r1348759823


##
docs/content/posts/2023-10-10-release-1.18.0.md:
##
@@ -0,0 +1,542 @@
+---
+authors:
+- JingGe:
+  name: "Jing Ge"
+  twitter: jingengineer
+- KonstantinKnauf:
+  name: "Konstantin Knauf"
+  twitter: snntrable
+- SergeyNuyanzin:
+  name: "Sergey Nuyanzin"
+  twitter: uckamello
+- QingshengRen:
+  name: "Qingsheng Ren"
+  twitter: renqstuite
+date: "2023-10-10T08:00:00Z"
+subtitle: ""
+title: Announcing the Release of Apache Flink 1.18
+aliases:
+- /news/2023/10/10/release-1.18.0.html
+---
+
+The Apache Flink PMC is pleased to announce the release of Apache Flink 
1.18.0. As usual, we are looking at a packed 
+release with a wide variety of improvements and new features. Overall, 176 
people contributed to this release completing 
+18 FLIPS and 700+ issues. Thank you!
+
+Let's dive into the highlights.
+
+# Towards a Streaming Lakehouse
+
+## Flink SQL Improvements
+
+### Introduce Flink JDBC Driver For Sql Gateway 
+
+Flink 1.18 comes with a JDBC Driver for the Flink SQL Gateway. So, you can now 
use any SQL Client that supports JDBC to 
+interact with your tables via Flink SQL. Here is an example using 
[SQLLine](https://julianhyde.github.io/sqlline/manual.html). 
+
+```shell
+sqlline> !connect jdbc:flink://localhost:8083
+```
+
+```shell
+sqlline version 1.12.0
+sqlline> !connect jdbc:flink://localhost:8083
+Enter username for jdbc:flink://localhost:8083:
+Enter password for jdbc:flink://localhost:8083:
+0: jdbc:flink://localhost:8083> CREATE TABLE T(
+. . . . . . . . . . . . . . .)>  a INT,
+. . . . . . . . . . . . . . .)>  b VARCHAR(10)
+. . . . . . . . . . . . . . .)>  ) WITH (
+. . . . . . . . . . . . . . .)>  'connector' = 'filesystem',
+. . . . . . . . . . . . . . .)>  'path' = 'file:///tmp/T.csv',
+. . . . . . . . . . . . . . .)>  'format' = 'csv'
+. . . . . . . . . . . . . . .)>  );
+No rows affected (0.122 seconds)
+0: jdbc:flink://localhost:8083> INSERT INTO T VALUES (1, 'Hi'), (2, 'Hello');
++--+
+|  job id  |
++--+
+| fbade1ab4450fc57ebd5269fdf60dcfd |
++--+
+1 row selected (1.282 seconds)
+0: jdbc:flink://localhost:8083> SELECT * FROM T;
++---+---+
+| a |   b   |
++---+---+
+| 1 | Hi|
+| 2 | Hello |
++---+---+
+2 rows selected (1.955 seconds)
+0: jdbc:flink://localhost:8083>
+```
+
+**More Information**
+* 
[Documentation](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/jdbcdriver/)
 
+* [FLIP-293: Introduce Flink Jdbc Driver For Sql 
Gateway](https://cwiki.apache.org/confluence/display/FLINK/FLIP-293%3A+Introduce+Flink+Jdbc+Driver+For+Sql+Gateway)
+
+
+### Stored Procedures
+
+Stored Procedures provide a convenient way to encapsulate complex logic to 
perform data manipulation or administrative 
+tasks in Apache Flink itself. Therefore, Flink introduces the support for 
calling stored procedures. 
+Flink now allows catalog developers to develop their own built-in stored 
procedures and then enables users to call these
+predefined stored procedures.
+
+**More Information**
+* 
[Documentation](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/procedures/)
+* [FLIP-311: Support Call Stored 
Procedure](https://cwiki.apache.org/confluence/display/FLINK/FLIP-311%3A+Support+Call+Stored+Procedure)
+
+### Extended DDL Support
+
+From this release onwards, Flink supports
+
+- `REPLACE TABLE AS SELECT`
+- `CREATE OR REPLACE TABLE AS SELECT`
+
+and both these commands and previously supported `CREATE TABLE AS` can now 
support atomicity provided the underlying 
+connector supports this.
+
+Moreover, Apache Flink now supports TRUNCATE TABLE in batch execution mode. As 
before, the underlying connector needs 
+to implement and provide this capability
+
+And, finally, we have also added support for adding, dropping and listing 
partitions via
+
+- `ALTER TABLE ADD PARTITION`
+- `ALTER TABLE DROP PARTITION`
+- `SHOW PARTITIONS`
+
+**More Information**
+- [Documentation on 
TRUNCATE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/truncate/)
+- [Documentation on CREATE OR 
REPLACE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/create/#create-or-replace-table)
+- [Documentation on ALTER 
TABLE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/alter/#alter-table)
+- [FLIP-302: Support TRUNCATE TABLE statement in batch 
mode](https://cwiki.apache.org/confluence/display/FLINK/FLIP-302%3A+Support+TRUNCATE+TABLE+statement+in+batch+mode)
+- [FLIP-303: Support REPLACE TABLE AS SELECT 
statement](https://cwiki.apache.org/confluence/display/FLINK/FLIP-303%3A+Support+REPLACE+TABLE+AS+SELECT+statement)
+- [FLIP-305: Support atomic for CREATE TABLE AS SELECT(CTAS) 
statement](https://cwiki.apache.org/confluence/display/FLINK

Re: [PR] add announcement blog post for Flink 1.18 [flink-web]

2023-10-06 Thread via GitHub


snuyanzin commented on code in PR #680:
URL: https://github.com/apache/flink-web/pull/680#discussion_r1348759823


##
docs/content/posts/2023-10-10-release-1.18.0.md:
##
@@ -0,0 +1,542 @@
+---
+authors:
+- JingGe:
+  name: "Jing Ge"
+  twitter: jingengineer
+- KonstantinKnauf:
+  name: "Konstantin Knauf"
+  twitter: snntrable
+- SergeyNuyanzin:
+  name: "Sergey Nuyanzin"
+  twitter: uckamello
+- QingshengRen:
+  name: "Qingsheng Ren"
+  twitter: renqstuite
+date: "2023-10-10T08:00:00Z"
+subtitle: ""
+title: Announcing the Release of Apache Flink 1.18
+aliases:
+- /news/2023/10/10/release-1.18.0.html
+---
+
+The Apache Flink PMC is pleased to announce the release of Apache Flink 
1.18.0. As usual, we are looking at a packed 
+release with a wide variety of improvements and new features. Overall, 176 
people contributed to this release completing 
+18 FLIPS and 700+ issues. Thank you!
+
+Let's dive into the highlights.
+
+# Towards a Streaming Lakehouse
+
+## Flink SQL Improvements
+
+### Introduce Flink JDBC Driver For Sql Gateway 
+
+Flink 1.18 comes with a JDBC Driver for the Flink SQL Gateway. So, you can now 
use any SQL Client that supports JDBC to 
+interact with your tables via Flink SQL. Here is an example using 
[SQLLine](https://julianhyde.github.io/sqlline/manual.html). 
+
+```shell
+sqlline> !connect jdbc:flink://localhost:8083
+```
+
+```shell
+sqlline version 1.12.0
+sqlline> !connect jdbc:flink://localhost:8083
+Enter username for jdbc:flink://localhost:8083:
+Enter password for jdbc:flink://localhost:8083:
+0: jdbc:flink://localhost:8083> CREATE TABLE T(
+. . . . . . . . . . . . . . .)>  a INT,
+. . . . . . . . . . . . . . .)>  b VARCHAR(10)
+. . . . . . . . . . . . . . .)>  ) WITH (
+. . . . . . . . . . . . . . .)>  'connector' = 'filesystem',
+. . . . . . . . . . . . . . .)>  'path' = 'file:///tmp/T.csv',
+. . . . . . . . . . . . . . .)>  'format' = 'csv'
+. . . . . . . . . . . . . . .)>  );
+No rows affected (0.122 seconds)
+0: jdbc:flink://localhost:8083> INSERT INTO T VALUES (1, 'Hi'), (2, 'Hello');
++--+
+|  job id  |
++--+
+| fbade1ab4450fc57ebd5269fdf60dcfd |
++--+
+1 row selected (1.282 seconds)
+0: jdbc:flink://localhost:8083> SELECT * FROM T;
++---+---+
+| a |   b   |
++---+---+
+| 1 | Hi|
+| 2 | Hello |
++---+---+
+2 rows selected (1.955 seconds)
+0: jdbc:flink://localhost:8083>
+```
+
+**More Information**
+* 
[Documentation](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/jdbcdriver/)
 
+* [FLIP-293: Introduce Flink Jdbc Driver For Sql 
Gateway](https://cwiki.apache.org/confluence/display/FLINK/FLIP-293%3A+Introduce+Flink+Jdbc+Driver+For+Sql+Gateway)
+
+
+### Stored Procedures
+
+Stored Procedures provide a convenient way to encapsulate complex logic to 
perform data manipulation or administrative 
+tasks in Apache Flink itself. Therefore, Flink introduces the support for 
calling stored procedures. 
+Flink now allows catalog developers to develop their own built-in stored 
procedures and then enables users to call these
+predefined stored procedures.
+
+**More Information**
+* 
[Documentation](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/procedures/)
+* [FLIP-311: Support Call Stored 
Procedure](https://cwiki.apache.org/confluence/display/FLINK/FLIP-311%3A+Support+Call+Stored+Procedure)
+
+### Extended DDL Support
+
+From this release onwards, Flink supports
+
+- `REPLACE TABLE AS SELECT`
+- `CREATE OR REPLACE TABLE AS SELECT`
+
+and both these commands and previously supported `CREATE TABLE AS` can now 
support atomicity provided the underlying 
+connector supports this.
+
+Moreover, Apache Flink now supports TRUNCATE TABLE in batch execution mode. As 
before, the underlying connector needs 
+to implement and provide this capability
+
+And, finally, we have also added support for adding, dropping and listing 
partitions via
+
+- `ALTER TABLE ADD PARTITION`
+- `ALTER TABLE DROP PARTITION`
+- `SHOW PARTITIONS`
+
+**More Information**
+- [Documentation on 
TRUNCATE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/truncate/)
+- [Documentation on CREATE OR 
REPLACE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/create/#create-or-replace-table)
+- [Documentation on ALTER 
TABLE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/alter/#alter-table)
+- [FLIP-302: Support TRUNCATE TABLE statement in batch 
mode](https://cwiki.apache.org/confluence/display/FLINK/FLIP-302%3A+Support+TRUNCATE+TABLE+statement+in+batch+mode)
+- [FLIP-303: Support REPLACE TABLE AS SELECT 
statement](https://cwiki.apache.org/confluence/display/FLINK/FLIP-303%3A+Support+REPLACE+TABLE+AS+SELECT+statement)
+- [FLIP-305: Support atomic for CREATE TABLE AS SELECT(CTAS) 
statement](https://cwiki.apache.org/confluence/display/FLINK

[jira] [Created] (FLINK-33198) Add timestamp with local time zone support in Avro converters

2023-10-06 Thread Zhenqiu Huang (Jira)
Zhenqiu Huang created FLINK-33198:
-

 Summary: Add timestamp with local time zone support in Avro 
converters
 Key: FLINK-33198
 URL: https://issues.apache.org/jira/browse/FLINK-33198
 Project: Flink
  Issue Type: Improvement
  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
Reporter: Zhenqiu Huang
 Fix For: 1.18.1


Currently, RowDataToAvroConverters doesn't handle with LogicType 
TIMESTAMP_WITH_LOCAL_TIME_ZONE. We should add the corresponding conversion.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33182) Allow metadata columns in NduAnalyzer with ChangelogNormalize

2023-10-06 Thread lincoln lee (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33182?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17772592#comment-17772592
 ] 

lincoln lee commented on FLINK-33182:
-

[~twalthr] I've thought about this carefully, and for upsert sources, while the 
non-determinism of the metadata column doesn't affect the +U message, if there 
is a -D message, there is no guarantee of correctness if we relax the strict 
restriction on the metadata (since the metadata column in the -D message is 
carried by the message itself, not as the changelognormalize node is handled). 
So, if we want to relax the restriction, we need to additionally make sure that 
the source does not contain -D message, WDYT?

> Allow metadata columns in NduAnalyzer with ChangelogNormalize
> -
>
> Key: FLINK-33182
> URL: https://issues.apache.org/jira/browse/FLINK-33182
> Project: Flink
>  Issue Type: Sub-task
>  Components: Table SQL / Planner
>Reporter: Timo Walther
>Priority: Major
>
> Currently, the NduAnalyzer is very strict about metadata columns in updating 
> sources. However, for upsert sources (like Kafka) that contain an incomplete 
> changelog, the planner always adds a ChangelogNormalize node. 
> ChangelogNormalize will make sure that metadata columns can be considered 
> deterministic. So the NduAnalyzer should be satisfied in this case. 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-33198) Add timestamp with local time zone support in Avro converters

2023-10-06 Thread Gyula Fora (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33198?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gyula Fora reassigned FLINK-33198:
--

Assignee: Zhenqiu Huang

> Add timestamp with local time zone support in Avro converters
> -
>
> Key: FLINK-33198
> URL: https://issues.apache.org/jira/browse/FLINK-33198
> Project: Flink
>  Issue Type: Improvement
>  Components: Formats (JSON, Avro, Parquet, ORC, SequenceFile)
>Reporter: Zhenqiu Huang
>Assignee: Zhenqiu Huang
>Priority: Minor
> Fix For: 1.18.1
>
>
> Currently, RowDataToAvroConverters doesn't handle with LogicType 
> TIMESTAMP_WITH_LOCAL_TIME_ZONE. We should add the corresponding conversion.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33109) Watermark alignment not applied after recovery from checkpoint

2023-10-06 Thread Yordan Pavlov (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yordan Pavlov updated FLINK-33109:
--
Attachment: WatermarkTest.scala

> Watermark alignment not applied after recovery from checkpoint
> --
>
> Key: FLINK-33109
> URL: https://issues.apache.org/jira/browse/FLINK-33109
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.17.1
>Reporter: Yordan Pavlov
>Priority: Major
> Attachments: WatermarkTest.scala, image-2023-09-18-15-40-06-868.png, 
> image-2023-09-18-15-46-16-106.png
>
>
> I am observing a problem where after recovery from a checkpoint the Kafka 
> source watermarks would start to diverge not honoring the watermark alignment 
> setting I have applied.
> I have a Kafka source which reads a topic with 32 partitions. I am applying 
> the following watermark strategy:
> {code:java}
> new EventAwareWatermarkStrategy[KeyedKafkaSourceMessage]](msg => 
> msg.value.getTimestamp)
>       .withWatermarkAlignment("alignment-sources-group", 
> time.Duration.ofMillis(sourceWatermarkAlignmentBlocks)){code}
>  
> This works great up until my job needs to recover from checkpoint. Once the 
> recovery takes place, no alignment is taking place any more. This can best be 
> illustrated by looking at the watermark metrics for various operators in the 
> image:
> !image-2023-09-18-15-40-06-868.png!
>  
> You can see how the watermarks disperse after the recovery. Trying to debug 
> the problem I noticed that before the failure there would be calls in
>  
> {code:java}
> SourceCoordinator::announceCombinedWatermark() 
> {code}
> after the recovery, no calls get there, so no value for 
> {code:java}
> watermarkAlignmentParams.getMaxAllowedWatermarkDrift(){code}
> is ever read. I can manually fix the problem If I stop the job, clear all 
> state from Zookeeper and then manually start Flink providing the last 
> checkpoint with 
> {code:java}
> '–fromSavepoint'{code}
>  flag. This would cause the SourceCoordinator to be constructed properly and 
> watermark drift to be checked. Once recovery manually watermarks would again 
> converge to the allowed drift as seen in the metrics:
> !image-2023-09-18-15-46-16-106.png!
>  
> Let me know If I can be helpful by providing any more information.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33187] Don't record duplicate event if no change [flink-kubernetes-operator]

2023-10-06 Thread via GitHub


gyfora commented on code in PR #683:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/683#discussion_r1348864871


##
flink-kubernetes-operator-autoscaler/src/main/java/org/apache/flink/kubernetes/operator/autoscaler/ScalingExecutor.java:
##
@@ -94,14 +94,15 @@ public boolean scaleResource(
 var scalingEnabled = conf.get(SCALING_ENABLED);
 
 var scalingReport = scalingReport(scalingSummaries, scalingEnabled);
-eventRecorder.triggerEvent(
+eventRecorder.triggerEventByInterval(

Review Comment:
   We should only control the events by interval if `scalingEnabled == false` 
(in advisor mode). Regular scaling should always trigger the event



##
flink-kubernetes-operator-autoscaler/src/main/java/org/apache/flink/kubernetes/operator/autoscaler/config/AutoScalerOptions.java:
##
@@ -168,4 +168,10 @@ private static ConfigOptions.OptionBuilder 
autoScalerConfig(String key) {
 .defaultValues()
 .withDescription(
 "A (semicolon-separated) list of vertex ids in 
hexstring for which to disable scaling. Caution: For non-sink vertices this 
will still scale their downstream operators until 
https://issues.apache.org/jira/browse/FLINK-31215 is implemented.");
+
+public static final ConfigOption SCALING_REPORT_INTERVAL =
+autoScalerConfig("scaling.report.interval")
+.longType()

Review Comment:
   This should be a `Duration` type config



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-33109) Watermark alignment not applied after recovery from checkpoint

2023-10-06 Thread Yordan Pavlov (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33109?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17772616#comment-17772616
 ] 

Yordan Pavlov commented on FLINK-33109:
---

Hi [~fanrui] 

sorry for the delay, running this image gives me the following error, shortly 
after start


{code:java}
2023-10-06 14:54:20.725 [AsyncOperations-thread-1] INFO  
o.a.flink.streaming.runtime.tasks.AsyncCheckpointRunnable  - 
TumblingEventTimeWindows (10/32)#0 - asynchronous part of checkpoint 1 could 
not be completed.
java.util.concurrent.ExecutionException: java.lang.NoSuchMethodError: 'void 
org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle.(java.util.UUID,
 org.apache.flink.runtime.state.KeyGroupRange, long, java.util.List, 
java.util.List, org.apache.flink.runtime.state.StreamStateHandle, long)'
    at java.base/java.util.concurrent.FutureTask.report(Unknown Source)
    at java.base/java.util.concurrent.FutureTask.get(Unknown Source)
    at 
org.apache.flink.util.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:544)
    at 
org.apache.flink.streaming.api.operators.OperatorSnapshotFinalizer.(OperatorSnapshotFinalizer.java:54)
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.finalizeNonFinishedSnapshots(AsyncCheckpointRunnable.java:191)
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.run(AsyncCheckpointRunnable.java:124)
    at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown 
Source)
    at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown 
Source)
    at java.base/java.lang.Thread.run(Unknown Source)
Caused by: java.lang.NoSuchMethodError: 'void 
org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle.(java.util.UUID,
 org.apache.flink.runtime.state.KeyGroupRange, long, java.util.List, 
java.util.List, org.apache.flink.runtime.state.StreamStateHandle, long)'
    at 
org.apache.flink.contrib.streaming.state.snapshot.RocksNativeFullSnapshotStrategy$RocksDBNativeFullSnapshotOperation.get(RocksNativeFullSnapshotStrategy.java:198)
    at 
org.apache.flink.runtime.state.SnapshotStrategyRunner$1.callInternal(SnapshotStrategyRunner.java:91)
    at 
org.apache.flink.runtime.state.SnapshotStrategyRunner$1.callInternal(SnapshotStrategyRunner.java:88)
    at 
org.apache.flink.runtime.state.AsyncSnapshotCallable.call(AsyncSnapshotCallable.java:78)
    at java.base/java.util.concurrent.FutureTask.run(Unknown Source)
    at 
org.apache.flink.util.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:541)
    ... 6 common frames omitted{code}
so I couldn't test if has the problem I am facing.

Attached you can find a simple program which would exhibit the problem with the 
latest stable image 1.17.1: [^WatermarkTest.scala]
what I would observe is that different sources would converge shortly after 
start having a small watermark alignment value. However if I kill the task 
manager and restart from a checkpoint they would start to diverge as if no 
watermark alignment is present.

> Watermark alignment not applied after recovery from checkpoint
> --
>
> Key: FLINK-33109
> URL: https://issues.apache.org/jira/browse/FLINK-33109
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.17.1
>Reporter: Yordan Pavlov
>Priority: Major
> Attachments: WatermarkTest.scala, image-2023-09-18-15-40-06-868.png, 
> image-2023-09-18-15-46-16-106.png
>
>
> I am observing a problem where after recovery from a checkpoint the Kafka 
> source watermarks would start to diverge not honoring the watermark alignment 
> setting I have applied.
> I have a Kafka source which reads a topic with 32 partitions. I am applying 
> the following watermark strategy:
> {code:java}
> new EventAwareWatermarkStrategy[KeyedKafkaSourceMessage]](msg => 
> msg.value.getTimestamp)
>       .withWatermarkAlignment("alignment-sources-group", 
> time.Duration.ofMillis(sourceWatermarkAlignmentBlocks)){code}
>  
> This works great up until my job needs to recover from checkpoint. Once the 
> recovery takes place, no alignment is taking place any more. This can best be 
> illustrated by looking at the watermark metrics for various operators in the 
> image:
> !image-2023-09-18-15-40-06-868.png!
>  
> You can see how the watermarks disperse after the recovery. Trying to debug 
> the problem I noticed that before the failure there would be calls in
>  
> {code:java}
> SourceCoordinator::announceCombinedWatermark() 
> {code}
> after the recovery, no calls get there, so no value for 
> {code:java}
> watermarkAlignmentParams.getMaxAllowedWatermarkDrift(){code}
> is ever read. I can manually fix the problem If I stop the job, clear all 
> state from Zookeeper and then manually start Flink prov

[jira] [Updated] (FLINK-33109) Watermark alignment not applied after recovery from checkpoint

2023-10-06 Thread Yordan Pavlov (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yordan Pavlov updated FLINK-33109:
--
Attachment: WatermarkTest-1.scala

> Watermark alignment not applied after recovery from checkpoint
> --
>
> Key: FLINK-33109
> URL: https://issues.apache.org/jira/browse/FLINK-33109
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.17.1
>Reporter: Yordan Pavlov
>Priority: Major
> Attachments: WatermarkTest-1.scala, 
> image-2023-09-18-15-40-06-868.png, image-2023-09-18-15-46-16-106.png
>
>
> I am observing a problem where after recovery from a checkpoint the Kafka 
> source watermarks would start to diverge not honoring the watermark alignment 
> setting I have applied.
> I have a Kafka source which reads a topic with 32 partitions. I am applying 
> the following watermark strategy:
> {code:java}
> new EventAwareWatermarkStrategy[KeyedKafkaSourceMessage]](msg => 
> msg.value.getTimestamp)
>       .withWatermarkAlignment("alignment-sources-group", 
> time.Duration.ofMillis(sourceWatermarkAlignmentBlocks)){code}
>  
> This works great up until my job needs to recover from checkpoint. Once the 
> recovery takes place, no alignment is taking place any more. This can best be 
> illustrated by looking at the watermark metrics for various operators in the 
> image:
> !image-2023-09-18-15-40-06-868.png!
>  
> You can see how the watermarks disperse after the recovery. Trying to debug 
> the problem I noticed that before the failure there would be calls in
>  
> {code:java}
> SourceCoordinator::announceCombinedWatermark() 
> {code}
> after the recovery, no calls get there, so no value for 
> {code:java}
> watermarkAlignmentParams.getMaxAllowedWatermarkDrift(){code}
> is ever read. I can manually fix the problem If I stop the job, clear all 
> state from Zookeeper and then manually start Flink providing the last 
> checkpoint with 
> {code:java}
> '–fromSavepoint'{code}
>  flag. This would cause the SourceCoordinator to be constructed properly and 
> watermark drift to be checked. Once recovery manually watermarks would again 
> converge to the allowed drift as seen in the metrics:
> !image-2023-09-18-15-46-16-106.png!
>  
> Let me know If I can be helpful by providing any more information.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33109) Watermark alignment not applied after recovery from checkpoint

2023-10-06 Thread Yordan Pavlov (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33109?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yordan Pavlov updated FLINK-33109:
--
Attachment: (was: WatermarkTest.scala)

> Watermark alignment not applied after recovery from checkpoint
> --
>
> Key: FLINK-33109
> URL: https://issues.apache.org/jira/browse/FLINK-33109
> Project: Flink
>  Issue Type: Bug
>  Components: Runtime / Coordination
>Affects Versions: 1.17.1
>Reporter: Yordan Pavlov
>Priority: Major
> Attachments: WatermarkTest-1.scala, 
> image-2023-09-18-15-40-06-868.png, image-2023-09-18-15-46-16-106.png
>
>
> I am observing a problem where after recovery from a checkpoint the Kafka 
> source watermarks would start to diverge not honoring the watermark alignment 
> setting I have applied.
> I have a Kafka source which reads a topic with 32 partitions. I am applying 
> the following watermark strategy:
> {code:java}
> new EventAwareWatermarkStrategy[KeyedKafkaSourceMessage]](msg => 
> msg.value.getTimestamp)
>       .withWatermarkAlignment("alignment-sources-group", 
> time.Duration.ofMillis(sourceWatermarkAlignmentBlocks)){code}
>  
> This works great up until my job needs to recover from checkpoint. Once the 
> recovery takes place, no alignment is taking place any more. This can best be 
> illustrated by looking at the watermark metrics for various operators in the 
> image:
> !image-2023-09-18-15-40-06-868.png!
>  
> You can see how the watermarks disperse after the recovery. Trying to debug 
> the problem I noticed that before the failure there would be calls in
>  
> {code:java}
> SourceCoordinator::announceCombinedWatermark() 
> {code}
> after the recovery, no calls get there, so no value for 
> {code:java}
> watermarkAlignmentParams.getMaxAllowedWatermarkDrift(){code}
> is ever read. I can manually fix the problem If I stop the job, clear all 
> state from Zookeeper and then manually start Flink providing the last 
> checkpoint with 
> {code:java}
> '–fromSavepoint'{code}
>  flag. This would cause the SourceCoordinator to be constructed properly and 
> watermark drift to be checked. Once recovery manually watermarks would again 
> converge to the allowed drift as seen in the metrics:
> !image-2023-09-18-15-46-16-106.png!
>  
> Let me know If I can be helpful by providing any more information.
>  



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-33109) Watermark alignment not applied after recovery from checkpoint

2023-10-06 Thread Yordan Pavlov (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33109?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17772616#comment-17772616
 ] 

Yordan Pavlov edited comment on FLINK-33109 at 10/6/23 3:44 PM:


Hi [~fanrui] 

sorry for the delay, running this image gives me the following error, shortly 
after start
{code:java}
2023-10-06 14:54:20.725 [AsyncOperations-thread-1] INFO  
o.a.flink.streaming.runtime.tasks.AsyncCheckpointRunnable  - 
TumblingEventTimeWindows (10/32)#0 - asynchronous part of checkpoint 1 could 
not be completed.
java.util.concurrent.ExecutionException: java.lang.NoSuchMethodError: 'void 
org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle.(java.util.UUID,
 org.apache.flink.runtime.state.KeyGroupRange, long, java.util.List, 
java.util.List, org.apache.flink.runtime.state.StreamStateHandle, long)'
    at java.base/java.util.concurrent.FutureTask.report(Unknown Source)
    at java.base/java.util.concurrent.FutureTask.get(Unknown Source)
    at 
org.apache.flink.util.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:544)
    at 
org.apache.flink.streaming.api.operators.OperatorSnapshotFinalizer.(OperatorSnapshotFinalizer.java:54)
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.finalizeNonFinishedSnapshots(AsyncCheckpointRunnable.java:191)
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.run(AsyncCheckpointRunnable.java:124)
    at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown 
Source)
    at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown 
Source)
    at java.base/java.lang.Thread.run(Unknown Source)
Caused by: java.lang.NoSuchMethodError: 'void 
org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle.(java.util.UUID,
 org.apache.flink.runtime.state.KeyGroupRange, long, java.util.List, 
java.util.List, org.apache.flink.runtime.state.StreamStateHandle, long)'
    at 
org.apache.flink.contrib.streaming.state.snapshot.RocksNativeFullSnapshotStrategy$RocksDBNativeFullSnapshotOperation.get(RocksNativeFullSnapshotStrategy.java:198)
    at 
org.apache.flink.runtime.state.SnapshotStrategyRunner$1.callInternal(SnapshotStrategyRunner.java:91)
    at 
org.apache.flink.runtime.state.SnapshotStrategyRunner$1.callInternal(SnapshotStrategyRunner.java:88)
    at 
org.apache.flink.runtime.state.AsyncSnapshotCallable.call(AsyncSnapshotCallable.java:78)
    at java.base/java.util.concurrent.FutureTask.run(Unknown Source)
    at 
org.apache.flink.util.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:541)
    ... 6 common frames omitted{code}
so I couldn't test if has the problem I am facing.

Attached you can find a simple program which would exhibit the problem with the 
latest stable image 1.17.1:  [^WatermarkTest-1.scala]
what I would observe is that different sources would converge shortly after 
start having a small watermark alignment value. However if I kill the task 
manager and restart from a checkpoint they would start to diverge as if no 
watermark alignment is present.


was (Author: yordanpavlov):
Hi [~fanrui] 

sorry for the delay, running this image gives me the following error, shortly 
after start


{code:java}
2023-10-06 14:54:20.725 [AsyncOperations-thread-1] INFO  
o.a.flink.streaming.runtime.tasks.AsyncCheckpointRunnable  - 
TumblingEventTimeWindows (10/32)#0 - asynchronous part of checkpoint 1 could 
not be completed.
java.util.concurrent.ExecutionException: java.lang.NoSuchMethodError: 'void 
org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle.(java.util.UUID,
 org.apache.flink.runtime.state.KeyGroupRange, long, java.util.List, 
java.util.List, org.apache.flink.runtime.state.StreamStateHandle, long)'
    at java.base/java.util.concurrent.FutureTask.report(Unknown Source)
    at java.base/java.util.concurrent.FutureTask.get(Unknown Source)
    at 
org.apache.flink.util.concurrent.FutureUtils.runIfNotDoneAndGet(FutureUtils.java:544)
    at 
org.apache.flink.streaming.api.operators.OperatorSnapshotFinalizer.(OperatorSnapshotFinalizer.java:54)
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.finalizeNonFinishedSnapshots(AsyncCheckpointRunnable.java:191)
    at 
org.apache.flink.streaming.runtime.tasks.AsyncCheckpointRunnable.run(AsyncCheckpointRunnable.java:124)
    at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown 
Source)
    at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown 
Source)
    at java.base/java.lang.Thread.run(Unknown Source)
Caused by: java.lang.NoSuchMethodError: 'void 
org.apache.flink.runtime.state.IncrementalRemoteKeyedStateHandle.(java.util.UUID,
 org.apache.flink.runtime.state.KeyGroupRange, long, java.util.List, 
java.util.List, org.apache.flink.runtime.state.StreamStateHandle, long)'
    at 
org.apache.flink.contrib.streaming.state.snapshot.RocksNativeFullSnap

[jira] [Created] (FLINK-33199) ArchitectureTests should test for canonical class names instead of Class objects

2023-10-06 Thread Alexander Fedulov (Jira)
Alexander Fedulov created FLINK-33199:
-

 Summary: ArchitectureTests should test for canonical class names 
instead of Class objects
 Key: FLINK-33199
 URL: https://issues.apache.org/jira/browse/FLINK-33199
 Project: Flink
  Issue Type: Improvement
  Components: Tests
Reporter: Alexander Fedulov


Currently architecture tests rely on importing such classes as 
MiniClusterExtension. This introduces a production scope dependency on 
flink-test-utils which in turn depends on flink-streaming-java. This is 
problematic because adding architecture tests to any direct or transitive 
dependency of flink-streaming-java creates a dependency cycle.

Example: https://github.com/apache/flink/pull/22850#discussion_r1243343382

In general, since architecture tests are supposed to be used freely in any 
submodule, it is desirable to reduce its dependency surface as much as possible 
to prevent such cycles. 

This can be achieved by moving away from using Class objects and employing 
fully qualified type names checks instead.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-33199) ArchitectureTests should test for canonical class names instead of Class objects

2023-10-06 Thread Alexander Fedulov (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33199?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Fedulov reassigned FLINK-33199:
-

Assignee: Alexander Fedulov

> ArchitectureTests should test for canonical class names instead of Class 
> objects
> 
>
> Key: FLINK-33199
> URL: https://issues.apache.org/jira/browse/FLINK-33199
> Project: Flink
>  Issue Type: Improvement
>  Components: Tests
>Reporter: Alexander Fedulov
>Assignee: Alexander Fedulov
>Priority: Major
>
> Currently architecture tests rely on importing such classes as 
> MiniClusterExtension. This introduces a production scope dependency on 
> flink-test-utils which in turn depends on flink-streaming-java. This is 
> problematic because adding architecture tests to any direct or transitive 
> dependency of flink-streaming-java creates a dependency cycle.
> Example: https://github.com/apache/flink/pull/22850#discussion_r1243343382
> In general, since architecture tests are supposed to be used freely in any 
> submodule, it is desirable to reduce its dependency surface as much as 
> possible to prevent such cycles. 
> This can be achieved by moving away from using Class objects and employing 
> fully qualified type names checks instead.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-33199) ArchitectureTests should test for fully qualified class names instead of Class objects

2023-10-06 Thread Alexander Fedulov (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-33199?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alexander Fedulov updated FLINK-33199:
--
Summary: ArchitectureTests should test for fully qualified class names 
instead of Class objects  (was: ArchitectureTests should test for canonical 
class names instead of Class objects)

> ArchitectureTests should test for fully qualified class names instead of 
> Class objects
> --
>
> Key: FLINK-33199
> URL: https://issues.apache.org/jira/browse/FLINK-33199
> Project: Flink
>  Issue Type: Improvement
>  Components: Tests
>Reporter: Alexander Fedulov
>Assignee: Alexander Fedulov
>Priority: Major
>
> Currently architecture tests rely on importing such classes as 
> MiniClusterExtension. This introduces a production scope dependency on 
> flink-test-utils which in turn depends on flink-streaming-java. This is 
> problematic because adding architecture tests to any direct or transitive 
> dependency of flink-streaming-java creates a dependency cycle.
> Example: https://github.com/apache/flink/pull/22850#discussion_r1243343382
> In general, since architecture tests are supposed to be used freely in any 
> submodule, it is desirable to reduce its dependency surface as much as 
> possible to prevent such cycles. 
> This can be achieved by moving away from using Class objects and employing 
> fully qualified type names checks instead.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-31631][FileSystems] shade guava in gs-fs filesystem [flink]

2023-10-06 Thread via GitHub


singhravidutt commented on PR #23489:
URL: https://github.com/apache/flink/pull/23489#issuecomment-1751097059

   @flinkbot run azure


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33187] Don't record duplicate event if no change [flink-kubernetes-operator]

2023-10-06 Thread via GitHub


clarax commented on code in PR #683:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/683#discussion_r1349036555


##
flink-kubernetes-operator-autoscaler/src/main/java/org/apache/flink/kubernetes/operator/autoscaler/config/AutoScalerOptions.java:
##
@@ -168,4 +168,10 @@ private static ConfigOptions.OptionBuilder 
autoScalerConfig(String key) {
 .defaultValues()
 .withDescription(
 "A (semicolon-separated) list of vertex ids in 
hexstring for which to disable scaling. Caution: For non-sink vertices this 
will still scale their downstream operators until 
https://issues.apache.org/jira/browse/FLINK-31215 is implemented.");
+
+public static final ConfigOption SCALING_REPORT_INTERVAL =
+autoScalerConfig("scaling.report.interval")
+.longType()

Review Comment:
   updated.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33187] Don't record duplicate event if no change [flink-kubernetes-operator]

2023-10-06 Thread via GitHub


clarax commented on code in PR #683:
URL: 
https://github.com/apache/flink-kubernetes-operator/pull/683#discussion_r1349038012


##
flink-kubernetes-operator-autoscaler/src/main/java/org/apache/flink/kubernetes/operator/autoscaler/ScalingExecutor.java:
##
@@ -94,14 +94,15 @@ public boolean scaleResource(
 var scalingEnabled = conf.get(SCALING_ENABLED);
 
 var scalingReport = scalingReport(scalingSummaries, scalingEnabled);
-eventRecorder.triggerEvent(
+eventRecorder.triggerEventByInterval(

Review Comment:
   updated.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-31371) Support subscribing non-existed topics in Pulsar source

2023-10-06 Thread Enzo Dechaene (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31371?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17772666#comment-17772666
 ] 

Enzo Dechaene commented on FLINK-31371:
---

Hi [~syhily], any news for this ticket ?

> Support subscribing non-existed topics in Pulsar source
> ---
>
> Key: FLINK-31371
> URL: https://issues.apache.org/jira/browse/FLINK-31371
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Pulsar
>Affects Versions: 1.15.3
>Reporter: Enzo Dechaene
>Assignee: Yufan Sheng
>Priority: Major
>
> *Describe the bug*
> With a Pulsar 2.8.4 server, a Flink stream containing Pulsar sources or sinks 
> will fail at startup if the topic doesn't exist.
>  
> *To Reproduce*
> Create a stream with :
>  * Flink 1.15.2
>  * Pulsar 2.8.4
>  * with a Pulsar source or sink linked to a non existant topic
>  * Start the stream
>  
> *Expected behavior*
> If the topic doesn't exist, it should be created at the first connection of 
> the source or sink without error.
>  
> *Additional context*
> In the TopicListSubscriber class of the connector, the method 
> getSubscribedTopicPartitions() try to get the metadata of a topic by doing 
> that :
>  
> {code:java}
> TopicMetadata metadata = queryTopicMetadata(pulsarAdmin, topic);{code}
>  
> If the topic doesn't exist, I get a NullPointerException on the metadata
> We created a previous 
> [ticket|https://github.com/streamnative/pulsar-flink/issues/366] on the 
> Pulsar connector and it was fixed



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-33011) Operator deletes HA data unexpectedly

2023-10-06 Thread Mason Chen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33011?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17772350#comment-17772350
 ] 

Mason Chen edited comment on FLINK-33011 at 10/6/23 7:06 PM:
-

[~gyfora] -could you backport this to 1.6? We are also hitting this bug in 1.6-

It has been already backported. I'll update the ticket fix versions. Thanks for 
the fix!


was (Author: mason6345):
[~gyfora] could you backport this to 1.6? We are also hitting this bug in 1.6

> Operator deletes HA data unexpectedly
> -
>
> Key: FLINK-33011
> URL: https://issues.apache.org/jira/browse/FLINK-33011
> Project: Flink
>  Issue Type: Bug
>  Components: Kubernetes Operator
>Affects Versions: 1.17.1, kubernetes-operator-1.6.0
> Environment: Flink: 1.17.1
> Flink Kubernetes Operator: 1.6.0
>Reporter: Ruibin Xing
>Assignee: Gyula Fora
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: kubernetes-operator-1.7.0
>
> Attachments: flink_operator_logs_0831.csv
>
>
> We encountered a problem where the operator unexpectedly deleted HA data.
> The timeline is as follows:
> 12:08 We submitted the first spec, which suspended the job with savepoint 
> upgrade mode.
> 12:08 The job was suspended, while the HA data was preserved, and the log 
> showed the observed job deployment status was MISSING.
> 12:10 We submitted the second spec, which deployed the job with the last 
> state upgrade mode.
> 12:10 Logs showed the operator deleted both the Flink deployment and the HA 
> data again.
> 12:10 The job failed to start because the HA data was missing.
> According to the log, the deletion was triggered by 
> https://github.com/apache/flink-kubernetes-operator/blob/a728ba768e20236184e2b9e9e45163304b8b196c/flink-kubernetes-operator/src/main/java/org/apache/flink/kubernetes/operator/reconciler/deployment/ApplicationReconciler.java#L168
> I think this would only be triggered if the job deployment status wasn't 
> MISSING. But the log before the deletion showed the observed job status was 
> MISSING at that moment.
> Related logs:
>  
> {code:java}
> 2023-08-30 12:08:48.190 + o.a.f.k.o.s.AbstractFlinkService [INFO 
> ][default/pipeline-pipeline-se-3] Cluster shutdown completed.
> 2023-08-30 12:10:27.010 + o.a.f.k.o.o.d.ApplicationObserver [INFO 
> ][default/pipeline-pipeline-se-3] Observing JobManager deployment. Previous 
> status: MISSING
> 2023-08-30 12:10:27.533 + o.a.f.k.o.l.AuditUtils         [INFO 
> ][default/pipeline-pipeline-se-3] >>> Event  | Info    | SPECCHANGED     | 
> UPGRADE change(s) detected (Diff: FlinkDeploymentSpec[image : 
> docker-registry.randomcompany.com/octopus/pipeline-pipeline-online:0835137c-362
>  -> 
> docker-registry.randomcompany.com/octopus/pipeline-pipeline-online:23db7ae8-365,
>  podTemplate.metadata.labels.app.kubernetes.io~1version : 
> 0835137cd803b7258695eb53a6ec520cb62a48a7 -> 
> 23db7ae84bdab8d91fa527fe2f8f2fce292d0abc, job.state : suspended -> running, 
> job.upgradeMode : last-state -> savepoint, restartNonce : 1545 -> 1547]), 
> starting reconciliation.
> 2023-08-30 12:10:27.679 + o.a.f.k.o.s.NativeFlinkService [INFO 
> ][default/pipeline-pipeline-se-3] Deleting JobManager deployment and HA 
> metadata.
> {code}
> A more complete log file is attached. Thanks.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-33011) Operator deletes HA data unexpectedly

2023-10-06 Thread Mason Chen (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33011?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17772350#comment-17772350
 ] 

Mason Chen edited comment on FLINK-33011 at 10/6/23 7:07 PM:
-

[~gyfora] -could you backport this to 1.6? We are also hitting this bug in 1.6-

It has been already backported. Would be good to update the ticket fix version. 
Thanks for the fix!


was (Author: mason6345):
[~gyfora] -could you backport this to 1.6? We are also hitting this bug in 1.6-

It has been already backported. I'll update the ticket fix versions. Thanks for 
the fix!

> Operator deletes HA data unexpectedly
> -
>
> Key: FLINK-33011
> URL: https://issues.apache.org/jira/browse/FLINK-33011
> Project: Flink
>  Issue Type: Bug
>  Components: Kubernetes Operator
>Affects Versions: 1.17.1, kubernetes-operator-1.6.0
> Environment: Flink: 1.17.1
> Flink Kubernetes Operator: 1.6.0
>Reporter: Ruibin Xing
>Assignee: Gyula Fora
>Priority: Blocker
>  Labels: pull-request-available
> Fix For: kubernetes-operator-1.7.0
>
> Attachments: flink_operator_logs_0831.csv
>
>
> We encountered a problem where the operator unexpectedly deleted HA data.
> The timeline is as follows:
> 12:08 We submitted the first spec, which suspended the job with savepoint 
> upgrade mode.
> 12:08 The job was suspended, while the HA data was preserved, and the log 
> showed the observed job deployment status was MISSING.
> 12:10 We submitted the second spec, which deployed the job with the last 
> state upgrade mode.
> 12:10 Logs showed the operator deleted both the Flink deployment and the HA 
> data again.
> 12:10 The job failed to start because the HA data was missing.
> According to the log, the deletion was triggered by 
> https://github.com/apache/flink-kubernetes-operator/blob/a728ba768e20236184e2b9e9e45163304b8b196c/flink-kubernetes-operator/src/main/java/org/apache/flink/kubernetes/operator/reconciler/deployment/ApplicationReconciler.java#L168
> I think this would only be triggered if the job deployment status wasn't 
> MISSING. But the log before the deletion showed the observed job status was 
> MISSING at that moment.
> Related logs:
>  
> {code:java}
> 2023-08-30 12:08:48.190 + o.a.f.k.o.s.AbstractFlinkService [INFO 
> ][default/pipeline-pipeline-se-3] Cluster shutdown completed.
> 2023-08-30 12:10:27.010 + o.a.f.k.o.o.d.ApplicationObserver [INFO 
> ][default/pipeline-pipeline-se-3] Observing JobManager deployment. Previous 
> status: MISSING
> 2023-08-30 12:10:27.533 + o.a.f.k.o.l.AuditUtils         [INFO 
> ][default/pipeline-pipeline-se-3] >>> Event  | Info    | SPECCHANGED     | 
> UPGRADE change(s) detected (Diff: FlinkDeploymentSpec[image : 
> docker-registry.randomcompany.com/octopus/pipeline-pipeline-online:0835137c-362
>  -> 
> docker-registry.randomcompany.com/octopus/pipeline-pipeline-online:23db7ae8-365,
>  podTemplate.metadata.labels.app.kubernetes.io~1version : 
> 0835137cd803b7258695eb53a6ec520cb62a48a7 -> 
> 23db7ae84bdab8d91fa527fe2f8f2fce292d0abc, job.state : suspended -> running, 
> job.upgradeMode : last-state -> savepoint, restartNonce : 1545 -> 1547]), 
> starting reconciliation.
> 2023-08-30 12:10:27.679 + o.a.f.k.o.s.NativeFlinkService [INFO 
> ][default/pipeline-pipeline-se-3] Deleting JobManager deployment and HA 
> metadata.
> {code}
> A more complete log file is attached. Thanks.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33175) Nightly builds from S3 are not available for download, breaking all connector tests

2023-10-06 Thread Jing Ge (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17772680#comment-17772680
 ] 

Jing Ge commented on FLINK-33175:
-

[~mapohl] the file is ok and not public accessible: 
[https://apache-flink.slack.com/archives/C051F2AFNJW/p1695206313618819?thread_ts=1695122850.132129&cid=C051F2AFNJW]

I just followed the release guideline and didn't change anything else. Do you 
know why the file can be downloaded from S3 previously? How is the credential 
configured?

> Nightly builds from S3 are not available for download, breaking all connector 
> tests
> ---
>
> Key: FLINK-33175
> URL: https://issues.apache.org/jira/browse/FLINK-33175
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Common
>Reporter: Martijn Visser
>Priority: Blocker
>
> All downloads of Flink binaries fail with:
> {code:java}
> Run wget -q -c 
> https://s3.amazonaws.com/flink-nightly/flink-1.18-SNAPSHOT-bin-scala_2.12.tgz 
> -O - | tar -xz
> gzip: stdin: unexpected end of file
> tar: Child returned status 1
> tar: Error is not recoverable: exiting now
> Error: Process completed with exit code 2.
> {code}
> This goes for 1.18, but also 1.17 and 1.16



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Comment Edited] (FLINK-33175) Nightly builds from S3 are not available for download, breaking all connector tests

2023-10-06 Thread Jing Ge (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33175?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17772680#comment-17772680
 ] 

Jing Ge edited comment on FLINK-33175 at 10/6/23 8:18 PM:
--

[~mapohl] the file is ok and not public accessible: 
[https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=53296&view=l[…]-51bc-8163-03e54ba9ce5b&t=2f31de30-3dd6-5bd6-d86d-dcdc2d7cc3de|https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=53296&view=logs&j=585d8b77-fa33-51bc-8163-03e54ba9ce5b&t=2f31de30-3dd6-5bd6-d86d-dcdc2d7cc3de]

I just followed the release guideline and didn't change anything else. Do you 
know why the file can be downloaded from S3 previously? How is the credential 
configured?


was (Author: jingge):
[~mapohl] the file is ok and not public accessible: 
[https://apache-flink.slack.com/archives/C051F2AFNJW/p1695206313618819?thread_ts=1695122850.132129&cid=C051F2AFNJW]

I just followed the release guideline and didn't change anything else. Do you 
know why the file can be downloaded from S3 previously? How is the credential 
configured?

> Nightly builds from S3 are not available for download, breaking all connector 
> tests
> ---
>
> Key: FLINK-33175
> URL: https://issues.apache.org/jira/browse/FLINK-33175
> Project: Flink
>  Issue Type: Bug
>  Components: Connectors / Common
>Reporter: Martijn Visser
>Priority: Blocker
>
> All downloads of Flink binaries fail with:
> {code:java}
> Run wget -q -c 
> https://s3.amazonaws.com/flink-nightly/flink-1.18-SNAPSHOT-bin-scala_2.12.tgz 
> -O - | tar -xz
> gzip: stdin: unexpected end of file
> tar: Child returned status 1
> tar: Error is not recoverable: exiting now
> Error: Process completed with exit code 2.
> {code}
> This goes for 1.18, but also 1.17 and 1.16



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32732) auto offset reset should be exposed to user

2023-10-06 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32732?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32732:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> auto offset reset should be exposed to user
> ---
>
> Key: FLINK-32732
> URL: https://issues.apache.org/jira/browse/FLINK-32732
> Project: Flink
>  Issue Type: Improvement
>  Components: Connectors / Kafka
>Affects Versions: 1.16.1
>Reporter: xiaogang zhou
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> {code:java}
> // code placeholder
> maybeOverride(
> ConsumerConfig.AUTO_OFFSET_RESET_CONFIG,
> 
> startingOffsetsInitializer.getAutoOffsetResetStrategy().name().toLowerCase(),
> true); {code}
> now flink override the auto.offset.reset with the scan.startup.mode config, 
> and user's explicit config does not take effect. I think maybe we should 
> expose this to customer?
>  
> I think after consuming kafka records from earliest to latest, the 
> scan.startup.mode should no longer influence the kafka scan behave. So I 
> suggest change the override to false.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-31615) Fix some parts forgot to translate in "Table API" page of "Table API & SQL"

2023-10-06 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31615?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-31615:
---
  Labels: auto-deprioritized-minor chinese-translation  (was: 
chinese-translation stale-minor)
Priority: Not a Priority  (was: Minor)

This issue was labeled "stale-minor" 7 days ago and has not received any 
updates so it is being deprioritized. If this ticket is actually Minor, please 
raise the priority and ask a committer to assign you the issue or revive the 
public discussion.


> Fix some parts forgot to translate in "Table API" page of "Table API & SQL" 
> 
>
> Key: FLINK-31615
> URL: https://issues.apache.org/jira/browse/FLINK-31615
> Project: Flink
>  Issue Type: Improvement
>  Components: Documentation
>Reporter: Hang Ruan
>Priority: Not a Priority
>  Labels: auto-deprioritized-minor, chinese-translation
>
> The query_state_warning in "Table API" page of "Table API & SQL"  is still in 
> English. And some comments in codes are in English.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Updated] (FLINK-32293) Support vector with long index

2023-10-06 Thread Flink Jira Bot (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-32293?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Flink Jira Bot updated FLINK-32293:
---
Labels: pull-request-available stale-major  (was: pull-request-available)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help 
the community manage its development. I see this issues has been marked as 
Major but is unassigned and neither itself nor its Sub-Tasks have been updated 
for 60 days. I have gone ahead and added a "stale-major" to the issue". If this 
ticket is a Major, please either assign yourself or give an update. Afterwards, 
please remove the label or in 7 days the issue will be deprioritized.


> Support vector with long index
> --
>
> Key: FLINK-32293
> URL: https://issues.apache.org/jira/browse/FLINK-32293
> Project: Flink
>  Issue Type: New Feature
>  Components: Library / Machine Learning
>Reporter: Zhipeng Zhang
>Priority: Major
>  Labels: pull-request-available, stale-major
>
> Currently in Flink ML, we only support sparse and dense vector with `int` as 
> index and `double` as value.
>  
> However, there are real-world cases that the index of a vector could exceed 
> the range of `INT.MAX`. Thus we need to support vector with `long` index.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-33200) ItemAt Expression validation fail in Table API due to type mismatch

2023-10-06 Thread Zhenqiu Huang (Jira)
Zhenqiu Huang created FLINK-33200:
-

 Summary: ItemAt Expression validation fail in Table API due to 
type mismatch
 Key: FLINK-33200
 URL: https://issues.apache.org/jira/browse/FLINK-33200
 Project: Flink
  Issue Type: Bug
  Components: Table SQL / API
Reporter: Zhenqiu Huang
 Fix For: 1.8.4


The table schema is defined as below:

public static final DataType DATA_TYPE = DataTypes.ROW(
DataTypes.FIELD("id", DataTypes.STRING()),
DataTypes.FIELD("events", 
DataTypes.ARRAY(DataTypes.MAP(DataTypes.STRING(), DataTypes.STRING(
);

public static final Schema SCHEMA = 
Schema.newBuilder().fromRowDataType(DATA_TYPE).build();


inTable.select(Expressions.$("events").at(1).at("eventType").as("firstEventType")

The validation fail as "eventType" is inferred as 
BasicTypeInfo.STRING_TYPE_INFO, the table key internally is a 
StringDataTypeInfo. The validation fail at 

case mti: MapTypeInfo[_, _] =>
if (key.resultType == mti.getKeyTypeInfo) {
  ValidationSuccess
} else {
  ValidationFailure(
s"Map entry access needs a valid key of type " +
  s"'${mti.getKeyTypeInfo}', found '${key.resultType}'.")
}








--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Commented] (FLINK-33200) ItemAt Expression validation fail in Table API due to type mismatch

2023-10-06 Thread Zhenqiu Huang (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-33200?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17772723#comment-17772723
 ] 

Zhenqiu Huang commented on FLINK-33200:
---

[~lzljs3620320]

As StringDataTypeInfo is the default key type in Flink table, shall we loose 
the validation to 
 
if (key.resultType == StringDataTypeInfo || key.resultType == 
BasicTypeInfo.STRING_TYPE_INFO)

> ItemAt Expression validation fail in Table API due to type mismatch
> ---
>
> Key: FLINK-33200
> URL: https://issues.apache.org/jira/browse/FLINK-33200
> Project: Flink
>  Issue Type: Bug
>  Components: Table SQL / API
>Reporter: Zhenqiu Huang
>Priority: Minor
> Fix For: 1.8.4
>
>
> The table schema is defined as below:
> public static final DataType DATA_TYPE = DataTypes.ROW(
> DataTypes.FIELD("id", DataTypes.STRING()),
> DataTypes.FIELD("events", 
> DataTypes.ARRAY(DataTypes.MAP(DataTypes.STRING(), DataTypes.STRING(
> );
> public static final Schema SCHEMA = 
> Schema.newBuilder().fromRowDataType(DATA_TYPE).build();
> inTable.select(Expressions.$("events").at(1).at("eventType").as("firstEventType")
> The validation fail as "eventType" is inferred as 
> BasicTypeInfo.STRING_TYPE_INFO, the table key internally is a 
> StringDataTypeInfo. The validation fail at 
> case mti: MapTypeInfo[_, _] =>
> if (key.resultType == mti.getKeyTypeInfo) {
>   ValidationSuccess
> } else {
>   ValidationFailure(
> s"Map entry access needs a valid key of type " +
>   s"'${mti.getKeyTypeInfo}', found '${key.resultType}'.")
> }



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] add announcement blog post for Flink 1.18 [flink-web]

2023-10-06 Thread via GitHub


lsyldliu commented on code in PR #680:
URL: https://github.com/apache/flink-web/pull/680#discussion_r1349437120


##
docs/content/posts/2023-10-10-release-1.18.0.md:
##
@@ -0,0 +1,542 @@
+---
+authors:
+- JingGe:
+  name: "Jing Ge"
+  twitter: jingengineer
+- KonstantinKnauf:
+  name: "Konstantin Knauf"
+  twitter: snntrable
+- SergeyNuyanzin:
+  name: "Sergey Nuyanzin"
+  twitter: uckamello
+- QingshengRen:
+  name: "Qingsheng Ren"
+  twitter: renqstuite
+date: "2023-10-10T08:00:00Z"
+subtitle: ""
+title: Announcing the Release of Apache Flink 1.18
+aliases:
+- /news/2023/10/10/release-1.18.0.html
+---
+
+The Apache Flink PMC is pleased to announce the release of Apache Flink 
1.18.0. As usual, we are looking at a packed 
+release with a wide variety of improvements and new features. Overall, 176 
people contributed to this release completing 
+18 FLIPS and 700+ issues. Thank you!
+
+Let's dive into the highlights.
+
+# Towards a Streaming Lakehouse
+
+## Flink SQL Improvements
+
+### Introduce Flink JDBC Driver For Sql Gateway 
+
+Flink 1.18 comes with a JDBC Driver for the Flink SQL Gateway. So, you can now 
use any SQL Client that supports JDBC to 
+interact with your tables via Flink SQL. Here is an example using 
[SQLLine](https://julianhyde.github.io/sqlline/manual.html). 
+
+```shell
+sqlline> !connect jdbc:flink://localhost:8083
+```
+
+```shell
+sqlline version 1.12.0
+sqlline> !connect jdbc:flink://localhost:8083
+Enter username for jdbc:flink://localhost:8083:
+Enter password for jdbc:flink://localhost:8083:
+0: jdbc:flink://localhost:8083> CREATE TABLE T(
+. . . . . . . . . . . . . . .)>  a INT,
+. . . . . . . . . . . . . . .)>  b VARCHAR(10)
+. . . . . . . . . . . . . . .)>  ) WITH (
+. . . . . . . . . . . . . . .)>  'connector' = 'filesystem',
+. . . . . . . . . . . . . . .)>  'path' = 'file:///tmp/T.csv',
+. . . . . . . . . . . . . . .)>  'format' = 'csv'
+. . . . . . . . . . . . . . .)>  );
+No rows affected (0.122 seconds)
+0: jdbc:flink://localhost:8083> INSERT INTO T VALUES (1, 'Hi'), (2, 'Hello');
++--+
+|  job id  |
++--+
+| fbade1ab4450fc57ebd5269fdf60dcfd |
++--+
+1 row selected (1.282 seconds)
+0: jdbc:flink://localhost:8083> SELECT * FROM T;
++---+---+
+| a |   b   |
++---+---+
+| 1 | Hi|
+| 2 | Hello |
++---+---+
+2 rows selected (1.955 seconds)
+0: jdbc:flink://localhost:8083>
+```
+
+**More Information**
+* 
[Documentation](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/jdbcdriver/)
 
+* [FLIP-293: Introduce Flink Jdbc Driver For Sql 
Gateway](https://cwiki.apache.org/confluence/display/FLINK/FLIP-293%3A+Introduce+Flink+Jdbc+Driver+For+Sql+Gateway)
+
+
+### Stored Procedures
+
+Stored Procedures provide a convenient way to encapsulate complex logic to 
perform data manipulation or administrative 
+tasks in Apache Flink itself. Therefore, Flink introduces the support for 
calling stored procedures. 
+Flink now allows catalog developers to develop their own built-in stored 
procedures and then enables users to call these
+predefined stored procedures.
+
+**More Information**
+* 
[Documentation](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/procedures/)
+* [FLIP-311: Support Call Stored 
Procedure](https://cwiki.apache.org/confluence/display/FLINK/FLIP-311%3A+Support+Call+Stored+Procedure)
+
+### Extended DDL Support
+
+From this release onwards, Flink supports
+
+- `REPLACE TABLE AS SELECT`
+- `CREATE OR REPLACE TABLE AS SELECT`
+
+and both these commands and previously supported `CREATE TABLE AS` can now 
support atomicity provided the underlying 
+connector supports this.
+
+Moreover, Apache Flink now supports TRUNCATE TABLE in batch execution mode. As 
before, the underlying connector needs 
+to implement and provide this capability
+
+And, finally, we have also added support for adding, dropping and listing 
partitions via
+
+- `ALTER TABLE ADD PARTITION`
+- `ALTER TABLE DROP PARTITION`
+- `SHOW PARTITIONS`
+
+**More Information**
+- [Documentation on 
TRUNCATE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/truncate/)
+- [Documentation on CREATE OR 
REPLACE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/create/#create-or-replace-table)
+- [Documentation on ALTER 
TABLE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/alter/#alter-table)
+- [FLIP-302: Support TRUNCATE TABLE statement in batch 
mode](https://cwiki.apache.org/confluence/display/FLINK/FLIP-302%3A+Support+TRUNCATE+TABLE+statement+in+batch+mode)
+- [FLIP-303: Support REPLACE TABLE AS SELECT 
statement](https://cwiki.apache.org/confluence/display/FLINK/FLIP-303%3A+Support+REPLACE+TABLE+AS+SELECT+statement)
+- [FLIP-305: Support atomic for CREATE TABLE AS SELECT(CTAS) 
statement](https://cwiki.apache.org/confluence/display/FLINK/

Re: [PR] [FLINK-26088][Connectors/ElasticSearch] Add Elasticsearch 8.0 support [flink-connector-elasticsearch]

2023-10-06 Thread via GitHub


rinkako commented on PR #53:
URL: 
https://github.com/apache/flink-connector-elasticsearch/pull/53#issuecomment-1751562763

   hi @mtfelisb , I read the code and found that it has no implementation for 
`DynamicTableSink`, hence this es8 sink cannot be used in flink-sql. Would you 
have any plan to work on this point? 😄


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33046][FLIP-333] Redesign Apache Flink website [flink-web]

2023-10-06 Thread via GitHub


Myasuka commented on PR #676:
URL: https://github.com/apache/flink-web/pull/676#issuecomment-1751567081

   I found a strange behavior that when I click the nav title on the right, it 
will hide the head of lines:
   
![image](https://github.com/apache/flink-web/assets/1709104/9d641aab-1db4-4115-b4f2-03d387057c95)
   What I expect:
   
![image](https://github.com/apache/flink-web/assets/1709104/33aa64c4-a9a3-46ce-84f2-f31476ff23f3)
   
   The problematic link 
https://website-refresh.d193kg429zpv7e.amplifyapp.com/what-is-flink/flink-applications/#state
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33053][zookeeper] Manually remove the leader watcher after ret… [flink]

2023-10-06 Thread via GitHub


KarmaGYZ commented on code in PR #23415:
URL: https://github.com/apache/flink/pull/23415#discussion_r1349440903


##
flink-runtime/src/main/java/org/apache/flink/runtime/leaderretrieval/ZooKeeperLeaderRetrievalDriver.java:
##
@@ -122,6 +125,19 @@ public void close() throws Exception {
 
client.getConnectionStateListenable().removeListener(connectionStateListener);
 
 cache.close();
+
+try {
+if (client.getZookeeperClient().isConnected()
+&& 
!connectionInformationPath.contains(RESOURCE_MANAGER_NODE)) {

Review Comment:
   Also an exclusive curator framework will definitely harm the performance and 
add additional pressure to ZooKeeper.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33053][zookeeper] Manually remove the leader watcher after ret… [flink]

2023-10-06 Thread via GitHub


KarmaGYZ commented on code in PR #23415:
URL: https://github.com/apache/flink/pull/23415#discussion_r1349440740


##
flink-runtime/src/main/java/org/apache/flink/runtime/leaderretrieval/ZooKeeperLeaderRetrievalDriver.java:
##
@@ -122,6 +125,19 @@ public void close() throws Exception {
 
client.getConnectionStateListenable().removeListener(connectionStateListener);
 
 cache.close();
+
+try {
+if (client.getZookeeperClient().isConnected()
+&& 
!connectionInformationPath.contains(RESOURCE_MANAGER_NODE)) {

Review Comment:
   I agree that this is a rather tricky fix. As Tison mentioned, JobMaster and 
Dispatcher will use the same curator framework at the JobManager side. Close 
the watch on resource manager when JobMaster exit will influence other 
JobMasters and Dispatcher, thus I arbitrarily exclude the RM here. We can also 
construct an exclusive curator framework for each JobMaster, but I think a 
cleaner solution should be implemented within curator or zookeeper, then we 
should remove this fix. WDYT?



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33053][zookeeper] Manually remove the leader watcher after ret… [flink]

2023-10-06 Thread via GitHub


KarmaGYZ commented on code in PR #23415:
URL: https://github.com/apache/flink/pull/23415#discussion_r1349440903


##
flink-runtime/src/main/java/org/apache/flink/runtime/leaderretrieval/ZooKeeperLeaderRetrievalDriver.java:
##
@@ -122,6 +125,19 @@ public void close() throws Exception {
 
client.getConnectionStateListenable().removeListener(connectionStateListener);
 
 cache.close();
+
+try {
+if (client.getZookeeperClient().isConnected()
+&& 
!connectionInformationPath.contains(RESOURCE_MANAGER_NODE)) {

Review Comment:
   Also an exclusive curator framework per JobMaster will definitely harm the 
performance and add additional pressure to ZooKeeper.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Commented] (FLINK-31275) Flink supports reporting and storage of source/sink tables relationship

2023-10-06 Thread Fang Yong (Jira)


[ 
https://issues.apache.org/jira/browse/FLINK-31275?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17772734#comment-17772734
 ] 

Fang Yong commented on FLINK-31275:
---

[~ZhenqiuHuang] Thanks for your attention, FLIP-314 is on voting 
https://lists.apache.org/thread/dxdqjc0dd8rf1vbdg755zo1n2zj1tj8d

> Flink supports reporting and storage of source/sink tables relationship
> ---
>
> Key: FLINK-31275
> URL: https://issues.apache.org/jira/browse/FLINK-31275
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Planner
>Affects Versions: 1.18.0
>Reporter: Fang Yong
>Priority: Major
>
> Currently flink generates job id in `JobGraph` which can identify a job. On 
> the other hand, flink create source/sink table in planner. We need to create 
> relations between source and sink tables for the job with an identifier id



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] add announcement blog post for Flink 1.18 [flink-web]

2023-10-06 Thread via GitHub


Myasuka commented on code in PR #680:
URL: https://github.com/apache/flink-web/pull/680#discussion_r1349442579


##
docs/content/posts/2023-10-10-release-1.18.0.md:
##
@@ -0,0 +1,542 @@
+---
+authors:
+- JingGe:
+  name: "Jing Ge"
+  twitter: jingengineer
+- KonstantinKnauf:
+  name: "Konstantin Knauf"
+  twitter: snntrable
+- SergeyNuyanzin:
+  name: "Sergey Nuyanzin"
+  twitter: uckamello
+- QingshengRen:
+  name: "Qingsheng Ren"
+  twitter: renqstuite
+date: "2023-10-10T08:00:00Z"
+subtitle: ""
+title: Announcing the Release of Apache Flink 1.18
+aliases:
+- /news/2023/10/10/release-1.18.0.html
+---
+
+The Apache Flink PMC is pleased to announce the release of Apache Flink 
1.18.0. As usual, we are looking at a packed 
+release with a wide variety of improvements and new features. Overall, 176 
people contributed to this release completing 
+18 FLIPS and 700+ issues. Thank you!
+
+Let's dive into the highlights.
+
+# Towards a Streaming Lakehouse
+
+## Flink SQL Improvements
+
+### Introduce Flink JDBC Driver For Sql Gateway 
+
+Flink 1.18 comes with a JDBC Driver for the Flink SQL Gateway. So, you can now 
use any SQL Client that supports JDBC to 
+interact with your tables via Flink SQL. Here is an example using 
[SQLLine](https://julianhyde.github.io/sqlline/manual.html). 
+
+```shell
+sqlline> !connect jdbc:flink://localhost:8083
+```
+
+```shell
+sqlline version 1.12.0
+sqlline> !connect jdbc:flink://localhost:8083
+Enter username for jdbc:flink://localhost:8083:
+Enter password for jdbc:flink://localhost:8083:
+0: jdbc:flink://localhost:8083> CREATE TABLE T(
+. . . . . . . . . . . . . . .)>  a INT,
+. . . . . . . . . . . . . . .)>  b VARCHAR(10)
+. . . . . . . . . . . . . . .)>  ) WITH (
+. . . . . . . . . . . . . . .)>  'connector' = 'filesystem',
+. . . . . . . . . . . . . . .)>  'path' = 'file:///tmp/T.csv',
+. . . . . . . . . . . . . . .)>  'format' = 'csv'
+. . . . . . . . . . . . . . .)>  );
+No rows affected (0.122 seconds)
+0: jdbc:flink://localhost:8083> INSERT INTO T VALUES (1, 'Hi'), (2, 'Hello');
++--+
+|  job id  |
++--+
+| fbade1ab4450fc57ebd5269fdf60dcfd |
++--+
+1 row selected (1.282 seconds)
+0: jdbc:flink://localhost:8083> SELECT * FROM T;
++---+---+
+| a |   b   |
++---+---+
+| 1 | Hi|
+| 2 | Hello |
++---+---+
+2 rows selected (1.955 seconds)
+0: jdbc:flink://localhost:8083>
+```
+
+**More Information**
+* 
[Documentation](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/jdbcdriver/)
 
+* [FLIP-293: Introduce Flink Jdbc Driver For Sql 
Gateway](https://cwiki.apache.org/confluence/display/FLINK/FLIP-293%3A+Introduce+Flink+Jdbc+Driver+For+Sql+Gateway)
+
+
+### Stored Procedures
+
+Stored Procedures provide a convenient way to encapsulate complex logic to 
perform data manipulation or administrative 
+tasks in Apache Flink itself. Therefore, Flink introduces the support for 
calling stored procedures. 
+Flink now allows catalog developers to develop their own built-in stored 
procedures and then enables users to call these
+predefined stored procedures.
+
+**More Information**
+* 
[Documentation](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/procedures/)
+* [FLIP-311: Support Call Stored 
Procedure](https://cwiki.apache.org/confluence/display/FLINK/FLIP-311%3A+Support+Call+Stored+Procedure)
+
+### Extended DDL Support
+
+From this release onwards, Flink supports
+
+- `REPLACE TABLE AS SELECT`
+- `CREATE OR REPLACE TABLE AS SELECT`
+
+and both these commands and previously supported `CREATE TABLE AS` can now 
support atomicity provided the underlying 
+connector supports this.
+
+Moreover, Apache Flink now supports TRUNCATE TABLE in batch execution mode. As 
before, the underlying connector needs 
+to implement and provide this capability
+
+And, finally, we have also added support for adding, dropping and listing 
partitions via
+
+- `ALTER TABLE ADD PARTITION`
+- `ALTER TABLE DROP PARTITION`
+- `SHOW PARTITIONS`
+
+**More Information**
+- [Documentation on 
TRUNCATE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/truncate/)
+- [Documentation on CREATE OR 
REPLACE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/create/#create-or-replace-table)
+- [Documentation on ALTER 
TABLE](https://nightlies.apache.org/flink/flink-docs-master/docs/dev/table/sql/alter/#alter-table)
+- [FLIP-302: Support TRUNCATE TABLE statement in batch 
mode](https://cwiki.apache.org/confluence/display/FLINK/FLIP-302%3A+Support+TRUNCATE+TABLE+statement+in+batch+mode)
+- [FLIP-303: Support REPLACE TABLE AS SELECT 
statement](https://cwiki.apache.org/confluence/display/FLINK/FLIP-303%3A+Support+REPLACE+TABLE+AS+SELECT+statement)
+- [FLIP-305: Support atomic for CREATE TABLE AS SELECT(CTAS) 
statement](https://cwiki.apache.org/confluence/display/FLINK/F

[jira] [Updated] (FLINK-31275) Flink supports reporting and storage of source/sink tables relationship

2023-10-06 Thread Fang Yong (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31275?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fang Yong updated FLINK-31275:
--
Description: FLIP-314 has been accepted 
https://cwiki.apache.org/confluence/display/FLINK/FLIP-314%3A+Support+Customized+Job+Lineage+Listener
  (was: Currently flink generates job id in `JobGraph` which can identify a 
job. On the other hand, flink create source/sink table in planner. We need to 
create relations between source and sink tables for the job with an identifier 
id)

> Flink supports reporting and storage of source/sink tables relationship
> ---
>
> Key: FLINK-31275
> URL: https://issues.apache.org/jira/browse/FLINK-31275
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Planner
>Affects Versions: 1.18.0
>Reporter: Fang Yong
>Priority: Major
>
> FLIP-314 has been accepted 
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-314%3A+Support+Customized+Job+Lineage+Listener



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Assigned] (FLINK-31275) Flink supports reporting and storage of source/sink tables relationship

2023-10-06 Thread Fang Yong (Jira)


 [ 
https://issues.apache.org/jira/browse/FLINK-31275?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Fang Yong reassigned FLINK-31275:
-

Assignee: Fang Yong

> Flink supports reporting and storage of source/sink tables relationship
> ---
>
> Key: FLINK-31275
> URL: https://issues.apache.org/jira/browse/FLINK-31275
> Project: Flink
>  Issue Type: Improvement
>  Components: Table SQL / Planner
>Affects Versions: 1.18.0
>Reporter: Fang Yong
>Assignee: Fang Yong
>Priority: Major
>
> FLIP-314 has been accepted 
> https://cwiki.apache.org/confluence/display/FLINK/FLIP-314%3A+Support+Customized+Job+Lineage+Listener



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33166][core][config] Make root logger level configurable [flink]

2023-10-06 Thread via GitHub


X-czh commented on PR #23476:
URL: https://github.com/apache/flink/pull/23476#issuecomment-1751601862

   @FangYongs, could you help review it when you have time? Many thanks.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



[jira] [Created] (FLINK-33201) Memory leak in CachingTopicSelector

2023-10-06 Thread Sunyeop Lee (Jira)
Sunyeop Lee created FLINK-33201:
---

 Summary: Memory leak in CachingTopicSelector
 Key: FLINK-33201
 URL: https://issues.apache.org/jira/browse/FLINK-33201
 Project: Flink
  Issue Type: Bug
  Components: Connectors / Kafka
 Environment: I am out of office now, so this is what I remember (the 
flink version may not be correct). Because I already identified the cause, this 
should not matter anyway.

EKS 1.24, x86_64, Bottlerocket OS, flink 1.14, scala 2.12
Reporter: Sunyeop Lee
 Attachments: 273084767-29bc0d8a-7445-4a74-a6e1-7c836775c7b1.png

Pull Request available at: 
https://github.com/apache/flink-connector-kafka/pull/55

 

In the CachingTopicSelector, a memory leak may occur when the internal logic 
fails to check the cache size due to a race condition. 
([https://github.com/apache/flink-connector-kafka/blob/d89a082180232bb79e3c764228c4e7dbb9eb6b8b/flink-connector-kafka/src/main/java/org/apache/flink/connector/kafka/sink/KafkaRecordSerializationSchemaBuilder.java#L287-L289)]

 

By analyzing a Java heap dump, I identified a memory leak in the 
CachingTopicSelector. As in the screenshot, cache has 47,769 elements. If the 
internal logic were functioning correctly, the number of elements should be 
less than or equal to CACHE_RESET_SIZE (which is 5).



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


[jira] [Created] (FLINK-33202) FLIP-327: Support switching from batch to stream mode to improve throughput when processing backlog data

2023-10-06 Thread Xuannan Su (Jira)
Xuannan Su created FLINK-33202:
--

 Summary: FLIP-327: Support switching from batch to stream mode to 
improve throughput when processing backlog data
 Key: FLINK-33202
 URL: https://issues.apache.org/jira/browse/FLINK-33202
 Project: Flink
  Issue Type: New Feature
  Components: Runtime / Task
Reporter: Xuannan Su
 Fix For: 1.19.0


Umbrella issue for 
[https://cwiki.apache.org/confluence/display/FLINK/FLIP-327%3A+Support+switching+from+batch+to+stream+mode+to+improve+throughput+when+processing+backlog+data|https://cwiki.apache.org/confluence/display/FLINK/FLIP-309%3A+Support+using+larger+checkpointing+interval+when+source+is+processing+backlog]
h4.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)


Re: [PR] [FLINK-33060] Fix the javadoc of ListState interfaces about not allowing null value [flink]

2023-10-06 Thread via GitHub


Zakelly commented on code in PR #23441:
URL: https://github.com/apache/flink/pull/23441#discussion_r1349473352


##
flink-core/src/main/java/org/apache/flink/api/common/state/AppendingState.java:
##
@@ -60,8 +60,11 @@ public interface AppendingState extends State {
  *
  * If null is passed in, the state value will remain unchanged.
  *
+ * Except for list state, where null value is not allowed to be passed 
in.

Review Comment:
   Agreed. I believe it would be better to eliminate the description of the 
null value at the moment, considering the implementations have already become 
varied. We can consider refactoring the interfaces in the upcoming major 
version. What do you think? @davidradl 



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



Re: [PR] [FLINK-33161] Java17 profile for benchmarking [flink-benchmarks]

2023-10-06 Thread via GitHub


Zakelly commented on PR #80:
URL: https://github.com/apache/flink-benchmarks/pull/80#issuecomment-1751624381

   I'm not sure if I have the right person, @pnowojski , but would you please 
review this one?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org