Re: [PR] [docs] Polish quickstart guide & migrate maven links from ververica to apache [flink-cdc]
gtk96 commented on PR #3343: URL: https://github.com/apache/flink-cdc/pull/3343#issuecomment-2146793364 cc @Jiabao-Sun @GOODBOY008 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-35473) FLIP-457: Improve Table/SQL Configuration for Flink 2.0
[ https://issues.apache.org/jira/browse/FLINK-35473?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17851938#comment-17851938 ] Jane Chan commented on FLINK-35473: --- [~lincoln.86xy] Thank you for the reminder. I'm planning to open a PR in the coming days, and it would be great if you could help review it. > FLIP-457: Improve Table/SQL Configuration for Flink 2.0 > --- > > Key: FLINK-35473 > URL: https://issues.apache.org/jira/browse/FLINK-35473 > Project: Flink > Issue Type: Improvement > Components: Table SQL / API >Affects Versions: 1.20.0 >Reporter: Jane Chan >Assignee: Jane Chan >Priority: Major > Fix For: 1.20.0 > > > This is the parent task for > [FLIP-457|https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=307136992]. -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-34559] Limit Global & Local Aggregation buffers [flink]
twalthr commented on code in PR #24869: URL: https://github.com/apache/flink/pull/24869#discussion_r1625562520 ## flink-core/src/main/java/org/apache/flink/api/common/ExecutionConfig.java: ## @@ -276,6 +277,16 @@ public long getLatencyTrackingInterval() { return configuration.get(MetricOptions.LATENCY_INTERVAL).toMillis(); } +@PublicEvolving +public Optional getGlobalAggregationBufferSize() { Review Comment: You are mixing layers here. Global aggregation is a concept of the SQL layer and thus should only change classes in the `flink-table` module. Configuration doesn't need to be added to `ExecutionConfig`. You can pass parameters to the runtime class during planning in the corresponding ExecNode. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[PR] Update ttlConfig in state.md [flink]
lizhigong opened a new pull request, #24884: URL: https://github.com/apache/flink/pull/24884 Duration.ofSeconds(1) is no longer accepted by the API, use Time.seconds(1) instead. ## What is the purpose of the change Change typos. ## Brief change log In the documentation of [Working With State] (https://nightlies.apache.org/flink/flink-docs-master/docs/dev/datastream/fault-tolerance/state/#state-time-to-live-ttl), the code uses `Duration.ofSeconds(1)`, however, the accepted type is `org.apache.flink.api.common.time.Time`, thus this should be changed to `Time.seconds(1)`. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] Update ttlConfig in state.md [flink]
lizhigong closed pull request #24884: Update ttlConfig in state.md URL: https://github.com/apache/flink/pull/24884 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-35517) CI pipeline triggered by pull request seems unstable
Weijie Guo created FLINK-35517: -- Summary: CI pipeline triggered by pull request seems unstable Key: FLINK-35517 URL: https://issues.apache.org/jira/browse/FLINK-35517 Project: Flink Issue Type: Bug Components: Build System / CI Affects Versions: 1.20.0 Reporter: Weijie Guo Flink CI pipeline triggered by pull request seems sort of unstable. For example, https://github.com/apache/flink/pull/24883 was filed 15 hours ago, but CI report is UNKNOWN. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-35517) CI pipeline triggered by pull request seems unstable
[ https://issues.apache.org/jira/browse/FLINK-35517?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Weijie Guo updated FLINK-35517: --- Priority: Blocker (was: Major) > CI pipeline triggered by pull request seems unstable > > > Key: FLINK-35517 > URL: https://issues.apache.org/jira/browse/FLINK-35517 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Priority: Blocker > > Flink CI pipeline triggered by pull request seems sort of unstable. > For example, https://github.com/apache/flink/pull/24883 was filed 15 hours > ago, but CI report is UNKNOWN. -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-35464] Fixes operator state backwards compatibility from CDC 3.0.x [flink-cdc]
PatrickRen commented on code in PR #3369: URL: https://github.com/apache/flink-cdc/pull/3369#discussion_r1625602137 ## flink-cdc-migration-tests/pom.xml: ## @@ -19,13 +19,10 @@ limitations under the License. xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"; xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd";> 4.0.0 - Review Comment: Actually you can keep the section here. Removing `flink-cdc-migration-tests` from the section in the parent pom should be enough. ## pom.xml: ## @@ -40,7 +40,6 @@ limitations under the License. flink-cdc-connect flink-cdc-runtime flink-cdc-e2e-tests -flink-cdc-migration-tests Review Comment: Is it possible to use profiles to control whether to build `flink-cdc-migration-tests` module? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [FLINK-35149][cdc-composer] Fix DataSinkTranslator#sinkTo ignoring pre-write topology if not TwoPhaseCommittingSink [flink-cdc]
PatrickRen merged PR #3233: URL: https://github.com/apache/flink-cdc/pull/3233 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [FLINK-35149][cdc-composer] Fix DataSinkTranslator#sinkTo ignoring pre-write topology if not TwoPhaseCommittingSink [flink-cdc]
PatrickRen commented on PR #3233: URL: https://github.com/apache/flink-cdc/pull/3233#issuecomment-2146945078 @loserwang1024 Could you cherry-pick this commit to release-3.1 branch? Thanks -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [FLINK-35282][python] Upgrade pyarrow and beam [flink]
hlteoh37 merged PR #24875: URL: https://github.com/apache/flink/pull/24875 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-35149) Fix DataSinkTranslator#sinkTo ignoring pre-write topology if not TwoPhaseCommittingSink
[ https://issues.apache.org/jira/browse/FLINK-35149?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17851955#comment-17851955 ] Qingsheng Ren commented on FLINK-35149: --- flink-cdc master: 33891869a9fffa2abf8b8ae03915d0ddccdaf5ec > Fix DataSinkTranslator#sinkTo ignoring pre-write topology if not > TwoPhaseCommittingSink > --- > > Key: FLINK-35149 > URL: https://issues.apache.org/jira/browse/FLINK-35149 > Project: Flink > Issue Type: Bug > Components: Flink CDC >Affects Versions: cdc-3.1.0 >Reporter: Hongshun Wang >Assignee: Hongshun Wang >Priority: Minor > Labels: pull-request-available > Fix For: cdc-3.2.0 > > > Current , when sink is not instanceof TwoPhaseCommittingSink, use > input.transform rather than stream. It means that pre-write topology will be > ignored. > {code:java} > private void sinkTo( > DataStream input, > Sink sink, > String sinkName, > OperatorID schemaOperatorID) { > DataStream stream = input; > // Pre write topology > if (sink instanceof WithPreWriteTopology) { > stream = ((WithPreWriteTopology) > sink).addPreWriteTopology(stream); > } > if (sink instanceof TwoPhaseCommittingSink) { > addCommittingTopology(sink, stream, sinkName, schemaOperatorID); > } else { > input.transform( > SINK_WRITER_PREFIX + sinkName, > CommittableMessageTypeInfo.noOutput(), > new DataSinkWriterOperatorFactory<>(sink, schemaOperatorID)); > } > } {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (FLINK-35149) Fix DataSinkTranslator#sinkTo ignoring pre-write topology if not TwoPhaseCommittingSink
[ https://issues.apache.org/jira/browse/FLINK-35149?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Qingsheng Ren reassigned FLINK-35149: - Assignee: Hongshun Wang > Fix DataSinkTranslator#sinkTo ignoring pre-write topology if not > TwoPhaseCommittingSink > --- > > Key: FLINK-35149 > URL: https://issues.apache.org/jira/browse/FLINK-35149 > Project: Flink > Issue Type: Bug > Components: Flink CDC >Affects Versions: cdc-3.1.0 >Reporter: Hongshun Wang >Assignee: Hongshun Wang >Priority: Minor > Labels: pull-request-available > Fix For: cdc-3.2.0 > > > Current , when sink is not instanceof TwoPhaseCommittingSink, use > input.transform rather than stream. It means that pre-write topology will be > ignored. > {code:java} > private void sinkTo( > DataStream input, > Sink sink, > String sinkName, > OperatorID schemaOperatorID) { > DataStream stream = input; > // Pre write topology > if (sink instanceof WithPreWriteTopology) { > stream = ((WithPreWriteTopology) > sink).addPreWriteTopology(stream); > } > if (sink instanceof TwoPhaseCommittingSink) { > addCommittingTopology(sink, stream, sinkName, schemaOperatorID); > } else { > input.transform( > SINK_WRITER_PREFIX + sinkName, > CommittableMessageTypeInfo.noOutput(), > new DataSinkWriterOperatorFactory<>(sink, schemaOperatorID)); > } > } {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (FLINK-35282) PyFlink Support for Apache Beam > 2.49
[ https://issues.apache.org/jira/browse/FLINK-35282?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hong Liang Teoh resolved FLINK-35282. - Resolution: Fixed > PyFlink Support for Apache Beam > 2.49 > -- > > Key: FLINK-35282 > URL: https://issues.apache.org/jira/browse/FLINK-35282 > Project: Flink > Issue Type: Improvement > Components: API / Python >Affects Versions: 1.19.0, 1.18.1 >Reporter: APA >Priority: Major > Labels: pull-request-available > Fix For: 1.20.0 > > > From what I see PyFlink still has the requirement of Apache Beam => 2.43.0 > and <= 2.49.0 which subsequently results in a requirement of PyArrow <= > 12.0.0. That keeps us exposed to > [https://nvd.nist.gov/vuln/detail/CVE-2023-47248] > I'm not deep enough familiar with the PyFlink code base to understand why > Apache Beam's upper dependency limit can't be lifted. From all the existing > issues I haven't seen one addressing this. Therefore I created one now. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-35282) PyFlink Support for Apache Beam > 2.49
[ https://issues.apache.org/jira/browse/FLINK-35282?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hong Liang Teoh updated FLINK-35282: Fix Version/s: (was: 1.19.1) > PyFlink Support for Apache Beam > 2.49 > -- > > Key: FLINK-35282 > URL: https://issues.apache.org/jira/browse/FLINK-35282 > Project: Flink > Issue Type: Improvement > Components: API / Python >Affects Versions: 1.19.0, 1.18.1 >Reporter: APA >Priority: Major > Labels: pull-request-available > Fix For: 1.20.0 > > > From what I see PyFlink still has the requirement of Apache Beam => 2.43.0 > and <= 2.49.0 which subsequently results in a requirement of PyArrow <= > 12.0.0. That keeps us exposed to > [https://nvd.nist.gov/vuln/detail/CVE-2023-47248] > I'm not deep enough familiar with the PyFlink code base to understand why > Apache Beam's upper dependency limit can't be lifted. From all the existing > issues I haven't seen one addressing this. Therefore I created one now. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-35282) PyFlink Support for Apache Beam > 2.49
[ https://issues.apache.org/jira/browse/FLINK-35282?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17851954#comment-17851954 ] Hong Liang Teoh commented on FLINK-35282: - merged commit [{{91a9e06}}|https://github.com/apache/flink/commit/91a9e06d3cb611e048274088e56b2c110cd29926] into apache:master > PyFlink Support for Apache Beam > 2.49 > -- > > Key: FLINK-35282 > URL: https://issues.apache.org/jira/browse/FLINK-35282 > Project: Flink > Issue Type: Improvement > Components: API / Python >Affects Versions: 1.19.0, 1.18.1 >Reporter: APA >Priority: Major > Labels: pull-request-available > Fix For: 1.20.0 > > > From what I see PyFlink still has the requirement of Apache Beam => 2.43.0 > and <= 2.49.0 which subsequently results in a requirement of PyArrow <= > 12.0.0. That keeps us exposed to > [https://nvd.nist.gov/vuln/detail/CVE-2023-47248] > I'm not deep enough familiar with the PyFlink code base to understand why > Apache Beam's upper dependency limit can't be lifted. From all the existing > issues I haven't seen one addressing this. Therefore I created one now. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-35149) Fix DataSinkTranslator#sinkTo ignoring pre-write topology if not TwoPhaseCommittingSink
[ https://issues.apache.org/jira/browse/FLINK-35149?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Qingsheng Ren updated FLINK-35149: -- Fix Version/s: cdc-3.1.1 > Fix DataSinkTranslator#sinkTo ignoring pre-write topology if not > TwoPhaseCommittingSink > --- > > Key: FLINK-35149 > URL: https://issues.apache.org/jira/browse/FLINK-35149 > Project: Flink > Issue Type: Bug > Components: Flink CDC >Affects Versions: cdc-3.1.0 >Reporter: Hongshun Wang >Assignee: Hongshun Wang >Priority: Minor > Labels: pull-request-available > Fix For: cdc-3.2.0, cdc-3.1.1 > > > Current , when sink is not instanceof TwoPhaseCommittingSink, use > input.transform rather than stream. It means that pre-write topology will be > ignored. > {code:java} > private void sinkTo( > DataStream input, > Sink sink, > String sinkName, > OperatorID schemaOperatorID) { > DataStream stream = input; > // Pre write topology > if (sink instanceof WithPreWriteTopology) { > stream = ((WithPreWriteTopology) > sink).addPreWriteTopology(stream); > } > if (sink instanceof TwoPhaseCommittingSink) { > addCommittingTopology(sink, stream, sinkName, schemaOperatorID); > } else { > input.transform( > SINK_WRITER_PREFIX + sinkName, > CommittableMessageTypeInfo.noOutput(), > new DataSinkWriterOperatorFactory<>(sink, schemaOperatorID)); > } > } {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (FLINK-35282) PyFlink Support for Apache Beam > 2.49
[ https://issues.apache.org/jira/browse/FLINK-35282?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hong Liang Teoh reassigned FLINK-35282: --- Assignee: Antonio Vespoli > PyFlink Support for Apache Beam > 2.49 > -- > > Key: FLINK-35282 > URL: https://issues.apache.org/jira/browse/FLINK-35282 > Project: Flink > Issue Type: Improvement > Components: API / Python >Affects Versions: 1.19.0, 1.18.1 >Reporter: APA >Assignee: Antonio Vespoli >Priority: Major > Labels: pull-request-available > Fix For: 1.20.0 > > > From what I see PyFlink still has the requirement of Apache Beam => 2.43.0 > and <= 2.49.0 which subsequently results in a requirement of PyArrow <= > 12.0.0. That keeps us exposed to > [https://nvd.nist.gov/vuln/detail/CVE-2023-47248] > I'm not deep enough familiar with the PyFlink code base to understand why > Apache Beam's upper dependency limit can't be lifted. From all the existing > issues I haven't seen one addressing this. Therefore I created one now. -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-35129][cdc][postgres] Add checkpoint cycle option for commit offsets [flink-cdc]
leonardBang merged PR #3349: URL: https://github.com/apache/flink-cdc/pull/3349 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-35517) CI pipeline triggered by pull request seems unstable
[ https://issues.apache.org/jira/browse/FLINK-35517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17851962#comment-17851962 ] Weijie Guo commented on FLINK-35517: I have reached out [~lorenzo.affetti] to see if this is related to the Ververica machine? cc [~jingge] > CI pipeline triggered by pull request seems unstable > > > Key: FLINK-35517 > URL: https://issues.apache.org/jira/browse/FLINK-35517 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Priority: Blocker > > Flink CI pipeline triggered by pull request seems sort of unstable. > For example, https://github.com/apache/flink/pull/24883 was filed 15 hours > ago, but CI report is UNKNOWN. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (FLINK-35129) Postgres source commits the offset after every multiple checkpoint cycles.
[ https://issues.apache.org/jira/browse/FLINK-35129?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Leonard Xu resolved FLINK-35129. Resolution: Implemented Implemented via master: 5b28d1a579919b29acac6acded46d9bee5596bde > Postgres source commits the offset after every multiple checkpoint cycles. > -- > > Key: FLINK-35129 > URL: https://issues.apache.org/jira/browse/FLINK-35129 > Project: Flink > Issue Type: Improvement > Components: Flink CDC >Affects Versions: cdc-3.1.0 >Reporter: Hongshun Wang >Assignee: Muhammet Orazov >Priority: Major > Labels: pull-request-available > Fix For: cdc-3.2.0 > > > After entering the Stream phase, the offset consumed by the global slot is > committed upon the completion of each checkpoint, preventing log files from > being unable to be recycled continuously, which could lead to insufficient > disk space. > However, the job can only restart from the latest checkpoint or savepoint. if > restored from an earlier state, WAL may already have been recycled. > > The way to solve it is to commit the offset after every multiple checkpoint > cycles. The number of checkpoint cycles is determine by connector option, and > the default value is 3. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (FLINK-35518) CI Bot doesn't run on PRs
Piotr Nowojski created FLINK-35518: -- Summary: CI Bot doesn't run on PRs Key: FLINK-35518 URL: https://issues.apache.org/jira/browse/FLINK-35518 Project: Flink Issue Type: Bug Components: Build System / CI Affects Versions: 1.20.0 Reporter: Piotr Nowojski Doesn't want to run on my PR/branch. I was doing force-pushes, rebases, asking flink bot to run, closed and opened new PR, but nothing helped https://github.com/apache/flink/pull/24868 https://github.com/apache/flink/pull/24883 I've heard others were having similar problems recently. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-35518) CI Bot doesn't run on PRs - status UNKNOWN
[ https://issues.apache.org/jira/browse/FLINK-35518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Piotr Nowojski updated FLINK-35518: --- Summary: CI Bot doesn't run on PRs - status UNKNOWN (was: CI Bot doesn't run on PRs) > CI Bot doesn't run on PRs - status UNKNOWN > -- > > Key: FLINK-35518 > URL: https://issues.apache.org/jira/browse/FLINK-35518 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Piotr Nowojski >Priority: Critical > > Doesn't want to run on my PR/branch. I was doing force-pushes, rebases, > asking flink bot to run, closed and opened new PR, but nothing helped > https://github.com/apache/flink/pull/24868 > https://github.com/apache/flink/pull/24883 > I've heard others were having similar problems recently. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-35518) CI Bot doesn't run on PRs - status UNKNOWN
[ https://issues.apache.org/jira/browse/FLINK-35518?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17851969#comment-17851969 ] Weijie Guo commented on FLINK-35518: Yes, I have created a ticket(FLINK-35517) also intend to track this issue. > CI Bot doesn't run on PRs - status UNKNOWN > -- > > Key: FLINK-35518 > URL: https://issues.apache.org/jira/browse/FLINK-35518 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Piotr Nowojski >Priority: Critical > > Doesn't want to run on my PR/branch. I was doing force-pushes, rebases, > asking flink bot to run, closed and opened new PR, but nothing helped > https://github.com/apache/flink/pull/24868 > https://github.com/apache/flink/pull/24883 > I've heard others were having similar problems recently. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (FLINK-35517) CI pipeline triggered by pull request seems unstable
[ https://issues.apache.org/jira/browse/FLINK-35517?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jing Ge reassigned FLINK-35517: --- Assignee: Jing Ge > CI pipeline triggered by pull request seems unstable > > > Key: FLINK-35517 > URL: https://issues.apache.org/jira/browse/FLINK-35517 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Assignee: Jing Ge >Priority: Blocker > > Flink CI pipeline triggered by pull request seems sort of unstable. > For example, https://github.com/apache/flink/pull/24883 was filed 15 hours > ago, but CI report is UNKNOWN. -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-35464] Fixes operator state backwards compatibility from CDC 3.0.x [flink-cdc]
yuxiqian commented on PR #3369: URL: https://github.com/apache/flink-cdc/pull/3369#issuecomment-2147045294 Thanks for the tips! Addressed your comments to simplify pom files. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-35519) Flink Job fails with SingleValueAggFunction received more than one element
Dawid Wysakowicz created FLINK-35519: Summary: Flink Job fails with SingleValueAggFunction received more than one element Key: FLINK-35519 URL: https://issues.apache.org/jira/browse/FLINK-35519 Project: Flink Issue Type: Bug Components: Table SQL / Planner Affects Versions: 1.19.0 Reporter: Dawid Wysakowicz When running a query: {code} select (SELECT t.id FROM raw_pagerduty_users, UNNEST(teams) AS t(id, type, summary, self, html_url)) from raw_pagerduty_users; {code} it is translated to: {code} Sink(table=[default_catalog.default_database.sink], fields=[EXPR$0]) +- Calc(select=[$f0 AS EXPR$0]) +- Join(joinType=[LeftOuterJoin], where=[true], select=[c, $f0], leftInputSpec=[NoUniqueKey], rightInputSpec=[JoinKeyContainsUniqueKey]) :- Exchange(distribution=[single]) : +- Calc(select=[c]) : +- TableSourceScan(table=[[default_catalog, default_database, raw_pagerduty_users, project=[c, teams], metadata=[]]], fields=[c, teams])(reuse_id=[1]) +- Exchange(distribution=[single]) +- GroupAggregate(select=[SINGLE_VALUE(id) AS $f0]) +- Exchange(distribution=[single]) +- Calc(select=[id]) +- Correlate(invocation=[$UNNEST_ROWS$1($cor0.teams)], correlate=[table($UNNEST_ROWS$1($cor0.teams))], select=[c,teams,id,type,summary,self,html_url], rowType=[RecordType(BIGINT c, RecordType:peek_no_expand(VARCHAR(2147483647) id, VARCHAR(2147483647) type, VARCHAR(2147483647) summary, VARCHAR(2147483647) self, VARCHAR(2147483647) html_url) ARRAY teams, VARCHAR(2147483647) id, VARCHAR(2147483647) type, VARCHAR(2147483647) summary, VARCHAR(2147483647) self, VARCHAR(2147483647) html_url)], joinType=[INNER]) +- Reused(reference_id=[1]) {code} and it fails with: {code} java.lang.RuntimeException: SingleValueAggFunction received more than one element. at GroupAggsHandler$150.accumulate(Unknown Source) at org.apache.flink.table.runtime.operators.aggregate.GroupAggFunction.processElement(GroupAggFunction.java:151) at org.apache.flink.table.runtime.operators.aggregate.GroupAggFunction.processElement(GroupAggFunction.java:43) at org.apache.flink.streaming.api.operators.KeyedProcessOperator.processElement(KeyedProcessOperator.java:83) at org.apache.flink.streaming.runtime.io.RecordProcessorUtils.lambda$getRecordProcessor$0(RecordProcessorUtils.java:60) at org.apache.flink.streaming.runtime.tasks.OneInputStreamTask$StreamTaskNetworkOutput.emitRecord(OneInputStreamTask.java:237) at org.apache.flink.streaming.runtime.io.AbstractStreamTaskNetworkInput.processElement(AbstractStreamTaskNetworkInput.java:146) at org.apache.flink.streaming.runtime.io.AbstractStreamTaskNetworkInput.emitNext(AbstractStreamTaskNetworkInput.java:110) at org.apache.flink.streaming.runtime.io.StreamOneInputProcessor.processInput(StreamOneInputProcessor.java:65) at org.apache.flink.streaming.runtime.tasks.StreamTask.processInput(StreamTask.java:571) at org.apache.flink.streaming.runtime.tasks.mailbox.MailboxProcessor.runMailboxLoop(MailboxProcessor.java:231) at org.apache.flink.streaming.runtime.tasks.StreamTask.runMailboxLoop(StreamTask.java:900) at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:849) at org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:953) at org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:932) at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:746) at org.apache.flink.runtime.taskmanager.Task.run(Task.java:562) at java.base/java.lang.Thread.run(Thread.java:829) {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-35305]Amazon SQS Sink Connector [flink-connector-aws]
hlteoh37 commented on code in PR #141: URL: https://github.com/apache/flink-connector-aws/pull/141#discussion_r1625636303 ## docs/content.zh/docs/connectors/datastream/sqs.md: ## @@ -0,0 +1,134 @@ +--- +title: DynamoDB +weight: 5 +type: docs +aliases: +- /zh/dev/connectors/dynamodb.html Review Comment: These are wrong and should be changed to `sqs.html` ## docs/content.zh/docs/connectors/datastream/sqs.md: ## @@ -0,0 +1,134 @@ +--- +title: DynamoDB +weight: 5 +type: docs +aliases: +- /zh/dev/connectors/dynamodb.html +--- + + +# Amazon SQS Sink + +The SQS sink writes to [Amazon SQS](https://aws.amazon.com/sqs) using the [AWS v2 SDK for Java](https://docs.aws.amazon.com/sdk-for-java/latest/developer-guide/home.html). Follow the instructions from the [Amazon SQS Developer Guide](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/welcome.html) +to setup a SQS. + +To use the connector, add the following Maven dependency to your project: + +{{< connector_artifact flink-connector-sqs sqs >}} + +{{< tabs "ec24a4ae-6a47-11ed-a1eb-0242ac120002" >}} +{{< tab "Java" >}} +```java +Properties sinkProperties = new Properties(); +// Required +sinkProperties.put(AWSConfigConstants.AWS_REGION, "eu-west-1"); Review Comment: Is there a reason we need to specify this? Can we figure this out via the SQS URL? ## flink-connector-aws/flink-connector-sqs/src/main/java/org.apache.flink.connector.sqs/sink/SqsConfigConstants.java: ## @@ -0,0 +1,31 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + *http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.connector.sqs.sink; + +import org.apache.flink.annotation.PublicEvolving; + +/** Defaults for {@link SqsSinkWriter}. */ +@PublicEvolving +public class SqsConfigConstants { + +public static final String BASE_SQS_USER_AGENT_PREFIX_FORMAT = Review Comment: Could we use `ConfigOption` instead of strings here? Example: https://github.com/apache/flink-connector-aws/blob/38aafb44d3a8200e4ff41d87e0780338f40da258/flink-connector-aws/flink-connector-aws-kinesis-streams/src/main/java/org/apache/flink/connector/kinesis/source/config/KinesisStreamsSourceConfigConstants.java#L41-L46 ## docs/content.zh/docs/connectors/datastream/sqs.md: ## @@ -0,0 +1,134 @@ +--- +title: DynamoDB +weight: 5 +type: docs +aliases: +- /zh/dev/connectors/dynamodb.html +--- + + +# Amazon SQS Sink + +The SQS sink writes to [Amazon SQS](https://aws.amazon.com/sqs) using the [AWS v2 SDK for Java](https://docs.aws.amazon.com/sdk-for-java/latest/developer-guide/home.html). Follow the instructions from the [Amazon SQS Developer Guide](https://docs.aws.amazon.com/AWSSimpleQueueService/latest/SQSDeveloperGuide/welcome.html) +to setup a SQS. Review Comment: nit: setup an SQS message queue. ## flink-connector-aws/flink-connector-sqs/src/main/java/org.apache.flink.connector.sqs/sink/SqsSinkWriter.java: ## @@ -0,0 +1,257 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + *http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.connector.sqs.sink; + +import org.apache.flink.annotation.Internal; +import org.apache.flink.annotation.VisibleForTesting; +import org.apache.flink.api.connector.sink2.Sink; +import org.apache.flink.connector.aws.sink.throwable.AWSExceptionHandler; +import org.apache.flink.connector.aws.util.AWSClientUtil; +import org.apache.flink.connector.aws.util.AWSGeneralUtil; +import org.apache.flink.connector.base.sink.thr
[jira] [Created] (FLINK-35520) Nightly build can't compile as problems were detected from NoticeFileChecker
Weijie Guo created FLINK-35520: -- Summary: Nightly build can't compile as problems were detected from NoticeFileChecker Key: FLINK-35520 URL: https://issues.apache.org/jira/browse/FLINK-35520 Project: Flink Issue Type: Bug Components: Build System / CI Affects Versions: 1.20.0 Reporter: Weijie Guo -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-35520) Nightly build can't compile as problems were detected from NoticeFileChecker
[ https://issues.apache.org/jira/browse/FLINK-35520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Weijie Guo updated FLINK-35520: --- Description: 09:26:55,177 ERROR org.apache.flink.tools.ci.licensecheck.JarFileChecker [] - Jar file /tmp/flink-validation-deployment/org/apache/flink/flink-python/1.20-SNAPSHOT/flink-python-1.20-20240604.084637-1.jar contains a LICENSE file in an unexpected location: /LICENSE [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45805] was: [] - Jar file /tmp/flink-validation-deployment/org/apache/flink/flink-python/1.20-SNAPSHOT/flink-python-1.20-20240604.084637-1.jar contains a LICENSE file in an unexpected location: /LICENSE https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45805 > Nightly build can't compile as problems were detected from NoticeFileChecker > > > Key: FLINK-35520 > URL: https://issues.apache.org/jira/browse/FLINK-35520 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Priority: Blocker > > 09:26:55,177 ERROR org.apache.flink.tools.ci.licensecheck.JarFileChecker [] - > Jar file > /tmp/flink-validation-deployment/org/apache/flink/flink-python/1.20-SNAPSHOT/flink-python-1.20-20240604.084637-1.jar > contains a LICENSE file in an unexpected location: /LICENSE > > [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45805] -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-35305]Amazon SQS Sink Connector [flink-connector-aws]
hlteoh37 commented on PR #141: URL: https://github.com/apache/flink-connector-aws/pull/141#issuecomment-2147097522 We seem to be having quite a few `.` in the class folders. Can we change them to `/` instead? e.g. `flink-connector-aws/flink-connector-sqs/src/test/java/org.apache.flink/connector.sqs/sink/SqsExceptionClassifiersTest.java -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-35520) Nightly build can't compile as problems were detected from NoticeFileChecker
[ https://issues.apache.org/jira/browse/FLINK-35520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Weijie Guo updated FLINK-35520: --- Description: [] - Jar file /tmp/flink-validation-deployment/org/apache/flink/flink-python/1.20-SNAPSHOT/flink-python-1.20-20240604.084637-1.jar contains a LICENSE file in an unexpected location: /LICENSE https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45805 > Nightly build can't compile as problems were detected from NoticeFileChecker > > > Key: FLINK-35520 > URL: https://issues.apache.org/jira/browse/FLINK-35520 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Priority: Blocker > > [] - Jar file > /tmp/flink-validation-deployment/org/apache/flink/flink-python/1.20-SNAPSHOT/flink-python-1.20-20240604.084637-1.jar > contains a LICENSE file in an unexpected location: /LICENSE > > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45805 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-35520) Nightly build can't compile
[ https://issues.apache.org/jira/browse/FLINK-35520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Weijie Guo updated FLINK-35520: --- Summary: Nightly build can't compile (was: Nightly build can't compile as problems were detected from NoticeFileChecker) > Nightly build can't compile > --- > > Key: FLINK-35520 > URL: https://issues.apache.org/jira/browse/FLINK-35520 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Priority: Blocker > > 09:26:55,177 ERROR org.apache.flink.tools.ci.licensecheck.JarFileChecker [] - > Jar file > /tmp/flink-validation-deployment/org/apache/flink/flink-python/1.20-SNAPSHOT/flink-python-1.20-20240604.084637-1.jar > contains a LICENSE file in an unexpected location: /LICENSE > > [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45805] -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-35305]Amazon SQS Sink Connector [flink-connector-aws]
hlteoh37 commented on code in PR #141: URL: https://github.com/apache/flink-connector-aws/pull/141#discussion_r1625705602 ## flink-connector-aws/flink-connector-sqs/src/main/java/org.apache.flink.connector.sqs/sink/SqsSinkElementConverter.java: ## @@ -0,0 +1,105 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one or more + * contributor license agreements. See the NOTICE file distributed with + * this work for additional information regarding copyright ownership. + * The ASF licenses this file to You under the Apache License, Version 2.0 + * (the "License"); you may not use this file except in compliance with + * the License. You may obtain a copy of the License at + * + *http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.connector.sqs.sink; + +import org.apache.flink.annotation.Internal; +import org.apache.flink.api.common.serialization.SerializationSchema; +import org.apache.flink.api.connector.sink2.Sink; +import org.apache.flink.api.connector.sink2.SinkWriter; +import org.apache.flink.connector.base.sink.writer.ElementConverter; +import org.apache.flink.metrics.MetricGroup; +import org.apache.flink.metrics.groups.UnregisteredMetricsGroup; +import org.apache.flink.util.FlinkRuntimeException; +import org.apache.flink.util.Preconditions; +import org.apache.flink.util.SimpleUserCodeClassLoader; +import org.apache.flink.util.UserCodeClassLoader; + +import software.amazon.awssdk.services.sqs.model.SendMessageBatchRequestEntry; + +import java.nio.charset.StandardCharsets; +import java.util.UUID; + +/** + * An implementation of the {@link ElementConverter} that uses the AWS SQS SDK v2. The user only + * needs to provide a {@link SerializationSchema} of the {@code InputT} to transform it into a + * {@link SendMessageBatchRequestEntry} that may be persisted. + */ +@Internal +public class SqsSinkElementConverter +implements ElementConverter { + +/** A serialization schema to specify how the input element should be serialized. */ +private final SerializationSchema serializationSchema; + +private SqsSinkElementConverter(SerializationSchema serializationSchema) { +this.serializationSchema = serializationSchema; +} + +@Override +public SendMessageBatchRequestEntry apply(InputT element, SinkWriter.Context context) { +final byte[] messageBody = serializationSchema.serialize(element); +return SendMessageBatchRequestEntry.builder() +.id(UUID.randomUUID().toString()) +.messageBody(new String(messageBody, StandardCharsets.UTF_8)) +.build(); +} + +@Override +public void open(Sink.InitContext context) { +try { +serializationSchema.open( +new SerializationSchema.InitializationContext() { +@Override +public MetricGroup getMetricGroup() { +return new UnregisteredMetricsGroup(); +} + +@Override +public UserCodeClassLoader getUserCodeClassLoader() { +return SimpleUserCodeClassLoader.create( + SqsSinkElementConverter.class.getClassLoader()); +} +}); +} catch (Exception e) { +throw new FlinkRuntimeException("Failed to initialize serialization schema.", e); +} +} + +public static Builder builder() { +return new Builder<>(); +} + +/** A builder for the SqsSinkElementConverter. */ +public static class Builder { + +private SerializationSchema serializationSchema; + +public Builder setSerializationSchema( +SerializationSchema serializationSchema) { +this.serializationSchema = serializationSchema; +return this; +} + +public SqsSinkElementConverter build() { +Preconditions.checkNotNull( +serializationSchema, +"No SerializationSchema was supplied to the " + "SQS Sink builder."); Review Comment: nit: `+` not needed here -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-35520) Nightly build can't compile as license check failed
[ https://issues.apache.org/jira/browse/FLINK-35520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Weijie Guo updated FLINK-35520: --- Summary: Nightly build can't compile as license check failed (was: Nightly build can't compile) > Nightly build can't compile as license check failed > --- > > Key: FLINK-35520 > URL: https://issues.apache.org/jira/browse/FLINK-35520 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Priority: Blocker > > 09:26:55,177 ERROR org.apache.flink.tools.ci.licensecheck.JarFileChecker [] - > Jar file > /tmp/flink-validation-deployment/org/apache/flink/flink-python/1.20-SNAPSHOT/flink-python-1.20-20240604.084637-1.jar > contains a LICENSE file in an unexpected location: /LICENSE > > [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45805] -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-35520) Nightly build can't compile as license check failed
[ https://issues.apache.org/jira/browse/FLINK-35520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Weijie Guo updated FLINK-35520: --- Description: [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45807|https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45805] (was: 09:26:55,177 ERROR org.apache.flink.tools.ci.licensecheck.JarFileChecker [] - Jar file /tmp/flink-validation-deployment/org/apache/flink/flink-python/1.20-SNAPSHOT/flink-python-1.20-20240604.084637-1.jar contains a LICENSE file in an unexpected location: /LICENSE [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45805]) > Nightly build can't compile as license check failed > --- > > Key: FLINK-35520 > URL: https://issues.apache.org/jira/browse/FLINK-35520 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Priority: Blocker > > [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45807|https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45805] -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-35520) Nightly build can't compile as license check failed
[ https://issues.apache.org/jira/browse/FLINK-35520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Weijie Guo updated FLINK-35520: --- Description: https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45808 (was: [https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45807|https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45805]) > Nightly build can't compile as license check failed > --- > > Key: FLINK-35520 > URL: https://issues.apache.org/jira/browse/FLINK-35520 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Priority: Blocker > > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45808 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-35520) Nightly build can't compile as license check failed
[ https://issues.apache.org/jira/browse/FLINK-35520?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17851983#comment-17851983 ] Weijie Guo commented on FLINK-35520: Hi [~antoniovespoli], would you mind taking a look? I found this issue after 91a9e06d merged. > Nightly build can't compile as license check failed > --- > > Key: FLINK-35520 > URL: https://issues.apache.org/jira/browse/FLINK-35520 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Priority: Blocker > > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45808 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Comment Edited] (FLINK-35520) Nightly build can't compile as license check failed
[ https://issues.apache.org/jira/browse/FLINK-35520?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17851983#comment-17851983 ] Weijie Guo edited comment on FLINK-35520 at 6/4/24 9:53 AM: Hi [~antoniovespoli], would you mind taking a look? I found this issue after FLINK-35282 merged. was (Author: weijie guo): Hi [~antoniovespoli], would you mind taking a look? I found this issue after 91a9e06d merged. > Nightly build can't compile as license check failed > --- > > Key: FLINK-35520 > URL: https://issues.apache.org/jira/browse/FLINK-35520 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Priority: Blocker > > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45808 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-35520) master can't compile as license check failed
[ https://issues.apache.org/jira/browse/FLINK-35520?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Weijie Guo updated FLINK-35520: --- Summary: master can't compile as license check failed (was: Nightly build can't compile as license check failed) > master can't compile as license check failed > > > Key: FLINK-35520 > URL: https://issues.apache.org/jira/browse/FLINK-35520 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Priority: Blocker > > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45808 -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-35435] Add timeout Configuration to Async Sink [flink]
vahmed-hamdy commented on PR #24839: URL: https://github.com/apache/flink/pull/24839#issuecomment-2147156792 @flinkbot run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-35517) CI pipeline triggered by pull request seems unstable
[ https://issues.apache.org/jira/browse/FLINK-35517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17851994#comment-17851994 ] Jing Ge commented on FLINK-35517: - It should work again. [~Weijie Guo] please let me know if you still have any issues. Thanks! > CI pipeline triggered by pull request seems unstable > > > Key: FLINK-35517 > URL: https://issues.apache.org/jira/browse/FLINK-35517 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Assignee: Jing Ge >Priority: Blocker > > Flink CI pipeline triggered by pull request seems sort of unstable. > For example, https://github.com/apache/flink/pull/24883 was filed 15 hours > ago, but CI report is UNKNOWN. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-35517) CI pipeline triggered by pull request seems unstable
[ https://issues.apache.org/jira/browse/FLINK-35517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17851995#comment-17851995 ] Rui Fan commented on FLINK-35517: - Hi [~jingge] , may I know whether all old PRs can be recovered as well? Or only new PR works now? > CI pipeline triggered by pull request seems unstable > > > Key: FLINK-35517 > URL: https://issues.apache.org/jira/browse/FLINK-35517 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Assignee: Jing Ge >Priority: Blocker > > Flink CI pipeline triggered by pull request seems sort of unstable. > For example, https://github.com/apache/flink/pull/24883 was filed 15 hours > ago, but CI report is UNKNOWN. -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-20398][e2e] Migrate test_batch_sql.sh to Java e2e tests framework [flink]
JingGe commented on code in PR #24471: URL: https://github.com/apache/flink/pull/24471#discussion_r1624540081 ## flink-end-to-end-tests/flink-batch-sql-test/src/test/java/org/apache/flink/sql/tests/Generator.java: ## @@ -0,0 +1,82 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.sql.tests; + +import org.apache.flink.connector.testframe.source.FromElementsSource; +import org.apache.flink.table.data.GenericRowData; +import org.apache.flink.table.data.RowData; +import org.apache.flink.table.data.StringData; +import org.apache.flink.table.data.TimestampData; +import org.apache.flink.types.Row; + +import java.time.Instant; +import java.time.LocalDateTime; +import java.time.ZoneOffset; +import java.util.ArrayList; +import java.util.List; + +class Generator implements FromElementsSource.ElementsSupplier { +private static final long serialVersionUID = -8455653458083514261L; +private final List elements; + +static Generator create( +int numKeys, float rowsPerKeyAndSecond, int durationSeconds, int offsetSeconds) { +final int stepMs = (int) (1000 / rowsPerKeyAndSecond); +final long durationMs = durationSeconds * 1000L; +final long offsetMs = offsetSeconds * 2000L; +final List elements = new ArrayList<>(); +int keyIndex = 0; +long ms = 0; +while (ms < durationMs) { +elements.add(createRow(keyIndex++, ms, offsetMs)); Review Comment: The original implementation was only for batch in `BatchSQLTestProgram`. This PR is for the migration that should not implicitly bring big change for the data generation that might cause performance issue later. In addition, the new implementation is still in the `flink-batch-sql-test` module which should be used only for batch. Not sure if there are already similar generators in the stream-sql-test. If not, a new jira task could created and add the new implementation to the stream sql test. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [FLINK-20398][e2e] Migrate test_batch_sql.sh to Java e2e tests framework [flink]
JingGe commented on code in PR #24471: URL: https://github.com/apache/flink/pull/24471#discussion_r1624540081 ## flink-end-to-end-tests/flink-batch-sql-test/src/test/java/org/apache/flink/sql/tests/Generator.java: ## @@ -0,0 +1,82 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.sql.tests; + +import org.apache.flink.connector.testframe.source.FromElementsSource; +import org.apache.flink.table.data.GenericRowData; +import org.apache.flink.table.data.RowData; +import org.apache.flink.table.data.StringData; +import org.apache.flink.table.data.TimestampData; +import org.apache.flink.types.Row; + +import java.time.Instant; +import java.time.LocalDateTime; +import java.time.ZoneOffset; +import java.util.ArrayList; +import java.util.List; + +class Generator implements FromElementsSource.ElementsSupplier { +private static final long serialVersionUID = -8455653458083514261L; +private final List elements; + +static Generator create( +int numKeys, float rowsPerKeyAndSecond, int durationSeconds, int offsetSeconds) { +final int stepMs = (int) (1000 / rowsPerKeyAndSecond); +final long durationMs = durationSeconds * 1000L; +final long offsetMs = offsetSeconds * 2000L; +final List elements = new ArrayList<>(); +int keyIndex = 0; +long ms = 0; +while (ms < durationMs) { +elements.add(createRow(keyIndex++, ms, offsetMs)); Review Comment: The original implementation was only for batch in `BatchSQLTestProgram`. This PR is for the migration that should not implicitly bring big change for the data generation that might cause performance issue later. In addition, the new implementation is still in the `flink-batch-sql-test` module which should be used only for batch. Not sure if there are already similar generators in the stream-sql-test. If not, a new jira task could be created and add the new generator implementation to the stream sql test. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [FLINK-20398][e2e] Migrate test_batch_sql.sh to Java e2e tests framework [flink]
JingGe commented on code in PR #24471: URL: https://github.com/apache/flink/pull/24471#discussion_r1624540081 ## flink-end-to-end-tests/flink-batch-sql-test/src/test/java/org/apache/flink/sql/tests/Generator.java: ## @@ -0,0 +1,82 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.sql.tests; + +import org.apache.flink.connector.testframe.source.FromElementsSource; +import org.apache.flink.table.data.GenericRowData; +import org.apache.flink.table.data.RowData; +import org.apache.flink.table.data.StringData; +import org.apache.flink.table.data.TimestampData; +import org.apache.flink.types.Row; + +import java.time.Instant; +import java.time.LocalDateTime; +import java.time.ZoneOffset; +import java.util.ArrayList; +import java.util.List; + +class Generator implements FromElementsSource.ElementsSupplier { +private static final long serialVersionUID = -8455653458083514261L; +private final List elements; + +static Generator create( +int numKeys, float rowsPerKeyAndSecond, int durationSeconds, int offsetSeconds) { +final int stepMs = (int) (1000 / rowsPerKeyAndSecond); +final long durationMs = durationSeconds * 1000L; +final long offsetMs = offsetSeconds * 2000L; +final List elements = new ArrayList<>(); +int keyIndex = 0; +long ms = 0; +while (ms < durationMs) { +elements.add(createRow(keyIndex++, ms, offsetMs)); Review Comment: The original implementation was only for batch in `BatchSQLTestProgram`. This PR is for the migration that should not implicitly bring big change for the data generation that might cause performance issue later. In addition, the new implementation is still in the `flink-batch-sql-test` module which should be used only for batch. Not sure if there are already similar generators in the stream-sql-test. If not, a new jira task could be created and add the new implementation to the stream sql test. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Comment Edited] (FLINK-35517) CI pipeline triggered by pull request seems unstable
[ https://issues.apache.org/jira/browse/FLINK-35517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17851995#comment-17851995 ] Rui Fan edited comment on FLINK-35517 at 6/4/24 10:52 AM: -- Hi [~jingge] , thanks for the quick fix. :) May I know whether all old PRs can be recovered as well? Or only new PR works now? was (Author: fanrui): Hi [~jingge] , may I know whether all old PRs can be recovered as well? Or only new PR works now? > CI pipeline triggered by pull request seems unstable > > > Key: FLINK-35517 > URL: https://issues.apache.org/jira/browse/FLINK-35517 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Assignee: Jing Ge >Priority: Blocker > > Flink CI pipeline triggered by pull request seems sort of unstable. > For example, https://github.com/apache/flink/pull/24883 was filed 15 hours > ago, but CI report is UNKNOWN. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (FLINK-32229) Implement metrics and logging for Initial implementation
[ https://issues.apache.org/jira/browse/FLINK-32229?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Hong Liang Teoh reassigned FLINK-32229: --- Assignee: Burak Ozakinci > Implement metrics and logging for Initial implementation > > > Key: FLINK-32229 > URL: https://issues.apache.org/jira/browse/FLINK-32229 > Project: Flink > Issue Type: Sub-task >Reporter: Hong Liang Teoh >Assignee: Burak Ozakinci >Priority: Major > > Add/Ensure Kinesis specific metrics for MillisBehindLatest/numRecordsIn are > published. > More metrics here: > [https://cwiki.apache.org/confluence/display/FLINK/FLIP-33%3A+Standardize+Connector+Metrics] -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Reopened] (FLINK-35282) PyFlink Support for Apache Beam > 2.49
[ https://issues.apache.org/jira/browse/FLINK-35282?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sergey Nuyanzin reopened FLINK-35282: - I noticed that it was reverted by [~hong] at [caa68c2481f7c483d0364206d36654af26f2074f|https://github.com/apache/flink/commit/caa68c2481f7c483d0364206d36654af26f2074f] for that reason it would make sense to reopen the issue as well BTW I have a fix for license check issue at https://github.com/snuyanzin/flink/commit/5a4f4d0eb785050552c73fbfc74214f85ee278b0 We could try is once build becomes more stable > PyFlink Support for Apache Beam > 2.49 > -- > > Key: FLINK-35282 > URL: https://issues.apache.org/jira/browse/FLINK-35282 > Project: Flink > Issue Type: Improvement > Components: API / Python >Affects Versions: 1.19.0, 1.18.1 >Reporter: APA >Assignee: Antonio Vespoli >Priority: Major > Labels: pull-request-available > Fix For: 1.20.0 > > > From what I see PyFlink still has the requirement of Apache Beam => 2.43.0 > and <= 2.49.0 which subsequently results in a requirement of PyArrow <= > 12.0.0. That keeps us exposed to > [https://nvd.nist.gov/vuln/detail/CVE-2023-47248] > I'm not deep enough familiar with the PyFlink code base to understand why > Apache Beam's upper dependency limit can't be lifted. From all the existing > issues I haven't seen one addressing this. Therefore I created one now. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-32229) Implement metrics and logging for Initial implementation
[ https://issues.apache.org/jira/browse/FLINK-32229?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17851998#comment-17851998 ] Hong Liang Teoh commented on FLINK-32229: - [~chalixar] sorry I didn't manage to see this, and [~burakoz] has requested to work on this. Could this be worked out between you and [~burakoz] ? Maybe we can collaborate by reviewing the PR? > Implement metrics and logging for Initial implementation > > > Key: FLINK-32229 > URL: https://issues.apache.org/jira/browse/FLINK-32229 > Project: Flink > Issue Type: Sub-task >Reporter: Hong Liang Teoh >Assignee: Burak Ozakinci >Priority: Major > > Add/Ensure Kinesis specific metrics for MillisBehindLatest/numRecordsIn are > published. > More metrics here: > [https://cwiki.apache.org/confluence/display/FLINK/FLIP-33%3A+Standardize+Connector+Metrics] -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Closed] (FLINK-35518) CI Bot doesn't run on PRs - status UNKNOWN
[ https://issues.apache.org/jira/browse/FLINK-35518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jing Ge closed FLINK-35518. --- Resolution: Duplicate > CI Bot doesn't run on PRs - status UNKNOWN > -- > > Key: FLINK-35518 > URL: https://issues.apache.org/jira/browse/FLINK-35518 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Piotr Nowojski >Assignee: Jing Ge >Priority: Critical > > Doesn't want to run on my PR/branch. I was doing force-pushes, rebases, > asking flink bot to run, closed and opened new PR, but nothing helped > https://github.com/apache/flink/pull/24868 > https://github.com/apache/flink/pull/24883 > I've heard others were having similar problems recently. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-35520) master can't compile as license check failed
[ https://issues.apache.org/jira/browse/FLINK-35520?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17852001#comment-17852001 ] Sergey Nuyanzin commented on FLINK-35520: - so far it was reverted as mentioned in comments FLINK-35282 PS: I have a fix for this at https://github.com/snuyanzin/flink/commit/5a4f4d0eb785050552c73fbfc74214f85ee278b0 which could be tried when build will be more stable > master can't compile as license check failed > > > Key: FLINK-35520 > URL: https://issues.apache.org/jira/browse/FLINK-35520 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Priority: Blocker > > https://dev.azure.com/apache-flink/apache-flink/_build/results?buildId=60024&view=logs&j=52b61abe-a3cc-5bde-cc35-1bbe89bb7df5&t=54421a62-0c80-5aad-3319-094ff69180bb&l=45808 -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Assigned] (FLINK-35518) CI Bot doesn't run on PRs - status UNKNOWN
[ https://issues.apache.org/jira/browse/FLINK-35518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jing Ge reassigned FLINK-35518: --- Assignee: Jing Ge > CI Bot doesn't run on PRs - status UNKNOWN > -- > > Key: FLINK-35518 > URL: https://issues.apache.org/jira/browse/FLINK-35518 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Piotr Nowojski >Assignee: Jing Ge >Priority: Critical > > Doesn't want to run on my PR/branch. I was doing force-pushes, rebases, > asking flink bot to run, closed and opened new PR, but nothing helped > https://github.com/apache/flink/pull/24868 > https://github.com/apache/flink/pull/24883 > I've heard others were having similar problems recently. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Resolved] (FLINK-35517) CI pipeline triggered by pull request seems unstable
[ https://issues.apache.org/jira/browse/FLINK-35517?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jing Ge resolved FLINK-35517. - Fix Version/s: 1.20.0 Resolution: Fixed > CI pipeline triggered by pull request seems unstable > > > Key: FLINK-35517 > URL: https://issues.apache.org/jira/browse/FLINK-35517 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Assignee: Jing Ge >Priority: Blocker > Fix For: 1.20.0 > > > Flink CI pipeline triggered by pull request seems sort of unstable. > For example, https://github.com/apache/flink/pull/24883 was filed 15 hours > ago, but CI report is UNKNOWN. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-35517) CI pipeline triggered by pull request seems unstable
[ https://issues.apache.org/jira/browse/FLINK-35517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17852002#comment-17852002 ] Jing Ge commented on FLINK-35517: - [~fanrui] all PRs will be recovered. > CI pipeline triggered by pull request seems unstable > > > Key: FLINK-35517 > URL: https://issues.apache.org/jira/browse/FLINK-35517 > Project: Flink > Issue Type: Bug > Components: Build System / CI >Affects Versions: 1.20.0 >Reporter: Weijie Guo >Assignee: Jing Ge >Priority: Blocker > Fix For: 1.20.0 > > > Flink CI pipeline triggered by pull request seems sort of unstable. > For example, https://github.com/apache/flink/pull/24883 was filed 15 hours > ago, but CI report is UNKNOWN. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-35512) ArtifactFetchManagerTest unit tests fail
[ https://issues.apache.org/jira/browse/FLINK-35512?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17852005#comment-17852005 ] Elphas Toringepi commented on FLINK-35512: -- [~hong] +1, I reproduced the error by running {noformat} ./mvnw clean package{noformat} > ArtifactFetchManagerTest unit tests fail > > > Key: FLINK-35512 > URL: https://issues.apache.org/jira/browse/FLINK-35512 > Project: Flink > Issue Type: Technical Debt >Affects Versions: 1.19.1 >Reporter: Hong Liang Teoh >Priority: Major > Fix For: 1.19.1 > > > The below three tests from *ArtifactFetchManagerTest* seem to fail > consistently: > * ArtifactFetchManagerTest.testFileSystemFetchWithAdditionalUri > * ArtifactFetchManagerTest.testMixedArtifactFetch > * ArtifactFetchManagerTest.testHttpFetch > The error printed is > {code:java} > java.lang.AssertionError: > Expecting actual not to be empty > at > org.apache.flink.client.program.artifact.ArtifactFetchManagerTest.getFlinkClientsJar(ArtifactFetchManagerTest.java:248) > at > org.apache.flink.client.program.artifact.ArtifactFetchManagerTest.testMixedArtifactFetch(ArtifactFetchManagerTest.java:146) > at java.lang.reflect.Method.invoke(Method.java:498) > at java.util.concurrent.RecursiveAction.exec(RecursiveAction.java:189) > at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289) > at > java.util.concurrent.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1056) > at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1692) > at > java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:175) > {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[PR] [FLINK-35516][Connector/Files] Update the Experimental annotation for files connector [flink]
RocMarshal opened a new pull request, #24885: URL: https://github.com/apache/flink/pull/24885 ## What is the purpose of the change Update the Experimental annotation for files connector ## Brief change log Update the Experimental annotation for files connector ## Verifying this change This change is already covered by existing tests ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (yes / **no**) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (yes / **no**) - The serializers: (yes / **no** / don't know) - The runtime per-record code paths (performance sensitive): (yes / **no** / don't know) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (yes / **no** / don't know) - The S3 file system connector: (yes / **no** / don't know) ## Documentation - Does this pull request introduce a new feature? (yes / **no**) - If yes, how is the feature documented? (**not applicable** / docs / JavaDocs / not documented) -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-35516) Update the Experimental annotation to PublicEvolving for files connector
[ https://issues.apache.org/jira/browse/FLINK-35516?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-35516: --- Labels: pull-request-available (was: ) > Update the Experimental annotation to PublicEvolving for files connector > - > > Key: FLINK-35516 > URL: https://issues.apache.org/jira/browse/FLINK-35516 > Project: Flink > Issue Type: Technical Debt > Components: Connectors / FileSystem >Reporter: RocMarshal >Assignee: RocMarshal >Priority: Minor > Labels: pull-request-available > > as described in https://issues.apache.org/jira/browse/FLINK-35496 > We should update the annotations for the stable APIs in files connector. -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-35516][Connector/Files] Update the Experimental annotation for files connector [flink]
flinkbot commented on PR #24885: URL: https://github.com/apache/flink/pull/24885#issuecomment-2147260985 ## CI report: * d303ba9557f4cbef16b70565a97ba35b829f9bc6 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [FLINK-26050] Manually compact small SST files [flink]
rkhachatryan merged PR #24880: URL: https://github.com/apache/flink/pull/24880 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Closed] (FLINK-35506) disable kafka auto-commit and rely on flink’s checkpointing if both are enabled
[ https://issues.apache.org/jira/browse/FLINK-35506?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] elon_X closed FLINK-35506. -- Resolution: Not A Problem > disable kafka auto-commit and rely on flink’s checkpointing if both are > enabled > --- > > Key: FLINK-35506 > URL: https://issues.apache.org/jira/browse/FLINK-35506 > Project: Flink > Issue Type: Improvement > Components: Connectors / Kafka >Affects Versions: 1.16.1 >Reporter: elon_X >Priority: Major > Attachments: image-2024-06-03-23-39-28-270.png > > > When I use KafkaSource for consuming topics and set the Kafka parameter > {{{}enable.auto.commit=true{}}}, while also enabling checkpointing for the > task, I notice that both will commit offsets. Should Kafka's auto-commit be > disabled when enabling Flink checkpointing, similar to how it's done with > FlinkKafkaConsumer? > > *How to reproduce* > > {code:java} > // code placeholder > Properties kafkaParams = new Properties(); > kafkaParams.put("enable.auto.commit", "true"); > kafkaParams.put("auto.offset.reset", "latest"); > kafkaParams.put("fetch.min.bytes", "4096"); > kafkaParams.put("sasl.mechanism", "PLAIN"); > kafkaParams.put("security.protocol", "SASL_PLAINTEXT"); > kafkaParams.put("bootstrap.servers", bootStrap); > kafkaParams.put("group.id", expoGroupId); > kafkaParams.put("sasl.jaas.config","org.apache.kafka.common.security.plain.PlainLoginModule > required username=\"" + username + "\" password=\"" + password + "\";"); > KafkaSource source = KafkaSource > .builder() > .setBootstrapServers(bootStrap) > .setProperties(kafkaParams) > .setGroupId(expoGroupId) > .setTopics(Arrays.asList(expoTopic)) > .setValueOnlyDeserializer(new SimpleStringSchema()) > .setStartingOffsets(OffsetsInitializer.latest()) > .build(); > StreamExecutionEnvironment env = > StreamExecutionEnvironment.getExecutionEnvironment(); > env.fromSource(source, WatermarkStrategy.noWatermarks(), "kafka-source") > .filter(r -> true); > env.enableCheckpointing(3000 * 1000); > env.getCheckpointConfig().setMinPauseBetweenCheckpoints(1000 * 1000); > env.getCheckpointConfig().setCheckpointingMode(CheckpointingMode.AT_LEAST_ONCE); > env.getCheckpointConfig().setCheckpointTimeout(1000 * 300); > env.execute("kafka-consumer"); {code} > > > the kafka client's > org.apache.kafka.clients.consumer.internals.ConsumerCoordinator continuously > committing offsets. > !image-2024-06-03-23-39-28-270.png! -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-32081][checkpoint] Compatibility between file-merging on and off across job runs [flink]
ljz2051 commented on code in PR #24873: URL: https://github.com/apache/flink/pull/24873#discussion_r1625978087 ## flink-tests/src/test/java/org/apache/flink/test/checkpointing/SnapshotFileMergingCompatibilityITCase.java: ## @@ -0,0 +1,126 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.test.checkpointing; + +import org.apache.flink.configuration.CheckpointingOptions; +import org.apache.flink.configuration.Configuration; +import org.apache.flink.contrib.streaming.state.EmbeddedRocksDBStateBackend; +import org.apache.flink.core.execution.RestoreMode; +import org.apache.flink.runtime.testutils.MiniClusterResourceConfiguration; +import org.apache.flink.test.util.MiniClusterWithClientResource; +import org.apache.flink.util.TestLogger; + +import org.junit.jupiter.api.io.TempDir; +import org.junit.jupiter.params.ParameterizedTest; +import org.junit.jupiter.params.provider.MethodSource; + +import java.nio.file.Path; +import java.util.Arrays; +import java.util.Collection; + +import static org.apache.flink.test.checkpointing.ResumeCheckpointManuallyITCase.runJobAndGetExternalizedCheckpoint; +import static org.assertj.core.api.Assertions.assertThat; + +/** + * FileMerging Compatibility IT case which tests recovery from a checkpoint created in different + * fileMerging mode (i.e. fileMerging enabled/disabled). + */ +public class SnapshotFileMergingCompatibilityITCase extends TestLogger { + +public static Collection parameters() { +return Arrays.asList( +new Object[][] { +{RestoreMode.CLAIM, true}, +{RestoreMode.CLAIM, false}, +{RestoreMode.NO_CLAIM, true}, +{RestoreMode.NO_CLAIM, false} +}); +} + +@ParameterizedTest(name = "RestoreMode = {0}, fileMergingAcrossBoundary = {1}") +@MethodSource("parameters") +public void testSwitchFromDisablingToEnablingFileMerging( +RestoreMode restoreMode, boolean fileMergingAcrossBoundary, @TempDir Path checkpointDir) +throws Exception { +testSwitchingFileMerging( +checkpointDir, false, true, restoreMode, fileMergingAcrossBoundary); +} + +@ParameterizedTest(name = "RestoreMode = {0}, fileMergingAcrossBoundary = {1}") +@MethodSource("parameters") +public void testSwitchFromEnablingToDisablingFileMerging( +RestoreMode restoreMode, boolean fileMergingAcrossBoundary, @TempDir Path checkpointDir) +throws Exception { +testSwitchingFileMerging( +checkpointDir, true, false, restoreMode, fileMergingAcrossBoundary); +} + +private void testSwitchingFileMerging( +Path checkpointDir, +boolean firstFileMergingSwitch, +boolean secondFileMergingSwitch, +RestoreMode restoreMode, +boolean fileMergingAcrossBoundary) +throws Exception { +final Configuration config = new Configuration(); +config.set(CheckpointingOptions.CHECKPOINTS_DIRECTORY, checkpointDir.toUri().toString()); +config.set(CheckpointingOptions.INCREMENTAL_CHECKPOINTS, true); +config.set(CheckpointingOptions.FILE_MERGING_ACROSS_BOUNDARY, fileMergingAcrossBoundary); +config.set(CheckpointingOptions.FILE_MERGING_ENABLED, firstFileMergingSwitch); +MiniClusterWithClientResource firstCluster = +new MiniClusterWithClientResource( +new MiniClusterResourceConfiguration.Builder() +.setConfiguration(config) +.setNumberTaskManagers(2) +.setNumberSlotsPerTaskManager(2) +.build()); +EmbeddedRocksDBStateBackend stateBackend1 = new EmbeddedRocksDBStateBackend(); +stateBackend1.configure(config, Thread.currentThread().getContextClassLoader()); +firstCluster.before(); +String externalCheckpoint; +try { +externalCheckpoint = +runJobAndGetExternalizedCheckpoint( +stateBackend1, null, firstCluster, restoreM
Re: [PR] [FLINK-32081][checkpoint] Compatibility between file-merging on and off across job runs [flink]
ljz2051 commented on code in PR #24873: URL: https://github.com/apache/flink/pull/24873#discussion_r1625978087 ## flink-tests/src/test/java/org/apache/flink/test/checkpointing/SnapshotFileMergingCompatibilityITCase.java: ## @@ -0,0 +1,126 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.test.checkpointing; + +import org.apache.flink.configuration.CheckpointingOptions; +import org.apache.flink.configuration.Configuration; +import org.apache.flink.contrib.streaming.state.EmbeddedRocksDBStateBackend; +import org.apache.flink.core.execution.RestoreMode; +import org.apache.flink.runtime.testutils.MiniClusterResourceConfiguration; +import org.apache.flink.test.util.MiniClusterWithClientResource; +import org.apache.flink.util.TestLogger; + +import org.junit.jupiter.api.io.TempDir; +import org.junit.jupiter.params.ParameterizedTest; +import org.junit.jupiter.params.provider.MethodSource; + +import java.nio.file.Path; +import java.util.Arrays; +import java.util.Collection; + +import static org.apache.flink.test.checkpointing.ResumeCheckpointManuallyITCase.runJobAndGetExternalizedCheckpoint; +import static org.assertj.core.api.Assertions.assertThat; + +/** + * FileMerging Compatibility IT case which tests recovery from a checkpoint created in different + * fileMerging mode (i.e. fileMerging enabled/disabled). + */ +public class SnapshotFileMergingCompatibilityITCase extends TestLogger { + +public static Collection parameters() { +return Arrays.asList( +new Object[][] { +{RestoreMode.CLAIM, true}, +{RestoreMode.CLAIM, false}, +{RestoreMode.NO_CLAIM, true}, +{RestoreMode.NO_CLAIM, false} +}); +} + +@ParameterizedTest(name = "RestoreMode = {0}, fileMergingAcrossBoundary = {1}") +@MethodSource("parameters") +public void testSwitchFromDisablingToEnablingFileMerging( +RestoreMode restoreMode, boolean fileMergingAcrossBoundary, @TempDir Path checkpointDir) +throws Exception { +testSwitchingFileMerging( +checkpointDir, false, true, restoreMode, fileMergingAcrossBoundary); +} + +@ParameterizedTest(name = "RestoreMode = {0}, fileMergingAcrossBoundary = {1}") +@MethodSource("parameters") +public void testSwitchFromEnablingToDisablingFileMerging( +RestoreMode restoreMode, boolean fileMergingAcrossBoundary, @TempDir Path checkpointDir) +throws Exception { +testSwitchingFileMerging( +checkpointDir, true, false, restoreMode, fileMergingAcrossBoundary); +} + +private void testSwitchingFileMerging( +Path checkpointDir, +boolean firstFileMergingSwitch, +boolean secondFileMergingSwitch, +RestoreMode restoreMode, +boolean fileMergingAcrossBoundary) +throws Exception { +final Configuration config = new Configuration(); +config.set(CheckpointingOptions.CHECKPOINTS_DIRECTORY, checkpointDir.toUri().toString()); +config.set(CheckpointingOptions.INCREMENTAL_CHECKPOINTS, true); +config.set(CheckpointingOptions.FILE_MERGING_ACROSS_BOUNDARY, fileMergingAcrossBoundary); +config.set(CheckpointingOptions.FILE_MERGING_ENABLED, firstFileMergingSwitch); +MiniClusterWithClientResource firstCluster = +new MiniClusterWithClientResource( +new MiniClusterResourceConfiguration.Builder() +.setConfiguration(config) +.setNumberTaskManagers(2) +.setNumberSlotsPerTaskManager(2) +.build()); +EmbeddedRocksDBStateBackend stateBackend1 = new EmbeddedRocksDBStateBackend(); +stateBackend1.configure(config, Thread.currentThread().getContextClassLoader()); +firstCluster.before(); +String externalCheckpoint; +try { +externalCheckpoint = +runJobAndGetExternalizedCheckpoint( +stateBackend1, null, firstCluster, restoreM
Re: [PR] [FLINK-35353][docs-zh]Translate "Profiler" page into Chinese [flink]
drymatini commented on PR #24822: URL: https://github.com/apache/flink/pull/24822#issuecomment-2147503847 Hi @JingGe , thank you for the review, I have already made adjustment based on your opinion, please check my commit again. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [FLINK-35475][runtime] Introduce isInternalSorterSupport to OperatorAttributes [flink]
jeyhunkarimov commented on PR #24874: URL: https://github.com/apache/flink/pull/24874#issuecomment-2147546699 Thanks @Sxnan LGTM! -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[PR] [hotfix][docs] Fix typo in Log Example Given on upgrade.md File [flink-kubernetes-operator]
nacisimsek opened a new pull request, #835: URL: https://github.com/apache/flink-kubernetes-operator/pull/835 The ID in the savepoint folder name that is passed as a parameter, and the ID in the folder name that is given as the log example do NOT match. Expected ID in log example: `aec3dd08e76d` Actual ID given in the log example: `2f40a9c8e4b9` -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [FLINK-35501] Use common IO thread pool for RocksDB data transfer [flink]
rkhachatryan commented on PR #24882: URL: https://github.com/apache/flink/pull/24882#issuecomment-2147612702 @flinkbot run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [FLINK-20398][e2e] Migrate test_batch_sql.sh to Java e2e tests framework [flink]
XComp commented on code in PR #24471: URL: https://github.com/apache/flink/pull/24471#discussion_r1626077906 ## flink-end-to-end-tests/flink-batch-sql-test/src/test/java/org/apache/flink/sql/tests/Generator.java: ## @@ -0,0 +1,82 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, software + * distributed under the License is distributed on an "AS IS" BASIS, + * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + * See the License for the specific language governing permissions and + * limitations under the License. + */ + +package org.apache.flink.sql.tests; + +import org.apache.flink.connector.testframe.source.FromElementsSource; +import org.apache.flink.table.data.GenericRowData; +import org.apache.flink.table.data.RowData; +import org.apache.flink.table.data.StringData; +import org.apache.flink.table.data.TimestampData; +import org.apache.flink.types.Row; + +import java.time.Instant; +import java.time.LocalDateTime; +import java.time.ZoneOffset; +import java.util.ArrayList; +import java.util.List; + +class Generator implements FromElementsSource.ElementsSupplier { +private static final long serialVersionUID = -8455653458083514261L; +private final List elements; + +static Generator create( +int numKeys, float rowsPerKeyAndSecond, int durationSeconds, int offsetSeconds) { +final int stepMs = (int) (1000 / rowsPerKeyAndSecond); +final long durationMs = durationSeconds * 1000L; +final long offsetMs = offsetSeconds * 2000L; +final List elements = new ArrayList<>(); +int keyIndex = 0; +long ms = 0; +while (ms < durationMs) { +elements.add(createRow(keyIndex++, ms, offsetMs)); Review Comment: Jing actually has a good point on the memory consumption. I missed that one. 👍 We should continue generating the records on-the-fly to be closer to what the original test did. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-35282) PyFlink Support for Apache Beam > 2.49
[ https://issues.apache.org/jira/browse/FLINK-35282?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17852074#comment-17852074 ] Hong Liang Teoh commented on FLINK-35282: - Yes. Reopned JIRA > PyFlink Support for Apache Beam > 2.49 > -- > > Key: FLINK-35282 > URL: https://issues.apache.org/jira/browse/FLINK-35282 > Project: Flink > Issue Type: Improvement > Components: API / Python >Affects Versions: 1.19.0, 1.18.1 >Reporter: APA >Assignee: Antonio Vespoli >Priority: Major > Labels: pull-request-available > Fix For: 1.20.0 > > > From what I see PyFlink still has the requirement of Apache Beam => 2.43.0 > and <= 2.49.0 which subsequently results in a requirement of PyArrow <= > 12.0.0. That keeps us exposed to > [https://nvd.nist.gov/vuln/detail/CVE-2023-47248] > I'm not deep enough familiar with the PyFlink code base to understand why > Apache Beam's upper dependency limit can't be lifted. From all the existing > issues I haven't seen one addressing this. Therefore I created one now. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] (FLINK-35282) PyFlink Support for Apache Beam > 2.49
[ https://issues.apache.org/jira/browse/FLINK-35282 ] Hong Liang Teoh deleted comment on FLINK-35282: - was (Author: hong): Yes. Reopned JIRA > PyFlink Support for Apache Beam > 2.49 > -- > > Key: FLINK-35282 > URL: https://issues.apache.org/jira/browse/FLINK-35282 > Project: Flink > Issue Type: Improvement > Components: API / Python >Affects Versions: 1.19.0, 1.18.1 >Reporter: APA >Assignee: Antonio Vespoli >Priority: Major > Labels: pull-request-available > Fix For: 1.20.0 > > > From what I see PyFlink still has the requirement of Apache Beam => 2.43.0 > and <= 2.49.0 which subsequently results in a requirement of PyArrow <= > 12.0.0. That keeps us exposed to > [https://nvd.nist.gov/vuln/detail/CVE-2023-47248] > I'm not deep enough familiar with the PyFlink code base to understand why > Apache Beam's upper dependency limit can't be lifted. From all the existing > issues I haven't seen one addressing this. Therefore I created one now. -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-34487][ci] Adds Python Wheels nightly GHA workflow [flink]
XComp commented on code in PR #24426: URL: https://github.com/apache/flink/pull/24426#discussion_r1626117295 ## .github/workflows/nightly.yml: ## @@ -94,3 +94,46 @@ jobs: s3_bucket: ${{ secrets.IT_CASE_S3_BUCKET }} s3_access_key: ${{ secrets.IT_CASE_S3_ACCESS_KEY }} s3_secret_key: ${{ secrets.IT_CASE_S3_SECRET_KEY }} + + build_python_wheels: +name: "Build Python Wheels on ${{ matrix.os_name }}" +runs-on: ${{ matrix.os }} +strategy: + fail-fast: false + matrix: +include: + - os: ubuntu-latest +os_name: linux + - os: macos-latest +os_name: macos +steps: + - name: "Checkout the repository" +uses: actions/checkout@v4 +with: + fetch-depth: 0 + persist-credentials: false + - name: "Stringify workflow name" +uses: "./.github/actions/stringify" +id: stringify_workflow +with: + value: ${{ github.workflow }} + - name: "Build python wheels for ${{ matrix.os_name }}" +uses: pypa/cibuildwheel@v2.16.5 Review Comment: Looks fine from my side. But I am not familiar with the whole buildwheel logic. @HuangXingBo can you do another pass over it and approve the changes once more? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [hotfix] A couple of hotfixes [flink]
JingGe commented on code in PR #24883: URL: https://github.com/apache/flink/pull/24883#discussion_r1626162425 ## flink-runtime/src/test/java/org/apache/flink/runtime/checkpoint/StateAssignmentOperationTest.java: ## @@ -563,7 +563,7 @@ private InflightDataGateOrPartitionRescalingDescriptor gate( } @Test -void testChannelStateAssignmentTwoGatesPartiallyDownscaling() +public void testChannelStateAssignmentTwoGatesPartiallyDownscaling() Review Comment: NIT: With Junit 5, methods are not required to be public. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [FLINK-35501] Use common IO thread pool for RocksDB data transfer [flink]
rkhachatryan commented on PR #24882: URL: https://github.com/apache/flink/pull/24882#issuecomment-2147786034 @flinkbot run azure -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [hotfix] A couple of hotfixes [flink]
pnowojski commented on code in PR #24883: URL: https://github.com/apache/flink/pull/24883#discussion_r1626244851 ## flink-runtime/src/test/java/org/apache/flink/runtime/checkpoint/StateAssignmentOperationTest.java: ## @@ -563,7 +563,7 @@ private InflightDataGateOrPartitionRescalingDescriptor gate( } @Test -void testChannelStateAssignmentTwoGatesPartiallyDownscaling() +public void testChannelStateAssignmentTwoGatesPartiallyDownscaling() Review Comment: Ahhh, that explains my confusion when I was cherry-picking some things between branches. Anyway, next time I will just skip this. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [hotfix] A couple of hotfixes [flink]
pnowojski merged PR #24883: URL: https://github.com/apache/flink/pull/24883 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Commented] (FLINK-31533) CREATE TABLE AS SELECT should support to define partition
[ https://issues.apache.org/jira/browse/FLINK-31533?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17852110#comment-17852110 ] Sergio Peña commented on FLINK-31533: - Hi [~luoyuxia] [~aitozi], I'd like to make some contribution to Flink to extend the CTAS statement to allow a custom schema definition (columns, partition & primary keys, watermarks, etc). I noticed this task and FLINK-31534 are meant for a subset of those changes. Are you still considering in working on this at some point? I see these comments were made a year ago, but I don't want to step on others people's work if there's some progress or interest on it. Are you ok if I take this task? I'm going to write a FLIP with the proposal about the semantics for the schema definition, which is similar to Mysql CTAS. > CREATE TABLE AS SELECT should support to define partition > - > > Key: FLINK-31533 > URL: https://issues.apache.org/jira/browse/FLINK-31533 > Project: Flink > Issue Type: Sub-task > Components: Table SQL / API >Reporter: luoyuxia >Priority: Major > -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Created] (FLINK-35521) Flink FileSystem SQL Connector Generating SUCESS File Multiple Times
EMERSON WANG created FLINK-35521: Summary: Flink FileSystem SQL Connector Generating SUCESS File Multiple Times Key: FLINK-35521 URL: https://issues.apache.org/jira/browse/FLINK-35521 Project: Flink Issue Type: Improvement Components: Connectors / FileSystem Affects Versions: 1.18.1 Environment: Our PyFlink SQL jobs are running in AWS EKS environment. Reporter: EMERSON WANG Our Flink table SQL job received data from the Kafka streams and then sinked all partitioned data into the associated parquet files under the same S3 folder through the filesystem SQL connector. For the S3 filesystem SQL connector, sink.partition-commit.policy.kind was set to 'success-file' and sink.partition-commit.trigger was set to 'partition-time'. We found that _SUCCESS file in the S3 folder was generated multiple times after multiple partitions are committed. Because all partitioned parquet files and _SUCCESS file are in the same S3 folder and _SUCCESS file is used to trigger the downstream application, we really like the _SUCCESS file to be generated only once instead of multiple times after all partitions are committed and all parquet files are ready to be processed. Thus, one _SUCCESS file can be used to trigger the downstream application only once instead of multiple times. We knew we could set sink.partition-commit.trigger to 'process-time' to generate _SUCCESS file only once in the S3 folder; however, 'process-time' would not meet our business requirements. We'd request the FileSystem SQL connector should support to the following new user case: Even if sink.partition-commit.trigger is set to 'partition-time', _SUCCESS file will be generated only once after all partitions are committed and all output files are ready to be processed, and will be used to trigger the downstream application only once instead of multiple times. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-35521) Flink FileSystem SQL Connector Generating SUCESS File Multiple Times
[ https://issues.apache.org/jira/browse/FLINK-35521?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] EMERSON WANG updated FLINK-35521: - Description: Our Flink table SQL job received data from the Kafka streams and then sinked all partitioned data into the associated parquet files under the same S3 folder through the filesystem SQL connector. For the S3 filesystem SQL connector, sink.partition-commit.policy.kind was set to 'success-file' and sink.partition-commit.trigger was set to 'partition-time'. We found that _SUCCESS file in the S3 folder was generated multiple times after multiple partitions are committed. Because all partitioned parquet files and _SUCCESS file are in the same S3 folder and _SUCCESS file is used to trigger the downstream application, we really like the _SUCCESS file to be generated only once instead of multiple times after all partitions are committed and all parquet files are ready to be processed. Thus, one _SUCCESS file can be used to trigger the downstream application only once instead of multiple times. We knew we could set sink.partition-commit.trigger to 'process-time' to generate _SUCCESS file only once in the S3 folder; however, 'process-time' would not meet our business requirements. We'd request the FileSystem SQL connector should support to the following new user case: Even if sink.partition-commit.trigger is set to 'partition-time', _SUCCESS file will be generated only once after all partitions are committed and all output files are ready to be processed, and will be used to trigger the downstream application only once instead of multiple times. was: Our Flink table SQL job received data from the Kafka streams and then sinked all partitioned data into the associated parquet files under the same S3 folder through the filesystem SQL connector. For the S3 filesystem SQL connector, sink.partition-commit.policy.kind was set to 'success-file' and sink.partition-commit.trigger was set to 'partition-time'. We found that _SUCCESS file in the S3 folder was generated multiple times after multiple partitions are committed. Because all partitioned parquet files and _SUCCESS file are in the same S3 folder and _SUCCESS file is used to trigger the downstream application, we really like the _SUCCESS file to be generated only once instead of multiple times after all partitions are committed and all parquet files are ready to be processed. Thus, one _SUCCESS file can be used to trigger the downstream application only once instead of multiple times. We knew we could set sink.partition-commit.trigger to 'process-time' to generate _SUCCESS file only once in the S3 folder; however, 'process-time' would not meet our business requirements. We'd request the FileSystem SQL connector should support to the following new user case: Even if sink.partition-commit.trigger is set to 'partition-time', _SUCCESS file will be generated only once after all partitions are committed and all output files are ready to be processed, and will be used to trigger the downstream application only once instead of multiple times. > Flink FileSystem SQL Connector Generating SUCESS File Multiple Times > > > Key: FLINK-35521 > URL: https://issues.apache.org/jira/browse/FLINK-35521 > Project: Flink > Issue Type: Improvement > Components: Connectors / FileSystem >Affects Versions: 1.18.1 > Environment: Our PyFlink SQL jobs are running in AWS EKS environment. >Reporter: EMERSON WANG >Priority: Major > > Our Flink table SQL job received data from the Kafka streams and then sinked > all partitioned data into the associated parquet files under the same S3 > folder through the filesystem SQL connector. > For the S3 filesystem SQL connector, sink.partition-commit.policy.kind was > set to 'success-file' and sink.partition-commit.trigger was set to > 'partition-time'. We found that _SUCCESS file in the S3 folder was generated > multiple times after multiple partitions are committed. > Because all partitioned parquet files and _SUCCESS file are in the same S3 > folder and _SUCCESS file is used to trigger the downstream application, we > really like the _SUCCESS file to be generated only once instead of multiple > times after all partitions are committed and all parquet files are ready to > be processed. Thus, one _SUCCESS file can be used to trigger the downstream > application only once instead of multiple times. > We knew we could set sink.partition-commit.trigger to 'process-time' to > generate _SUCCESS file only once in the S3 folder; however, 'process-time' > would not meet our business requirements. > We'd request the FileSystem SQL connector should support to the following new > user case: > Even if sink.partition-commit.trigger is set to 'part
[jira] [Updated] (FLINK-35515) Upgrade hive version to 4.0.0
[ https://issues.apache.org/jira/browse/FLINK-35515?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Martijn Visser updated FLINK-35515: --- Fix Version/s: (was: 1.18.2) > Upgrade hive version to 4.0.0 > - > > Key: FLINK-35515 > URL: https://issues.apache.org/jira/browse/FLINK-35515 > Project: Flink > Issue Type: Improvement > Components: Connectors / Hive >Affects Versions: 1.18.1 >Reporter: vikasap >Priority: Major > > Hive version 4.0.0 was released recently. However none of the major flink > versions will work with this. Filing this so that major flink version's > flink-sql and table api will be able to work with the new version of hive > metastore. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Updated] (FLINK-35515) Upgrade hive version to 4.0.0
[ https://issues.apache.org/jira/browse/FLINK-35515?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Martijn Visser updated FLINK-35515: --- Issue Type: New Feature (was: Improvement) > Upgrade hive version to 4.0.0 > - > > Key: FLINK-35515 > URL: https://issues.apache.org/jira/browse/FLINK-35515 > Project: Flink > Issue Type: New Feature > Components: Connectors / Hive >Affects Versions: 1.18.1 >Reporter: vikasap >Priority: Major > > Hive version 4.0.0 was released recently. However none of the major flink > versions will work with this. Filing this so that major flink version's > flink-sql and table api will be able to work with the new version of hive > metastore. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-24605) org.apache.flink.kafka.shaded.org.apache.kafka.clients.consumer.NoOffsetForPartitionException: Undefined offset with no reset policy for partitions
[ https://issues.apache.org/jira/browse/FLINK-24605?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17852128#comment-17852128 ] EMERSON WANG commented on FLINK-24605: -- We got the same exception when scan.startup.mode was set to 'group-offsets' and properties.auto.offset.reset was set to 'latest'. We had to work around as follows: To start the Flink job for the first time, we had to set scan.startup.mode to 'latest', let the job run for a few minutes, then stopped the job, reset scan.startup.mode to 'group-offsets' and finally restart the job. We'd appreciate it very much if you could resolve this ticket as soon as possible. > org.apache.flink.kafka.shaded.org.apache.kafka.clients.consumer.NoOffsetForPartitionException: > Undefined offset with no reset policy for partitions > --- > > Key: FLINK-24605 > URL: https://issues.apache.org/jira/browse/FLINK-24605 > Project: Flink > Issue Type: Bug > Components: Connectors / Kafka >Affects Versions: 1.14.0 >Reporter: Abhijit Talukdar >Priority: Major > > Getting below issue when using 'scan.startup.mode' = 'group-offsets'. > > WITH ( > 'connector' = 'kafka', > 'topic' = 'ss7gsm-signaling-event', > 'properties.bootstrap.servers' = '**:9093', > 'properties.group.id' = 'ss7gsm-signaling-event-T5', > 'value.format' = 'avro-confluent', > 'value.avro-confluent.schema-registry.url' = 'https://***:9099', > {color:#ff8b00}'scan.startup.mode' = 'group-offsets',{color} > {color:#ff8b00} 'properties.auto.offset.reset' = 'earliest',{color} > 'properties.security.protocol'= 'SASL_SSL', > 'properties.ssl.truststore.location'= '/*/*/ca-certs.jks', > 'properties.ssl.truststore.password'= '*', > 'properties.sasl.kerberos.service.name'= 'kafka' > ) > > 'ss7gsm-signaling-event-T5' is a new group id. If the group id is present in > ZK then it works otherwise getting below exception. > 'properties.auto.offset.reset' property is ignored. > > 021-10-20 22:18:28,267 INFO > org.apache.flink.kafka.shaded.org.apache.kafka.clients.consumer.ConsumerConfig > [] - ConsumerConfig values: 021-10-20 22:18:28,267 INFO > org.apache.flink.kafka.shaded.org.apache.kafka.clients.consumer.ConsumerConfig > [] - ConsumerConfig values: > allow.auto.create.topics = false > auto.commit.interval.ms = 5000 > {color:#FF} +*auto.offset.reset = none*+{color} > bootstrap.servers = [.xxx.com:9093] > > > Exception: > > 021-10-20 22:18:28,620 INFO > org.apache.flink.runtime.executiongraph.ExecutionGraph [] - Source: > KafkaSource-hiveSignaling.signaling_stg.ss7gsm_signaling_event_flink_k -> > Sink: Collect table sink (1/1) (89b175333242fab8914271ad7638ba92) switched > from INITIALIZING to RUNNING.021-10-20 22:18:28,620 INFO > org.apache.flink.runtime.executiongraph.ExecutionGraph [] - Source: > KafkaSource-hiveSignaling.signaling_stg.ss7gsm_signaling_event_flink_k -> > Sink: Collect table sink (1/1) (89b175333242fab8914271ad7638ba92) switched > from INITIALIZING to RUNNING.2021-10-20 22:18:28,621 INFO > org.apache.flink.connector.kafka.source.enumerator.KafkaSourceEnumerator [] - > Assigning splits to readers \{0=[[Partition: ss7gsm-signaling-event-2, > StartingOffset: -3, StoppingOffset: -9223372036854775808], [Partition: > ss7gsm-signaling-event-8, StartingOffset: -3, StoppingOffset: > -9223372036854775808], [Partition: ss7gsm-signaling-event-7, StartingOffset: > -3, StoppingOffset: -9223372036854775808], [Partition: > ss7gsm-signaling-event-9, StartingOffset: -3, StoppingOffset: > -9223372036854775808], [Partition: ss7gsm-signaling-event-5, StartingOffset: > -3, StoppingOffset: -9223372036854775808], [Partition: > ss7gsm-signaling-event-6, StartingOffset: -3, StoppingOffset: > -9223372036854775808], [Partition: ss7gsm-signaling-event-0, StartingOffset: > -3, StoppingOffset: -9223372036854775808], [Partition: > ss7gsm-signaling-event-4, StartingOffset: -3, StoppingOffset: > -9223372036854775808], [Partition: ss7gsm-signaling-event-1, StartingOffset: > -3, StoppingOffset: -9223372036854775808], [Partition: > ss7gsm-signaling-event-3, StartingOffset: -3, StoppingOffset: > -9223372036854775808]]}2021-10-20 22:18:28,716 INFO > org.apache.flink.runtime.executiongraph.ExecutionGraph [] - Source: > KafkaSource-hiveSignaling.signaling_stg.ss7gsm_signaling_event_flink_k -> > Sink: Collect table sink (1/1) (89b175333242fab8914271ad7638ba92) switched > from RUNNING to FAILED on xx.xxx.xxx.xxx:42075-d80607 @ xx.xxx.com > (dataPort=34120).java.lang.RuntimeException: One or more fetchers have > encountered exception at > org.apache.flink.connector.base.source.reader.fetcher.SplitFetcherManager.checkError
[jira] [Commented] (FLINK-35502) compress the checkpoint metadata generated by ZK/ETCD HA Services
[ https://issues.apache.org/jira/browse/FLINK-35502?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17852132#comment-17852132 ] Mingliang Liu commented on FLINK-35502: --- I guess in my day job I don't see user requests that need to recover from any point in the past days. I think it works just fine to recover from recent checkpoints in the past days. And compressing is a good improvement as data is getting large. > compress the checkpoint metadata generated by ZK/ETCD HA Services > - > > Key: FLINK-35502 > URL: https://issues.apache.org/jira/browse/FLINK-35502 > Project: Flink > Issue Type: Improvement >Reporter: Ying Z >Priority: Major > > In the implementation of Flink HA, the metadata of checkpoints is stored in > either Zookeeper (ZK HA) or ETCD (K8S HA), such as: > {code:java} > checkpointID-0036044: > checkpointID-0036045: > ... > ... {code} > However, neither of these are designed to store excessive amounts of data. If > the > [state.checkpoints.num-retained]([https://nightlies.apache.org/flink/flink-docs-release-1.19/zh/docs/deployment/config/#state-checkpoints-num-retained]) > setting is set too large, it can easily cause abnormalities in ZK/ETCD. > The error log when set state.checkpoints.num-retained to 1500: > {code:java} > Caused by: org.apache.flink.util.SerializedThrowable: > io.fabric8.kubernetes.client.KubernetesClientException: Failure executing: > PUT at: > https://xxx/api/v1/namespaces/default/configmaps/xxx-jobmanager-leader. > Message: ConfigMap "xxx-jobmanager-leader" is invalid: 0J: > Too long: must have at most 1048576 bytes. Received status: > Status(apiVersion=v1, code=422, > details=StatusDetails(causes=(StatusCause(field=[J, message=Too long: must > have at most 1048576 bytes, reason=FieldValueTooLong, > additionalProperties={})l, group=null, kind=ConfigMap, > name=xxx-jobmanager-leader, retryAfterSeconds=null, uid=null, > additionalProperties=(}), kind=Status, message=ConfigMap > "xxx-jobmanager-leader" is invalid: [): Too long: must have at most 1048576 > bytes, metadata=ListMeta(_continue=null, remainingItemCount=null, > resourceVersion=null, selfLink=null, additionalProperties={}), > reason=Invalid, status=Failure, additionalProperties=(}). {code} > In Flink's code, all checkpoint metadata are updated at the same time, and > The checkpoint metadata contains many repeated bytes, therefore it can > achieve a very good compression ratio. > Therefore, I suggest compressing the data when writing checkpoints and > decompressing it when reading, to reduce storage pressure and improve IO > efficiency. > Here is the sample code, and reduce the metadata size from 1M bytes to 30K. > {code:java} > // Map -> Json > ObjectMapper objectMapper = new ObjectMapper(); > String checkpointJson = objectMapper.writeValueAsString(checkpointMap); // // > copress and base64 > String compressedBase64 = compressAndEncode(checkpointJson); > compressedData.put("checkpoint-all", compressedBase64);{code} > {code:java} > private static String compressAndEncode(String data) throws IOException { > ByteArrayOutputStream outputStream = new ByteArrayOutputStream(); > try (GZIPOutputStream gzipOutputStream = new > GZIPOutputStream(outputStream)) > { gzipOutputStream.write(data.getBytes(StandardCharsets.UTF_8)); > } > byte[] compressedData = outputStream.toByteArray(); > return Base64.getEncoder().encodeToString(compressedData); > } {code} -- This message was sent by Atlassian Jira (v8.20.10#820010)
[PR] [FLINK-34172] Add support for altering a distribution via ALTER TABLE [flink]
jnh5y opened a new pull request, #24886: URL: https://github.com/apache/flink/pull/24886 ## What is the purpose of the change This PR implements the SQL parser changes for ALTER TABLE to support ADD, MODIFY, and DROP DISTRIBUTION statements. ## Brief change log The SQL Parser has been updated. The `AlterSchemaConverter` has been updated to pass the changes in a DISTRIBUTION through to the `Operation`. ## Verifying this change This change added tests and can be verified as follows: *(example:)* - *Added integration tests for end-to-end deployment with large payloads (100MB)* - *Extended integration test for recovery after master (JobManager) failure* - *Added test that validates that TaskInfo is transferred only once across recoveries* - *Manually verified the change by running a 4 node cluster with 2 JobManagers and 4 TaskManagers, a stateful streaming program, and killing one JobManager and two TaskManagers during the execution, verifying that recovery happens correctly.* ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): (no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: (no) - The serializers: (no) - The runtime per-record code paths (performance sensitive): (no) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (no) - The S3 file system connector: (no) ## Documentation - Does this pull request introduce a new feature? (yes) - If yes, how is the feature documented? (not applicable / docs / JavaDocs / not documented) -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-34172) Add support for altering a distribution via ALTER TABLE
[ https://issues.apache.org/jira/browse/FLINK-34172?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-34172: --- Labels: pull-request-available (was: ) > Add support for altering a distribution via ALTER TABLE > > > Key: FLINK-34172 > URL: https://issues.apache.org/jira/browse/FLINK-34172 > Project: Flink > Issue Type: Sub-task >Reporter: Jim Hughes >Assignee: Jim Hughes >Priority: Major > Labels: pull-request-available > Fix For: 1.20.0 > > -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-34172] Add support for altering a distribution via ALTER TABLE [flink]
flinkbot commented on PR #24886: URL: https://github.com/apache/flink/pull/24886#issuecomment-2148475342 ## CI report: * 4081f6d00def12b386970d94681364dd99d089a6 UNKNOWN Bot commands The @flinkbot bot supports the following commands: - `@flinkbot run azure` re-run the last Azure build -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [FLINK-35504] Improve Elasticsearch 8 connector observability [flink-connector-elasticsearch]
liuml07 commented on PR #106: URL: https://github.com/apache/flink-connector-elasticsearch/pull/106#issuecomment-2148526213 @reswqa Could you help review and merge? Thank you! -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Created] (FLINK-35522) The source task may get stuck after a failover occurs in batch jobs
xingbe created FLINK-35522: -- Summary: The source task may get stuck after a failover occurs in batch jobs Key: FLINK-35522 URL: https://issues.apache.org/jira/browse/FLINK-35522 Project: Flink Issue Type: Bug Components: Runtime / Coordination Affects Versions: 1.18.1, 1.19.0, 1.17.2, 1.20.0 Reporter: xingbe Fix For: 1.20.0 If the source task does not get assigned a split because the SplitEnumerator has no more splits, and a failover occurs during the closing process, the SourceCoordinatorContext will not resend the NoMoreSplit event to the newly started source task, causing the source vertex to remain stuck indefinitely. This case may only occur in batch jobs where speculative execution has been enabled. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-35522) The source task may get stuck after a failover occurs in batch jobs
[ https://issues.apache.org/jira/browse/FLINK-35522?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17852240#comment-17852240 ] xingbe commented on FLINK-35522: I have a solution to fix it, Could you please assign this ticket to me? [~zhuzh] > The source task may get stuck after a failover occurs in batch jobs > --- > > Key: FLINK-35522 > URL: https://issues.apache.org/jira/browse/FLINK-35522 > Project: Flink > Issue Type: Bug > Components: Runtime / Coordination >Affects Versions: 1.17.2, 1.19.0, 1.18.1, 1.20.0 >Reporter: xingbe >Priority: Major > Fix For: 1.20.0 > > > If the source task does not get assigned a split because the SplitEnumerator > has no more splits, and a failover occurs during the closing process, the > SourceCoordinatorContext will not resend the NoMoreSplit event to the newly > started source task, causing the source vertex to remain stuck indefinitely. > This case may only occur in batch jobs where speculative execution has been > enabled. -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-35354] Support host mapping in Flink tikv cdc [flink-cdc]
Mrart commented on PR #3336: URL: https://github.com/apache/flink-cdc/pull/3336#issuecomment-2148692911 @leonardBang Can you help me review it again? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Assigned] (FLINK-35522) The source task may get stuck after a failover occurs in batch jobs
[ https://issues.apache.org/jira/browse/FLINK-35522?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Zhu Zhu reassigned FLINK-35522: --- Assignee: xingbe > The source task may get stuck after a failover occurs in batch jobs > --- > > Key: FLINK-35522 > URL: https://issues.apache.org/jira/browse/FLINK-35522 > Project: Flink > Issue Type: Bug > Components: Runtime / Coordination >Affects Versions: 1.17.2, 1.19.0, 1.18.1, 1.20.0 >Reporter: xingbe >Assignee: xingbe >Priority: Major > Fix For: 1.20.0 > > > If the source task does not get assigned a split because the SplitEnumerator > has no more splits, and a failover occurs during the closing process, the > SourceCoordinatorContext will not resend the NoMoreSplit event to the newly > started source task, causing the source vertex to remain stuck indefinitely. > This case may only occur in batch jobs where speculative execution has been > enabled. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-35522) The source task may get stuck after a failover occurs in batch jobs
[ https://issues.apache.org/jira/browse/FLINK-35522?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17852241#comment-17852241 ] Zhu Zhu commented on FLINK-35522: - Thanks for reporting this problem and volunteering to fix it! [~xiasun] The ticket is assigned to you. > The source task may get stuck after a failover occurs in batch jobs > --- > > Key: FLINK-35522 > URL: https://issues.apache.org/jira/browse/FLINK-35522 > Project: Flink > Issue Type: Bug > Components: Runtime / Coordination >Affects Versions: 1.17.2, 1.19.0, 1.18.1, 1.20.0 >Reporter: xingbe >Assignee: xingbe >Priority: Major > Fix For: 1.20.0 > > > If the source task does not get assigned a split because the SplitEnumerator > has no more splits, and a failover occurs during the closing process, the > SourceCoordinatorContext will not resend the NoMoreSplit event to the newly > started source task, causing the source vertex to remain stuck indefinitely. > This case may only occur in batch jobs where speculative execution has been > enabled. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[jira] [Commented] (FLINK-35512) ArtifactFetchManagerTest unit tests fail
[ https://issues.apache.org/jira/browse/FLINK-35512?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17852243#comment-17852243 ] Rob Young commented on FLINK-35512: --- The additional artifact looks arbitrary from the ArtifactFetchManager's point of view, so instead of depending on output of the build, maybe we should use a file controlled by the test like: {code:java} - private File getFlinkClientsJar() throws IOException { - return TestingUtils.getFileFromTargetDir( - ArtifactFetchManager.class, - p -> - org.apache.flink.util.FileUtils.isJarFile(p) - && p.toFile().getName().startsWith("flink-clients") - && !p.toFile().getName().contains("test-utils")); + private File createArbitraryArtifact() throws IOException { + Path tempFile = Files.createTempFile(tempDir, "arbitrary", ".jar"); + Files.write(tempFile, UUID.randomUUID().toString().getBytes(StandardCharsets.UTF_8)); + return tempFile.toFile(); } {code} usages of `File sourceFile = TestingUtils.getClassFile(getClass());` could also be replaced with this so all test inputs are generated by the test > ArtifactFetchManagerTest unit tests fail > > > Key: FLINK-35512 > URL: https://issues.apache.org/jira/browse/FLINK-35512 > Project: Flink > Issue Type: Technical Debt >Affects Versions: 1.19.1 >Reporter: Hong Liang Teoh >Priority: Major > Fix For: 1.19.1 > > > The below three tests from *ArtifactFetchManagerTest* seem to fail > consistently: > * ArtifactFetchManagerTest.testFileSystemFetchWithAdditionalUri > * ArtifactFetchManagerTest.testMixedArtifactFetch > * ArtifactFetchManagerTest.testHttpFetch > The error printed is > {code:java} > java.lang.AssertionError: > Expecting actual not to be empty > at > org.apache.flink.client.program.artifact.ArtifactFetchManagerTest.getFlinkClientsJar(ArtifactFetchManagerTest.java:248) > at > org.apache.flink.client.program.artifact.ArtifactFetchManagerTest.testMixedArtifactFetch(ArtifactFetchManagerTest.java:146) > at java.lang.reflect.Method.invoke(Method.java:498) > at java.util.concurrent.RecursiveAction.exec(RecursiveAction.java:189) > at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289) > at > java.util.concurrent.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1056) > at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1692) > at > java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:175) > {code} > -- This message was sent by Atlassian Jira (v8.20.10#820010)
Re: [PR] [FLINK-31664] Implement ARRAY_INTERSECT function [flink]
liuyongvs commented on code in PR #24526: URL: https://github.com/apache/flink/pull/24526#discussion_r1626831685 ## flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/functions/CollectionFunctionsITCase.java: ## @@ -1723,6 +1724,83 @@ private Stream arrayExceptTestCases() { + "ARRAY_EXCEPT(, )")); } +private Stream arrayIntersectTestCases() { +return Stream.of( + TestSetSpec.forFunction(BuiltInFunctionDefinitions.ARRAY_INTERSECT) +.onFieldsWithData( +new Integer[] {1, 1, 2}, +null, +new Row[] {Row.of(true, 1), Row.of(true, 2), null}, +new Integer[] {null, null, 1}, +new Map[] { +CollectionUtil.map(entry(1, "a"), entry(2, "b")), +CollectionUtil.map(entry(3, "c"), entry(4, "d")) +}, +new Integer[][] {new Integer[] {1, 2, 3}}) +.andDataTypes( +DataTypes.ARRAY(DataTypes.INT()), +DataTypes.ARRAY(DataTypes.INT()), +DataTypes.ARRAY( +DataTypes.ROW(DataTypes.BOOLEAN(), DataTypes.INT())), +DataTypes.ARRAY(DataTypes.INT()), +DataTypes.ARRAY(DataTypes.MAP(DataTypes.INT(), DataTypes.STRING())), + DataTypes.ARRAY(DataTypes.ARRAY(DataTypes.INT( +// ARRAY +.testResult( +$("f0").arrayIntersect(new Integer[] {1, null, 4}), +"ARRAY_INTERSECT(f0, ARRAY[1, NULL, 4])", +new Integer[] {1, 1}, Review Comment: @snuyanzin fixed -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [FLINK-31664] Implement ARRAY_INTERSECT function [flink]
liuyongvs commented on code in PR #24526: URL: https://github.com/apache/flink/pull/24526#discussion_r1626831821 ## flink-table/flink-table-planner/src/test/java/org/apache/flink/table/planner/functions/CollectionFunctionsITCase.java: ## @@ -1723,6 +1724,83 @@ private Stream arrayExceptTestCases() { + "ARRAY_EXCEPT(, )")); } +private Stream arrayIntersectTestCases() { +return Stream.of( + TestSetSpec.forFunction(BuiltInFunctionDefinitions.ARRAY_INTERSECT) +.onFieldsWithData( +new Integer[] {1, 1, 2}, +null, +new Row[] {Row.of(true, 1), Row.of(true, 2), null}, +new Integer[] {null, null, 1}, +new Map[] { +CollectionUtil.map(entry(1, "a"), entry(2, "b")), +CollectionUtil.map(entry(3, "c"), entry(4, "d")) +}, +new Integer[][] {new Integer[] {1, 2, 3}}) +.andDataTypes( +DataTypes.ARRAY(DataTypes.INT()), +DataTypes.ARRAY(DataTypes.INT()), +DataTypes.ARRAY( +DataTypes.ROW(DataTypes.BOOLEAN(), DataTypes.INT())), +DataTypes.ARRAY(DataTypes.INT()), +DataTypes.ARRAY(DataTypes.MAP(DataTypes.INT(), DataTypes.STRING())), + DataTypes.ARRAY(DataTypes.ARRAY(DataTypes.INT( +// ARRAY +.testResult( +$("f0").arrayIntersect(new Integer[] {1, null, 4}), +"ARRAY_INTERSECT(f0, ARRAY[1, NULL, 4])", +new Integer[] {1, 1}, +DataTypes.ARRAY(DataTypes.INT())) +.testResult( +$("f0").arrayIntersect(new Integer[] {3, 4}), +"ARRAY_INTERSECT(f0, ARRAY[3, 4])", +new Integer[] {}, +DataTypes.ARRAY(DataTypes.INT())) +.testResult( +$("f1").arrayIntersect(new Integer[] {1, null, 4}), +"ARRAY_INTERSECT(f1, ARRAY[1, NULL, 4])", +null, +DataTypes.ARRAY(DataTypes.INT())) +// ARRAY> +.testResult( +$("f2").arrayIntersect( +new Row[] { +null, Row.of(true, 2), +}), +"ARRAY_INTERSECT(f2, ARRAY[NULL, ROW(TRUE, 2)])", +new Row[] {Row.of(true, 2), null}, +DataTypes.ARRAY( +DataTypes.ROW(DataTypes.BOOLEAN(), DataTypes.INT( +// arrayOne contains null elements +.testResult( +$("f3").arrayIntersect(new Integer[] {null, 42}), +"ARRAY_INTERSECT(f3, ARRAY[null, 42])", +new Integer[] {null, null}, Review Comment: fixed @snuyanzin -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[jira] [Updated] (FLINK-35522) The source task may get stuck after a failover occurs in batch jobs
[ https://issues.apache.org/jira/browse/FLINK-35522?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] ASF GitHub Bot updated FLINK-35522: --- Labels: pull-request-available (was: ) > The source task may get stuck after a failover occurs in batch jobs > --- > > Key: FLINK-35522 > URL: https://issues.apache.org/jira/browse/FLINK-35522 > Project: Flink > Issue Type: Bug > Components: Runtime / Coordination >Affects Versions: 1.17.2, 1.19.0, 1.18.1, 1.20.0 >Reporter: xingbe >Assignee: xingbe >Priority: Major > Labels: pull-request-available > Fix For: 1.20.0 > > > If the source task does not get assigned a split because the SplitEnumerator > has no more splits, and a failover occurs during the closing process, the > SourceCoordinatorContext will not resend the NoMoreSplit event to the newly > started source task, causing the source vertex to remain stuck indefinitely. > This case may only occur in batch jobs where speculative execution has been > enabled. -- This message was sent by Atlassian Jira (v8.20.10#820010)
[PR] [FLINK-35522][runtime] Fix the issue that the source task may get stuck in speculative execution mode. [flink]
SinBex opened a new pull request, #24887: URL: https://github.com/apache/flink/pull/24887 ## What is the purpose of the change If the source task does not get assigned a split because the SplitEnumerator has no more splits, and a failover occurs during the closing process, the SourceCoordinatorContext will not resend the NoMoreSplit event to the newly started source task, causing the source vertex to remain stuck indefinitely. This case may only occur in batch jobs where speculative execution has been enabled. ## Brief change log - fix the issue that the source task may get stuck in speculative execution mode. ## Verifying this change This change added tests and can be verified as follows: - *Added it case to verify the issue.* ## Does this pull request potentially affect one of the following parts: - Dependencies (does it add or upgrade a dependency): ( no) - The public API, i.e., is any changed class annotated with `@Public(Evolving)`: ( no) - The serializers: (no) - The runtime per-record code paths (performance sensitive): ( no ) - Anything that affects deployment or recovery: JobManager (and its components), Checkpointing, Kubernetes/Yarn, ZooKeeper: (no ) - The S3 file system connector: (no ) ## Documentation - Does this pull request introduce a new feature? (no) -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [FLINK-31664] Implement ARRAY_INTERSECT function [flink]
liuyongvs commented on PR #24526: URL: https://github.com/apache/flink/pull/24526#issuecomment-2148729746 @snuyanzin fix your review, thanks very much -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[PR] [BP-3.1.1][FLINK-35149][cdc-composer] Fix DataSinkTranslator#sinkTo ignoring pre-write topology if not TwoPhaseCommittingSink (#3233) [flink-cdc]
loserwang1024 opened a new pull request, #3386: URL: https://github.com/apache/flink-cdc/pull/3386 BP https://issues.apache.org/jira/browse/FLINK-35149 to 3.1.1 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: [PR] [BP-3.1.1][FLINK-35149][cdc-composer] Fix DataSinkTranslator#sinkTo ignoring pre-write topology if not TwoPhaseCommittingSink (#3233) [flink-cdc]
loserwang1024 closed pull request #3386: [BP-3.1.1][FLINK-35149][cdc-composer] Fix DataSinkTranslator#sinkTo ignoring pre-write topology if not TwoPhaseCommittingSink (#3233) URL: https://github.com/apache/flink-cdc/pull/3386 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
[PR] [BP-3.1.1][FLINK-35149][cdc-composer] Fix DataSinkTranslator#sinkTo ignoring pre-write topology if not TwoPhaseCommittingSink (#3233) [flink-cdc]
loserwang1024 opened a new pull request, #3387: URL: https://github.com/apache/flink-cdc/pull/3387 BP https://issues.apache.org/jira/browse/FLINK-35149 to 3.1.1 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@flink.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org