Beam High Priority Issue Report (40)
This is your daily summary of Beam's current high priority issues that may need attention. See https://beam.apache.org/contribute/issue-priorities for the meaning and expectations around issue priorities. Unassigned P1 Issues: https://github.com/apache/beam/issues/25140 [Bug]: GenerateSequence is broken on SDF https://github.com/apache/beam/issues/24971 [Bug]: Messages are not published when a connection is closed with JmsIO https://github.com/apache/beam/issues/24776 [Bug]: Race condition in Python SDK Harness ProcessBundleProgress https://github.com/apache/beam/issues/24655 [Bug]: Pipeline fusion should break at @RequiresStableInput boundary https://github.com/apache/beam/issues/24389 [Failing Test]: HadoopFormatIOElasticTest.classMethod ExceptionInInitializerError ContainerFetchException https://github.com/apache/beam/issues/24367 [Bug]: workflow.tar.gz cannot be passed to flink runner https://github.com/apache/beam/issues/24313 [Flaky]: apache_beam/runners/portability/portable_runner_test.py::PortableRunnerTestWithSubprocesses::test_pardo_state_with_custom_key_coder https://github.com/apache/beam/issues/24267 [Failing Test]: Timeout waiting to lock gradle https://github.com/apache/beam/issues/23944 beam_PreCommit_Python_Cron regularily failing - test_pardo_large_input flaky https://github.com/apache/beam/issues/23709 [Flake]: Spark batch flakes in ParDoLifecycleTest.testTeardownCalledAfterExceptionInProcessElement and ParDoLifecycleTest.testTeardownCalledAfterExceptionInStartBundle https://github.com/apache/beam/issues/22969 Discrepancy in behavior of `DoFn.process()` when `yield` is combined with `return` statement, or vice versa https://github.com/apache/beam/issues/22961 [Bug]: WriteToBigQuery silently skips most of records without job fail https://github.com/apache/beam/issues/22913 [Bug]: beam_PostCommit_Java_ValidatesRunner_Flink is flakes in org.apache.beam.sdk.transforms.GroupByKeyTest$BasicTests.testAfterProcessingTimeContinuationTriggerUsingState https://github.com/apache/beam/issues/21713 404s in BigQueryIO don't get output to Failed Inserts PCollection https://github.com/apache/beam/issues/21695 DataflowPipelineResult does not raise exception for unsuccessful states. https://github.com/apache/beam/issues/21643 FnRunnerTest with non-trivial (order 1000 elements) numpy input flakes in non-cython environment https://github.com/apache/beam/issues/21469 beam_PostCommit_XVR_Flink flaky: Connection refused https://github.com/apache/beam/issues/21424 Java VR (Dataflow, V2, Streaming) failing: ParDoTest$TimestampTests/OnWindowExpirationTests https://github.com/apache/beam/issues/21262 Python AfterAny, AfterAll do not follow spec https://github.com/apache/beam/issues/21260 Python DirectRunner does not emit data at GC time https://github.com/apache/beam/issues/21121 apache_beam.examples.streaming_wordcount_it_test.StreamingWordCountIT.test_streaming_wordcount_it flakey https://github.com/apache/beam/issues/21104 Flaky: apache_beam.runners.portability.fn_api_runner.fn_runner_test.FnApiRunnerTestWithGrpcAndMultiWorkers https://github.com/apache/beam/issues/20976 apache_beam.runners.portability.flink_runner_test.FlinkRunnerTestOptimized.test_flink_metrics is flaky https://github.com/apache/beam/issues/20974 Python GHA PreCommits flake with grpc.FutureTimeoutError on SDK harness startup https://github.com/apache/beam/issues/20689 Kafka commitOffsetsInFinalize OOM on Flink https://github.com/apache/beam/issues/20108 Python direct runner doesn't emit empty pane when it should https://github.com/apache/beam/issues/19814 Flink streaming flakes in ParDoLifecycleTest.testTeardownCalledAfterExceptionInStartBundleStateful and ParDoLifecycleTest.testTeardownCalledAfterExceptionInProcessElementStateful https://github.com/apache/beam/issues/19241 Python Dataflow integration tests should export the pipeline Job ID and console output to Jenkins Test Result section P1 Issues with no update in the last week: https://github.com/apache/beam/issues/23875 [Bug]: beam.Row.__eq__ returns true for unequal rows https://github.com/apache/beam/issues/23525 [Bug]: Default PubsubMessage coder will drop message id and orderingKey https://github.com/apache/beam/issues/22605 [Bug]: Beam Python failure for dataflow_exercise_metrics_pipeline_test.ExerciseMetricsPipelineTest.test_metrics_it https://github.com/apache/beam/issues/22115 [Bug]: apache_beam.runners.portability.portable_runner_test.PortableRunnerTestWithSubprocesses is flaky https://github.com/apache/beam/issues/21714 PulsarIOTest.testReadFromSimpleTopic is very flaky https://github.com/apache/beam/issues/21708 beam_PostCommit_Java_DataflowV2, testBigQueryStorageWrite30MProto failing consistently https://github.com/apache/beam/issues/21706 Flaky timeout in github Python unit test action StatefulDoFnOnDirectRunnerTest.test_dynamic_timer_clear_then_set_timer https://github.com/apache/beam/issues/21700 --dataflowServiceOpt
[GitHub] [beam-site] robertwb merged pull request #639: Update beam-site for release 2.44.0.
robertwb merged PR #639: URL: https://github.com/apache/beam-site/pull/639 -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: dev-unsubscr...@beam.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org
Re: Beam SQL Alias issue while using With Clause
Hi Andrew, Thank you for your MR. I am parricated to help us to solve the issue. I rerun our tests and they are partially passing now with your fix. However, there is one more issue with the WITH clause. When i run following query somehow beam lost type of column WITH tempTable AS (SELECT * FROM PCOLLECTION WHERE PCOLLECTION.`user_info`.`name` = 'User1') SELECT * FROM tempTable I havent test on Beam Master. I run with your latest patch on our code base. This is the output 14:00:30.095 [Test worker] INFO o.a.b.sdk.extensions.sql.impl.CalciteQueryPlanner - SQL: WITH `tempTable` AS (SELECT `PCOLLECTION`.`id`, `PCOLLECTION`.`value`, `PCOLLECTION`.`user_info` FROM `beam`.`PCOLLECTION` AS `PCOLLECTION` WHERE `PCOLLECTION`.`user_info`.`name` = 'User1') (SELECT `tempTable`.`id`, `tempTable`.`value`, `tempTable`.`user_info` FROM `tempTable` AS `tempTable`) 14:00:30.106 [Test worker] DEBUG o.a.b.v.calcite.v1_28_0.org.apache.calcite.sql2rel - Plan after converting SqlNode to RelNode LogicalProject(id=[$0], value=[$1], user_info=[$2]) LogicalFilter(condition=[=($2.name, 'User1')]) BeamIOSourceRel(table=[[beam, PCOLLECTION]]) 14:00:30.107 [Test worker] DEBUG o.a.b.v.calcite.v1_28_0.org.apache.calcite.sql2rel - Plan after converting SqlNode to RelNode LogicalProject(id=[$0], value=[$1], user_info=[$2]) LogicalFilter(condition=[=($2.name, 'User1')]) BeamIOSourceRel(table=[[beam, PCOLLECTION]]) 14:00:30.109 [Test worker] INFO o.a.b.sdk.extensions.sql.impl.CalciteQueryPlanner - SQLPlan> LogicalProject(id=[$0], value=[$1], user_info=[ROW($2)]) LogicalFilter(condition=[=($2.name, 'User1')]) LogicalProject(id=[$0], value=[$1], name=[$2.name]) BeamIOSourceRel(table=[[beam, PCOLLECTION]]) 14:00:30.173 [Test worker] DEBUG o.a.b.v.c.v.org.apache.calcite.plan.RelOptPlanner - PLANNER = org.apache.beam.vendor.calcite.v1_28_0.org.apache.calcite.plan.volcano.IterativeRuleDriver@1c510081; COST = {inf} 14:00:30.173 [Test worker] DEBUG o.a.b.v.c.v.org.apache.calcite.plan.RelOptPlanner - Pop match: rule [BeamEnumerableConverterRule(in:BEAM_LOGICAL,out:ENUMERABLE)] rels [#27] 14:00:30.173 [Test worker] DEBUG o.a.b.v.c.v.org.apache.calcite.plan.RelOptPlanner - call#118: Apply rule [BeamEnumerableConverterRule(in:BEAM_LOGICAL,out:ENUMERABLE)] to [rel#27:BeamIOSourceRel.BEAM_LOGICAL(table=[beam, PCOLLECTION])] 14:00:30.174 [Test worker] DEBUG o.a.b.v.c.v.org.apache.calcite.plan.RelOptPlanner - Transform to: rel#41 via BeamEnumerableConverterRule(in:BEAM_LOGICAL,out:ENUMERABLE) 14:00:30.175 [Test worker] DEBUG o.a.b.v.c.v.org.apache.calcite.plan.RelOptPlanner - call#118 generated 1 successors: [rel#41:BeamEnumerableConverter.ENUMERABLE(input=BeamIOSourceRel#27)] 14:00:30.175 [Test worker] DEBUG o.a.b.v.c.v.org.apache.calcite.plan.RelOptPlanner - PLANNER = org.apache.beam.vendor.calcite.v1_28_0.org.apache.calcite.plan.volcano.IterativeRuleDriver@1c510081; COST = {inf} 14:00:30.175 [Test worker] DEBUG o.a.b.v.c.v.org.apache.calcite.plan.RelOptPlanner - Pop match: rule [ProjectToCalcRule] rels [#33] 14:00:30.175 [Test worker] DEBUG o.a.b.v.c.v.org.apache.calcite.plan.RelOptPlanner - call#136: Apply rule [ProjectToCalcRule] to [rel#33:LogicalProject.NONE(input=RelSubset#32,inputs=0..1,exprs=[$2.name])] 14:00:30.177 [Test worker] DEBUG o.a.b.v.c.v.org.apache.calcite.plan.RelOptPlanner - Transform to: rel#44 via ProjectToCalcRule 14:00:30.178 [Test worker] DEBUG o.a.b.v.c.v.org.apache.calcite.plan.RelOptPlanner - call#136 generated 1 successors: [rel#44:LogicalCalc.NONE(input=RelSubset#32,expr#0..2={inputs},expr#3=$ t2.name,proj#0..1={exprs},2=$t3)] 14:00:30.178 [Test worker] DEBUG o.a.b.v.c.v.org.apache.calcite.plan.RelOptPlanner - PLANNER = org.apache.beam.vendor.calcite.v1_28_0.org.apache.calcite.plan.volcano.IterativeRuleDriver@1c510081; COST = {inf} 14:00:30.178 [Test worker] DEBUG o.a.b.v.c.v.org.apache.calcite.plan.RelOptPlanner - Pop match: rule [FilterToCalcRule] rels [#35] 14:00:30.178 [Test worker] DEBUG o.a.b.v.c.v.org.apache.calcite.plan.RelOptPlanner - call#160: Apply rule [FilterToCalcRule] to [rel#35:LogicalFilter.NONE(input=RelSubset#34,condition==($2.name, 'User1'))] fieldList must not be null, type = VARCHAR java.lang.AssertionError: fieldList must not be null, type = VARCHAR at org.apache.beam.vendor.calcite.v1_28_0.org.apache.calcite.rel.type.RelDataTypeImpl.getFieldList(RelDataTypeImpl.java:164) at org.apache.beam.vendor.calcite.v1_28_0.org.apache.calcite.rex.RexFieldAccess.checkValid(RexFieldAccess.java:76) at org.apache.beam.vendor.calcite.v1_28_0.org.apache.calcite.rex.RexFieldAccess.(RexFieldAccess.java:64) at org.apache.beam.vendor.calcite.v1_28_0.org.apache.calcite.rex.RexShuttle.visitFieldAccess(RexShuttle.java:208) at org.apache.beam.vendor.calcite.v1_28_0.org.apache.calcite.rex.RexProgramBuilder$RegisterShuttle.visitFieldAccess(RexProgramBuilder.java:911) at org.apache.beam.vendor.calcite.v1_28_0.org.apache.calcite.rex.RexProgramBuilder$RegisterShuttle.visitFieldAccess(RexProgramBuil