Agreed, let's not tie our releases to other projects if we can help it. On Thu, Mar 27, 2025 at 1:44 AM Jean-Baptiste Onofré <j...@nanthrax.net> wrote:
> Hi Ajantha > > I will help on the Flink PR. However I don’t think we need to hold 1.9.0 > for Spark and Flink updates. We can always include that in the next release > cycle. > > I propose to focus on issues and check the milestone on GH. > > Regards > JB > > Le jeu. 27 mars 2025 à 07:11, Ajantha Bhat <ajanthab...@gmail.com> a > écrit : > >> Latest update on the release: >> >> As discussed in the last community sync, I have waited for the Spark 4.0 >> release. But the RC3 has failed. >> Flink 2.0 has been released, there is an open PR >> <https://github.com/apache/iceberg/pull/12527> to include it at the >> Iceberg. It didn't make any progress this week. >> >> We have recently found a few issues and small requirements for Auth >> manager, >> Added them to the milestone: >> https://github.com/apache/iceberg/milestone/53 >> <https://github.com/apache/iceberg/milestone/53> >> If we get good review support, I think 2-3 days should be sufficient to >> close these and prepare a release. >> >> >> - Ajantha >> >> On Thu, Mar 20, 2025 at 1:42 PM Alex Dutra <alex.du...@dremio.com.invalid> >> wrote: >> >>> Hi Yuya, >>> >>> Thanks for reporting this issue, which is indeed a defect in the SigV4 >>> auth manager. The fix was hopefully easy: >>> >>> https://github.com/apache/iceberg/pull/12582 >>> >>> Sorry for the inconvenience, >>> >>> Alex >>> >>> On Thu, Mar 20, 2025 at 5:20 AM Yuya Ebihara < >>> yuya.ebih...@starburstdata.com> wrote: >>> >>>> S3 Tables tests in Trino with 1.9.0 nightly release throws >>>> ClassCastException (RESTSigV4AuthSession → OAuth2Util$AuthSession). The >>>> test works fine in 1.8.1. >>>> I’m going to check if adjusting our test settings based on this PR ( >>>> https://github.com/apache/iceberg/pull/11995) can fix the issue. >>>> >>>> >>>> https://github.com/trinodb/trino/actions/runs/13960201796/job/39080056606?pr=25331 >>>> Error: >>>> io.trino.plugin.iceberg.catalog.rest.TestIcebergS3TablesConnectorSmokeTest >>>> -- Time elapsed: 3.038 s <<< ERROR! >>>> io.trino.testing.QueryFailedException: class >>>> org.apache.iceberg.aws.RESTSigV4AuthSession cannot be cast to class >>>> org.apache.iceberg.rest.auth.OAuth2Util$AuthSession >>>> (org.apache.iceberg.aws.RESTSigV4AuthSession and >>>> org.apache.iceberg.rest.auth.OAuth2Util$AuthSession are in unnamed module >>>> of loader 'app') >>>> at >>>> io.trino.testing.AbstractTestingTrinoClient.execute(AbstractTestingTrinoClient.java:138) >>>> at >>>> io.trino.testing.DistributedQueryRunner.executeInternal(DistributedQueryRunner.java:565) >>>> at >>>> io.trino.testing.DistributedQueryRunner.execute(DistributedQueryRunner.java:548) >>>> at io.trino.testing.QueryRunner.execute(QueryRunner.java:82) >>>> at >>>> io.trino.plugin.iceberg.SchemaInitializer.accept(SchemaInitializer.java:54) >>>> at >>>> io.trino.plugin.iceberg.IcebergQueryRunner$Builder.lambda$build$3(IcebergQueryRunner.java:178) >>>> at java.base/java.util.Optional.ifPresent(Optional.java:178) >>>> at >>>> io.trino.plugin.iceberg.IcebergQueryRunner$Builder.build(IcebergQueryRunner.java:178) >>>> at >>>> io.trino.plugin.iceberg.catalog.rest.TestIcebergS3TablesConnectorSmokeTest.createQueryRunner(TestIcebergS3TablesConnectorSmokeTest.java:83) >>>> at >>>> io.trino.testing.AbstractTestQueryFramework.init(AbstractTestQueryFramework.java:119) >>>> at java.base/java.lang.reflect.Method.invoke(Method.java:580) >>>> at >>>> java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:507) >>>> at >>>> java.base/java.util.concurrent.ForkJoinPool$WorkQueue.tryRemoveAndExec(ForkJoinPool.java:1501) >>>> at >>>> java.base/java.util.concurrent.ForkJoinPool.helpJoin(ForkJoinPool.java:2274) >>>> at >>>> java.base/java.util.concurrent.ForkJoinTask.awaitDone(ForkJoinTask.java:495) >>>> at >>>> java.base/java.util.concurrent.ForkJoinTask.join(ForkJoinTask.java:662) >>>> at >>>> java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:507) >>>> at >>>> java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1460) >>>> at >>>> java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:2036) >>>> at >>>> java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:189) >>>> Suppressed: java.lang.Exception: SQL: CREATE SCHEMA IF NOT EXISTS "tpch" >>>> at >>>> io.trino.testing.DistributedQueryRunner.executeInternal(DistributedQueryRunner.java:572) >>>> ... 18 more >>>> Caused by: java.lang.ClassCastException: class >>>> org.apache.iceberg.aws.RESTSigV4AuthSession cannot be cast to class >>>> org.apache.iceberg.rest.auth.OAuth2Util$AuthSession >>>> (org.apache.iceberg.aws.RESTSigV4AuthSession and >>>> org.apache.iceberg.rest.auth.OAuth2Util$AuthSession are in unnamed module >>>> of loader 'app') >>>> at >>>> org.apache.iceberg.rest.auth.OAuth2Manager.contextualSession(OAuth2Manager.java:142) >>>> at >>>> org.apache.iceberg.rest.auth.OAuth2Manager.contextualSession(OAuth2Manager.java:40) >>>> at >>>> org.apache.iceberg.aws.RESTSigV4AuthManager.contextualSession(RESTSigV4AuthManager.java:68) >>>> at >>>> org.apache.iceberg.rest.RESTSessionCatalog.loadNamespaceMetadata(RESTSessionCatalog.java:608) >>>> at >>>> org.apache.iceberg.catalog.SessionCatalog.namespaceExists(SessionCatalog.java:358) >>>> at >>>> org.apache.iceberg.rest.RESTSessionCatalog.namespaceExists(RESTSessionCatalog.java:595) >>>> at >>>> io.trino.plugin.iceberg.catalog.rest.TrinoRestCatalog.namespaceExists(TrinoRestCatalog.java:162) >>>> at >>>> io.trino.plugin.iceberg.IcebergMetadata.schemaExists(IcebergMetadata.java:512) >>>> at >>>> io.trino.plugin.base.classloader.ClassLoaderSafeConnectorMetadata.schemaExists(ClassLoaderSafeConnectorMetadata.java:193) >>>> at >>>> io.trino.tracing.TracingConnectorMetadata.schemaExists(TracingConnectorMetadata.java:125) >>>> at >>>> io.trino.metadata.MetadataManager.lambda$schemaExists$1(MetadataManager.java:248) >>>> at >>>> java.base/java.util.stream.MatchOps$1MatchSink.accept(MatchOps.java:90) >>>> at >>>> java.base/java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:215) >>>> at >>>> java.base/java.util.Spliterators$ArraySpliterator.tryAdvance(Spliterators.java:1034) >>>> at >>>> java.base/java.util.stream.ReferencePipeline.forEachWithCancel(ReferencePipeline.java:147) >>>> at >>>> java.base/java.util.stream.AbstractPipeline.copyIntoWithCancel(AbstractPipeline.java:588) >>>> at >>>> java.base/java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:574) >>>> at >>>> java.base/java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:560) >>>> at >>>> java.base/java.util.stream.MatchOps$MatchOp.evaluateSequential(MatchOps.java:230) >>>> at >>>> java.base/java.util.stream.MatchOps$MatchOp.evaluateSequential(MatchOps.java:196) >>>> at >>>> java.base/java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:265) >>>> at >>>> java.base/java.util.stream.ReferencePipeline.anyMatch(ReferencePipeline.java:672) >>>> at >>>> io.trino.metadata.MetadataManager.schemaExists(MetadataManager.java:248) >>>> at >>>> io.trino.tracing.TracingMetadata.schemaExists(TracingMetadata.java:165) >>>> at >>>> io.trino.execution.CreateSchemaTask.internalExecute(CreateSchemaTask.java:119) >>>> at io.trino.execution.CreateSchemaTask.execute(CreateSchemaTask.java:82) >>>> at io.trino.execution.CreateSchemaTask.execute(CreateSchemaTask.java:54) >>>> at >>>> io.trino.execution.DataDefinitionExecution.start(DataDefinitionExecution.java:152) >>>> at >>>> io.trino.execution.SqlQueryManager.createQuery(SqlQueryManager.java:272) >>>> at >>>> io.trino.dispatcher.LocalDispatchQuery.startExecution(LocalDispatchQuery.java:150) >>>> at >>>> io.trino.dispatcher.LocalDispatchQuery.lambda$waitForMinimumWorkers$2(LocalDispatchQuery.java:134) >>>> at >>>> io.airlift.concurrent.MoreFutures.lambda$addSuccessCallback$12(MoreFutures.java:570) >>>> at io.airlift.concurrent.MoreFutures$3.onSuccess(MoreFutures.java:545) >>>> at >>>> com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1139) >>>> at io.trino.$gen.Trino_testversion____20250320_014939_2038.run(Unknown >>>> Source) >>>> at >>>> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1144) >>>> at >>>> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:642) >>>> at java.base/java.lang.Thread.run(Thread.java:1575) >>>> >>>> BR, >>>> Yuya >>>> >>>> On Wed, Mar 19, 2025 at 12:38 AM Manu Zhang <owenzhang1...@gmail.com> >>>> wrote: >>>> >>>>> Hi Ajantha, >>>>> >>>>> Thanks for driving the release. Can we include >>>>> https://github.com/apache/iceberg/pull/12120? >>>>> >>>>> On Tue, Mar 18, 2025 at 3:18 AM Steve Loughran >>>>> <ste...@cloudera.com.invalid> wrote: >>>>> >>>>>> >>>>>> Can I get this reviewed and merged; gives all hadoop filesystems with >>>>>> bulk delete calls the ability to issue bulk deletes up to their page >>>>>> sizes; >>>>>> off by default. Tested all the way through iceberg to AWS S3 london. >>>>>> >>>>>> https://github.com/apache/iceberg/pull/10233 >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On Mon, 17 Mar 2025 at 12:32, Yuya Ebihara < >>>>>> yuya.ebih...@starburstdata.com> wrote: >>>>>> >>>>>>> Hi, can we include https://github.com/apache/iceberg/pull/12264 to >>>>>>> fix the S3-compatible storage issue? >>>>>>> We downgraded the problematic library in >>>>>>> https://github.com/apache/iceberg/pull/12339, but the issue is >>>>>>> still present in the main branch. >>>>>>> >>>>>> >>>>>> I'd go with downgrading the AWS SDK to 2.29. x and then get on the >>>>>> relevant AWS SDK discussion to express your concerns: >>>>>> https://github.com/aws/aws-sdk-java-v2/discussions/5802 >>>>>> >>>>>> The problem here is that there's such a broad set of implementations >>>>>> of the S3 API that it'll take testing to see how even the suggestions >>>>>> from >>>>>> the SDK team work everywhere -and we now have explicit confirmation that >>>>>> the SDK team leave all such testing to downstream users. >>>>>> >>>>>> >>>>>> >>>>>> *The AWS SDKs and CLI are designed for usage with official AWS >>>>>> services.We may introduce and enable new features by default, such as >>>>>> these >>>>>> new default integrity protections,prior to them being supported or >>>>>> otherwise handled by third-party service implementations.* >>>>>> >>>>>> I think ASF projects need to make clear how dangerous this is -that >>>>>> projects will end up shipping releases which don't work, and the "set an >>>>>> env var or a system property" workarounds are not enough. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>> >>>>>>> On Mon, Mar 17, 2025 at 8:47 PM Ajantha Bhat <ajanthab...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> We decided to have a fast follow up on the things that missed >>>>>>>> release train of 1.8.0 during community sync. >>>>>>>> >>>>>>>> More details here: >>>>>>>> https://lists.apache.org/thread/wvz5sd7pmh5ww1yqhsxpt1kwf993276j >>>>>>>> >>>>>>>> On Mon, Mar 17, 2025 at 4:53 PM Russell Spitzer < >>>>>>>> russell.spit...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Can you please rehash the plan? I thought we just did a release >>>>>>>>> last month and were aiming for a three month schedule. I may have >>>>>>>>> missed >>>>>>>>> something >>>>>>>>> >>>>>>>>> On Mon, Mar 17, 2025 at 6:00 AM Ajantha Bhat < >>>>>>>>> ajanthab...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> Hey everyone, >>>>>>>>>> >>>>>>>>>> Following the plan from the 1.8.0 release, I'll be handling the >>>>>>>>>> 1.9.0 release. We have several major updates ready: >>>>>>>>>> >>>>>>>>>> - *Partition stats APIs:* All core APIs for partition >>>>>>>>>> statistics have been merged, unblocking engines like Dremio, >>>>>>>>>> Trino, and >>>>>>>>>> Hive that were waiting for this feature. >>>>>>>>>> - *REST catalog authentication:* The refactoring PR for the >>>>>>>>>> REST catalog authentication manager has been merged, improving >>>>>>>>>> authentication support. >>>>>>>>>> - *Spark 3.3 support removed:* We've officially dropped >>>>>>>>>> support for Spark 3.3. >>>>>>>>>> - *InternalData support in the core module:* The core module >>>>>>>>>> now supports InternalData, leveraging the internal Parquet and >>>>>>>>>> Avro readers >>>>>>>>>> added in the previous release. This allows metadata to be written >>>>>>>>>> in >>>>>>>>>> Parquet. >>>>>>>>>> - *Bug fixes:* Many important bug fixes have been merged. >>>>>>>>>> >>>>>>>>>> A 1.9.0 milestone has been created with additional "good-to-have" >>>>>>>>>> issues: >>>>>>>>>> https://github.com/apache/iceberg/milestone/53 >>>>>>>>>> >>>>>>>>>> If there's anything urgent that needs to be included for this >>>>>>>>>> release, please let me know or reply to this thread. >>>>>>>>>> I'm aiming to start the release cut by the end of this week. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> - Ajantha >>>>>>>>>> >>>>>>>>>