+1 (non-binding) - Verify Signature - Verify Checksums - Verify License Documentation - Verify Build and Test in Java 8 and 11
On Mon, Oct 10, 2022 at 7:11 PM Daniel Weeks <dwe...@apache.org> wrote: > +1 (binding) > > Verified sigs/sums/license/build/test (java 8) > > Everything checks out, > -Dan > > On Mon, Oct 10, 2022 at 3:25 PM Steven Wu <stevenz...@gmail.com> wrote: > >> Never mind. Missed the information that this 1.0.0 is based on the latest >> 0.14.1 release, which doesn't contain the PR 5318. I thought it was based >> on the latest master branch. >> >> +1 (non-binding) >> - Verify signature >> - Verify checksum >> - Tried SQL insert and query with Flink 1.15 >> >> On Mon, Oct 10, 2022 at 3:20 PM Steven Wu <stevenz...@gmail.com> wrote: >> >>> Ryan, >>> >>> It seems that this PR (merged on July 28) for the Flink FLIP-27 source >>> is not included. https://github.com/apache/iceberg/pull/5318 >>> >>> The commit ID still contains the old file in the old location. The new >>> location should be "flink/source/IcebergTableSource.java" with FLIP-27 >>> config support. >>> >>> https://github.com/apache/iceberg/blob/e2bb9ad7e792efca419fa7c4a1afde7c4c44fa01/flink/v1.15/flink/src/main/java/org/apache/iceberg/flink/IcebergTableSource.java >>> >>> Thanks, >>> Steven >>> >>> On Mon, Oct 10, 2022 at 11:37 AM Szehon Ho <szehon.apa...@gmail.com> >>> wrote: >>> >>>> Whoops, sorry for the noise, I made a typo and was using the wrong >>>> scala version of the iceberg-spark-runtime jar, this works. >>>> >>>> +1 (non-binding) >>>> - Verify signature >>>> - Verify checksum >>>> - Verify license documentation >>>> - Tried with Spark 3.3 >>>> - Ran unit tests >>>> >>>> Thanks >>>> Szehon >>>> >>>> >>>> >>>> >>>> On Mon, Oct 10, 2022 at 11:26 AM Szehon Ho <szehon.apa...@gmail.com> >>>> wrote: >>>> >>>>> Hi, >>>>> >>>>> I get a NoClassDefFoundError from IcebergSparkExtensions when running >>>>> Spark 3.3, with iceberg-spark-runtime-3.3_2.12-1.0.0.jar. I noticed this >>>>> jar doesn't contain scala classes, unlike previous jars >>>>> iceberg-spark-runtime-3.3_2.12-0.14.1.jar. >>>>> >>>>> scala> spark.sql("show databases").show >>>>> java.lang.NoClassDefFoundError: scala/collection/SeqOps >>>>> at >>>>> org.apache.iceberg.spark.extensions.IcebergSparkSessionExtensions.$anonfun$apply$2(IcebergSparkSessionExtensions.scala:50) >>>>> at >>>>> org.apache.spark.sql.SparkSessionExtensions.$anonfun$buildResolutionRules$1(SparkSessionExtensions.scala:152) >>>>> at >>>>> scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:286) >>>>> at >>>>> scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62) >>>>> at >>>>> scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55) >>>>> at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49) >>>>> at scala.collection.TraversableLike.map(TraversableLike.scala:286) >>>>> at scala.collection.TraversableLike.map$(TraversableLike.scala:279) >>>>> at scala.collection.AbstractTraversable.map(Traversable.scala:108) >>>>> at >>>>> org.apache.spark.sql.SparkSessionExtensions.buildResolutionRules(SparkSessionExtensions.scala:152) >>>>> at >>>>> org.apache.spark.sql.internal.BaseSessionStateBuilder.customResolutionRules(BaseSessionStateBuilder.scala:216) >>>>> at >>>>> org.apache.spark.sql.hive.HiveSessionStateBuilder$$anon$1.<init>(HiveSessionStateBuilder.scala:94) >>>>> at >>>>> org.apache.spark.sql.hive.HiveSessionStateBuilder.analyzer(HiveSessionStateBuilder.scala:85) >>>>> at >>>>> org.apache.spark.sql.internal.BaseSessionStateBuilder.$anonfun$build$2(BaseSessionStateBuilder.scala:360) >>>>> at >>>>> org.apache.spark.sql.internal.SessionState.analyzer$lzycompute(SessionState.scala:87) >>>>> at >>>>> org.apache.spark.sql.internal.SessionState.analyzer(SessionState.scala:87) >>>>> at >>>>> org.apache.spark.sql.execution.QueryExecution.$anonfun$analyzed$1(QueryExecution.scala:76) >>>>> at >>>>> org.apache.spark.sql.catalyst.QueryPlanningTracker.measurePhase(QueryPlanningTracker.scala:111) >>>>> at >>>>> org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$2(QueryExecution.scala:185) >>>>> at >>>>> org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:510) >>>>> at >>>>> org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$1(QueryExecution.scala:185) >>>>> at >>>>> org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779) >>>>> at >>>>> org.apache.spark.sql.execution.QueryExecution.executePhase(QueryExecution.scala:184) >>>>> at >>>>> org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:76) >>>>> at >>>>> org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:74) >>>>> at >>>>> org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:66) >>>>> at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:99) >>>>> at >>>>> org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779) >>>>> at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97) >>>>> at >>>>> org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:622) >>>>> at >>>>> org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779) >>>>> at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:617) >>>>> ... 47 elided >>>>> Caused by: java.lang.ClassNotFoundException: scala.collection.SeqOps >>>>> at >>>>> java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:581) >>>>> at >>>>> java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178) >>>>> at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:522) >>>>> ... 79 more >>>>> >>>>> Note, I usually verify by copying the spark-runtime jar to spark jars >>>>> dir (can't usually get --packages flag to work as indicated on >>>>> https://iceberg.apache.org/how-to-release/#verifying-with-spark, as >>>>> version is not released yet), so let me know if I am using the wrong jar? >>>>> >>>>> Thanks >>>>> Szehon >>>>> >>>>> On Mon, Oct 10, 2022 at 9:22 AM Eduard Tudenhoefner <edu...@tabular.io> >>>>> wrote: >>>>> >>>>>> +1 (non-binding) >>>>>> >>>>>> - validated checksum and signature >>>>>> - checked license docs & ran RAT checks >>>>>> - ran build and tests with JDK11 >>>>>> >>>>>> >>>>>> Eduard >>>>>> >>>>>> On Mon, Oct 10, 2022 at 8:01 AM Ajantha Bhat <ajanthab...@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> +1 (non-binding) >>>>>>> >>>>>>> >>>>>>> - Verified the Spark runtime jar contents. >>>>>>> - Checked license docs, ran RAT checks. >>>>>>> - Validated checksum and signature. >>>>>>> >>>>>>> >>>>>>> Thanks, >>>>>>> Ajantha >>>>>>> >>>>>>> On Mon, Oct 10, 2022 at 10:45 AM Prashant Singh < >>>>>>> prashant010...@gmail.com> wrote: >>>>>>> >>>>>>>> Hello Everyone, >>>>>>>> >>>>>>>> Wanted to know your thoughts on whether we should also include the >>>>>>>> following bug fixes in this release as well: >>>>>>>> >>>>>>>> 1. MERGE INTO nullability fix, leads to query failure otherwise: >>>>>>>> *Reported instances :* >>>>>>>> a. >>>>>>>> https://stackoverflow.com/questions/73424454/spark-iceberg-merge-into-issue-caused-by-org-apache-spark-sql-analysisexcep >>>>>>>> b. https://github.com/apache/iceberg/issues/5739 >>>>>>>> c. >>>>>>>> https://github.com/apache/iceberg/issues/5424#issuecomment-1220688298 >>>>>>>> >>>>>>>> *PR's (Merged):* >>>>>>>> a. https://github.com/apache/iceberg/pull/5880 >>>>>>>> b. https://github.com/apache/iceberg/pull/5679 >>>>>>>> >>>>>>>> 2. QueryFailure when running RewriteManifestProcedure on Date / >>>>>>>> Timestamp partitioned table when >>>>>>>> `spark.sql.datetime.java8API.enabled` is true. >>>>>>>> *Reported instances :* >>>>>>>> a. https://github.com/apache/iceberg/issues/5104 >>>>>>>> b. >>>>>>>> https://apache-iceberg.slack.com/archives/C025PH0G1D4/p1663982635731469 >>>>>>>> >>>>>>>> *PR* : >>>>>>>> a. https://github.com/apache/iceberg/pull/5860 >>>>>>>> >>>>>>>> Regards, >>>>>>>> Prashant Singh >>>>>>>> >>>>>>>> On Mon, Oct 10, 2022 at 4:15 AM Ryan Blue <b...@apache.org> wrote: >>>>>>>> >>>>>>>>> +1 (binding) >>>>>>>>> >>>>>>>>> - Checked license docs, ran RAT checks >>>>>>>>> - Validated checksum and signature >>>>>>>>> - Built and tested with Java 11 >>>>>>>>> - Built binary artifacts with Java 8 >>>>>>>>> >>>>>>>>> >>>>>>>>> On Sun, Oct 9, 2022 at 3:42 PM Ryan Blue <b...@apache.org> wrote: >>>>>>>>> >>>>>>>>>> Hi Everyone, >>>>>>>>>> >>>>>>>>>> I propose that we release the following RC as the official Apache >>>>>>>>>> Iceberg 1.0.0 release. >>>>>>>>>> >>>>>>>>>> The commit ID is e2bb9ad7e792efca419fa7c4a1afde7c4c44fa01 >>>>>>>>>> * This corresponds to the tag: apache-iceberg-1.0.0-rc0 >>>>>>>>>> * >>>>>>>>>> https://github.com/apache/iceberg/commits/apache-iceberg-1.0.0-rc0 >>>>>>>>>> * >>>>>>>>>> https://github.com/apache/iceberg/tree/e2bb9ad7e792efca419fa7c4a1afde7c4c44fa01 >>>>>>>>>> >>>>>>>>>> The release tarball, signature, and checksums are here: >>>>>>>>>> * >>>>>>>>>> https://dist.apache.org/repos/dist/dev/iceberg/apache-iceberg-1.0.0-rc0 >>>>>>>>>> >>>>>>>>>> You can find the KEYS file here: >>>>>>>>>> * https://dist.apache.org/repos/dist/dev/iceberg/KEYS >>>>>>>>>> >>>>>>>>>> Convenience binary artifacts are staged on Nexus. The Maven >>>>>>>>>> repository URL is: >>>>>>>>>> * >>>>>>>>>> https://repository.apache.org/content/repositories/orgapacheiceberg-1106/ >>>>>>>>>> >>>>>>>>>> Please download, verify, and test. >>>>>>>>>> >>>>>>>>>> This release is based on the latest 0.14.1 release. It includes >>>>>>>>>> changes to remove deprecated APIs and the following additional bug >>>>>>>>>> fixes: >>>>>>>>>> * Increase metrics limit to 100 columns >>>>>>>>>> * Bump Spark patch versions for CVE-2022-33891 >>>>>>>>>> * Exclude Scala from Spark runtime Jars >>>>>>>>>> >>>>>>>>>> Please vote in the next 72 hours. >>>>>>>>>> >>>>>>>>>> [ ] +1 Release this as Apache Iceberg 1.0.0 >>>>>>>>>> [ ] +0 >>>>>>>>>> [ ] -1 Do not release this because... >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Ryan Blue >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Ryan Blue >>>>>>>>> >>>>>>>> -- *Rushan Jiang*