+1 (non-binding)
- Verified Signature
- Verified Checksum
- Verified License
- Verified build and test with Java 11
- Test with Apache Spark 3.3

Best,

Yufei

`This is not a contribution`


On Mon, Oct 10, 2022 at 7:07 PM Rushan Jiang <rush...@andrew.cmu.edu> wrote:

> +1 (non-binding)
> - Verify Signature
> - Verify Checksums
> - Verify License Documentation
> - Verify Build and Test in Java 8 and 11
>
> On Mon, Oct 10, 2022 at 7:11 PM Daniel Weeks <dwe...@apache.org> wrote:
>
>> +1 (binding)
>>
>> Verified sigs/sums/license/build/test (java 8)
>>
>> Everything checks out,
>> -Dan
>>
>> On Mon, Oct 10, 2022 at 3:25 PM Steven Wu <stevenz...@gmail.com> wrote:
>>
>>> Never mind. Missed the information that this 1.0.0 is based on the
>>> latest 0.14.1 release, which doesn't contain the PR 5318. I thought it was
>>> based on the latest master branch.
>>>
>>> +1 (non-binding)
>>> - Verify signature
>>> - Verify checksum
>>> - Tried SQL insert and query with Flink 1.15
>>>
>>> On Mon, Oct 10, 2022 at 3:20 PM Steven Wu <stevenz...@gmail.com> wrote:
>>>
>>>> Ryan,
>>>>
>>>> It seems that this PR (merged on July 28) for the Flink FLIP-27 source
>>>> is not included. https://github.com/apache/iceberg/pull/5318
>>>>
>>>> The commit ID still contains the old file in the old location. The new
>>>> location should be "flink/source/IcebergTableSource.java" with FLIP-27
>>>> config support.
>>>>
>>>> https://github.com/apache/iceberg/blob/e2bb9ad7e792efca419fa7c4a1afde7c4c44fa01/flink/v1.15/flink/src/main/java/org/apache/iceberg/flink/IcebergTableSource.java
>>>>
>>>> Thanks,
>>>> Steven
>>>>
>>>> On Mon, Oct 10, 2022 at 11:37 AM Szehon Ho <szehon.apa...@gmail.com>
>>>> wrote:
>>>>
>>>>> Whoops, sorry for the noise, I made a typo and was using the wrong
>>>>> scala version of the iceberg-spark-runtime jar, this works.
>>>>>
>>>>> +1 (non-binding)
>>>>> - Verify signature
>>>>> - Verify checksum
>>>>> - Verify license documentation
>>>>> - Tried with Spark 3.3
>>>>> - Ran unit tests
>>>>>
>>>>> Thanks
>>>>> Szehon
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On Mon, Oct 10, 2022 at 11:26 AM Szehon Ho <szehon.apa...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> I get a NoClassDefFoundError from IcebergSparkExtensions when running
>>>>>> Spark 3.3, with iceberg-spark-runtime-3.3_2.12-1.0.0.jar.  I noticed this
>>>>>> jar doesn't contain scala classes, unlike previous jars
>>>>>> iceberg-spark-runtime-3.3_2.12-0.14.1.jar.
>>>>>>
>>>>>> scala> spark.sql("show databases").show
>>>>>> java.lang.NoClassDefFoundError: scala/collection/SeqOps
>>>>>>   at
>>>>>> org.apache.iceberg.spark.extensions.IcebergSparkSessionExtensions.$anonfun$apply$2(IcebergSparkSessionExtensions.scala:50)
>>>>>>   at
>>>>>> org.apache.spark.sql.SparkSessionExtensions.$anonfun$buildResolutionRules$1(SparkSessionExtensions.scala:152)
>>>>>>   at
>>>>>> scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:286)
>>>>>>   at
>>>>>> scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)
>>>>>>   at
>>>>>> scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)
>>>>>>   at
>>>>>> scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)
>>>>>>   at scala.collection.TraversableLike.map(TraversableLike.scala:286)
>>>>>>   at scala.collection.TraversableLike.map$(TraversableLike.scala:279)
>>>>>>   at scala.collection.AbstractTraversable.map(Traversable.scala:108)
>>>>>>   at
>>>>>> org.apache.spark.sql.SparkSessionExtensions.buildResolutionRules(SparkSessionExtensions.scala:152)
>>>>>>   at
>>>>>> org.apache.spark.sql.internal.BaseSessionStateBuilder.customResolutionRules(BaseSessionStateBuilder.scala:216)
>>>>>>   at
>>>>>> org.apache.spark.sql.hive.HiveSessionStateBuilder$$anon$1.<init>(HiveSessionStateBuilder.scala:94)
>>>>>>   at
>>>>>> org.apache.spark.sql.hive.HiveSessionStateBuilder.analyzer(HiveSessionStateBuilder.scala:85)
>>>>>>   at
>>>>>> org.apache.spark.sql.internal.BaseSessionStateBuilder.$anonfun$build$2(BaseSessionStateBuilder.scala:360)
>>>>>>   at
>>>>>> org.apache.spark.sql.internal.SessionState.analyzer$lzycompute(SessionState.scala:87)
>>>>>>   at
>>>>>> org.apache.spark.sql.internal.SessionState.analyzer(SessionState.scala:87)
>>>>>>   at
>>>>>> org.apache.spark.sql.execution.QueryExecution.$anonfun$analyzed$1(QueryExecution.scala:76)
>>>>>>   at
>>>>>> org.apache.spark.sql.catalyst.QueryPlanningTracker.measurePhase(QueryPlanningTracker.scala:111)
>>>>>>   at
>>>>>> org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$2(QueryExecution.scala:185)
>>>>>>   at
>>>>>> org.apache.spark.sql.execution.QueryExecution$.withInternalError(QueryExecution.scala:510)
>>>>>>   at
>>>>>> org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$1(QueryExecution.scala:185)
>>>>>>   at
>>>>>> org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779)
>>>>>>   at
>>>>>> org.apache.spark.sql.execution.QueryExecution.executePhase(QueryExecution.scala:184)
>>>>>>   at
>>>>>> org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:76)
>>>>>>   at
>>>>>> org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:74)
>>>>>>   at
>>>>>> org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:66)
>>>>>>   at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:99)
>>>>>>   at
>>>>>> org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779)
>>>>>>   at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:97)
>>>>>>   at
>>>>>> org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:622)
>>>>>>   at
>>>>>> org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:779)
>>>>>>   at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:617)
>>>>>>   ... 47 elided
>>>>>> Caused by: java.lang.ClassNotFoundException: scala.collection.SeqOps
>>>>>>   at
>>>>>> java.base/jdk.internal.loader.BuiltinClassLoader.loadClass(BuiltinClassLoader.java:581)
>>>>>>   at
>>>>>> java.base/jdk.internal.loader.ClassLoaders$AppClassLoader.loadClass(ClassLoaders.java:178)
>>>>>>   at java.base/java.lang.ClassLoader.loadClass(ClassLoader.java:522)
>>>>>>   ... 79 more
>>>>>>
>>>>>> Note, I usually verify by copying the spark-runtime jar to spark jars
>>>>>> dir (can't usually get --packages flag to work as indicated on
>>>>>> https://iceberg.apache.org/how-to-release/#verifying-with-spark, as
>>>>>> version is not released yet), so let me know if I am using the wrong jar?
>>>>>>
>>>>>> Thanks
>>>>>> Szehon
>>>>>>
>>>>>> On Mon, Oct 10, 2022 at 9:22 AM Eduard Tudenhoefner <
>>>>>> edu...@tabular.io> wrote:
>>>>>>
>>>>>>> +1 (non-binding)
>>>>>>>
>>>>>>>    - validated checksum and signature
>>>>>>>    - checked license docs & ran RAT checks
>>>>>>>    - ran build and tests with JDK11
>>>>>>>
>>>>>>>
>>>>>>> Eduard
>>>>>>>
>>>>>>> On Mon, Oct 10, 2022 at 8:01 AM Ajantha Bhat <ajanthab...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> +1 (non-binding)
>>>>>>>>
>>>>>>>>
>>>>>>>>    - Verified the Spark runtime jar contents.
>>>>>>>>    - Checked license docs, ran RAT checks.
>>>>>>>>    - Validated checksum and signature.
>>>>>>>>
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> Ajantha
>>>>>>>>
>>>>>>>> On Mon, Oct 10, 2022 at 10:45 AM Prashant Singh <
>>>>>>>> prashant010...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Hello Everyone,
>>>>>>>>>
>>>>>>>>> Wanted to know your thoughts on whether we should also include the
>>>>>>>>> following bug fixes in this release as well:
>>>>>>>>>
>>>>>>>>> 1. MERGE INTO nullability fix, leads to query failure otherwise:
>>>>>>>>> *Reported instances :*
>>>>>>>>> a.
>>>>>>>>> https://stackoverflow.com/questions/73424454/spark-iceberg-merge-into-issue-caused-by-org-apache-spark-sql-analysisexcep
>>>>>>>>> b. https://github.com/apache/iceberg/issues/5739
>>>>>>>>> c.
>>>>>>>>> https://github.com/apache/iceberg/issues/5424#issuecomment-1220688298
>>>>>>>>>
>>>>>>>>> *PR's (Merged):*
>>>>>>>>> a. https://github.com/apache/iceberg/pull/5880
>>>>>>>>> b. https://github.com/apache/iceberg/pull/5679
>>>>>>>>>
>>>>>>>>> 2.  QueryFailure when running RewriteManifestProcedure on Date /
>>>>>>>>> Timestamp partitioned table when
>>>>>>>>> `spark.sql.datetime.java8API.enabled` is true.
>>>>>>>>> *Reported instances :*
>>>>>>>>> a. https://github.com/apache/iceberg/issues/5104
>>>>>>>>> b.
>>>>>>>>> https://apache-iceberg.slack.com/archives/C025PH0G1D4/p1663982635731469
>>>>>>>>>
>>>>>>>>> *PR* :
>>>>>>>>> a. https://github.com/apache/iceberg/pull/5860
>>>>>>>>>
>>>>>>>>> Regards,
>>>>>>>>> Prashant Singh
>>>>>>>>>
>>>>>>>>> On Mon, Oct 10, 2022 at 4:15 AM Ryan Blue <b...@apache.org> wrote:
>>>>>>>>>
>>>>>>>>>> +1 (binding)
>>>>>>>>>>
>>>>>>>>>>    - Checked license docs, ran RAT checks
>>>>>>>>>>    - Validated checksum and signature
>>>>>>>>>>    - Built and tested with Java 11
>>>>>>>>>>    - Built binary artifacts with Java 8
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Sun, Oct 9, 2022 at 3:42 PM Ryan Blue <b...@apache.org> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi Everyone,
>>>>>>>>>>>
>>>>>>>>>>> I propose that we release the following RC as the official
>>>>>>>>>>> Apache Iceberg 1.0.0 release.
>>>>>>>>>>>
>>>>>>>>>>> The commit ID is e2bb9ad7e792efca419fa7c4a1afde7c4c44fa01
>>>>>>>>>>> * This corresponds to the tag: apache-iceberg-1.0.0-rc0
>>>>>>>>>>> *
>>>>>>>>>>> https://github.com/apache/iceberg/commits/apache-iceberg-1.0.0-rc0
>>>>>>>>>>> *
>>>>>>>>>>> https://github.com/apache/iceberg/tree/e2bb9ad7e792efca419fa7c4a1afde7c4c44fa01
>>>>>>>>>>>
>>>>>>>>>>> The release tarball, signature, and checksums are here:
>>>>>>>>>>> *
>>>>>>>>>>> https://dist.apache.org/repos/dist/dev/iceberg/apache-iceberg-1.0.0-rc0
>>>>>>>>>>>
>>>>>>>>>>> You can find the KEYS file here:
>>>>>>>>>>> * https://dist.apache.org/repos/dist/dev/iceberg/KEYS
>>>>>>>>>>>
>>>>>>>>>>> Convenience binary artifacts are staged on Nexus. The Maven
>>>>>>>>>>> repository URL is:
>>>>>>>>>>> *
>>>>>>>>>>> https://repository.apache.org/content/repositories/orgapacheiceberg-1106/
>>>>>>>>>>>
>>>>>>>>>>> Please download, verify, and test.
>>>>>>>>>>>
>>>>>>>>>>> This release is based on the latest 0.14.1 release. It includes
>>>>>>>>>>> changes to remove deprecated APIs and the following additional bug 
>>>>>>>>>>> fixes:
>>>>>>>>>>> * Increase metrics limit to 100 columns
>>>>>>>>>>> * Bump Spark patch versions for CVE-2022-33891
>>>>>>>>>>> * Exclude Scala from Spark runtime Jars
>>>>>>>>>>>
>>>>>>>>>>> Please vote in the next 72 hours.
>>>>>>>>>>>
>>>>>>>>>>> [ ] +1 Release this as Apache Iceberg 1.0.0
>>>>>>>>>>> [ ] +0
>>>>>>>>>>> [ ] -1 Do not release this because...
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> Ryan Blue
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Ryan Blue
>>>>>>>>>>
>>>>>>>>>
>
> --
> *Rushan Jiang*
>

Reply via email to