+1 (non-binding) Verified all steps.
Just a bit of heads up in case anyone encounters the same issue, when running unit tests on my laptop, my local custom HADOOP_HOME and HADOOP_CONF_DIR settings somehow caused 30 iceberg-flink unit tests to fail and "TestFlinkIcebergSink" to hang. After unsetting these 2 env variables, all unit tests passed. I don't think it should block the RC. On Thu, Nov 5, 2020 at 8:50 AM Daniel Weeks <dwe...@apache.org> wrote: > +1 (binding) > > Verified all steps. > > On Thu, Nov 5, 2020 at 8:31 AM Anton Okolnychyi > <aokolnyc...@apple.com.invalid> wrote: > >> +1 (binding) >> >> Verified the RC locally. >> >> - Anton >> >> On 5 Nov 2020, at 08:15, Edgar Rodriguez < >> edgar.rodrig...@airbnb.com.INVALID> wrote: >> >> +1 non-binding for RC4. Tested with internal tests in cluster, validated >> Spark write and Hive reads. >> >> On Thu, Nov 5, 2020 at 5:56 AM Mass Dosage <massdos...@gmail.com> wrote: >> >>> +1 non-binding on RC4. I tested out the Hive read path on a distributed >>> cluster using HadoopTables. >>> >>> On Thu, 5 Nov 2020 at 04:46, Dongjoon Hyun <dongjoon.h...@gmail.com> >>> wrote: >>> >>>> +1 for 0.10.0 RC4. >>>> >>>> Bests, >>>> Dongjoon. >>>> >>>> On Wed, Nov 4, 2020 at 7:17 PM Jingsong Li <jingsongl...@gmail.com> >>>> wrote: >>>> >>>>> +1 >>>>> >>>>> 1. Download the source tarball, signature (.asc), and checksum >>>>> (.sha512): OK >>>>> 2. Import gpg keys: download KEYS and run gpg --import >>>>> /path/to/downloaded/KEYS (optional if this hasn’t changed) : OK >>>>> 3. Verify the signature by running: gpg --verify >>>>> apache-iceberg-xx.tar.gz.asc: OK >>>>> 4. Verify the checksum by running: sha512sum -c >>>>> apache-iceberg-xx.tar.gz.sha512 : OK >>>>> 5. Untar the archive and go into the source directory: tar xzf >>>>> apache-iceberg-xx.tar.gz && cd apache-iceberg-xx: OK >>>>> 6. Run RAT checks to validate license headers: dev/check-license: OK >>>>> 7. Build and test the project: ./gradlew build (use Java 8) : OK >>>>> >>>>> Best, >>>>> Jingsong >>>>> >>>>> On Thu, Nov 5, 2020 at 7:38 AM Ryan Blue <rb...@netflix.com.invalid> >>>>> wrote: >>>>> >>>>>> +1 >>>>>> >>>>>> - Validated checksum and signature >>>>>> - Ran license checks >>>>>> - Built and ran tests >>>>>> - Queried a Hadoop FS table created with 0.9.0 in Spark 3.0.1 >>>>>> - Created a Hive table from Spark 3.0.1 >>>>>> - Tested metadata tables from Spark >>>>>> - Tested Hive and Hadoop table reads in Hive 2.3.7 >>>>>> >>>>>> I was able to read both Hadoop and Hive tables created in Spark from >>>>>> Hive using: >>>>>> >>>>>> add jar /home/blue/Downloads/iceberg-hive-runtime-0.10.0.jar; >>>>>> create external table hadoop_table >>>>>> stored by 'org.apache.iceberg.mr.hive.HiveIcebergStorageHandler' >>>>>> location 'file:/home/blue/tmp/hadoop-warehouse/default/test'; >>>>>> select * from hadoop_table; >>>>>> >>>>>> set iceberg.mr.catalog=hive; >>>>>> select * from hive_table; >>>>>> >>>>>> The hive_table needed engine.hive.enabled=true set in table >>>>>> properties by Spark using: >>>>>> >>>>>> alter table hive_table set tblproperties ('engine.hive.enabled'='true') >>>>>> >>>>>> Hive couldn’t read the #snapshots metadata table for Hadoop. It >>>>>> failed with this error: >>>>>> >>>>>> Failed with exception >>>>>> java.io.IOException:org.apache.hadoop.hive.ql.metadata.HiveException: >>>>>> java.lang.ClassCastException: java.lang.Long cannot be cast to >>>>>> java.time.OffsetDateTime >>>>>> >>>>>> I also couldn’t read the Hadoop table once iceberg.mr.catalog was >>>>>> set in my environment, so I think we have a bit more work to do to clean >>>>>> up >>>>>> Hive table configuration. >>>>>> >>>>>> On Wed, Nov 4, 2020 at 12:54 AM Ryan Murray <rym...@dremio.com> >>>>>> wrote: >>>>>> >>>>>>> +1 (non-binding) >>>>>>> >>>>>>> 1. Download the source tarball, signature (.asc), and checksum >>>>>>> (.sha512): OK >>>>>>> 2. Import gpg keys: download KEYS and run gpg --import >>>>>>> /path/to/downloaded/KEYS (optional if this hasn’t changed) : OK >>>>>>> 3. Verify the signature by running: gpg --verify >>>>>>> apache-iceberg-xx.tar.gz.asc: I got a warning "gpg: WARNING: This key >>>>>>> is >>>>>>> not certified with a trusted signature! gpg: There is no >>>>>>> indication that the signature belongs to the owner." but it passed >>>>>>> 4. Verify the checksum by running: sha512sum -c >>>>>>> apache-iceberg-xx.tar.gz.sha512 : OK >>>>>>> 5. Untar the archive and go into the source directory: tar xzf >>>>>>> apache-iceberg-xx.tar.gz && cd apache-iceberg-xx: OK >>>>>>> 6. Run RAT checks to validate license headers: dev/check-license: OK >>>>>>> 7. Build and test the project: ./gradlew build (use Java 8 & Java >>>>>>> 11) : OK >>>>>>> >>>>>>> >>>>>>> On Wed, Nov 4, 2020 at 2:56 AM OpenInx <open...@gmail.com> wrote: >>>>>>> >>>>>>>> +1 for 0.10.0 RC4 >>>>>>>> >>>>>>>> 1. Download the source tarball, signature (.asc), and checksum >>>>>>>> (.sha512): OK >>>>>>>> 2. Import gpg keys: download KEYS and run gpg --import >>>>>>>> /path/to/downloaded/KEYS (optional if this hasn’t changed) : OK >>>>>>>> 3. Verify the signature by running: gpg --verify >>>>>>>> apache-iceberg-xx.tar.gz.asc: OK >>>>>>>> 4. Verify the checksum by running: sha512sum -c >>>>>>>> apache-iceberg-xx.tar.gz.sha512 : OK >>>>>>>> 5. Untar the archive and go into the source directory: tar xzf >>>>>>>> apache-iceberg-xx.tar.gz && cd apache-iceberg-xx: OK >>>>>>>> 6. Run RAT checks to validate license headers: dev/check-license: OK >>>>>>>> 7. Build and test the project: ./gradlew build (use Java 8) : OK >>>>>>>> >>>>>>>> On Wed, Nov 4, 2020 at 8:25 AM Anton Okolnychyi < >>>>>>>> aokolnyc...@apple.com.invalid> wrote: >>>>>>>> >>>>>>>>> Hi everyone, >>>>>>>>> >>>>>>>>> I propose the following RC to be released as official Apache >>>>>>>>> Iceberg 0.10.0 release. >>>>>>>>> >>>>>>>>> The commit id is d39fad00b7dded98121368309f381473ec21e85f >>>>>>>>> * This corresponds to the tag: apache-iceberg-0.10.0-rc4 >>>>>>>>> * >>>>>>>>> https://github.com/apache/iceberg/commits/apache-iceberg-0.10.0-rc4 >>>>>>>>> * >>>>>>>>> https://github.com/apache/iceberg/tree/d39fad00b7dded98121368309f381473ec21e85f >>>>>>>>> >>>>>>>>> The release tarball, signature, and checksums are here: >>>>>>>>> * >>>>>>>>> https://dist.apache.org/repos/dist/dev/iceberg/apache-iceberg-0.10.0-rc4/ >>>>>>>>> >>>>>>>>> You can find the KEYS file here (make sure to import the new key >>>>>>>>> that was used to sign the release): >>>>>>>>> * https://dist.apache.org/repos/dist/dev/iceberg/KEYS >>>>>>>>> >>>>>>>>> Convenience binary artifacts are staged in Nexus. The Maven >>>>>>>>> repository URL is: >>>>>>>>> * >>>>>>>>> https://repository.apache.org/content/repositories/orgapacheiceberg-1012 >>>>>>>>> >>>>>>>>> This release includes important changes: >>>>>>>>> >>>>>>>>> * Flink support >>>>>>>>> * Hive read support >>>>>>>>> * ORC support fixes and improvements >>>>>>>>> * Application of row-level delete files on read >>>>>>>>> * Snapshot partition summary >>>>>>>>> * Ability to load LocationProvider dynamically >>>>>>>>> * Sort spec >>>>>>>>> >>>>>>>>> Please download, verify, and test. >>>>>>>>> >>>>>>>>> Please vote in the next 72 hours. >>>>>>>>> >>>>>>>>> [ ] +1 Release this as Apache Iceberg 0.10.0 >>>>>>>>> [ ] +0 >>>>>>>>> [ ] -1 Do not release this because… >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> Anton >>>>>>>>> >>>>>>>> >>>>>> >>>>>> -- >>>>>> Ryan Blue >>>>>> Software Engineer >>>>>> Netflix >>>>>> >>>>> >>>>> >>>>> -- >>>>> Best, Jingsong Lee >>>>> >>>> >> >> -- >> Edgar R >> >> >> -- John Zhuge