+1 (binding) 1. Download the source tarball, signature (.asc), and checksum (.sha512): OK 2. Import gpg keys: download KEYS and run gpg --import /path/to/downloaded/KEYS (optional if this hasn’t changed) : OK 3. Verify the signature by running: gpg --verify apache-iceberg-xx-incubating.tar.gz.asc: OK 4. Verify the checksum by running: shasum -a 256 -c apache-iceberg-0.12.1.tar.gz.sha512 apache-iceberg-0.12.1.tar.gz : OK 5. Untar the archive and go into the source directory: tar xzf apache-iceberg-xx-incubating.tar.gz && cd apache-iceberg-xx-incubating: OK 6. Run RAT checks to validate license headers: dev/check-license: OK 7. Build and test the project: ./gradlew build (use Java 8) : OK 8. Check the flink works fine by the following command line:
./bin/sql-client.sh embedded -j /Users/openinx/Downloads/apache-iceberg-0.12.1/flink-runtime/build/libs/iceberg-flink-runtime-0.12.1.jar shell CREATE CATALOG hadoop_prod WITH ( 'type'='iceberg', 'catalog-type'='hadoop', 'warehouse'='file:///Users/openinx/test/iceberg-warehouse' ); CREATE TABLE `hadoop_prod`.`default`.`flink_table` ( id BIGINT, data STRING ); INSERT INTO `hadoop_prod`.`default`.`flink_table` VALUES (1, 'AAA'); SELECT * FROM `hadoop_prod`.`default`.`flink_table`; +----+------+ | id | data | +----+------+ | 1 | AAA | +----+------+ 1 row in set Thanks all for the work. On Fri, Nov 5, 2021 at 2:20 PM Cheng Pan <[email protected]> wrote: > +1 (non-binding) > > The integration test based on the master branch of Apache Kyuubi > (Incubating) passed. > > https://github.com/apache/incubator-kyuubi/pull/1338 > > Thanks, > Cheng Pan > > On Fri, Nov 5, 2021 at 1:19 PM Kyle Bendickson <[email protected]> wrote: > > > > > > +1 (binding) > > > > - Validated checksums, signatures, and licenses > > - Ran all of the unit tests > > - Imported Files from Orc tables via Spark stored procedure, with > floating point type columns and inspected the metrics afterwards > > - Registered and used bucketed UDFs for various types such as integer > and byte > > - Created and dropped tables > > - Ran MERGE INTO queries using Spark DDL > > - Verified ability to read tables with parquet files with nested map > type schema from various versions (both before and after Parquet 1.11.0 -> > 1.11.1 upgrade) > > - Tried to set a tblproperty to null (received error as expected) > > - Full unit test suite > > - Ran several Flink queries, both batch and streaming. > > - Tested against a custom catalog > > > > My spark configuration was very similar to Ryan’s. I used Flink 1.12.1 > on a docker-compose setup via the Flink SQL client with 2 task managers. > > > > In addition to testing with a custom catalog, I also tested with HMS / > Hive catalog with HDFS as storage as well as Hadoop Catalog with data on > (local) HDFS. > > > > I’ve not gotten the Hive3 errors despite running unit tests several > times. > > > > - Kyle (@kbendick) > > > > > > On Thu, Nov 4, 2021 at 9:57 PM Daniel Weeks <[email protected]> wrote: > >> > >> +1 (binding) > >> > >> Verified sigs, sums, license, build and test. > >> > >> -Dan > >> > >> On Thu, Nov 4, 2021 at 4:30 PM Ryan Blue <[email protected]> wrote: > >>> > >>> +1 (binding) > >>> > >>> Validated checksums, checked signature, ran tests (still a couple > failing in Hive3) > >>> Staged binaries from the release tarball > >>> Tested Spark metadata tables > >>> Used rewrite_manifests stored procedure in Spark > >>> Updated to v2 using SET TBLPROPERTIES > >>> Dropped and added partition fields > >>> Replaced a table with itself using INSERT OVERWRITE > >>> Tested custom catalogs > >>> > >>> Here’s my Spark config script in case anyone else wants to validate: > >>> > >>> /home/blue/Apps/spark-3.1.1-bin-hadoop3.2/bin/spark-shell \ > >>> --conf spark.jars.repositories= > https://repository.apache.org/content/repositories/orgapacheiceberg-1019/ > \ > >>> --packages org.apache.iceberg:iceberg-spark3-runtime:0.12.1 \ > >>> --conf > spark.sql.extensions=org.apache.iceberg.spark.extensions.IcebergSparkSessionExtensions > \ > >>> --conf > spark.sql.catalog.local=org.apache.iceberg.spark.SparkCatalog \ > >>> --conf spark.sql.catalog.local.type=hadoop \ > >>> --conf > spark.sql.catalog.local.warehouse=/home/blue/tmp/hadoop-warehouse \ > >>> --conf spark.sql.catalog.local.default-namespace=default \ > >>> --conf > spark.sql.catalog.prodhive=org.apache.iceberg.spark.SparkCatalog \ > >>> --conf spark.sql.catalog.prodhive.type=hive \ > >>> --conf > spark.sql.catalog.prodhive.warehouse=/home/blue/tmp/prod-warehouse \ > >>> --conf spark.sql.catalog.prodhive.default-namespace=default \ > >>> --conf spark.sql.defaultCatalog=local > >>> > >>> > >>> On Thu, Nov 4, 2021 at 1:02 PM Jack Ye <[email protected]> wrote: > >>>> > >>>> +1, non-binding > >>>> > >>>> ran checksum, build, unit tests, AWS integration tests and verified > fixes in EMR 6.4.0. > >>>> > >>>> Best, > >>>> Jack Ye > >>>> > >>>> On Tue, Nov 2, 2021 at 7:16 PM Kyle Bendickson <[email protected]> > wrote: > >>>>> > >>>>> Hi everyone, > >>>>> > >>>>> > >>>>> I propose the following RC to be released as the official Apache > Iceberg 0.12.1 release. > >>>>> > >>>>> > >>>>> The commit id is d4052a73f14b63e1f519aaa722971dc74f8c9796 > >>>>> > >>>>> * This corresponds to the tag: apache-iceberg-0.12.1-rc0 > >>>>> > >>>>> * > https://github.com/apache/iceberg/commits/apache-iceberg-0.12.1-rc0 > >>>>> > >>>>> * > https://github.com/apache/iceberg/tree/d4052a73f14b63e1f519aaa722971dc74f8c9796 > >>>>> > >>>>> > >>>>> The release tarball, signature, and checksums are here: > >>>>> > >>>>> * > https://dist.apache.org/repos/dist/dev/iceberg/apache-iceberg-0.12.1-rc0/ > >>>>> > >>>>> > >>>>> You can find the KEYS file here: > >>>>> > >>>>> * https://dist.apache.org/repos/dist/dev/iceberg/KEYS > >>>>> > >>>>> > >>>>> Convenience binary artifacts are staged in Nexus. The Maven > repository URL is: > >>>>> > >>>>> * > https://repository.apache.org/content/repositories/orgapacheiceberg-1019/ > >>>>> > >>>>> > >>>>> This release includes the following changes: > >>>>> > >>>>> > https://github.com/apache/iceberg/compare/apache-iceberg-0.12.0...apache-iceberg-0.12.1-rc0 > >>>>> > >>>>> > >>>>> Please download, verify, and test. > >>>>> > >>>>> > >>>>> Please vote in the next 72 hours. > >>>>> > >>>>> > >>>>> [ ] +1 Release this as Apache Iceberg <VERSION> > >>>>> > >>>>> [ ] +0 > >>>>> > >>>>> [ ] -1 Do not release this because... > >>>>> > >>>>> -- > >>>>> Best, > >>>>> Kyle Bendickson > >>>>> Github: @kbendick > >>> > >>> > >>> > >>> -- > >>> Ryan Blue > >>> Tabular >
