+1 (non-binding)

The integration test based on the master branch of Apache Kyuubi
(Incubating) passed.

https://github.com/apache/incubator-kyuubi/pull/1338

Thanks,
Cheng Pan

On Fri, Nov 5, 2021 at 1:19 PM Kyle Bendickson <k...@tabular.io> wrote:
>
>
> +1 (binding)
>
> - Validated checksums, signatures, and licenses
> -  Ran all of the unit tests
> - Imported Files from Orc tables via Spark stored procedure, with floating 
> point type columns and inspected the metrics afterwards
> - Registered and used bucketed UDFs for various types such as integer and byte
> - Created and dropped tables
> - Ran MERGE INTO queries using Spark DDL
> - Verified ability to read tables with parquet files with nested map type 
> schema from various versions (both before and after Parquet 1.11.0 -> 1.11.1 
> upgrade)
> - Tried to set a tblproperty to null (received error as expected)
> - Full unit test suite
> - Ran several Flink queries, both batch and streaming.
> - Tested against a custom catalog
>
> My spark configuration was very similar to Ryan’s. I used Flink 1.12.1 on a 
> docker-compose setup via the Flink SQL client with 2 task managers.
>
> In addition to testing with a custom catalog, I also tested with HMS / Hive 
> catalog with HDFS as storage as well as Hadoop Catalog with data on (local) 
> HDFS.
>
> I’ve not gotten the Hive3 errors despite running unit tests several times.
>
> - Kyle (@kbendick)
>
>
> On Thu, Nov 4, 2021 at 9:57 PM Daniel Weeks <dwe...@apache.org> wrote:
>>
>> +1 (binding)
>>
>> Verified sigs, sums, license, build and test.
>>
>> -Dan
>>
>> On Thu, Nov 4, 2021 at 4:30 PM Ryan Blue <b...@tabular.io> wrote:
>>>
>>> +1 (binding)
>>>
>>> Validated checksums, checked signature, ran tests (still a couple failing 
>>> in Hive3)
>>> Staged binaries from the release tarball
>>> Tested Spark metadata tables
>>> Used rewrite_manifests stored procedure in Spark
>>> Updated to v2 using SET TBLPROPERTIES
>>> Dropped and added partition fields
>>> Replaced a table with itself using INSERT OVERWRITE
>>> Tested custom catalogs
>>>
>>> Here’s my Spark config script in case anyone else wants to validate:
>>>
>>> /home/blue/Apps/spark-3.1.1-bin-hadoop3.2/bin/spark-shell \
>>>     --conf 
>>> spark.jars.repositories=https://repository.apache.org/content/repositories/orgapacheiceberg-1019/
>>>  \
>>>     --packages org.apache.iceberg:iceberg-spark3-runtime:0.12.1 \
>>>     --conf 
>>> spark.sql.extensions=org.apache.iceberg.spark.extensions.IcebergSparkSessionExtensions
>>>  \
>>>     --conf spark.sql.catalog.local=org.apache.iceberg.spark.SparkCatalog \
>>>     --conf spark.sql.catalog.local.type=hadoop \
>>>     --conf 
>>> spark.sql.catalog.local.warehouse=/home/blue/tmp/hadoop-warehouse \
>>>     --conf spark.sql.catalog.local.default-namespace=default \
>>>     --conf spark.sql.catalog.prodhive=org.apache.iceberg.spark.SparkCatalog 
>>> \
>>>     --conf spark.sql.catalog.prodhive.type=hive \
>>>     --conf 
>>> spark.sql.catalog.prodhive.warehouse=/home/blue/tmp/prod-warehouse \
>>>     --conf spark.sql.catalog.prodhive.default-namespace=default \
>>>     --conf spark.sql.defaultCatalog=local
>>>
>>>
>>> On Thu, Nov 4, 2021 at 1:02 PM Jack Ye <yezhao...@gmail.com> wrote:
>>>>
>>>> +1, non-binding
>>>>
>>>> ran checksum, build, unit tests, AWS integration tests and verified fixes 
>>>> in EMR 6.4.0.
>>>>
>>>> Best,
>>>> Jack Ye
>>>>
>>>> On Tue, Nov 2, 2021 at 7:16 PM Kyle Bendickson <k...@tabular.io> wrote:
>>>>>
>>>>> Hi everyone,
>>>>>
>>>>>
>>>>> I propose the following RC to be released as the official Apache Iceberg 
>>>>> 0.12.1 release.
>>>>>
>>>>>
>>>>> The commit id is d4052a73f14b63e1f519aaa722971dc74f8c9796
>>>>>
>>>>> * This corresponds to the tag: apache-iceberg-0.12.1-rc0
>>>>>
>>>>> * https://github.com/apache/iceberg/commits/apache-iceberg-0.12.1-rc0
>>>>>
>>>>> * 
>>>>> https://github.com/apache/iceberg/tree/d4052a73f14b63e1f519aaa722971dc74f8c9796
>>>>>
>>>>>
>>>>> The release tarball, signature, and checksums are here:
>>>>>
>>>>> * 
>>>>> https://dist.apache.org/repos/dist/dev/iceberg/apache-iceberg-0.12.1-rc0/
>>>>>
>>>>>
>>>>> You can find the KEYS file here:
>>>>>
>>>>> * https://dist.apache.org/repos/dist/dev/iceberg/KEYS
>>>>>
>>>>>
>>>>> Convenience binary artifacts are staged in Nexus. The Maven repository 
>>>>> URL is:
>>>>>
>>>>> * 
>>>>> https://repository.apache.org/content/repositories/orgapacheiceberg-1019/
>>>>>
>>>>>
>>>>> This release includes the following changes:
>>>>>
>>>>> https://github.com/apache/iceberg/compare/apache-iceberg-0.12.0...apache-iceberg-0.12.1-rc0
>>>>>
>>>>>
>>>>> Please download, verify, and test.
>>>>>
>>>>>
>>>>> Please vote in the next 72 hours.
>>>>>
>>>>>
>>>>> [ ] +1 Release this as Apache Iceberg <VERSION>
>>>>>
>>>>> [ ] +0
>>>>>
>>>>> [ ] -1 Do not release this because...
>>>>>
>>>>> --
>>>>> Best,
>>>>> Kyle Bendickson
>>>>> Github: @kbendick
>>>
>>>
>>>
>>> --
>>> Ryan Blue
>>> Tabular

Reply via email to