https://github.com/apache/iceberg/pull/2954 should be ready to merge. The CI passed.
On Mon, Aug 9, 2021 at 9:08 AM Wing Yew Poon <wyp...@cloudera.com> wrote: > Ryan, > Thanks for the review. Let me look into implementing your refactoring > suggestion. > - Wing Yew > > > On Mon, Aug 9, 2021 at 8:41 AM Ryan Blue <b...@tabular.io> wrote: > >> Yeah, I agree. We should fix this for the 0.12.0 release. That said, I >> plan to continue testing this RC because it won't change that much since >> this affects the Spark extensions in 3.1. Other engines and Spark 3.0 or >> older should be fine. >> >> I left a comment on the PR. I think it looks good, but we should try to >> refactor to make sure we don't have more issues like this. I think when we >> update our extensions to be compatible with multiple Spark versions, we >> should introduce a factory method to create the Catalyst plan node and use >> that everywhere. That will hopefully cut down on the number of times this >> happens. >> >> Thank you, Wing Yew! >> >> On Sun, Aug 8, 2021 at 2:52 PM Carl Steinbach <cwsteinb...@gmail.com> >> wrote: >> >>> Hi Wing Yew, >>> >>> I will create a new RC once this patch is committed. >>> >>> Thanks. >>> >>> - Carl >>> >>> On Sat, Aug 7, 2021 at 4:29 PM Wing Yew Poon <wyp...@cloudera.com.invalid> >>> wrote: >>> >>>> Sorry to bring this up so late, but this just came up: there is a Spark >>>> 3.1 (runtime) compatibility issue (not found by existing tests), which I >>>> have a fix for in https://github.com/apache/iceberg/pull/2954. I think >>>> it would be really helpful if it can go into 0.12.0. >>>> - Wing Yew >>>> >>>> >>>> On Fri, Aug 6, 2021 at 11:36 AM Jack Ye <yezhao...@gmail.com> wrote: >>>> >>>>> +1 (non-binding) >>>>> >>>>> Verified release test and AWS integration test, issue found in test >>>>> but not blocking for release ( >>>>> https://github.com/apache/iceberg/pull/2948) >>>>> >>>>> Verified Spark 3.1 and 3.0 operations and new SQL extensions and >>>>> procedures on EMR. >>>>> >>>>> Thanks, >>>>> Jack Ye >>>>> >>>>> On Fri, Aug 6, 2021 at 1:19 AM Kyle Bendickson <kjbendick...@gmail.com> >>>>> wrote: >>>>> >>>>>> +1 (binding) >>>>>> >>>>>> I verified: >>>>>> - KEYS signature & checksum >>>>>> - ./gradlew clean build (tests, etc) >>>>>> - Ran Spark jobs on Kubernetes after building from the tarball at >>>>>> https://dist.apache.org/repos/dist/dev/iceberg/apache-iceberg-0.12.0-rc2/ >>>>>> - Spark 3.1.1 batch jobs against both Hadoop and Hive tables, >>>>>> using HMS for Hive catalog >>>>>> - Verified default FileIO and S3FileIO >>>>>> - Basic read and writes >>>>>> - Jobs using Spark procedures (remove unreachable files) >>>>>> - Special mention: verified that Spark catalogs can override hadoop >>>>>> configurations using configs prefixed with >>>>>> "spark.sql.catalog.(catalog-name).hadoop." >>>>>> - one of my contributions to this release that has been asked >>>>>> about by several customers internally >>>>>> - tested using >>>>>> `spark.sql.catalog.(catalog-name).hadoop.fs.s3a.impl` for two catalogs, >>>>>> both values respected as opposed to the default globally configured value >>>>>> >>>>>> Thank you Carl! >>>>>> >>>>>> - Kyle, Data OSS Dev @ Apple =) >>>>>> >>>>>> On Thu, Aug 5, 2021 at 11:49 PM Szehon Ho <szehon.apa...@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> +1 (non-binding) >>>>>>> >>>>>>> * Verify Signature Keys >>>>>>> * Verify Checksum >>>>>>> * dev/check-license >>>>>>> * Build >>>>>>> * Run tests (though some timeout failures, on Hive MR test..) >>>>>>> >>>>>>> Thanks >>>>>>> Szehon >>>>>>> >>>>>>> On Thu, Aug 5, 2021 at 2:23 PM Daniel Weeks <dwe...@apache.org> >>>>>>> wrote: >>>>>>> >>>>>>>> +1 (binding) >>>>>>>> >>>>>>>> I verified sigs/sums, license, build, and test >>>>>>>> >>>>>>>> -Dan >>>>>>>> >>>>>>>> On Wed, Aug 4, 2021 at 2:53 PM Ryan Murray <rym...@gmail.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> After some wrestling w/ Spark I discovered that the problem was >>>>>>>>> with my test. Some SparkSession apis changed. so all good here now. >>>>>>>>> >>>>>>>>> +1 (non-binding) >>>>>>>>> >>>>>>>>> On Wed, Aug 4, 2021 at 11:29 PM Ryan Murray <rym...@gmail.com> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Thanks for the help Carl, got it sorted out. The gpg check now >>>>>>>>>> works. For those who were interested I used a canned wget command in >>>>>>>>>> my >>>>>>>>>> history and it pulled the RC0 :-) >>>>>>>>>> >>>>>>>>>> Will have a PR to fix the Nessie Catalog soon. >>>>>>>>>> >>>>>>>>>> Best, >>>>>>>>>> Ryan >>>>>>>>>> >>>>>>>>>> On Wed, Aug 4, 2021 at 9:21 PM Carl Steinbach < >>>>>>>>>> cwsteinb...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> Hi Ryan, >>>>>>>>>>> >>>>>>>>>>> Can you please run the following command to see which keys in >>>>>>>>>>> your public keyring are associated with my UID? >>>>>>>>>>> >>>>>>>>>>> % gpg --list-keys c...@apache.org >>>>>>>>>>> pub rsa4096/5A5C7F6EB9542945 2021-07-01 [SC] >>>>>>>>>>> 160F51BE45616B94103ED24D5A5C7F6EB9542945 >>>>>>>>>>> uid [ultimate] Carl W. Steinbach (CODE SIGNING >>>>>>>>>>> KEY) <c...@apache.org> >>>>>>>>>>> sub rsa4096/4158EB8A4F03D2AA 2021-07-01 [E] >>>>>>>>>>> >>>>>>>>>>> Thanks. >>>>>>>>>>> >>>>>>>>>>> - Carl >>>>>>>>>>> >>>>>>>>>>> On Wed, Aug 4, 2021 at 11:12 AM Ryan Murray <rym...@gmail.com> >>>>>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi all, >>>>>>>>>>>> >>>>>>>>>>>> Unfortunately I have to give -1 >>>>>>>>>>>> >>>>>>>>>>>> I had trouble w/ the keys: >>>>>>>>>>>> >>>>>>>>>>>> gpg: assuming signed data in 'apache-iceberg-0.12.0.tar.gz' >>>>>>>>>>>> gpg: Signature made Mon 02 Aug 2021 03:36:30 CEST >>>>>>>>>>>> gpg: using RSA key >>>>>>>>>>>> FAFEB6EAA60C95E2BB5E26F01FF0803CB78D539F >>>>>>>>>>>> gpg: Can't check signature: No public key >>>>>>>>>>>> >>>>>>>>>>>> And I have discovered a bug in NessieCatalog. It is unclear >>>>>>>>>>>> what is wrong but the NessieCatalog doesn't play nice w/ Spark3.1. >>>>>>>>>>>> I will >>>>>>>>>>>> raise a patch ASAP to fix it. Very sorry for the inconvenience. >>>>>>>>>>>> >>>>>>>>>>>> Best, >>>>>>>>>>>> Ryan >>>>>>>>>>>> >>>>>>>>>>>> On Wed, Aug 4, 2021 at 3:20 AM Carl Steinbach <c...@apache.org> >>>>>>>>>>>> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Hi everyone, >>>>>>>>>>>>> >>>>>>>>>>>>> I propose that we release RC2 as the official Apache Iceberg >>>>>>>>>>>>> 0.12.0 release. Please note that RC0 and RC1 were DOA. >>>>>>>>>>>>> >>>>>>>>>>>>> The commit id for RC2 is >>>>>>>>>>>>> 7c2fcfd893ab71bee41242b46e894e6187340070 >>>>>>>>>>>>> * This corresponds to the tag: apache-iceberg-0.12.0-rc2 >>>>>>>>>>>>> * >>>>>>>>>>>>> https://github.com/apache/iceberg/commits/apache-iceberg-0.12.0-rc2 >>>>>>>>>>>>> * >>>>>>>>>>>>> https://github.com/apache/iceberg/tree/7c2fcfd893ab71bee41242b46e894e6187340070 >>>>>>>>>>>>> >>>>>>>>>>>>> The release tarball, signature, and checksums are here: >>>>>>>>>>>>> * >>>>>>>>>>>>> https://dist.apache.org/repos/dist/dev/iceberg/apache-iceberg-0.12.0-rc2/ >>>>>>>>>>>>> >>>>>>>>>>>>> You can find the KEYS file here: >>>>>>>>>>>>> * https://dist.apache.org/repos/dist/dev/iceberg/KEYS >>>>>>>>>>>>> >>>>>>>>>>>>> Convenience binary artifacts are staged in Nexus. The Maven >>>>>>>>>>>>> repository URL is: >>>>>>>>>>>>> * >>>>>>>>>>>>> https://repository.apache.org/content/repositories/orgapacheiceberg-1017/ >>>>>>>>>>>>> >>>>>>>>>>>>> Please download, verify, and test. >>>>>>>>>>>>> >>>>>>>>>>>>> Please vote in the next 72 hours. >>>>>>>>>>>>> >>>>>>>>>>>>> [ ] +1 Release this as Apache Iceberg 0.12.0 >>>>>>>>>>>>> [ ] +0 >>>>>>>>>>>>> [ ] -1 Do not release this because... >>>>>>>>>>>>> >>>>>>>>>>>> >> >> -- >> Ryan Blue >> Tabular >> >