Hi All, The vote has failed. I will create RC2 in a couple of days.
Maxim Gekk Software Engineer Databricks, Inc. On Wed, May 11, 2022 at 4:23 AM Hyukjin Kwon <gurwls...@gmail.com> wrote: > I expect to see RC2 too. I guess he just sticks to the standard, leaving > the vote open till the end. > It hasn't got enough +1s anyway :-). > > On Wed, 11 May 2022 at 10:17, Holden Karau <hol...@pigscanfly.ca> wrote: > >> Technically release don't follow vetos (see >> https://www.apache.org/foundation/voting.html ) it's up to the RM if >> they get the minimum number of binding +1s (although they are encouraged to >> cancel the release if any serious issues are raised). >> >> That being said I'll add my -1 based on the issues reported in this >> thread. >> >> On Tue, May 10, 2022 at 6:07 PM Sean Owen <sro...@gmail.com> wrote: >> >>> There's a -1 vote here, so I think this RC fails anyway. >>> >>> On Fri, May 6, 2022 at 10:30 AM Gengliang Wang <ltn...@gmail.com> wrote: >>> >>>> Hi Maxim, >>>> >>>> Thanks for the work! >>>> There is a bug fix from Bruce merged on branch-3.3 right after the RC1 >>>> is cut: >>>> SPARK-39093: Dividing interval by integral can result in codegen >>>> compilation error >>>> <https://github.com/apache/spark/commit/fd998c8a6783c0c8aceed8dcde4017cd479e42c8> >>>> >>>> So -1 from me. We should have RC2 to include the fix. >>>> >>>> Thanks >>>> Gengliang >>>> >>>> On Fri, May 6, 2022 at 6:15 PM Maxim Gekk >>>> <maxim.g...@databricks.com.invalid> wrote: >>>> >>>>> Hi Dongjoon, >>>>> >>>>> > https://issues.apache.org/jira/projects/SPARK/versions/12350369 >>>>> > Since RC1 is started, could you move them out from the 3.3.0 >>>>> milestone? >>>>> >>>>> I have removed the 3.3.0 label from Fix version(s). Thank you, >>>>> Dongjoon. >>>>> >>>>> Maxim Gekk >>>>> >>>>> Software Engineer >>>>> >>>>> Databricks, Inc. >>>>> >>>>> >>>>> On Fri, May 6, 2022 at 11:06 AM Dongjoon Hyun <dongjoon.h...@gmail.com> >>>>> wrote: >>>>> >>>>>> Hi, Sean. >>>>>> It's interesting. I didn't see those failures from my side. >>>>>> >>>>>> Hi, Maxim. >>>>>> In the following link, there are 17 in-progress and 6 to-do JIRA >>>>>> issues which look irrelevant to this RC1 vote. >>>>>> >>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369 >>>>>> >>>>>> Since RC1 is started, could you move them out from the 3.3.0 >>>>>> milestone? >>>>>> Otherwise, we cannot distinguish new real blocker issues from those >>>>>> obsolete JIRA issues. >>>>>> >>>>>> Thanks, >>>>>> Dongjoon. >>>>>> >>>>>> >>>>>> On Thu, May 5, 2022 at 11:46 AM Adam Binford <adam...@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> I looked back at the first one (SPARK-37618), it expects/assumes a >>>>>>> 0022 umask to correctly test the behavior. I'm not sure how to get that >>>>>>> to >>>>>>> not fail or be ignored with a more open umask. >>>>>>> >>>>>>> On Thu, May 5, 2022 at 1:56 PM Sean Owen <sro...@gmail.com> wrote: >>>>>>> >>>>>>>> I'm seeing test failures; is anyone seeing ones like this? This is >>>>>>>> Java 8 / Scala 2.12 / Ubuntu 22.04: >>>>>>>> >>>>>>>> - SPARK-37618: Sub dirs are group writable when removing from >>>>>>>> shuffle service enabled *** FAILED *** >>>>>>>> [OWNER_WRITE, GROUP_READ, GROUP_WRITE, GROUP_EXECUTE, >>>>>>>> OTHERS_READ, OWNER_READ, OTHERS_EXECUTE, OWNER_EXECUTE] contained >>>>>>>> GROUP_WRITE (DiskBlockManagerSuite.scala:155) >>>>>>>> >>>>>>>> - Check schemas for expression examples *** FAILED *** >>>>>>>> 396 did not equal 398 Expected 396 blocks in result file but got >>>>>>>> 398. Try regenerating the result files. >>>>>>>> (ExpressionsSchemaSuite.scala:161) >>>>>>>> >>>>>>>> Function 'bloom_filter_agg', Expression class >>>>>>>> 'org.apache.spark.sql.catalyst.expressions.aggregate.BloomFilterAggregate' >>>>>>>> "" did not start with " >>>>>>>> Examples: >>>>>>>> " (ExpressionInfoSuite.scala:142) >>>>>>>> >>>>>>>> On Thu, May 5, 2022 at 6:01 AM Maxim Gekk >>>>>>>> <maxim.g...@databricks.com.invalid> wrote: >>>>>>>> >>>>>>>>> Please vote on releasing the following candidate as Apache Spark >>>>>>>>> version 3.3.0. >>>>>>>>> >>>>>>>>> The vote is open until 11:59pm Pacific time May 10th and passes >>>>>>>>> if a majority +1 PMC votes are cast, with a minimum of 3 +1 votes. >>>>>>>>> >>>>>>>>> [ ] +1 Release this package as Apache Spark 3.3.0 >>>>>>>>> [ ] -1 Do not release this package because ... >>>>>>>>> >>>>>>>>> To learn more about Apache Spark, please see http://spark >>>>>>>>> .apache.org/ >>>>>>>>> >>>>>>>>> The tag to be voted on is v3.3.0-rc1 (commit >>>>>>>>> 482b7d54b522c4d1e25f3e84eabbc78126f22a3d): >>>>>>>>> https://github.com/apache/spark/tree/v3.3.0-rc1 >>>>>>>>> >>>>>>>>> The release files, including signatures, digests, etc. can be >>>>>>>>> found at: >>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc1-bin/ >>>>>>>>> >>>>>>>>> Signatures used for Spark RCs can be found in this file: >>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS >>>>>>>>> >>>>>>>>> The staging repository for this release can be found at: >>>>>>>>> >>>>>>>>> https://repository.apache.org/content/repositories/orgapachespark-1402 >>>>>>>>> >>>>>>>>> The documentation corresponding to this release can be found at: >>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc1-docs/ >>>>>>>>> >>>>>>>>> The list of bug fixes going into 3.3.0 can be found at the >>>>>>>>> following URL: >>>>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369 >>>>>>>>> >>>>>>>>> This release is using the release script of the tag v3.3.0-rc1. >>>>>>>>> >>>>>>>>> >>>>>>>>> FAQ >>>>>>>>> >>>>>>>>> ========================= >>>>>>>>> How can I help test this release? >>>>>>>>> ========================= >>>>>>>>> If you are a Spark user, you can help us test this release by >>>>>>>>> taking >>>>>>>>> an existing Spark workload and running on this release candidate, >>>>>>>>> then >>>>>>>>> reporting any regressions. >>>>>>>>> >>>>>>>>> If you're working in PySpark you can set up a virtual env and >>>>>>>>> install >>>>>>>>> the current RC and see if anything important breaks, in the >>>>>>>>> Java/Scala >>>>>>>>> you can add the staging repository to your projects resolvers and >>>>>>>>> test >>>>>>>>> with the RC (make sure to clean up the artifact cache before/after >>>>>>>>> so >>>>>>>>> you don't end up building with a out of date RC going forward). >>>>>>>>> >>>>>>>>> =========================================== >>>>>>>>> What should happen to JIRA tickets still targeting 3.3.0? >>>>>>>>> =========================================== >>>>>>>>> The current list of open tickets targeted at 3.3.0 can be found >>>>>>>>> at: >>>>>>>>> https://issues.apache.org/jira/projects/SPARK and search for >>>>>>>>> "Target Version/s" = 3.3.0 >>>>>>>>> >>>>>>>>> Committers should look at those and triage. Extremely important bug >>>>>>>>> fixes, documentation, and API tweaks that impact compatibility >>>>>>>>> should >>>>>>>>> be worked on immediately. Everything else please retarget to an >>>>>>>>> appropriate release. >>>>>>>>> >>>>>>>>> ================== >>>>>>>>> But my bug isn't fixed? >>>>>>>>> ================== >>>>>>>>> In order to make timely releases, we will typically not hold the >>>>>>>>> release unless the bug in question is a regression from the >>>>>>>>> previous >>>>>>>>> release. That being said, if there is something which is a >>>>>>>>> regression >>>>>>>>> that has not been correctly targeted please ping me or a committer >>>>>>>>> to >>>>>>>>> help target the issue. >>>>>>>>> >>>>>>>>> Maxim Gekk >>>>>>>>> >>>>>>>>> Software Engineer >>>>>>>>> >>>>>>>>> Databricks, Inc. >>>>>>>>> >>>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Adam Binford >>>>>>> >>>>>> >> >> -- >> Twitter: https://twitter.com/holdenkarau >> Books (Learning Spark, High Performance Spark, etc.): >> https://amzn.to/2MaRAG9 <https://amzn.to/2MaRAG9> >> YouTube Live Streams: https://www.youtube.com/user/holdenkarau >> >