Do we have everything we want for RC2 targeted to 3.3.0 for tracking? On Wed, May 11, 2022 at 6:44 AM Maxim Gekk <maxim.g...@databricks.com.invalid> wrote:
> Hi All, > > The vote has failed. I will create RC2 in a couple of days. > > Maxim Gekk > > Software Engineer > > Databricks, Inc. > > > On Wed, May 11, 2022 at 4:23 AM Hyukjin Kwon <gurwls...@gmail.com> wrote: > >> I expect to see RC2 too. I guess he just sticks to the standard, leaving >> the vote open till the end. >> It hasn't got enough +1s anyway :-). >> >> On Wed, 11 May 2022 at 10:17, Holden Karau <hol...@pigscanfly.ca> wrote: >> >>> Technically release don't follow vetos (see >>> https://www.apache.org/foundation/voting.html ) it's up to the RM if >>> they get the minimum number of binding +1s (although they are encouraged to >>> cancel the release if any serious issues are raised). >>> >>> That being said I'll add my -1 based on the issues reported in this >>> thread. >>> >>> On Tue, May 10, 2022 at 6:07 PM Sean Owen <sro...@gmail.com> wrote: >>> >>>> There's a -1 vote here, so I think this RC fails anyway. >>>> >>>> On Fri, May 6, 2022 at 10:30 AM Gengliang Wang <ltn...@gmail.com> >>>> wrote: >>>> >>>>> Hi Maxim, >>>>> >>>>> Thanks for the work! >>>>> There is a bug fix from Bruce merged on branch-3.3 right after the RC1 >>>>> is cut: >>>>> SPARK-39093: Dividing interval by integral can result in codegen >>>>> compilation error >>>>> <https://github.com/apache/spark/commit/fd998c8a6783c0c8aceed8dcde4017cd479e42c8> >>>>> >>>>> So -1 from me. We should have RC2 to include the fix. >>>>> >>>>> Thanks >>>>> Gengliang >>>>> >>>>> On Fri, May 6, 2022 at 6:15 PM Maxim Gekk >>>>> <maxim.g...@databricks.com.invalid> wrote: >>>>> >>>>>> Hi Dongjoon, >>>>>> >>>>>> > https://issues.apache.org/jira/projects/SPARK/versions/12350369 >>>>>> > Since RC1 is started, could you move them out from the 3.3.0 >>>>>> milestone? >>>>>> >>>>>> I have removed the 3.3.0 label from Fix version(s). Thank you, >>>>>> Dongjoon. >>>>>> >>>>>> Maxim Gekk >>>>>> >>>>>> Software Engineer >>>>>> >>>>>> Databricks, Inc. >>>>>> >>>>>> >>>>>> On Fri, May 6, 2022 at 11:06 AM Dongjoon Hyun < >>>>>> dongjoon.h...@gmail.com> wrote: >>>>>> >>>>>>> Hi, Sean. >>>>>>> It's interesting. I didn't see those failures from my side. >>>>>>> >>>>>>> Hi, Maxim. >>>>>>> In the following link, there are 17 in-progress and 6 to-do JIRA >>>>>>> issues which look irrelevant to this RC1 vote. >>>>>>> >>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369 >>>>>>> >>>>>>> Since RC1 is started, could you move them out from the 3.3.0 >>>>>>> milestone? >>>>>>> Otherwise, we cannot distinguish new real blocker issues from those >>>>>>> obsolete JIRA issues. >>>>>>> >>>>>>> Thanks, >>>>>>> Dongjoon. >>>>>>> >>>>>>> >>>>>>> On Thu, May 5, 2022 at 11:46 AM Adam Binford <adam...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> I looked back at the first one (SPARK-37618), it expects/assumes a >>>>>>>> 0022 umask to correctly test the behavior. I'm not sure how to get >>>>>>>> that to >>>>>>>> not fail or be ignored with a more open umask. >>>>>>>> >>>>>>>> On Thu, May 5, 2022 at 1:56 PM Sean Owen <sro...@gmail.com> wrote: >>>>>>>> >>>>>>>>> I'm seeing test failures; is anyone seeing ones like this? This is >>>>>>>>> Java 8 / Scala 2.12 / Ubuntu 22.04: >>>>>>>>> >>>>>>>>> - SPARK-37618: Sub dirs are group writable when removing from >>>>>>>>> shuffle service enabled *** FAILED *** >>>>>>>>> [OWNER_WRITE, GROUP_READ, GROUP_WRITE, GROUP_EXECUTE, >>>>>>>>> OTHERS_READ, OWNER_READ, OTHERS_EXECUTE, OWNER_EXECUTE] contained >>>>>>>>> GROUP_WRITE (DiskBlockManagerSuite.scala:155) >>>>>>>>> >>>>>>>>> - Check schemas for expression examples *** FAILED *** >>>>>>>>> 396 did not equal 398 Expected 396 blocks in result file but got >>>>>>>>> 398. Try regenerating the result files. >>>>>>>>> (ExpressionsSchemaSuite.scala:161) >>>>>>>>> >>>>>>>>> Function 'bloom_filter_agg', Expression class >>>>>>>>> 'org.apache.spark.sql.catalyst.expressions.aggregate.BloomFilterAggregate' >>>>>>>>> "" did not start with " >>>>>>>>> Examples: >>>>>>>>> " (ExpressionInfoSuite.scala:142) >>>>>>>>> >>>>>>>>> On Thu, May 5, 2022 at 6:01 AM Maxim Gekk >>>>>>>>> <maxim.g...@databricks.com.invalid> wrote: >>>>>>>>> >>>>>>>>>> Please vote on releasing the following candidate as Apache Spark >>>>>>>>>> version 3.3.0. >>>>>>>>>> >>>>>>>>>> The vote is open until 11:59pm Pacific time May 10th and passes >>>>>>>>>> if a majority +1 PMC votes are cast, with a minimum of 3 +1 votes >>>>>>>>>> . >>>>>>>>>> >>>>>>>>>> [ ] +1 Release this package as Apache Spark 3.3.0 >>>>>>>>>> [ ] -1 Do not release this package because ... >>>>>>>>>> >>>>>>>>>> To learn more about Apache Spark, please see http://spark >>>>>>>>>> .apache.org/ >>>>>>>>>> >>>>>>>>>> The tag to be voted on is v3.3.0-rc1 (commit >>>>>>>>>> 482b7d54b522c4d1e25f3e84eabbc78126f22a3d): >>>>>>>>>> https://github.com/apache/spark/tree/v3.3.0-rc1 >>>>>>>>>> >>>>>>>>>> The release files, including signatures, digests, etc. can be >>>>>>>>>> found at: >>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc1-bin/ >>>>>>>>>> >>>>>>>>>> Signatures used for Spark RCs can be found in this file: >>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS >>>>>>>>>> >>>>>>>>>> The staging repository for this release can be found at: >>>>>>>>>> >>>>>>>>>> https://repository.apache.org/content/repositories/orgapachespark-1402 >>>>>>>>>> >>>>>>>>>> The documentation corresponding to this release can be found at: >>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc1-docs/ >>>>>>>>>> >>>>>>>>>> The list of bug fixes going into 3.3.0 can be found at the >>>>>>>>>> following URL: >>>>>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369 >>>>>>>>>> >>>>>>>>>> This release is using the release script of the tag v3.3.0-rc1. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> FAQ >>>>>>>>>> >>>>>>>>>> ========================= >>>>>>>>>> How can I help test this release? >>>>>>>>>> ========================= >>>>>>>>>> If you are a Spark user, you can help us test this release by >>>>>>>>>> taking >>>>>>>>>> an existing Spark workload and running on this release candidate, >>>>>>>>>> then >>>>>>>>>> reporting any regressions. >>>>>>>>>> >>>>>>>>>> If you're working in PySpark you can set up a virtual env and >>>>>>>>>> install >>>>>>>>>> the current RC and see if anything important breaks, in the >>>>>>>>>> Java/Scala >>>>>>>>>> you can add the staging repository to your projects resolvers and >>>>>>>>>> test >>>>>>>>>> with the RC (make sure to clean up the artifact cache >>>>>>>>>> before/after so >>>>>>>>>> you don't end up building with a out of date RC going forward). >>>>>>>>>> >>>>>>>>>> =========================================== >>>>>>>>>> What should happen to JIRA tickets still targeting 3.3.0? >>>>>>>>>> =========================================== >>>>>>>>>> The current list of open tickets targeted at 3.3.0 can be found >>>>>>>>>> at: >>>>>>>>>> https://issues.apache.org/jira/projects/SPARK and search for >>>>>>>>>> "Target Version/s" = 3.3.0 >>>>>>>>>> >>>>>>>>>> Committers should look at those and triage. Extremely important >>>>>>>>>> bug >>>>>>>>>> fixes, documentation, and API tweaks that impact compatibility >>>>>>>>>> should >>>>>>>>>> be worked on immediately. Everything else please retarget to an >>>>>>>>>> appropriate release. >>>>>>>>>> >>>>>>>>>> ================== >>>>>>>>>> But my bug isn't fixed? >>>>>>>>>> ================== >>>>>>>>>> In order to make timely releases, we will typically not hold the >>>>>>>>>> release unless the bug in question is a regression from the >>>>>>>>>> previous >>>>>>>>>> release. That being said, if there is something which is a >>>>>>>>>> regression >>>>>>>>>> that has not been correctly targeted please ping me or a >>>>>>>>>> committer to >>>>>>>>>> help target the issue. >>>>>>>>>> >>>>>>>>>> Maxim Gekk >>>>>>>>>> >>>>>>>>>> Software Engineer >>>>>>>>>> >>>>>>>>>> Databricks, Inc. >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Adam Binford >>>>>>>> >>>>>>> >>> >>> -- >>> Twitter: https://twitter.com/holdenkarau >>> Books (Learning Spark, High Performance Spark, etc.): >>> https://amzn.to/2MaRAG9 <https://amzn.to/2MaRAG9> >>> YouTube Live Streams: https://www.youtube.com/user/holdenkarau >>> >> -- Twitter: https://twitter.com/holdenkarau Books (Learning Spark, High Performance Spark, etc.): https://amzn.to/2MaRAG9 <https://amzn.to/2MaRAG9> YouTube Live Streams: https://www.youtube.com/user/holdenkarau