Hi All,

The vote has failed. I will create RC2 in a couple of days.

Maxim Gekk

Software Engineer

Databricks, Inc.


On Wed, May 11, 2022 at 4:23 AM Hyukjin Kwon <gurwls...@gmail.com> wrote:

> I expect to see RC2 too. I guess he just sticks to the standard, leaving
> the vote open till the end.
> It hasn't got enough +1s anyway :-).
>
> On Wed, 11 May 2022 at 10:17, Holden Karau <hol...@pigscanfly.ca> wrote:
>
>> Technically release don't follow vetos (see
>> https://www.apache.org/foundation/voting.html ) it's up to the RM if
>> they get the minimum number of binding +1s (although they are encouraged to
>> cancel the release if any serious issues are raised).
>>
>> That being said I'll add my -1 based on the issues reported in this
>> thread.
>>
>> On Tue, May 10, 2022 at 6:07 PM Sean Owen <sro...@gmail.com> wrote:
>>
>>> There's a -1 vote here, so I think this RC fails anyway.
>>>
>>> On Fri, May 6, 2022 at 10:30 AM Gengliang Wang <ltn...@gmail.com> wrote:
>>>
>>>> Hi Maxim,
>>>>
>>>> Thanks for the work!
>>>> There is a bug fix from Bruce merged on branch-3.3 right after the RC1
>>>> is cut:
>>>> SPARK-39093: Dividing interval by integral can result in codegen
>>>> compilation error
>>>> <https://github.com/apache/spark/commit/fd998c8a6783c0c8aceed8dcde4017cd479e42c8>
>>>>
>>>> So -1 from me. We should have RC2 to include the fix.
>>>>
>>>> Thanks
>>>> Gengliang
>>>>
>>>> On Fri, May 6, 2022 at 6:15 PM Maxim Gekk
>>>> <maxim.g...@databricks.com.invalid> wrote:
>>>>
>>>>> Hi Dongjoon,
>>>>>
>>>>>  > https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>>> > Since RC1 is started, could you move them out from the 3.3.0
>>>>> milestone?
>>>>>
>>>>> I have removed the 3.3.0 label from Fix version(s). Thank you,
>>>>> Dongjoon.
>>>>>
>>>>> Maxim Gekk
>>>>>
>>>>> Software Engineer
>>>>>
>>>>> Databricks, Inc.
>>>>>
>>>>>
>>>>> On Fri, May 6, 2022 at 11:06 AM Dongjoon Hyun <dongjoon.h...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hi, Sean.
>>>>>> It's interesting. I didn't see those failures from my side.
>>>>>>
>>>>>> Hi, Maxim.
>>>>>> In the following link, there are 17 in-progress and 6 to-do JIRA
>>>>>> issues which look irrelevant to this RC1 vote.
>>>>>>
>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>>>>
>>>>>> Since RC1 is started, could you move them out from the 3.3.0
>>>>>> milestone?
>>>>>> Otherwise, we cannot distinguish new real blocker issues from those
>>>>>> obsolete JIRA issues.
>>>>>>
>>>>>> Thanks,
>>>>>> Dongjoon.
>>>>>>
>>>>>>
>>>>>> On Thu, May 5, 2022 at 11:46 AM Adam Binford <adam...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> I looked back at the first one (SPARK-37618), it expects/assumes a
>>>>>>> 0022 umask to correctly test the behavior. I'm not sure how to get that 
>>>>>>> to
>>>>>>> not fail or be ignored with a more open umask.
>>>>>>>
>>>>>>> On Thu, May 5, 2022 at 1:56 PM Sean Owen <sro...@gmail.com> wrote:
>>>>>>>
>>>>>>>> I'm seeing test failures; is anyone seeing ones like this? This is
>>>>>>>> Java 8 / Scala 2.12 / Ubuntu 22.04:
>>>>>>>>
>>>>>>>> - SPARK-37618: Sub dirs are group writable when removing from
>>>>>>>> shuffle service enabled *** FAILED ***
>>>>>>>>   [OWNER_WRITE, GROUP_READ, GROUP_WRITE, GROUP_EXECUTE,
>>>>>>>> OTHERS_READ, OWNER_READ, OTHERS_EXECUTE, OWNER_EXECUTE] contained
>>>>>>>> GROUP_WRITE (DiskBlockManagerSuite.scala:155)
>>>>>>>>
>>>>>>>> - Check schemas for expression examples *** FAILED ***
>>>>>>>>   396 did not equal 398 Expected 396 blocks in result file but got
>>>>>>>> 398. Try regenerating the result files. 
>>>>>>>> (ExpressionsSchemaSuite.scala:161)
>>>>>>>>
>>>>>>>>  Function 'bloom_filter_agg', Expression class
>>>>>>>> 'org.apache.spark.sql.catalyst.expressions.aggregate.BloomFilterAggregate'
>>>>>>>> "" did not start with "
>>>>>>>>       Examples:
>>>>>>>>   " (ExpressionInfoSuite.scala:142)
>>>>>>>>
>>>>>>>> On Thu, May 5, 2022 at 6:01 AM Maxim Gekk
>>>>>>>> <maxim.g...@databricks.com.invalid> wrote:
>>>>>>>>
>>>>>>>>> Please vote on releasing the following candidate as Apache Spark
>>>>>>>>>  version 3.3.0.
>>>>>>>>>
>>>>>>>>> The vote is open until 11:59pm Pacific time May 10th and passes
>>>>>>>>> if a majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>>>>>>>>
>>>>>>>>> [ ] +1 Release this package as Apache Spark 3.3.0
>>>>>>>>> [ ] -1 Do not release this package because ...
>>>>>>>>>
>>>>>>>>> To learn more about Apache Spark, please see http://spark
>>>>>>>>> .apache.org/
>>>>>>>>>
>>>>>>>>> The tag to be voted on is v3.3.0-rc1 (commit
>>>>>>>>> 482b7d54b522c4d1e25f3e84eabbc78126f22a3d):
>>>>>>>>> https://github.com/apache/spark/tree/v3.3.0-rc1
>>>>>>>>>
>>>>>>>>> The release files, including signatures, digests, etc. can be
>>>>>>>>> found at:
>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc1-bin/
>>>>>>>>>
>>>>>>>>> Signatures used for Spark RCs can be found in this file:
>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>>>>>
>>>>>>>>> The staging repository for this release can be found at:
>>>>>>>>>
>>>>>>>>> https://repository.apache.org/content/repositories/orgapachespark-1402
>>>>>>>>>
>>>>>>>>> The documentation corresponding to this release can be found at:
>>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc1-docs/
>>>>>>>>>
>>>>>>>>> The list of bug fixes going into 3.3.0 can be found at the
>>>>>>>>> following URL:
>>>>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>>>>>>>
>>>>>>>>> This release is using the release script of the tag v3.3.0-rc1.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> FAQ
>>>>>>>>>
>>>>>>>>> =========================
>>>>>>>>> How can I help test this release?
>>>>>>>>> =========================
>>>>>>>>> If you are a Spark user, you can help us test this release by
>>>>>>>>> taking
>>>>>>>>> an existing Spark workload and running on this release candidate,
>>>>>>>>> then
>>>>>>>>> reporting any regressions.
>>>>>>>>>
>>>>>>>>> If you're working in PySpark you can set up a virtual env and
>>>>>>>>> install
>>>>>>>>> the current RC and see if anything important breaks, in the
>>>>>>>>> Java/Scala
>>>>>>>>> you can add the staging repository to your projects resolvers and
>>>>>>>>> test
>>>>>>>>> with the RC (make sure to clean up the artifact cache before/after
>>>>>>>>> so
>>>>>>>>> you don't end up building with a out of date RC going forward).
>>>>>>>>>
>>>>>>>>> ===========================================
>>>>>>>>> What should happen to JIRA tickets still targeting 3.3.0?
>>>>>>>>> ===========================================
>>>>>>>>> The current list of open tickets targeted at 3.3.0 can be found
>>>>>>>>> at:
>>>>>>>>> https://issues.apache.org/jira/projects/SPARK and search for
>>>>>>>>> "Target Version/s" = 3.3.0
>>>>>>>>>
>>>>>>>>> Committers should look at those and triage. Extremely important bug
>>>>>>>>> fixes, documentation, and API tweaks that impact compatibility
>>>>>>>>> should
>>>>>>>>> be worked on immediately. Everything else please retarget to an
>>>>>>>>> appropriate release.
>>>>>>>>>
>>>>>>>>> ==================
>>>>>>>>> But my bug isn't fixed?
>>>>>>>>> ==================
>>>>>>>>> In order to make timely releases, we will typically not hold the
>>>>>>>>> release unless the bug in question is a regression from the
>>>>>>>>> previous
>>>>>>>>> release. That being said, if there is something which is a
>>>>>>>>> regression
>>>>>>>>> that has not been correctly targeted please ping me or a committer
>>>>>>>>> to
>>>>>>>>> help target the issue.
>>>>>>>>>
>>>>>>>>> Maxim Gekk
>>>>>>>>>
>>>>>>>>> Software Engineer
>>>>>>>>>
>>>>>>>>> Databricks, Inc.
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Adam Binford
>>>>>>>
>>>>>>
>>
>> --
>> Twitter: https://twitter.com/holdenkarau
>> Books (Learning Spark, High Performance Spark, etc.):
>> https://amzn.to/2MaRAG9  <https://amzn.to/2MaRAG9>
>> YouTube Live Streams: https://www.youtube.com/user/holdenkarau
>>
>

Reply via email to