Hi Dongjoon,

I think that priority should be given to user and dev Apache Spark communities 
and decision made based on what mostly benefits both communities. Said that I 
am OK with all 3 possible scenarios and will go with the community decision and 
Spark policies.

This is the list in the order of my preference (based on my thoughts on how 
community will benefit).

1. Upgrade Spark dependency on 3.5 branch on Apache Parquet to 1.15.2 and 
handle parquet dependency as any other dependency with CVE that EOL minor 
version used by the Spark. I assume (based on the existing PR) that such 
upgrade is safe to complete and there are no any known blocking issues (like 
conflicting dependencies, incompatible Java version, deadlock, functional 
regression, etc). Please correct me if I'm wrong.

2. Keep it as is. There may be a confusion on whether CVE-2025-46762 and 
CVE-2025-30065 affect Spark or not. So, it may be good to document that they 
are not (see 
https://github.com/apache/parquet-java/pull/3196#issuecomment-2823186647).

3. Upgrade Spark dependency on 3.5 branch on Apache Parquet to 1.13.2. It 
depends on the availability of 1.13.2 and it looks that Apache Parquet 
maintains only single minor release version with 1.13.x and 1.14.x being EOL (I 
put this option as the last one with the assumption that CVE-2025-46762 and 
CVE-2025-30065  do not impact Spark. If they do, it would be my second choice).


Thank you,

Vlad

On May 28, 2025, at 5:00 PM, Dongjoon Hyun <dongj...@apache.org> wrote:


From Vlad's claims, the following guess is incorrect because the link is an 
upgrade from Apache ORC 1.9.5 to 1.9.6 which is a maintenance version upgrade.

I guess that the similar argument applies to ORC upgrade
(https://github.com/apache/spark/pull/50813).
[SPARK-52025][BUILD][3.5] Upgrade ORC to 1.9.6

FYI, the Apache ORC community maintains all release branches for 3 years under 
the Semantic Versioning policy. Since 3-years is longer than Apache Spark's 
support period, we can say that  Apache ORC branch-1.9 has been maintained for 
Apache Spark branch-3.5. If there is a user request, Apache ORC community 
provides the maintenance release during that period.

To Vlad, this is not a single ASF project issue. I believe the best scenario 
for all ASF projects is that the Apache Parquet community releases 1.13.2 to 
provide the CVE and bug fixes properly and safely. Then, Apache Spark 3.5.x can 
upgrade to Apache Parquet 1.13.2 from 1.13.1.

WDYT, Vlad?

Dongjoon.

On 2025/05/28 04:09:39 Hyukjin Kwon wrote:
It's written in https://spark.apache.org/versioning-policy.html.
Spark follows semver. Improvements go to the feature version. Bug fixes or
critical stuff go to maintenance version.

On Wed, 28 May 2025 at 12:38, Jungtaek Lim <kabhwan.opensou...@gmail.com>
wrote:

+1 (non-binding) pending the discussion on CVEs (not the performance
improvement) in the current version of Apache Parquet.

On Tue, May 27, 2025 at 11:19 AM L. C. Hsieh <vii...@gmail.com> wrote:

+1

On Mon, May 26, 2025 at 6:51 PM Wenchen Fan <cloud0...@gmail.com> wrote:

+1. When this release is out, let's also update the release process
document to introduce the new way of making releases with GitHub Action
jobs.

On Tue, May 27, 2025 at 6:22 AM Dongjoon Hyun <dongj...@apache.org>
wrote:

+1 from my side.

Thank you, Hyukjin.

Dongjoon

On 2025/05/26 22:19:22 Hyukjin Kwon wrote:
Thanks guys. BTW for clarification, this is the preparation of more
frequent releases so we don't have to wait so long for each release.
Let's
prepare this first, and roll it faster

On Tue, 27 May 2025 at 01:52, Yang Jie <yangji...@apache.org> wrote:

+1

On 2025/05/26 01:10:23 Hyukjin Kwon wrote:
The key issue was fixed.

On Mon, 26 May 2025 at 10:05, Hyukjin Kwon <gurwls...@apache.org>
wrote:

Probably should avoid backporting it for improvements but If
there is a
CVE that directly affects Spark, let's upgrade.

On Mon, 26 May 2025 at 00:27, Rozov, Vlad
<vro...@amazon.com.invalid>
wrote:

Should parquet version be upgraded to 1.15.1 or 1.15.2? There
are 10
CVEs
in the current 1.13.1 and even though they may not impact
Spark there
are
other improvements (better performance) that will benefit
Spark users.

Thank you,

Vlad

On May 24, 2025, at 8:02 PM, Hyukjin Kwon <
gurwls...@apache.org>
wrote:

Oh let me check. Thanks for letting me know.

On Sun, May 25, 2025 at 12:00 PM Dongjoon Hyun <
dongj...@apache.org>
wrote:

I saw 38 commits to make this work. Thank you for driving
this,
Hyukjin.

BTW, your key seems to be new and is not in
https://dist.apache.org/repos/dist/dev/spark/KEYS yet.
Could you
double-check?

$ curl -LO https://dist.apache.org/repos/dist/dev/spark/KEYS
$ gpg --import KEYS
$ gpg --verify spark-3.5.6-bin-hadoop3.tgz.asc
gpg: assuming signed data in 'spark-3.5.6-bin-hadoop3.tgz'
gpg: Signature made Thu May 22 23:49:54 2025 PDT
gpg:                using RSA key
0FE4571297AB84440673665669600C8338F65970
gpg:                issuer "gurwls...@apache.org"
gpg: Can't check signature: No public key

Dongjoon.

On 2025/05/23 17:56:25 Allison Wang wrote:
+1

On Fri, May 23, 2025 at 10:15 AM Hyukjin Kwon <
gurwls...@apache.org>
wrote:

Oh it's actually a test and also to release. Let me know
if you
have
any
concern!

On Fri, May 23, 2025 at 11:25 PM Mridul Muralidharan <
mri...@gmail.com>
wrote:

Hi Hyukjin,

 This thread is to test the automated release, right ?
Not to actually release it ?

Regards,
Mridul

On Fri, May 23, 2025 at 8:26 AM Ruifeng Zheng <
ruife...@apache.org>
wrote:

+1

On Fri, May 23, 2025 at 5:27 PM Hyukjin Kwon <
gurwls...@apache.org

wrote:

Please vote on releasing the following candidate as
Apache
Spark
version 3.5.6.

The vote is open until May 27 (PST)  and passes if a
majority
+1
PMC
votes are cast, with
a minimum of 3 +1 votes.

[ ] +1 Release this package as Apache Spark 3.5.6
[ ] -1 Do not release this package because ...

To learn more about Apache Spark, please see
https://spark.apache.org/

The tag to be voted on is v3.5.6-rc5 (commit
303c18c74664f161b9b969ac343784c088b47593):




https://github.com/apache/spark/tree/303c18c74664f161b9b969ac343784c088b47593

The release files, including signatures, digests,
etc. can be
found at:

https://dist.apache.org/repos/dist/dev/spark/v3.5.6-rc1-bin/

Signatures used for Spark RCs can be found in this
file:
https://dist.apache.org/repos/dist/dev/spark/KEYS

The staging repository for this release can be found
at:



https://repository.apache.org/content/repositories/orgapachespark-1495/

The documentation corresponding to this release can
be found
at:

https://dist.apache.org/repos/dist/dev/spark/v3.5.6-rc1-docs/

The list of bug fixes going into 3.5.6 can be found
at the
following
URL:

https://issues.apache.org/jira/projects/SPARK/versions/12355703

FAQ

=========================
How can I help test this release?
=========================

If you are a Spark user, you can help us test this
release by
taking
an existing Spark workload and running on this release
candidate,
then
reporting any regressions.

If you're working in PySpark you can set up a virtual
env and
install
the current RC via "pip install



https://dist.apache.org/repos/dist/dev/spark/v3.5.6-rc1-bin/pyspark-3.5.6.tar.gz
"
and see if anything important breaks.
In the Java/Scala, you can add the staging repository
to your
projects
resolvers and test
with the RC (make sure to clean up the artifact cache
before/after so
you don't end up building with a out of date RC going
forward).





---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org






---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org




---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org


---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org




---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org


Reply via email to