[ https://issues.apache.org/jira/browse/SPARK-44157?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17930648#comment-17930648 ]
Sakthi commented on SPARK-44157: -------------------------------- This is mostly because probably Spark historically depended on the older version ({{{}protobuf-java-2.5.0{}}}), which was likely retained for compatibility reasons or due to Python-side dependencies. > Outdated JARs in PySpark package > -------------------------------- > > Key: SPARK-44157 > URL: https://issues.apache.org/jira/browse/SPARK-44157 > Project: Spark > Issue Type: Bug > Components: Build, PySpark > Affects Versions: 3.4.1 > Reporter: Adrian Gonzalez-Martin > Priority: Minor > Labels: pyspark > > The JARs which ship embedded within PySpark's package in PyPi don't seem > aligned with the deps specified in Spark's own `pom.xml`. > For example, in Spark's `pom.xml`, `protobuf-java` is set to `3.21.12`: > [https://github.com/apache/spark/blob/6b1ff22dde1ead51cbf370be6e48a802daae58b6/pom.xml#L127] > However, if we look at the JARs embedded within PySpark tarball, the version > of `protobuf-java` is `2.5.0` (i.e. > `..../site-packages/pyspark/jars/protobuf-java-2.5.0.jar`). Same seems to > apply to all other dependencies. > This introduces a set of CVEs which are fixed on upstream Spark, but are > still present in PySpark (e.g. `CVE-2022-3509`, `CVE-2021-22569`, ` > CVE-2015-5237` and a few others). As well as potentially introduce a source > of conflict whenever there's a breaking change on these deps. -- This message was sent by Atlassian Jira (v8.20.10#820010) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org