+1 On Wed, Oct 26, 2016 at 11:26 AM, Reynold Xin <r...@databricks.com> wrote:
> We can do the following concrete proposal: > > 1. Plan to remove support for Java 7 / Scala 2.10 in Spark 2.2.0 (Mar/Apr > 2017). > > 2. In Spark 2.1.0 release, aggressively and explicitly announce the > deprecation of Java 7 / Scala 2.10 support. > > (a) It should appear in release notes, documentations that mention how to > build Spark > > (b) and a warning should be shown every time SparkContext is started using > Scala 2.10 or Java 7. > > > > On Wed, Oct 26, 2016 at 7:50 PM, Dongjoon Hyun <dongj...@apache.org> > wrote: > >> Hi, Daniel. >> >> I guess that kind of works will start sufficiently in 2.1.0 after PMC's >> annoucement/reminder on mailing list. >> >> Bests, >> Dongjoon. >> >> >> On Wednesday, October 26, 2016, Daniel Siegmann < >> dsiegm...@securityscorecard.io> wrote: >> >>> Is the deprecation of JDK 7 and Scala 2.10 documented anywhere outside >>> the release notes for Spark 2.0.0? I do not consider release notes to be >>> sufficient public notice for deprecation of supported platforms - this >>> should be noted in the documentation somewhere. Here are on the only >>> mentions I could find: >>> >>> At http://spark.apache.org/downloads.html it says: >>> >>> "*Note: Starting version 2.0, Spark is built with Scala 2.11 by >>> default. Scala 2.10 users should download the Spark source package and >>> build with Scala 2.10 support >>> <http://spark.apache.org/docs/latest/building-spark.html#building-for-scala-210>."* >>> >>> At http://spark.apache.org/docs/latest/#downloading it says: >>> >>> "Spark runs on Java 7+, Python 2.6+/3.4+ and R 3.1+. For the Scala API, >>> Spark 2.0.1 uses Scala 2.11. You will need to use a compatible Scala >>> version (2.11.x)." >>> >>> At http://spark.apache.org/docs/latest/programming-guide.html#l >>> inking-with-spark it says: >>> >>> - "Spark 2.0.1 is built and distributed to work with Scala 2.11 by >>> default. (Spark can be built to work with other versions of Scala, too.) >>> To >>> write applications in Scala, you will need to use a compatible Scala >>> version (e.g. 2.11.X)." >>> - "Spark 2.0.1 works with Java 7 and higher. If you are using Java >>> 8, Spark supports lambda expressions >>> >>> <http://docs.oracle.com/javase/tutorial/java/javaOO/lambdaexpressions.html> >>> for concisely writing functions, otherwise you can use the classes in the >>> org.apache.spark.api.java.function >>> >>> <http://spark.apache.org/docs/latest/api/java/index.html?org/apache/spark/api/java/function/package-summary.html> >>> package." >>> - "Spark 2.0.1 works with Python 2.6+ or Python 3.4+. It can use the >>> standard CPython interpreter, so C libraries like NumPy can be used. It >>> also works with PyPy 2.3+." >>> >>> >>> >