+1

On Wed, Oct 26, 2016 at 11:26 AM, Reynold Xin <r...@databricks.com> wrote:

> We can do the following concrete proposal:
>
> 1. Plan to remove support for Java 7 / Scala 2.10 in Spark 2.2.0 (Mar/Apr
> 2017).
>
> 2. In Spark 2.1.0 release, aggressively and explicitly announce the
> deprecation of Java 7 / Scala 2.10 support.
>
> (a) It should appear in release notes, documentations that mention how to
> build Spark
>
> (b) and a warning should be shown every time SparkContext is started using
> Scala 2.10 or Java 7.
>
>
>
> On Wed, Oct 26, 2016 at 7:50 PM, Dongjoon Hyun <dongj...@apache.org>
> wrote:
>
>> Hi, Daniel.
>>
>> I guess that kind of works will start sufficiently in 2.1.0 after PMC's
>> annoucement/reminder on mailing list.
>>
>> Bests,
>> Dongjoon.
>>
>>
>> On Wednesday, October 26, 2016, Daniel Siegmann <
>> dsiegm...@securityscorecard.io> wrote:
>>
>>> Is the deprecation of JDK 7 and Scala 2.10 documented anywhere outside
>>> the release notes for Spark 2.0.0? I do not consider release notes to be
>>> sufficient public notice for deprecation of supported platforms - this
>>> should be noted in the documentation somewhere. Here are on the only
>>> mentions I could find:
>>>
>>> At http://spark.apache.org/downloads.html it says:
>>>
>>> "*Note: Starting version 2.0, Spark is built with Scala 2.11 by
>>> default. Scala 2.10 users should download the Spark source package and
>>> build with Scala 2.10 support
>>> <http://spark.apache.org/docs/latest/building-spark.html#building-for-scala-210>."*
>>>
>>> At http://spark.apache.org/docs/latest/#downloading it says:
>>>
>>> "Spark runs on Java 7+, Python 2.6+/3.4+ and R 3.1+. For the Scala API,
>>> Spark 2.0.1 uses Scala 2.11. You will need to use a compatible Scala
>>> version (2.11.x)."
>>>
>>> At http://spark.apache.org/docs/latest/programming-guide.html#l
>>> inking-with-spark it says:
>>>
>>>    - "Spark 2.0.1 is built and distributed to work with Scala 2.11 by
>>>    default. (Spark can be built to work with other versions of Scala, too.) 
>>> To
>>>    write applications in Scala, you will need to use a compatible Scala
>>>    version (e.g. 2.11.X)."
>>>    - "Spark 2.0.1 works with Java 7 and higher. If you are using Java
>>>    8, Spark supports lambda expressions
>>>    
>>> <http://docs.oracle.com/javase/tutorial/java/javaOO/lambdaexpressions.html>
>>>    for concisely writing functions, otherwise you can use the classes in the
>>>    org.apache.spark.api.java.function
>>>    
>>> <http://spark.apache.org/docs/latest/api/java/index.html?org/apache/spark/api/java/function/package-summary.html>
>>>    package."
>>>    - "Spark 2.0.1 works with Python 2.6+ or Python 3.4+. It can use the
>>>    standard CPython interpreter, so C libraries like NumPy can be used. It
>>>    also works with PyPy 2.3+."
>>>
>>>
>>>
>

Reply via email to