So to clarify, only scala 2.12 is supported in Spark 3?
________________________________ From: Ryan Blue <rb...@netflix.com.invalid> Sent: Tuesday, November 6, 2018 1:24 PM To: d_t...@apple.com Cc: Sean Owen; Spark Dev List; cdelg...@apple.com Subject: Re: Make Scala 2.12 as default Scala version in Spark 3.0 +1 to Scala 2.12 as the default in Spark 3.0. On Tue, Nov 6, 2018 at 11:50 AM DB Tsai <d_t...@apple.com<mailto:d_t...@apple.com>> wrote: +1 on dropping Scala 2.11 in Spark 3.0 to simplify the build. As Scala 2.11 will not support Java 11 unless we make a significant investment, if we decide not to drop Scala 2.11 in Spark 3.0, what we can do is have only Scala 2.12 build support Java 11 while Scala 2.11 support Java 8. But I agree with Sean that this can make the decencies really complicated; hence I support to drop Scala 2.11 in Spark 3.0 directly. DB Tsai | Siri Open Source Technologies [not a contribution] | Apple, Inc On Nov 6, 2018, at 11:38 AM, Sean Owen <sro...@gmail.com<mailto:sro...@gmail.com>> wrote: I think we should make Scala 2.12 the default in Spark 3.0. I would also prefer to drop Scala 2.11 support in 3.0. In theory, not dropping 2.11 support it means we'd support Scala 2.11 for years, the lifetime of Spark 3.x. In practice, we could drop 2.11 support in a 3.1.0 or 3.2.0 release, kind of like what happened with 2.10 in 2.x. Java (9-)11 support also complicates this. I think getting it to work will need some significant dependency updates, and I worry not all will be available for 2.11 or will present some knotty problems. We'll find out soon if that forces the issue. Also note that Scala 2.13 is pretty close to release, and we'll want to support it soon after release, perhaps sooner than the long delay before 2.12 was supported (because it was hard!). It will probably be out well before Spark 3.0. Cross-compiling for 3 Scala versions sounds like too much. 3.0 could support 2.11 and 2.12, and 3.1 support 2.12 and 2.13, or something. But if 2.13 support is otherwise attainable at the release of Spark 3.0, I wonder if that too argues for dropping 2.11 support. Finally I'll say that Spark itself isn't dropping 2.11 support for a while, no matter what; it still exists in the 2.4.x branch of course. People who can't update off Scala 2.11 can stay on Spark 2.x, note. Sean On Tue, Nov 6, 2018 at 1:13 PM DB Tsai <d_t...@apple.com<mailto:d_t...@apple.com>> wrote: We made Scala 2.11 as default Scala version in Spark 2.0. Now, the next Spark version will be 3.0, so it's a great time to discuss should we make Scala 2.12 as default Scala version in Spark 3.0. Scala 2.11 is EOL, and it came out 4.5 ago; as a result, it's unlikely to support JDK 11 in Scala 2.11 unless we're willing to sponsor the needed work per discussion in Scala community, https://github.com/scala/scala-dev/issues/559#issuecomment-436160166 We have initial support of Scala 2.12 in Spark 2.4. If we decide to make Scala 2.12 as default for Spark 3.0 now, we will have ample time to work on bugs and issues that we may run into. What do you think? Thanks, DB Tsai | Siri Open Source Technologies [not a contribution] | Apple, Inc --------------------------------------------------------------------- To unsubscribe e-mail: dev-unsubscr...@spark.apache.org<mailto:dev-unsubscr...@spark.apache.org> -- Ryan Blue Software Engineer Netflix