It's not making 2.12 the default, but not dropping 2.11. Supporting 2.13 could mean supporting 3 Scala versions at once, which I claim is just too much. I think the options are likely:
- Support 2.11, 2.12 in Spark 3.0. Deprecate 2.11 and make 2.12 the default. Add 2.13 support in 3.x and drop 2.11 in the same release - Deprecate 2.11 right now via announcement and/or Spark 2.4.1 soon. Drop 2.11 support in Spark 3.0, and support only 2.12. - (same as above, but add Spark 2.13 support if possible for Spark 3.0) On Wed, Nov 7, 2018 at 12:32 PM Mark Hamstra <m...@clearstorydata.com> wrote: > > I'm not following "exclude Scala 2.13". Is there something inherent in making > 2.12 the default Scala version in Spark 3.0 that would prevent us from > supporting the option of building with 2.13? > > On Tue, Nov 6, 2018 at 5:48 PM Sean Owen <sro...@gmail.com> wrote: >> >> That's possible here, sure. The issue is: would you exclude Scala 2.13 >> support in 3.0 for this, if it were otherwise ready to go? >> I think it's not a hard rule that something has to be deprecated >> previously to be removed in a major release. The notice is helpful, >> sure, but there are lots of ways to provide that notice to end users. >> Lots of things are breaking changes in a major release. Or: deprecate >> in Spark 2.4.1, if desired? >> >> On Tue, Nov 6, 2018 at 7:36 PM Wenchen Fan <cloud0...@gmail.com> wrote: >> > >> > We make Scala 2.11 the default one in Spark 2.0, then drop Scala 2.10 in >> > Spark 2.3. Shall we follow it and drop Scala 2.11 at some point of Spark >> > 3.x? >> > >> > On Wed, Nov 7, 2018 at 8:55 AM Reynold Xin <r...@databricks.com> wrote: >> >> >> >> Have we deprecated Scala 2.11 already in an existing release? >> >> --------------------------------------------------------------------- >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >> --------------------------------------------------------------------- To unsubscribe e-mail: dev-unsubscr...@spark.apache.org