+1 on Xiangrui’s plan. On Thu, May 30, 2019 at 7:55 AM shane knapp <skn...@berkeley.edu> wrote:
> I don't have a good sense of the overhead of continuing to support >> Python 2; is it large enough to consider dropping it in Spark 3.0? >> >> from the build/test side, it will actually be pretty easy to continue > support for python2.7 for spark 2.x as the feature sets won't be expanding. > > that being said, i will be cracking a bottle of champagne when i can > delete all of the ansible and anaconda configs for python2.x. :) > > shane > -- > Shane Knapp > UC Berkeley EECS Research / RISELab Staff Technical Lead > https://rise.cs.berkeley.edu >