Makes sense Thank you Jeff !
On Sun, Dec 10, 2017 at 11:24 PM Jeff Zhang <zjf...@gmail.com> wrote: > > I am afraid currently there's no way to make ipython as default of > %pyspark, but you can use %ipyspark to use ipython without this warning > message. > > But making ipython as default is on my plan, for now I try to keep > backward compatibility as much as possible, so only use ipython when it is > available, otherwise still use the old python interpreter implementation. > I will change ipython as default and the original python implementation as > fallback when ipython interpreter become much more mature. > > > > > Ruslan Dautkhanov <dautkha...@gmail.com>于2017年12月11日周一 下午1:20写道: > >> Getting "IPython is available, use IPython for PySparkInterpreter" >> warning after starting pyspark interpreter. >> >> How do I default %pyspark to ipython? >> >> Tried to change to >> "class": "org.apache.zeppelin.spark.PySparkInterpreter", >> to >> "class": "org.apache.zeppelin.spark.IPySparkInterpreter", >> in interpreter.json but this gets overwritten back to PySparkInterpreter. >> >> Also tried to change to zeppelin.pyspark.python to ipython with no luck >> too. >> >> Is there is a documented way to default pyspark interpreter to ipython? >> Glanced over PR-2474 but can't quickly get what I am missing. >> >> >> Thanks. >> >> >>