Hmm, it should work with you run `PYSPARK_DRIVER_PYTHON="ipython" spark/bin/pyspark`
PYTHONSTARTUP is a PYTHON environment variable https://docs.python.org/2/using/cmdline.html#envvar-PYTHONSTARTUP On Sun, Jul 26, 2015 at 4:06 PM -0700, "Zerony Zhao" <bw.li...@gmail.com> wrote: Hello everyone, I have a newbie question. $SPARK_HOME/bin/pyspark will create SparkContext automatically. Welcome to ____ __ / __/__ ___ _____/ /__ _\ \/ _ \/ _ `/ __/ '_/ /__ / .__/\_,_/_/ /_/\_\ version 1.4.1 /_/ Using Python version 2.7.3 (default, Jun 22 2015 19:33:41) SparkContext available as sc, HiveContext available as sqlContext. But When using ipython as a driver, PYSPARK_DRIVER_PYTHON="ipython" spark/bin/pyspark , does not create SparkContext automatically. I have to execute execfile('spark_home/python/pyspark/shell.py') is it by design? I read the bash script bin/pyspark, I noticed the line: export PYTHONSTARTUP="$SPARK_HOME/python/pyspark/shell.py" But I searched the whole spark source code, the variable PYTHONSTARTUP is never used, I could not understand when PYTHONSTARTUP is executed. Thank you.