Thank you for the replies. It makes sense for scala/java, but in python the JVM is launched when the spark context is initialised, so it should be able to set it, I assume.
On Wed, Oct 1, 2014 at 6:24 PM, Andrew Or <and...@databricks.com> wrote: > Hi Tamas, > > Yes, Marcelo is right. The reason why it doesn't make sense to set > "spark.driver.memory" in your SparkConf is because your application code, by > definition, is the driver. This means by the time you get to the code that > initializes your SparkConf, your driver JVM has already started with some > heap size, and you can't easily change the size of the JVM once it has > started. Note that this is true regardless of the deploy mode (client or > cluster). > > Alternatives to set this include the following: (1) You can set > "spark.driver.memory" in your `spark-defaults.conf` on the node that submits > the application, (2) You can use the --driver-memory command line option if > you are using Spark submit (bin/pyspark goes through this path, as you have > discovered on your own). > > Does that make sense? > > > 2014-10-01 10:17 GMT-07:00 Tamas Jambor <jambo...@gmail.com>: > >> when you say "respective backend code to launch it", I thought this is >> the way to do that. >> >> thanks, >> Tamas >> >> On Wed, Oct 1, 2014 at 6:13 PM, Marcelo Vanzin <van...@cloudera.com> >> wrote: >> > Because that's not how you launch apps in cluster mode; you have to do >> > it through the command line, or by calling directly the respective >> > backend code to launch it. >> > >> > (That being said, it would be nice to have a programmatic way of >> > launching apps that handled all this - this has been brought up in a >> > few different contexts, but I don't think there's an "official" >> > solution yet.) >> > >> > On Wed, Oct 1, 2014 at 9:59 AM, Tamas Jambor <jambo...@gmail.com> wrote: >> >> thanks Marcelo. >> >> >> >> What's the reason it is not possible in cluster mode, either? >> >> >> >> On Wed, Oct 1, 2014 at 5:42 PM, Marcelo Vanzin <van...@cloudera.com> >> >> wrote: >> >>> You can't set up the driver memory programatically in client mode. In >> >>> that mode, the same JVM is running the driver, so you can't modify >> >>> command line options anymore when initializing the SparkContext. >> >>> >> >>> (And you can't really start cluster mode apps that way, so the only >> >>> way to set this is through the command line / config files.) >> >>> >> >>> On Wed, Oct 1, 2014 at 9:26 AM, jamborta <jambo...@gmail.com> wrote: >> >>>> Hi all, >> >>>> >> >>>> I cannot figure out why this command is not setting the driver memory >> >>>> (it is >> >>>> setting the executor memory): >> >>>> >> >>>> conf = (SparkConf() >> >>>> .setMaster("yarn-client") >> >>>> .setAppName("test") >> >>>> .set("spark.driver.memory", "1G") >> >>>> .set("spark.executor.memory", "1G") >> >>>> .set("spark.executor.instances", 2) >> >>>> .set("spark.executor.cores", 4)) >> >>>> sc = SparkContext(conf=conf) >> >>>> >> >>>> whereas if I run the spark console: >> >>>> ./bin/pyspark --driver-memory 1G >> >>>> >> >>>> it sets it correctly. Seemingly they both generate the same commands >> >>>> in the >> >>>> logs. >> >>>> >> >>>> thanks a lot, >> >>>> >> >>>> >> >>>> >> >>>> >> >>>> >> >>>> -- >> >>>> View this message in context: >> >>>> http://apache-spark-user-list.1001560.n3.nabble.com/spark-driver-memory-is-not-set-pyspark-1-1-0-tp15498.html >> >>>> Sent from the Apache Spark User List mailing list archive at >> >>>> Nabble.com. >> >>>> >> >>>> --------------------------------------------------------------------- >> >>>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >> >>>> For additional commands, e-mail: user-h...@spark.apache.org >> >>>> >> >>> >> >>> >> >>> >> >>> -- >> >>> Marcelo >> > >> > >> > >> > -- >> > Marcelo >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >> For additional commands, e-mail: user-h...@spark.apache.org >> > --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org