Hi,

How to set the number of executors and tasks in a Spark Streaming job in
Mesos? I have the following settings but my job still shows me 11 active
tasks and 11 executors. Any idea as to why this is happening
?

 sparkConf.set("spark.mesos.coarse", "true")
      sparkConf.set("spark.cores.max", "128")
      sparkConf.set("spark.default.parallelism", "100")
      //sparkConf.set("spark.locality.wait", "0")
      sparkConf.set("spark.executor.memory", "32g")
      sparkConf.set("spark.streaming.unpersist", "true")
      sparkConf.set("spark.shuffle.io.numConnectionsPerPeer", "1")
      sparkConf.set("spark.rdd.compress", "true")
      sparkConf.set("spark.shuffle.memoryFraction", ".6")
      sparkConf.set("spark.storage.memoryFraction", ".2")
      sparkConf.set("spark.shuffle.spill", "true")
      sparkConf.set("spark.shuffle.spill.compress", "true")
      sparkConf.set("spark.streaming.receiver.writeAheadLog.enable", "true")
      sparkConf.set("spark.streaming.blockInterval", "400")



--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/How-to-set-the-number-of-executors-and-tasks-in-a-Spark-Streaming-job-in-Mesos-tp24348.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to