Hi, *Release of Spark:* 1.6.0, i downloaded it and made a built using 'sbt/sbt assembly'
*command for submitting your app: *bin/spark-submit --master spark://shams-machine:7077 --executor-cores 2 --class in.myapp.email.combiner.CombinerRealtime /opt/dev/workspace-luna/combiner_spark/target/combiner-0.0.1-SNAPSHOT.jar 2>&1 & *code snippet of your app: *i developed a lot chained transormations and connected with Kafka, MongoDB, Cassandra. But tested all of them using *local[2] *setting in *conf.setMaster *method. Everything is working there. *pastebin of log:* http://pastebin.com/0LjTWLfm Thanks Shams On Thu, Mar 10, 2016 at 8:11 PM, Ted Yu <yuzhih...@gmail.com> wrote: > Can you provide a bit more information ? > > Release of Spark > command for submitting your app > code snippet of your app > pastebin of log > > Thanks > > On Thu, Mar 10, 2016 at 6:32 AM, Shams ul Haque <sham...@cashcare.in> > wrote: > >> Hi, >> >> I have developed a spark realtime app and started spark-standalone on my >> laptop. But when i tried to submit that app in Spark it is always >> in WAITING state & Cores is always Zero. >> >> I have set: >> export SPARK_WORKER_CORES="2" >> export SPARK_EXECUTOR_CORES="1" >> >> in spark-env.sh, but still nothing happend. And same log entry in: >> *TaskSchedulerImpl:70 - Initial job has not accepted any resources* >> >> So, does i need a seperate machine for all this? >> >> Please help me to sort that out. >> >> Thanks >> Shams >> > >