Hi I am using the below command to run a spark job and I get an error like
"Container preempted by scheduler"

I am not sure if it's related to the wrong usage of Memory:

nohup ~/spark1.3/bin/spark-submit \ --num-executors 50 \ --master yarn \
--deploy-mode cluster \ --queue adhoc \ --driver-memory 18G \
--executor-memory 12G \ --class main.ru.<custom>.bigdata.externalchurn.Main
\ --conf "spark.task.maxFailures=100" \ --conf
"spark.yarn.max.executor.failures=10000" \ --conf "spark.executor.cores=1"
\ --conf "spark.akka.frameSize=50" \ --conf
"spark.storage.memoryFraction=0.5" \ --conf
"spark.driver.maxResultSize=10G" \
~/external-flow/externalChurn-1.0-SNAPSHOT-shaded.jar \
prepareTraining=true \ prepareTrainingMNP=true \ prepareMap=false \
bouldozerMode=true \ &> ~/external-flow/run.log & echo "STARTED" tail -f
~/external-flow/run.log

Thanks,

Reply via email to