Hello Sean & Akhil, I tried running the stop-all.sh script on my master and I got this message:
localhost: Permission denied (publickey,gssapi-keyex,gssapi-with-mic). chown: changing ownership of `/opt/cloudera/parcels/CDH-5.3.0-1.cdh5.3.0.p0.30/lib/spark/logs': Operation not permitted no org.apache.spark.deploy.master.Master to stop I am running Spark (Yarn) via Cloudera Manager. I tried stopping it from Cloudera Manager first, but it looked like it was only stopping the history server, so I started Spark again and tried ./stop-all.sh and got the above message. Also, what is the command for shutting down storage or can I simply stop hdfs in Cloudera Manager? Thank you for the help! On Sat, Jan 17, 2015 at 12:58 PM, Su She <suhsheka...@gmail.com> wrote: > Thanks Akhil and Sean for the responses. > > I will try shutting down spark, then storage and then the instances. > Initially, when hdfs was in safe mode, I waited for >1 hour and the problem > still persisted. I will try this new method. > > Thanks! > > > > On Sat, Jan 17, 2015 at 2:03 AM, Sean Owen <so...@cloudera.com> wrote: > >> You would not want to turn off storage underneath Spark. Shut down >> Spark first, then storage, then shut down the instances. Reverse the >> order when restarting. >> >> HDFS will be in safe mode for a short time after being started before >> it becomes writeable. I would first check that it's not just that. >> Otherwise, find out why the cluster went into safe mode from the logs, >> fix it, and then leave safe mode. >> >> On Sat, Jan 17, 2015 at 9:03 AM, Akhil Das <ak...@sigmoidanalytics.com> >> wrote: >> > Safest way would be to first shutdown HDFS and then shutdown Spark (call >> > stop-all.sh would do) and then shutdown the machines. >> > >> > You can execute the following command to disable safe mode: >> > >> >> hadoop fs -safemode leave >> > >> > >> > >> > Thanks >> > Best Regards >> > >> > On Sat, Jan 17, 2015 at 8:31 AM, Su She <suhsheka...@gmail.com> wrote: >> >> >> >> Hello Everyone, >> >> >> >> I am encountering trouble running Spark applications when I shut down >> my >> >> EC2 instances. Everything else seems to work except Spark. When I try >> >> running a simple Spark application, like sc.parallelize() I get the >> message >> >> that hdfs name node is in safemode. >> >> >> >> Has anyone else had this issue? Is there a proper protocol I should be >> >> following to turn off my spark nodes? >> >> >> >> Thank you! >> >> >> >> >> > >> > >