Re: configuring .sparkStaging with group rwx

2021-02-25 Thread Yuri Oleynikov (‫יורי אולייניקוב‬‎)
owning the cluster jobs run fine... however, when > trying to run pyspark with a different user it fails because the > .sparkStaging/application_* is written with 700 so the user cannot write to > that directory > > how to configure spark/yarn cluster so .sparkStaging is wr

configuring .sparkStaging with group rwx

2021-02-25 Thread Bulldog20630405
we have a spark cluster running on with multiple users... when running with the user owning the cluster jobs run fine... however, when trying to run pyspark with a different user it fails because the .sparkStaging/application_* is written with 700 so the user cannot write to that directory

Re: Environment variable for deleting .sparkStaging

2020-02-13 Thread mailfordebu
Any feedback please? Thanks, Debu Sent from my iPhone > On 13-Feb-2020, at 6:36 PM, Debabrata Ghosh wrote: > >  > Greetings All ! > > I have got plenty of application directories lying around sparkStaging , such > as .sparkStaging/application_1580703507814_0074 > &g

Environment variable for deleting .sparkStaging

2020-02-13 Thread Debabrata Ghosh
Greetings All ! I have got plenty of application directories lying around sparkStaging , such as .sparkStaging/application_1580703507814_0074 Would you please be able to help advise me which variable I need to set in spark-env.sh so that the sparkStaging applications aren't preserved afte

sparkStaging

2014-07-10 Thread Koert Kuipers
in spark 1.0.0 using yarn-client mode i am seeing that the sparkStaging directories do not get cleaned up. for example i run: $ spark-submit --class org.apache.spark.examples.SparkPi spark-examples-1.0.0-hadoop2.3.0-cdh5.0.2.jar 10 after which i have this directory left behind with one file in

RE: HDFS folder .sparkStaging not deleted and filled up HDFS in yarn mode

2014-06-23 Thread Andrew Lee
t: 5f48721, github.com/apache/spark/pull/586 From: alee...@hotmail.com To: user@spark.apache.org Subject: RE: HDFS folder .sparkStaging not deleted and filled up HDFS in yarn mode Date: Wed, 18 Jun 2014 11:24:36 -0700 Forgot to mention that I am using spark-submit to submit jobs, and a verbose

RE: HDFS folder .sparkStaging not deleted and filled up HDFS in yarn mode

2014-06-18 Thread Andrew Lee
Forgot to mention that I am using spark-submit to submit jobs, and a verbose mode print out looks like this with the SparkPi examples.The .sparkStaging won't be deleted. My thoughts is that this should be part of the staging and should be cleaned up as well when sc gets termi

HDFS folder .sparkStaging not deleted and filled up HDFS in yarn mode

2014-06-18 Thread Andrew Lee
Hi All, Have anyone ran into the same problem? By looking at the source code in official release (rc11),this property settings is set to false by default, however, I'm seeing the .sparkStaging folder remains on the HDFS and causing it to fill up the disk pretty fast since SparkContext de