Re: Num of executors and cores

2016-07-26 Thread Mail.com
Hi, In spark submit, I specify --master yarn-client. When I go to executors in UI I do see all the 12 different executors assigned. But for the stage when I drill down to Tasks I saw only 8 tasks with index 0-7. I ran again increasing the number of executors as 15 and I now see 12 tasks for the

Re: Num of executors and cores

2016-07-26 Thread Jacek Laskowski
Hi, Where's this yarn-client mode specified? When you said "However, when I run the job I see that the stage which reads the directory has only 8 tasks." -- how do you see 8 tasks for a stage? It appears you're in local[*] mode on a 8-core machine (like me) and that's why I'm asking such basic que

Re: Num of executors and cores

2016-07-26 Thread Mail.com
More of jars and files and app name. It runs on yarn-client mode. Thanks, Pradeep > On Jul 26, 2016, at 7:10 AM, Jacek Laskowski wrote: > > Hi, > > What's ""? What master URL do you use? > > Pozdrawiam, > Jacek Laskowski > > https://medium.com/@jaceklaskowski/ > Mastering Apache Spark ht

Re: Num of executors and cores

2016-07-26 Thread Jacek Laskowski
Hi, What's ""? What master URL do you use? Pozdrawiam, Jacek Laskowski https://medium.com/@jaceklaskowski/ Mastering Apache Spark http://bit.ly/mastering-apache-spark Follow me at https://twitter.com/jaceklaskowski On Tue, Jul 26, 2016 at 2:18 AM, Mail.com wrote: > Hi All, > > I have a di

Num of executors and cores

2016-07-25 Thread Mail.com
Hi All, I have a directory which has 12 files. I want to read the entire file so I am reading it as wholeTextFiles(dirpath, numPartitions). I run spark-submit as --num-executors 12 --executor-cores 1 and numPartitions 12. However, when I run the job I see that the stage which reads the direct