Then in theory every user can fire multiple spark-submit jobs. do you cap it with settings in $SPARK_HOME/conf/spark-defaults.conf , but I guess in reality every user submits one job only.
This is an interesting model for two reasons: - It uses parallel processing across all the nodes or most of the nodes to minimise the processing time - it requires less intervention Dr Mich Talebzadeh LinkedIn * https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* http://talebzadehmich.wordpress.com On 19 May 2016 at 21:33, Mathieu Longtin <math...@closetwork.org> wrote: > Driver memory is default. Executor memory depends on job, the caller > decides how much memory to use. We don't specify --num-executors as we want > all cores assigned to the local master, since they were started by the > current user. No local executor. --master=spark://localhost:someport. 1 > core per executor. > > On Thu, May 19, 2016 at 4:12 PM Mich Talebzadeh <mich.talebza...@gmail.com> > wrote: > >> Thanks Mathieu >> >> So it would be interesting to see what resources allocated in your case, >> especially the num-executors and executor-cores. I gather every node has >> enough memory and cores. >> >> >> >> ${SPARK_HOME}/bin/spark-submit \ >> >> --master local[2] \ >> >> --driver-memory 4g \ >> >> --num-executors=1 \ >> >> --executor-memory=4G \ >> >> --executor-cores=2 \ >> >> Dr Mich Talebzadeh >> >> >> >> LinkedIn * >> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* >> >> >> >> http://talebzadehmich.wordpress.com >> >> >> >> On 19 May 2016 at 21:02, Mathieu Longtin <math...@closetwork.org> wrote: >> >>> The driver (the process started by spark-submit) runs locally. The >>> executors run on any of thousands of servers. So far, I haven't tried more >>> than 500 executors. >>> >>> Right now, I run a master on the same server as the driver. >>> >>> On Thu, May 19, 2016 at 3:49 PM Mich Talebzadeh < >>> mich.talebza...@gmail.com> wrote: >>> >>>> ok so you are using some form of NFS mounted file system shared among >>>> the nodes and basically you start the processes through spark-submit. >>>> >>>> In Stand-alone mode, a simple cluster manager included with Spark. It >>>> does the management of resources so it is not clear to me what you are >>>> referring as worker manager here? >>>> >>>> This is my take from your model. >>>> The application will go and grab all the cores in the cluster. >>>> You only have one worker that lives within the driver JVM process. >>>> The Driver node runs on the same host that the cluster manager is >>>> running. The Driver requests the Cluster Manager for resources to run >>>> tasks. In this case there is only one executor for the Driver? The Executor >>>> runs tasks for the Driver. >>>> >>>> >>>> HTH >>>> >>>> Dr Mich Talebzadeh >>>> >>>> >>>> >>>> LinkedIn * >>>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >>>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* >>>> >>>> >>>> >>>> http://talebzadehmich.wordpress.com >>>> >>>> >>>> >>>> On 19 May 2016 at 20:37, Mathieu Longtin <math...@closetwork.org> >>>> wrote: >>>> >>>>> No master and no node manager, just the processes that do actual work. >>>>> >>>>> We use the "stand alone" version because we have a shared file system >>>>> and a way of allocating computing resources already (Univa Grid Engine). >>>>> If >>>>> an executor were to die, we have other ways of restarting it, we don't >>>>> need >>>>> the worker manager to deal with it. >>>>> >>>>> On Thu, May 19, 2016 at 3:16 PM Mich Talebzadeh < >>>>> mich.talebza...@gmail.com> wrote: >>>>> >>>>>> Hi Mathieu >>>>>> >>>>>> What does this approach provide that the norm lacks? >>>>>> >>>>>> So basically each node has its master in this model. >>>>>> >>>>>> Are these supposed to be individual stand alone servers? >>>>>> >>>>>> >>>>>> Thanks >>>>>> >>>>>> >>>>>> Dr Mich Talebzadeh >>>>>> >>>>>> >>>>>> >>>>>> LinkedIn * >>>>>> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw >>>>>> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* >>>>>> >>>>>> >>>>>> >>>>>> http://talebzadehmich.wordpress.com >>>>>> >>>>>> >>>>>> >>>>>> On 19 May 2016 at 18:45, Mathieu Longtin <math...@closetwork.org> >>>>>> wrote: >>>>>> >>>>>>> First a bit of context: >>>>>>> We use Spark on a platform where each user start workers as needed. >>>>>>> This has the advantage that all permission management is handled by the >>>>>>> OS, >>>>>>> so the users can only read files they have permission to. >>>>>>> >>>>>>> To do this, we have some utility that does the following: >>>>>>> - start a master >>>>>>> - start worker managers on a number of servers >>>>>>> - "submit" the Spark driver program >>>>>>> - the driver then talks to the master, tell it how many executors it >>>>>>> needs >>>>>>> - the master tell the worker nodes to start executors and talk to >>>>>>> the driver >>>>>>> - the executors are started >>>>>>> >>>>>>> From here on, the master doesn't do much, neither do the process >>>>>>> manager on the worker nodes. >>>>>>> >>>>>>> What I would like to do is simplify this to: >>>>>>> - Start the driver program >>>>>>> - Start executors on a number of servers, telling them where to find >>>>>>> the driver >>>>>>> - The executors connect directly to the driver >>>>>>> >>>>>>> Is there a way I could do this without the master and worker >>>>>>> managers? >>>>>>> >>>>>>> Thanks! >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Mathieu Longtin >>>>>>> 1-514-803-8977 >>>>>>> >>>>>> >>>>>> -- >>>>> Mathieu Longtin >>>>> 1-514-803-8977 >>>>> >>>> >>>> -- >>> Mathieu Longtin >>> 1-514-803-8977 >>> >> >> -- > Mathieu Longtin > 1-514-803-8977 >