Re: overcommit: cpus / vcores

2018-10-16 Thread Khaled Zaouk
Hi Peter, I actually meant the spark configuration that you put in your spark-submit program (such as --conf spark.executor.instances= ..., --conf spark.executor.memory= ..., etc...). I advice you to check the number of partitions that you get in each stage of your workload the Spark GUI while th

Re: overcommit: cpus / vcores

2018-10-15 Thread Peter Liu
Hi Khaled, I have attached the spark streaming config below in (a). In case of the 100vcore run (see the initial email), I used 50 executors where each executor has 2 vcores and 3g memory. For 70 vcore case, 35 executors, for 80 vcore case, 40 executors. In the yarn config (yarn-site.xml, (b) bel

Re: overcommit: cpus / vcores

2018-10-15 Thread Khaled Zaouk
Hi Peter, What parameters are you putting in your spark streaming configuration? What are you putting as number of executor instances and how many cores per executor are you setting in your Spark job? Best, Khaled On Mon, Oct 15, 2018 at 9:18 PM Peter Liu wrote: > Hi there, > > I have a syste

re: overcommit: cpus / vcores

2018-10-15 Thread Peter Liu
Hi there, I have a system with 80 vcores and a relatively light spark streaming workload. Overcomming the vcore resource (i.e. > 80) in the config (see (a) below) seems to help to improve the average spark batch time (see (b) below). Is there any best practice guideline on resource overcommit wit