I am running spark1.3 on Cloudera hadoop 5.4
--
Jan Sterba
https://twitter.com/honzasterba | http://flickr.com/honzasterba |
http://500px.com/honzasterba


On Thu, Feb 25, 2016 at 4:22 PM, Yin Yang <yy201...@gmail.com> wrote:
> Which Spark / hadoop release are you running ?
>
> Thanks
>
> On Thu, Feb 25, 2016 at 4:28 AM, Jan Štěrba <i...@jansterba.com> wrote:
>>
>> Hello,
>>
>> I have quite a weird behaviour that I can't quite wrap my head around.
>> I am running Spark on a Hadoop YARN cluster. I have Spark configured
>> in such a way that it utilizes all free vcores in the cluster (setting
>> max vcores per executor and number of executors to use all vcores in
>> cluster).
>>
>> Once oozie launcher job and spark AM claim their job, there should be
>> free resources for 8 spark executor nodes but in spark UI I only see 7
>> active nodes (there should be two spark nodes per one hadoop host). I
>> have checked what containers are running on each hadoop nodes and
>> discovered that one node is in deed running more spark containers than
>> is reported in spark UI.
>>
>> This behaviour is very strange to me and I have no idea what to make
>> of it or how to debug it.
>>
>> Any thoughts?
>>
>> Thanks.
>>
>> --
>> Jan Sterba
>> https://twitter.com/honzasterba | http://flickr.com/honzasterba |
>> http://500px.com/honzasterba
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
>>
>

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to