Spark master keeps running out of RAM

2016-03-31 Thread Dillian Murphey
Why would the spark master run out of RAM if I have too many slaves? Is this a flaw in the coding? I'm just a user of spark. The developer that set this up left the company, so I'm starting from the top here. So I noticed if I spawn lots of jobs, my spark master ends up crashing due to low memo

Someone explain the /mnt and /mnt2 folders on spark-ec2 slaves

2016-03-25 Thread Dillian Murphey
I have a 40GB ephemeral disk on /mnt and another one on /mnt2 The person that set this up has left. I'm aware of having maybe 1 ebs disk, but I guess this was launched with 2 ebs volumes using the --ebsxyz? Or are those two instance storages part of the AMI? tnx

Testing spark with AWS spot instances

2016-03-24 Thread Dillian Murphey
I'm very new to apache spark. I'm just a user not a developer. I'm running a cluster with many spot instances. Am I correct in understanding that spark can handle an unlimited number of spot instance failures and restarts? Sometimes all the spot instances will dissapear without warning, and then

Re: Slaves died, but jobs not picked up by standby slaves

2016-03-24 Thread Dillian Murphey
Never mind. What I was missing was waiting long enough :-O. Sry bout that. On Thu, Mar 24, 2016 at 11:20 AM, Dillian Murphey wrote: > Had 15 slaves. > > Added 10 more. > > Shut down some slaves in the first bunch of 15. > > The 10 slaves I added are sitting there idle.

Slaves died, but jobs not picked up by standby slaves

2016-03-24 Thread Dillian Murphey
Had 15 slaves. Added 10 more. Shut down some slaves in the first bunch of 15. The 10 slaves I added are sitting there idle. Spark did not assign idle cores to pick up the slack. What am I missing? Thanks for any help. Confused. :-i

Master is listing DEAD slaves, can they be cleaned up?

2015-12-01 Thread Dillian Murphey
On the status page on port 8080 I see the slaves that have been stopped. They are marked as DEAD. If I implement an autoscaling system on aws, which is the plan, then I can see there be hundreds of slaves going up and down. Is there a way to force a cleanup of dead slaves so they are not piling u

Re: Adding more slaves to a running cluster

2015-11-25 Thread Dillian Murphey
everything I've found so far seems to indicate it isn't supported yet. But yet here I am with 1.5 and it at least appears to be working. Am I missing something? On Tue, Nov 24, 2015 at 4:40 PM, Dillian Murphey wrote: > What's the current status on adding slaves to a running clu

Adding more slaves to a running cluster

2015-11-24 Thread Dillian Murphey
What's the current status on adding slaves to a running cluster? I want to leverage spark-ec2 and autoscaling groups. I want to launch slaves as spot instances when I need to do some heavy lifting, but I don't want to bring down my cluster in order to add nodes. Can this be done by just running