Re: spark-streaming "Could not compute split" exception

2014-09-10 Thread Tim Smith
I had a similar issue and many others - all were basically symptoms for yarn killing the container for high memory usage. Haven't gotten to root cause yet. On Tue, Sep 9, 2014 at 3:18 PM, Marcelo Vanzin wrote: > Your executor is exiting or crashing unexpectedly: > > On Tue, Sep 9, 2014 at 3:13 P

Re: spark-streaming "Could not compute split" exception

2014-09-09 Thread Marcelo Vanzin
Your executor is exiting or crashing unexpectedly: On Tue, Sep 9, 2014 at 3:13 PM, Penny Espinoza wrote: > org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: Exit > code from container container_1410224367331_0006_01_03 is : 1 > 2014-09-09 21:47:26,345 WARN > org.apache.hadoo

Re: spark-streaming "Could not compute split" exception

2014-09-09 Thread Penny Espinoza
The node manager log looks like this - not exactly sure what this means, but the container messages seem to indicate there is still plenty of memory. 2014-09-09 21:47:00,718 INFO org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl: Memory usage of ProcessTre

Re: spark-streaming "Could not compute split" exception

2014-09-09 Thread Marcelo Vanzin
This has all the symptoms of Yarn killing your executors due to them exceeding their memory limits. Could you check your RM/NM logs to see if that's the case? (The error was because of an executor at domU-12-31-39-0B-F1-D1.compute-1.internal, so you can check that NM's log file.) If that's the ca

spark-streaming "Could not compute split" exception

2014-09-09 Thread Penny Espinoza
Hey - I have a Spark 1.0.2 job (using spark-streaming-kafka) that runs successfully using master = local[4]. However, when I run it on a Hadoop 2.2 EMR cluster using master yarn-client, it fails after running for about 5 minutes. My main method does something like this: 1. gets streaming