On 22-Dec-08, at 10:24 AM, Nathan Marz wrote:

I have been experiencing some unusual behavior from Hadoop recently. When trying to run a job, some of the tasks fail with:

java.io.IOException: Task process exit with nonzero status of 1.
        at org.apache.hadoop.mapred.TaskRunner.runChild(TaskRunner.java:462)
        at org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:403


Not all the tasks fail, but enough tasks fail such that the job fails. Unfortunately, there are no further logs for these tasks. Trying to retrieve the logs produces:

HTTP ERROR: 410

Failed to retrieve stdout log for task: attempt_200811101232_0218_m_000001_0

RequestURI=/tasklog


It seems like the tasktracker isn't able to even start the tasks on those machines. Has anyone seen anything like this before?

I see this on jobs that also get the "too many open files" task errors, or on subsequent jobs. I've always assumed that it's another manifestation of the same problem. Once I start getting these errors, I keep getting them until I shut down the cluster, although I don't always get enough to cause a job to fail. I haven't bothered restarting individual boxes or services.

I haven't been able to reproduce it consistently, but it seems to happen when I have many small input files; a job with one large input file broke after I split the input up. I'm using Streaming.

Karl Anderson
[email protected]
http://monkey.org/~kra



Reply via email to