Hi,
The actual useful part of the error is:
Execution Error, return code 2 from
org.apache.hadoop.hive.ql.exec.mr.MapRedTask

If you do a search for this plus "EC2" in Google you will find a couple of
results that point to memory exhaustion issues. You should try increasing
the configurated memory size.

Since you are using a t2.micro you should really try using a bigger Amazon
instance size. This might probably be a lot more useful than trying
different configurations.
 On Jul 24, 2014 7:08 AM, "Sarfraz Ramay" <sarfraz.ra...@gmail.com> wrote:

> Can anyone please help with this ?
>
> [image: Inline image 1]
>
>
> i followed the advice here
> http://stackoverflow.com/questions/20390217/mapreduce-job-in-headless-environment-fails-n-times-due-to-am-container-exceptio
>
> and added to mapred-site.xml following properties but still getting the
> same error.
>
> <property>
>     <name>mapred.child.java.opts</name>
>     <value>-Djava.awt.headless=true</value></property><!-- add headless to 
> default -Xmx1024m --><property>
>     <name>yarn.app.mapreduce.am.command-opts</name>
>     <value>-Djava.awt.headless=true -Xmx1024m</value></property><property>
>     <name>yarn.app.mapreduce.am.admin-command-opts</name>
>     <value>-Djava.awt.headless=true</value></property
>
>
>
> Regards,
> Sarfraz Rasheed Ramay (DIT)
> Dublin, Ireland.
>
>
> On Tue, Jul 22, 2014 at 8:19 AM, Sarfraz Ramay <sarfraz.ra...@gmail.com>
> wrote:
>
>> Hi,
>>
>> I am using Hive 0.13.1 and Hadoop 2-2.0 on amazon EC2 t2.micro instances.
>> I have 4 instances, master has the namenode and yarn, secondarynode is a
>> separate instance and two slaves are on separate instances each.
>>
>> It was working fine till now but it started to break when i tried to run
>> the following query on tpch generated 3GB data. same worked ok on 1GB
>>
>> SELECT
>>   l_orderkey
>>   , sum(l_extendedprice*(1-l_discount)) as revenue
>>   , o_orderdate
>>   , o_shippriority
>> FROM
>> customer c JOIN orders o
>>     ON (c.c_custkey = o.o_custkey)
>> JOIN lineitem l
>>     on (l.l_orderkey = o.o_orderkey)
>> WHERE
>>  o_orderdate < '1995-03-15' and l_shipdate > '1995-03-15'
>> AND c.c_mktsegment = 'AUTOMOBILE'
>> GROUP BY
>> l_orderkey, o_orderdate, o_shippriority
>> HAVING
>> sum(l_extendedprice*(1-l_discount)) > 38500 --average revenue
>> --LIMIT 10;
>>
>> i have tried many things but nothing seems to work. I am also attaching
>> my mapred-site.xml and yarn-site.xml files for reference plus the error
>> log. I have also tried to limit the memory settings in mapred-site.xml and
>> yarn-site but nothing seems to be working. For full log details please find
>> attached hive.log file. Please help!
>>
>> Hadoop job information for Stage-7: number of mappers: 9; number of
>> reducers: 0
>> 2014-07-22 06:39:31,643 Stage-7 map = 0%,  reduce = 0%
>> 2014-07-22 06:39:43,940 Stage-7 map = 6%,  reduce = 0%, Cumulative CPU
>> 5.34 sec
>> 2014-07-22 06:39:45,002 Stage-7 map = 11%,  reduce = 0%, Cumulative CPU
>> 6.94 sec
>> 2014-07-22 06:40:08,373 Stage-7 map = 17%,  reduce = 0%, Cumulative CPU
>> 12.6 sec
>> 2014-07-22 06:40:10,417 Stage-7 map = 22%,  reduce = 0%, Cumulative CPU
>> 14.06 sec
>> 2014-07-22 06:40:22,732 Stage-7 map = 28%,  reduce = 0%, Cumulative CPU
>> 24.46 sec
>> 2014-07-22 06:40:25,843 Stage-7 map = 33%,  reduce = 0%, Cumulative CPU
>> 25.74 sec
>> 2014-07-22 06:40:33,039 Stage-7 map = 44%,  reduce = 0%, Cumulative CPU
>> 33.32 sec
>> 2014-07-22 06:40:38,709 Stage-7 map = 56%,  reduce = 0%, Cumulative CPU
>> 37.19 sec
>> 2014-07-22 06:41:07,648 Stage-7 map = 61%,  reduce = 0%, Cumulative CPU
>> 42.83 sec
>> 2014-07-22 06:41:15,900 Stage-7 map = 56%,  reduce = 0%, Cumulative CPU
>> 39.49 sec
>> 2014-07-22 06:41:27,299 Stage-7 map = 67%,  reduce = 0%, Cumulative CPU
>> 46.07 sec
>> 2014-07-22 06:41:28,342 Stage-7 map = 56%,  reduce = 0%, Cumulative CPU
>> 40.9 sec
>> 2014-07-22 06:41:43,753 Stage-7 map = 61%,  reduce = 0%, Cumulative CPU
>> 42.84 sec
>> 2014-07-22 06:41:45,801 Stage-7 map = 100%,  reduce = 0%, Cumulative CPU
>> 37.19 sec
>> MapReduce Total cumulative CPU time: 37 seconds 190 msec
>> Ended Job = job_1406011031680_0002 with errors
>> Error during job, obtaining debugging information...
>> Job Tracking URL:
>> http://ec2-54-77-76-145.eu-west-1.compute.amazonaws.com:8088/proxy/application_1406011031680_0002/
>> Examining task ID: task_1406011031680_0002_m_000001 (and more) from job
>> job_1406011031680_0002
>> Examining task ID: task_1406011031680_0002_m_000005 (and more) from job
>> job_1406011031680_0002
>>
>> Task with the most failures(4):
>> -----
>> Task ID:
>>   task_1406011031680_0002_m_000008
>>
>> URL:
>>
>> http://ec2-54-77-76-145.eu-west-1.compute.amazonaws.com:8088/taskdetails.jsp?jobid=job_1406011031680_0002&tipid=task_1406011031680_0002_m_000008
>> -----
>> Diagnostic Messages for this Task:
>> Exception from container-launch:
>> org.apache.hadoop.util.Shell$ExitCodeException:
>>         at org.apache.hadoop.util.Shell.runCommand(Shell.java:464)
>>         at org.apache.hadoop.util.Shell.run(Shell.java:379)
>>         at
>> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:589)
>>         at
>> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)
>>         at
>> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:283)
>>         at
>> org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:79)
>>         at java.util.concurrent.FutureTask.run(FutureTask.java:262)
>>         at
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>         at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>         at java.lang.Thread.run(Thread.java:744)
>>
>>
>> FAILED: Execution Error, return code 2 from
>> org.apache.hadoop.hive.ql.exec.mr.MapRedTask
>> MapReduce Jobs Launched:
>> Job 0: Map: 3  Reduce: 1   Cumulative CPU: 24.58 sec   HDFS Read:
>> 593821601 HDFS Write: 14518009 SUCCESS
>> Job 1: Map: 9   Cumulative CPU: 37.19 sec   HDFS Read: 1342219615 HDFS
>> Write: 821879 FAIL
>> Total MapReduce CPU Time Spent: 1 minutes 1 seconds 770 msec
>> hive (default)> exit;
>>
>>
>>
>> Regards,
>> Sarfraz Rasheed Ramay (DIT)
>> Dublin, Ireland.
>>
>
>

Reply via email to