I use hive under hadoop 2.2.0, first I start hive
[hadoop@master sbin]$ hive
14/04/21 19:06:32 INFO Configuration.deprecation: mapred.input.dir.recursive is 
deprecated. Instead, use mapreduce.input.fileinputformat.input.dir.recursive
14/04/21 19:06:32 INFO Configuration.deprecation: mapred.max.split.size is 
deprecated. Instead, use mapreduce.input.fileinputformat.split.maxsize
14/04/21 19:06:32 INFO Configuration.deprecation: mapred.min.split.size is 
deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize
14/04/21 19:06:32 INFO Configuration.deprecation: 
mapred.min.split.size.per.rack is deprecated. Instead, use 
mapreduce.input.fileinputformat.split.minsize.per.rack
14/04/21 19:06:32 INFO Configuration.deprecation: 
mapred.min.split.size.per.node is deprecated. Instead, use 
mapreduce.input.fileinputformat.split.minsize.per.node
14/04/21 19:06:32 INFO Configuration.deprecation: mapred.reduce.tasks is 
deprecated. Instead, use mapreduce.job.reduces
14/04/21 19:06:32 INFO Configuration.deprecation: 
mapred.reduce.tasks.speculative.execution is deprecated. Instead, use 
mapreduce.reduce.speculative
14/04/21 19:06:32 WARN conf.Configuration: 
org.apache.hadoop.hive.conf.LoopingByteArrayInputStream@2128d0:an attempt to 
override final parameter: mapreduce.job.end-notification.max.retry.interval;  
Ignoring.
14/04/21 19:06:32 WARN conf.Configuration: 
org.apache.hadoop.hive.conf.LoopingByteArrayInputStream@2128d0:an attempt to 
override final parameter: mapreduce.job.end-notification.max.attempts;  
Ignoring.
Logging initialized using configuration in 
jar:file:/home/software/hive-0.11.0/lib/hive-common-0.11.0.jar!/hive-log4j.properties
Hive history 
file=/tmp/hadoop/hive_job_log_hadoop_7623@master_201404211906_2069310090.txt
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in 
[jar:file:/home/software/hadoop-2.2.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in 
[jar:file:/home/software/hive-0.11.0/lib/slf4j-log4j12-1.6.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]

Then I creat a table
hive> create table test(id STRING);
OK
Time taken: 17.277 seconds

Then  I insert some date into test  
hive> load data inpath 'a.txt' overwrite into table test;
Loading data to table default.test
rmr: DEPRECATED: Please use 'rm -r' instead.
Deleted /user/hive/warehouse/test
Table default.test stats: [num_partitions: 0, num_files: 1, num_rows: 0, 
total_size: 19, raw_data_size: 0]
OK
Time taken: 1.855 seconds

hive> select * from test;
OK
China
US
Australia
Time taken: 0.526 seconds, Fetched: 3 row(s)

Now I use count command, I expected the result value is 3, but it runs failure! 
 Why? Where is wrong? I am puzzled with it for several days. Anyone could tell 
me how to correct it?
hive> select count(*) from test;
Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks determined at compile time: 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapred.reduce.tasks=<number>
Starting Job = job_1398132272370_0001, Tracking URL = 
http://master:8088/proxy/application_1398132272370_0001/
Kill Command = /home/software/hadoop-2.2.0/bin/hadoop job  -kill 
job_1398132272370_0001
Hadoop job information for Stage-1: number of mappers: 0; number of reducers: 0
2014-04-21 19:15:56,684 Stage-1 map = 0%,  reduce = 0%
Ended Job = job_1398132272370_0001 with errors
Error during job, obtaining debugging information...
FAILED: Execution Error, return code 2 from 
org.apache.hadoop.hive.ql.exec.MapRedTask
MapReduce Jobs Launched: 
Job 0:  HDFS Read: 0 HDFS Write: 0 FAIL
Total MapReduce CPU Time Spent: 0 msec
hive> 


Error information under 
http://172.11.12.6:8088/cluster/app/application_1398132272370_0001
User:  hadoop
Name:  select count(*) from test(Stage-1)
Application Type:  MAPREDUCE
State:  FAILED
FinalStatus:  FAILED
Started:  21-Apr-2014 19:14:55
Elapsed:  57sec
Tracking URL:  History
Diagnostics:  
Application application_1398132272370_0001 failed 2 times due to AM Container 
for appattempt_1398132272370_0001_000002 exited with exitCode: 1 due to: 
Exception from container-launch:
org.apache.hadoop.util.Shell$ExitCodeException:
at org.apache.hadoop.util.Shell.runCommand(Shell.java:464)
at org.apache.hadoop.util.Shell.run(Shell.java:379)
at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:589)
at 
org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchContainer(DefaultContainerExecutor.java:195)
at 
org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:283)
at 
org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.ContainerLaunch.call(ContainerLaunch.java:79)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
at java.util.concurrent.FutureTask.run(FutureTask.java:166)
at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1110)
at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:603)
at java.lang.Thread.run(Thread.java:722)
.Failing this attempt.. Failing the application 
---------------------------------------------------------------------------------------------------
Confidentiality Notice: The information contained in this e-mail and any 
accompanying attachment(s) 
is intended only for the use of the intended recipient and may be confidential 
and/or privileged of 
Neusoft Corporation, its subsidiaries and/or its affiliates. If any reader of 
this communication is 
not the intended recipient, unauthorized use, forwarding, printing,  storing, 
disclosure or copying 
is strictly prohibited, and may be unlawful.If you have received this 
communication in error,please 
immediately notify the sender by return e-mail, and delete the original message 
and all copies from 
your system. Thank you. 
---------------------------------------------------------------------------------------------------

Reply via email to