Hi, everyone:
I have set up a hadoop cluster on THREE windows7 machines with Cygwin, and made 
several test, which were all passed, with hadoop-test-1.1.2.jar and 
hadoop-examples-1.1.2.jar. 
Then I tried to run Hive 0.10.0 on my cluster ( also in Cygwin ). I could 
create tables, show them, load data into them and "select *" from them. But 
when I tried "select count(*)" from my table, I've got the following exception. 
My question is: what is that HIVE_PLANxxxxxx file? how is it created? where is 
it placed?
Would anyone give me some infomation? 
......
Total MapReduce jobs = 1
Launching Job 1 out of 1
Number of reduce tasks determined at compile time: 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapred.reduce.tasks=<number>
Starting Job = job_201306070901_0001, Tracking URL = 
http://hdfs-namenode:50030/jobdetails.jsp?jobid=job_201306070901_0001
Kill Command = C:\hadoop-1.1.2\\bin\hadoop.cmd job  -kill job_201306070901_0001
Hadoop job information for Stage-1: number of mappers: 13; number of reducers: 1
2013-06-07 09:02:19,296 Stage-1 map = 0%,  reduce = 0%
2013-06-07 09:02:51,745 Stage-1 map = 100%,  reduce = 100%
Ended Job = job_201306070901_0001 with errors
Error during job, obtaining debugging information...
Job Tracking URL: 
http://hdfs-namenode:50030/jobdetails.jsp?jobid=job_201306070901_0001
Examining task ID: task_201306070901_0001_m_000014 (and more) from job 
job_201306070901_0001

Task with the most failures(4): 
-----
Task ID:
  task_201306070901_0001_m_000006

URL:
  
http://hdfs-namenode:50030/taskdetails.jsp?jobid=job_201306070901_0001&tipid=task_201306070901_0001_m_000006
-----
Diagnostic Messages for this Task:
java.lang.RuntimeException: java.io.FileNotFoundException: 
HIVE_PLANc632c8e2-257d-4cd4-b833-a09c7d249b2c (Access is denied)
        at 
org.apache.hadoop.hive.ql.exec.Utilities.getMapRedWork(Utilities.java:226)
        at 
org.apache.hadoop.hive.ql.io.HiveInputFormat.init(HiveInputFormat.java:255)
        at 
org.apache.hadoop.hive.ql.io.HiveInputFormat.pushProjectionsAndFilters(HiveInputFormat.java:381)
        at 
org.apache.hadoop.hive.ql.io.HiveInputFormat.pushProjectionsAndFilters(HiveInputFormat.java:374)
        at 
org.apache.hadoop.hive.ql.io.CombineHiveInputFormat.getRecordReader(CombineHiveInputFormat.java:536)
        at 
org.apache.hadoop.mapred.MapTask$TrackedRecordReader.<init>(MapTask.java:197)
        at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:418)
        at org.apache.hadoop.mapred.MapTask.run(MapTask.java:372)
        at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Unknown Source)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1149)
        at org.apache.hadoop.mapred.Child.main(Child.java:249)
Caused by: java.io.FileNotFoundException: 
HIVE_PLANc632c8e2-257d-4cd4-b833-a09c7d249b2c (Access is denied)
        at java.io.FileInputStream.open(Native Method)
        at java.io.FileInputStream.<init>(Unknown Source)
        at java.io.FileInputStream.<init>(Unknown Source)
        at 
org.apache.hadoop.hive.ql.exec.Utilities.getMapRedWork(Utilities.java:217)
        ... 12 more


FAILED: Execution Error, return code 2 from 
org.apache.hadoop.hive.ql.exec.MapRedTask
MapReduce Jobs Launched: 
Job 0: Map: 13  Reduce: 1   HDFS Read: 0 HDFS Write: 0 FAIL
Total MapReduce CPU Time Spent: 0 msec
                                          

Reply via email to