perhaps deleteOnExit() is set somewhere From: Jeff Zhang <zjf...@gmail.com<mailto:zjf...@gmail.com>> Reply-To: User <user@hive.apache.org<mailto:user@hive.apache.org>> Date: Thursday, June 25, 2015 at 2:09 AM To: User <user@hive.apache.org<mailto:user@hive.apache.org>> Subject: Re: hive -e run tez query error
The drop table command between these 2 query fails due to permission issue. If I solve the permission issue, the tez error will disappear. But what confuse me is that I suppose the failure of drop table command should not affect the following tez jobs. But from the hdfs-audit logs, tez_staging_directory is deleted between the first query and the second query. But besides the audit log, I don't see anywhere else about the deleting of tez_staging_directory including hive log & yarn app log. Does anyone know whether hive will delete the tez staging directory in this case ? Here's the hive log and hdfs-audit log. hive.log 2015-06-25 15:54:33,623 INFO [main]: metastore.HiveMetaStore (HiveMetaStore.java:logInfo(732)) - 0: drop_table : db=TESTDEFAULT tbl=tmp_recomm_prdt_detail0 2015-06-25 15:54:33,623 INFO [main]: HiveMetaStore.audit (HiveMetaStore.java:logAuditEvent(358)) - ugi=lujian ip=unknown-ip-addr cmd=drop_table : db=TESTDEFAULT tbl=tmp_recomm_prdt_detail0 2015-06-25 15:54:33,658 INFO [main]: metastore.hivemetastoressimpl (HiveMetaStoreFsImpl.java:deleteDir(41)) - deleting hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/hive/warehouse/testdefault.db/tmp_recomm_prdt_detail0<http://yhd-jqhadoop2.int.yihaodian.com:8020/user/hive/warehouse/testdefault.db/tmp_recomm_prdt_detail0> 2015-06-25 15:54:33,660 INFO [main]: fs.TrashPolicyDefault (TrashPolicyDefault.java:initialize(92)) - Namenode trash configuration: Deletion interval = 1440 minutes, Emptier interval = 0 minutes. 2015-06-25 15:54:33,669 ERROR [main]: hive.log (MetaStoreUtils.java:logAndThrowMetaException(1173)) - Got exception: java.io.IOException Failed to move to trash: hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/hive/warehouse/testdefault.db/tmp_recomm_prdt_detail0<http://yhd-jqhadoop2.int.yihaodian.com:8020/user/hive/warehouse/testdefault.db/tmp_recomm_prdt_detail0> java.io.IOException: Failed to move to trash: hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/hive/warehouse/testdefault.db/tmp_recomm_prdt_detail0<http://yhd-jqhadoop2.int.yihaodian.com:8020/user/hive/warehouse/testdefault.db/tmp_recomm_prdt_detail0> at org.apache.hadoop.fs.TrashPolicyDefault.moveToTrash(TrashPolicyDefault.java:160) at org.apache.hadoop.fs.Trash.moveToTrash(Trash.java:109) .. (remove the full stacktrace) 2015-06-25 15:54:33,670 ERROR [main]: hive.log (MetaStoreUtils.java:logAndThrowMetaException(1174)) - Converting exception to MetaException 2015-06-25 15:54:33,670 ERROR [main]: metastore.HiveMetaStore (HiveMetaStore.java:deleteTableData(1557)) - Failed to delete table directory: hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/hive/warehouse/testdefault.db/tmp_recomm_prdt_detail0<http://yhd-jqhadoop2.int.yihaodian.com:8020/user/hive/warehouse/testdefault.db/tmp_recomm_prdt_detail0> Got exception: java.io.IOException Failed to move to trash: hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/hive/warehouse/testdefault.db/tmp_recomm_prdt_detail0<http://yhd-jqhadoop2.int.yihaodian.com:8020/user/hive/warehouse/testdefault.db/tmp_recomm_prdt_detail0> 2015-06-25 15:54:33,670 INFO [main]: log.PerfLogger (PerfLogger.java:PerfLogEnd(148)) - </PERFLOG method=runTasks start=1435218873582 end=1435218873670 duration=88 from=org.apache.hadoop.hive.ql.Driver> 2015-06-25 15:54:33,670 INFO [main]: log.PerfLogger (PerfLogger.java:PerfLogEnd(148)) - </PERFLOG method=Driver.execute start=1435218873582 end=1435218873670 duration=88 from=org.apache.hadoop.hive.ql.Driver> 2015-06-25 15:54:33,671 INFO [main]: ql.Driver (SessionState.java:printInfo(852)) - OK 2015-06-25 15:54:33,671 INFO [main]: log.PerfLogger (PerfLogger.java:PerfLogBegin(121)) - <PERFLOG method=releaseLocks from=org.apache.hadoop.hive.ql.Driver> 2015-06-25 15:54:33,671 INFO [main]: log.PerfLogger (PerfLogger.java:PerfLogEnd(148)) - </PERFLOG method=releaseLocks start=1435218873671 end=1435218873671 duration=0 from=org.apache.hadoop.hive.ql.Driver> 2015-06-25 15:54:33,671 INFO [main]: log.PerfLogger (PerfLogger.java:PerfLogEnd(148)) - </PERFLOG method=Driver.run start=1435218873554 end=1435218873671 duration=117 from=org.apache.hadoop.hive.ql.Driver> 2015-06-25 15:54:33,671 INFO [main]: CliDriver (SessionState.java:printInfo(852)) - Time taken: 0.117 seconds 2015-06-25 15:54:33,671 INFO [main]: log.PerfLogger (PerfLogger.java:PerfLogBegin(121)) - <PERFLOG method=Driver.run from=org.apache.hadoop.hive.ql.Driver> 2015-06-25 15:54:33,671 INFO [main]: log.PerfLogger (PerfLogger.java:PerfLogBegin(121)) - <PERFLOG method=TimeToSubmit from=org.apache.hadoop.hive.ql.Driver> 2015-06-25 15:54:33,671 INFO [main]: log.PerfLogger (PerfLogger.java:PerfLogBegin(121)) - <PERFLOG method=compile from=org.apache.hadoop.hive.ql.Driver> 2015-06-25 15:54:33,672 INFO [main]: log.PerfLogger (PerfLogger.java:PerfLogBegin(121)) - <PERFLOG method=parse from=org.apache.hadoop.hive.ql.Driver> 2015-06-25 15:54:33,672 INFO [main]: parse.ParseDriver (ParseDriver.java:parse(185)) - Parsing command: create table TESTDEFAULT.tmp_recomm_prdt_detail0(track_id bigint,cart_time string,product_id bigint, session_id string,gu_id string,url string,referer string,province_id bigint,page_type bigint, button_position string,link_position string,ip string,m_session_id string,link_id bigint, button_id bigint,algorithm_id bigint,from_value string,ref_value string,step string,platform_type_id int,pm_info_id bigint,refer_page_value string,url_page_value string) 2015-06-25 15:54:33,674 INFO [main]: parse.ParseDriver (ParseDriver.java:parse(206)) - Parse Completed 2015-06-25 15:54:33,674 INFO [main]: log.PerfLogger (PerfLogger.java:PerfLogEnd(148)) - </PERFLOG method=parse start=1435218873672 end=1435218873674 duration=2 from=org.apache.hadoop.hive.ql.Driver> 2015-06-25 15:54:33,674 INFO [main]: log.PerfLogger (PerfLogger.java:PerfLogBegin(121)) - <PERFLOG method=semanticAnalyze from=org.apache.hadoop.hive.ql.Driver> 2015-06-25 15:54:33,675 INFO [main]: parse.CalcitePlanner (SemanticAnalyzer.java:analyzeInternal(10146)) - Starting Semantic Analysis hdfs-audit.log 2015-06-25 15:54:33,673 INFO FSNamesystem.audit: allowed=true ugi=lujian (auth:SIMPLE) ip=/10.17.28.11<http://10.17.28.11> cmd=delete src=/user/lujian/lujian/_tez_session_dir/abb91da9-ac07-4024-a09f-8622ee1caedf dst=null perm=null On Tue, Jun 23, 2015 at 2:02 PM, r7raul1...@163.com<mailto:r7raul1...@163.com> <r7raul1...@163.com<mailto:r7raul1...@163.com>> wrote: Sorry , I made a mistake, use hivecli to execute same query on tez ,throw the same exception. ________________________________ r7raul1...@163.com<mailto:r7raul1...@163.com> From: r7raul1...@163.com<mailto:r7raul1...@163.com> Date: 2015-06-23 13:53 To: user<mailto:user@hive.apache.org> Subject: hive -e run tez query error When I use hive 1.1.0 on tez 0.53 in hadoop 2.3.0: hive -v -e "set hive.execution.engine=tez;set mapred.job.queue.name<http://mapred.job.queue.name>=bi_etl;drop table TESTTMP.a_start;create table TESTTMP.a_start(id bigint);insert overwrite table TESTTMP.a_start select id from tandem.p_city;drop table TESTTMP.a_end;create table TESTTMP.a_end(id bigint);insert overwrite table TESTTMP.a_end select id from TESTTMP.a_start;" Logging initialized using configuration in jar:file:/usr/local/src/apache-hive/lib/hive-common-1.1.0.jar!/hive-log4j.properties SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/usr/local/src/apache-hive/lib/hive-jdbc-1.1.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/usr/lib/zookeeper/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] set hive.execution.engine=tez set mapred.job.queue.name<http://mapred.job.queue.name>=bi_etl drop table TESTTMP.a_start OK Time taken: 1.702 seconds create table TESTTMP.a_start(id bigint) OK Time taken: 0.311 seconds insert overwrite table TESTTMP.a_start select id from tandem.p_city Query ID = lujian_20150623134848_31ba3183-74b8-4c3a-96ae-5c8b650b99df Total jobs = 1 Launching Job 1 out of 1 Status: Running (Executing on YARN cluster with App id application_1433219182593_252390) -------------------------------------------------------------------------------- VERTICES STATUS TOTAL COMPLETED RUNNING PENDING FAILED KILLED -------------------------------------------------------------------------------- Map 1 .......... SUCCEEDED 1 1 0 0 0 0 -------------------------------------------------------------------------------- VERTICES: 01/01 [==========================>>] 100% ELAPSED TIME: 13.55 s -------------------------------------------------------------------------------- Loading data to table testtmp.a_start Moved: 'hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/hive/warehouse/testtmp.db/a_start/000000_0<http://yhd-jqhadoop2.int.yihaodian.com:8020/user/hive/warehouse/testtmp.db/a_start/000000_0>' to trash at: hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/.Trash/Current<http://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/.Trash/Current> Table testtmp.a_start stats: [numFiles=1, numRows=531, totalSize=2154, rawDataSize=1623] OK Time taken: 25.586 seconds drop table TESTTMP.a_end OK Time taken: 0.232 seconds create table TESTTMP.a_end(id bigint) OK Time taken: 0.068 seconds insert overwrite table TESTTMP.a_end select id from TESTTMP.a_start Query ID = lujian_20150623134949_bff735c9-6abc-47e7-a9f7-f7e2be7e43e9 Total jobs = 1 Launching Job 1 out of 1 Status: Running (Executing on YARN cluster with App id application_1433219182593_252390) -------------------------------------------------------------------------------- VERTICES STATUS TOTAL COMPLETED RUNNING PENDING FAILED KILLED -------------------------------------------------------------------------------- Map 1 FAILED 1 0 0 1 4 0 -------------------------------------------------------------------------------- VERTICES: 00/01 [>>--------------------------] 0% ELAPSED TIME: 15.46 s -------------------------------------------------------------------------------- Status: Failed Vertex failed, vertexName=Map 1, vertexId=vertex_1433219182593_252390_2_00, diagnostics=[Task failed, taskId=task_1433219182593_252390_2_00_000000, diagnostics=[TaskAttempt 0 failed, info=[Container container_1433219182593_252390_01_000003 finished with diagnostics set to [Container failed. File does not exist: hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/lujian/_tez_session_dir/63de23a2-1cff-4434-96ad-1304089fb489/.tez/application_1433219182593_252390/tez-conf.pb<http://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/lujian/_tez_session_dir/63de23a2-1cff-4434-96ad-1304089fb489/.tez/application_1433219182593_252390/tez-conf.pb> ]], TaskAttempt 1 failed, info=[Container container_1433219182593_252390_01_000004 finished with diagnostics set to [Container failed. File does not exist: hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/lujian/_tez_session_dir/63de23a2-1cff-4434-96ad-1304089fb489/.tez/application_1433219182593_252390/tez-conf.pb<http://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/lujian/_tez_session_dir/63de23a2-1cff-4434-96ad-1304089fb489/.tez/application_1433219182593_252390/tez-conf.pb> ]], TaskAttempt 2 failed, info=[Container container_1433219182593_252390_01_000005 finished with diagnostics set to [Container failed. File does not exist: hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/lujian/_tez_session_dir/63de23a2-1cff-4434-96ad-1304089fb489/.tez/application_1433219182593_252390/tez-conf.pb<http://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/lujian/_tez_session_dir/63de23a2-1cff-4434-96ad-1304089fb489/.tez/application_1433219182593_252390/tez-conf.pb> ]], TaskAttempt 3 failed, info=[Container container_1433219182593_252390_01_000006 finished with diagnostics set to [Container failed. File does not exist: hdfs://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/lujian/_tez_session_dir/63de23a2-1cff-4434-96ad-1304089fb489/.tez/application_1433219182593_252390/<http://yhd-jqhadoop2.int.yihaodian.com:8020/user/lujian/lujian/_tez_session_dir/63de23a2-1cff-4434-96ad-1304089fb489/.tez/application_1433219182593_252390/>tez-conf.pb ]]], Vertex failed as one or more tasks failed. failedTasks:1, Vertex vertex_1433219182593_252390_2_00 [Map 1] killed/failed due to:null] DAG failed due to vertex failure. failedVertices:1 killedVertices:0 FAILED: Execution Error, return code 2 from org.apache.hadoop.hive.ql.exec.tez.TezTask use hivecli to execute same query ,no exception throw. use hive -e to execute same query on mr ,no exeception throw ________________________________ r7raul1...@163.com<mailto:r7raul1...@163.com> -- Best Regards Jeff Zhang