[ https://issues.apache.org/jira/browse/HIVE-8235?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14158695#comment-14158695 ]
Alan Gates commented on HIVE-8235: ---------------------------------- [~mmokhtar] can you try to rerun this now that HIVE-8290 has been committed? > Insert into partitioned bucketed sorted tables fails with "this file is > already being created by" > ------------------------------------------------------------------------------------------------- > > Key: HIVE-8235 > URL: https://issues.apache.org/jira/browse/HIVE-8235 > Project: Hive > Issue Type: Bug > Components: Query Processor > Affects Versions: 0.14.0 > Environment: cn105 > Reporter: Mostafa Mokhtar > Assignee: Alan Gates > Priority: Critical > Fix For: 0.14.0 > > Attachments: insert_into_partitioned_bucketed_table.txt.tar.gz.zip > > > When loading into a partitioned bucketed sorted table the query fails with > {code} > Caused by: > org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException): > Failed to create file > [/tmp/hive/mmokhtar/621d7923-90d1-4d9d-a4c6-b3bb075c7a8c/hive_2014-09-22_23-25-11_678_1598300430132235708-1/_task_tmp.-ext-10000/ss_sold_date=1998-01-02/_tmp.000000_3/delta_0123305_0123305/bucket_00000] > for [DFSClient_attempt_1406566393272_6085_r_000144_3_-1677753045_12] for > client [172.21.128.111], because this file is already being created by > [DFSClient_attempt_1406566393272_6085_r_000031_3_-1506661042_12] on > [172.21.128.122] > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.recoverLeaseInternal(FSNamesystem.java:2543) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInternal(FSNamesystem.java:2308) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFileInt(FSNamesystem.java:2237) > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.startFile(FSNamesystem.java:2190) > at > org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.create(NameNodeRpcServer.java:520) > at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.create(ClientNamenodeProtocolServerSideTranslatorPB.java:354) > at > org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013) > at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009) > at java.security.AccessController.doPrivileged(Native Method) > at javax.security.auth.Subject.doAs(Subject.java:415) > at > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1557) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007) > at org.apache.hadoop.ipc.Client.call(Client.java:1410) > at org.apache.hadoop.ipc.Client.call(Client.java:1363) > at > org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206) > at com.sun.proxy.$Proxy15.create(Unknown Source) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:190) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:103) > at com.sun.proxy.$Proxy15.create(Unknown Source) > at > org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.create(ClientNamenodeProtocolTranslatorPB.java:258) > at > org.apache.hadoop.hdfs.DFSOutputStream.newStreamForCreate(DFSOutputStream.java:1600) > at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1465) > at org.apache.hadoop.hdfs.DFSClient.create(DFSClient.java:1390) > at > org.apache.hadoop.hdfs.DistributedFileSystem$6.doCall(DistributedFileSystem.java:394) > at > org.apache.hadoop.hdfs.DistributedFileSystem$6.doCall(DistributedFileSystem.java:390) > at > org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81) > at > org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:390) > at > org.apache.hadoop.hdfs.DistributedFileSystem.create(DistributedFileSystem.java:334) > at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:906) > at org.apache.hadoop.fs.FileSystem.create(FileSystem.java:887) > at > org.apache.hadoop.hive.ql.io.orc.WriterImpl.getStream(WriterImpl.java:1966) > at > org.apache.hadoop.hive.ql.io.orc.WriterImpl.flushStripe(WriterImpl.java:1983) > at > org.apache.hadoop.hive.ql.io.orc.WriterImpl.close(WriterImpl.java:2287) > at > org.apache.hadoop.hive.ql.io.orc.OrcRecordUpdater.close(OrcRecordUpdater.java:356) > at org.apache.hadoop.hive.ql.exec.FileSinkOperator$FSPaths.closeWriter > {code} > DDL > {code} > CREATE TABLE store_sales_sb ( > ss_sold_date_sk int, > ss_sold_time_sk int, > ss_item_sk int, > ss_customer_sk int, > ss_cdemo_sk int, > ss_hdemo_sk int, > ss_addr_sk int, > ss_store_sk int, > ss_promo_sk int, > ss_ticket_number int, > ss_quantity int, > ss_wholesale_cost float, > ss_list_price float, > ss_sales_price float, > ss_ext_discount_amt float, > ss_ext_sales_price float, > ss_ext_wholesale_cost float, > ss_ext_list_price float, > ss_ext_tax float, > ss_coupon_amt float, > ss_net_paid float, > ss_net_paid_inc_tax float, > ss_net_profit float) > PARTITIONED BY(ss_sold_date string) > CLUSTERED BY(ss_item_sk,ss_ticket_number) SORTED > BY(ss_item_sk,ss_ticket_number) INTO 32 BUCKETS > STORED AS ORC; > {code} > DML > {code} > insert overwrite table store_sales_sb partition (ss_sold_date) > select > ss.ss_sold_date_sk, > ss.ss_sold_time_sk, > ss.ss_item_sk, > ss.ss_customer_sk, > ss.ss_cdemo_sk, > ss.ss_hdemo_sk, > ss.ss_addr_sk, > ss.ss_store_sk, > ss.ss_promo_sk, > ss.ss_ticket_number, > ss.ss_quantity, > ss.ss_wholesale_cost, > ss.ss_list_price, > ss.ss_sales_price, > ss.ss_ext_discount_amt, > ss.ss_ext_sales_price, > ss.ss_ext_wholesale_cost, > ss.ss_ext_list_price, > ss.ss_ext_tax, > ss.ss_coupon_amt, > ss.ss_net_paid, > ss.ss_net_paid_inc_tax, > ss.ss_net_profit, > ss.ss_sold_date > from store_sales ss; > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)