[ 
https://issues.apache.org/jira/browse/HIVE-9389?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15422883#comment-15422883
 ] 

Murali Parimi commented on HIVE-9389:
-------------------------------------

I don't think this issue is fixed in version 1.2 as well. Is someone tracking 
this?

> INSERT OVERWRITE DIRECTORY fails to delete old data files
> ---------------------------------------------------------
>
>                 Key: HIVE-9389
>                 URL: https://issues.apache.org/jira/browse/HIVE-9389
>             Project: Hive
>          Issue Type: Bug
>    Affects Versions: 0.13.1
>         Environment: CDH 5.3.0, non-secure hdfs, perm checking off
>            Reporter: Andy Skelton
>
> {code:sql}
> FROM myview INSERT OVERWRITE DIRECTORY 'hdfs://nameservice/path/' SELECT 
> COUNT(DISTINCT mycol);
> {code}
> This always produces one row. Sometimes the output is two files, {{000000_0}} 
> and {{000000_1}}, one of which is empty. Sometimes we have seen new results 
> in {{000000_0}} while old results remain in {{000000_1}}.
> We were alerted to this because Sqoop was exporting the output files in order 
> by filename, writing first the new value and then overwriting with the old 
> value, triggering an alert because the value stopped increasing in our 
> database.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to