Andy Skelton created HIVE-9389:
----------------------------------

             Summary: INSERT OVERWRITE DIRECTORY fails to delete old data files
                 Key: HIVE-9389
                 URL: https://issues.apache.org/jira/browse/HIVE-9389
             Project: Hive
          Issue Type: Bug
    Affects Versions: 0.13.1
         Environment: CDH 5.3.0, non-secure hdfs, perm checking off
            Reporter: Andy Skelton


{code:sql}
FROM myview INSERT OVERWRITE DIRECTORY 'hdfs://nameservice/path/' SELECT 
COUNT(DISTINCT mycol);
{code}

This always produces one row. Sometimes the output is two files, {{000000_0}} 
and {{000000_1}}, one of which is empty. Sometimes we have seen new results in 
{{000000_0}} while old results remain in {{000000_1}}.

We were alerted to this because Sqoop was exporting the output files in order 
by filename, writing first the new value and then overwriting with the old 
value, triggering an alert because the value stopped increasing in our database.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to