----------------------------------------------------------- This is an automatically generated e-mail. To reply, visit: https://reviews.apache.org/r/26457/ -----------------------------------------------------------
(Updated Oct. 8, 2014, 9 p.m.) Review request for hive. Repository: hive-git Description ------- run the test mvn -Phadoop-2 test -Dtest=TestCliDriver -Dqfile=alter_merge_2_orc.q to reproduce it. Simply, this query does three data loads which generates three orc files, ALTER TABLE CONCATENATE tries to merge orc pieces into a single one which is final file to queried. Output \hive\itests\qtest\target\qfile-results\clientpositive\alter_merge_2_orc.q.out shows # records as 600 that is wrong as opposed to 610 expected. Because OrcFileMergeOperator only closes last orc file, the 1st and 2nd orc files still remain in table directory due to failure of deleting unclosed file for old data clean when MoveTask tries to copy merged orc file from scratch dir to table dir. Eventually the query goes to old data(1st and 2nd orc files). Diffs ----- ql/src/java/org/apache/hadoop/hive/ql/exec/OrcFileMergeOperator.java 0fb1c57 Diff: https://reviews.apache.org/r/26457/diff/ Testing (updated) ------- Cleared a bunch of orc file merge UT falures. Thanks, XIAOBING ZHOU