[ https://issues.apache.org/jira/browse/HIVE-18910?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16391946#comment-16391946 ]
Hive QA commented on HIVE-18910: -------------------------------- Here are the results of testing the latest attachment: https://issues.apache.org/jira/secure/attachment/12913657/HIVE-18910.1.patch {color:red}ERROR:{color} -1 due to build exiting with an error Test results: https://builds.apache.org/job/PreCommit-HIVE-Build/9557/testReport Console output: https://builds.apache.org/job/PreCommit-HIVE-Build/9557/console Test logs: http://104.198.109.242/logs/PreCommit-HIVE-Build-9557/ Messages: {noformat} Executing org.apache.hive.ptest.execution.TestCheckPhase Executing org.apache.hive.ptest.execution.PrepPhase Tests exited with: NonZeroExitCodeException Command 'bash /data/hiveptest/working/scratch/source-prep.sh' failed with exit status 1 and output '+ date '+%Y-%m-%d %T.%3N' 2018-03-08 21:39:07.418 + [[ -n /usr/lib/jvm/java-8-openjdk-amd64 ]] + export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64 + JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64 + export PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games + PATH=/usr/lib/jvm/java-8-openjdk-amd64/bin/:/usr/local/bin:/usr/bin:/bin:/usr/local/games:/usr/games + export 'ANT_OPTS=-Xmx1g -XX:MaxPermSize=256m ' + ANT_OPTS='-Xmx1g -XX:MaxPermSize=256m ' + export 'MAVEN_OPTS=-Xmx1g ' + MAVEN_OPTS='-Xmx1g ' + cd /data/hiveptest/working/ + tee /data/hiveptest/logs/PreCommit-HIVE-Build-9557/source-prep.txt + [[ false == \t\r\u\e ]] + mkdir -p maven ivy + [[ git = \s\v\n ]] + [[ git = \g\i\t ]] + [[ -z master ]] + [[ -d apache-github-source-source ]] + [[ ! -d apache-github-source-source/.git ]] + [[ ! -d apache-github-source-source ]] + date '+%Y-%m-%d %T.%3N' 2018-03-08 21:39:07.421 + cd apache-github-source-source + git fetch origin >From https://github.com/apache/hive 7edb1d6..9b36ffa master -> origin/master + git reset --hard HEAD HEAD is now at 7edb1d6 HIVE-18861 : druid-hdfs-storage is pulling in hadoop-aws-2.7.x and aws SDK, creating classpath problems on hadoop 3.x (Steve Loughran via Ashutosh Chauhan) + git clean -f -d + git checkout master Already on 'master' Your branch is behind 'origin/master' by 1 commit, and can be fast-forwarded. (use "git pull" to update your local branch) + git reset --hard origin/master HEAD is now at 9b36ffa HIVE-18571 : stats issues for MM tables; ACID doesn't check state for CTAS (Sergey Shelukhin, reviewed by Eugene Koifman) + git merge --ff-only origin/master Already up-to-date. + date '+%Y-%m-%d %T.%3N' 2018-03-08 21:39:15.725 + rm -rf ../yetus_PreCommit-HIVE-Build-9557 + mkdir ../yetus_PreCommit-HIVE-Build-9557 + git gc + cp -R . ../yetus_PreCommit-HIVE-Build-9557 + mkdir /data/hiveptest/logs/PreCommit-HIVE-Build-9557/yetus + patchCommandPath=/data/hiveptest/working/scratch/smart-apply-patch.sh + patchFilePath=/data/hiveptest/working/scratch/build.patch + [[ -f /data/hiveptest/working/scratch/build.patch ]] + chmod +x /data/hiveptest/working/scratch/smart-apply-patch.sh + /data/hiveptest/working/scratch/smart-apply-patch.sh /data/hiveptest/working/scratch/build.patch error: a/hcatalog/streaming/src/java/org/apache/hive/hcatalog/streaming/AbstractRecordWriter.java: does not exist in index error: a/hcatalog/streaming/src/java/org/apache/hive/hcatalog/streaming/mutate/worker/BucketIdResolver.java: does not exist in index error: a/hcatalog/streaming/src/java/org/apache/hive/hcatalog/streaming/mutate/worker/BucketIdResolverImpl.java: does not exist in index error: a/hcatalog/streaming/src/java/org/apache/hive/hcatalog/streaming/mutate/worker/MutatorCoordinator.java: does not exist in index error: a/hcatalog/streaming/src/test/org/apache/hive/hcatalog/streaming/mutate/worker/TestBucketIdResolverImpl.java: does not exist in index error: a/ql/src/java/org/apache/hadoop/hive/ql/exec/ReduceSinkOperator.java: does not exist in index error: a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/keyseries/VectorKeySeriesSerializedImpl.java: does not exist in index error: a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkCommonOperator.java: does not exist in index error: a/ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java: does not exist in index error: a/ql/src/java/org/apache/hadoop/hive/ql/metadata/Table.java: does not exist in index error: a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/ConvertJoinMapJoin.java: does not exist in index error: a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/FixedBucketPruningOptimizer.java: does not exist in index error: a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/PrunerOperatorFactory.java: does not exist in index error: a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/metainfo/annotation/OpTraitsRulesProcFactory.java: does not exist in index error: a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java: does not exist in index error: a/ql/src/java/org/apache/hadoop/hive/ql/optimizer/spark/SparkMapJoinOptimizer.java: does not exist in index error: a/ql/src/java/org/apache/hadoop/hive/ql/plan/OpTraits.java: does not exist in index error: a/ql/src/test/queries/clientpositive/tez_dynpart_hashjoin_3.q: does not exist in index error: a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out: does not exist in index error: a/serde/src/java/org/apache/hadoop/hive/serde2/binarysortable/BinarySortableSerDe.java: does not exist in index error: a/serde/src/java/org/apache/hadoop/hive/serde2/objectinspector/ObjectInspectorUtils.java: does not exist in index error: a/serde/src/test/org/apache/hadoop/hive/serde2/objectinspector/TestObjectInspectorUtils.java: does not exist in index error: a/standalone-metastore/src/gen/thrift/gen-javabean/org/apache/hadoop/hive/metastore/api/hive_metastoreConstants.java: does not exist in index error: a/storage-api/src/java/org/apache/hive/common/util/Murmur3.java: does not exist in index error: patch failed: ql/src/java/org/apache/hadoop/hive/ql/exec/ReduceSinkOperator.java:31 Falling back to three-way merge... Applied patch to 'ql/src/java/org/apache/hadoop/hive/ql/exec/ReduceSinkOperator.java' cleanly. error: patch failed: ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java:51 Falling back to three-way merge... Applied patch to 'ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java' with conflicts. error: patch failed: ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java:44 Falling back to three-way merge... Applied patch to 'ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java' cleanly. Going to apply patch with: git apply -p1 error: patch failed: ql/src/java/org/apache/hadoop/hive/ql/exec/ReduceSinkOperator.java:31 Falling back to three-way merge... Applied patch to 'ql/src/java/org/apache/hadoop/hive/ql/exec/ReduceSinkOperator.java' cleanly. error: patch failed: ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java:51 Falling back to three-way merge... Applied patch to 'ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java' with conflicts. error: patch failed: ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java:44 Falling back to three-way merge... Applied patch to 'ql/src/java/org/apache/hadoop/hive/ql/optimizer/physical/Vectorizer.java' cleanly. U ql/src/java/org/apache/hadoop/hive/ql/exec/vector/reducesink/VectorReduceSinkObjectHashOperator.java + exit 1 ' {noformat} This message is automatically generated. ATTACHMENT ID: 12913657 - PreCommit-HIVE-Build > Migrate to Murmur hash for shuffle and bucketing > ------------------------------------------------ > > Key: HIVE-18910 > URL: https://issues.apache.org/jira/browse/HIVE-18910 > Project: Hive > Issue Type: Task > Reporter: Deepak Jaiswal > Assignee: Deepak Jaiswal > Priority: Major > Attachments: HIVE-18910.1.patch > > > Hive uses JAVA hash which is not as good as murmur for better distribution > and efficiency in bucketing a table. > Migrate to murmur hash but still keep backward compatibility for existing > users so that they dont have to reload the existing tables. -- This message was sent by Atlassian JIRA (v7.6.3#76005)