GitHub user jerryshao opened a pull request: https://github.com/apache/hive/pull/364
HIVE-16391: Add a new classifier for hive-exec to be used by Spark This fix adding a new classifier for hive-exec artifact (`core-spark`), which is specifically used for Spark. Details in [SPARK-20202](https://issues.apache.org/jira/browse/SPARK-20202). This is because original hive-exec packages many transitive dependencies into shaded jar without relocation, this makes conflicts in Spark. Spark only needs to relocate protobuf and kryo jar. So here propose to add a new classifier to generate a new artifact only for Spark. You can merge this pull request into a Git repository by running: $ git pull https://github.com/jerryshao/hive 1.2-spark-fix Alternatively you can review and apply these changes as the patch at: https://github.com/apache/hive/pull/364.patch To close this pull request, make a commit to your master/trunk branch with (at least) the following in the commit message: This closes #364 ---- commit bb27b260d82fa0a77d9fea3c123f2af8f1ea88aa Author: jerryshao <sshao@...> Date: 2018-06-05T06:59:37Z HIVE-16391: Add a new classifier for hive-exec to be used by Spark ---- ---