[ https://issues.apache.org/jira/browse/HIVE-20020?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16526295#comment-16526295 ]
BELUGA BEHR commented on HIVE-20020: ------------------------------------ Just to echo what [~johndee] said in regards to the {{MultiDelimitSerDe}} SerDe, it is confusing because as it stands, the following scenarios exists: # Create table with Serde (/) # Execute SELECT * FROM <table> LIMIT 10 (/) # Execute SELECT * FROM TABLE WHERE ... LIMIT 10 (x) This is very confusing and inconsistent. The last one fails because the first two operations do not require a MapReduce/Spark job. All of the work happens with HS2 and it has access to the hive-contrib JAR in its classpath, but the JAR file is not sent along into the cluster for MapReduce/Spark jobs. > Hive contrib jar should not be in lib > ------------------------------------- > > Key: HIVE-20020 > URL: https://issues.apache.org/jira/browse/HIVE-20020 > Project: Hive > Issue Type: Improvement > Components: Contrib > Reporter: Johndee Burks > Priority: Trivial > > Currently the way hive is packaged it includes hive-contrib-<version>.jar in > lib, we should not include it here because it is picked up by services like > HS2. This creates a situation in which experimental features such as the > [MultiDelimitSerDe|https://github.com/apache/hive/blob/master/contrib/src/java/org/apache/hadoop/hive/contrib/serde2/MultiDelimitSerDe.java] > are accessible without understanding how to really install and use it. For > example you can create a table using HS2 via beeline with the aforementioned > SerDe and it will work as long you do not do M/R jobs. The M/R jobs do not > work because the SerDe is not in aux to get shipped into distcache. I propose > we do not package it this way and if someone would like to leverage an > experimental feature they can add it manually to their environment. -- This message was sent by Atlassian JIRA (v7.6.3#76005)