You should be building Spark without Hive. For Spark 2.3.0, the command is:

./dev/make-distribution.sh --name "hadoop2-without-hive" --tgz
"-Pyarn,hadoop-provided,hadoop-2.7,parquet-provided,orc-provided
<https://cwiki.apache.org/confluence/display/Hive/hadoop-2.7,parquet-provided,orc-provided>
"

If you check the distribution after running the command, it shouldn't
contain any Hive jars.

On Tue, Jun 19, 2018 at 7:18 AM, Sachin janani <sachin.janani...@gmail.com>
wrote:

> It shows following exception :
>
>
>
>
> *java.lang.NoSuchFieldError: HIVE_STATS_JDBC_TIMEOUTat
> org.apache.spark.sql.hive.HiveUtils$.formatTimeVarsForHiveClient(HiveUtils.scala:205)at
> org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:286)at
> org.apache.spark.sql.hive.HiveExternalCatalog.client$lzycompute(HiveExternalCatalog.scala:66)*
> *at
> org.apache.spark.sql.hive.HiveExternalCatalog.client(HiveExternalCatalog.scala:65)*
>
>
> After looking at jira SPARK-13446
> <https://issues.apache.org/jira/browse/SPARK-13446> it seems that it is
> fixed but as per the source code it is not. So I resolved by changing spark
> code and rebuilding the spark binaries again but now it shows new error
> NoSuchMethodError. As per my preliminary investigation it seems that Spark
> is build with Hive 1.2.1 which is causing this issues. Can you please let
> me know if i am missing anything?
>
>
> Regards,
> Sachin Janani
>
> On Tue, Jun 19, 2018 at 5:38 PM, Sahil Takiar <takiar.sa...@gmail.com>
> wrote:
>
>> I updated the doc to reflect that Hive 3.0.0 works with Spark 2.3.0. What
>> issues are you seeing?
>>
>> On Tue, Jun 19, 2018 at 7:03 AM, Sachin janani <
>> sachin.janani...@gmail.com> wrote:
>>
>>> This is the same link which I followed. As per this link for
>>> spark-2.3.0 we need to use hive master instead of hive 3.0.0. Also we
>>> need to custom build spark without hive dependencies but after trying
>>> all this it shows some compatibility issues.
>>>
>>>
>>> Regards,
>>> Sachin Janani
>>>
>>> On Tue, Jun 19, 2018 at 5:02 PM, Sahil Takiar <takiar.sa...@gmail.com>
>>> wrote:
>>> > Yes, Hive 3.0.0 works with Spark 2.3.0 - this section of the wiki has
>>> > details on which Hive releases support which Spark versions.
>>> >
>>> > On Tue, Jun 19, 2018 at 5:59 AM, Sachin janani <
>>> sachin.janani...@gmail.com>
>>> > wrote:
>>> >>
>>> >> Hi,
>>> >> I am trying to run hive on spark by following the steps mentioned
>>> >> here-
>>> >> https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spa
>>> rk%3A+Getting+Started
>>> >> , but getting many compatibility issues like NoSuchMethodError,
>>> >> NoSuchFieldException etc. So just need to know if it works and whether
>>> >> someone tried it out,
>>> >>
>>> >>
>>> >> Thanks and Regards,
>>> >> --
>>> >> Sachin Janani
>>> >
>>> >
>>> >
>>> >
>>> > --
>>> > Sahil Takiar
>>> > Software Engineer
>>> > takiar.sa...@gmail.com | (510) 673-0309
>>>
>>>
>>>
>>> --
>>> Sachin Janani
>>>
>>
>>
>>
>> --
>> Sahil Takiar
>> Software Engineer
>> takiar.sa...@gmail.com | (510) 673-0309
>>
>
>
>
> --
> *Sachin Janani*
>
>



-- 
Sahil Takiar
Software Engineer
takiar.sa...@gmail.com | (510) 673-0309

Reply via email to