You should be building Spark without Hive. For Spark 2.3.0, the command is:
./dev/make-distribution.sh --name "hadoop2-without-hive" --tgz "-Pyarn,hadoop-provided,hadoop-2.7,parquet-provided,orc-provided <https://cwiki.apache.org/confluence/display/Hive/hadoop-2.7,parquet-provided,orc-provided> " If you check the distribution after running the command, it shouldn't contain any Hive jars. On Tue, Jun 19, 2018 at 7:18 AM, Sachin janani <sachin.janani...@gmail.com> wrote: > It shows following exception : > > > > > *java.lang.NoSuchFieldError: HIVE_STATS_JDBC_TIMEOUTat > org.apache.spark.sql.hive.HiveUtils$.formatTimeVarsForHiveClient(HiveUtils.scala:205)at > org.apache.spark.sql.hive.HiveUtils$.newClientForMetadata(HiveUtils.scala:286)at > org.apache.spark.sql.hive.HiveExternalCatalog.client$lzycompute(HiveExternalCatalog.scala:66)* > *at > org.apache.spark.sql.hive.HiveExternalCatalog.client(HiveExternalCatalog.scala:65)* > > > After looking at jira SPARK-13446 > <https://issues.apache.org/jira/browse/SPARK-13446> it seems that it is > fixed but as per the source code it is not. So I resolved by changing spark > code and rebuilding the spark binaries again but now it shows new error > NoSuchMethodError. As per my preliminary investigation it seems that Spark > is build with Hive 1.2.1 which is causing this issues. Can you please let > me know if i am missing anything? > > > Regards, > Sachin Janani > > On Tue, Jun 19, 2018 at 5:38 PM, Sahil Takiar <takiar.sa...@gmail.com> > wrote: > >> I updated the doc to reflect that Hive 3.0.0 works with Spark 2.3.0. What >> issues are you seeing? >> >> On Tue, Jun 19, 2018 at 7:03 AM, Sachin janani < >> sachin.janani...@gmail.com> wrote: >> >>> This is the same link which I followed. As per this link for >>> spark-2.3.0 we need to use hive master instead of hive 3.0.0. Also we >>> need to custom build spark without hive dependencies but after trying >>> all this it shows some compatibility issues. >>> >>> >>> Regards, >>> Sachin Janani >>> >>> On Tue, Jun 19, 2018 at 5:02 PM, Sahil Takiar <takiar.sa...@gmail.com> >>> wrote: >>> > Yes, Hive 3.0.0 works with Spark 2.3.0 - this section of the wiki has >>> > details on which Hive releases support which Spark versions. >>> > >>> > On Tue, Jun 19, 2018 at 5:59 AM, Sachin janani < >>> sachin.janani...@gmail.com> >>> > wrote: >>> >> >>> >> Hi, >>> >> I am trying to run hive on spark by following the steps mentioned >>> >> here- >>> >> https://cwiki.apache.org/confluence/display/Hive/Hive+on+Spa >>> rk%3A+Getting+Started >>> >> , but getting many compatibility issues like NoSuchMethodError, >>> >> NoSuchFieldException etc. So just need to know if it works and whether >>> >> someone tried it out, >>> >> >>> >> >>> >> Thanks and Regards, >>> >> -- >>> >> Sachin Janani >>> > >>> > >>> > >>> > >>> > -- >>> > Sahil Takiar >>> > Software Engineer >>> > takiar.sa...@gmail.com | (510) 673-0309 >>> >>> >>> >>> -- >>> Sachin Janani >>> >> >> >> >> -- >> Sahil Takiar >> Software Engineer >> takiar.sa...@gmail.com | (510) 673-0309 >> > > > > -- > *Sachin Janani* > > -- Sahil Takiar Software Engineer takiar.sa...@gmail.com | (510) 673-0309