Hi ,

You mean hive-site.xml file right ?,I did placed the hive-site.xml in spark 
conf but not sure how spark certain date functions like interval is still 
working .
Hive 0.14 don't have interval function but how spark is managing to do that ?
Does spark has its own date functions ? I am using spark-sql shell for your 
information.

Can I set hive context.sql in spark-Sql shell ? As we do in traditional spark 
Scala application.

Thanks
Sri

Sent from my iPhone

> On 21 May 2016, at 02:24, Mich Talebzadeh <mich.talebza...@gmail.com> wrote:
> 
> Sou want to use hive version 0.14 when using Spark 1.6?
> 
> Go to directory $SPARK_HOME/conf and create a softlink to hive-core.xml file
> 
> cd $SPARK_HOME
> hduser@rhes564: /usr/lib/spark-1.6.1-bin-hadoop2.6> cd conf
> hduser@rhes564: /usr/lib/spark-1.6.1-bin-hadoop2.6/conf> ls -ltr
> 
> lrwxrwxrwx  1 hduser hadoop   32 May  3 17:48 hive-site.xml -> 
> /usr/lib/hive/conf/hive-site.xml
> -
> 
> You can see the softlink in mine. Just create one as below
> 
> ln -s /usr/lib/hive/conf/hive-site.xml hive-site.xml
> 
> 
> That should work
> 
> HTH
> 
> Dr Mich Talebzadeh
>  
> LinkedIn  
> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
>  
> http://talebzadehmich.wordpress.com
>  
> 
>> On 21 May 2016 at 00:57, kali.tumm...@gmail.com <kali.tumm...@gmail.com> 
>> wrote:
>> Hi All ,
>> 
>> Is there a way to ask spark and spark-sql to use Hive 0.14 version instead
>> of inbuilt hive 1.2.1.
>> 
>> I am testing spark-sql locally by downloading spark 1.6 from internet , I
>> want to execute my hive queries in spark sql using hive version 0.14 can I
>> go back to previous version just for a simple test.
>> 
>> Please share out the steps involved.
>> 
>> 
>> Thanks
>> Sri
>> 
>> 
>> 
>> --
>> View this message in context: 
>> http://apache-spark-user-list.1001560.n3.nabble.com/set-spark-1-6-with-Hive-0-14-tp26989.html
>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>> 
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
> 

Reply via email to