Using this URL made it work: jdbc:hive2://myhost.example.com:21050/;auth=noSasl
On Wed, Aug 31, 2016 at 11:13 AM, Abhi Basu <9000r...@gmail.com> wrote: > Except spark-sql is geared more towards developers and our users are > looking for a SQL engine like hive (except faster). :) > > > > On Wed, Aug 31, 2016 at 11:11 AM, Pradeep Reddy < > pradeepreddy.a...@gmail.com> wrote: > >> You could use the JDBC interpreter to setup a new interpreter for Impala. >> That said, Impala is just using your hive metastore and enabling >> proprietary caching for high performance on your hive database tables >> rather than doing map reduce translation of hive queries. Running Spark SQL >> on spark interpreter can be considered as a reasonable alternative to >> running impala queries through jdbc. >> >> http://www.cloudera.com/documentation/archive/impala/2-x/2- >> 1-x/topics/impala_jdbc.html >> >> Thanks, >> Pradeep >> >> >> On Wed, Aug 31, 2016 at 10:45 AM, Abhi Basu <9000r...@gmail.com> wrote: >> >>> How do I setup a connection to impala? Do I need to point to >>> impala-jdbc jar in dependencies? >>> >>> Thanks, >>> >>> Abhi >>> >>> On Wed, Aug 31, 2016 at 10:36 AM, Abhi Basu <9000r...@gmail.com> wrote: >>> >>>> OK, got it. Added the hadoop jar to dependencies and it started working. >>>> >>>> Thanks. >>>> >>>> On Wed, Aug 31, 2016 at 10:24 AM, Abhi Basu <9000r...@gmail.com> wrote: >>>> >>>>> So, path to the jars like /usr/lib/hive/* ? >>>>> >>>>> On Wed, Aug 31, 2016 at 9:53 AM, Jeff Zhang <zjf...@gmail.com> wrote: >>>>> >>>>>> You don't need to copy these jars manually, just specify them in the >>>>>> interpreter setting page. >>>>>> >>>>>> On Wed, Aug 31, 2016 at 9:52 PM, Abhi Basu <9000r...@gmail.com> >>>>>> wrote: >>>>>> >>>>>>> Where do these jars have to be placed? >>>>>>> >>>>>>> I thought copying the hive-site.xml and pointing to hadoop conf >>>>>>> folder in zeppelin conf should be enough (like before). >>>>>>> >>>>>>> Thanks, >>>>>>> >>>>>>> Abhi >>>>>>> >>>>>>> On Tue, Aug 30, 2016 at 6:59 PM, Jeff Zhang <zjf...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> You need add the following 2 dependencies in the interpreter >>>>>>>> setting page. >>>>>>>> >>>>>>>> https://zeppelin.apache.org/docs/0.6.1/interpreter/hive.html >>>>>>>> #dependencies >>>>>>>> >>>>>>>> org.apache.hive:hive-jdbc:0.14.0 >>>>>>>> org.apache.hadoop:hadoop-common:2.6.0 >>>>>>>> >>>>>>>> >>>>>>>> On Wed, Aug 31, 2016 at 2:39 AM, Abhi Basu <9000r...@gmail.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Folks: >>>>>>>>> >>>>>>>>> Seems like a config issue. >>>>>>>>> >>>>>>>>> 1. Copied hive-site.xml into /ZEPP_HOME/conf folder >>>>>>>>> 2. Added following to config file: >>>>>>>>> >>>>>>>>> export JAVA_HOME=/...../...export HADOOP_CONF_DIR=/etc/hadoop/conf >>>>>>>>> >>>>>>>>> >>>>>>>>> I am using Zeppelin after a while, and looks like Hive interpreter >>>>>>>>> is part of JDBC interpreter now. >>>>>>>>> Interpreter properties seem to be set correctly: >>>>>>>>> PropertyValue >>>>>>>>> hive.driver org.apache.hive.jdbc.HiveDriver >>>>>>>>> hive.url jdbc:hive2://localhost:10000 >>>>>>>>> hive.user hiveUser >>>>>>>>> hive.password hivePassword >>>>>>>>> >>>>>>>>> When I run %hive from Zeppelin, I get a hive jdbc driver not found >>>>>>>>> error. How do I fix this? Also, how do I configure for Impala within >>>>>>>>> the >>>>>>>>> JDBC section of interpreters. >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> >>>>>>>>> Abhi >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Abhi Basu >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Best Regards >>>>>>>> >>>>>>>> Jeff Zhang >>>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Abhi Basu >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Best Regards >>>>>> >>>>>> Jeff Zhang >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> Abhi Basu >>>>> >>>> >>>> >>>> >>>> -- >>>> Abhi Basu >>>> >>> >>> >>> >>> -- >>> Abhi Basu >>> >> >> > > > -- > Abhi Basu > -- Abhi Basu