I put it in both and I am still getting the same error.
On Thu, May 2, 2013 at 1:55 PM, kulkarni.swar...@gmail.com < kulkarni.swar...@gmail.com> wrote: > But that would still use the HADOOP_CLASSPATH right? > > > On Thu, May 2, 2013 at 12:52 PM, Cyril Bogus <cyrilbo...@gmail.com> wrote: > >> But right now I am just trying to run it as standalone (no need to check >> for the packages I assume) with hadoop's hdfs in order to do some indexing >> on data already present in the hdfs >> >> >> On Thu, May 2, 2013 at 1:50 PM, Cyril Bogus <cyrilbo...@gmail.com> wrote: >> >>> Actually two the one from hadoop (which is the same as from the one in >>> the hive package) and the one from mahout 0.7 which is newer antlr 3.2 >>> >>> >>> On Thu, May 2, 2013 at 1:47 PM, kulkarni.swar...@gmail.com < >>> kulkarni.swar...@gmail.com> wrote: >>> >>>> Do you have a different version of antlr jar in your classpath other >>>> than the one packaged with hive? >>>> >>>> >>>> On Thu, May 2, 2013 at 12:38 PM, Cyril Bogus <cyrilbo...@gmail.com>wrote: >>>> >>>>> I am using the default setup for the hive-site.xml so the meta store >>>>> is in /user/hive/warehouse in the hdfs (Which I have setup as specified >>>>> under Getting started in the hive website. >>>>> >>>>> Here is the output from the command. >>>>> >>>>> >>>>> >>>>> hive -hiveconf hive.root.logger=INFO,console -e "show databases" >>>>> WARNING: org.apache.hadoop.metrics.jvm.EventCounter is deprecated. >>>>> Please use org.apache.hadoop.log.metrics.EventCounter in all the >>>>> log4j.properties files. >>>>> Logging initialized using configuration in >>>>> jar:file:/home/hive/lib/hive-common-0.10.0.jar!/hive-log4j.properties >>>>> 13/05/02 13:37:09 INFO SessionState: Logging initialized using >>>>> configuration in >>>>> jar:file:/home/hive/lib/hive-common-0.10.0.jar!/hive-log4j.properties >>>>> Hive history >>>>> file=/tmp/cyrille/hive_job_log_cyrille_201305021337_431171577.txt >>>>> 13/05/02 13:37:09 INFO exec.HiveHistory: Hive history >>>>> file=/tmp/cyrille/hive_job_log_cyrille_201305021337_431171577.txt >>>>> 13/05/02 13:37:09 INFO ql.Driver: <PERFLOG method=Driver.run> >>>>> 13/05/02 13:37:09 INFO ql.Driver: <PERFLOG method=TimeToSubmit> >>>>> 13/05/02 13:37:09 INFO ql.Driver: <PERFLOG method=compile> >>>>> 13/05/02 13:37:09 INFO parse.ParseDriver: Parsing command: show >>>>> databases >>>>> 13/05/02 13:37:09 INFO ql.Driver: </PERFLOG method=compile >>>>> start=1367516229487 end=1367516229681 duration=194> >>>>> >>>>> Exception in thread "main" java.lang.NoSuchFieldError: type >>>>> at >>>>> org.apache.hadoop.hive.ql.parse.HiveLexer.mKW_SHOW(HiveLexer.java:1305) >>>>> at >>>>> org.apache.hadoop.hive.ql.parse.HiveLexer.mTokens(HiveLexer.java:6439) >>>>> at org.antlr.runtime.Lexer.nextToken(Lexer.java:84) >>>>> at >>>>> org.antlr.runtime.CommonTokenStream.fillBuffer(CommonTokenStream.java:95) >>>>> at org.antlr.runtime.CommonTokenStream.LT >>>>> (CommonTokenStream.java:238) >>>>> at >>>>> org.apache.hadoop.hive.ql.parse.HiveParser.statement(HiveParser.java:573) >>>>> at >>>>> org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:439) >>>>> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:416) >>>>> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:335) >>>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:893) >>>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:259) >>>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:216) >>>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:412) >>>>> at >>>>> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:347) >>>>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:706) >>>>> >>>>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:613) >>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>> at >>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>>> at >>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>> at java.lang.reflect.Method.invoke(Method.java:616) >>>>> at org.apache.hadoop.util.RunJar.main(RunJar.java:156) >>>>> >>>>> >>>>> >>>>> On Thu, May 2, 2013 at 1:25 PM, Sanjay Subramanian < >>>>> sanjay.subraman...@wizecommerce.com> wrote: >>>>> >>>>>> Can u share your hive-site.xml ? What meta store r u using ? >>>>>> Also try this to get additional debug messages that u can use to >>>>>> analyze the problem >>>>>> >>>>>> From your linux command prompt run the following and tell us what u >>>>>> see. Also hive-site.xml please >>>>>> >>>>>> /path/to/hive -hiveconf hive.root.logger=INFO,console -e "show >>>>>> databases" >>>>>> >>>>>> From: Cyril Bogus <cyrilbo...@gmail.com> >>>>>> Reply-To: "user@hive.apache.org" <user@hive.apache.org> >>>>>> Date: Thursday, May 2, 2013 10:19 AM >>>>>> To: "user@hive.apache.org" <user@hive.apache.org> >>>>>> Subject: Getting Started >>>>>> >>>>>> Hi, >>>>>> >>>>>> I am currently running hadoop 1.0.4 and hive 0.10.0 >>>>>> also I have HADOOP_HOME set to /home/hadoop and HIVE_HOME to >>>>>> /home/hive along with JAVA_HOME also to the right location. >>>>>> and I would like to run the hive command line but I keep getting the >>>>>> following error when I try to run a simple query like show databases; >>>>>> >>>>>> >>>>>> hive >>>>>> WARNING: org.apache.hadoop.metrics.jvm.EventCounter is deprecated. >>>>>> Please use org.apache.hadoop.log.metrics.EventCounter in all the >>>>>> log4j.properties files. >>>>>> Logging initialized using configuration in >>>>>> jar:file:/home/hive/lib/hive-common-0.10.0.jar!/hive-log4j.properties >>>>>> Hive history >>>>>> file=/tmp/cyrille/hive_job_log_cyrille_201305021317_1253522258.txt >>>>>> hive> show databases; >>>>>> Exception in thread "main" java.lang.NoSuchFieldError: type >>>>>> at >>>>>> org.apache.hadoop.hive.ql.parse.HiveLexer.mKW_SHOW(HiveLexer.java:1305) >>>>>> at >>>>>> org.apache.hadoop.hive.ql.parse.HiveLexer.mTokens(HiveLexer.java:6439) >>>>>> at org.antlr.runtime.Lexer.nextToken(Lexer.java:84) >>>>>> at >>>>>> org.antlr.runtime.CommonTokenStream.fillBuffer(CommonTokenStream.java:95) >>>>>> at org.antlr.runtime.CommonTokenStream.LT >>>>>> (CommonTokenStream.java:238) >>>>>> at >>>>>> org.apache.hadoop.hive.ql.parse.HiveParser.statement(HiveParser.java:573) >>>>>> at >>>>>> org.apache.hadoop.hive.ql.parse.ParseDriver.parse(ParseDriver.java:439) >>>>>> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:416) >>>>>> at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:335) >>>>>> at org.apache.hadoop.hive.ql.Driver.run(Driver.java:893) >>>>>> at >>>>>> org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:259) >>>>>> at >>>>>> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:216) >>>>>> at >>>>>> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:412) >>>>>> at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:755) >>>>>> at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:613) >>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>>>> at >>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>>>>> at >>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>>> at java.lang.reflect.Method.invoke(Method.java:616) >>>>>> at org.apache.hadoop.util.RunJar.main(RunJar.java:156) >>>>>> >>>>>> Thank you in advance for the input >>>>>> >>>>>> CONFIDENTIALITY NOTICE >>>>>> ====================== >>>>>> This email message and any attachments are for the exclusive use of >>>>>> the intended recipient(s) and may contain confidential and privileged >>>>>> information. Any unauthorized review, use, disclosure or distribution is >>>>>> prohibited. If you are not the intended recipient, please contact the >>>>>> sender by reply email and destroy all copies of the original message >>>>>> along >>>>>> with any attachments, from your computer system. If you are the intended >>>>>> recipient, please be advised that the content of this message is subject >>>>>> to >>>>>> access, review and disclosure by the sender's Email System Administrator. >>>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> Swarnim >>>> >>> >>> >> > > > -- > Swarnim >