I was able to figure out where the problem is exactly. It's spark. because when I start the hiveserver2 manually and run query it work fine. but when I try to access the hive through spark's thrift port it does not work. throws the above mentioned error.
Please help me to fix this. On Sun, Dec 27, 2015 at 11:15 AM, Dasun Hegoda <dasunheg...@gmail.com> wrote: > Yes, didn't work for me > > > On Sun, Dec 27, 2015 at 10:56 AM, Ted Yu <yuzhih...@gmail.com> wrote: > >> Have you seen this ? >> >> >> http://stackoverflow.com/questions/30705576/python-cannot-connect-hiveserver2 >> >> On Sat, Dec 26, 2015 at 9:09 PM, Dasun Hegoda <dasunheg...@gmail.com> >> wrote: >> >>> I'm running apache-hive-1.2.1-bin and spark-1.5.1-bin-hadoop2.6. spark >>> as the hive engine. When I try to connect through JasperStudio using thrift >>> port I get below error. I'm running ubuntu 14.04. >>> >>> 15/12/26 23:36:20 ERROR server.TThreadPoolServer: Error occurred >>> during processing of message. >>> java.lang.RuntimeException: >>> org.apache.thrift.transport.TSaslTransportException: No data or no sasl >>> data in the stream >>> at >>> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:219) >>> at >>> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:268) >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>> at java.lang.Thread.run(Thread.java:745) >>> Caused by: org.apache.thrift.transport.TSaslTransportException: No >>> data or no sasl data in the stream >>> at >>> org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:328) >>> at >>> org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java:41) >>> at >>> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:216) >>> ... 4 more >>> 15/12/26 23:36:20 INFO thrift.ThriftCLIService: Client protocol >>> version: HIVE_CLI_SERVICE_PROTOCOL_V5 >>> 15/12/26 23:36:20 INFO session.SessionState: Created local >>> directory: /tmp/c670ff55-01bb-4f6f-a375-d22a13c44eaf_resources >>> 15/12/26 23:36:20 INFO session.SessionState: Created HDFS directory: >>> /tmp/hive/anonymous/c670ff55-01bb-4f6f-a375-d22a13c44eaf >>> 15/12/26 23:36:20 INFO session.SessionState: Created local >>> directory: /tmp/hduser/c670ff55-01bb-4f6f-a375-d22a13c44eaf >>> 15/12/26 23:36:20 INFO session.SessionState: Created HDFS directory: >>> /tmp/hive/anonymous/c670ff55-01bb-4f6f-a375-d22a13c44eaf/_tmp_space.db >>> 15/12/26 23:36:20 INFO thriftserver.SparkExecuteStatementOperation: >>> Running query 'use default' with d842cd88-2fda-42b2-b943-468017e95f37 >>> 15/12/26 23:36:20 INFO parse.ParseDriver: Parsing command: use >>> default >>> 15/12/26 23:36:20 INFO parse.ParseDriver: Parse Completed >>> 15/12/26 23:36:20 INFO log.PerfLogger: <PERFLOG method=Driver.run >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO log.PerfLogger: <PERFLOG method=TimeToSubmit >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO log.PerfLogger: <PERFLOG method=compile >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO log.PerfLogger: <PERFLOG method=parse >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO parse.ParseDriver: Parsing command: use >>> default >>> 15/12/26 23:36:20 INFO parse.ParseDriver: Parse Completed >>> 15/12/26 23:36:20 INFO log.PerfLogger: </PERFLOG method=parse >>> start=1451190980590 end=1451190980591 duration=1 >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO log.PerfLogger: <PERFLOG >>> method=semanticAnalyze from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO metastore.HiveMetaStore: 2: get_database: >>> default >>> 15/12/26 23:36:20 INFO HiveMetaStore.audit: ugi=hduser >>> ip=unknown-ip-addr cmd=get_database: default >>> 15/12/26 23:36:20 INFO metastore.HiveMetaStore: 2: Opening raw store >>> with implemenation class:org.apache.hadoop.hive.metastore.ObjectStore >>> 15/12/26 23:36:20 INFO metastore.ObjectStore: ObjectStore, >>> initialize called >>> 15/12/26 23:36:20 INFO DataNucleus.Query: Reading in results for >>> query "org.datanucleus.store.rdbms.query.SQLQuery@0" since the >>> connection used is closing >>> 15/12/26 23:36:20 INFO metastore.MetaStoreDirectSql: Using direct >>> SQL, underlying DB is DERBY >>> 15/12/26 23:36:20 INFO metastore.ObjectStore: Initialized ObjectStore >>> 15/12/26 23:36:20 INFO ql.Driver: Semantic Analysis Completed >>> 15/12/26 23:36:20 INFO log.PerfLogger: </PERFLOG >>> method=semanticAnalyze start=1451190980592 end=1451190980620 duration=28 >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO ql.Driver: Returning Hive schema: >>> Schema(fieldSchemas:null, properties:null) >>> 15/12/26 23:36:20 INFO log.PerfLogger: </PERFLOG method=compile >>> start=1451190980588 end=1451190980621 duration=33 >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO ql.Driver: Concurrency mode is disabled, not >>> creating a lock manager >>> 15/12/26 23:36:20 INFO log.PerfLogger: <PERFLOG >>> method=Driver.execute from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO ql.Driver: Starting >>> command(queryId=hduser_20151226233620_6bc633ef-5c6f-49e4-9300-f79fdf0c357b): >>> use default >>> 15/12/26 23:36:20 INFO log.PerfLogger: </PERFLOG method=TimeToSubmit >>> start=1451190980588 end=1451190980622 duration=34 >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO log.PerfLogger: <PERFLOG method=runTasks >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO log.PerfLogger: <PERFLOG >>> method=task.DDL.Stage-0 from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO ql.Driver: Starting task [Stage-0:DDL] in >>> serial mode >>> 15/12/26 23:36:20 INFO metastore.HiveMetaStore: 2: get_database: >>> default >>> 15/12/26 23:36:20 INFO HiveMetaStore.audit: ugi=hduser >>> ip=unknown-ip-addr cmd=get_database: default >>> 15/12/26 23:36:20 INFO metastore.HiveMetaStore: 2: get_database: >>> default >>> 15/12/26 23:36:20 INFO HiveMetaStore.audit: ugi=hduser >>> ip=unknown-ip-addr cmd=get_database: default >>> 15/12/26 23:36:20 INFO log.PerfLogger: </PERFLOG method=runTasks >>> start=1451190980622 end=1451190980637 duration=15 >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO log.PerfLogger: </PERFLOG >>> method=Driver.execute start=1451190980621 end=1451190980637 duration=16 >>> from=org.apache.hadoop.hive.ql.Driver> >>> OK >>> 15/12/26 23:36:20 INFO ql.Driver: OK >>> 15/12/26 23:36:20 INFO log.PerfLogger: <PERFLOG method=releaseLocks >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO log.PerfLogger: </PERFLOG method=releaseLocks >>> start=1451190980639 end=1451190980639 duration=0 >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO log.PerfLogger: </PERFLOG method=Driver.run >>> start=1451190980587 end=1451190980639 duration=52 >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:36:20 INFO thriftserver.SparkExecuteStatementOperation: >>> Result Schema: List(result#28) >>> 15/12/26 23:36:20 INFO thriftserver.SparkExecuteStatementOperation: >>> Running query 'SELECT * FROM service' with >>> 37916038-9856-43eb-8b73-920f9faf738f >>> 15/12/26 23:36:20 INFO parse.ParseDriver: Parsing command: SELECT * >>> FROM service >>> 15/12/26 23:36:20 INFO parse.ParseDriver: Parse Completed >>> 15/12/26 23:36:20 INFO metastore.HiveMetaStore: 2: get_table : >>> db=default tbl=service >>> 15/12/26 23:36:20 INFO HiveMetaStore.audit: ugi=hduser >>> ip=unknown-ip-addr cmd=get_table : db=default tbl=service >>> 15/12/26 23:36:20 ERROR thriftserver.SparkExecuteStatementOperation: >>> Error executing query, currentState RUNNING, >>> org.apache.spark.sql.AnalysisException: no such table service; line >>> 1 pos 14 >>> at >>> org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42) >>> at >>> org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.getTable(Analyzer.scala:260) >>> at >>> org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$7.applyOrElse(Analyzer.scala:268) >>> at >>> org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$$anonfun$apply$7.applyOrElse(Analyzer.scala:264) >>> at >>> org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:57) >>> at >>> org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$resolveOperators$1.apply(LogicalPlan.scala:57) >>> at >>> org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:51) >>> at >>> org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:56) >>> at >>> org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:54) >>> at >>> org.apache.spark.sql.catalyst.plans.logical.LogicalPlan$$anonfun$1.apply(LogicalPlan.scala:54) >>> at >>> org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$4.apply(TreeNode.scala:249) >>> at scala.collection.Iterator$$anon$11.next(Iterator.scala:328) >>> at scala.collection.Iterator$class.foreach(Iterator.scala:727) >>> at scala.collection.AbstractIterator.foreach(Iterator.scala:1157) >>> at >>> scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48) >>> at >>> scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:103) >>> at >>> scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:47) >>> at scala.collection.TraversableOnce$class.to >>> (TraversableOnce.scala:273) >>> at scala.collection.AbstractIterator.to(Iterator.scala:1157) >>> at >>> scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:265) >>> at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1157) >>> at >>> scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:252) >>> at scala.collection.AbstractIterator.toArray(Iterator.scala:1157) >>> at >>> org.apache.spark.sql.catalyst.trees.TreeNode.transformChildren(TreeNode.scala:279) >>> at >>> org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.resolveOperators(LogicalPlan.scala:54) >>> at >>> org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:264) >>> at >>> org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveRelations$.apply(Analyzer.scala:254) >>> at >>> org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:83) >>> at >>> org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1$$anonfun$apply$1.apply(RuleExecutor.scala:80) >>> at >>> scala.collection.LinearSeqOptimized$class.foldLeft(LinearSeqOptimized.scala:111) >>> at scala.collection.immutable.List.foldLeft(List.scala:84) >>> at >>> org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:80) >>> at >>> org.apache.spark.sql.catalyst.rules.RuleExecutor$$anonfun$execute$1.apply(RuleExecutor.scala:72) >>> at scala.collection.immutable.List.foreach(List.scala:318) >>> at >>> org.apache.spark.sql.catalyst.rules.RuleExecutor.execute(RuleExecutor.scala:72) >>> at >>> org.apache.spark.sql.SQLContext$QueryExecution.analyzed$lzycompute(SQLContext.scala:916) >>> at >>> org.apache.spark.sql.SQLContext$QueryExecution.analyzed(SQLContext.scala:916) >>> at >>> org.apache.spark.sql.SQLContext$QueryExecution.assertAnalyzed(SQLContext.scala:914) >>> at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:132) >>> at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:51) >>> at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:725) >>> at >>> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.runInternal(SparkExecuteStatementOperation.scala:224) >>> at >>> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.run(SparkExecuteStatementOperation.scala:144) >>> at >>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:388) >>> at >>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:369) >>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>> at >>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>> at >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>> at java.lang.reflect.Method.invoke(Method.java:606) >>> at >>> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78) >>> at >>> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36) >>> at >>> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63) >>> at java.security.AccessController.doPrivileged(Native Method) >>> at javax.security.auth.Subject.doAs(Subject.java:415) >>> at >>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) >>> at >>> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59) >>> at com.sun.proxy.$Proxy47.executeStatement(Unknown Source) >>> at >>> org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:261) >>> at >>> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:486) >>> at >>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1313) >>> at >>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1298) >>> at >>> org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) >>> at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39) >>> at >>> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56) >>> at >>> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:285) >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>> at java.lang.Thread.run(Thread.java:745) >>> 15/12/26 23:36:20 WARN thrift.ThriftCLIService: Error executing >>> statement: >>> org.apache.hive.service.cli.HiveSQLException: >>> org.apache.spark.sql.AnalysisException: no such table service; line 1 pos 14 >>> at >>> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.runInternal(SparkExecuteStatementOperation.scala:259) >>> at >>> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.run(SparkExecuteStatementOperation.scala:144) >>> at >>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatementInternal(HiveSessionImpl.java:388) >>> at >>> org.apache.hive.service.cli.session.HiveSessionImpl.executeStatement(HiveSessionImpl.java:369) >>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>> at >>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) >>> at >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>> at java.lang.reflect.Method.invoke(Method.java:606) >>> at >>> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:78) >>> at >>> org.apache.hive.service.cli.session.HiveSessionProxy.access$000(HiveSessionProxy.java:36) >>> at >>> org.apache.hive.service.cli.session.HiveSessionProxy$1.run(HiveSessionProxy.java:63) >>> at java.security.AccessController.doPrivileged(Native Method) >>> at javax.security.auth.Subject.doAs(Subject.java:415) >>> at >>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) >>> at >>> org.apache.hive.service.cli.session.HiveSessionProxy.invoke(HiveSessionProxy.java:59) >>> at com.sun.proxy.$Proxy47.executeStatement(Unknown Source) >>> at >>> org.apache.hive.service.cli.CLIService.executeStatement(CLIService.java:261) >>> at >>> org.apache.hive.service.cli.thrift.ThriftCLIService.ExecuteStatement(ThriftCLIService.java:486) >>> at >>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1313) >>> at >>> org.apache.hive.service.cli.thrift.TCLIService$Processor$ExecuteStatement.getResult(TCLIService.java:1298) >>> at >>> org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39) >>> at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39) >>> at >>> org.apache.hive.service.auth.TSetIpAddressProcessor.process(TSetIpAddressProcessor.java:56) >>> at >>> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:285) >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>> at java.lang.Thread.run(Thread.java:745) >>> 15/12/26 23:41:08 ERROR server.TThreadPoolServer: Error occurred >>> during processing of message. >>> java.lang.RuntimeException: >>> org.apache.thrift.transport.TSaslTransportException: No data or no sasl >>> data in the stream >>> at >>> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:219) >>> at >>> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:268) >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>> at java.lang.Thread.run(Thread.java:745) >>> Caused by: org.apache.thrift.transport.TSaslTransportException: No >>> data or no sasl data in the stream >>> at >>> org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:328) >>> at >>> org.apache.thrift.transport.TSaslServerTransport.open(TSaslServerTransport.java:41) >>> at >>> org.apache.thrift.transport.TSaslServerTransport$Factory.getTransport(TSaslServerTransport.java:216) >>> ... 4 more >>> 15/12/26 23:41:08 INFO thrift.ThriftCLIService: Client protocol >>> version: HIVE_CLI_SERVICE_PROTOCOL_V5 >>> 15/12/26 23:41:08 INFO session.SessionState: Created local >>> directory: /tmp/aa7ce472-0284-4469-823b-748ef786ab73_resources >>> 15/12/26 23:41:08 INFO session.SessionState: Created HDFS directory: >>> /tmp/hive/anonymous/aa7ce472-0284-4469-823b-748ef786ab73 >>> 15/12/26 23:41:08 INFO session.SessionState: Created local >>> directory: /tmp/hduser/aa7ce472-0284-4469-823b-748ef786ab73 >>> 15/12/26 23:41:08 INFO session.SessionState: Created HDFS directory: >>> /tmp/hive/anonymous/aa7ce472-0284-4469-823b-748ef786ab73/_tmp_space.db >>> 15/12/26 23:41:08 INFO thriftserver.SparkExecuteStatementOperation: >>> Running query 'use default' with 6a274f01-2a83-44b9-b970-2154792af7a2 >>> 15/12/26 23:41:08 INFO parse.ParseDriver: Parsing command: use >>> default >>> 15/12/26 23:41:08 INFO parse.ParseDriver: Parse Completed >>> 15/12/26 23:41:08 INFO log.PerfLogger: <PERFLOG method=Driver.run >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO log.PerfLogger: <PERFLOG method=TimeToSubmit >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO log.PerfLogger: <PERFLOG method=compile >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO log.PerfLogger: <PERFLOG method=parse >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO parse.ParseDriver: Parsing command: use >>> default >>> 15/12/26 23:41:08 INFO parse.ParseDriver: Parse Completed >>> 15/12/26 23:41:08 INFO log.PerfLogger: </PERFLOG method=parse >>> start=1451191268389 end=1451191268390 duration=1 >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO log.PerfLogger: <PERFLOG >>> method=semanticAnalyze from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO metastore.HiveMetaStore: 2: get_database: >>> default >>> 15/12/26 23:41:08 INFO HiveMetaStore.audit: ugi=hduser >>> ip=unknown-ip-addr cmd=get_database: default >>> 15/12/26 23:41:08 INFO ql.Driver: Semantic Analysis Completed >>> 15/12/26 23:41:08 INFO log.PerfLogger: </PERFLOG >>> method=semanticAnalyze start=1451191268390 end=1451191268397 duration=7 >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO ql.Driver: Returning Hive schema: >>> Schema(fieldSchemas:null, properties:null) >>> 15/12/26 23:41:08 INFO log.PerfLogger: </PERFLOG method=compile >>> start=1451191268387 end=1451191268398 duration=11 >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO ql.Driver: Concurrency mode is disabled, not >>> creating a lock manager >>> 15/12/26 23:41:08 INFO log.PerfLogger: <PERFLOG >>> method=Driver.execute from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO ql.Driver: Starting >>> command(queryId=hduser_20151226234108_27b4ad3d-0f88-4a81-83f6-eaf0ef49cd22): >>> use default >>> 15/12/26 23:41:08 INFO log.PerfLogger: </PERFLOG method=TimeToSubmit >>> start=1451191268387 end=1451191268399 duration=12 >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO log.PerfLogger: <PERFLOG method=runTasks >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO log.PerfLogger: <PERFLOG >>> method=task.DDL.Stage-0 from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO ql.Driver: Starting task [Stage-0:DDL] in >>> serial mode >>> 15/12/26 23:41:08 INFO metastore.HiveMetaStore: 2: get_database: >>> default >>> 15/12/26 23:41:08 INFO HiveMetaStore.audit: ugi=hduser >>> ip=unknown-ip-addr cmd=get_database: default >>> 15/12/26 23:41:08 INFO metastore.HiveMetaStore: 2: get_database: >>> default >>> 15/12/26 23:41:08 INFO HiveMetaStore.audit: ugi=hduser >>> ip=unknown-ip-addr cmd=get_database: default >>> 15/12/26 23:41:08 INFO log.PerfLogger: </PERFLOG method=runTasks >>> start=1451191268399 end=1451191268412 duration=13 >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO log.PerfLogger: </PERFLOG >>> method=Driver.execute start=1451191268398 end=1451191268412 duration=14 >>> from=org.apache.hadoop.hive.ql.Driver> >>> OK >>> 15/12/26 23:41:08 INFO ql.Driver: OK >>> 15/12/26 23:41:08 INFO log.PerfLogger: <PERFLOG method=releaseLocks >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO log.PerfLogger: </PERFLOG method=releaseLocks >>> start=1451191268413 end=1451191268413 duration=0 >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO log.PerfLogger: </PERFLOG method=Driver.run >>> start=1451191268387 end=1451191268413 duration=26 >>> from=org.apache.hadoop.hive.ql.Driver> >>> 15/12/26 23:41:08 INFO thriftserver.SparkExecuteStatementOperation: >>> Result Schema: List(result#43) >>> >>> >>> Below is the apache-hive-1.2.1-bin/conf/hive-site.xml >>> >>> >>> <description> >>> The cluster manager to connect to >>> </description> >>> </property> >>> >>> <property> >>> <name>spark.serializer</name> >>> <value>org.apache.spark.serializer.KryoSerializer</value> >>> <description> >>> Class to use for serializing objects that will be sent >>> over the network >>> </description> >>> </property> >>> >>> >>> >>> <property> >>> <name>hive.server2.authentication</name> >>> <value>NONE</value> >>> <description> >>> Client authentication types. >>> NONE: no authentication check >>> LDAP: LDAP/AD based authentication >>> KERBEROS: Kerberos/GSSAPI authentication >>> CUSTOM: Custom authentication provider >>> (Use with property >>> hive.server2.custom.authentication.class) >>> </description> >>> </property> >>> >>> <property> >>> <name>hive.metastore.sasl.enabled</name> >>> <value>false</value> >>> <description>If true, the metastore thrift interface will be >>> secured with SASL. Clients must authenticate with Kerberos.</description> >>> </property> >>> >>> <!--Hive server --> >>> <property> >>> <name>hive.server2.thrift.port</name> >>> <value>10000</value> >>> <description>Port number of HiveServer2 Thrift interface. >>> Can be overridden by setting >>> $HIVE_SERVER2_THRIFT_PORT</description> >>> </property> >>> >>> <property> >>> <name>hive.server2.thrift.bind.host</name> >>> <value>192.168.7.87</value> >>> <description>Bind host on which to run the HiveServer2 Thrift >>> interface. >>> Can be overridden by setting >>> $HIVE_SERVER2_THRIFT_BIND_HOST</description> >>> </property> >>> >>> How can I fix this? >>> >>> -- >>> Regards, >>> Dasun Hegoda, Software Engineer >>> www.dasunhegoda.com | dasunheg...@gmail.com >>> >> >> > > > -- > Regards, > Dasun Hegoda, Software Engineer > www.dasunhegoda.com | dasunheg...@gmail.com > -- Regards, Dasun Hegoda, Software Engineer www.dasunhegoda.com | dasunheg...@gmail.com