ok what is the exact spark code that is causing the issue. can you show it in its entirety?
HTH Dr Mich Talebzadeh LinkedIn * https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>* http://talebzadehmich.wordpress.com On 31 May 2016 at 06:31, ayan guha <guha.a...@gmail.com> wrote: > No there is no semicolon. > > This is the query: > > 16/05/31 14:34:29 INFO SparkExecuteStatementOperation: Running query > 'DESCRIBE EXTENDED `sds.unhealthy_om_delta`' with > e24282a8-43d1-4c3a-a3f3-2645761ed40f > > > On Tue, May 31, 2016 at 3:10 PM, Raju Bairishetti <raju....@gmail.com> > wrote: > >> >> >> On Tue, May 31, 2016 at 1:02 PM, ayan guha <guha.a...@gmail.com> wrote: >> >>> Hi >>> >>> While running spark thrift, we are getting 2 issues. >>> >>> 1.. 16/05/31 14:36:18 WARN ThriftCLIService: Error executing statement: >>> org.apache.hive.service.cli.HiveSQLException: >>> org.apache.spark.sql.AnalysisException: Table not found: >>> sds.unhealthy_om_delta; >>> >> >> Are you using *;* (semi colon) at the end of query like >> *sqlcontext.sql(query;)*? You should not mention *;* at the end of >> query >> >> at >>> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org >>> $apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala: >>> 246) >>> >>> Unfortunately, the table exists and I can see it from beeline. >>> >>> This error is happening from a front end, where the front end service is >>> launched by a different user. However, we do not restrict read access to >>> anybody. >>> >>> 2. org.apache.hive.service.cli.HiveSQLException: >>> java.lang.RuntimeException: [1.20] failure: end of input expected >>> >>> SHOW TABLES IN sds LIKE '.*' >>> ^ >>> at >>> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org >>> $apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala: >>> 246) >>> >>> It seems a pure Hive error, and looks like wrong syntax.. Any suggestion >>> what is the correct syntax? >>> >>> Both issues are coming while running a 3rd party tool (datameer) >>> connecting to Spark Thrift Server. Spark Version 1.6 on HDP 2.4. >>> >>> >>> TIA... >>> >>> -- >>> Best Regards, >>> Ayan Guha >>> >> >> >> >> -- >> Thanks, >> Raju Bairishetti, >> >> www.lazada.com >> >> >> > > > -- > Best Regards, > Ayan Guha >