ok what is the exact spark code that is causing the issue.

can you show it in its entirety?

HTH

Dr Mich Talebzadeh



LinkedIn * 
https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
<https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*



http://talebzadehmich.wordpress.com



On 31 May 2016 at 06:31, ayan guha <guha.a...@gmail.com> wrote:

> No there is no semicolon.
>
> This is the query:
>
> 16/05/31 14:34:29 INFO SparkExecuteStatementOperation: Running query
> 'DESCRIBE EXTENDED `sds.unhealthy_om_delta`' with
> e24282a8-43d1-4c3a-a3f3-2645761ed40f
>
>
> On Tue, May 31, 2016 at 3:10 PM, Raju Bairishetti <raju....@gmail.com>
> wrote:
>
>>
>>
>> On Tue, May 31, 2016 at 1:02 PM, ayan guha <guha.a...@gmail.com> wrote:
>>
>>> Hi
>>>
>>> While running spark thrift, we are getting 2 issues.
>>>
>>> 1.. 16/05/31 14:36:18 WARN ThriftCLIService: Error executing statement:
>>> org.apache.hive.service.cli.HiveSQLException:
>>> org.apache.spark.sql.AnalysisException: Table not found:
>>> sds.unhealthy_om_delta;
>>>
>>
>> Are you using *;* (semi colon) at the end of query like
>> *sqlcontext.sql(query;)*?   You should not mention *;* at the end of
>> query
>>
>>         at
>>> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org
>>> $apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:
>>> 246)
>>>
>>> Unfortunately, the table exists and I can see it from beeline.
>>>
>>> This error is happening from a front end, where the front end service is
>>> launched by a different user. However, we do not restrict read access to
>>> anybody.
>>>
>>> 2. org.apache.hive.service.cli.HiveSQLException:
>>> java.lang.RuntimeException: [1.20] failure: end of input expected
>>>
>>> SHOW TABLES IN sds LIKE '.*'
>>>                    ^
>>>         at
>>> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org
>>> $apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:
>>> 246)
>>>
>>> It seems a pure Hive error, and looks like wrong syntax.. Any suggestion
>>> what is the correct syntax?
>>>
>>> Both issues are coming while running a 3rd party tool (datameer)
>>> connecting to Spark Thrift Server. Spark Version 1.6 on HDP 2.4.
>>>
>>>
>>> TIA...
>>>
>>> --
>>> Best Regards,
>>> Ayan Guha
>>>
>>
>>
>>
>> --
>> Thanks,
>> Raju Bairishetti,
>>
>> www.lazada.com
>>
>>
>>
>
>
> --
> Best Regards,
> Ayan Guha
>

Reply via email to