Unfortunately, I do not have it, as it is 3rd party code :(

But essentially I am trying to overwrite data to a hive table from a source

On Tue, May 31, 2016 at 4:01 PM, Mich Talebzadeh <mich.talebza...@gmail.com>
wrote:

> ok what is the exact spark code that is causing the issue.
>
> can you show it in its entirety?
>
> HTH
>
> Dr Mich Talebzadeh
>
>
>
> LinkedIn * 
> https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw
> <https://www.linkedin.com/profile/view?id=AAEAAAAWh2gBxianrbJd6zP6AcPCCdOABUrV8Pw>*
>
>
>
> http://talebzadehmich.wordpress.com
>
>
>
> On 31 May 2016 at 06:31, ayan guha <guha.a...@gmail.com> wrote:
>
>> No there is no semicolon.
>>
>> This is the query:
>>
>> 16/05/31 14:34:29 INFO SparkExecuteStatementOperation: Running query
>> 'DESCRIBE EXTENDED `sds.unhealthy_om_delta`' with
>> e24282a8-43d1-4c3a-a3f3-2645761ed40f
>>
>>
>> On Tue, May 31, 2016 at 3:10 PM, Raju Bairishetti <raju....@gmail.com>
>> wrote:
>>
>>>
>>>
>>> On Tue, May 31, 2016 at 1:02 PM, ayan guha <guha.a...@gmail.com> wrote:
>>>
>>>> Hi
>>>>
>>>> While running spark thrift, we are getting 2 issues.
>>>>
>>>> 1.. 16/05/31 14:36:18 WARN ThriftCLIService: Error executing statement:
>>>> org.apache.hive.service.cli.HiveSQLException:
>>>> org.apache.spark.sql.AnalysisException: Table not found:
>>>> sds.unhealthy_om_delta;
>>>>
>>>
>>> Are you using *;* (semi colon) at the end of query like
>>> *sqlcontext.sql(query;)*?   You should not mention *;* at the end of
>>> query
>>>
>>>         at
>>>> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org
>>>> $apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:
>>>> 246)
>>>>
>>>> Unfortunately, the table exists and I can see it from beeline.
>>>>
>>>> This error is happening from a front end, where the front end service
>>>> is launched by a different user. However, we do not restrict read access to
>>>> anybody.
>>>>
>>>> 2. org.apache.hive.service.cli.HiveSQLException:
>>>> java.lang.RuntimeException: [1.20] failure: end of input expected
>>>>
>>>> SHOW TABLES IN sds LIKE '.*'
>>>>                    ^
>>>>         at
>>>> org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org
>>>> $apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:
>>>> 246)
>>>>
>>>> It seems a pure Hive error, and looks like wrong syntax.. Any
>>>> suggestion what is the correct syntax?
>>>>
>>>> Both issues are coming while running a 3rd party tool (datameer)
>>>> connecting to Spark Thrift Server. Spark Version 1.6 on HDP 2.4.
>>>>
>>>>
>>>> TIA...
>>>>
>>>> --
>>>> Best Regards,
>>>> Ayan Guha
>>>>
>>>
>>>
>>>
>>> --
>>> Thanks,
>>> Raju Bairishetti,
>>>
>>> www.lazada.com
>>>
>>>
>>>
>>
>>
>> --
>> Best Regards,
>> Ayan Guha
>>
>
>


-- 
Best Regards,
Ayan Guha

Reply via email to