Still any inputs on that one ?
Nobody met it ?

Regards,


Loïc

Loïc CHANEL
System Big Data engineer
MS&T - Worldline Analytics Platform - Worldline (Villeurbanne, France)

2017-08-07 18:26 GMT+02:00 Loïc Chanel <loic.cha...@telecomnancy.net>:

> Hi,
>
> As I tried to run some queries with JSON SerDe from Spark SQL client, I
> encountered that error :
>
> 17/08/07 18:20:40 ERROR SparkSQLDriver: Failed in [create external table
> client_project.test_ext(DocVersion string, DriverID string) row format
> serde 'org.apache.hive.hcatalog.data.JsonSerDe' WITH SERDEPROPERTIES
> ("ignore.malformed.json" = "true") location '/etl/client/ct/aftermarket/
> processing/proj/envt=M5']
> org.apache.spark.sql.execution.QueryExecutionException: FAILED: Execution
> Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Cannot
> validate serde: org.apache.hive.hcatalog.data.JsonSerDe
>         at org.apache.spark.sql.hive.client.ClientWrapper$$anonfun$
> runHive$1.apply(ClientWrapper.scala:455)
>         at org.apache.spark.sql.hive.client.ClientWrapper$$anonfun$
> runHive$1.apply(ClientWrapper.scala:440)
>         at org.apache.spark.sql.hive.client.ClientWrapper$$anonfun$
> withHiveState$1.apply(ClientWrapper.scala:278)
>         at org.apache.spark.sql.hive.client.ClientWrapper.
> retryLocked(ClientWrapper.scala:233)
>         at org.apache.spark.sql.hive.client.ClientWrapper.
> withHiveState(ClientWrapper.scala:270)
>         at org.apache.spark.sql.hive.client.ClientWrapper.runHive(
> ClientWrapper.scala:440)
>         at org.apache.spark.sql.hive.client.ClientWrapper.
> runSqlHive(ClientWrapper.scala:430)
>         at org.apache.spark.sql.hive.HiveContext.runSqlHive(
> HiveContext.scala:561)
>         at org.apache.spark.sql.hive.execution.HiveNativeCommand.
> run(HiveNativeCommand.scala:33)
>         at org.apache.spark.sql.execution.ExecutedCommand.
> sideEffectResult$lzycompute(commands.scala:57)
>         at org.apache.spark.sql.execution.ExecutedCommand.
> sideEffectResult(commands.scala:57)
>         at org.apache.spark.sql.execution.ExecutedCommand.
> doExecute(commands.scala:69)
>         at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> execute$5.apply(SparkPlan.scala:140)
>         at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> execute$5.apply(SparkPlan.scala:138)
>         at org.apache.spark.rdd.RDDOperationScope$.withScope(
> RDDOperationScope.scala:147)
>         at org.apache.spark.sql.execution.SparkPlan.execute(
> SparkPlan.scala:138)
>         at org.apache.spark.sql.SQLContext$QueryExecution.
> toRdd$lzycompute(SQLContext.scala:933)
>         at org.apache.spark.sql.SQLContext$QueryExecution.
> toRdd(SQLContext.scala:933)
>         at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:144)
>         at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:129)
>         at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:51)
>         at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:725)
>         at org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.
> run(SparkSQLDriver.scala:62)
>         at org.apache.spark.sql.hive.thriftserver.
> SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:308)
>         at org.apache.hadoop.hive.cli.CliDriver.processLine(
> CliDriver.java:376)
>         at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(
> SparkSQLCLIDriver.scala:226)
>         at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(
> SparkSQLCLIDriver.scala)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:62)
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:497)
>         at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> deploy$SparkSubmit$$runMain(SparkSubmit.scala:685)
>         at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
> SparkSubmit.scala:180)
>         at org.apache.spark.deploy.SparkSubmit$.submit(
> SparkSubmit.scala:205)
>         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
> scala:120)
>         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
> org.apache.spark.sql.execution.QueryExecutionException: FAILED: Execution
> Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. Cannot
> validate serde: org.apache.hive.hcatalog.data.JsonSerDe
>         at org.apache.spark.sql.hive.client.ClientWrapper$$anonfun$
> runHive$1.apply(ClientWrapper.scala:455)
>         at org.apache.spark.sql.hive.client.ClientWrapper$$anonfun$
> runHive$1.apply(ClientWrapper.scala:440)
>         at org.apache.spark.sql.hive.client.ClientWrapper$$anonfun$
> withHiveState$1.apply(ClientWrapper.scala:278)
>         at org.apache.spark.sql.hive.client.ClientWrapper.
> retryLocked(ClientWrapper.scala:233)
>         at org.apache.spark.sql.hive.client.ClientWrapper.
> withHiveState(ClientWrapper.scala:270)
>         at org.apache.spark.sql.hive.client.ClientWrapper.runHive(
> ClientWrapper.scala:440)
>         at org.apache.spark.sql.hive.client.ClientWrapper.
> runSqlHive(ClientWrapper.scala:430)
>         at org.apache.spark.sql.hive.HiveContext.runSqlHive(
> HiveContext.scala:561)
>         at org.apache.spark.sql.hive.execution.HiveNativeCommand.
> run(HiveNativeCommand.scala:33)
>         at org.apache.spark.sql.execution.ExecutedCommand.
> sideEffectResult$lzycompute(commands.scala:57)
>         at org.apache.spark.sql.execution.ExecutedCommand.
> sideEffectResult(commands.scala:57)
>         at org.apache.spark.sql.execution.ExecutedCommand.
> doExecute(commands.scala:69)
>         at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> execute$5.apply(SparkPlan.scala:140)
>         at org.apache.spark.sql.execution.SparkPlan$$anonfun$
> execute$5.apply(SparkPlan.scala:138)
>         at org.apache.spark.rdd.RDDOperationScope$.withScope(
> RDDOperationScope.scala:147)
>         at org.apache.spark.sql.execution.SparkPlan.execute(
> SparkPlan.scala:138)
>         at org.apache.spark.sql.SQLContext$QueryExecution.
> toRdd$lzycompute(SQLContext.scala:933)
>         at org.apache.spark.sql.SQLContext$QueryExecution.
> toRdd(SQLContext.scala:933)
>         at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:144)
>         at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:129)
>         at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:51)
>         at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:725)
>         at org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.
> run(SparkSQLDriver.scala:62)
>         at org.apache.spark.sql.hive.thriftserver.
> SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:308)
>         at org.apache.hadoop.hive.cli.CliDriver.processLine(
> CliDriver.java:376)
>         at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(
> SparkSQLCLIDriver.scala:226)
>         at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(
> SparkSQLCLIDriver.scala)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:62)
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:497)
>         at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$
> deploy$SparkSubmit$$runMain(SparkSubmit.scala:685)
>         at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(
> SparkSubmit.scala:180)
>         at org.apache.spark.deploy.SparkSubmit$.submit(
> SparkSubmit.scala:205)
>         at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.
> scala:120)
>         at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>
> I double-checked, my classpath is clean, I have both
> of /usr/hdp/current/hive-webhcat/share/hcatalog/hive-hcatalog-core.jar
> and /usr/hdp/2.3.4.0-3485/hive/lib/hive-serde.jar lib in it.
>
> Does someone know that problem ? Any inputs about where it could come from
> ?
>
> Thanks in advance for your help !
> Regards,
>
>
> Loïc
>
> Loïc CHANEL
> System Big Data engineer
> MS&T - Worldline Analytics Platform - Worldline (Villeurbanne, France)
>

Reply via email to