yes, have placed spark-assembly jar in hive lib folder.
hive.log--- bmit.2317151720491931059.properties --class org.apache.hive.spark.client.RemoteDriver /opt/cluster/apache-hive-1.2.0-SNAPSHOT-bin/lib/hive-exec-1.2.0-SNAPSHOT.jar --remote-host M151 --remote-port 56996 --conf hive.spark.client.connect.timeout=10000 --conf hive.spark.client.server.connect.timeout=90000 --conf hive.spark.client.channel.log.level=null --conf hive.spark.client.rpc.max.size=52428800 --conf hive.spark.client.rpc.threads=8 --conf hive.spark.client.secret.bits=256 2015-03-02 20:33:39,893 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - Warning: Ignoring non-spark config property: hive.spark.client.connect.timeout=10000 2015-03-02 20:33:39,894 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - Warning: Ignoring non-spark config property: hive.spark.client.rpc.threads=8 2015-03-02 20:33:39,894 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - Warning: Ignoring non-spark config property: hive.spark.client.rpc.max.size=52428800 2015-03-02 20:33:39,894 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - Warning: Ignoring non-spark config property: hive.spark.client.secret.bits=256 2015-03-02 20:33:39,894 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - Warning: Ignoring non-spark config property: hive.spark.client.server.connect.timeout=90000 2015-03-02 20:33:40,002 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - 15/03/02 20:33:40 INFO client.RemoteDriver: Connecting to: M151:56996 2015-03-02 20:33:40,005 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - Exception in thread "main" java.lang.NoSuchFieldError: SPARK_RPC_CLIENT_CONNECT_TIMEOUT 2015-03-02 20:33:40,005 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - at org.apache.hive.spark.client.rpc.RpcConfiguration.<clinit>(RpcConfiguration.java:46) 2015-03-02 20:33:40,005 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - at org.apache.hive.spark.client.RemoteDriver.<init>(RemoteDriver.java:139) 2015-03-02 20:33:40,005 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - at org.apache.hive.spark.client.RemoteDriver.main(RemoteDriver.java:544) 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - at java.lang.reflect.Method.invoke(Method.java:601) 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:569) 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:166) 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:189) 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:110) 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(553)) - at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala) 2015-03-02 20:33:40,410 WARN [Driver]: client.SparkClientImpl (SparkClientImpl.java:run(411)) - Child process exited with code 1. 2015-03-02 20:35:08,950 WARN [main]: client.SparkClientImpl (SparkClientImpl.java:<init>(98)) - Error while waiting for client to connect. java.util.concurrent.ExecutionException: java.util.concurrent.TimeoutException: Timed out waiting for client connection. at io.netty.util.concurrent.AbstractFuture.get(AbstractFuture.java:37) at org.apache.hive.spark.client.SparkClientImpl.<init>(SparkClientImpl.java:96) at org.apache.hive.spark.client.SparkClientFactory.createClient(SparkClientFactory.java:80) at org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient.<init>(RemoteHiveSparkClient.java:88) at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.createHiveSparkClient(HiveSparkClientFactory.java:65) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:55) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionManagerImpl.getSession(SparkSessionManagerImpl.java:116) at org.apache.hadoop.hive.ql.exec.spark.SparkUtilities.getSparkSession(SparkUtilities.java:114) at org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:95) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:88) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1642) at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1402) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1187) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1053) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1043) at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:207) at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:159) at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:370) at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:754) at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675) at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:615) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:601) at org.apache.hadoop.util.RunJar.run(RunJar.java:221) at org.apache.hadoop.util.RunJar.main(RunJar.java:136) Caused by: java.util.concurrent.TimeoutException: Timed out waiting for client connection. at org.apache.hive.spark.client.rpc.RpcServer$2.run(RpcServer.java:134) at io.netty.util.concurrent.PromiseTask$RunnableAdapter.call(PromiseTask.java:38) at io.netty.util.concurrent.ScheduledFutureTask.run(ScheduledFutureTask.java:123) at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:380) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:357) at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:116) at java.lang.Thread.run(Thread.java:722) 2015-03-02 20:35:08,979 ERROR [main]: exec.Task (SessionState.java:printError(944)) - Failed to execute spark task, with exception 'org.apache.hadoop.hive.ql.metadata.HiveException(Failed to create spark client.)' org.apache.hadoop.hive.ql.metadata.HiveException: Failed to create spark client. at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:57) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionManagerImpl.getSession(SparkSessionManagerImpl.java:116) at org.apache.hadoop.hive.ql.exec.spark.SparkUtilities.getSparkSession(SparkUtilities.java:114) at org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute(SparkTask.java:95) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:88) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1642) at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1402) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1187) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1053) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1043) at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:207) at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:159) at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:370) at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:754) at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675) at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:615) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:601) at org.apache.hadoop.util.RunJar.run(RunJar.java:221) at org.apache.hadoop.util.RunJar.main(RunJar.java:136) Caused by: java.lang.RuntimeException: java.util.concurrent.ExecutionException: java.util.concurrent.TimeoutException: Timed out waiting for client connection. at com.google.common.base.Throwables.propagate(Throwables.java:156) at org.apache.hive.spark.client.SparkClientImpl.<init>(SparkClientImpl.java:106) at org.apache.hive.spark.client.SparkClientFactory.createClient(SparkClientFactory.java:80) at org.apache.hadoop.hive.ql.exec.spark.RemoteHiveSparkClient.<init>(RemoteHiveSparkClient.java:88) at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory.createHiveSparkClient(HiveSparkClientFactory.java:65) at org.apache.hadoop.hive.ql.exec.spark.session.SparkSessionImpl.open(SparkSessionImpl.java:55) ... 22 more Caused by: java.util.concurrent.ExecutionException: java.util.concurrent.TimeoutException: Timed out waiting for client connection. at io.netty.util.concurrent.AbstractFuture.get(AbstractFuture.java:37) at org.apache.hive.spark.client.SparkClientImpl.<init>(SparkClientImpl.java:96) ... 26 more Caused by: java.util.concurrent.TimeoutException: Timed out waiting for client connection. at org.apache.hive.spark.client.rpc.RpcServer$2.run(RpcServer.java:134) at io.netty.util.concurrent.PromiseTask$RunnableAdapter.call(PromiseTask.java:38) at io.netty.util.concurrent.ScheduledFutureTask.run(ScheduledFutureTask.java:123) at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:380) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:357) at io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:116) at java.lang.Thread.run(Thread.java:722) and i do not find spark.log thanks. On 2015/3/2 22:39, Xuefu Zhang wrote:
Could you check your hive.log and spark.log for more detailed error message? Quick check though, do you have spark-assembly.jar in your hive lib folder? Thanks, Xuefu On Mon, Mar 2, 2015 at 5:14 AM, scwf <wangf...@huawei.com <mailto:wangf...@huawei.com>> wrote: Hi all, anyone met this error: HiveException(Failed to create spark client.) M151:/opt/cluster/apache-hive-__1.2.0-SNAPSHOT-bin # bin/hive Logging initialized using configuration in jar:file:/opt/cluster/apache-__hive-1.2.0-SNAPSHOT-bin/lib/__hive-common-1.2.0-SNAPSHOT.__jar!/hive-log4j.properties [INFO] Unable to bind key for unsupported operation: backward-delete-word [INFO] Unable to bind key for unsupported operation: backward-delete-word [INFO] Unable to bind key for unsupported operation: down-history [INFO] Unable to bind key for unsupported operation: up-history [INFO] Unable to bind key for unsupported operation: up-history [INFO] Unable to bind key for unsupported operation: down-history [INFO] Unable to bind key for unsupported operation: up-history [INFO] Unable to bind key for unsupported operation: down-history [INFO] Unable to bind key for unsupported operation: up-history [INFO] Unable to bind key for unsupported operation: down-history [INFO] Unable to bind key for unsupported operation: up-history [INFO] Unable to bind key for unsupported operation: down-history hive> set spark.home=/opt/cluster/spark-__1.3.0-bin-hadoop2-without-__hive; hive> set hive.execution.engine=spark; hive> set spark.master=spark://9.91.8.__151:7070 <http://9.91.8.151:7070>; hive> select count(1) from src; Query ID = root_20150302203333_4bed4c2a-__b9a5-4d99-a485-67570e2712b7 Total jobs = 1 Launching Job 1 out of 1 In order to change the average load for a reducer (in bytes): set hive.exec.reducers.bytes.per.__reducer=<number> In order to limit the maximum number of reducers: set hive.exec.reducers.max=<__number> In order to set a constant number of reducers: set mapreduce.job.reduces=<number> Failed to execute spark task, with exception 'org.apache.hadoop.hive.ql.__metadata.HiveException(Failed to create spark client.)' FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.__exec.spark.SparkTask thanks