It seems that the remote spark context failed to come up. I saw you're using Spark standalone cluster. Please make sure spark cluster is up. You may try spark.master=local first.
On Mon, Mar 2, 2015 at 5:15 PM, scwf <wangf...@huawei.com> wrote: > yes, have placed spark-assembly jar in hive lib folder. > > hive.log--- > bmit.2317151720491931059.properties --class > org.apache.hive.spark.client.RemoteDriver > /opt/cluster/apache-hive-1.2.0-SNAPSHOT-bin/lib/hive-exec-1.2.0-SNAPSHOT.jar > --remote-host M151 --remote-port 56996 --conf > hive.spark.client.connect.timeout=10000 > --conf hive.spark.client.server.connect.timeout=90000 --conf > hive.spark.client.channel.log.level=null --conf > hive.spark.client.rpc.max.size=52428800 > --conf hive.spark.client.rpc.threads=8 --conf > hive.spark.client.secret.bits=256 > 2015-03-02 20:33:39,893 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - Warning: Ignoring non-spark config > property: hive.spark.client.connect.timeout=10000 > 2015-03-02 20:33:39,894 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - Warning: Ignoring non-spark config > property: hive.spark.client.rpc.threads=8 > 2015-03-02 20:33:39,894 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - Warning: Ignoring non-spark config > property: hive.spark.client.rpc.max.size=52428800 > 2015-03-02 20:33:39,894 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - Warning: Ignoring non-spark config > property: hive.spark.client.secret.bits=256 > 2015-03-02 20:33:39,894 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - Warning: Ignoring non-spark config > property: hive.spark.client.server.connect.timeout=90000 > 2015-03-02 20:33:40,002 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - 15/03/02 20:33:40 INFO > client.RemoteDriver: Connecting to: M151:56996 > 2015-03-02 20:33:40,005 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - Exception in thread "main" > java.lang.NoSuchFieldError: SPARK_RPC_CLIENT_CONNECT_TIMEOUT > 2015-03-02 20:33:40,005 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - at org.apache.hive.spark.client. > rpc.RpcConfiguration.<clinit>(RpcConfiguration.java:46) > 2015-03-02 20:33:40,005 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - at org.apache.hive.spark.client. > RemoteDriver.<init>(RemoteDriver.java:139) > 2015-03-02 20:33:40,005 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - at org.apache.hive.spark.client. > RemoteDriver.main(RemoteDriver.java:544) > 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - at sun.reflect. > NativeMethodAccessorImpl.invoke0(Native Method) > 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - at sun.reflect. > NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - at sun.reflect. > DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - at java.lang.reflect.Method. > invoke(Method.java:601) > 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - at org.apache.spark.deploy. > SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain( > SparkSubmit.scala:569) > 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - at org.apache.spark.deploy. > SparkSubmit$.doRunMain$1(SparkSubmit.scala:166) > 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - at org.apache.spark.deploy. > SparkSubmit$.submit(SparkSubmit.scala:189) > 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - at org.apache.spark.deploy. > SparkSubmit$.main(SparkSubmit.scala:110) > 2015-03-02 20:33:40,006 INFO [stderr-redir-1]: client.SparkClientImpl > (SparkClientImpl.java:run(553)) - at org.apache.spark.deploy. > SparkSubmit.main(SparkSubmit.scala) > 2015-03-02 20:33:40,410 WARN [Driver]: client.SparkClientImpl > (SparkClientImpl.java:run(411)) - Child process exited with code 1. > 2015-03-02 20:35:08,950 WARN [main]: client.SparkClientImpl > (SparkClientImpl.java:<init>(98)) - Error while waiting for client to > connect. > java.util.concurrent.ExecutionException: > java.util.concurrent.TimeoutException: > Timed out waiting for client connection. > at io.netty.util.concurrent.AbstractFuture.get( > AbstractFuture.java:37) > at org.apache.hive.spark.client.SparkClientImpl.<init>( > SparkClientImpl.java:96) > at org.apache.hive.spark.client.SparkClientFactory.createClient( > SparkClientFactory.java:80) > at org.apache.hadoop.hive.ql.exec.spark. > RemoteHiveSparkClient.<init>(RemoteHiveSparkClient.java:88) > at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory. > createHiveSparkClient(HiveSparkClientFactory.java:65) > at org.apache.hadoop.hive.ql.exec.spark.session. > SparkSessionImpl.open(SparkSessionImpl.java:55) > at org.apache.hadoop.hive.ql.exec.spark.session. > SparkSessionManagerImpl.getSession(SparkSessionManagerImpl.java:116) > at org.apache.hadoop.hive.ql.exec.spark.SparkUtilities. > getSparkSession(SparkUtilities.java:114) > at org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute( > SparkTask.java:95) > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160) > at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential( > TaskRunner.java:88) > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1642) > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1402) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1187) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1053) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1043) > at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd( > CliDriver.java:207) > at org.apache.hadoop.hive.cli.CliDriver.processCmd( > CliDriver.java:159) > at org.apache.hadoop.hive.cli.CliDriver.processLine( > CliDriver.java:370) > at org.apache.hadoop.hive.cli.CliDriver.executeDriver( > CliDriver.java:754) > at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:615) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:57) > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.apache.hadoop.util.RunJar.run(RunJar.java:221) > at org.apache.hadoop.util.RunJar.main(RunJar.java:136) > Caused by: java.util.concurrent.TimeoutException: Timed out waiting for > client connection. > at org.apache.hive.spark.client.rpc.RpcServer$2.run(RpcServer. > java:134) > at io.netty.util.concurrent.PromiseTask$RunnableAdapter. > call(PromiseTask.java:38) > at io.netty.util.concurrent.ScheduledFutureTask.run( > ScheduledFutureTask.java:123) > at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks( > SingleThreadEventExecutor.java:380) > at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:357) > at io.netty.util.concurrent.SingleThreadEventExecutor$2. > run(SingleThreadEventExecutor.java:116) > at java.lang.Thread.run(Thread.java:722) > 2015-03-02 20:35:08,979 ERROR [main]: exec.Task > (SessionState.java:printError(944)) - Failed to execute spark task, with > exception 'org.apache.hadoop.hive.ql.metadata.HiveException(Failed to > create spark client.)' > org.apache.hadoop.hive.ql.metadata.HiveException: Failed to create spark > client. > at org.apache.hadoop.hive.ql.exec.spark.session. > SparkSessionImpl.open(SparkSessionImpl.java:57) > at org.apache.hadoop.hive.ql.exec.spark.session. > SparkSessionManagerImpl.getSession(SparkSessionManagerImpl.java:116) > at org.apache.hadoop.hive.ql.exec.spark.SparkUtilities. > getSparkSession(SparkUtilities.java:114) > at org.apache.hadoop.hive.ql.exec.spark.SparkTask.execute( > SparkTask.java:95) > at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160) > at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential( > TaskRunner.java:88) > at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1642) > at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1402) > at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1187) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1053) > at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1043) > at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd( > CliDriver.java:207) > at org.apache.hadoop.hive.cli.CliDriver.processCmd( > CliDriver.java:159) > at org.apache.hadoop.hive.cli.CliDriver.processLine( > CliDriver.java:370) > at org.apache.hadoop.hive.cli.CliDriver.executeDriver( > CliDriver.java:754) > at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:675) > at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:615) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at sun.reflect.NativeMethodAccessorImpl.invoke( > NativeMethodAccessorImpl.java:57) > at sun.reflect.DelegatingMethodAccessorImpl.invoke( > DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:601) > at org.apache.hadoop.util.RunJar.run(RunJar.java:221) > at org.apache.hadoop.util.RunJar.main(RunJar.java:136) > Caused by: java.lang.RuntimeException: > java.util.concurrent.ExecutionException: > java.util.concurrent.TimeoutException: Timed out waiting for client > connection. > at com.google.common.base.Throwables.propagate( > Throwables.java:156) > at org.apache.hive.spark.client.SparkClientImpl.<init>( > SparkClientImpl.java:106) > at org.apache.hive.spark.client.SparkClientFactory.createClient( > SparkClientFactory.java:80) > at org.apache.hadoop.hive.ql.exec.spark. > RemoteHiveSparkClient.<init>(RemoteHiveSparkClient.java:88) > at org.apache.hadoop.hive.ql.exec.spark.HiveSparkClientFactory. > createHiveSparkClient(HiveSparkClientFactory.java:65) > at org.apache.hadoop.hive.ql.exec.spark.session. > SparkSessionImpl.open(SparkSessionImpl.java:55) > ... 22 more > Caused by: java.util.concurrent.ExecutionException: > java.util.concurrent.TimeoutException: > Timed out waiting for client connection. > at io.netty.util.concurrent.AbstractFuture.get( > AbstractFuture.java:37) > at org.apache.hive.spark.client.SparkClientImpl.<init>( > SparkClientImpl.java:96) > ... 26 more > Caused by: java.util.concurrent.TimeoutException: Timed out waiting for > client connection. > at org.apache.hive.spark.client.rpc.RpcServer$2.run(RpcServer. > java:134) > at io.netty.util.concurrent.PromiseTask$RunnableAdapter. > call(PromiseTask.java:38) > at io.netty.util.concurrent.ScheduledFutureTask.run( > ScheduledFutureTask.java:123) > at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks( > SingleThreadEventExecutor.java:380) > at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:357) > at io.netty.util.concurrent.SingleThreadEventExecutor$2. > run(SingleThreadEventExecutor.java:116) > at java.lang.Thread.run(Thread.java:722) > > > and i do not find spark.log > > thanks. > > > On 2015/3/2 22:39, Xuefu Zhang wrote: > >> Could you check your hive.log and spark.log for more detailed error >> message? Quick check though, do you have spark-assembly.jar in your hive >> lib folder? >> >> Thanks, >> Xuefu >> >> On Mon, Mar 2, 2015 at 5:14 AM, scwf <wangf...@huawei.com <mailto: >> wangf...@huawei.com>> wrote: >> >> Hi all, >> anyone met this error: HiveException(Failed to create spark >> client.) >> >> M151:/opt/cluster/apache-hive-__1.2.0-SNAPSHOT-bin # bin/hive >> >> Logging initialized using configuration in >> jar:file:/opt/cluster/apache-__hive-1.2.0-SNAPSHOT-bin/lib/_ >> _hive-common-1.2.0-SNAPSHOT.__jar!/hive-log4j.properties >> [INFO] Unable to bind key for unsupported operation: >> backward-delete-word >> [INFO] Unable to bind key for unsupported operation: >> backward-delete-word >> [INFO] Unable to bind key for unsupported operation: down-history >> [INFO] Unable to bind key for unsupported operation: up-history >> [INFO] Unable to bind key for unsupported operation: up-history >> [INFO] Unable to bind key for unsupported operation: down-history >> [INFO] Unable to bind key for unsupported operation: up-history >> [INFO] Unable to bind key for unsupported operation: down-history >> [INFO] Unable to bind key for unsupported operation: up-history >> [INFO] Unable to bind key for unsupported operation: down-history >> [INFO] Unable to bind key for unsupported operation: up-history >> [INFO] Unable to bind key for unsupported operation: down-history >> hive> set spark.home=/opt/cluster/spark- >> __1.3.0-bin-hadoop2-without-__hive; >> hive> set hive.execution.engine=spark; >> hive> set spark.master=spark://9.91.8.__151:7070 < >> http://9.91.8.151:7070>; >> hive> select count(1) from src; >> Query ID = root_20150302203333_4bed4c2a-__b9a5-4d99-a485-67570e2712b7 >> Total jobs = 1 >> Launching Job 1 out of 1 >> In order to change the average load for a reducer (in bytes): >> set hive.exec.reducers.bytes.per.__reducer=<number> >> In order to limit the maximum number of reducers: >> set hive.exec.reducers.max=<__number> >> In order to set a constant number of reducers: >> set mapreduce.job.reduces=<number> >> Failed to execute spark task, with exception >> 'org.apache.hadoop.hive.ql.__metadata.HiveException(Failed to create >> spark client.)' >> FAILED: Execution Error, return code 1 from >> org.apache.hadoop.hive.ql.__exec.spark.SparkTask >> >> thanks >> >> >> > >