*It seems that the submit number depend on stage of the query.*
*This query include three stages.*

If queue resources is still *not enough after submit threee
applications,** Hive
client will close.*
*"**Failed to execute spark task, with exception
'org.apache.hadoop.hive.ql.metadata.HiveException(Failed to create spark
client.)'*
*FAILED: Execution Error, return code 1 from
org.apache.hadoop.hive.ql.exec.spark.SparkTask**"*
*And this time, the port(eg **34682**)  kill in hive client(eg *
*10.179.12.140**) use to **communicate with RSC **will  lost.*

*The reources of queue is free **after awhile, the AM of three applications
will fast fail beacause of "**15/12/10 12:28:43 INFO client.RemoteDriver:
Connecting to: 10.179.12.140:34682...java.net.ConnectException: Connection
refused: /10.179.12.140:34682 <http://10.179.12.140:34682>**"*

*So, The application will fail if the queue resources if not **enough at
point of the query be submited, even if the resources is free **after
awhile.*
*Do you have more idea about this question?*

*Attch the query*
set hive.execution.engine=spark;
set spark.yarn.queue=tms;
set spark.app.name=t_ad_tms_heartbeat_ok_3;
insert overwrite table t_ad_tms_heartbeat_ok partition(ds=20151208)
SELECT
            NVL(a.qimei, b.qimei) AS qimei,
            NVL(b.first_ip,a.user_ip) AS first_ip,
            NVL(a.user_ip, b.last_ip) AS last_ip,
            NVL(b.first_date, a.ds) AS first_date,
            NVL(a.ds, b.last_date) AS last_date,
            NVL(b.first_chid, a.chid) AS first_chid,
            NVL(a.chid, b.last_chid) AS last_chid,
            NVL(b.first_lc, a.lc) AS first_lc,
            NVL(a.lc, b.last_lc) AS last_lc,
            NVL(a.guid, b.guid) AS guid,
            NVL(a.sn, b.sn) AS sn,
            NVL(a.vn, b.vn) AS vn,
                        NVL(a.vc, b.vc) AS vc,
                        NVL(a.mo, b.mo) AS mo,
                        NVL(a.rl, b.rl) AS rl,
                        NVL(a.os, b.os) AS os,
                        NVL(a.rv, b.rv) AS rv,
                        NVL(a.qv, b.qv) AS qv,
                        NVL(a.imei, b.imei) AS imei,
                        NVL(a.romid, b.romid) AS romid,
            NVL(a.bn, b.bn) AS bn,
                        NVL(a.account_type, b.account_type) AS account_type,
                        NVL(a.account, b.account) AS account
    FROM
        (SELECT
ds,user_ip,guid,sn,vn,vc,mo,rl,chid,lcid,os,rv,qv,imei,qimei,lc,romid,bn,account_type,account
        FROM    t_od_tms_heartbeat_ok
        WHERE   ds = 20151208) a
    FULL OUTER JOIN
        (SELECT
qimei,first_ip,last_ip,first_date,last_date,first_chid,last_chid,first_lc,last_lc,guid,sn,vn,vc,mo,rl,os,rv,qv,imei,romid,bn,account_type,account
        FROM    t_ad_tms_heartbeat_ok
        WHERE   last_date > 20150611
        AND     ds = 20151207) b
        ON   a.qimei=b.qimei;

*Thanks.*
*Best wishes.*

2015-12-09 19:51 GMT+08:00 Jone Zhang <joyoungzh...@gmail.com>:

> But in some cases all of the applications will fail which caused
>> by SparkContext did not initialize after waiting for 150000 ms.
>> See attchment (hive.spark.client.server.connect.timeout is set to 5min).
>
>
> *The error log is different  from original mail*
>
> Container: container_1448873753366_113453_01_000001 on 10.247.169.134_8041
>
> ============================================================================
> LogType: stderr
> LogLength: 3302
> Log Contents:
> Please use CMSClassUnloadingEnabled in place of CMSPermGenSweepingEnabled
> in the future
> Please use CMSClassUnloadingEnabled in place of CMSPermGenSweepingEnabled
> in the future
> 15/12/09 02:11:48 INFO yarn.ApplicationMaster: Registered signal handlers
> for [TERM, HUP, INT]
> 15/12/09 02:11:48 INFO yarn.ApplicationMaster: ApplicationAttemptId:
> appattempt_1448873753366_113453_000001
> 15/12/09 02:11:49 INFO spark.SecurityManager: Changing view acls to: mqq
> 15/12/09 02:11:49 INFO spark.SecurityManager: Changing modify acls to: mqq
> 15/12/09 02:11:49 INFO spark.SecurityManager: SecurityManager:
> authentication disabled; ui acls disabled; users with view permissions:
> Set(mqq); users with modify permissions: Set(mqq)
> 15/12/09 02:11:49 INFO yarn.ApplicationMaster: Starting the user
> application in a separate Thread
> 15/12/09 02:11:49 INFO yarn.ApplicationMaster: Waiting for spark context
> initialization
> 15/12/09 02:11:49 INFO yarn.ApplicationMaster: Waiting for spark context
> initialization ...
> 15/12/09 02:11:49 INFO client.RemoteDriver: Connecting to:
> 10.179.12.140:58013
> 15/12/09 02:11:49 ERROR yarn.ApplicationMaster: User class threw
> exception: java.util.concurrent.ExecutionException:
> java.net.ConnectException: Connection refused: /10.179.12.140:58013
> java.util.concurrent.ExecutionException: java.net.ConnectException:
> Connection refused: /10.179.12.140:58013
>         at
> io.netty.util.concurrent.AbstractFuture.get(AbstractFuture.java:37)
>         at
> org.apache.hive.spark.client.RemoteDriver.<init>(RemoteDriver.java:156)
>         at
> org.apache.hive.spark.client.RemoteDriver.main(RemoteDriver.java:556)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>         at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:606)
>         at
> org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:483)
> Caused by: java.net.ConnectException: Connection refused: /
> 10.179.12.140:58013
>         at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
>         at
> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:739)
>         at
> io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:208)
>         at
> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:287)
>         at
> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:528)
>         at
> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:468)
>         at
> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:382)
>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:354)
>         at
> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:116)
>         at java.lang.Thread.run(Thread.java:745)
> 15/12/09 02:11:49 INFO yarn.ApplicationMaster: Final app status: FAILED,
> exitCode: 15, (reason: User class threw exception:
> java.util.concurrent.ExecutionException: java.net.ConnectException:
> Connection refused: /10.179.12.140:58013)
> 15/12/09 02:11:59 ERROR yarn.ApplicationMaster: SparkContext did not
> initialize after waiting for 150000 ms. Please check earlier log output for
> errors. Failing the application.
> 15/12/09 02:11:59 INFO util.Utils: Shutdown hook called
>
> 2015-12-09 19:22 GMT+08:00 Jone Zhang <joyoungzh...@gmail.com>:
>
>> Hive version is 1.2.1
>> Spark version is 1.4.1
>> Hadoop version is 2.5.1
>>
>> The application_1448873753366_121062 will success in the above mail.
>>
>> But in some cases all of the applications will fail which caused by 
>> SparkContext
>> did not initialize after waiting for 150000 ms.
>> See attchment (hive.spark.client.server.connect.timeout is set to 5min).
>>
>> Thanks.
>> Best wishes.
>>
>> 2015-12-09 17:56 GMT+08:00 Jone Zhang <joyoungzh...@gmail.com>:
>>
>>> *Hi, Xuefu:*
>>>
>>> *See attachment 1*
>>> *When the queue resources is not enough.*
>>> *The application application_1448873753366_121022 will pending.*
>>> *Two minutes later, the application application_1448873753366_121055
>>> will be submited and pending.*
>>> *And then application_1448873753366_121062.*
>>>
>>> *See attachment 2*
>>> *When the queue resources is free.*
>>> *The application  application_1448873753366_121062 begin to running.*
>>> *Application_1448873753366_121022 and application_1448873753366_121055
>>>  will failed fast.*
>>>
>>> *Logs of Application_1448873753366_121022 as follows(same as *
>>> *application_1448873753366_121055**):*
>>> Container: container_1448873753366_121022_03_000001 on
>>> 10.226.136.122_8041
>>>
>>> ============================================================================
>>> LogType: stderr
>>> LogLength: 4664
>>> Log Contents:
>>> Please use CMSClassUnloadingEnabled in place of
>>> CMSPermGenSweepingEnabled in the future
>>> Please use CMSClassUnloadingEnabled in place of
>>> CMSPermGenSweepingEnabled in the future
>>> 15/12/09 16:29:45 INFO yarn.ApplicationMaster: Registered signal
>>> handlers for [TERM, HUP, INT]
>>> 15/12/09 16:29:46 INFO yarn.ApplicationMaster: ApplicationAttemptId:
>>> appattempt_1448873753366_121022_000003
>>> 15/12/09 16:29:47 INFO spark.SecurityManager: Changing view acls to: mqq
>>> 15/12/09 16:29:47 INFO spark.SecurityManager: Changing modify acls to:
>>> mqq
>>> 15/12/09 16:29:47 INFO spark.SecurityManager: SecurityManager:
>>> authentication disabled; ui acls disabled; users with view permissions:
>>> Set(mqq); users with modify permissions: Set(mqq)
>>> 15/12/09 16:29:47 INFO yarn.ApplicationMaster: Starting the user
>>> application in a separate Thread
>>> 15/12/09 16:29:47 INFO yarn.ApplicationMaster: Waiting for spark context
>>> initialization
>>> 15/12/09 16:29:47 INFO yarn.ApplicationMaster: Waiting for spark context
>>> initialization ...
>>> 15/12/09 16:29:47 INFO client.RemoteDriver: Connecting to:
>>> 10.179.12.140:38842
>>> 15/12/09 16:29:48 WARN rpc.Rpc: Invalid log level null, reverting to
>>> default.
>>> 15/12/09 16:29:48 ERROR yarn.ApplicationMaster: User class threw
>>> exception: java.util.concurrent.ExecutionException:
>>> javax.security.sasl.SaslException: Client closed before SASL negotiation
>>> finished.
>>> java.util.concurrent.ExecutionException:
>>> javax.security.sasl.SaslException: Client closed before SASL negotiation
>>> finished.
>>>         at
>>> io.netty.util.concurrent.AbstractFuture.get(AbstractFuture.java:37)
>>>         at
>>> org.apache.hive.spark.client.RemoteDriver.<init>(RemoteDriver.java:156)
>>>         at
>>> org.apache.hive.spark.client.RemoteDriver.main(RemoteDriver.java:556)
>>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>         at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>         at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>         at java.lang.reflect.Method.invoke(Method.java:606)
>>>         at
>>> org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:483)
>>> Caused by: javax.security.sasl.SaslException: Client closed before SASL
>>> negotiation finished.
>>>         at
>>> org.apache.hive.spark.client.rpc.Rpc$SaslClientHandler.dispose(Rpc.java:449)
>>>         at
>>> org.apache.hive.spark.client.rpc.SaslHandler.channelInactive(SaslHandler.java:90)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:233)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:219)
>>>         at
>>> io.netty.channel.ChannelInboundHandlerAdapter.channelInactive(ChannelInboundHandlerAdapter.java:75)
>>>         at
>>> org.apache.hive.spark.client.rpc.KryoMessageCodec.channelInactive(KryoMessageCodec.java:127)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:233)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:219)
>>>         at
>>> io.netty.channel.ChannelInboundHandlerAdapter.channelInactive(ChannelInboundHandlerAdapter.java:75)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:233)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:219)
>>>         at
>>> io.netty.channel.DefaultChannelPipeline.fireChannelInactive(DefaultChannelPipeline.java:769)
>>>         at
>>> io.netty.channel.AbstractChannel$AbstractUnsafe$5.run(AbstractChannel.java:567)
>>>         at
>>> io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:380)
>>>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:357)
>>>         at
>>> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:116)
>>>         at java.lang.Thread.run(Thread.java:745)
>>> 15/12/09 16:29:48 INFO yarn.ApplicationMaster: Final app status: FAILED,
>>> exitCode: 15, (reason: User class threw exception:
>>> java.util.concurrent.ExecutionException: javax.security.sasl.SaslException:
>>> Client closed before SASL negotiation finished.)
>>> 15/12/09 16:29:57 ERROR yarn.ApplicationMaster: SparkContext did not
>>> initialize after waiting for 150000 ms. Please check earlier log output for
>>> errors. Failing the application.
>>> 15/12/09 16:29:57 INFO yarn.ApplicationMaster: Unregistering
>>> ApplicationMaster with FAILED (diag message: User class threw exception:
>>> java.util.concurrent.ExecutionException: javax.security.sasl.SaslException:
>>> Client closed before SASL negotiation finished.)
>>> 15/12/09 16:29:57 INFO yarn.ApplicationMaster: Deleting staging
>>> directory .sparkStaging/application_1448873753366_121022
>>> 15/12/09 16:29:57 INFO util.Utils: Shutdown hook called
>>>
>>> LogType: stdout
>>> LogLength: 216
>>> Log Contents:
>>> Java HotSpot(TM) 64-Bit Server VM warning: ignoring option
>>> UseCompressedStrings; support was removed in 7.0
>>> Java HotSpot(TM) 64-Bit Server VM warning: ignoring option
>>> UseCompressedStrings; support was removed in 7.0
>>>
>>>
>>>
>>> Container: container_1448873753366_121022_01_000001 on 10.226.148.5_8041
>>>
>>> ==========================================================================
>>> LogType: stderr
>>> LogLength: 4284
>>> Log Contents:
>>> Please use CMSClassUnloadingEnabled in place of
>>> CMSPermGenSweepingEnabled in the future
>>> Please use CMSClassUnloadingEnabled in place of
>>> CMSPermGenSweepingEnabled in the future
>>> 15/12/09 16:29:17 INFO yarn.ApplicationMaster: Registered signal
>>> handlers for [TERM, HUP, INT]
>>> 15/12/09 16:29:18 INFO yarn.ApplicationMaster: ApplicationAttemptId:
>>> appattempt_1448873753366_121022_000001
>>> 15/12/09 16:29:18 INFO spark.SecurityManager: Changing view acls to: mqq
>>> 15/12/09 16:29:18 INFO spark.SecurityManager: Changing modify acls to:
>>> mqq
>>> 15/12/09 16:29:18 INFO spark.SecurityManager: SecurityManager:
>>> authentication disabled; ui acls disabled; users with view permissions:
>>> Set(mqq); users with modify permissions: Set(mqq)
>>> 15/12/09 16:29:19 INFO yarn.ApplicationMaster: Starting the user
>>> application in a separate Thread
>>> 15/12/09 16:29:19 INFO yarn.ApplicationMaster: Waiting for spark context
>>> initialization
>>> 15/12/09 16:29:19 INFO yarn.ApplicationMaster: Waiting for spark context
>>> initialization ...
>>> 15/12/09 16:29:19 INFO client.RemoteDriver: Connecting to:
>>> 10.179.12.140:38842
>>> 15/12/09 16:29:20 WARN rpc.Rpc: Invalid log level null, reverting to
>>> default.
>>> 15/12/09 16:29:20 ERROR yarn.ApplicationMaster: User class threw
>>> exception: java.util.concurrent.ExecutionException:
>>> javax.security.sasl.SaslException: Client closed before SASL negotiation
>>> finished.
>>> java.util.concurrent.ExecutionException:
>>> javax.security.sasl.SaslException: Client closed before SASL negotiation
>>> finished.
>>>         at
>>> io.netty.util.concurrent.AbstractFuture.get(AbstractFuture.java:37)
>>>         at
>>> org.apache.hive.spark.client.RemoteDriver.<init>(RemoteDriver.java:156)
>>>         at
>>> org.apache.hive.spark.client.RemoteDriver.main(RemoteDriver.java:556)
>>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>         at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>         at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>         at java.lang.reflect.Method.invoke(Method.java:606)
>>>         at
>>> org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:483)
>>> Caused by: javax.security.sasl.SaslException: Client closed before SASL
>>> negotiation finished.
>>>         at
>>> org.apache.hive.spark.client.rpc.Rpc$SaslClientHandler.dispose(Rpc.java:449)
>>>         at
>>> org.apache.hive.spark.client.rpc.SaslHandler.channelInactive(SaslHandler.java:90)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:233)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:219)
>>>         at
>>> io.netty.channel.ChannelInboundHandlerAdapter.channelInactive(ChannelInboundHandlerAdapter.java:75)
>>>         at
>>> org.apache.hive.spark.client.rpc.KryoMessageCodec.channelInactive(KryoMessageCodec.java:127)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:233)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:219)
>>>         at
>>> io.netty.channel.ChannelInboundHandlerAdapter.channelInactive(ChannelInboundHandlerAdapter.java:75)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:233)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:219)
>>>         at
>>> io.netty.channel.DefaultChannelPipeline.fireChannelInactive(DefaultChannelPipeline.java:769)
>>>         at
>>> io.netty.channel.AbstractChannel$AbstractUnsafe$5.run(AbstractChannel.java:567)
>>>         at
>>> io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:380)
>>>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:357)
>>>         at
>>> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:116)
>>>         at java.lang.Thread.run(Thread.java:745)
>>> 15/12/09 16:29:20 INFO yarn.ApplicationMaster: Final app status: FAILED,
>>> exitCode: 15, (reason: User class threw exception:
>>> java.util.concurrent.ExecutionException: javax.security.sasl.SaslException:
>>> Client closed before SASL negotiation finished.)
>>> 15/12/09 16:29:29 ERROR yarn.ApplicationMaster: SparkContext did not
>>> initialize after waiting for 150000 ms. Please check earlier log output for
>>> errors. Failing the application.
>>> 15/12/09 16:29:29 INFO util.Utils: Shutdown hook called
>>>
>>> LogType: stdout
>>> LogLength: 216
>>> Log Contents:
>>> Java HotSpot(TM) 64-Bit Server VM warning: ignoring option
>>> UseCompressedStrings; support was removed in 7.0
>>> Java HotSpot(TM) 64-Bit Server VM warning: ignoring option
>>> UseCompressedStrings; support was removed in 7.0
>>>
>>>
>>>
>>> Container: container_1448873753366_121022_02_000001 on 10.239.243.69_8041
>>>
>>> ===========================================================================
>>> LogType: stderr
>>> LogLength: 4284
>>> Log Contents:
>>> Please use CMSClassUnloadingEnabled in place of
>>> CMSPermGenSweepingEnabled in the future
>>> Please use CMSClassUnloadingEnabled in place of
>>> CMSPermGenSweepingEnabled in the future
>>> 15/12/09 16:29:31 INFO yarn.ApplicationMaster: Registered signal
>>> handlers for [TERM, HUP, INT]
>>> 15/12/09 16:29:32 INFO yarn.ApplicationMaster: ApplicationAttemptId:
>>> appattempt_1448873753366_121022_000002
>>> 15/12/09 16:29:33 INFO spark.SecurityManager: Changing view acls to: mqq
>>> 15/12/09 16:29:33 INFO spark.SecurityManager: Changing modify acls to:
>>> mqq
>>> 15/12/09 16:29:33 INFO spark.SecurityManager: SecurityManager:
>>> authentication disabled; ui acls disabled; users with view permissions:
>>> Set(mqq); users with modify permissions: Set(mqq)
>>> 15/12/09 16:29:33 INFO yarn.ApplicationMaster: Starting the user
>>> application in a separate Thread
>>> 15/12/09 16:29:33 INFO yarn.ApplicationMaster: Waiting for spark context
>>> initialization
>>> 15/12/09 16:29:33 INFO yarn.ApplicationMaster: Waiting for spark context
>>> initialization ...
>>> 15/12/09 16:29:33 INFO client.RemoteDriver: Connecting to:
>>> 10.179.12.140:38842
>>> 15/12/09 16:29:33 WARN rpc.Rpc: Invalid log level null, reverting to
>>> default.
>>> 15/12/09 16:29:33 ERROR yarn.ApplicationMaster: User class threw
>>> exception: java.util.concurrent.ExecutionException:
>>> javax.security.sasl.SaslException: Client closed before SASL negotiation
>>> finished.
>>> java.util.concurrent.ExecutionException:
>>> javax.security.sasl.SaslException: Client closed before SASL negotiation
>>> finished.
>>>         at
>>> io.netty.util.concurrent.AbstractFuture.get(AbstractFuture.java:37)
>>>         at
>>> org.apache.hive.spark.client.RemoteDriver.<init>(RemoteDriver.java:156)
>>>         at
>>> org.apache.hive.spark.client.RemoteDriver.main(RemoteDriver.java:556)
>>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>         at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>>         at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>         at java.lang.reflect.Method.invoke(Method.java:606)
>>>         at
>>> org.apache.spark.deploy.yarn.ApplicationMaster$$anon$2.run(ApplicationMaster.scala:483)
>>> Caused by: javax.security.sasl.SaslException: Client closed before SASL
>>> negotiation finished.
>>>         at
>>> org.apache.hive.spark.client.rpc.Rpc$SaslClientHandler.dispose(Rpc.java:449)
>>>         at
>>> org.apache.hive.spark.client.rpc.SaslHandler.channelInactive(SaslHandler.java:90)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:233)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:219)
>>>         at
>>> io.netty.channel.ChannelInboundHandlerAdapter.channelInactive(ChannelInboundHandlerAdapter.java:75)
>>>         at
>>> org.apache.hive.spark.client.rpc.KryoMessageCodec.channelInactive(KryoMessageCodec.java:127)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:233)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:219)
>>>         at
>>> io.netty.channel.ChannelInboundHandlerAdapter.channelInactive(ChannelInboundHandlerAdapter.java:75)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.invokeChannelInactive(AbstractChannelHandlerContext.java:233)
>>>         at
>>> io.netty.channel.AbstractChannelHandlerContext.fireChannelInactive(AbstractChannelHandlerContext.java:219)
>>>         at
>>> io.netty.channel.DefaultChannelPipeline.fireChannelInactive(DefaultChannelPipeline.java:769)
>>>         at
>>> io.netty.channel.AbstractChannel$AbstractUnsafe$5.run(AbstractChannel.java:567)
>>>         at
>>> io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:380)
>>>         at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:357)
>>>         at
>>> io.netty.util.concurrent.SingleThreadEventExecutor$2.run(SingleThreadEventExecutor.java:116)
>>>         at java.lang.Thread.run(Thread.java:745)
>>> 15/12/09 16:29:33 INFO yarn.ApplicationMaster: Final app status: FAILED,
>>> exitCode: 15, (reason: User class threw exception:
>>> java.util.concurrent.ExecutionException: javax.security.sasl.SaslException:
>>> Client closed before SASL negotiation finished.)
>>> 15/12/09 16:29:43 ERROR yarn.ApplicationMaster: SparkContext did not
>>> initialize after waiting for 150000 ms. Please check earlier log output for
>>> errors. Failing the application.
>>> 15/12/09 16:29:43 INFO util.Utils: Shutdown hook called
>>>
>>> LogType: stdout
>>> LogLength: 216
>>> Log Contents:
>>> Java HotSpot(TM) 64-Bit Server VM warning: ignoring option
>>> UseCompressedStrings; support was removed in 7.0
>>> Java HotSpot(TM) 64-Bit Server VM warning: ignoring option
>>> UseCompressedStrings; support was removed in 7.0
>>>
>>>
>>> *I find submit cycle is base
>>> on hive.spark.client.server.connect.timeout,which is 120000ms in our team.*
>>> *And spark.yarn.am.waitTime is 150s.*
>>>
>>> *My question is *
>>> *1.Is a bug or design of submit more times?How do i set up the value of
>>> hive.spark.client.server.connect.timeout when the queue resources is not
>>> enough?*
>>> *2. Why the application will fast fail,caused by "SparkContext did not
>>> initialize after waiting for 150000 ms"?*
>>>
>>> *Thanks.*
>>> *Best wishes.*
>>>
>>>
>>
>>
>

Reply via email to