I'm guessing you might have conflicting versions of libthrift on your
classpath

On Tue, 1 Oct 2019, 08:44 Jeff Zhang, <zjf...@gmail.com> wrote:

> It looks like you are using pyspark, could you try just start scala spark
> interpreter via `%spark` ? First let's figure out whether it is related
> with pyspark.
>
>
>
> Manuel Sopena Ballesteros <manuel...@garvan.org.au> 于2019年10月1日周二
> 下午3:29写道:
>
>> Dear Zeppelin community,
>>
>>
>>
>> I would like to ask for advice in regards an error I am having with
>> thrift.
>>
>>
>>
>> I am getting quite a lot of these errors while running my notebooks
>>
>>
>>
>> org.apache.thrift.transport.TTransportException at
>> org.apache.thrift.transport.TIOStreamTransport.read(TIOStreamTransport.java:132)
>> at org.apache.thrift.transport.TTransport.readAll(TTransport.java:86) at
>> org.apache.thrift.protocol.TBinaryProtocol.readAll(TBinaryProtocol.java:429)
>> at
>> org.apache.thrift.protocol.TBinaryProtocol.readI32(TBinaryProtocol.java:318)
>> at
>> org.apache.thrift.protocol.TBinaryProtocol.readMessageBegin(TBinaryProtocol.java:219)
>> at org.apache.thrift.TServiceClient.receiveBase(TServiceClient.java:77) at
>> org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$Client.recv_interpret(RemoteInterpreterService.java:274)
>> at
>> org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$Client.interpret(RemoteInterpreterService.java:258)
>> at
>> org.apache.zeppelin.interpreter.remote.RemoteInterpreter$4.call(RemoteInterpreter.java:233)
>> at
>> org.apache.zeppelin.interpreter.remote.RemoteInterpreter$4.call(RemoteInterpreter.java:229)
>> at
>> org.apache.zeppelin.interpreter.remote.RemoteInterpreterProcess.callRemoteFunction(RemoteInterpreterProcess.java:135)
>> at
>> org.apache.zeppelin.interpreter.remote.RemoteInterpreter.interpret(RemoteInterpreter.java:228)
>> at org.apache.zeppelin.notebook.Paragraph.jobRun(Paragraph.java:437) at
>> org.apache.zeppelin.scheduler.Job.run(Job.java:188) at
>> org.apache.zeppelin.scheduler.RemoteScheduler$JobRunner.run(RemoteScheduler.java:307)
>> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>> at java.util.concurrent.FutureTask.run(FutureTask.java:266) at
>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>> at
>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>> at
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>> at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>> at java.lang.Thread.run(Thread.java:745)
>>
>>
>>
>> And this is the Spark driver application logs:
>>
>> …
>>
>>
>> ===============================================================================
>>
>> YARN executor launch context:
>>
>>   env:
>>
>>     CLASSPATH ->
>> {{PWD}}<CPS>{{PWD}}/__spark_conf__<CPS>{{PWD}}/__spark_libs__/*<CPS>$HADOOP_CONF_DIR<CPS>/usr/hdp/3.1.0.0-78/hadoop/*<CPS>/usr/hdp/3.1.0.0-78/hadoop/lib/*<CPS>/usr/hdp/current/hadoop-hdfs-client/*<CPS>/usr/hdp/current/hadoop-hdfs-client/lib/*<CPS>/usr/hdp/current/hadoop-yarn-client/*<CPS>/usr/hdp/current/hadoop-yarn-client/lib/*<CPS>$PWD/mr-framework/hadoop/share/hadoop/mapreduce/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/lib/*:$PWD/mr-framework/hadoop/share/hadoop/common/*:$PWD/mr-framework/hadoop/share/hadoop/common/lib/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/lib/*:$PWD/mr-framework/hadoop/share/hadoop/tools/lib/*:/usr/hdp/3.1.0.0-78/hadoop/lib/hadoop-lzo-0.6.0.3.1.0.0-78.jar:/etc/hadoop/conf/secure<CPS>{{PWD}}/__spark_conf__/__hadoop_conf__
>>
>>     SPARK_YARN_STAGING_DIR ->
>> hdfs://gl-hdp-ctrl01-mlx.mlx:8020/user/mansop/.sparkStaging/application_1568954689585_0052
>>
>>     SPARK_USER -> mansop
>>
>>     PYTHONPATH ->
>> /usr/hdp/current/spark2-client/python/lib/py4j-0.10.7-src.zip:/usr/hdp/current/spark2-client/python/:<CPS>{{PWD}}/pyspark.zip<CPS>{{PWD}}/py4j-0.10.7-src.zip
>>
>>
>>
>>   command:
>>
>>
>> LD_LIBRARY_PATH="/usr/hdp/current/hadoop-client/lib/native:/usr/hdp/current/hadoop-client/lib/native/Linux-amd64-64:$LD_LIBRARY_PATH"
>> \
>>
>>       {{JAVA_HOME}}/bin/java \
>>
>>       -server \
>>
>>       -Xmx1024m \
>>
>>       '-XX:+UseNUMA' \
>>
>>       -Djava.io.tmpdir={{PWD}}/tmp \
>>
>>       '-Dspark.history.ui.port=18081' \
>>
>>       -Dspark.yarn.app.container.log.dir=<LOG_DIR> \
>>
>>       -XX:OnOutOfMemoryError='kill %p' \
>>
>>       org.apache.spark.executor.CoarseGrainedExecutorBackend \
>>
>>       --driver-url \
>>
>>       spark://coarsegrainedschedu...@r640-1-12-mlx.mlx:35602 \
>>
>>       --executor-id \
>>
>>       <executorId> \
>>
>>       --hostname \
>>
>>       <hostname> \
>>
>>       --cores \
>>
>>       1 \
>>
>>       --app-id \
>>
>>       application_1568954689585_0052 \
>>
>>       --user-class-path \
>>
>>       file:$PWD/__app__.jar \
>>
>>       1><LOG_DIR>/stdout \
>>
>>       2><LOG_DIR>/stderr
>>
>>
>>
>>   resources:
>>
>>     __app__.jar -> resource { scheme: "hdfs" host:
>> "gl-hdp-ctrl01-mlx.mlx" port: 8020 file:
>> "/user/mansop/.sparkStaging/application_1568954689585_0052/spark-interpreter-0.8.0.3.1.0.0-78.jar"
>> } size: 20433040 timestamp: 1569804142906 type: FILE visibility: PRIVATE
>>
>>     __spark_conf__ -> resource { scheme: "hdfs" host:
>> "gl-hdp-ctrl01-mlx.mlx" port: 8020 file:
>> "/user/mansop/.sparkStaging/application_1568954689585_0052/__spark_conf__.zip"
>> } size: 277725 timestamp: 1569804143239 type: ARCHIVE visibility: PRIVATE
>>
>>     sparkr -> resource { scheme: "hdfs" host: "gl-hdp-ctrl01-mlx.mlx"
>> port: 8020 file:
>> "/user/mansop/.sparkStaging/application_1568954689585_0052/sparkr.zip" }
>> size: 688255 timestamp: 1569804142991 type: ARCHIVE visibility: PRIVATE
>>
>>     log4j_yarn_cluster.properties -> resource { scheme: "hdfs" host:
>> "gl-hdp-ctrl01-mlx.mlx" port: 8020 file:
>> "/user/mansop/.sparkStaging/application_1568954689585_0052/log4j_yarn_cluster.properties"
>> } size: 1018 timestamp: 1569804142955 type: FILE visibility: PRIVATE
>>
>>     pyspark.zip -> resource { scheme: "hdfs" host:
>> "gl-hdp-ctrl01-mlx.mlx" port: 8020 file:
>> "/user/mansop/.sparkStaging/application_1568954689585_0052/pyspark.zip" }
>> size: 550570 timestamp: 1569804143018 type: FILE visibility: PRIVATE
>>
>>     __spark_libs__ -> resource { scheme: "hdfs" host:
>> "gl-hdp-ctrl01-mlx.mlx" port: 8020 file:
>> "/hdp/apps/3.1.0.0-78/spark2/spark2-hdp-yarn-archive.tar.gz" } size:
>> 280293050 timestamp: 1568938921259 type: ARCHIVE visibility: PUBLIC
>>
>>     py4j-0.10.7-src.zip -> resource { scheme: "hdfs" host:
>> "gl-hdp-ctrl01-mlx.mlx" port: 8020 file:
>> "/user/mansop/.sparkStaging/application_1568954689585_0052/py4j-0.10.7-src.zip"
>> } size: 42437 timestamp: 1569804143043 type: FILE visibility: PRIVATE
>>
>>     __hive_libs__ -> resource { scheme: "hdfs" host:
>> "gl-hdp-ctrl01-mlx.mlx" port: 8020 file:
>> "/hdp/apps/3.1.0.0-78/spark2/spark2-hdp-hive-archive.tar.gz" } size:
>> 43807162 timestamp: 1568938925069 type: ARCHIVE visibility: PUBLIC
>>
>>
>>
>>
>> ===============================================================================
>>
>> INFO [2019-09-30 10:42:37,303] ({main}
>> RMProxy.java[newProxyInstance]:133) - Connecting to ResourceManager at
>> gl-hdp-ctrl03-mlx.mlx/10.0.1.248:8030
>>
>> INFO [2019-09-30 10:42:37,324] ({main} Logging.scala[logInfo]:54) -
>> Registering the ApplicationMaster
>>
>> INFO [2019-09-30 10:42:37,454] ({main}
>> Configuration.java[getConfResourceAsInputStream]:2756) - found resource
>> resource-types.xml at file:/etc/hadoop/3.1.0.0-78/0/resource-types.xml
>>
>> INFO [2019-09-30 10:42:37,470] ({main} Logging.scala[logInfo]:54) - Will
>> request 2 executor container(s), each with 1 core(s) and 1408 MB memory
>> (including 384 MB of overhead)
>>
>> INFO [2019-09-30 10:42:37,474] ({dispatcher-event-loop-14}
>> Logging.scala[logInfo]:54) - ApplicationMaster registered as
>> NettyRpcEndpointRef(spark://yar...@r640-1-12-mlx.mlx:35602)
>>
>> INFO [2019-09-30 10:42:37,485] ({main} Logging.scala[logInfo]:54) -
>> Submitted 2 unlocalized container requests.
>>
>> INFO [2019-09-30 10:42:37,518] ({main} Logging.scala[logInfo]:54) -
>> Started progress reporter thread with (heartbeat : 3000, initial allocation
>> : 200) intervals
>>
>> INFO [2019-09-30 10:42:37,619] ({Reporter} Logging.scala[logInfo]:54) -
>> Launching container container_e01_1568954689585_0052_01_000002 on host
>> r640-1-12-mlx.mlx for executor with ID 1
>>
>> INFO [2019-09-30 10:42:37,621] ({Reporter} Logging.scala[logInfo]:54) -
>> Launching container container_e01_1568954689585_0052_01_000003 on host
>> r640-1-13-mlx.mlx for executor with ID 2
>>
>> INFO [2019-09-30 10:42:37,623] ({Reporter} Logging.scala[logInfo]:54) -
>> Received 2 containers from YARN, launching executors on 2 of them.
>>
>> INFO [2019-09-30 10:42:39,481] ({dispatcher-event-loop-51}
>> Logging.scala[logInfo]:54) - Registered executor
>> NettyRpcEndpointRef(spark-client://Executor) (10.0.1.12:54340) with ID 1
>>
>> INFO [2019-09-30 10:42:39,553] ({dispatcher-event-loop-62}
>> Logging.scala[logInfo]:54) - Registering block manager
>> r640-1-12-mlx.mlx:33043 with 408.9 MB RAM, BlockManagerId(1,
>> r640-1-12-mlx.mlx, 33043, None)
>>
>> INFO [2019-09-30 10:42:40,003] ({dispatcher-event-loop-9}
>> Logging.scala[logInfo]:54) - Registered executor
>> NettyRpcEndpointRef(spark-client://Executor) (10.0.1.13:33812) with ID 2
>>
>> INFO [2019-09-30 10:42:40,023] ({pool-6-thread-2}
>> Logging.scala[logInfo]:54) - SchedulerBackend is ready for scheduling
>> beginning after reached minRegisteredResourcesRatio: 0.8
>>
>> INFO [2019-09-30 10:42:40,025] ({pool-6-thread-2}
>> Logging.scala[logInfo]:54) - YarnClusterScheduler.postStartHook done
>>
>> INFO [2019-09-30 10:42:40,072] ({dispatcher-event-loop-11}
>> Logging.scala[logInfo]:54) - Registering block manager
>> r640-1-13-mlx.mlx:34105 with 408.9 MB RAM, BlockManagerId(2,
>> r640-1-13-mlx.mlx, 34105, None)
>>
>> INFO [2019-09-30 10:42:41,779] ({pool-6-thread-2}
>> SparkShims.java[loadShims]:54) - Initializing shims for Spark 2.x
>>
>> INFO [2019-09-30 10:42:41,840] ({pool-6-thread-2}
>> Py4JUtils.java[createGatewayServer]:44) - Launching GatewayServer at
>> 127.0.0.1:36897
>>
>> INFO [2019-09-30 10:42:41,852] ({pool-6-thread-2}
>> PySparkInterpreter.java[createGatewayServerAndStartScript]:265) -
>> pythonExec: /home/mansop/anaconda2/bin/python
>>
>> INFO [2019-09-30 10:42:41,862] ({pool-6-thread-2}
>> PySparkInterpreter.java[setupPySparkEnv]:236) - PYTHONPATH:
>> /usr/hdp/current/spark2-client/python/lib/py4j-0.10.7-src.zip:/usr/hdp/current/spark2-client/python/::/d1/hadoop/yarn/local/usercache/mansop/appcache/application_1568954689585_0052/container_e01_1568954689585_0052_01_000001/pyspark.zip:/d1/hadoop/yarn/local/usercache/mansop/appcache/application_1568954689585_0052/container_e01_1568954689585_0052_01_000001/py4j-0.10.7-src.zip
>>
>> ERROR [2019-09-30 10:43:09,061] ({SIGTERM handler}
>> SignalUtils.scala[apply$mcZ$sp]:43) - RECEIVED SIGNAL TERM
>>
>> INFO [2019-09-30 10:43:09,068] ({shutdown-hook-0}
>> Logging.scala[logInfo]:54) - Invoking stop() from shutdown hook
>>
>> INFO [2019-09-30 10:43:09,082] ({shutdown-hook-0}
>> AbstractConnector.java[doStop]:318) - Stopped Spark@505439b3
>> {HTTP/1.1,[http/1.1]}{0.0.0.0:0}
>>
>> INFO [2019-09-30 10:43:09,085] ({shutdown-hook-0}
>> Logging.scala[logInfo]:54) - Stopped Spark web UI at
>> http://r640-1-12-mlx.mlx:42446
>>
>> INFO [2019-09-30 10:43:09,140] ({dispatcher-event-loop-52}
>> Logging.scala[logInfo]:54) - Driver requested a total number of 0
>> executor(s).
>>
>> INFO [2019-09-30 10:43:09,142] ({shutdown-hook-0}
>> Logging.scala[logInfo]:54) - Shutting down all executors
>>
>> INFO [2019-09-30 10:43:09,144] ({dispatcher-event-loop-51}
>> Logging.scala[logInfo]:54) - Asking each executor to shut down
>>
>> INFO [2019-09-30 10:43:09,151] ({shutdown-hook-0}
>> Logging.scala[logInfo]:54) - Stopping SchedulerExtensionServices
>>
>> (serviceOption=None,
>>
>> services=List(),
>>
>> started=false)
>>
>> ERROR [2019-09-30 10:43:09,155] ({Reporter} Logging.scala[logError]:91) -
>> Exception from Reporter thread.
>>
>> org.apache.hadoop.yarn.exceptions.ApplicationAttemptNotFoundException:
>> Application attempt appattempt_1568954689585_0052_000001 doesn't exist in
>> ApplicationMasterService cache.
>>
>>                at
>> org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:404)
>>
>>                at
>> org.apache.hadoop.yarn.api.impl.pb.service.ApplicationMasterProtocolPBServiceImpl.allocate(ApplicationMasterProtocolPBServiceImpl.java:60)
>>
>>                at
>> org.apache.hadoop.yarn.proto.ApplicationMasterProtocol$ApplicationMasterProtocolService$2.callBlockingMethod(ApplicationMasterProtocol.java:99)
>>
>>                at
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:524)
>>
>>                at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1025)
>>
>>                at
>> org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:876)
>>
>>                at
>> org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:822)
>>
>>                at java.security.AccessController.doPrivileged(Native
>> Method)
>>
>>                at javax.security.auth.Subject.doAs(Subject.java:422)
>>
>>                at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1730)
>>
>>                at
>> org.apache.hadoop.ipc.Server$Handler.run(Server.java:2682)
>>
>>
>>
>>                at
>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>>
>>                at
>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>>
>>                at
>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>>
>>                at
>> java.lang.reflect.Constructor.newInstance(Constructor.java:423)
>>
>>                at
>> org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53)
>>
>>                at
>> org.apache.hadoop.yarn.ipc.RPCUtil.instantiateYarnException(RPCUtil.java:75)
>>
>>                at
>> org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:116)
>>
>>                at
>> org.apache.hadoop.yarn.api.impl.pb.client.ApplicationMasterProtocolPBClientImpl.allocate(ApplicationMasterProtocolPBClientImpl.java:79)
>>
>>                at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>> Method)
>>
>>                at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>
>>                at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>
>>                at java.lang.reflect.Method.invoke(Method.java:498)
>>
>>                at
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422)
>>
>>                at
>> org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165)
>>
>>                at
>> org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157)
>>
>>                at
>> org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
>>
>>                at
>> org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359)
>>
>>                at com.sun.proxy.$Proxy21.allocate(Unknown Source)
>>
>>                at
>> org.apache.hadoop.yarn.client.api.impl.AMRMClientImpl.allocate(AMRMClientImpl.java:320)
>>
>>                at
>> org.apache.spark.deploy.yarn.YarnAllocator.allocateResources(YarnAllocator.scala:268)
>>
>>                at
>> org.apache.spark.deploy.yarn.ApplicationMaster$$anon$3.run(ApplicationMaster.scala:556)
>>
>> Caused by:
>> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.yarn.exceptions.ApplicationAttemptNotFoundException):
>> Application attempt appattempt_1568954689585_0052_000001 doesn't exist in
>> ApplicationMasterService cache.
>>
>>                at
>> org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:404)
>>
>>                at
>> org.apache.hadoop.yarn.api.impl.pb.service.ApplicationMasterProtocolPBServiceImpl.allocate(ApplicationMasterProtocolPBServiceImpl.java:60)
>>
>>                at
>> org.apache.hadoop.yarn.proto.ApplicationMasterProtocol$ApplicationMasterProtocolService$2.callBlockingMethod(ApplicationMasterProtocol.java:99)
>>
>>                at
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:524)
>>
>>                at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1025)
>>
>>                at
>> org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:876)
>>
>>                at
>> org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:822)
>>
>>                at java.security.AccessController.doPrivileged(Native
>> Method)
>>
>>                at javax.security.auth.Subject.doAs(Subject.java:422)
>>
>>                at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1730)
>>
>>                at
>> org.apache.hadoop.ipc.Server$Handler.run(Server.java:2682)
>>
>>
>>
>>                at
>> org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1497)
>>
>>                at org.apache.hadoop.ipc.Client.call(Client.java:1443)
>>
>>                at org.apache.hadoop.ipc.Client.call(Client.java:1353)
>>
>>                at
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:228)
>>
>>                at
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:116)
>>
>>                at com.sun.proxy.$Proxy20.allocate(Unknown Source)
>>
>>                at
>> org.apache.hadoop.yarn.api.impl.pb.client.ApplicationMasterProtocolPBClientImpl.allocate(ApplicationMasterProtocolPBClientImpl.java:77)
>>
>>                ... 13 more
>>
>> INFO [2019-09-30 10:43:09,164] ({Reporter} Logging.scala[logInfo]:54) -
>> Final app status: FAILED, exitCode: 12, (reason: Application attempt
>> appattempt_1568954689585_0052_000001 doesn't exist in
>> ApplicationMasterService cache.
>>
>>                at
>> org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:404)
>>
>>                at
>> org.apache.hadoop.yarn.api.impl.pb.service.ApplicationMasterProtocolPBServiceImpl.allocate(ApplicationMasterProtocolPBServiceImpl.java:60)
>>
>>                at
>> org.apache.hadoop.yarn.proto.ApplicationMasterProtocol$ApplicationMasterProtocolService$2.callBlockingMethod(ApplicationMasterProtocol.java:99)
>>
>>                at
>> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:524)
>>
>>                at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1025)
>>
>>                at
>> org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:876)
>>
>>                at
>> org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:822)
>>
>>                at java.security.AccessController.doPrivileged(Native
>> Method)
>>
>>                at javax.security.auth.Subject.doAs(Subject.java:422)
>>
>>                at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1730)
>>
>>                at
>> org.apache.hadoop.ipc.Server$Handler.run(Server.java:2682)
>>
>> )
>>
>> INFO [2019-09-30 10:43:09,166] ({dispatcher-event-loop-54}
>> Logging.scala[logInfo]:54) - MapOutputTrackerMasterEndpoint stopped!
>>
>> INFO [2019-09-30 10:43:09,236] ({shutdown-hook-0}
>> Logging.scala[logInfo]:54) - MemoryStore cleared
>>
>> INFO [2019-09-30 10:43:09,237] ({shutdown-hook-0}
>> Logging.scala[logInfo]:54) - BlockManager stopped
>>
>> INFO [2019-09-30 10:43:09,237] ({shutdown-hook-0}
>> Logging.scala[logInfo]:54) - BlockManagerMaster stopped
>>
>> INFO [2019-09-30 10:43:09,241] ({dispatcher-event-loop-73}
>> Logging.scala[logInfo]:54) - OutputCommitCoordinator stopped!
>>
>> INFO [2019-09-30 10:43:09,252] ({shutdown-hook-0}
>> Logging.scala[logInfo]:54) - Successfully stopped SparkContext
>>
>> INFO [2019-09-30 10:43:09,253] ({shutdown-hook-0}
>> Logging.scala[logInfo]:54) - Shutdown hook called
>>
>> INFO [2019-09-30 10:43:09,254] ({shutdown-hook-0}
>> Logging.scala[logInfo]:54) - Deleting directory
>> /d1/hadoop/yarn/local/usercache/mansop/appcache/application_1568954689585_0052/spark-ba80cda3-812a-4cf0-b1f6-6e9eb52952b2
>>
>> INFO [2019-09-30 10:43:09,254] ({shutdown-hook-0}
>> Logging.scala[logInfo]:54) - Deleting directory
>> /d0/hadoop/yarn/local/usercache/mansop/appcache/application_1568954689585_0052/spark-43078781-8f1c-4cd6-a8da-e81b32892cf8
>>
>> INFO [2019-09-30 10:43:09,255] ({shutdown-hook-0}
>> Logging.scala[logInfo]:54) - Deleting directory
>> /d0/hadoop/yarn/local/usercache/mansop/appcache/application_1568954689585_0052/spark-43078781-8f1c-4cd6-a8da-e81b32892cf8/pyspark-9138f7ad-3f15-42c6-9bf3-e3e72d5d4086
>>
>>
>>
>> How can I continue troubleshooting in order to find out what this error
>> means?
>>
>>
>>
>> Thank you very much
>>
>>
>> NOTICE
>> Please consider the environment before printing this email. This message
>> and any attachments are intended for the addressee named and may contain
>> legally privileged/confidential/copyright information. If you are not the
>> intended recipient, you should not read, use, disclose, copy or distribute
>> this communication. If you have received this message in error please
>> notify us at once by return email and then delete both messages. We accept
>> no liability for the distribution of viruses or similar in electronic
>> communications. This notice should not be removed.
>>
>
>
> --
> Best Regards
>
> Jeff Zhang
>

Reply via email to