LD_LIBRARY_PATH points to /opt/eTOX_spark/lib/ the location of the nattive
libraries.


Oriol.

2015-11-25 17:39 GMT+01:00 Ted Yu <yuzhih...@gmail.com>:

> In your spark-env, did you set LD_LIBRARY_PATH ?
>
> Cheers
>
> On Wed, Nov 25, 2015 at 7:32 AM, Oriol López Massaguer <
> oriol.lo...@gmail.com> wrote:
>
>> Hello;
>>
>> I'm trying to use a native library in Spark.
>>
>> I was using a simple standalone cluster with one master and worker.
>>
>> According to the documentation I edited the spark-defautls.conf by
>> setting:
>>
>> spark.driver.extraClassPath=/opt/eTOX_spark/lib/org.RDKit.jar
>> spark.driver.extraLibraryPath=/opt/eTOX_spark/lib/
>> spark.executor.extraLibraryPath=/opt/eTOX_spark/lib/
>>
>> In the path /opt/eTOX_spark/lib/ there are 3 so files wich are wrapped in
>> org.RDKit.jar.
>>
>> But when I try so submit a job that uses the native library I get:
>>
>> Exception in thread "main" java.lang.UnsatisfiedLinkError:
>> org.RDKit.RDKFuncsJNI.RWMol_MolFromSmiles__SWIG_3(Ljava/lang/String;)J
>>     at org.RDKit.RDKFuncsJNI.RWMol_MolFromSmiles__SWIG_3(Native Method)
>>     at org.RDKit.RWMol.MolFromSmiles(RWMol.java:426)
>>     at models.spark.sources.eTOX_DB$.main(eTOX.scala:54)
>>     at models.spark.sources.eTOX_DB.main(eTOX.scala)
>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>     at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>     at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>     at java.lang.reflect.Method.invoke(Method.java:497)
>>     at
>> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:727)
>>     at
>> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
>>     at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
>>     at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
>>     at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>>
>> I use the submit.sh with the following parameters:
>>
>>  /opt/spark/bin/spark-submit --verbose --class
>> "models.spark.sources.eTOX_DB"  --master
>> spark://localhost.localdomain:7077
>> target/scala-2.10/etox_spark_2.10-1.0.jar
>>
>> the full output is:
>>
>> Using properties file: /opt/spark/conf/spark-defaults.conf
>> Adding default property:
>> spark.driver.extraLibraryPath=/opt/eTOX_spark/lib/
>> Adding default property:
>> spark.driver.extraClassPath=/opt/eTOX_spark/lib/org.RDKit.jar
>> Adding default property:
>> spark.executor.extraLibraryPath=/opt/eTOX_spark/lib/
>> Parsed arguments:
>>   master                  spark://localhost.localdomain:7077
>>   deployMode              null
>>   executorMemory          null
>>   executorCores           null
>>   totalExecutorCores      null
>>   propertiesFile          /opt/spark/conf/spark-defaults.conf
>>   driverMemory            null
>>   driverCores             null
>>   driverExtraClassPath    /opt/eTOX_spark/lib/org.RDKit.jar
>>   driverExtraLibraryPath  /opt/eTOX_spark/lib/
>>   driverExtraJavaOptions  null
>>   supervise               false
>>   queue                   null
>>   numExecutors            null
>>   files                   null
>>   pyFiles                 null
>>   archives                null
>>   mainClass               models.spark.sources.eTOX_DB
>>   primaryResource
>> file:/opt/eTOX_spark/target/scala-2.10/etox_spark_2.10-1.0.jar
>>   name                    models.spark.sources.eTOX_DB
>>   childArgs               []
>>   jars                    null
>>   packages                null
>>   packagesExclusions      null
>>   repositories            null
>>   verbose                 true
>>
>> Spark properties used, including those specified through
>>  --conf and those from the properties file
>> /opt/spark/conf/spark-defaults.conf:
>>   spark.executor.extraLibraryPath -> /opt/eTOX_spark/lib/
>>   spark.driver.extraLibraryPath -> /opt/eTOX_spark/lib/
>>   spark.driver.extraClassPath -> /opt/eTOX_spark/lib/org.RDKit.jar
>>
>>
>> Main class:
>> models.spark.sources.eTOX_DB
>> Arguments:
>>
>> System properties:
>> spark.executor.extraLibraryPath -> /opt/eTOX_spark/lib/
>> spark.driver.extraLibraryPath -> /opt/eTOX_spark/lib/
>> SPARK_SUBMIT -> true
>> spark.app.name -> models.spark.sources.eTOX_DB
>> spark.jars ->
>> file:/opt/eTOX_spark/target/scala-2.10/etox_spark_2.10-1.0.jar
>> spark.submit.deployMode -> client
>> spark.master -> spark://localhost.localdomain:7077
>> spark.driver.extraClassPath -> /opt/eTOX_spark/lib/org.RDKit.jar
>> Classpath elements:
>> file:/opt/eTOX_spark/target/scala-2.10/etox_spark_2.10-1.0.jar
>>
>>
>> Buffer(/opt/jdk1.8.0_45/jre/lib/amd64/libzip.so)
>> Loading libraries
>> Buffer(/opt/jdk1.8.0_45/jre/lib/amd64/libzip.so, /opt/eTOX_spark/lib/
>> libboost_thread.1.48.0.so, /opt/eTOX_spark/lib/libboost_system.1.48.0.so,
>> /opt/eTOX_spark/lib/libGraphMolWrap.so)
>> Loading libraries
>> Using Spark's default log4j profile:
>> org/apache/spark/log4j-defaults.properties
>> 15/11/25 16:27:32 INFO SparkContext: Running Spark version 1.6.0-SNAPSHOT
>> 15/11/25 16:27:33 WARN NativeCodeLoader: Unable to load native-hadoop
>> library for your platform... using builtin-java classes where applicable
>> 15/11/25 16:27:33 WARN Utils: Your hostname, localhost.localdomain
>> resolves to a loopback address: 127.0.0.1; using 10.0.2.15 instead (on
>> interface enp0s3)
>> 15/11/25 16:27:33 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to
>> another address
>> 15/11/25 16:27:33 INFO SecurityManager: Changing view acls to: user
>> 15/11/25 16:27:33 INFO SecurityManager: Changing modify acls to: user
>> 15/11/25 16:27:33 INFO SecurityManager: SecurityManager: authentication
>> disabled; ui acls disabled; users with view permissions: Set(user); users
>> with modify permissions: Set(user)
>> 15/11/25 16:27:34 INFO Utils: Successfully started service 'sparkDriver'
>> on port 35799.
>> 15/11/25 16:27:34 INFO Slf4jLogger: Slf4jLogger started
>> 15/11/25 16:27:34 INFO Remoting: Starting remoting
>> 15/11/25 16:27:34 INFO Remoting: Remoting started; listening on addresses
>> :[akka.tcp://sparkDriverActorSystem@10.0.2.15:48344]
>> 15/11/25 16:27:34 INFO Utils: Successfully started service
>> 'sparkDriverActorSystem' on port 48344.
>> 15/11/25 16:27:34 INFO SparkEnv: Registering MapOutputTracker
>> 15/11/25 16:27:34 INFO SparkEnv: Registering BlockManagerMaster
>> 15/11/25 16:27:34 INFO DiskBlockManager: Created local directory at
>> /tmp/blockmgr-ffbf1759-2e79-4ecd-a0d8-5d7f28d3c132
>> 15/11/25 16:27:34 INFO MemoryStore: MemoryStore started with capacity
>> 736.1 MB
>> 15/11/25 16:27:34 INFO HttpFileServer: HTTP File server directory is
>> /tmp/spark-3f269496-c3aa-4bbc-8955-1ab88a179420/httpd-4ea1a9f9-2ce6-41b3-82ef-96e9094f5aa4
>> 15/11/25 16:27:34 INFO HttpServer: Starting HTTP Server
>> 15/11/25 16:27:35 INFO Utils: Successfully started service 'HTTP file
>> server' on port 50803.
>> 15/11/25 16:27:35 INFO SparkEnv: Registering OutputCommitCoordinator
>> 15/11/25 16:27:35 INFO Utils: Successfully started service 'SparkUI' on
>> port 4040.
>> 15/11/25 16:27:35 INFO SparkUI: Started SparkUI at http://10.0.2.15:4040
>> 15/11/25 16:27:35 INFO SparkContext: Added JAR
>> file:/opt/eTOX_spark/target/scala-2.10/etox_spark_2.10-1.0.jar at
>> http://10.0.2.15:50803/jars/etox_spark_2.10-1.0.jar with timestamp
>> 1448465255275
>> 15/11/25 16:27:35 INFO AppClient$ClientEndpoint: Connecting to master
>> spark://localhost.localdomain:7077...
>> 15/11/25 16:27:35 INFO SparkDeploySchedulerBackend: Connected to Spark
>> cluster with app ID app-20151125162735-0003
>> 15/11/25 16:27:35 INFO AppClient$ClientEndpoint: Executor added:
>> app-20151125162735-0003/0 on worker-20151125161159-10.0.2.15-48408 (
>> 10.0.2.15:48408) with 4 cores
>> 15/11/25 16:27:35 INFO SparkDeploySchedulerBackend: Granted executor ID
>> app-20151125162735-0003/0 on hostPort 10.0.2.15:48408 with 4 cores,
>> 1024.0 MB RAM
>> 15/11/25 16:27:35 INFO Utils: Successfully started service
>> 'org.apache.spark.network.netty.NettyBlockTransferService' on port 35769.
>> 15/11/25 16:27:35 INFO NettyBlockTransferService: Server created on 35769
>> 15/11/25 16:27:35 INFO BlockManagerMaster: Trying to register BlockManager
>> 15/11/25 16:27:35 INFO BlockManagerMasterEndpoint: Registering block
>> manager 10.0.2.15:35769 with 736.1 MB RAM, BlockManagerId(driver,
>> 10.0.2.15, 35769)
>> 15/11/25 16:27:35 INFO AppClient$ClientEndpoint: Executor updated:
>> app-20151125162735-0003/0 is now LOADING
>> 15/11/25 16:27:35 INFO AppClient$ClientEndpoint: Executor updated:
>> app-20151125162735-0003/0 is now RUNNING
>> 15/11/25 16:27:35 INFO BlockManagerMaster: Registered BlockManager
>> 15/11/25 16:27:35 INFO SparkDeploySchedulerBackend: SchedulerBackend is
>> ready for scheduling beginning after reached minRegisteredResourcesRatio:
>> 0.0
>> 15/11/25 16:27:36 INFO ParquetRelation: Listing
>> file:/opt/etox_reports_2015_1/spark_data/etox/allfindings.parquet on driver
>> 15/11/25 16:27:37 INFO SparkContext: Starting job: load at eTOX.scala:52
>> 15/11/25 16:27:37 INFO DAGScheduler: Got job 0 (load at eTOX.scala:52)
>> with 2 output partitions
>> 15/11/25 16:27:37 INFO DAGScheduler: Final stage: ResultStage 0 (load at
>> eTOX.scala:52)
>> 15/11/25 16:27:37 INFO DAGScheduler: Parents of final stage: List()
>> 15/11/25 16:27:37 INFO DAGScheduler: Missing parents: List()
>> 15/11/25 16:27:37 INFO DAGScheduler: Submitting ResultStage 0
>> (MapPartitionsRDD[1] at load at eTOX.scala:52), which has no missing parents
>> 15/11/25 16:27:37 INFO MemoryStore: Ensuring 1048576 bytes of free space
>> for block broadcast_0(free: 771883008, max: 771883008)
>> 15/11/25 16:27:37 INFO MemoryStore: Ensuring 63000 bytes of free space
>> for block broadcast_0(free: 771883008, max: 771883008)
>> 15/11/25 16:27:37 INFO MemoryStore: Block broadcast_0 stored as values in
>> memory (estimated size 61.5 KB, free 61.5 KB)
>> 15/11/25 16:27:37 INFO MemoryStore: Ensuring 21090 bytes of free space
>> for block broadcast_0_piece0(free: 771820008, max: 771883008)
>> 15/11/25 16:27:37 INFO MemoryStore: Block broadcast_0_piece0 stored as
>> bytes in memory (estimated size 20.6 KB, free 82.1 KB)
>> 15/11/25 16:27:37 INFO BlockManagerInfo: Added broadcast_0_piece0 in
>> memory on 10.0.2.15:35769 (size: 20.6 KB, free: 736.1 MB)
>> 15/11/25 16:27:37 INFO SparkContext: Created broadcast 0 from broadcast
>> at DAGScheduler.scala:1003
>> 15/11/25 16:27:37 INFO DAGScheduler: Submitting 2 missing tasks from
>> ResultStage 0 (MapPartitionsRDD[1] at load at eTOX.scala:52)
>> 15/11/25 16:27:37 INFO TaskSchedulerImpl: Adding task set 0.0 with 2 tasks
>> 15/11/25 16:27:39 INFO SparkDeploySchedulerBackend: Registered executor
>> NettyRpcEndpointRef(null) (10.0.2.15:50472) with ID 0
>> 15/11/25 16:27:39 INFO TaskSetManager: Starting task 0.0 in stage 0.0
>> (TID 0, 10.0.2.15, partition 0,PROCESS_LOCAL, 2137 bytes)
>> 15/11/25 16:27:39 INFO TaskSetManager: Starting task 1.0 in stage 0.0
>> (TID 1, 10.0.2.15, partition 1,PROCESS_LOCAL, 2285 bytes)
>> 15/11/25 16:27:39 INFO BlockManagerMasterEndpoint: Registering block
>> manager 10.0.2.15:39373 with 736.1 MB RAM, BlockManagerId(0, 10.0.2.15,
>> 39373)
>> 15/11/25 16:27:39 INFO BlockManagerInfo: Added broadcast_0_piece0 in
>> memory on 10.0.2.15:39373 (size: 20.6 KB, free: 736.1 MB)
>> 15/11/25 16:27:40 INFO TaskSetManager: Finished task 0.0 in stage 0.0
>> (TID 0) in 1097 ms on 10.0.2.15 (1/2)
>> 15/11/25 16:27:42 INFO TaskSetManager: Finished task 1.0 in stage 0.0
>> (TID 1) in 3050 ms on 10.0.2.15 (2/2)
>> 15/11/25 16:27:42 INFO DAGScheduler: ResultStage 0 (load at
>> eTOX.scala:52) finished in 4.739 s
>> 15/11/25 16:27:42 INFO TaskSchedulerImpl: Removed TaskSet 0.0, whose
>> tasks have all completed, from pool
>> 15/11/25 16:27:42 INFO DAGScheduler: Job 0 finished: load at
>> eTOX.scala:52, took 5.157662 s
>> 15/11/25 16:27:42 INFO MemoryStore: Ensuring 1048576 bytes of free space
>> for block broadcast_1(free: 771798918, max: 771883008)
>> 15/11/25 16:27:42 INFO MemoryStore: Ensuring 63280 bytes of free space
>> for block broadcast_1(free: 771798918, max: 771883008)
>> 15/11/25 16:27:42 INFO MemoryStore: Block broadcast_1 stored as values in
>> memory (estimated size 61.8 KB, free 143.9 KB)
>> 15/11/25 16:27:42 INFO MemoryStore: Ensuring 19788 bytes of free space
>> for block broadcast_1_piece0(free: 771735638, max: 771883008)
>> 15/11/25 16:27:42 INFO MemoryStore: Block broadcast_1_piece0 stored as
>> bytes in memory (estimated size 19.3 KB, free 163.2 KB)
>> 15/11/25 16:27:42 INFO BlockManagerInfo: Added broadcast_1_piece0 in
>> memory on 10.0.2.15:35769 (size: 19.3 KB, free: 736.1 MB)
>> 15/11/25 16:27:42 INFO SparkContext: Created broadcast 1 from count at
>> eTOX.scala:53
>> 15/11/25 16:27:42 INFO deprecation: mapred.min.split.size is deprecated.
>> Instead, use mapreduce.input.fileinputformat.split.minsize
>> 15/11/25 16:27:43 INFO ParquetRelation: Reading Parquet file(s) from
>> file:/opt/etox_reports_2015_1/spark_data/etox/allfindings.parquet/part-r-00000-a0018ffe-9aaf-4446-a718-79097835d08d.gz.parquet
>> 15/11/25 16:27:43 INFO SparkContext: Starting job: count at eTOX.scala:53
>> 15/11/25 16:27:43 INFO DAGScheduler: Registering RDD 4 (count at
>> eTOX.scala:53)
>> 15/11/25 16:27:43 INFO DAGScheduler: Got job 1 (count at eTOX.scala:53)
>> with 1 output partitions
>> 15/11/25 16:27:43 INFO DAGScheduler: Final stage: ResultStage 2 (count at
>> eTOX.scala:53)
>> 15/11/25 16:27:43 INFO DAGScheduler: Parents of final stage:
>> List(ShuffleMapStage 1)
>> 15/11/25 16:27:43 INFO DAGScheduler: Missing parents:
>> List(ShuffleMapStage 1)
>> 15/11/25 16:27:43 INFO DAGScheduler: Submitting ShuffleMapStage 1
>> (MapPartitionsRDD[4] at count at eTOX.scala:53), which has no missing
>> parents
>> 15/11/25 16:27:43 INFO MemoryStore: Ensuring 1048576 bytes of free space
>> for block broadcast_2(free: 771715850, max: 771883008)
>> 15/11/25 16:27:43 INFO MemoryStore: Ensuring 12840 bytes of free space
>> for block broadcast_2(free: 771715850, max: 771883008)
>> 15/11/25 16:27:43 INFO MemoryStore: Block broadcast_2 stored as values in
>> memory (estimated size 12.5 KB, free 175.8 KB)
>> 15/11/25 16:27:43 INFO MemoryStore: Ensuring 6175 bytes of free space for
>> block broadcast_2_piece0(free: 771703010, max: 771883008)
>> 15/11/25 16:27:43 INFO MemoryStore: Block broadcast_2_piece0 stored as
>> bytes in memory (estimated size 6.0 KB, free 181.8 KB)
>> 15/11/25 16:27:43 INFO BlockManagerInfo: Added broadcast_2_piece0 in
>> memory on 10.0.2.15:35769 (size: 6.0 KB, free: 736.1 MB)
>> 15/11/25 16:27:43 INFO SparkContext: Created broadcast 2 from broadcast
>> at DAGScheduler.scala:1003
>> 15/11/25 16:27:43 INFO DAGScheduler: Submitting 1 missing tasks from
>> ShuffleMapStage 1 (MapPartitionsRDD[4] at count at eTOX.scala:53)
>> 15/11/25 16:27:43 INFO TaskSchedulerImpl: Adding task set 1.0 with 1 tasks
>> 15/11/25 16:27:43 INFO TaskSetManager: Starting task 0.0 in stage 1.0
>> (TID 2, 10.0.2.15, partition 0,PROCESS_LOCAL, 2300 bytes)
>> 15/11/25 16:27:43 INFO BlockManagerInfo: Added broadcast_2_piece0 in
>> memory on 10.0.2.15:39373 (size: 6.0 KB, free: 736.1 MB)
>> 15/11/25 16:27:43 INFO BlockManagerInfo: Added broadcast_1_piece0 in
>> memory on 10.0.2.15:39373 (size: 19.3 KB, free: 736.1 MB)
>> 15/11/25 16:27:44 INFO TaskSetManager: Finished task 0.0 in stage 1.0
>> (TID 2) in 1000 ms on 10.0.2.15 (1/1)
>> 15/11/25 16:27:44 INFO TaskSchedulerImpl: Removed TaskSet 1.0, whose
>> tasks have all completed, from pool
>> 15/11/25 16:27:44 INFO DAGScheduler: ShuffleMapStage 1 (count at
>> eTOX.scala:53) finished in 1.003 s
>> 15/11/25 16:27:44 INFO DAGScheduler: looking for newly runnable stages
>> 15/11/25 16:27:44 INFO DAGScheduler: running: Set()
>> 15/11/25 16:27:44 INFO DAGScheduler: waiting: Set(ResultStage 2)
>> 15/11/25 16:27:44 INFO DAGScheduler: failed: Set()
>> 15/11/25 16:27:44 INFO DAGScheduler: Submitting ResultStage 2
>> (MapPartitionsRDD[7] at count at eTOX.scala:53), which has no missing
>> parents
>> 15/11/25 16:27:44 INFO MemoryStore: Ensuring 1048576 bytes of free space
>> for block broadcast_3(free: 771696835, max: 771883008)
>> 15/11/25 16:27:44 INFO MemoryStore: Ensuring 14032 bytes of free space
>> for block broadcast_3(free: 771696835, max: 771883008)
>> 15/11/25 16:27:44 INFO MemoryStore: Block broadcast_3 stored as values in
>> memory (estimated size 13.7 KB, free 195.5 KB)
>> 15/11/25 16:27:44 INFO MemoryStore: Ensuring 6698 bytes of free space for
>> block broadcast_3_piece0(free: 771682803, max: 771883008)
>> 15/11/25 16:27:44 INFO MemoryStore: Block broadcast_3_piece0 stored as
>> bytes in memory (estimated size 6.5 KB, free 202.1 KB)
>> 15/11/25 16:27:44 INFO BlockManagerInfo: Added broadcast_3_piece0 in
>> memory on 10.0.2.15:35769 (size: 6.5 KB, free: 736.1 MB)
>> 15/11/25 16:27:44 INFO SparkContext: Created broadcast 3 from broadcast
>> at DAGScheduler.scala:1003
>> 15/11/25 16:27:44 INFO DAGScheduler: Submitting 1 missing tasks from
>> ResultStage 2 (MapPartitionsRDD[7] at count at eTOX.scala:53)
>> 15/11/25 16:27:44 INFO TaskSchedulerImpl: Adding task set 2.0 with 1 tasks
>> 15/11/25 16:27:44 INFO TaskSetManager: Starting task 0.0 in stage 2.0
>> (TID 3, 10.0.2.15, partition 0,NODE_LOCAL, 2060 bytes)
>> 15/11/25 16:27:44 INFO BlockManagerInfo: Added broadcast_3_piece0 in
>> memory on 10.0.2.15:39373 (size: 6.5 KB, free: 736.1 MB)
>> 15/11/25 16:27:44 INFO MapOutputTrackerMasterEndpoint: Asked to send map
>> output locations for shuffle 0 to 10.0.2.15:50472
>> 15/11/25 16:27:44 INFO MapOutputTrackerMaster: Size of output statuses
>> for shuffle 0 is 138 bytes
>> 15/11/25 16:27:44 INFO TaskSetManager: Finished task 0.0 in stage 2.0
>> (TID 3) in 190 ms on 10.0.2.15 (1/1)
>> 15/11/25 16:27:44 INFO TaskSchedulerImpl: Removed TaskSet 2.0, whose
>> tasks have all completed, from pool
>> 15/11/25 16:27:44 INFO DAGScheduler: ResultStage 2 (count at
>> eTOX.scala:53) finished in 0.195 s
>> 15/11/25 16:27:44 INFO DAGScheduler: Job 1 finished: count at
>> eTOX.scala:53, took 1.249107 s
>> Num Findings: 2775141
>> Exception in thread "main" java.lang.UnsatisfiedLinkError:
>> org.RDKit.RDKFuncsJNI.RWMol_MolFromSmiles__SWIG_3(Ljava/lang/String;)J
>>     at org.RDKit.RDKFuncsJNI.RWMol_MolFromSmiles__SWIG_3(Native Method)
>>     at org.RDKit.RWMol.MolFromSmiles(RWMol.java:426)
>>     at models.spark.sources.eTOX_DB$.main(eTOX.scala:54)
>>     at models.spark.sources.eTOX_DB.main(eTOX.scala)
>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>     at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>     at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>     at java.lang.reflect.Method.invoke(Method.java:497)
>>     at
>> org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:727)
>>     at
>> org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:181)
>>     at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:206)
>>     at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:121)
>>     at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>> 15/11/25 16:27:44 INFO SparkContext: Invoking stop() from shutdown hook
>> 15/11/25 16:27:44 INFO BlockManagerInfo: Removed broadcast_3_piece0 on
>> 10.0.2.15:35769 in memory (size: 6.5 KB, free: 736.1 MB)
>> 15/11/25 16:27:44 INFO BlockManagerInfo: Removed broadcast_3_piece0 on
>> 10.0.2.15:39373 in memory (size: 6.5 KB, free: 736.1 MB)
>> 15/11/25 16:27:44 INFO ContextCleaner: Cleaned accumulator 12
>> 15/11/25 16:27:44 INFO BlockManagerInfo: Removed broadcast_2_piece0 on
>> 10.0.2.15:35769 in memory (size: 6.0 KB, free: 736.1 MB)
>> 15/11/25 16:27:44 INFO BlockManagerInfo: Removed broadcast_2_piece0 on
>> 10.0.2.15:39373 in memory (size: 6.0 KB, free: 736.1 MB)
>> 15/11/25 16:27:44 INFO SparkUI: Stopped Spark web UI at
>> http://10.0.2.15:4040
>> 15/11/25 16:27:44 INFO ContextCleaner: Cleaned accumulator 11
>> 15/11/25 16:27:44 INFO ContextCleaner: Cleaned shuffle 0
>> 15/11/25 16:27:44 INFO DAGScheduler: Stopping DAGScheduler
>> 15/11/25 16:27:44 INFO SparkDeploySchedulerBackend: Shutting down all
>> executors
>> 15/11/25 16:27:44 INFO SparkDeploySchedulerBackend: Asking each executor
>> to shut down
>> 15/11/25 16:27:44 INFO MapOutputTrackerMasterEndpoint:
>> MapOutputTrackerMasterEndpoint stopped!
>> 15/11/25 16:27:44 INFO MemoryStore: MemoryStore cleared
>> 15/11/25 16:27:44 INFO BlockManager: BlockManager stopped
>> 15/11/25 16:27:44 INFO BlockManagerMaster: BlockManagerMaster stopped
>> 15/11/25 16:27:44 INFO
>> OutputCommitCoordinator$OutputCommitCoordinatorEndpoint:
>> OutputCommitCoordinator stopped!
>> 15/11/25 16:27:44 INFO RemoteActorRefProvider$RemotingTerminator:
>> Shutting down remote daemon.
>> 15/11/25 16:27:44 INFO SparkContext: Successfully stopped SparkContext
>> 15/11/25 16:27:44 INFO ShutdownHookManager: Shutdown hook called
>> 15/11/25 16:27:44 INFO RemoteActorRefProvider$RemotingTerminator: Remote
>> daemon shut down; proceeding with flushing remote transports.
>> 15/11/25 16:27:44 INFO ShutdownHookManager: Deleting directory
>> /tmp/spark-3f269496-c3aa-4bbc-8955-1ab88a179420
>>
>>
>>
>> Oriol.
>>
>
>

Reply via email to