I'm getting an exception when I try to submit a job (through prediction.io, if 
you know it):
[INFO] [Runner$] Submission command: 
/home/pio/PredictionIO/vendors/spark-1.5.1/bin/spark-submit --class 
io.prediction.tools.imprt.FileToEvents --files 
file:/home/pio/PredictionIO/conf/log4j.properties,file:/home/pio/PredictionIO/vendors/hbase-1.0.0/conf/hbase-site.xml
 --driver-class-path 
/home/pio/PredictionIO/conf:/home/pio/PredictionIO/vendors/hbase-1.0.0/conf 
file:/home/pio/PredictionIO/lib/pio-assembly-0.9.5.jar --appid 1 --input 
file:/home/pio/pio-ur/my_events.json --env 
PIO_STORAGE_SOURCES_HBASE_TYPE=hbase,PIO_ENV_LOADED=1,PIO_STORAGE_REPOSITORIES_METADATA_NAME=pio_meta,PIO_FS_BASEDIR=/home/pio/.pio_store,PIO_STORAGE_SOURCES_HBASE_HOME=/home/pio/PredictionIO/vendors/hbase-1.0.0,PIO_HOME=/home/pio/PredictionIO,PIO_FS_ENGINESDIR=/home/pio/.pio_store/engines,PIO_STORAGE_SOURCES_LOCALFS_PATH=/home/pio/.pio_store/models,PIO_STORAGE_SOURCES_ELASTICSEARCH_TYPE=elasticsearch,PIO_STORAGE_REPOSITORIES_METADATA_SOURCE=ELASTICSEARCH,PIO_STORAGE_REPOSITORIES_MODELDATA_SOURCE=LOCALFS,PIO_STORAGE_REPOSITORIES_EVENTDATA_NAME=pio_event,PIO_STORAGE_SOURCES_ELASTICSEARCH_HOME=/home/pio/PredictionIO/vendors/elasticsearch-1.4.4,PIO_FS_TMPDIR=/home/pio/.pio_store/tmp,PIO_STORAGE_REPOSITORIES_MODELDATA_NAME=pio_model,PIO_STORAGE_REPOSITORIES_EVENTDATA_SOURCE=HBASE,PIO_CONF_DIR=/home/pio/PredictionIO/conf,PIO_STORAGE_SOURCES_LOCALFS_TYPE=localfs[INFO]
 [Remoting] Starting remoting[INFO] [Remoting] Remoting started; listening on 
addresses :[akka.tcp://sparkDriver@192.168.56.150:52838][WARN] [MetricsSystem] 
Using default name DAGScheduler for source because spark.app.id is not 
set.[ERROR] [Executor] Exception in task 0.0 in stage 0.0 (TID 0)[WARN] 
[ThrowableSerializationWrapper] Task exception could not be deserialized[ERROR] 
[TaskResultGetter] Could not deserialize TaskEndReason: ClassNotFound with 
classloader org.apache.spark.util.MutableURLClassLoader@59e5ddf[WARN] 
[TaskSetManager] Lost task 0.0 in stage 0.0 (TID 0, localhost): 
UnknownReason[ERROR] [TaskSetManager] Task 0 in stage 0.0 failed 1 times; 
aborting jobException in thread "main" org.apache.spark.SparkException: Job 
aborted due to stage failure: Task 0 in stage 0.0 failed 1 times, most recent 
failure: Lost task 0.0 in stage 0.0 (TID 0, localhost): UnknownReasonDriver 
stacktrace: at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1283)
 at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1271)
 at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1270)
 at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) 
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1270) at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
 at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
 at scala.Option.foreach(Option.scala:236) at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:697)
 at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1496)
 at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1458)
 at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1447)
 at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) at 
org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:567) at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:1822) at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:1835) at 
org.apache.spark.SparkContext.runJob(SparkContext.scala:1912) at 
org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply$mcV$sp(PairRDDFunctions.scala:1055)
 at 
org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:998)
 at 
org.apache.spark.rdd.PairRDDFunctions$$anonfun$saveAsNewAPIHadoopDataset$1.apply(PairRDDFunctions.scala:998)
 at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147) 
at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:108) 
at org.apache.spark.rdd.RDD.withScope(RDD.scala:306) at 
org.apache.spark.rdd.PairRDDFunctions.saveAsNewAPIHadoopDataset(PairRDDFunctions.scala:998)
 at io.prediction.data.storage.hbase.HBPEvents.write(HBPEvents.scala:108) at 
io.prediction.tools.imprt.FileToEvents$$anonfun$main$1.apply(FileToEvents.scala:90)
 at 
io.prediction.tools.imprt.FileToEvents$$anonfun$main$1.apply(FileToEvents.scala:63)
 at scala.Option.map(Option.scala:145) at 
io.prediction.tools.imprt.FileToEvents$.main(FileToEvents.scala:63) at 
io.prediction.tools.imprt.FileToEvents.main(FileToEvents.scala) at 
sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:497) at 
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:672)
 at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180) at 
org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205) at 
org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:120) at 
org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)[pio@allinone 
pio-ur]$
Hoping to get some hints what's causing this, or at least how to get the 
TaskEndReason class added to the class path in case it can provide more 
information.
Thanks!

Reply via email to