Hello all,

I am trying to run a Spark job but while running the job I am getting missing 
class errors. First I got a missing class error for Joda DateTime, I added the 
dependency in my build.sbt and rebuilt the assembly jar and tried again. This 
time I am getting the same error for java.util.Date. I find that error odd as I 
shouldn’t have to package a core java class in the assembly package. Am I 
missing something? Shouldn’t it load the class from rt. jar?

4/10/27 11:54:09 WARN TaskSetManager: Lost task 0.0 in stage 2.0 (TID 5, 
localhost): java.io.IOException: Exception during preparation of SELECT "mpxn", 
"reading_date", "division", "insert_date", "insert_etl_id", "msn", 
"register_name", "sourcekey", "t0000", "t0030", "t0100", "t0130", "t0200", 
"t0230", "t0300", "t0330", "t0400", "t0430", "t0500", "t0530", "t0600", 
"t0630", "t0700", "t0730", "t0800", "t0830", "t0900", "t0930", "t1000", 
"t1030", "t1100", "t1130", "t1200", "t1230", "t1300", "t1330", "t1400", 
"t1430", "t1500", "t1530", "t1600", "t1630", "t1700", "t1730", "t1800", 
"t1830", "t1900", "t1930", "t2000", "t2030", "t2100", "t2130", "t2200", 
"t2230", "t2300", "t2330", "trilliant_smart_read_id", "update_date", 
"update_etl_id" FROM "ser_ingest"."profile_reads" WHERE token("mpxn") > ? AND 
token("mpxn") <= ? ALLOW FILTERING: class java.util.Date in JavaMirror with 
org.apache.spark.executor.ExecutorURLClassLoader@6b69dc09 of type class 
org.apache.spark.executor.ExecutorURLClassLoader with classpath 
[file:/tmp/transformation.jar] and parent being 
sun.misc.Launcher$AppClassLoader@2494310b of type class 
sun.misc.Launcher$AppClassLoader with classpath 
[file:/etc/spark/spark-1.1.0-bin-hadoop2.4/,file:/etc/spark/spark-1.1.0-bin-hadoop2.4/,file:/etc/spark/spark-1.1.0-bin-hadoop2.4/conf/,file:/etc/spark/spark-1.1.0-bin-hadoop2.4/lib/spark-assembly-1.1.0-hadoop2.4.0.jar]
 and parent being sun.misc.Launcher$ExtClassLoader@7d1fdbef of type class 
sun.misc.Launcher$ExtClassLoader with classpath 
[file:/usr/lib/jvm/jre1.7.0_60/lib/ext/localedata.jar,file:/usr/lib/jvm/jre1.7.0_60/lib/ext/sunec.jar,file:/usr/lib/jvm/jre1.7.0_60/lib/ext/zipfs.jar,file:/usr/lib/jvm/jre1.7.0_60/lib/ext/sunpkcs11.jar,file:/usr/lib/jvm/jre1.7.0_60/lib/ext/sunjce_provider.jar,file:/usr/lib/jvm/jre1.7.0_60/lib/ext/dnsns.jar]
 and parent being primordial classloader with boot classpath 
[/usr/lib/jvm/jre1.7.0_60/lib/resources.jar:/usr/lib/jvm/jre1.7.0_60/lib/rt.jar:/usr/lib/jvm/jre1.7.0_60/lib/sunrsasign.jar:/usr/lib/jvm/jre1.7.0_60/lib/jsse.jar:/usr/lib/jvm/jre1.7.0_60/lib/jce.jar:/usr/lib/jvm/jre1.7.0_60/lib/charsets.jar:/usr/lib/jvm/jre1.7.0_60/lib/jfr.jar:/usr/lib/jvm/jre1.7.0_60/classes]
 not found.


SBT file
import AssemblyKeys._

assemblySettings

jarName in assembly := "transformation.jar"

name := "dataplatform-readings-transformation"

version := "1.0"

scalaVersion := "2.10.4"

libraryDependencies ++= Seq(
  "joda-time" % "joda-time" % "2.5",
  "org.apache.spark" %% "spark-core" % "1.1.0" % "provided",
  "com.datastax.spark" %%       "spark-cassandra-connector" % "1.1.0-alpha4",
//  "com.datastax.spark" %%     "spark-cassandra-connector" % "1.0.3",
  "org.apache.spark" %% "spark-sql" % "1.1.0"
)

mergeStrategy in assembly <<= (mergeStrategy in assembly) { (old) => {
  case PathList("META-INF", "ECLIPSEF.RSA" ) => MergeStrategy.discard
  case PathList("META-INF", "mailcap" ) => MergeStrategy.discard
  case PathList("com", "esotericsoftware", xs @ _*) => MergeStrategy.last
  case PathList("org", "apache", xs @ _*) => MergeStrategy.last
  case PathList("plugin.properties" ) => MergeStrategy.discard
  case PathList("META-INF","maven","org.slf4j","slf4j-api","pom.properties" ) 
=> MergeStrategy.discard
  case PathList("META-INF","maven","org.slf4j","slf4j-api","pom.xml" ) => 
MergeStrategy.discard
  case x => old(x)
}
}

Thanks,
Saket

Reply via email to