i have the same issue running spark sql code from eclipse workspace. If you run your code from the command line (with a packaged jar) or from Intellij, I bet it should work.
IMHO This is some how related to eclipse env, but would love to know how to fix it (whether via eclipse conf, or via a patch in spark). On 03/01/2015 02:32 AM, Michael Armbrust wrote: > I think its possible that the problem is that the scala compiler is > not being loaded by the primordial classloader (but instead by some > child classloader) and thus the scala reflection mirror is failing to > initialize when it can't find it. Unfortunately, the only solution > that I know of is to load all required jars when the JVM starts. > > On Sat, Feb 28, 2015 at 5:26 PM, Ashish Nigam <ashnigamt...@gmail.com > <mailto:ashnigamt...@gmail.com>> wrote: > > Also, can scala version play any role here? > I am using scala 2.11.5 but all spark packages have dependency to > scala 2.11.2 > Just wanted to make sure that scala version is not an issue here. > > On Sat, Feb 28, 2015 at 9:18 AM, Ashish Nigam > <ashnigamt...@gmail.com <mailto:ashnigamt...@gmail.com>> wrote: > > Hi, > I wrote a very simple program in scala to convert an existing > RDD to SchemaRDD. > But createSchemaRDD function is throwing exception > > Exception in thread "main" scala.ScalaReflectionException: > class org.apache.spark.sql.catalyst.ScalaReflection in > JavaMirror with primordial classloader with boot classpath > [.....] not found > > > Here's more info on the versions I am using - > > <scala.binary.version>2.11</scala.binary.version> > <spark.version>1.2.1</spark.version> > <scala.version>2.11.5</scala.version> > > Please let me know how can I resolve this problem. > > Thanks > Ashish > > >