Hi Michael, We use Spark v1.1.1-rc1 with jdk 1.7.0_51 and scala 2.10.4.
On Tue, Nov 18, 2014 at 7:09 AM, Michael Armbrust <mich...@databricks.com> wrote: > What version of Spark SQL? > > On Sat, Nov 15, 2014 at 10:25 PM, Eric Zhen <zhpeng...@gmail.com> wrote: > >> Hi all, >> >> We run SparkSQL on TPCDS benchmark Q19 with spark.sql.codegen=true, we >> got exceptions as below, has anyone else saw these before? >> >> java.lang.ExceptionInInitializerError >> at >> org.apache.spark.sql.execution.SparkPlan.newProjection(SparkPlan.scala:92) >> at >> org.apache.spark.sql.execution.Exchange$$anonfun$execute$1$$anonfun$1.apply(Exchange.scala:51) >> at >> org.apache.spark.sql.execution.Exchange$$anonfun$execute$1$$anonfun$1.apply(Exchange.scala:48) >> at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:596) >> at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:596) >> at >> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35) >> at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262) >> at org.apache.spark.rdd.RDD.iterator(RDD.scala:229) >> at >> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68) >> at >> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41) >> at org.apache.spark.scheduler.Task.run(Task.scala:54) >> at >> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:178) >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >> at java.lang.Thread.run(Thread.java:744) >> Caused by: java.lang.NullPointerException >> at >> scala.reflect.internal.Types$TypeRef.computeHashCode(Types.scala:2358) >> at >> scala.reflect.internal.Types$UniqueType.<init>(Types.scala:1304) >> at scala.reflect.internal.Types$TypeRef.<init>(Types.scala:2341) >> at >> scala.reflect.internal.Types$NoArgsTypeRef.<init>(Types.scala:2137) >> at >> scala.reflect.internal.Types$TypeRef$$anon$6.<init>(Types.scala:2544) >> at scala.reflect.internal.Types$TypeRef$.apply(Types.scala:2544) >> at scala.reflect.internal.Types$class.typeRef(Types.scala:3615) >> at >> scala.reflect.internal.SymbolTable.typeRef(SymbolTable.scala:13) >> at >> scala.reflect.internal.Symbols$TypeSymbol.newTypeRef(Symbols.scala:2752) >> at >> scala.reflect.internal.Symbols$TypeSymbol.typeConstructor(Symbols.scala:2806) >> at >> scala.reflect.internal.Symbols$SymbolContextApiImpl.toTypeConstructor(Symbols.scala:103) >> at >> scala.reflect.internal.Symbols$TypeSymbol.toTypeConstructor(Symbols.scala:2698) >> at >> org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator$$typecreator1$1.apply(CodeGenerator.scala:46) >> at >> scala.reflect.api.TypeTags$WeakTypeTagImpl.tpe$lzycompute(TypeTags.scala:231) >> at >> scala.reflect.api.TypeTags$WeakTypeTagImpl.tpe(TypeTags.scala:231) >> at scala.reflect.api.TypeTags$class.typeOf(TypeTags.scala:335) >> at scala.reflect.api.Universe.typeOf(Universe.scala:59) >> at >> org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator.<init>(CodeGenerator.scala:46) >> at >> org.apache.spark.sql.catalyst.expressions.codegen.GenerateProjection$.<init>(GenerateProjection.scala:29) >> at >> org.apache.spark.sql.catalyst.expressions.codegen.GenerateProjection$.<clinit>(GenerateProjection.scala) >> ... 15 more >> -- >> Best Regards >> > > -- Best Regards