Interesting, I believe we have run that query with version 1.1.0 with
codegen turned on and not much has changed there.  Is the error
deterministic?

On Mon, Nov 17, 2014 at 7:04 PM, Eric Zhen <zhpeng...@gmail.com> wrote:

> Hi Michael,
>
> We use Spark v1.1.1-rc1 with jdk 1.7.0_51 and scala 2.10.4.
>
> On Tue, Nov 18, 2014 at 7:09 AM, Michael Armbrust <mich...@databricks.com>
> wrote:
>
>> What version of Spark SQL?
>>
>> On Sat, Nov 15, 2014 at 10:25 PM, Eric Zhen <zhpeng...@gmail.com> wrote:
>>
>>> Hi all,
>>>
>>> We run SparkSQL on TPCDS benchmark Q19 with  spark.sql.codegen=true, we
>>> got exceptions as below, has anyone else saw these before?
>>>
>>> java.lang.ExceptionInInitializerError
>>>         at
>>> org.apache.spark.sql.execution.SparkPlan.newProjection(SparkPlan.scala:92)
>>>         at
>>> org.apache.spark.sql.execution.Exchange$$anonfun$execute$1$$anonfun$1.apply(Exchange.scala:51)
>>>         at
>>> org.apache.spark.sql.execution.Exchange$$anonfun$execute$1$$anonfun$1.apply(Exchange.scala:48)
>>>         at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:596)
>>>         at org.apache.spark.rdd.RDD$$anonfun$13.apply(RDD.scala:596)
>>>         at
>>> org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
>>>         at
>>> org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262)
>>>         at org.apache.spark.rdd.RDD.iterator(RDD.scala:229)
>>>         at
>>> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:68)
>>>         at
>>> org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
>>>         at org.apache.spark.scheduler.Task.run(Task.scala:54)
>>>         at
>>> org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:178)
>>>         at
>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>>>         at
>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>>>         at java.lang.Thread.run(Thread.java:744)
>>> Caused by: java.lang.NullPointerException
>>>         at
>>> scala.reflect.internal.Types$TypeRef.computeHashCode(Types.scala:2358)
>>>         at
>>> scala.reflect.internal.Types$UniqueType.<init>(Types.scala:1304)
>>>         at scala.reflect.internal.Types$TypeRef.<init>(Types.scala:2341)
>>>         at
>>> scala.reflect.internal.Types$NoArgsTypeRef.<init>(Types.scala:2137)
>>>         at
>>> scala.reflect.internal.Types$TypeRef$$anon$6.<init>(Types.scala:2544)
>>>         at scala.reflect.internal.Types$TypeRef$.apply(Types.scala:2544)
>>>         at scala.reflect.internal.Types$class.typeRef(Types.scala:3615)
>>>         at
>>> scala.reflect.internal.SymbolTable.typeRef(SymbolTable.scala:13)
>>>         at
>>> scala.reflect.internal.Symbols$TypeSymbol.newTypeRef(Symbols.scala:2752)
>>>         at
>>> scala.reflect.internal.Symbols$TypeSymbol.typeConstructor(Symbols.scala:2806)
>>>         at
>>> scala.reflect.internal.Symbols$SymbolContextApiImpl.toTypeConstructor(Symbols.scala:103)
>>>         at
>>> scala.reflect.internal.Symbols$TypeSymbol.toTypeConstructor(Symbols.scala:2698)
>>>         at
>>> org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator$$typecreator1$1.apply(CodeGenerator.scala:46)
>>>         at
>>> scala.reflect.api.TypeTags$WeakTypeTagImpl.tpe$lzycompute(TypeTags.scala:231)
>>>         at
>>> scala.reflect.api.TypeTags$WeakTypeTagImpl.tpe(TypeTags.scala:231)
>>>         at scala.reflect.api.TypeTags$class.typeOf(TypeTags.scala:335)
>>>         at scala.reflect.api.Universe.typeOf(Universe.scala:59)
>>>         at
>>> org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator.<init>(CodeGenerator.scala:46)
>>>         at
>>> org.apache.spark.sql.catalyst.expressions.codegen.GenerateProjection$.<init>(GenerateProjection.scala:29)
>>>         at
>>> org.apache.spark.sql.catalyst.expressions.codegen.GenerateProjection$.<clinit>(GenerateProjection.scala)
>>>         ... 15 more
>>> --
>>> Best Regards
>>>
>>
>>
>
>
> --
> Best Regards
>

Reply via email to