Zoltan Haindrich created HIVE-20747:
---------------------------------------

             Summary: Running a single spark tests alone (spark_explainuser_1) 
results in NoSuchMethodError
                 Key: HIVE-20747
                 URL: https://issues.apache.org/jira/browse/HIVE-20747
             Project: Hive
          Issue Type: Bug
          Components: Spark, Tests
            Reporter: Zoltan Haindrich


reproduce:
{code}
time mvn install -Pitests -pl itests/qtest-spark/ 
-Dtest=TestMiniSparkOnYarnCliDriver#testCliDriver[spark_explainuser_1]  -am 
{code}

I think the actual error is misleading...the real exception in hive.log is:
{code}
2018-10-15T04:44:39,102 ERROR [5bad7b56-dbbe-4868-8006-0aeecf9eb6c3 main] 
status.SparkJobMonitor: Spark job[1] failed
java.util.concurrent.ExecutionException: Exception thrown by job
        at 
org.apache.spark.JavaFutureActionWrapper.getImpl(FutureAction.scala:337) 
~[spark-core_2.11-2.3.0.jar:2.3.0]
        at org.apache.spark.JavaFutureActionWrapper.get(FutureAction.scala:342) 
~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:404)
 ~[hive-exec-4.0.0-SNAPSHOT.jar:4.0.0-SNAPSHOT]
        at 
org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:365)
 ~[hive-exec-4.0.0-SNAPSHOT.jar:4.0.0-SNAPSHOT]
        at java.util.concurrent.FutureTask.run(FutureTask.java:266) 
~[?:1.8.0_181]
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) 
~[?:1.8.0_181]
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) 
~[?:1.8.0_181]
        at java.lang.Thread.run(Thread.java:748) ~[?:1.8.0_181]
Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: 
Task 0 in stage 1.0 failed 4 times, most recent failure: Lost task 0.3 in stage 
1.0 (TID 4, savara.lan, executor 1): java.lang.NoSuchMethodError: 
com.esotericsoftware.kryo.io.Output.writeVarInt(IZ)I
        at 
org.apache.hive.spark.HiveKryoRegistrator$HiveKeySerializer.write(HiveKryoRegistrator.java:44)
        at 
org.apache.hive.spark.HiveKryoRegistrator$HiveKeySerializer.write(HiveKryoRegistrator.java:41)
        at com.esotericsoftware.kryo.Kryo.writeClassAndObject(Kryo.java:568)
        at 
org.apache.spark.serializer.KryoSerializationStream.writeObject(KryoSerializer.scala:241)
        at 
org.apache.spark.serializer.SerializationStream.writeKey(Serializer.scala:132)
        at 
org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:240)
        at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:151)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)
        at org.apache.spark.scheduler.Task.run(Task.scala:109)
        at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
        at java.lang.Thread.run(Thread.java:748)

Driver stacktrace:
        at 
org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1599)
 ~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1587)
 ~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1586)
 ~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) 
~[scala-library-2.11.8.jar:?]
        at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) 
~[scala-library-2.11.8.jar:?]
        at 
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1586) 
~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:831)
 ~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:831)
 ~[spark-core_2.11-2.3.0.jar:2.3.0]
        at scala.Option.foreach(Option.scala:257) ~[scala-library-2.11.8.jar:?]
        at 
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:831)
 ~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1820)
 ~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1769)
 ~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1758)
 ~[spark-core_2.11-2.3.0.jar:2.3.0]
        at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48) 
~[spark-core_2.11-2.3.0.jar:2.3.0]
Caused by: java.lang.NoSuchMethodError: 
com.esotericsoftware.kryo.io.Output.writeVarInt(IZ)I
        at 
org.apache.hive.spark.HiveKryoRegistrator$HiveKeySerializer.write(HiveKryoRegistrator.java:44)
 ~[?:?]
        at 
org.apache.hive.spark.HiveKryoRegistrator$HiveKeySerializer.write(HiveKryoRegistrator.java:41)
 ~[?:?]
        at com.esotericsoftware.kryo.Kryo.writeClassAndObject(Kryo.java:568) 
~[kryo-2.21.jar:?]
        at 
org.apache.spark.serializer.KryoSerializationStream.writeObject(KryoSerializer.scala:241)
 ~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
org.apache.spark.serializer.SerializationStream.writeKey(Serializer.scala:132) 
~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:240)
 ~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:151)
 ~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96) 
~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53) 
~[spark-core_2.11-2.3.0.jar:2.3.0]
        at org.apache.spark.scheduler.Task.run(Task.scala:109) 
~[spark-core_2.11-2.3.0.jar:2.3.0]
        at 
org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345) 
~[spark-core_2.11-2.3.0.jar:2.3.0]
{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to