Supreeth Sharma created ZEPPELIN-3433: -----------------------------------------
Summary: Incorrect status shown for %spark2 interpreter Key: ZEPPELIN-3433 URL: https://issues.apache.org/jira/browse/ZEPPELIN-3433 Project: Zeppelin Issue Type: Bug Components: zeppelin-server Affects Versions: 0.8.0 Reporter: Supreeth Sharma Fix For: 0.8.0 %spark2 interpreter is showing the status as 'Finished' even when it has actually failed. Steps to reproduce : 1) Create spark.conf with incorrect jar file. {code:java} %spark2.conf spark.jars /tmp/zep_test/spark2-examples-assembl-2.3.0.3.0.0.0-1260.jar spark.app.name test_new_1 spark.executor.instances 2 {code} 2) Now create spark context by running any spark statement. {code:java} %spark sc.version {code} 3) Though spark context is not created and para didnt run successfully, the status is shown as 'Finished' without showing error message Below exceptions are seen in zeppelin logs, but are not shown on the UI. {code:java} INFO [2018-04-27 14:58:55,014] ({pool-3-thread-2} Logging.scala[logInfo]:54) - Bound SparkUI to 0.0.0.0, and started at http://ctr-e138-1518143905142-267605-01-000002.hwx.site:4040 ERROR [2018-04-27 14:58:55,030] ({pool-3-thread-2} Logging.scala[logError]:91) - Failed to add /tmp/zep_test/spark2-examples-assembl-2.3.0.3.0.0.0-1260.jar to Spark environment java.io.FileNotFoundException: Jar /tmp/zep_test/spark2-examples-assembl-2.3.0.3.0.0.0-1260.jar not found at org.apache.spark.SparkContext.addJarFile$1(SparkContext.scala:1814) at org.apache.spark.SparkContext.addJar(SparkContext.scala:1842) at org.apache.spark.SparkContext$$anonfun$12.apply(SparkContext.scala:457) at org.apache.spark.SparkContext$$anonfun$12.apply(SparkContext.scala:457) at scala.collection.immutable.List.foreach(List.scala:381) at org.apache.spark.SparkContext.<init>(SparkContext.scala:457) at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2493) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:930) at org.apache.spark.sql.SparkSession$Builder$$anonfun$7.apply(SparkSession.scala:921) at scala.Option.getOrElse(Option.scala:121) at org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:921) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:44) at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:39) at org.apache.zeppelin.spark.OldSparkInterpreter.createSparkSession(OldSparkInterpreter.java:345) at org.apache.zeppelin.spark.OldSparkInterpreter.getSparkSession(OldSparkInterpreter.java:219) at org.apache.zeppelin.spark.OldSparkInterpreter.open(OldSparkInterpreter.java:738) at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:61) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:611) at org.apache.zeppelin.scheduler.Job.run(Job.java:186) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:140) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) WARN [2018-04-27 14:58:55,074] ({pool-3-thread-2} Logging.scala[logWarning]:66) - Fair Scheduler configuration file not found so jobs will be scheduled in FIFO order. To use fair scheduling, configure pools in fairscheduler.xml or set spark.scheduler.allocation.file to a file that contains the configuration. {code} Correct status(ERROR) is shown for livy2 interpreter in this scenario. Issue is happening only with spark2 interpreter. -- This message was sent by Atlassian JIRA (v7.6.3#76005)