Hi guys,

after few years of working via spark livy I wanted to test out spark via
Yarn in cluster mode

I`m using HDP 2.6.5.x version of hadoop with spark 2.3.0, my HDP is
kerberized

I builded zeppelin from master branch and updated my interpreter
settings (see the attachment)
after running any spark paragraph I`m getting    

org.apache.zeppelin.interpreter.InterpreterException: 
org.apache.zeppelin.interpreter.InterpreterException: Fail to open 
SparkInterpreter

...

Caused by: scala.reflect.internal.MissingRequirementError: object 
java.lang.Object in compiler mirror not found.


I also attached logs from yarn and zeppelin to keep this mail


Also, when I switch from cluster mode to client mode, everything works
as expected...


If you have any idea, what I did wrong configuring this, please let me
know as I`m getting super frustrated here :(


Thanks a million

best regards

Miso

SPARK_HOME 	/usr/hdp/2.6.5.0-292/spark2 
master 	yarn 
spark.app.name 	test_drive 
spark.driver.cores 	1
spark.driver.memory 	1g 
spark.executor.cores 	1
spark.executor.memory 	1g 
spark.files 	
spark.jars 	
spark.jars.packages 	
zeppelin.spark.useHiveContext 	true 
zeppelin.spark.printREPLOutput 	true 
zeppelin.spark.maxResult 	1000
zeppelin.spark.enableSupportedVersionCheck 	true 
zeppelin.spark.uiWebUrl 	
zeppelin.spark.ui.hidden 	false 
spark.webui.yarn.useProxy 	false 
zeppelin.spark.scala.color 	true 
zeppelin.spark.deprecatedMsg.show 	true 
zeppelin.spark.concurrentSQL 	false 
zeppelin.spark.concurrentSQL.max 	10
zeppelin.spark.sql.stacktrace 	false 
zeppelin.spark.sql.interpolation 	false 
PYSPARK_PYTHON 	python 
PYSPARK_DRIVER_PYTHON 	python 
zeppelin.pyspark.useIPython 	true 
zeppelin.R.knitr 	true 
zeppelin.R.cmd 	R 
zeppelin.R.image.width 	100% 
zeppelin.R.render.options 	out.format = 'html', comment = NA, echo = FALSE, results = 'asis', message = F, warning = F, fig.retina = 2 
zeppelin.kotlin.shortenTypes 	true 
HADOOP_CONF_DIR 	/etc/hadoop/conf 
spark.submit.deployMode 	cluster 
SPARK_APP_JAR 	/usr/hdp/2.6.5.0-292/spark2/jars 
JAVA_HOME 	/usr/lib/jvm/java-8-oracle/jre/bin/java 
 INFO [2020-04-01 14:52:00,389] ({main} ZeppelinConfiguration.java[create]:163) - Load configuration from file:/opt/zeppelin-2020.03.31/conf/zeppelin-site.xml
 INFO [2020-04-01 14:52:00,459] ({main} ZeppelinConfiguration.java[create]:171) - Server Host: 0.0.0.0
 INFO [2020-04-01 14:52:00,459] ({main} ZeppelinConfiguration.java[create]:173) - Server Port: 9995
 INFO [2020-04-01 14:52:00,459] ({main} ZeppelinConfiguration.java[create]:177) - Context Path: /
 INFO [2020-04-01 14:52:00,459] ({main} ZeppelinConfiguration.java[create]:178) - Zeppelin Version: 0.9.0-SNAPSHOT
 INFO [2020-04-01 14:52:00,477] ({main} Log.java[initialized]:193) - Logging initialized @461ms to org.eclipse.jetty.util.log.Slf4jLog
 WARN [2020-04-01 14:52:00,728] ({main} ZeppelinConfiguration.java[getConfigFSDir]:631) - zeppelin.config.fs.dir is not specified, fall back to local conf directory zeppelin.conf.dir
 INFO [2020-04-01 14:52:00,756] ({main} Credentials.java[loadFromFile]:121) - /opt/zeppelin-server/conf/credentials.json
 INFO [2020-04-01 14:52:00,784] ({ImmediateThread-1585745520724} PluginManager.java[loadNotebookRepo]:60) - Loading NotebookRepo Plugin: org.apache.zeppelin.notebook.repo.VFSNotebookRepo
 INFO [2020-04-01 14:52:00,785] ({ImmediateThread-1585745520724} VFSNotebookRepo.java[setNotebookDirectory]:70) - Using notebookDir: /opt/zeppelin-server/notebook
 INFO [2020-04-01 14:52:00,827] ({main} ZeppelinServer.java[setupWebAppContext]:488) - warPath is: /opt/zeppelin-server/zeppelin-web-0.9.0-SNAPSHOT.war
 INFO [2020-04-01 14:52:00,827] ({main} ZeppelinServer.java[setupWebAppContext]:501) - ZeppelinServer Webapp path: /opt/zeppelin-server/webapps
 INFO [2020-04-01 14:52:00,849] ({main} ZeppelinServer.java[setupWebAppContext]:488) - warPath is: 
 INFO [2020-04-01 14:52:00,849] ({main} ZeppelinServer.java[setupWebAppContext]:501) - ZeppelinServer Webapp path: /opt/zeppelin-server/webapps/next
 INFO [2020-04-01 14:52:00,883] ({main} NotebookServer.java[<init>]:153) - NotebookServer instantiated: org.apache.zeppelin.socket.NotebookServer@7748410a
 INFO [2020-04-01 14:52:00,884] ({main} NotebookServer.java[setNotebook]:164) - Injected NotebookProvider
 INFO [2020-04-01 14:52:00,885] ({main} NotebookServer.java[setServiceLocator]:158) - Injected ServiceLocator: ServiceLocatorImpl(shared-locator,0,1884122755)
 INFO [2020-04-01 14:52:00,885] ({main} NotebookServer.java[setNotebookService]:171) - Injected NotebookServiceProvider
 INFO [2020-04-01 14:52:00,886] ({main} NotebookServer.java[setAuthorizationServiceProvider]:178) - Injected NotebookAuthorizationServiceProvider
 INFO [2020-04-01 14:52:00,886] ({main} NotebookServer.java[setConnectionManagerProvider]:184) - Injected ConnectionManagerProvider
 INFO [2020-04-01 14:52:00,886] ({main} ZeppelinServer.java[setupClusterManagerServer]:439) - Cluster mode is disabled
 INFO [2020-04-01 14:52:00,892] ({main} ZeppelinServer.java[main]:251) - Starting zeppelin server
 INFO [2020-04-01 14:52:00,893] ({main} Server.java[doStart]:370) - jetty-9.4.18.v20190429; built: 2019-04-29T20:42:08.989Z; git: e1bc35120a6617ee3df052294e433f3a25ce7097; jvm 1.8.0_201-b09
 INFO [2020-04-01 14:52:00,969] ({main} StandardDescriptorProcessor.java[visitServlet]:283) - NO JSP Support for /, did not find org.eclipse.jetty.jsp.JettyJspServlet
 INFO [2020-04-01 14:52:00,982] ({main} DefaultSessionIdManager.java[doStart]:365) - DefaultSessionIdManager workerName=node0
 INFO [2020-04-01 14:52:00,982] ({main} DefaultSessionIdManager.java[doStart]:370) - No SessionScavenger set, using defaults
 INFO [2020-04-01 14:52:00,984] ({main} HouseKeeper.java[startScavenging]:149) - node0 Scavenging every 600000ms
 INFO [2020-04-01 14:52:00,992] ({main} ContextHandler.java[log]:2349) - Initializing Shiro environment
 INFO [2020-04-01 14:52:00,992] ({main} EnvironmentLoader.java[initEnvironment]:133) - Starting Shiro environment initialization.
 INFO [2020-04-01 14:52:01,367] ({main} IniRealm.java[processDefinitions]:188) - IniRealm defined, but there is no [users] section defined.  This realm will not be populated with any users and it is assumed that they will be populated programatically.  Users must be defined for this Realm instance to be useful.
 INFO [2020-04-01 14:52:01,371] ({main} EnvironmentLoader.java[initEnvironment]:147) - Shiro environment initialized in 379 ms.
 INFO [2020-04-01 14:52:02,231] ({main} ContextHandler.java[doStart]:855) - Started o.e.j.w.WebAppContext@6150c3ec{zeppelin-web,/,jar:file:///opt/zeppelin-2020.03.31/zeppelin-web-0.9.0-SNAPSHOT.war!/,AVAILABLE}{/opt/zeppelin-server/zeppelin-web-0.9.0-SNAPSHOT.war}
 INFO [2020-04-01 14:52:02,242] ({main} StandardDescriptorProcessor.java[visitServlet]:283) - NO JSP Support for /next, did not find org.eclipse.jetty.jsp.JettyJspServlet
 INFO [2020-04-01 14:52:02,243] ({main} ContextHandler.java[log]:2349) - Initializing Shiro environment
 INFO [2020-04-01 14:52:02,243] ({main} EnvironmentLoader.java[initEnvironment]:133) - Starting Shiro environment initialization.
 INFO [2020-04-01 14:52:02,250] ({main} IniRealm.java[processDefinitions]:188) - IniRealm defined, but there is no [users] section defined.  This realm will not be populated with any users and it is assumed that they will be populated programatically.  Users must be defined for this Realm instance to be useful.
 INFO [2020-04-01 14:52:02,251] ({main} EnvironmentLoader.java[initEnvironment]:147) - Shiro environment initialized in 8 ms.
 INFO [2020-04-01 14:52:02,355] ({main} ContextHandler.java[doStart]:855) - Started o.e.j.w.WebAppContext@229f66ed{/next,file:///,AVAILABLE}{/}
 INFO [2020-04-01 14:52:02,369] ({main} AbstractConnector.java[doStart]:292) - Started ServerConnector@704a52ec{HTTP/1.1,[http/1.1]}{0.0.0.0:9995}
 INFO [2020-04-01 14:52:02,369] ({main} Server.java[doStart]:410) - Started @2355ms
 INFO [2020-04-01 14:52:04,615] ({qtp701141022-14} NotebookServer.java[onOpen]:243) - New connection from 10.100.212.124:54972
 WARN [2020-04-01 14:52:04,661] ({qtp701141022-14} ZeppelinConfiguration.java[getConfigFSDir]:631) - zeppelin.config.fs.dir is not specified, fall back to local conf directory zeppelin.conf.dir
 WARN [2020-04-01 14:52:04,661] ({qtp701141022-14} ZeppelinConfiguration.java[getConfigFSDir]:631) - zeppelin.config.fs.dir is not specified, fall back to local conf directory zeppelin.conf.dir
 WARN [2020-04-01 14:52:04,662] ({qtp701141022-14} ZeppelinConfiguration.java[getConfigFSDir]:631) - zeppelin.config.fs.dir is not specified, fall back to local conf directory zeppelin.conf.dir
 WARN [2020-04-01 14:52:04,662] ({qtp701141022-14} LocalConfigStorage.java[loadNotebookAuthorization]:83) - NotebookAuthorization file /opt/zeppelin-server/conf/notebook-authorization.json is not existed
 INFO [2020-04-01 14:52:07,369] ({main} ZeppelinServer.java[main]:265) - Done, zeppelin server started
 INFO [2020-04-01 14:52:08,462] ({qtp701141022-13} AbstractValidatingSessionManager.java[enableSessionValidation]:233) - Enabling session validation scheduler...
 INFO [2020-04-01 14:52:08,631] ({qtp701141022-13} InterpreterSettingManager.java[<init>]:186) - Using RecoveryStorage: org.apache.zeppelin.interpreter.recovery.NullRecoveryStorage
 INFO [2020-04-01 14:52:08,632] ({qtp701141022-13} InterpreterSettingManager.java[<init>]:192) - Using LifecycleManager: org.apache.zeppelin.interpreter.lifecycle.NullLifecycleManager
 INFO [2020-04-01 14:52:08,641] ({Thread-15} RemoteInterpreterEventServer.java[run]:104) - InterpreterEventServer is starting at 192.168.200.65:38198
 INFO [2020-04-01 14:52:09,135] ({qtp701141022-13} RemoteInterpreterEventServer.java[start]:128) - RemoteInterpreterEventServer is started
 INFO [2020-04-01 14:52:09,163] ({qtp701141022-13} InterpreterSettingManager.java[registerInterpreterSetting]:463) - Register InterpreterSettingTemplate: spark
 INFO [2020-04-01 14:52:09,168] ({qtp701141022-13} InterpreterSettingManager.java[registerInterpreterSetting]:463) - Register InterpreterSettingTemplate: java
 INFO [2020-04-01 14:52:09,176] ({qtp701141022-13} InterpreterSettingManager.java[registerInterpreterSetting]:463) - Register InterpreterSettingTemplate: submarine
 INFO [2020-04-01 14:52:09,178] ({qtp701141022-13} InterpreterSettingManager.java[registerInterpreterSetting]:463) - Register InterpreterSettingTemplate: jupyter
 INFO [2020-04-01 14:52:09,181] ({qtp701141022-13} InterpreterSettingManager.java[registerInterpreterSetting]:463) - Register InterpreterSettingTemplate: python
 INFO [2020-04-01 14:52:09,183] ({qtp701141022-13} InterpreterSettingManager.java[registerInterpreterSetting]:463) - Register InterpreterSettingTemplate: angular
 INFO [2020-04-01 14:52:09,185] ({qtp701141022-13} InterpreterSettingManager.java[registerInterpreterSetting]:463) - Register InterpreterSettingTemplate: file
 INFO [2020-04-01 14:52:09,187] ({qtp701141022-13} InterpreterSettingManager.java[registerInterpreterSetting]:463) - Register InterpreterSettingTemplate: md
 INFO [2020-04-01 14:52:09,190] ({qtp701141022-13} InterpreterSettingManager.java[registerInterpreterSetting]:463) - Register InterpreterSettingTemplate: livy
 INFO [2020-04-01 14:52:09,192] ({qtp701141022-13} InterpreterSettingManager.java[registerInterpreterSetting]:463) - Register InterpreterSettingTemplate: groovy
 INFO [2020-04-01 14:52:09,193] ({qtp701141022-13} InterpreterSettingManager.java[registerInterpreterSetting]:463) - Register InterpreterSettingTemplate: sh
 INFO [2020-04-01 14:52:09,195] ({qtp701141022-13} InterpreterSettingManager.java[registerInterpreterSetting]:463) - Register InterpreterSettingTemplate: jdbc
 INFO [2020-04-01 14:52:09,197] ({qtp701141022-13} InterpreterSettingManager.java[registerInterpreterSetting]:463) - Register InterpreterSettingTemplate: kotlin
 INFO [2020-04-01 14:52:09,199] ({qtp701141022-13} InterpreterSettingManager.java[registerInterpreterSetting]:463) - Register InterpreterSettingTemplate: r
 INFO [2020-04-01 14:52:09,200] ({qtp701141022-13} LocalConfigStorage.java[loadInterpreterSettings]:69) - Load Interpreter Setting from file: /opt/zeppelin-server/conf/interpreter.json
 INFO [2020-04-01 14:52:09,238] ({qtp701141022-13} InterpreterSettingManager.java[loadFromFile]:284) - Create Interpreter Setting python from interpreter.json
 INFO [2020-04-01 14:52:09,238] ({qtp701141022-13} InterpreterSettingManager.java[loadFromFile]:284) - Create Interpreter Setting kotlin from interpreter.json
 INFO [2020-04-01 14:52:09,238] ({qtp701141022-13} InterpreterSettingManager.java[loadFromFile]:284) - Create Interpreter Setting jdbc from interpreter.json
 INFO [2020-04-01 14:52:09,239] ({qtp701141022-13} InterpreterSettingManager.java[loadFromFile]:284) - Create Interpreter Setting angular from interpreter.json
 INFO [2020-04-01 14:52:09,239] ({qtp701141022-13} InterpreterSettingManager.java[loadFromFile]:284) - Create Interpreter Setting livy from interpreter.json
 INFO [2020-04-01 14:52:09,239] ({qtp701141022-13} InterpreterSettingManager.java[loadFromFile]:284) - Create Interpreter Setting r from interpreter.json
 INFO [2020-04-01 14:52:09,240] ({qtp701141022-13} InterpreterSettingManager.java[loadFromFile]:284) - Create Interpreter Setting java from interpreter.json
 INFO [2020-04-01 14:52:09,240] ({qtp701141022-13} InterpreterSettingManager.java[loadFromFile]:284) - Create Interpreter Setting jupyter from interpreter.json
 INFO [2020-04-01 14:52:09,240] ({qtp701141022-13} InterpreterSettingManager.java[loadFromFile]:284) - Create Interpreter Setting file from interpreter.json
 INFO [2020-04-01 14:52:09,240] ({qtp701141022-13} InterpreterSettingManager.java[loadFromFile]:284) - Create Interpreter Setting groovy from interpreter.json
 INFO [2020-04-01 14:52:09,241] ({qtp701141022-13} InterpreterSettingManager.java[loadFromFile]:284) - Create Interpreter Setting submarine from interpreter.json
 INFO [2020-04-01 14:52:09,241] ({qtp701141022-13} InterpreterSettingManager.java[loadFromFile]:284) - Create Interpreter Setting spark from interpreter.json
 INFO [2020-04-01 14:52:09,242] ({qtp701141022-13} InterpreterSettingManager.java[loadFromFile]:284) - Create Interpreter Setting sh from interpreter.json
 INFO [2020-04-01 14:52:09,242] ({qtp701141022-13} InterpreterSettingManager.java[loadFromFile]:284) - Create Interpreter Setting md from interpreter.json
 INFO [2020-04-01 14:52:09,260] ({qtp701141022-13} InterpreterSetting.java[setStatus]:740) - Set interpreter python status to DOWNLOADING_DEPENDENCIES
 INFO [2020-04-01 14:52:09,261] ({qtp701141022-13} InterpreterSetting.java[setStatus]:740) - Set interpreter kotlin status to DOWNLOADING_DEPENDENCIES
 INFO [2020-04-01 14:52:09,261] ({Thread-16} InterpreterSetting.java[setStatus]:740) - Set interpreter python status to READY
 INFO [2020-04-01 14:52:09,262] ({qtp701141022-13} InterpreterSetting.java[setStatus]:740) - Set interpreter jdbc status to DOWNLOADING_DEPENDENCIES
 INFO [2020-04-01 14:52:09,262] ({Thread-17} InterpreterSetting.java[setStatus]:740) - Set interpreter kotlin status to READY
 INFO [2020-04-01 14:52:09,262] ({qtp701141022-13} InterpreterSetting.java[setStatus]:740) - Set interpreter angular status to DOWNLOADING_DEPENDENCIES
 INFO [2020-04-01 14:52:09,262] ({Thread-18} InterpreterSetting.java[setStatus]:740) - Set interpreter jdbc status to READY
 INFO [2020-04-01 14:52:09,263] ({qtp701141022-13} InterpreterSetting.java[setStatus]:740) - Set interpreter livy status to DOWNLOADING_DEPENDENCIES
 INFO [2020-04-01 14:52:09,263] ({Thread-19} InterpreterSetting.java[setStatus]:740) - Set interpreter angular status to READY
 INFO [2020-04-01 14:52:09,264] ({qtp701141022-13} InterpreterSetting.java[setStatus]:740) - Set interpreter r status to DOWNLOADING_DEPENDENCIES
 INFO [2020-04-01 14:52:09,285] ({Thread-20} InterpreterSetting.java[setStatus]:740) - Set interpreter livy status to READY
 INFO [2020-04-01 14:52:09,285] ({Thread-21} InterpreterSetting.java[setStatus]:740) - Set interpreter r status to READY
 INFO [2020-04-01 14:52:09,285] ({qtp701141022-13} InterpreterSetting.java[setStatus]:740) - Set interpreter java status to DOWNLOADING_DEPENDENCIES
 INFO [2020-04-01 14:52:09,287] ({qtp701141022-13} InterpreterSetting.java[setStatus]:740) - Set interpreter jupyter status to DOWNLOADING_DEPENDENCIES
 INFO [2020-04-01 14:52:09,287] ({Thread-22} InterpreterSetting.java[setStatus]:740) - Set interpreter java status to READY
 INFO [2020-04-01 14:52:09,288] ({Thread-23} InterpreterSetting.java[setStatus]:740) - Set interpreter jupyter status to READY
 INFO [2020-04-01 14:52:09,288] ({qtp701141022-13} InterpreterSetting.java[setStatus]:740) - Set interpreter file status to DOWNLOADING_DEPENDENCIES
 INFO [2020-04-01 14:52:09,301] ({qtp701141022-13} InterpreterSetting.java[setStatus]:740) - Set interpreter groovy status to DOWNLOADING_DEPENDENCIES
 INFO [2020-04-01 14:52:09,301] ({Thread-24} InterpreterSetting.java[setStatus]:740) - Set interpreter file status to READY
 INFO [2020-04-01 14:52:09,301] ({qtp701141022-13} InterpreterSetting.java[setStatus]:740) - Set interpreter submarine status to DOWNLOADING_DEPENDENCIES
 INFO [2020-04-01 14:52:09,302] ({Thread-25} InterpreterSetting.java[setStatus]:740) - Set interpreter groovy status to READY
 INFO [2020-04-01 14:52:09,302] ({qtp701141022-13} InterpreterSetting.java[setStatus]:740) - Set interpreter spark status to DOWNLOADING_DEPENDENCIES
 INFO [2020-04-01 14:52:09,302] ({Thread-26} InterpreterSetting.java[setStatus]:740) - Set interpreter submarine status to READY
 INFO [2020-04-01 14:52:09,303] ({qtp701141022-13} InterpreterSetting.java[setStatus]:740) - Set interpreter sh status to DOWNLOADING_DEPENDENCIES
 INFO [2020-04-01 14:52:09,303] ({Thread-27} InterpreterSetting.java[setStatus]:740) - Set interpreter spark status to READY
 INFO [2020-04-01 14:52:09,304] ({qtp701141022-13} InterpreterSetting.java[setStatus]:740) - Set interpreter md status to DOWNLOADING_DEPENDENCIES
 INFO [2020-04-01 14:52:09,304] ({Thread-28} InterpreterSetting.java[setStatus]:740) - Set interpreter sh status to READY
 INFO [2020-04-01 14:52:09,304] ({qtp701141022-13} LocalConfigStorage.java[save]:59) - Save Interpreter Setting to /opt/zeppelin-server/conf/interpreter.json
 INFO [2020-04-01 14:52:09,304] ({Thread-29} InterpreterSetting.java[setStatus]:740) - Set interpreter md status to READY
 INFO [2020-04-01 14:52:09,350] ({qtp701141022-13} LuceneSearch.java[<init>]:94) - Use /tmp/zeppelin-index for storing lucene search index
 INFO [2020-04-01 14:52:09,467] ({qtp701141022-13} ShiroAuthenticationService.java[<init>]:70) - ShiroAuthenticationService is initialized
 INFO [2020-04-01 14:52:09,628] ({qtp701141022-13} LoginRestApi.java[postLogin]:234) - {"status":"OK","message":"","body":{"principal":"mvince","ticket":"c27568c2-d493-4a94-933f-3b5351bc3210","roles":"[]"}}
 INFO [2020-04-01 14:52:09,754] ({qtp701141022-15} StdSchedulerFactory.java[instantiate]:1184) - Using default implementation for ThreadExecutor
 INFO [2020-04-01 14:52:09,757] ({qtp701141022-15} SimpleThreadPool.java[initialize]:268) - Job execution threads will use class loader of thread: qtp701141022-15
 INFO [2020-04-01 14:52:09,770] ({qtp701141022-15} SchedulerSignalerImpl.java[<init>]:61) - Initialized Scheduler Signaller of type: class org.quartz.core.SchedulerSignalerImpl
 INFO [2020-04-01 14:52:09,771] ({qtp701141022-15} QuartzScheduler.java[<init>]:240) - Quartz Scheduler v.2.2.1 created.
 INFO [2020-04-01 14:52:09,772] ({qtp701141022-15} RAMJobStore.java[initialize]:155) - RAMJobStore initialized.
 INFO [2020-04-01 14:52:09,773] ({qtp701141022-15} QuartzScheduler.java[initialize]:305) - Scheduler meta-data: Quartz Scheduler (v2.2.1) 'DefaultQuartzScheduler' with instanceId 'NON_CLUSTERED'
  Scheduler class: 'org.quartz.core.QuartzScheduler' - running locally.
  NOT STARTED.
  Currently in standby mode.
  Number of jobs executed: 0
  Using thread pool 'org.quartz.simpl.SimpleThreadPool' - with 10 threads.
  Using job-store 'org.quartz.simpl.RAMJobStore' - which does not support persistence. and is not clustered.

 INFO [2020-04-01 14:52:09,774] ({qtp701141022-15} StdSchedulerFactory.java[instantiate]:1339) - Quartz scheduler 'DefaultQuartzScheduler' initialized from default resource file in Quartz package: 'quartz.properties'
 INFO [2020-04-01 14:52:09,774] ({qtp701141022-15} StdSchedulerFactory.java[instantiate]:1343) - Quartz scheduler version: 2.2.1
 INFO [2020-04-01 14:52:09,774] ({qtp701141022-15} QuartzScheduler.java[start]:575) - Scheduler DefaultQuartzScheduler_$_NON_CLUSTERED started.
 INFO [2020-04-01 14:52:09,775] ({Init CronJob Thread} QuartzSchedulerService.java[lambda$new$1]:65) - Starting init cronjobs
 INFO [2020-04-01 14:52:09,784] ({qtp701141022-15} Notebook.java[getNotesInfo]:557) - Start getNoteList
 INFO [2020-04-01 14:52:09,787] ({qtp701141022-15} Notebook.java[getNotesInfo]:581) - Finish getNoteList
 WARN [2020-04-01 14:52:09,815] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EFNUMC42 because its cron expression is empty.
 WARN [2020-04-01 14:52:09,847] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EA6SHT18 because its cron expression is empty.
 WARN [2020-04-01 14:52:09,860] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EGGA1EKB because its cron expression is empty.
 WARN [2020-04-01 14:52:09,869] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EF24ZWPW because its cron expression is empty.
 WARN [2020-04-01 14:52:09,874] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E8UCVQJR because its cron expression is empty.
 WARN [2020-04-01 14:52:09,883] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EJUT6WUF because its cron expression is empty.
 WARN [2020-04-01 14:52:09,888] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E97WF2EQ because its cron expression is empty.
 WARN [2020-04-01 14:52:09,893] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EP8ERVQK because its cron expression is empty.
 WARN [2020-04-01 14:52:09,905] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2ENZ4K8U8 because its cron expression is empty.
 INFO [2020-04-01 14:52:09,920] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:167) - Trigger cron for note:  loading_time_all_last, with cron expression: 0 16 10 * * ?
 WARN [2020-04-01 14:52:09,931] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EZ9G3JJU because its cron expression is empty.
 WARN [2020-04-01 14:52:09,932] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EB4N5R3F because its cron expression is empty.
 WARN [2020-04-01 14:52:09,935] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EDQW7GMK because its cron expression is empty.
 WARN [2020-04-01 14:52:09,950] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2BWJFTXKJ because its cron expression is empty.
 WARN [2020-04-01 14:52:09,966] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2BWJFTXKM because its cron expression is empty.
 INFO [2020-04-01 14:52:09,972] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:167) - Trigger cron for note:  profile_extended_last, with cron expression: 0 41 9 * * ?
 WARN [2020-04-01 14:52:09,975] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EPMFU6DA because its cron expression is empty.
 WARN [2020-04-01 14:52:09,977] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EKF8N63T because its cron expression is empty.
 WARN [2020-04-01 14:52:09,978] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EFDWHZB1 because its cron expression is empty.
 WARN [2020-04-01 14:52:09,989] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EDST5HC1 because its cron expression is empty.
 WARN [2020-04-01 14:52:09,998] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EM5CEYTJ because its cron expression is empty.
 INFO [2020-04-01 14:52:09,999] ({qtp701141022-19} InterpreterSetting.java[getOrCreateInterpreterGroup]:466) - Create InterpreterGroup with groupId: spark-mvince for user: mvince and note: 2ET5BSCWB
 INFO [2020-04-01 14:52:10,002] ({qtp701141022-19} InterpreterSetting.java[createInterpreters]:821) - Interpreter org.apache.zeppelin.spark.SparkInterpreter created for user: mvince, sessionId: shared_session
 INFO [2020-04-01 14:52:10,002] ({qtp701141022-19} InterpreterSetting.java[createInterpreters]:821) - Interpreter org.apache.zeppelin.spark.SparkSqlInterpreter created for user: mvince, sessionId: shared_session
 INFO [2020-04-01 14:52:10,002] ({qtp701141022-19} InterpreterSetting.java[createInterpreters]:821) - Interpreter org.apache.zeppelin.spark.PySparkInterpreter created for user: mvince, sessionId: shared_session
 INFO [2020-04-01 14:52:10,002] ({qtp701141022-19} InterpreterSetting.java[createInterpreters]:821) - Interpreter org.apache.zeppelin.spark.IPySparkInterpreter created for user: mvince, sessionId: shared_session
 INFO [2020-04-01 14:52:10,002] ({qtp701141022-19} InterpreterSetting.java[createInterpreters]:821) - Interpreter org.apache.zeppelin.spark.SparkRInterpreter created for user: mvince, sessionId: shared_session
 INFO [2020-04-01 14:52:10,002] ({qtp701141022-19} InterpreterSetting.java[createInterpreters]:821) - Interpreter org.apache.zeppelin.spark.SparkIRInterpreter created for user: mvince, sessionId: shared_session
 INFO [2020-04-01 14:52:10,002] ({qtp701141022-19} InterpreterSetting.java[createInterpreters]:821) - Interpreter org.apache.zeppelin.spark.SparkShinyInterpreter created for user: mvince, sessionId: shared_session
 INFO [2020-04-01 14:52:10,002] ({qtp701141022-19} InterpreterSetting.java[createInterpreters]:821) - Interpreter org.apache.zeppelin.spark.KotlinSparkInterpreter created for user: mvince, sessionId: shared_session
 INFO [2020-04-01 14:52:10,003] ({qtp701141022-19} ManagedInterpreterGroup.java[getOrCreateSession]:174) - Create Session: shared_session in InterpreterGroup: spark-mvince for user: mvince
 INFO [2020-04-01 14:52:10,003] ({qtp701141022-19} InterpreterSetting.java[getOrCreateInterpreterGroup]:466) - Create InterpreterGroup with groupId: livy-shared_process for user: mvince and note: 2ET5BSCWB
 INFO [2020-04-01 14:52:10,003] ({qtp701141022-19} InterpreterSetting.java[createInterpreters]:821) - Interpreter org.apache.zeppelin.livy.LivySparkInterpreter created for user: mvince, sessionId: mvince
 INFO [2020-04-01 14:52:10,003] ({qtp701141022-19} InterpreterSetting.java[createInterpreters]:821) - Interpreter org.apache.zeppelin.livy.LivySparkSQLInterpreter created for user: mvince, sessionId: mvince
 INFO [2020-04-01 14:52:10,003] ({qtp701141022-19} InterpreterSetting.java[createInterpreters]:821) - Interpreter org.apache.zeppelin.livy.LivyPySparkInterpreter created for user: mvince, sessionId: mvince
 INFO [2020-04-01 14:52:10,004] ({qtp701141022-19} InterpreterSetting.java[createInterpreters]:821) - Interpreter org.apache.zeppelin.livy.LivyPySpark3Interpreter created for user: mvince, sessionId: mvince
 INFO [2020-04-01 14:52:10,004] ({qtp701141022-19} InterpreterSetting.java[createInterpreters]:821) - Interpreter org.apache.zeppelin.livy.LivySparkRInterpreter created for user: mvince, sessionId: mvince
 INFO [2020-04-01 14:52:10,004] ({qtp701141022-19} InterpreterSetting.java[createInterpreters]:821) - Interpreter org.apache.zeppelin.livy.LivySharedInterpreter created for user: mvince, sessionId: mvince
 INFO [2020-04-01 14:52:10,004] ({qtp701141022-19} ManagedInterpreterGroup.java[getOrCreateSession]:174) - Create Session: mvince in InterpreterGroup: livy-shared_process for user: mvince
 WARN [2020-04-01 14:52:10,007] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EA84Z46N because its cron expression is empty.
 WARN [2020-04-01 14:52:10,012] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EJVKCCHQ because its cron expression is empty.
 WARN [2020-04-01 14:52:10,015] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EYT7Q6R8 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,020] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2ECMC9HSM because its cron expression is empty.
 WARN [2020-04-01 14:52:10,028] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EG4PG1WY because its cron expression is empty.
 WARN [2020-04-01 14:52:10,036] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2F2AVWJ77 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,036] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EBEQAYV2 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,039] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2F1CHQ4TT because its cron expression is empty.
 WARN [2020-04-01 14:52:10,041] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E8CQ9MS6 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,041] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2ET5BSCWB because its cron expression is empty.
 WARN [2020-04-01 14:52:10,046] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EN2D3N9X because its cron expression is empty.
 WARN [2020-04-01 14:52:10,050] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EBJZ7B3G because its cron expression is empty.
 WARN [2020-04-01 14:52:10,050] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EBQRAERC because its cron expression is empty.
 WARN [2020-04-01 14:52:10,052] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EP8AG4BU because its cron expression is empty.
 WARN [2020-04-01 14:52:10,055] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EF6X5CJF because its cron expression is empty.
 INFO [2020-04-01 14:52:10,063] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:167) - Trigger cron for note:  metrics_d0_last, with cron expression: 0 55 9 * * ?
 WARN [2020-04-01 14:52:10,067] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EJ23SKTS because its cron expression is empty.
 WARN [2020-04-01 14:52:10,070] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EK8SFBGG because its cron expression is empty.
 WARN [2020-04-01 14:52:10,072] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EE649P2Y because its cron expression is empty.
 INFO [2020-04-01 14:52:10,076] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:167) - Trigger cron for note: , with cron expression: 0 15 05 ? * *
 WARN [2020-04-01 14:52:10,081] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2ER62Y5VJ because its cron expression is empty.
 WARN [2020-04-01 14:52:10,084] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EVBRSEEH because its cron expression is empty.
 WARN [2020-04-01 14:52:10,084] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EA62S7HB because its cron expression is empty.
 WARN [2020-04-01 14:52:10,086] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E67RTQ67 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,088] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E61STQJ5 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,094] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EM2DXA7V because its cron expression is empty.
 WARN [2020-04-01 14:52:10,100] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EGD129A5 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,104] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2F2YS7PCE because its cron expression is empty.
 WARN [2020-04-01 14:52:10,110] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EAFDAS3E because its cron expression is empty.
 WARN [2020-04-01 14:52:10,112] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E6V3FRRP because its cron expression is empty.
 WARN [2020-04-01 14:52:10,114] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E9JUE7SP because its cron expression is empty.
 WARN [2020-04-01 14:52:10,119] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EC93CJFY because its cron expression is empty.
 INFO [2020-04-01 14:52:10,122] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:167) - Trigger cron for note:  metrics_by_dig_last, with cron expression: 0 50 9 * * ?
 WARN [2020-04-01 14:52:10,127] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EFQN2E5H because its cron expression is empty.
 WARN [2020-04-01 14:52:10,130] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2C57UKYWR because its cron expression is empty.
 WARN [2020-04-01 14:52:10,148] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2ED5VNXNT because its cron expression is empty.
 WARN [2020-04-01 14:52:10,149] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EFTZ976C because its cron expression is empty.
 WARN [2020-04-01 14:52:10,152] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EHAVGUBC because its cron expression is empty.
 WARN [2020-04-01 14:52:10,157] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EDZWEKC4 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,159] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EJZYWBKU because its cron expression is empty.
 WARN [2020-04-01 14:52:10,159] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EK8HTF88 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,193] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2F1S9ZY8Z because its cron expression is empty.
 WARN [2020-04-01 14:52:10,194] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EASTANXY because its cron expression is empty.
 WARN [2020-04-01 14:52:10,196] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EG2SXMMM because its cron expression is empty.
 WARN [2020-04-01 14:52:10,200] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EK2KDS2Z because its cron expression is empty.
 WARN [2020-04-01 14:52:10,202] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EF19QR3K because its cron expression is empty.
 WARN [2020-04-01 14:52:10,202] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EC2RVHSA because its cron expression is empty.
 WARN [2020-04-01 14:52:10,206] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EMJUV66E because its cron expression is empty.
 WARN [2020-04-01 14:52:10,208] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EBTQJJDQ because its cron expression is empty.
 WARN [2020-04-01 14:52:10,209] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EATUNUKB because its cron expression is empty.
 WARN [2020-04-01 14:52:10,213] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EYDJKFFY because its cron expression is empty.
 WARN [2020-04-01 14:52:10,219] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2C35YU814 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,219] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EA55FQHT because its cron expression is empty.
 WARN [2020-04-01 14:52:10,223] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EQNNCYC9 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,225] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EZ66TM57 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,225] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E97C8D5T because its cron expression is empty.
 WARN [2020-04-01 14:52:10,228] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EPZWH987 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,235] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EAUE9CX7 because its cron expression is empty.
 INFO [2020-04-01 14:52:10,239] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:167) - Trigger cron for note:  loading_time_first_last, with cron expression: 0 10 10 * * ?
 WARN [2020-04-01 14:52:10,241] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E7XHUXJK because its cron expression is empty.
 WARN [2020-04-01 14:52:10,246] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2C2AUG798 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,247] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EPKD684J because its cron expression is empty.
 WARN [2020-04-01 14:52:10,250] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EPANGK4J because its cron expression is empty.
 WARN [2020-04-01 14:52:10,252] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EW4HM8XQ because its cron expression is empty.
 WARN [2020-04-01 14:52:10,253] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EJ798ABT because its cron expression is empty.
 WARN [2020-04-01 14:52:10,265] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EKDR3PT7 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,270] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2BYEZ5EVK because its cron expression is empty.
 WARN [2020-04-01 14:52:10,276] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E72RM55E because its cron expression is empty.
 WARN [2020-04-01 14:52:10,283] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EJNNR6DK because its cron expression is empty.
 WARN [2020-04-01 14:52:10,286] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EHMR1ADA because its cron expression is empty.
 WARN [2020-04-01 14:52:10,291] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EMCU8KAC because its cron expression is empty.
 WARN [2020-04-01 14:52:10,300] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EWM84JXA because its cron expression is empty.
 WARN [2020-04-01 14:52:10,304] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EN1E1ATY because its cron expression is empty.
 WARN [2020-04-01 14:52:10,309] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EPP67VKS because its cron expression is empty.
 WARN [2020-04-01 14:52:10,357] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EPATQQDT because its cron expression is empty.
 INFO [2020-04-01 14:52:10,361] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:167) - Trigger cron for note:  fleet_overview_last, with cron expression: 0 35 9 * * ?
 WARN [2020-04-01 14:52:10,365] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EG3KSZ91 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,371] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EX6S4QSD because its cron expression is empty.
 WARN [2020-04-01 14:52:10,373] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EU19WKPQ because its cron expression is empty.
 WARN [2020-04-01 14:52:10,375] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EAGE4GN7 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,380] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2ESRDQTJP because its cron expression is empty.
 WARN [2020-04-01 14:52:10,387] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EW19CSPA because its cron expression is empty.
 WARN [2020-04-01 14:52:10,392] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EJE2WJK4 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,395] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EJJBKHPH because its cron expression is empty.
 WARN [2020-04-01 14:52:10,398] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EQ5G471X because its cron expression is empty.
 WARN [2020-04-01 14:52:10,401] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E859M6CV because its cron expression is empty.
 WARN [2020-04-01 14:52:10,403] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2ENQDFSC3 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,408] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EVH8FBP5 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,410] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EM8RB72W because its cron expression is empty.
 WARN [2020-04-01 14:52:10,410] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EBPNK6HD because its cron expression is empty.
 WARN [2020-04-01 14:52:10,414] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EZFM3GJA because its cron expression is empty.
 WARN [2020-04-01 14:52:10,415] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E9KJ2421 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,415] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EGNSJ24Y because its cron expression is empty.
 WARN [2020-04-01 14:52:10,419] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2ENGYC9FS because its cron expression is empty.
 WARN [2020-04-01 14:52:10,420] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E9CQC4EN because its cron expression is empty.
 WARN [2020-04-01 14:52:10,431] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2ETDNWNDD because its cron expression is empty.
 WARN [2020-04-01 14:52:10,433] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EKTB3H2B because its cron expression is empty.
 WARN [2020-04-01 14:52:10,434] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EFW2526S because its cron expression is empty.
 WARN [2020-04-01 14:52:10,437] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EQTYFAVQ because its cron expression is empty.
 WARN [2020-04-01 14:52:10,441] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2ECQJEPE3 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,445] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EKMN5G7V because its cron expression is empty.
 WARN [2020-04-01 14:52:10,447] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EFZR463V because its cron expression is empty.
 WARN [2020-04-01 14:52:10,449] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EJS1HGKH because its cron expression is empty.
 WARN [2020-04-01 14:52:10,449] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EBT34D6C because its cron expression is empty.
 INFO [2020-04-01 14:52:10,451] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:167) - Trigger cron for note:  levelup_last, with cron expression: 0 00 10 * * ?
 WARN [2020-04-01 14:52:10,453] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EHX3MGHH because its cron expression is empty.
 WARN [2020-04-01 14:52:10,456] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E5P699H6 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,457] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2E7WFV4CG because its cron expression is empty.
 WARN [2020-04-01 14:52:10,460] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EQMM2A9W because its cron expression is empty.
 WARN [2020-04-01 14:52:10,465] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2ENDE3NH9 because its cron expression is empty.
 WARN [2020-04-01 14:52:10,466] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EYD56B9B because its cron expression is empty.
 INFO [2020-04-01 14:52:10,469] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:167) - Trigger cron for note:  tutorial_last, with cron expression: 0 20 10 * * ?
 WARN [2020-04-01 14:52:10,470] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2ECHQRNZA because its cron expression is empty.
 WARN [2020-04-01 14:52:10,472] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EGT4EH9G because its cron expression is empty.
 WARN [2020-04-01 14:52:10,473] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2ERA45BAY because its cron expression is empty.
 WARN [2020-04-01 14:52:10,476] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2A94M5J1Z because its cron expression is empty.
 WARN [2020-04-01 14:52:10,481] ({Init CronJob Thread} QuartzSchedulerService.java[refreshCron]:132) - Skip refresh cron of note 2EYUV26VR because its cron expression is empty.
 INFO [2020-04-01 14:52:10,481] ({Init CronJob Thread} QuartzSchedulerService.java[lambda$new$1]:81) - Complete init cronjobs
 WARN [2020-04-01 14:52:11,107] ({qtp701141022-19} NotebookServer.java[onFailure]:2041) - HTTP 404 Not Found
 WARN [2020-04-01 14:52:11,154] ({qtp701141022-18} NotebookServer.java[onFailure]:2041) - HTTP 404 Not Found
 INFO [2020-04-01 14:52:26,883] ({qtp701141022-20} Helium.java[loadConf]:140) - Add helium local registry /opt/zeppelin-server/helium
 WARN [2020-04-01 14:52:26,884] ({qtp701141022-20} Helium.java[loadConf]:148) - /opt/zeppelin-server/conf/helium.json does not exists
 INFO [2020-04-01 14:52:29,117] ({qtp701141022-19} NotebookService.java[runParagraph]:293) - Start to run paragraph: paragraph_1571833609109_-338708939 of note: 2ET5BSCWB
 INFO [2020-04-01 14:52:29,118] ({qtp701141022-19} VFSNotebookRepo.java[save]:145) - Saving note 2ET5BSCWB to mvince/spark_test_2ET5BSCWB.zpln
 INFO [2020-04-01 14:52:29,139] ({qtp701141022-19} SchedulerFactory.java[<init>]:62) - Scheduler Thread Pool Size: 100
 INFO [2020-04-01 14:52:29,156] ({SchedulerFactory2} AbstractScheduler.java[runJob]:125) - Job paragraph_1571833609109_-338708939 started by scheduler RemoteInterpreter-spark-mvince-shared_session
 INFO [2020-04-01 14:52:29,157] ({SchedulerFactory2} Paragraph.java[jobRun]:407) - Run paragraph [paragraph_id: paragraph_1571833609109_-338708939, interpreter: org.apache.zeppelin.spark.SparkInterpreter, note_id: 2ET5BSCWB, user: mvince]
 INFO [2020-04-01 14:52:29,157] ({SchedulerFactory2} ManagedInterpreterGroup.java[getOrCreateInterpreterProcess]:62) - Create InterpreterProcess for InterpreterGroup: spark-mvince
 INFO [2020-04-01 14:52:29,159] ({SchedulerFactory2} PluginManager.java[loadInterpreterLauncher]:141) - Loading Interpreter Launcher Plugin: SparkInterpreterLauncher
 INFO [2020-04-01 14:52:29,163] ({SchedulerFactory2} StandardInterpreterLauncher.java[launch]:49) - Launching Interpreter: spark
 INFO [2020-04-01 14:52:29,944] ({SchedulerFactory2} SparkInterpreterLauncher.java[buildEnvFromProperties]:182) - Run Spark under secure mode with keytab: /etc/security/keytabs/zeppelin.server.kerberos.keytab, principal: zeppelin-dwh_prod_ht@mynetwork.LOCAL
 INFO [2020-04-01 14:52:29,959] ({SchedulerFactory2} ProcessLauncher.java[transition]:109) - Process state is transitioned to LAUNCHED
 INFO [2020-04-01 14:52:29,959] ({SchedulerFactory2} ProcessLauncher.java[launch]:96) - Process is launched: [/opt/zeppelin-server/bin/interpreter.sh, -d, /opt/zeppelin-server/interpreter/spark, -c, 192.168.200.65, -p, 38198, -r, :, -i, spark-mvince, -u, mvince, -l, /opt/zeppelin-server/local-repo/spark, -g, spark]
 INFO [2020-04-01 14:52:29,999] ({Exec Stream Pumper} ProcessLauncher.java[processLine]:188) - Interpreter launch command:  /usr/hdp/2.6.5.0-292/spark2/bin/spark-submit --class org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer --driver-class-path ":/opt/zeppelin-server/interpreter/spark/*::/opt/zeppelin-server/interpreter/zeppelin-interpreter-shaded-0.9.0-SNAPSHOT.jar:/opt/zeppelin-server/interpreter/spark/spark-interpreter-0.9.0-SNAPSHOT.jar:/usr/hdp/2.6.5.0-292/spark2/jars/*:/etc/hadoop/conf" --driver-java-options " -Dfile.encoding=UTF-8 -Dlog4j.configuration=log4j_yarn_cluster.properties -Dzeppelin.log.file='/opt/zeppelin-server/logs/zeppelin-interpreter-spark-mvince-mvince-zeppelin-dwh-prod-ht-edge-05.mynetwork-dc.local.log'" --master yarn --conf spark\.yarn\.dist\.archives\=\/usr\/hdp\/2\.6\.5\.0-292\/spark2\/R\/lib\/sparkr\.zip\#sparkr --conf spark\.yarn\.isPython\=true --conf spark\.app\.name\=test_drive --conf spark\.webui\.yarn\.useProxy\=false --conf spark\.driver\.cores\=1 --conf spark\.yarn\.maxAppAttempts\=1 --conf spark\.executor\.memory\=1g --conf spark\.files\=\/opt\/zeppelin-server\/conf\/log4j_yarn_cluster\.properties --conf spark\.driver\.memory\=1g --conf spark\.jars\=\/opt\/zeppelin-server\/interpreter\/spark\/scala-2\.11\/spark-scala-2\.11-0\.9\.0-SNAPSHOT\.jar\,\/opt\/zeppelin-server\/interpreter\/zeppelin-interpreter-shaded-0\.9\.0-SNAPSHOT\.jar --conf spark\.submit\.deployMode\=cluster --conf spark\.executor\.cores\=1 --conf spark\.yarn\.submit\.waitAppCompletion\=false --proxy-user mvince /opt/zeppelin-server/interpreter/spark/spark-interpreter-0.9.0-SNAPSHOT.jar 192.168.200.65 38198 "spark-mvince" :
 WARN [2020-04-01 14:52:35,887] ({Exec Default Executor} RemoteInterpreterManagedProcess.java[onProcessComplete]:263) - Process is exited with exit value 0
 INFO [2020-04-01 14:52:35,887] ({Exec Default Executor} ProcessLauncher.java[transition]:109) - Process state is transitioned to COMPLETED
 INFO [2020-04-01 14:52:41,903] ({pool-6-thread-1} ProcessLauncher.java[transition]:109) - Process state is transitioned to RUNNING
 INFO [2020-04-01 14:52:41,960] ({SchedulerFactory2} RemoteInterpreter.java[call]:171) - Create RemoteInterpreter org.apache.zeppelin.spark.SparkInterpreter
 INFO [2020-04-01 14:52:42,215] ({SchedulerFactory2} RemoteInterpreter.java[call]:171) - Create RemoteInterpreter org.apache.zeppelin.spark.SparkSqlInterpreter
 INFO [2020-04-01 14:52:42,218] ({SchedulerFactory2} RemoteInterpreter.java[call]:171) - Create RemoteInterpreter org.apache.zeppelin.spark.PySparkInterpreter
 INFO [2020-04-01 14:52:42,224] ({SchedulerFactory2} RemoteInterpreter.java[call]:171) - Create RemoteInterpreter org.apache.zeppelin.spark.IPySparkInterpreter
 INFO [2020-04-01 14:52:42,232] ({SchedulerFactory2} RemoteInterpreter.java[call]:171) - Create RemoteInterpreter org.apache.zeppelin.spark.SparkRInterpreter
 INFO [2020-04-01 14:52:42,237] ({SchedulerFactory2} RemoteInterpreter.java[call]:171) - Create RemoteInterpreter org.apache.zeppelin.spark.SparkIRInterpreter
 INFO [2020-04-01 14:52:42,240] ({SchedulerFactory2} RemoteInterpreter.java[call]:171) - Create RemoteInterpreter org.apache.zeppelin.spark.SparkShinyInterpreter
 INFO [2020-04-01 14:52:42,244] ({SchedulerFactory2} RemoteInterpreter.java[call]:171) - Create RemoteInterpreter org.apache.zeppelin.spark.KotlinSparkInterpreter
 INFO [2020-04-01 14:52:42,347] ({SchedulerFactory2} RemoteInterpreter.java[call]:141) - Open RemoteInterpreter org.apache.zeppelin.spark.SparkInterpreter
 INFO [2020-04-01 14:52:42,347] ({SchedulerFactory2} RemoteInterpreter.java[pushAngularObjectRegistryToRemote]:431) - Push local angular object registry from ZeppelinServer to remote interpreter group spark-mvince
 WARN [2020-04-01 14:52:44,434] ({SchedulerFactory2} NotebookServer.java[onStatusChange]:1901) - Job paragraph_1571833609109_-338708939 is finished, status: ERROR, exception: null, result: %text org.apache.zeppelin.interpreter.InterpreterException: org.apache.zeppelin.interpreter.InterpreterException: Fail to open SparkInterpreter
	at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:76)
	at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:668)
	at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:577)
	at org.apache.zeppelin.scheduler.Job.run(Job.java:172)
	at org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:130)
	at org.apache.zeppelin.scheduler.FIFOScheduler.lambda$runJobInScheduler$0(FIFOScheduler.java:39)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.zeppelin.interpreter.InterpreterException: Fail to open SparkInterpreter
	at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:114)
	at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:70)
	... 8 more
Caused by: scala.reflect.internal.MissingRequirementError: object java.lang.Object in compiler mirror not found.
	at scala.reflect.internal.MissingRequirementError$.signal(MissingRequirementError.scala:17)
	at scala.reflect.internal.MissingRequirementError$.notFound(MissingRequirementError.scala:18)
	at scala.reflect.internal.Mirrors$RootsBase.getModuleOrClass(Mirrors.scala:53)
	at scala.reflect.internal.Mirrors$RootsBase.getModuleOrClass(Mirrors.scala:45)
	at scala.reflect.internal.Mirrors$RootsBase.getModuleOrClass(Mirrors.scala:45)
	at scala.reflect.internal.Mirrors$RootsBase.getModuleOrClass(Mirrors.scala:66)
	at scala.reflect.internal.Mirrors$RootsBase.getClassByName(Mirrors.scala:102)
	at scala.reflect.internal.Mirrors$RootsBase.getRequiredClass(Mirrors.scala:105)
	at scala.reflect.internal.Definitions$DefinitionsClass.ObjectClass$lzycompute(Definitions.scala:257)
	at scala.reflect.internal.Definitions$DefinitionsClass.ObjectClass(Definitions.scala:257)
	at scala.reflect.internal.Definitions$DefinitionsClass.init(Definitions.scala:1394)
	at scala.tools.nsc.Global$Run.<init>(Global.scala:1215)
	at scala.tools.nsc.interpreter.IMain.compileSourcesKeepingRun(IMain.scala:432)
	at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.compileAndSaveRun(IMain.scala:855)
	at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.compile(IMain.scala:813)
	at scala.tools.nsc.interpreter.IMain.bind(IMain.scala:675)
	at scala.tools.nsc.interpreter.IMain.bind(IMain.scala:712)
	at scala.tools.nsc.interpreter.IMain$$anonfun$quietBind$1.apply(IMain.scala:711)
	at scala.tools.nsc.interpreter.IMain$$anonfun$quietBind$1.apply(IMain.scala:711)
	at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:214)
	at scala.tools.nsc.interpreter.IMain.quietBind(IMain.scala:711)
	at org.apache.zeppelin.spark.SparkScala211Interpreter$.loopPostInit$1(SparkScala211Interpreter.scala:164)
	at org.apache.zeppelin.spark.SparkScala211Interpreter$.org$apache$zeppelin$spark$SparkScala211Interpreter$$loopPostInit(SparkScala211Interpreter.scala:199)
	at org.apache.zeppelin.spark.SparkScala211Interpreter.open(SparkScala211Interpreter.scala:95)
	at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:98)
	... 9 more

 INFO [2020-04-01 14:52:44,435] ({SchedulerFactory2} VFSNotebookRepo.java[save]:145) - Saving note 2ET5BSCWB to mvince/spark_test_2ET5BSCWB.zpln
 INFO [2020-04-01 14:52:44,438] ({SchedulerFactory2} AbstractScheduler.java[runJob]:152) - Job paragraph_1571833609109_-338708939 finished by scheduler RemoteInterpreter-spark-mvince-shared_session
lrwxrwxrwx 1 mvince hadoop   88 Apr  1 14:52 __app__.jar -> /data/00/yarn/local/usercache/mvince/filecache/1883/spark-interpreter-0.9.0-SNAPSHOT.jar
-rw------- 1 mvince hadoop  364 Apr  1 14:52 container_tokens
-rwx------ 1 mvince hadoop 8177 Apr  1 14:52 launch_container.sh
lrwxrwxrwx 1 mvince hadoop   81 Apr  1 14:52 log4j_yarn_cluster.properties -> /data/00/yarn/local/usercache/mvince/filecache/1878/log4j_yarn_cluster.properties
lrwxrwxrwx 1 mvince hadoop   71 Apr  1 14:52 py4j-0.10.6-src.zip -> /data/02/yarn/local/usercache/mvince/filecache/1880/py4j-0.10.6-src.zip
lrwxrwxrwx 1 mvince hadoop   63 Apr  1 14:52 pyspark.zip -> /data/01/yarn/local/usercache/mvince/filecache/1879/pyspark.zip
lrwxrwxrwx 1 mvince hadoop   70 Apr  1 14:52 __spark_conf__ -> /data/04/yarn/local/usercache/mvince/filecache/1882/__spark_conf__.zip
drwxr-s--- 2 mvince hadoop 4096 Apr  1 14:52 __spark_libs__
lrwxrwxrwx 1 mvince hadoop   62 Apr  1 14:52 sparkr -> /data/04/yarn/local/usercache/mvince/filecache/1877/sparkr.zip
lrwxrwxrwx 1 mvince hadoop   87 Apr  1 14:52 spark-scala-2.11-0.9.0-SNAPSHOT.jar -> /data/03/yarn/local/usercache/mvince/filecache/1881/spark-scala-2.11-0.9.0-SNAPSHOT.jar
drwxr-s--- 2 mvince hadoop 4096 Apr  1 14:52 tmp
lrwxrwxrwx 1 mvince hadoop   98 Apr  1 14:52 zeppelin-interpreter-shaded-0.9.0-SNAPSHOT.jar -> /data/01/yarn/local/usercache/mvince/filecache/1884/zeppelin-interpreter-shaded-0.9.0-SNAPSHOT.jar
find -L . -maxdepth 5 -ls:
 61343077      4 drwxr-s---   4 mvince   hadoop       4096 Apr  1 14:52 .
 61343078      8 -rwx------   1 mvince   hadoop       8177 Apr  1 14:52 ./launch_container.sh
116262286      4 drwx------   3 mvince   mvince       4096 Apr  1 14:52 ./sparkr
116393329      4 drwx------   9 mvince   mvince       4096 Apr  1 14:52 ./sparkr/SparkR
117048744      4 drwx------   2 mvince   mvince       4096 Apr  1 14:52 ./sparkr/SparkR/profile
117048745      4 -r-x------   1 mvince   mvince        958 Apr  1 14:52 ./sparkr/SparkR/profile/general.R
117048746      4 -r-x------   1 mvince   mvince       1715 Apr  1 14:52 ./sparkr/SparkR/profile/shell.R
116655509      4 drwx------   3 mvince   mvince       4096 Apr  1 14:52 ./sparkr/SparkR/tests
116786560      4 drwx------   2 mvince   mvince       4096 Apr  1 14:52 ./sparkr/SparkR/tests/testthat
116786561      4 -r-x------   1 mvince   mvince       3561 Apr  1 14:52 ./sparkr/SparkR/tests/testthat/test_basic.R
117441940      4 drwx------   2 mvince   mvince       4096 Apr  1 14:52 ./sparkr/SparkR/help
117441944      4 -r-x------   1 mvince   mvince         60 Apr  1 14:52 ./sparkr/SparkR/help/aliases.rds
117441942      4 -r-x------   1 mvince   mvince        113 Apr  1 14:52 ./sparkr/SparkR/help/SparkR.rdx
117441941      0 -r-x------   1 mvince   mvince          0 Apr  1 14:52 ./sparkr/SparkR/help/AnIndex
117441943      0 -r-x------   1 mvince   mvince          0 Apr  1 14:52 ./sparkr/SparkR/help/SparkR.rdb
117441945      4 -r-x------   1 mvince   mvince         64 Apr  1 14:52 ./sparkr/SparkR/help/paths.rds
117310871      4 drwx------   2 mvince   mvince       4096 Apr  1 14:52 ./sparkr/SparkR/html
117310873      4 -r-x------   1 mvince   mvince        937 Apr  1 14:52 ./sparkr/SparkR/html/00Index.html
117310872      4 -r-x------   1 mvince   mvince       1319 Apr  1 14:52 ./sparkr/SparkR/html/R.css
116524385      4 drwx------   2 mvince   mvince       4096 Apr  1 14:52 ./sparkr/SparkR/Meta
116524387      4 -r-x------   1 mvince   mvince       2929 Apr  1 14:52 ./sparkr/SparkR/Meta/nsInfo.rds
116524386      4 -r-x------   1 mvince   mvince       1151 Apr  1 14:52 ./sparkr/SparkR/Meta/package.rds
116524388      4 -r-x------   1 mvince   mvince        188 Apr  1 14:52 ./sparkr/SparkR/Meta/hsearch.rds
116524390      4 -r-x------   1 mvince   mvince         38 Apr  1 14:52 ./sparkr/SparkR/Meta/links.rds
116524389      4 -r-x------   1 mvince   mvince        164 Apr  1 14:52 ./sparkr/SparkR/Meta/Rd.rds
116917684      4 drwx------   2 mvince   mvince       4096 Apr  1 14:52 ./sparkr/SparkR/worker
116917686      4 -r-x------   1 mvince   mvince       3991 Apr  1 14:52 ./sparkr/SparkR/worker/daemon.R
116917685     12 -r-x------   1 mvince   mvince       8949 Apr  1 14:52 ./sparkr/SparkR/worker/worker.R
116393330      4 -r-x------   1 mvince   mvince       1617 Apr  1 14:52 ./sparkr/SparkR/DESCRIPTION
116393331     16 -r-x------   1 mvince   mvince      12807 Apr  1 14:52 ./sparkr/SparkR/NAMESPACE
117179801      4 drwx------   2 mvince   mvince       4096 Apr  1 14:52 ./sparkr/SparkR/R
117179802     28 -r-x------   1 mvince   mvince      26662 Apr  1 14:52 ./sparkr/SparkR/R/SparkR.rdx
117179803   1084 -r-x------   1 mvince   mvince    1109793 Apr  1 14:52 ./sparkr/SparkR/R/SparkR.rdb
117179804      4 -r-x------   1 mvince   mvince       1056 Apr  1 14:52 ./sparkr/SparkR/R/SparkR
 61212086    532 -r-x------   1 mvince   mvince     541536 Apr  1 14:52 ./pyspark.zip
 11404735     76 -r-x------   1 mvince   mvince      75035 Apr  1 14:52 ./spark-scala-2.11-0.9.0-SNAPSHOT.jar
118100066      4 -r-x------   1 mvince   mvince       1019 Apr  1 14:52 ./log4j_yarn_cluster.properties
118362129  61100 -r-x------   1 mvince   mvince   62565313 Apr  1 14:52 ./__app__.jar
 18351295     80 -r-x------   1 mvince   mvince      80352 Apr  1 14:52 ./py4j-0.10.6-src.zip
 61736356      4 drwxr-s---   2 mvince   hadoop       4096 Apr  1 14:52 ./tmp
 61867326      4 drwxr-s---   2 mvince   hadoop       4096 Apr  1 14:52 ./__spark_libs__
  7340053    168 -r-xr-xr-x   1 yarn     hadoop     168693 Jan 11  2019 ./__spark_libs__/logstash-gelf-1.12.0.jar
110887458      4 -r-xr-xr-x   1 yarn     hadoop          2 Jun 29  2018 ./__spark_libs__/dummy.jar
117835091      4 drwx------   3 mvince   mvince       4096 Apr  1 14:52 ./__spark_conf__
117966262      4 drwx------   2 mvince   mvince       4096 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__
117966299      4 -r-x------   1 mvince   mvince        984 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/mapred-env.sh
117966266      4 -r-x------   1 mvince   mvince       2404 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/hadoop-metrics2.properties
117966304      4 -r-x------   1 mvince   mvince       2316 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/ssl-client.xml.example
117966282      4 -r-x------   1 mvince   mvince       1335 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/configuration.xsl
117966308      4 -r-x------   1 mvince   mvince       1000 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/ssl-server.xml
117966268     12 -r-x------   1 mvince   mvince       8994 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/core-site.xml
117966291     12 -r-x------   1 mvince   mvince      10938 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/hdfs-site.xml
117966294      4 -r-x------   1 mvince   mvince        320 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/hdfs_nn_jaas.conf
117966263      8 -r-x------   1 mvince   mvince       6586 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/hadoop-env.sh
117966305      4 -r-x------   1 mvince   mvince       1308 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/hadoop-policy.xml
117966297      8 -r-x------   1 mvince   mvince       4221 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/task-log4j.properties
117966301      4 -r-x------   1 mvince   mvince        307 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/mapred_jaas.conf
117966296      4 -r-x------   1 mvince   mvince        545 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/yarn_jaas.conf
117966270      4 -r-x------   1 mvince   mvince        884 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/ssl-client.xml
117966267     28 -r-x------   1 mvince   mvince      24889 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/yarn-site.xml
117966295      4 -r-x------   1 mvince   mvince       1249 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/container-executor.cfg
117966311      4 -r-x------   1 mvince   mvince        945 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/taskcontroller.cfg
117966287      8 -r-x------   1 mvince   mvince       5522 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/yarn-env.sh
117966279      4 -r-x------   1 mvince   mvince        320 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/yarn_nm_jaas.conf
117966292      4 -r-x------   1 mvince   mvince        320 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/hdfs_dn_jaas.conf
117966300      4 -r-x------   1 mvince   mvince       1602 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/health_check
117966310      4 -r-x------   1 mvince   mvince       2697 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/ssl-server.xml.example
117966278      4 -r-x------   1 mvince   mvince       2358 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/topology_script.py
117966309      4 -r-x------   1 mvince   mvince        320 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/hdfs_jn_jaas.conf
117966306      4 -r-x------   1 mvince   mvince        305 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/slaves
117966264      8 -r-x------   1 mvince   mvince       7601 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/mapred-site.xml
117966265     12 -r-x------   1 mvince   mvince      10570 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/log4j.properties
117966303      4 -r-x------   1 mvince   mvince       1311 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/hive-site.xml
117966298      4 -r-x------   1 mvince   mvince       2490 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/hadoop-metrics.properties
117966273      8 -r-x------   1 mvince   mvince       6332 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/capacity-scheduler.xml
117966307      4 -r-x------   1 mvince   mvince        667 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/topology_mappings.data
117966290      4 -r-x------   1 mvince   mvince        324 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/yarn_ats_jaas.conf
117966293      4 -r-x------   1 mvince   mvince       1020 Apr  1 14:52 ./__spark_conf__/__hadoop_conf__/commons-logging.properties
117835094    132 -r-x------   1 mvince   mvince     134635 Apr  1 14:52 ./__spark_conf__/__spark_hadoop_conf__.xml
117835093      8 -r-x------   1 mvince   mvince       4956 Apr  1 14:52 ./__spark_conf__/metrics.properties
117835099      8 -r-x------   1 mvince   mvince       4258 Apr  1 14:52 ./__spark_conf__/__spark_conf__.properties
117835092      4 -r-x------   1 mvince   mvince       1139 Apr  1 14:52 ./__spark_conf__/log4j.properties
 61343079      4 -rw-------   1 mvince   hadoop        364 Apr  1 14:52 ./container_tokens
 61474199  22868 -r-x------   1 mvince   mvince   23412838 Apr  1 14:52 ./zeppelin-interpreter-shaded-0.9.0-SNAPSHOT.jar
broken symlinks(find -L . -maxdepth 5 -type l -ls):
INFO [2020-04-01 14:52:39,894] ({main} Logging.scala[logInfo]:54) - Registered signal handler for TERM
 INFO [2020-04-01 14:52:39,897] ({main} Logging.scala[logInfo]:54) - Registered signal handler for HUP
 INFO [2020-04-01 14:52:39,899] ({main} Logging.scala[logInfo]:54) - Registered signal handler for INT
 INFO [2020-04-01 14:52:40,031] ({main} Logging.scala[logInfo]:54) - Changing view acls to: mvince
 INFO [2020-04-01 14:52:40,031] ({main} Logging.scala[logInfo]:54) - Changing modify acls to: mvince
 INFO [2020-04-01 14:52:40,031] ({main} Logging.scala[logInfo]:54) - Changing view acls groups to: 
 INFO [2020-04-01 14:52:40,032] ({main} Logging.scala[logInfo]:54) - Changing modify acls groups to: 
 INFO [2020-04-01 14:52:40,032] ({main} Logging.scala[logInfo]:54) - SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(mvince); groups with view permissions: Set(); users  with modify permissions: Set(mvince); groups with modify permissions: Set()
DEBUG [2020-04-01 14:52:40,043] ({main} Logging.scala[logDebug]:58) - Created SSL options for fs: SSLOptions{enabled=false, keyStore=None, keyStorePassword=None, trustStore=None, trustStorePassword=None, protocol=None, enabledAlgorithms=Set()}
DEBUG [2020-04-01 14:52:40,130] ({main} Shell.java[isSetsidSupported]:778) - setsid exited with exit code 0
DEBUG [2020-04-01 14:52:40,342] ({main} MutableMetricsFactory.java[newForField]:42) - field org.apache.hadoop.metrics2.lib.MutableRate org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginSuccess with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, always=false, sampleName=Ops, type=DEFAULT, valueName=Time, value=[Rate of successful kerberos logins and latency (milliseconds)])
DEBUG [2020-04-01 14:52:40,352] ({main} MutableMetricsFactory.java[newForField]:42) - field org.apache.hadoop.metrics2.lib.MutableRate org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginFailure with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, always=false, sampleName=Ops, type=DEFAULT, valueName=Time, value=[Rate of failed kerberos logins and latency (milliseconds)])
DEBUG [2020-04-01 14:52:40,352] ({main} MutableMetricsFactory.java[newForField]:42) - field org.apache.hadoop.metrics2.lib.MutableRate org.apache.hadoop.security.UserGroupInformation$UgiMetrics.getGroups with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, always=false, sampleName=Ops, type=DEFAULT, valueName=Time, value=[GetGroups])
DEBUG [2020-04-01 14:52:40,353] ({main} MutableMetricsFactory.java[newForField]:42) - field private org.apache.hadoop.metrics2.lib.MutableGaugeLong org.apache.hadoop.security.UserGroupInformation$UgiMetrics.renewalFailuresTotal with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, always=false, sampleName=Ops, type=DEFAULT, valueName=Time, value=[Renewal failures since startup])
DEBUG [2020-04-01 14:52:40,353] ({main} MutableMetricsFactory.java[newForField]:42) - field private org.apache.hadoop.metrics2.lib.MutableGaugeInt org.apache.hadoop.security.UserGroupInformation$UgiMetrics.renewalFailures with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, always=false, sampleName=Ops, type=DEFAULT, valueName=Time, value=[Renewal failures since last successful login])
DEBUG [2020-04-01 14:52:40,354] ({main} MetricsSystemImpl.java[register]:232) - UgiMetrics, User and group related metrics
DEBUG [2020-04-01 14:52:40,393] ({main} SecurityUtil.java[setTokenServiceUseIp]:116) - Setting hadoop.security.token.service.use_ip to true
DEBUG [2020-04-01 14:52:40,423] ({main} Groups.java[getUserToGroupsMappingService]:418) -  Creating new Groups object
DEBUG [2020-04-01 14:52:40,424] ({main} NativeCodeLoader.java[<clinit>]:46) - Trying to load the custom-built native-hadoop library...
DEBUG [2020-04-01 14:52:40,425] ({main} NativeCodeLoader.java[<clinit>]:50) - Loaded the native-hadoop library
DEBUG [2020-04-01 14:52:40,425] ({main} JniBasedUnixGroupsMapping.java[<clinit>]:50) - Using JniBasedUnixGroupsMapping for Group resolution
DEBUG [2020-04-01 14:52:40,425] ({main} JniBasedUnixGroupsMappingWithFallback.java[<init>]:45) - Group mapping impl=org.apache.hadoop.security.JniBasedUnixGroupsMapping
DEBUG [2020-04-01 14:52:40,490] ({main} Groups.java[<init>]:145) - Group mapping impl=org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback; cacheTimeout=300000; warningDeltaMs=5000
DEBUG [2020-04-01 14:52:40,495] ({main} UserGroupInformation.java[login]:247) - hadoop login
DEBUG [2020-04-01 14:52:40,496] ({main} UserGroupInformation.java[commit]:182) - hadoop login commit
DEBUG [2020-04-01 14:52:40,497] ({main} UserGroupInformation.java[commit]:196) - using kerberos user:null
DEBUG [2020-04-01 14:52:40,498] ({main} UserGroupInformation.java[commit]:212) - using local user:UnixPrincipal: mvince
DEBUG [2020-04-01 14:52:40,498] ({main} UserGroupInformation.java[commit]:218) - Using user: "UnixPrincipal: mvince" with name mvince
DEBUG [2020-04-01 14:52:40,499] ({main} UserGroupInformation.java[commit]:228) - User entry: "mvince"
DEBUG [2020-04-01 14:52:40,499] ({main} UserGroupInformation.java[loginUserFromSubject]:855) - Assuming keytab is managed externally since logged in from subject.
DEBUG [2020-04-01 14:52:40,501] ({main} UserGroupInformation.java[loginUserFromSubject]:894) - Reading credentials from location set in HADOOP_TOKEN_FILE_LOCATION: /data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/container_tokens
DEBUG [2020-04-01 14:52:40,505] ({main} UserGroupInformation.java[loginUserFromSubject]:905) - Loaded 3 tokens
DEBUG [2020-04-01 14:52:40,506] ({main} UserGroupInformation.java[loginUserFromSubject]:914) - UGI loginUser:mvince (auth:KERBEROS)
DEBUG [2020-04-01 14:52:40,577] ({main} Logging.scala[logDebug]:58) - creating UGI for user: mvince
DEBUG [2020-04-01 14:52:40,578] ({main} UserGroupInformation.java[logPrivilegedAction]:1896) - PrivilegedAction as:mvince (auth:SIMPLE) from:org.apache.spark.deploy.yarn.ApplicationMaster.doAsUser(ApplicationMaster.scala:814)
DEBUG [2020-04-01 14:52:40,622] ({main} UserGroupInformation.java[logPrivilegedAction]:1896) - PrivilegedAction as:mvince (auth:SIMPLE) from:org.apache.spark.deploy.yarn.ApplicationMaster.doAsUser(ApplicationMaster.scala:814)
 INFO [2020-04-01 14:52:40,622] ({main} Logging.scala[logInfo]:54) - Preparing Local resources
DEBUG [2020-04-01 14:52:40,899] ({main} GoogleHadoopFileSystemBase.java[<clinit>]:642) - GHFS version: 1.8.1.2.6.5.0-292
DEBUG [2020-04-01 14:52:40,985] ({main} DFSClient.java[<init>]:474) - dfs.client.use.legacy.blockreader.local = false
DEBUG [2020-04-01 14:52:40,985] ({main} DFSClient.java[<init>]:477) - dfs.client.read.shortcircuit = true
DEBUG [2020-04-01 14:52:40,985] ({main} DFSClient.java[<init>]:480) - dfs.client.domain.socket.data.traffic = false
DEBUG [2020-04-01 14:52:40,985] ({main} DFSClient.java[<init>]:483) - dfs.domain.socket.path = /var/lib/hadoop-hdfs/dn_socket
DEBUG [2020-04-01 14:52:41,054] ({main} SecurityUtil.java[setTokenService]:421) - Acquired token Kind: HDFS_DELEGATION_TOKEN, Service: 192.168.200.21:8020, Ident: (HDFS_DELEGATION_TOKEN token 3788277 for mvince)
DEBUG [2020-04-01 14:52:41,055] ({main} HAUtil.java[cloneDelegationTokenForLogicalUri]:323) - Mapped HA service delegation token for logical URI hdfs://dwhprodht/user/mvince/.sparkStaging/application_1573680659125_670094/__spark_conf__.zip to namenode dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020
DEBUG [2020-04-01 14:52:41,055] ({main} SecurityUtil.java[setTokenService]:421) - Acquired token Kind: HDFS_DELEGATION_TOKEN, Service: 192.168.200.22:8020, Ident: (HDFS_DELEGATION_TOKEN token 3788277 for mvince)
DEBUG [2020-04-01 14:52:41,055] ({main} HAUtil.java[cloneDelegationTokenForLogicalUri]:323) - Mapped HA service delegation token for logical URI hdfs://dwhprodht/user/mvince/.sparkStaging/application_1573680659125_670094/__spark_conf__.zip to namenode dwh-prod-ht-master-02.mynetwork-dc.local/192.168.200.22:8020
DEBUG [2020-04-01 14:52:41,056] ({main} DFSClient.java[<init>]:474) - dfs.client.use.legacy.blockreader.local = false
DEBUG [2020-04-01 14:52:41,056] ({main} DFSClient.java[<init>]:477) - dfs.client.read.shortcircuit = true
DEBUG [2020-04-01 14:52:41,056] ({main} DFSClient.java[<init>]:480) - dfs.client.domain.socket.data.traffic = false
DEBUG [2020-04-01 14:52:41,056] ({main} DFSClient.java[<init>]:483) - dfs.domain.socket.path = /var/lib/hadoop-hdfs/dn_socket
DEBUG [2020-04-01 14:52:41,067] ({main} RetryUtils.java[getDefaultRetryPolicy]:75) - multipleLinearRandomRetry = null
DEBUG [2020-04-01 14:52:41,142] ({Finalizer} NativeAzureFileSystem.java[finalize]:3880) - finalize() called.
DEBUG [2020-04-01 14:52:41,149] ({Finalizer} NativeAzureFileSystem.java[finalize]:3880) - finalize() called.
DEBUG [2020-04-01 14:52:41,211] ({main} Server.java[registerProtocolEngine]:275) - rpcKind=RPC_PROTOCOL_BUFFER, rpcRequestWrapperClass=class org.apache.hadoop.ipc.ProtobufRpcEngine$RpcRequestWrapper, rpcInvoker=org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker@191ae03f
DEBUG [2020-04-01 14:52:41,214] ({main} ClientCache.java[getClient]:63) - getting client out of cache: org.apache.hadoop.ipc.Client@35fe2125
DEBUG [2020-04-01 14:52:41,417] ({client DomainSocketWatcher} DomainSocketWatcher.java[run]:447) - org.apache.hadoop.net.unix.DomainSocketWatcher$2@2f0cb6f6: starting with interruptCheckPeriodMs = 60000
DEBUG [2020-04-01 14:52:41,421] ({main} DomainSocketFactory.java[<init>]:120) - The short-circuit local reads feature is enabled.
DEBUG [2020-04-01 14:52:41,425] ({main} DataTransferSaslUtil.java[getSaslPropertiesResolver]:183) - DataTransferProtocol not using SaslPropertiesResolver, no QOP found in configuration for dfs.data.transfer.protection
DEBUG [2020-04-01 14:52:41,452] ({main} Client.java[<init>]:465) - The ping interval is 60000 ms.
DEBUG [2020-04-01 14:52:41,454] ({main} Client.java[setupIOstreams]:737) - Connecting to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020
DEBUG [2020-04-01 14:52:41,465] ({main} UserGroupInformation.java[logPrivilegedAction]:1896) - PrivilegedAction as:mvince (auth:SIMPLE) from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:758)
DEBUG [2020-04-01 14:52:41,524] ({main} SaslRpcClient.java[sendSaslMessage]:457) - Sending sasl message state: NEGOTIATE

DEBUG [2020-04-01 14:52:41,530] ({main} SaslRpcClient.java[saslConnect]:389) - Received SASL message state: NEGOTIATE
auths {
  method: "TOKEN"
  mechanism: "DIGEST-MD5"
  protocol: ""
  serverId: "default"
  challenge: "realm=\"default\",nonce=\"kjhasdjknasjn+jkhdsjhdsjnjksdnjjksd\",qop=\"auth\",charset=utf-8,algorithm=md5-sess"
}
auths {
  method: "KERBEROS"
  mechanism: "GSSAPI"
  protocol: "nn"
  serverId: "dwh-prod-ht-master-01.mynetwork-dc.local"
}

DEBUG [2020-04-01 14:52:41,531] ({main} SaslRpcClient.java[getServerToken]:264) - Get token info proto:interface org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB info:@org.apache.hadoop.security.token.TokenInfo(value=class org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenSelector)
DEBUG [2020-04-01 14:52:41,536] ({main} SaslRpcClient.java[createSaslClient]:247) - Creating SASL DIGEST-MD5(TOKEN)  client to authenticate to service at default
DEBUG [2020-04-01 14:52:41,537] ({main} SaslRpcClient.java[selectSaslClient]:176) - Use TOKEN authentication for protocol ClientNamenodeProtocolPB
DEBUG [2020-04-01 14:52:41,538] ({main} SaslRpcClient.java[handle]:683) - SASL client callback: setting username: jdisajisadjosaosdjaijdsjasdoijsaiojsadijdsaijsdai==
DEBUG [2020-04-01 14:52:41,538] ({main} SaslRpcClient.java[handle]:688) - SASL client callback: setting userPassword
DEBUG [2020-04-01 14:52:41,539] ({main} SaslRpcClient.java[handle]:693) - SASL client callback: setting realm: default
DEBUG [2020-04-01 14:52:41,540] ({main} SaslRpcClient.java[sendSaslMessage]:457) - Sending sasl message state: INITIATE
token: "charset=utf-8,username=\"jdisajisadjosaosdjaijdsjasdoijsaiojsadijdsaijsdai==\",realm=\"default\",nonce=\"6humwQry+j9f3j28chdhsucjsx\",nc=00000001,cnonce=\"izPh9q+w/eyxfwxy4fKgWDcYJQcjISX80YHcSTEL\",digest-uri=\"/default\",maxbuf=65536,response=e544dcf1bbac8207d73e3e6727486369,qop=auth"
auths {
  method: "TOKEN"
  mechanism: "DIGEST-MD5"
  protocol: ""
  serverId: "default"
}

DEBUG [2020-04-01 14:52:41,541] ({main} SaslRpcClient.java[saslConnect]:389) - Received SASL message state: SUCCESS
token: "rspauth=6767897987898ashash89sx9sasa98sa8"

DEBUG [2020-04-01 14:52:41,541] ({main} Client.java[setupIOstreams]:781) - Negotiated QOP is :auth
DEBUG [2020-04-01 14:52:41,546] ({IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince} Client.java[run]:1009) - IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince: starting, having connections 1
DEBUG [2020-04-01 14:52:41,547] ({IPC Parameter Sending Thread #0} Client.java[run]:1072) - IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince sending #0
DEBUG [2020-04-01 14:52:41,548] ({IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince} Client.java[receiveRpcResponse]:1129) - IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince got value #0
DEBUG [2020-04-01 14:52:41,548] ({main} ProtobufRpcEngine.java[invoke]:254) - Call: getFileInfo took 116ms
DEBUG [2020-04-01 14:52:41,601] ({main} UserGroupInformation.java[logPrivilegedAction]:1896) - PrivilegedAction as:mvince (auth:SIMPLE) from:org.apache.spark.deploy.yarn.ApplicationMaster.doAsUser(ApplicationMaster.scala:814)
 INFO [2020-04-01 14:52:41,656] ({main} Logging.scala[logInfo]:54) - ApplicationAttemptId: appattempt_1573680659125_670094_000001
DEBUG [2020-04-01 14:52:41,657] ({main} Logging.scala[logDebug]:58) - Adding shutdown hook
 INFO [2020-04-01 14:52:41,667] ({main} Logging.scala[logInfo]:54) - Starting the user application in a separate Thread
 INFO [2020-04-01 14:52:41,698] ({main} Logging.scala[logInfo]:54) - Waiting for spark context initialization...
 WARN [2020-04-01 14:52:41,783] ({Driver} ZeppelinConfiguration.java[create]:159) - Failed to load configuration, proceeding with a default
 INFO [2020-04-01 14:52:41,824] ({Driver} ZeppelinConfiguration.java[create]:171) - Server Host: 127.0.0.1
 INFO [2020-04-01 14:52:41,824] ({Driver} ZeppelinConfiguration.java[create]:173) - Server Port: 8080
 INFO [2020-04-01 14:52:41,824] ({Driver} ZeppelinConfiguration.java[create]:177) - Context Path: /
 INFO [2020-04-01 14:52:41,824] ({Driver} ZeppelinConfiguration.java[create]:178) - Zeppelin Version: 0.9.0-SNAPSHOT
 INFO [2020-04-01 14:52:41,825] ({Driver} RemoteInterpreterServer.java[<init>]:161) - Starting remote interpreter server on port 0, intpEventServerAddress: 192.168.200.65:38198
 INFO [2020-04-01 14:52:41,854] ({Driver} RemoteInterpreterServer.java[<init>]:188) - Launching ThriftServer at 192.168.200.32:45956
 INFO [2020-04-01 14:52:42,206] ({pool-5-thread-1} RemoteInterpreterServer.java[createInterpreter]:408) - Instantiate interpreter org.apache.zeppelin.spark.SparkInterpreter
DEBUG [2020-04-01 14:52:42,206] ({pool-5-thread-1} InterpreterGroup.java[addInterpreterToSession]:91) - Add Interpreter org.apache.zeppelin.spark.SparkInterpreter to session shared_session
 INFO [2020-04-01 14:52:42,212] ({pool-5-thread-1} RemoteInterpreterServer.java[createInterpreter]:408) - Instantiate interpreter org.apache.zeppelin.spark.SparkSqlInterpreter
DEBUG [2020-04-01 14:52:42,212] ({pool-5-thread-1} InterpreterGroup.java[addInterpreterToSession]:91) - Add Interpreter org.apache.zeppelin.spark.SparkSqlInterpreter to session shared_session
 INFO [2020-04-01 14:52:42,218] ({pool-5-thread-1} RemoteInterpreterServer.java[createInterpreter]:408) - Instantiate interpreter org.apache.zeppelin.spark.PySparkInterpreter
DEBUG [2020-04-01 14:52:42,219] ({pool-5-thread-1} InterpreterGroup.java[addInterpreterToSession]:91) - Add Interpreter org.apache.zeppelin.spark.PySparkInterpreter to session shared_session
 INFO [2020-04-01 14:52:42,226] ({pool-5-thread-1} RemoteInterpreterServer.java[createInterpreter]:408) - Instantiate interpreter org.apache.zeppelin.spark.IPySparkInterpreter
DEBUG [2020-04-01 14:52:42,226] ({pool-5-thread-1} InterpreterGroup.java[addInterpreterToSession]:91) - Add Interpreter org.apache.zeppelin.spark.IPySparkInterpreter to session shared_session
 INFO [2020-04-01 14:52:42,231] ({pool-5-thread-1} RemoteInterpreterServer.java[createInterpreter]:408) - Instantiate interpreter org.apache.zeppelin.spark.SparkRInterpreter
DEBUG [2020-04-01 14:52:42,231] ({pool-5-thread-1} InterpreterGroup.java[addInterpreterToSession]:91) - Add Interpreter org.apache.zeppelin.spark.SparkRInterpreter to session shared_session
 INFO [2020-04-01 14:52:42,234] ({pool-5-thread-1} RemoteInterpreterServer.java[createInterpreter]:408) - Instantiate interpreter org.apache.zeppelin.spark.SparkIRInterpreter
DEBUG [2020-04-01 14:52:42,234] ({pool-5-thread-1} InterpreterGroup.java[addInterpreterToSession]:91) - Add Interpreter org.apache.zeppelin.spark.SparkIRInterpreter to session shared_session
 INFO [2020-04-01 14:52:42,238] ({pool-5-thread-1} RemoteInterpreterServer.java[createInterpreter]:408) - Instantiate interpreter org.apache.zeppelin.spark.SparkShinyInterpreter
DEBUG [2020-04-01 14:52:42,238] ({pool-5-thread-1} InterpreterGroup.java[addInterpreterToSession]:91) - Add Interpreter org.apache.zeppelin.spark.SparkShinyInterpreter to session shared_session
DEBUG [2020-04-01 14:52:42,243] ({pool-5-thread-1} KotlinSparkInterpreter.java[<init>]:62) - Creating KotlinSparkInterpreter
 INFO [2020-04-01 14:52:42,340] ({pool-5-thread-1} RemoteInterpreterServer.java[createInterpreter]:408) - Instantiate interpreter org.apache.zeppelin.spark.KotlinSparkInterpreter
DEBUG [2020-04-01 14:52:42,340] ({pool-5-thread-1} InterpreterGroup.java[addInterpreterToSession]:91) - Add Interpreter org.apache.zeppelin.spark.KotlinSparkInterpreter to session shared_session
 INFO [2020-04-01 14:52:42,395] ({pool-5-thread-2} SchedulerFactory.java[<init>]:62) - Scheduler Thread Pool Size: 100
DEBUG [2020-04-01 14:52:42,400] ({pool-5-thread-2} Interpreter.java[getProperty]:204) - key: zeppelin.spark.concurrentSQL, value: false
DEBUG [2020-04-01 14:52:42,429] ({pool-5-thread-2} RemoteInterpreterServer.java[interpret]:523) - st:
val x = "hello world teraz sem pisem zmeny velke, sem pisem sovje zmeny, tuktk tuk tu,"
print(x)
 INFO [2020-04-01 14:52:42,484] ({FIFOScheduler-interpreter_1074544391-Worker-1} AbstractScheduler.java[runJob]:125) - Job paragraph_1571833609109_-338708939 started by scheduler interpreter_1074544391
 INFO [2020-04-01 14:52:42,498] ({FIFOScheduler-interpreter_1074544391-Worker-1} SparkInterpreter.java[extractScalaVersion]:233) - Using Scala: version 2.11.8
DEBUG [2020-04-01 14:52:42,517] ({FIFOScheduler-interpreter_1074544391-Worker-1} Interpreter.java[getProperty]:204) - key: zeppelin.interpreter.localRepo, value: /opt/zeppelin-server/local-repo/spark
 INFO [2020-04-01 14:52:42,524] ({FIFOScheduler-interpreter_1074544391-Worker-1} SparkScala211Interpreter.scala[open]:62) - Scala shell repl output dir: /data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/tmp/spark8391684309472116098
DEBUG [2020-04-01 14:52:42,763] ({FIFOScheduler-interpreter_1074544391-Worker-1} BaseSparkScalaInterpreter.scala[getUserJars]:463) - User jar for spark repl: file:/data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/__app__.jar,file:/data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/spark-scala-2.11-0.9.0-SNAPSHOT.jar,file:/data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/zeppelin-interpreter-shaded-0.9.0-SNAPSHOT.jar,/data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/__app__.jar,/data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/spark-scala-2.11-0.9.0-SNAPSHOT.jar,/data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/zeppelin-interpreter-shaded-0.9.0-SNAPSHOT.jar
 INFO [2020-04-01 14:52:42,764] ({FIFOScheduler-interpreter_1074544391-Worker-1} SparkScala211Interpreter.scala[open]:77) - UserJars: file:/data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/__app__.jar:file:/data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/spark-scala-2.11-0.9.0-SNAPSHOT.jar:file:/data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/zeppelin-interpreter-shaded-0.9.0-SNAPSHOT.jar:/data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/__app__.jar:/data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/spark-scala-2.11-0.9.0-SNAPSHOT.jar:/data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/zeppelin-interpreter-shaded-0.9.0-SNAPSHOT.jar
[init] error: error while loading <root>, Error accessing /data/01/yarn/local/usercache/mvince/appcache/application_1573680659125_670094/container_e150_1573680659125_670094_01_000001/__spark_libs__/dummy.jar

Failed to initialize compiler: object java.lang.Object in compiler mirror not found.
** Note that as of 2.8 scala does not assume use of the java classpath.
** For the old behavior pass -usejavacp to scala, or if using a Settings
** object programmatically, settings.usejavacp.value = true.

Failed to initialize compiler: object java.lang.Object in compiler mirror not found.
** Note that as of 2.8 scala does not assume use of the java classpath.
** For the old behavior pass -usejavacp to scala, or if using a Settings
** object programmatically, settings.usejavacp.value = true.
ERROR [2020-04-01 14:52:44,394] ({FIFOScheduler-interpreter_1074544391-Worker-1} SparkInterpreter.java[open]:113) - Fail to open SparkInterpreter
scala.reflect.internal.MissingRequirementError: object java.lang.Object in compiler mirror not found.
	at scala.reflect.internal.MissingRequirementError$.signal(MissingRequirementError.scala:17)
	at scala.reflect.internal.MissingRequirementError$.notFound(MissingRequirementError.scala:18)
	at scala.reflect.internal.Mirrors$RootsBase.getModuleOrClass(Mirrors.scala:53)
	at scala.reflect.internal.Mirrors$RootsBase.getModuleOrClass(Mirrors.scala:45)
	at scala.reflect.internal.Mirrors$RootsBase.getModuleOrClass(Mirrors.scala:45)
	at scala.reflect.internal.Mirrors$RootsBase.getModuleOrClass(Mirrors.scala:66)
	at scala.reflect.internal.Mirrors$RootsBase.getClassByName(Mirrors.scala:102)
	at scala.reflect.internal.Mirrors$RootsBase.getRequiredClass(Mirrors.scala:105)
	at scala.reflect.internal.Definitions$DefinitionsClass.ObjectClass$lzycompute(Definitions.scala:257)
	at scala.reflect.internal.Definitions$DefinitionsClass.ObjectClass(Definitions.scala:257)
	at scala.reflect.internal.Definitions$DefinitionsClass.init(Definitions.scala:1394)
	at scala.tools.nsc.Global$Run.<init>(Global.scala:1215)
	at scala.tools.nsc.interpreter.IMain.compileSourcesKeepingRun(IMain.scala:432)
	at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.compileAndSaveRun(IMain.scala:855)
	at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.compile(IMain.scala:813)
	at scala.tools.nsc.interpreter.IMain.bind(IMain.scala:675)
	at scala.tools.nsc.interpreter.IMain.bind(IMain.scala:712)
	at scala.tools.nsc.interpreter.IMain$$anonfun$quietBind$1.apply(IMain.scala:711)
	at scala.tools.nsc.interpreter.IMain$$anonfun$quietBind$1.apply(IMain.scala:711)
	at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:214)
	at scala.tools.nsc.interpreter.IMain.quietBind(IMain.scala:711)
	at org.apache.zeppelin.spark.SparkScala211Interpreter$.loopPostInit$1(SparkScala211Interpreter.scala:164)
	at org.apache.zeppelin.spark.SparkScala211Interpreter$.org$apache$zeppelin$spark$SparkScala211Interpreter$$loopPostInit(SparkScala211Interpreter.scala:199)
	at org.apache.zeppelin.spark.SparkScala211Interpreter.open(SparkScala211Interpreter.scala:95)
	at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:98)
	at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:70)
	at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:668)
	at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:577)
	at org.apache.zeppelin.scheduler.Job.run(Job.java:172)
	at org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:130)
	at org.apache.zeppelin.scheduler.FIFOScheduler.lambda$runJobInScheduler$0(FIFOScheduler.java:39)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
 INFO [2020-04-01 14:52:44,396] ({FIFOScheduler-interpreter_1074544391-Worker-1} SparkInterpreter.java[close]:159) - Close SparkInterpreter
DEBUG [2020-04-01 14:52:44,404] ({FIFOScheduler-interpreter_1074544391-Worker-1} AbstractScheduler.java[runJob]:143) - Job Error, paragraph_1571833609109_-338708939, %text org.apache.zeppelin.interpreter.InterpreterException: org.apache.zeppelin.interpreter.InterpreterException: Fail to open SparkInterpreter
	at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:76)
	at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:668)
	at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:577)
	at org.apache.zeppelin.scheduler.Job.run(Job.java:172)
	at org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:130)
	at org.apache.zeppelin.scheduler.FIFOScheduler.lambda$runJobInScheduler$0(FIFOScheduler.java:39)
	at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
	at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
	at java.lang.Thread.run(Thread.java:748)
Caused by: org.apache.zeppelin.interpreter.InterpreterException: Fail to open SparkInterpreter
	at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:114)
	at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:70)
	... 8 more
Caused by: scala.reflect.internal.MissingRequirementError: object java.lang.Object in compiler mirror not found.
	at scala.reflect.internal.MissingRequirementError$.signal(MissingRequirementError.scala:17)
	at scala.reflect.internal.MissingRequirementError$.notFound(MissingRequirementError.scala:18)
	at scala.reflect.internal.Mirrors$RootsBase.getModuleOrClass(Mirrors.scala:53)
	at scala.reflect.internal.Mirrors$RootsBase.getModuleOrClass(Mirrors.scala:45)
	at scala.reflect.internal.Mirrors$RootsBase.getModuleOrClass(Mirrors.scala:45)
	at scala.reflect.internal.Mirrors$RootsBase.getModuleOrClass(Mirrors.scala:66)
	at scala.reflect.internal.Mirrors$RootsBase.getClassByName(Mirrors.scala:102)
	at scala.reflect.internal.Mirrors$RootsBase.getRequiredClass(Mirrors.scala:105)
	at scala.reflect.internal.Definitions$DefinitionsClass.ObjectClass$lzycompute(Definitions.scala:257)
	at scala.reflect.internal.Definitions$DefinitionsClass.ObjectClass(Definitions.scala:257)
	at scala.reflect.internal.Definitions$DefinitionsClass.init(Definitions.scala:1394)
	at scala.tools.nsc.Global$Run.<init>(Global.scala:1215)
	at scala.tools.nsc.interpreter.IMain.compileSourcesKeepingRun(IMain.scala:432)
	at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.compileAndSaveRun(IMain.scala:855)
	at scala.tools.nsc.interpreter.IMain$ReadEvalPrint.compile(IMain.scala:813)
	at scala.tools.nsc.interpreter.IMain.bind(IMain.scala:675)
	at scala.tools.nsc.interpreter.IMain.bind(IMain.scala:712)
	at scala.tools.nsc.interpreter.IMain$$anonfun$quietBind$1.apply(IMain.scala:711)
	at scala.tools.nsc.interpreter.IMain$$anonfun$quietBind$1.apply(IMain.scala:711)
	at scala.tools.nsc.interpreter.IMain.beQuietDuring(IMain.scala:214)
	at scala.tools.nsc.interpreter.IMain.quietBind(IMain.scala:711)
	at org.apache.zeppelin.spark.SparkScala211Interpreter$.loopPostInit$1(SparkScala211Interpreter.scala:164)
	at org.apache.zeppelin.spark.SparkScala211Interpreter$.org$apache$zeppelin$spark$SparkScala211Interpreter$$loopPostInit(SparkScala211Interpreter.scala:199)
	at org.apache.zeppelin.spark.SparkScala211Interpreter.open(SparkScala211Interpreter.scala:95)
	at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:98)
	... 9 more

 INFO [2020-04-01 14:52:44,404] ({FIFOScheduler-interpreter_1074544391-Worker-1} AbstractScheduler.java[runJob]:152) - Job paragraph_1571833609109_-338708939 finished by scheduler interpreter_1074544391
DEBUG [2020-04-01 14:52:44,597] ({pool-5-thread-2} RemoteInterpreterServer.java[resourcePoolGetAll]:1014) - Request resourcePoolGetAll from ZeppelinServer
DEBUG [2020-04-01 14:53:11,566] ({IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince} Client.java[close]:1226) - IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince: closed
DEBUG [2020-04-01 14:53:11,566] ({IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince} Client.java[run]:1027) - IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince: stopped, remaining connections 0
ERROR [2020-04-01 14:54:21,726] ({main} Logging.scala[logError]:91) - Uncaught exception: 
java.util.concurrent.TimeoutException: Futures timed out after [100000 milliseconds]
	at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
	at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
	at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:201)
	at org.apache.spark.deploy.yarn.ApplicationMaster.runDriver(ApplicationMaster.scala:498)
	at org.apache.spark.deploy.yarn.ApplicationMaster.org$apache$spark$deploy$yarn$ApplicationMaster$$runImpl(ApplicationMaster.scala:345)
	at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$2.apply$mcV$sp(ApplicationMaster.scala:260)
	at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$2.apply(ApplicationMaster.scala:260)
	at org.apache.spark.deploy.yarn.ApplicationMaster$$anonfun$run$2.apply(ApplicationMaster.scala:260)
	at org.apache.spark.deploy.yarn.ApplicationMaster$$anon$5.run(ApplicationMaster.scala:815)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:422)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1869)
	at org.apache.spark.deploy.yarn.ApplicationMaster.doAsUser(ApplicationMaster.scala:814)
	at org.apache.spark.deploy.yarn.ApplicationMaster.run(ApplicationMaster.scala:259)
	at org.apache.spark.deploy.yarn.ApplicationMaster$.main(ApplicationMaster.scala:839)
	at org.apache.spark.deploy.yarn.ApplicationMaster.main(ApplicationMaster.scala)
 INFO [2020-04-01 14:54:21,727] ({main} Logging.scala[logInfo]:54) - Final app status: FAILED, exitCode: 13, (reason: Uncaught exception: java.util.concurrent.TimeoutException: Futures timed out after [100000 milliseconds])
DEBUG [2020-04-01 14:54:21,727] ({main} Logging.scala[logDebug]:58) - shutting down user thread
 INFO [2020-04-01 14:54:21,732] ({pool-1-thread-1} Logging.scala[logInfo]:54) - Deleting staging directory hdfs://dwhprodht/user/mvince/.sparkStaging/application_1573680659125_670094
DEBUG [2020-04-01 14:54:21,734] ({pool-1-thread-1} Client.java[<init>]:465) - The ping interval is 60000 ms.
DEBUG [2020-04-01 14:54:21,734] ({pool-1-thread-1} Client.java[setupIOstreams]:737) - Connecting to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020
DEBUG [2020-04-01 14:54:21,735] ({pool-1-thread-1} UserGroupInformation.java[logPrivilegedAction]:1896) - PrivilegedAction as:mvince (auth:SIMPLE) from:org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:758)
DEBUG [2020-04-01 14:54:21,736] ({pool-1-thread-1} SaslRpcClient.java[sendSaslMessage]:457) - Sending sasl message state: NEGOTIATE

DEBUG [2020-04-01 14:54:21,737] ({pool-1-thread-1} SaslRpcClient.java[saslConnect]:389) - Received SASL message state: NEGOTIATE
auths {
  method: "TOKEN"
  mechanism: "DIGEST-MD5"
  protocol: ""
  serverId: "default"
  challenge: "realm=\"default\",nonce=\"4kzuNdnplBFu+kdsalpfelwp[fewo,wdodwo\",qop=\"auth\",charset=utf-8,algorithm=md5-sess"
}
auths {
  method: "KERBEROS"
  mechanism: "GSSAPI"
  protocol: "nn"
  serverId: "dwh-prod-ht-master-01.mynetwork-dc.local"
}

DEBUG [2020-04-01 14:54:21,737] ({pool-1-thread-1} SaslRpcClient.java[getServerToken]:264) - Get token info proto:interface org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolPB info:@org.apache.hadoop.security.token.TokenInfo(value=class org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenSelector)
DEBUG [2020-04-01 14:54:21,738] ({pool-1-thread-1} SaslRpcClient.java[createSaslClient]:247) - Creating SASL DIGEST-MD5(TOKEN)  client to authenticate to service at default
DEBUG [2020-04-01 14:54:21,738] ({pool-1-thread-1} SaslRpcClient.java[selectSaslClient]:176) - Use TOKEN authentication for protocol ClientNamenodeProtocolPB
DEBUG [2020-04-01 14:54:21,739] ({pool-1-thread-1} SaslRpcClient.java[handle]:683) - SASL client callback: setting username: jdisajisadjosaosdjaijdsjasdoijsaiojsadijdsaijsdai==
DEBUG [2020-04-01 14:54:21,739] ({pool-1-thread-1} SaslRpcClient.java[handle]:688) - SASL client callback: setting userPassword
DEBUG [2020-04-01 14:54:21,739] ({pool-1-thread-1} SaslRpcClient.java[handle]:693) - SASL client callback: setting realm: default
DEBUG [2020-04-01 14:54:21,740] ({pool-1-thread-1} SaslRpcClient.java[sendSaslMessage]:457) - Sending sasl message state: INITIATE
token: "charset=utf-8,username=\"jdisajisadjosaosdjaijdsjasdoijsaiojsadijdsaijsdai==\",realm=\"default\",nonce=\"4kzuNdnplBFu+kdsalpfelwp[fewo,wdodwo\",nc=00000001,cnonce=\"ufewijweij0932jd9dd+2sdads0asdi0\",digest-uri=\"/default\",maxbuf=65536,response=ad252d5ddf8d9be67a741cb9c2e27595,qop=auth"
auths {
  method: "TOKEN"
  mechanism: "DIGEST-MD5"
  protocol: ""
  serverId: "default"
}

DEBUG [2020-04-01 14:54:21,741] ({pool-1-thread-1} SaslRpcClient.java[saslConnect]:389) - Received SASL message state: SUCCESS
token: "rspauth=c72ffa1d7dd98d78dde18ffc147357cd"

DEBUG [2020-04-01 14:54:21,742] ({pool-1-thread-1} Client.java[setupIOstreams]:781) - Negotiated QOP is :auth
DEBUG [2020-04-01 14:54:21,742] ({IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince} Client.java[run]:1009) - IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince: starting, having connections 1
DEBUG [2020-04-01 14:54:21,743] ({IPC Parameter Sending Thread #1} Client.java[run]:1072) - IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince sending #1
DEBUG [2020-04-01 14:54:21,745] ({IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince} Client.java[receiveRpcResponse]:1129) - IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince got value #1
DEBUG [2020-04-01 14:54:21,745] ({pool-1-thread-1} ProtobufRpcEngine.java[invoke]:254) - Call: delete took 11ms
 INFO [2020-04-01 14:54:21,747] ({pool-1-thread-1} Logging.scala[logInfo]:54) - Shutdown hook called
DEBUG [2020-04-01 14:54:21,749] ({pool-1-thread-1} ClientCache.java[stopClient]:97) - stopping client from cache: org.apache.hadoop.ipc.Client@35fe2125
DEBUG [2020-04-01 14:54:21,749] ({pool-1-thread-1} ClientCache.java[stopClient]:103) - removing client from cache: org.apache.hadoop.ipc.Client@35fe2125
DEBUG [2020-04-01 14:54:21,749] ({pool-1-thread-1} ClientCache.java[stopClient]:110) - stopping actual client because no more references remain: org.apache.hadoop.ipc.Client@35fe2125
DEBUG [2020-04-01 14:54:21,749] ({pool-1-thread-1} Client.java[stop]:1279) - Stopping client
DEBUG [2020-04-01 14:54:21,751] ({IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince} Client.java[close]:1226) - IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince: closed
DEBUG [2020-04-01 14:54:21,751] ({IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince} Client.java[run]:1027) - IPC Client (176955204) connection to dwh-prod-ht-master-01.mynetwork-dc.local/192.168.200.21:8020 from mvince: stopped, remaining connections 0
DEBUG [2020-04-01 14:54:21,852] ({Thread-5} ShutdownHookManager.java[run]:84) - ShutdownHookManger complete shutdown.

Reply via email to