Ooops, sorry. No zeppelin-interpreter-log, but I did looked at the wrong log file.
No error on startup, and here is the stack trace when I query a hive table : INFO [2018-02-06 12:12:18,405] ({qtp1632392469-74} AbstractValidatingSessionManager.java[enableSessionValidation]:230) - Enabling session validation scheduler... WARN [2018-02-06 12:12:18,484] ({qtp1632392469-74} LoginRestApi.java[postLogin]:119) - {"status":"OK","message":"","body":{"principal":"alinz","ticket":"f4ebf4ec-d889-4d01-808a-ee8b9800c80e","roles":"[]"}} INFO [2018-02-06 12:12:18,710] ({qtp1632392469-74} NotebookServer.java[sendNote]:711) - New operation from 172.24.193.234 : 61997 : alinz : GET_NOTE : 2D1SETNNQ WARN [2018-02-06 12:12:18,764] ({qtp1632392469-74} GitNotebookRepo.java[revisionHistory]:157) - No Head found for 2D1SETNNQ, No HEAD exists and no explicit starting revision was specified INFO [2018-02-06 12:12:19,102] ({qtp1632392469-71} InterpreterFactory.java[createInterpretersForNote]:188) - Create interpreter instance spark for note 2D1SETNNQ INFO [2018-02-06 12:12:19,107] ({qtp1632392469-71} InterpreterFactory.java[createInterpretersForNote]:221) - Interpreter org.apache.zeppelin.spark.SparkInterpreter 1563969204 created INFO [2018-02-06 12:12:19,109] ({qtp1632392469-71} InterpreterFactory.java[createInterpretersForNote]:221) - Interpreter org.apache.zeppelin.spark.SparkSqlInterpreter 119386711 created INFO [2018-02-06 12:12:19,110] ({qtp1632392469-71} InterpreterFactory.java[createInterpretersForNote]:221) - Interpreter org.apache.zeppelin.spark.DepInterpreter 1706768324 created INFO [2018-02-06 12:12:19,111] ({qtp1632392469-71} InterpreterFactory.java[createInterpretersForNote]:221) - Interpreter org.apache.zeppelin.spark.PySparkInterpreter 261340605 created INFO [2018-02-06 12:12:19,112] ({qtp1632392469-71} InterpreterFactory.java[createInterpretersForNote]:221) - Interpreter org.apache.zeppelin.spark.SparkRInterpreter 276911901 created INFO [2018-02-06 12:12:19,111] ({qtp1632392469-71} InterpreterFactory.java[createInterpretersForNote]:221) - Interpreter org.apache.zeppelin.spark.PySparkInterpreter 261340605 created INFO [2018-02-06 12:12:19,112] ({qtp1632392469-71} InterpreterFactory.java[createInterpretersForNote]:221) - Interpreter org.apache.zeppelin.spark.SparkRInterpreter 276911901 created INFO [2018-02-06 12:13:38,419] ({pool-2-thread-2} SchedulerFactory.java[jobStarted]:131) - Job paragraph_1513703910594_823567073 started by scheduler org.apache.zeppelin.interpreter.remote.RemoteInterpretershared_session1977922636 INFO [2018-02-06 12:13:38,420] ({pool-2-thread-2} Paragraph.java[jobRun]:362) - run paragraph 20171219-181830_1097537317 using pyspark org.apache.zeppelin.interpreter.LazyOpenInterpreter@f93bdbd<mailto:org.apache.zeppelin.interpreter.LazyOpenInterpreter@f93bdbd> INFO [2018-02-06 12:13:38,428] ({pool-2-thread-2} RemoteInterpreterManagedProcess.java[start]:126) - Run interpreter process [/opt/zeppelin/bin/interpreter.sh, -d, /opt/zeppelin/interpreter/spark, -p, 43118, -u, alinz, -l, /opt/zeppelin/local-repo/2CYVF45A9] INFO [2018-02-06 12:13:39,474] ({pool-2-thread-2} RemoteInterpreter.java[init]:221) - Create remote interpreter org.apache.zeppelin.spark.SparkInterpreter INFO [2018-02-06 12:13:39,570] ({pool-2-thread-2} RemoteInterpreter.java[pushAngularObjectRegistryToRemote]:551) - Push local angular object registry from ZeppelinServer to remote interpreter group 2CYVF45A9:alinz: INFO [2018-02-06 12:13:39,585] ({pool-2-thread-2} RemoteInterpreter.java[init]:221) - Create remote interpreter org.apache.zeppelin.spark.SparkSqlInterpreter INFO [2018-02-06 12:13:39,587] ({pool-2-thread-2} RemoteInterpreter.java[init]:221) - Create remote interpreter org.apache.zeppelin.spark.DepInterpreter INFO [2018-02-06 12:13:39,592] ({pool-2-thread-2} RemoteInterpreter.java[init]:221) - Create remote interpreter org.apache.zeppelin.spark.PySparkInterpreter INFO [2018-02-06 12:13:39,603] ({pool-2-thread-2} RemoteInterpreter.java[init]:221) - Create remote interpreter org.apache.zeppelin.spark.SparkRInterpreter WARN [2018-02-06 12:13:53,749] ({pool-2-thread-2} NotebookServer.java[afterStatusChange]:2064) - Job 20171219-181830_1097537317 is finished, status: ERROR, exception: null, result: %text WARNING: Running python applications through 'pyspark' is deprecated as of Spark 1.0. se ./bin/spark-submit <python file> /tmp/zeppelin_pyspark-5465626759760627770.py:151: UserWarning: Unable to load inline matplotlib backend, falling back to Agg warnings.warn("Unable to load inline matplotlib backend, " %text Traceback (most recent call last): File "/opt/cloudera/parcels/CDH-5.11.1-1.cdh5.11.1.p0.4/lib/spark/python/lib/pyspark.zip/pyspark/sql/utils.py", line 45, in deco return f(*a, **kw) File "/opt/cloudera/parcels/CDH-5.11.1-1.cdh5.11.1.p0.4/lib/spark/python/lib/py4j-0.9-src.zip/py4j/protocol.py", line 308, in get_return_value format(target_id, ".", name), value) py4j.protocol.Py4JJavaError: An error occurred while calling o73.sql. : org.apache.spark.sql.AnalysisException: Table not found: `hiveDatabase`.`hiveTable`; at org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42) at org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$checkAnalysis$1.apply(CheckAnalysis.scala:54) at org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$checkAnalysis$1.apply(CheckAnalysis.scala:50) at org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:121) at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$foreachUp$1.apply(TreeNode.scala:120) at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$foreachUp$1.apply(TreeNode.scala:120) at scala.collection.immutable.List.foreach(List.scala:318) at org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:120) at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$foreachUp$1.apply(TreeNode.scala:120) at org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$foreachUp$1.apply(TreeNode.scala:120) at scala.collection.immutable.List.foreach(List.scala:318) at org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:120) at org.apache.spark.sql.catalyst.analysis.CheckAnalysis$class.checkAnalysis(CheckAnalysis.scala:50) at org.apache.spark.sql.catalyst.analysis.Analyzer.checkAnalysis(Analyzer.scala:44) at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:35) at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:133) at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:52) at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:829) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:231) at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:381) at py4j.Gateway.invoke(Gateway.java:259) at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133) at py4j.commands.CallCommand.execute(CallCommand.java:79) at py4j.GatewayConnection.run(GatewayConnection.java:209) at java.lang.Thread.run(Thread.java:745) (…) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/tmp/zeppelin_pyspark-5465626759760627770.py", line 355, in <module> exec(code, _zcUserQueryNameSpace) File "<stdin>", line 1, in <module> File "/opt/cloudera/parcels/CDH-5.11.1-1.cdh5.11.1.p0.4/lib/spark/python/lib/pyspark.zip/pyspark/sql/context.py", line 580, in sql return DataFrame(self._ssql_ctx.sql(sqlQuery), self) File "/opt/cloudera/parcels/CDH-5.11.1-1.cdh5.11.1.p0.4/lib/spark/python/lib/py4j-0.9-src.zip/py4j/java_gateway.py", line 813, in __call__ answer, self.gateway_client, self.target_id, self.name) File "/opt/cloudera/parcels/CDH-5.11.1-1.cdh5.11.1.p0.4/lib/spark/python/lib/pyspark.zip/pyspark/sql/utils.py", line 51, in deco raise AnalysisException(s.split(': ', 1)[1], stackTrace) pyspark.sql.utils.AnalysisException: 'Table not found: `hiveDatabase`.`hiveTable`;' INFO [2018-02-06 12:13:53,795] ({pool-2-thread-2} SchedulerFactory.java[jobFinished]:137) - Job paragraph_1513703910594_823567073 finished by scheduler org.apache.zeppelin.interpreter.remote.RemoteInterpretershared_session1977922636 De : Jeff Zhang [mailto:zjf...@gmail.com] Envoyé : mardi 6 février 2018 12:04 À : users@zeppelin.apache.org Objet : Re: How to have a native graphical representation (%sql) of a HiveContext? It is weird, there should be a interpreter log file with name 'zeppelin-interpreter-spark*.log' Are you sure you are looking at the correct log folder ? Because from the log you pasted, it is the log of 2017-12-08 LINZ, Arnaud <al...@bouyguestelecom.fr<mailto:al...@bouyguestelecom.fr>>于2018年2月6日周二 下午6:45写道: I’ve been searching for such a log, but I don’t see anything related to spark in …/zeppelin/logs The only logs I have are zeppelin-$USER-$HOST.log and zeppelin-$USER-$HOST.out but they really don’t contain anything useful. Log = INFO [2017-12-08 12:21:36,847] ({main} ZeppelinConfiguration.java[create]:101) - Load configuration from file:/etc/zeppelin/zeppelin-site.xml INFO [2017-12-08 12:21:36,896] ({main} ZeppelinConfiguration.java[create]:109) - Server Host: 0.0.0.0 INFO [2017-12-08 12:21:36,896] ({main} ZeppelinConfiguration.java[create]:113) - Server SSL Port: 8080 INFO [2017-12-08 12:21:36,896] ({main} ZeppelinConfiguration.java[create]:115) - Context Path: / INFO [2017-12-08 12:21:36,900] ({main} ZeppelinConfiguration.java[create]:116) - Zeppelin Version: 0.7.3 INFO [2017-12-08 12:21:36,917] ({main} Log.java[initialized]:186) - Logging initialized @275ms INFO [2017-12-08 12:21:36,979] ({main} ZeppelinServer.java[setupWebAppContext]:346) - ZeppelinServer Webapp path: /opt/zeppelin/webapps INFO [2017-12-08 12:21:37,196] ({main} IniRealm.java[processDefinitions]:188) - IniRealm defined, but there is no [users] section defined. This realm will not be populated with any users and it is assumed that they will be populated programatically. Users must be defined for this Realm instance to be useful. INFO [2017-12-08 12:21:37,197] ({main} AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or cacheManager properties have been set. Authorization cache cannot be obtained. INFO [2017-12-08 12:21:37,227] ({main} ZeppelinServer.java[main]:187) - Starting zeppelin server INFO [2017-12-08 12:21:37,229] ({main} Server.java[doStart]:327) - jetty-9.2.15.v20160210 INFO [2017-12-08 12:21:38,913] ({main} StandardDescriptorProcessor.java[visitServlet]:297) - NO JSP Support for /, did not find org.eclipse.jetty.jsp.JettyJspServlet INFO [2017-12-08 12:21:38,922] ({main} ContextHandler.java[log]:2052) - Initializing Shiro environment INFO [2017-12-08 12:21:38,922] ({main} EnvironmentLoader.java[initEnvironment]:128) - Starting Shiro environment initialization. INFO [2017-12-08 12:21:38,972] ({main} IniRealm.java[processDefinitions]:188) - IniRealm defined, but there is no [users] section defined. This realm will not be populated with any users and it is assumed that they will be populated programatically. Users must be defined for this Realm instance to be useful. INFO [2017-12-08 12:21:38,972] ({main} AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or cacheManager properties have been set. Authorization cache cannot be obtained. INFO [2017-12-08 12:21:38,977] ({main} EnvironmentLoader.java[initEnvironment]:141) - Shiro environment initialized in 55 ms. WARN [2017-12-08 12:21:39,115] ({main} Helium.java[loadConf]:101) - /etc/zeppelin/helium.json does not exists WARN [2017-12-08 12:21:39,402] ({main} Interpreter.java[register]:406) - Static initialization is deprecated for interpreter sql, You should change it to use interpreter-setting.json in your jar or interpreter/{interpreter}/interpreter-setting.json INFO [2017-12-08 12:21:39,403] ({main} InterpreterSettingManager.java[init]:305) - Interpreter psql.sql found. class=org.apache.zeppelin.postgresql.PostgreSqlInterpreter INFO [2017-12-08 12:21:39,497] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name ignite INFO [2017-12-08 12:21:39,497] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name python INFO [2017-12-08 12:21:39,497] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name jdbc INFO [2017-12-08 12:21:39,497] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name psql INFO [2017-12-08 12:21:39,497] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name lens INFO [2017-12-08 12:21:39,497] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name pig INFO [2017-12-08 12:21:39,498] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name flink INFO [2017-12-08 12:21:39,498] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name angular INFO [2017-12-08 12:21:39,498] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name livy INFO [2017-12-08 12:21:39,498] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name file INFO [2017-12-08 12:21:39,498] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name elasticsearch INFO [2017-12-08 12:21:39,498] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name cassandra INFO [2017-12-08 12:21:39,498] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name sh INFO [2017-12-08 12:21:39,498] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name spark INFO [2017-12-08 12:21:39,498] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name md INFO [2017-12-08 12:21:39,498] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name alluxio INFO [2017-12-08 12:21:39,499] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name bigquery INFO [2017-12-08 12:21:39,499] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name hbase INFO [2017-12-08 12:21:39,499] ({main} InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name kylin INFO [2017-12-08 12:21:39,532] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group md : id=2D18QE9AX, name=md INFO [2017-12-08 12:21:39,533] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group flink : id=2CYSDG6AU, name=flink INFO [2017-12-08 12:21:39,533] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group angular : id=2CZY6QWUG, name=angular INFO [2017-12-08 12:21:39,533] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group sh : id=2CYZFQZXG, name=sh INFO [2017-12-08 12:21:39,533] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group file : id=2D1RRAN3P, name=file INFO [2017-12-08 12:21:39,533] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group python : id=2D1A42TEJ, name=python INFO [2017-12-08 12:21:39,533] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group livy : id=2D26DBPPT, name=livy INFO [2017-12-08 12:21:39,533] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group psql : id=2CYQH5RKE, name=psql INFO [2017-12-08 12:21:39,533] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group kylin : id=2D1VSHNAX, name=kylin INFO [2017-12-08 12:21:39,533] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group lens : id=2D3AQXBDD, name=lens INFO [2017-12-08 12:21:39,533] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group jdbc : id=2CZ1RS873, name=jdbc INFO [2017-12-08 12:21:39,534] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group cassandra : id=2D338M3RA, name=cassandra INFO [2017-12-08 12:21:39,534] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group elasticsearch : id=2D372MTWM, name=elasticsearch INFO [2017-12-08 12:21:39,534] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group ignite : id=2D3A12WV4, name=ignite INFO [2017-12-08 12:21:39,534] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group alluxio : id=2D1D8DBB6, name=alluxio INFO [2017-12-08 12:21:39,534] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group hbase : id=2CZ72SGDR, name=hbase INFO [2017-12-08 12:21:39,534] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group pig : id=2CZJA495Z, name=pig INFO [2017-12-08 12:21:39,534] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group spark : id=2CYVF45A9, name=spark INFO [2017-12-08 12:21:39,534] ({main} InterpreterSettingManager.java[init]:366) - InterpreterSetting group bigquery : id=2D1KTF7YE, name=bigquery INFO [2017-12-08 12:21:39,547] ({main} InterpreterFactory.java[<init>]:130) - shiroEnabled: false INFO [2017-12-08 12:21:39,587] ({main} VfsLog.java[info]:138) - Using "/tmp/vfs_cache" as temporary files store. INFO [2017-12-08 12:21:39,621] ({main} GitNotebookRepo.java[<init>]:63) - Opening a git repo at '/opt/zeppelin/notebook' INFO [2017-12-08 12:21:39,840] ({main} NotebookAuthorization.java[loadFromFile]:96) - /etc/zeppelin/notebook-authorization.json INFO [2017-12-08 12:21:39,843] ({main} Credentials.java[loadFromFile]:102) - /etc/zeppelin/credentials.json INFO [2017-12-08 12:21:39,867] ({main} StdSchedulerFactory.java[instantiate]:1184) - Using default implementation for ThreadExecutor INFO [2017-12-08 12:21:39,870] ({main} SimpleThreadPool.java[initialize]:268) - Job execution threads will use class loader of thread: main INFO [2017-12-08 12:21:39,880] ({main} SchedulerSignalerImpl.java[<init>]:61) - Initialized Scheduler Signaller of type: class org.quartz.core.SchedulerSignalerImpl INFO [2017-12-08 12:21:39,881] ({main} QuartzScheduler.java[<init>]:240) - Quartz Scheduler v.2.2.1 created. INFO [2017-12-08 12:21:39,881] ({main} RAMJobStore.java[initialize]:155) - RAMJobStore initialized. INFO [2017-12-08 12:21:39,882] ({main} QuartzScheduler.java[initialize]:305) - Scheduler meta-data: Quartz Scheduler (v2.2.1) 'DefaultQuartzScheduler' with instanceId 'NON_CLUSTERED' Scheduler class: 'org.quartz.core.QuartzScheduler' - running locally. NOT STARTED. Currently in standby mode. Number of jobs executed: 0 Using thread pool 'org.quartz.simpl.SimpleThreadPool' - with 10 threads. Using job-store 'org.quartz.simpl.RAMJobStore' - which does not support persistence. and is not clustered. INFO [2017-12-08 12:21:39,882] ({main} StdSchedulerFactory.java[instantiate]:1339) - Quartz scheduler 'DefaultQuartzScheduler' initialized from default resource file in Quartz package: 'quartz.properties' INFO [2017-12-08 12:21:39,882] ({main} StdSchedulerFactory.java[instantiate]:1343) - Quartz scheduler version: 2.2.1 INFO [2017-12-08 12:21:39,882] ({main} QuartzScheduler.java[start]:575) - Scheduler DefaultQuartzScheduler_$_NON_CLUSTERED started. INFO [2017-12-08 12:21:39,982] ({main} FolderView.java[createFolder]:107) - Create folder Zeppelin Tutorial INFO [2017-12-08 12:21:39,982] ({main} FolderView.java[createFolder]:107) - Create folder / INFO [2017-12-08 12:21:39,983] ({main} Folder.java[setParent]:168) - Set parent of / to / INFO [2017-12-08 12:21:39,983] ({main} Folder.java[setParent]:168) - Set parent of Zeppelin Tutorial to / INFO [2017-12-08 12:21:39,983] ({main} Folder.java[addNote]:184) - Add note 2C2AUG798 to folder Zeppelin Tutorial INFO [2017-12-08 12:21:39,994] ({main} Folder.java[addNote]:184) - Add note 2BWJFTXKJ to folder Zeppelin Tutorial INFO [2017-12-08 12:21:40,000] ({main} Folder.java[addNote]:184) - Add note 2C35YU814 to folder Zeppelin Tutorial INFO [2017-12-08 12:21:40,002] ({main} Folder.java[addNote]:184) - Add note 2A94M5J1Z to folder Zeppelin Tutorial INFO [2017-12-08 12:21:40,009] ({main} Folder.java[addNote]:184) - Add note 2BYEZ5EVK to folder Zeppelin Tutorial INFO [2017-12-08 12:21:40,011] ({main} Folder.java[addNote]:184) - Add note 2C57UKYWR to folder Zeppelin Tutorial INFO [2017-12-08 12:21:40,011] ({main} Notebook.java[<init>]:127) - Notebook indexing started... INFO [2017-12-08 12:21:40,142] ({main} LuceneSearch.java[addIndexDocs]:305) - Indexing 6 notebooks took 130ms INFO [2017-12-08 12:21:40,142] ({main} Notebook.java[<init>]:129) - Notebook indexing finished: 6 indexed in 0s INFO [2017-12-08 12:21:40,245] ({main} ServerImpl.java[initDestination]:94) - Setting the server's publish address to be / INFO [2017-12-08 12:21:40,688] ({main} ContextHandler.java[doStart]:744) - Started o.e.j.w.WebAppContext@418e7838{/,file:/opt/zeppelin-0.7.3/webapps/webapp/,AVAILABLE}{/opt/zeppelin/zeppelin-web-0.7.3.war}<mailto:o.e.j.w.WebAppContext@418e7838%7b/,file:/opt/zeppelin-0.7.3/webapps/webapp/,AVAILABLE%7d%7b/opt/zeppelin/zeppelin-web-0.7.3.war%7d> INFO [2017-12-08 12:21:40,743] ({main} AbstractConnector.java[doStart]:266) - Started ServerConnector@2630dbc4{SSL-HTTP/1.1}{0.0.0.0:8080<http://0.0.0.0:8080>} INFO [2017-12-08 12:21:40,743] ({main} Server.java[doStart]:379) - Started @4103ms INFO [2017-12-08 12:21:40,743] ({main} ZeppelinServer.java[main]:197) - Done, zeppelin server started INFO [2017-12-08 12:21:42,392] ({qtp1632392469-71} NotebookServer.java[onOpen]:157) - New connection from 10.136.169.200 : 43470 INFO [2017-12-08 12:21:46,800] ({qtp1632392469-71} AbstractValidatingSessionManager.java[enableSessionValidation]:230) - Enabling session validation scheduler... INFO [2017-12-08 12:21:46,864] ({qtp1632392469-71} AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or cacheManager properties have been set. Authorization cache cannot be obtained. INFO [2017-12-08 12:21:46,896] ({qtp1632392469-71} AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or cacheManager properties have been set. Authorization cache cannot be obtained. INFO [2017-12-08 12:21:46,931] ({qtp1632392469-71} AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or cacheManager properties have been set. Authorization cache cannot be obtained. INFO [2017-12-08 12:21:46,958] ({qtp1632392469-71} AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or cacheManager properties have been set. Authorization cache cannot be obtained. INFO [2017-12-08 12:21:46,992] ({qtp1632392469-71} AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or cacheManager properties have been set. Authorization cache cannot be obtained. INFO [2017-12-08 12:21:47,019] ({qtp1632392469-71} AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or cacheManager properties have been set. Authorization cache cannot be obtained. INFO [2017-12-08 12:21:47,047] ({qtp1632392469-71} AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or cacheManager properties have been set. Authorization cache cannot be obtained. INFO [2017-12-08 12:21:47,076] ({qtp1632392469-71} AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or cacheManager properties have been set. Authorization cache cannot be obtained. WARN [2017-12-08 12:21:47,109] ({qtp1632392469-71} LoginRestApi.java[postLogin]:119) - {"status":"OK","message":"","body":{"principal":"zepptest","ticket":"e0c65401-e791-4fd9-bcc0-ec68b47f2b27","roles":"[ipausers]"}} INFO [2017-12-08 12:22:03,273] ({qtp1632392469-71} AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or cacheManager properties have been set. Authorization cache cannot be obtained. INFO [2017-12-08 12:22:03,290] ({qtp1632392469-77} AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or cacheManager properties have been set. Authorization cache cannot be obtained. INFO [2017-12-08 12:22:03,305] ({qtp1632392469-79} AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or cacheManager properties have been set. Authorization cache cannot be obtained. OUT Dec 08, 2017 12:20:03 PM com.sun.jersey.api.core.ScanningResourceConfig logClasses INFO: Root resource classes found: class org.apache.zeppelin.rest.ConfigurationsRestApi class org.apache.zeppelin.rest.InterpreterRestApi class org.apache.zeppelin.rest.CredentialRestApi class org.apache.zeppelin.rest.LoginRestApi class org.apache.zeppelin.rest.NotebookRestApi class org.apache.zeppelin.rest.NotebookRepoRestApi class org.apache.zeppelin.rest.SecurityRestApi class org.apache.zeppelin.rest.ZeppelinRestApi class org.apache.zeppelin.rest.HeliumRestApi Dec 08, 2017 12:20:03 PM com.sun.jersey.api.core.ScanningResourceConfig init INFO: No provider classes found. Dec 08, 2017 12:20:03 PM com.sun.jersey.server.impl.application.WebApplicationImpl _initiate INFO: Initiating Jersey application, version 'Jersey: 1.13 06/29/2012 05:14 PM' Dec 08, 2017 12:20:03 PM com.sun.jersey.spi.inject.Errors processErrorMessages WARNING: The following warnings have been detected with resource and/or provider classes: WARNING: A HTTP GET method, public javax.ws.rs.core.Response org.apache.zeppelin.rest.CredentialRestApi.getCredentials(java.lang.String) throws java.io.IOException,java.lang.IllegalArgumentException, should not consume any entity. WARNING: A HTTP GET method, public javax.ws.rs.core.Response org.apache.zeppelin.rest.InterpreterRestApi.listInterpreter(java.lang.String), should not consume any entity. WARNING: A sub-resource method, public javax.ws.rs.core.Response org.apache.zeppelin.rest.NotebookRestApi.createNote(java.lang.String) throws java.io.IOException, with URI template, "/", is treated as a resource method WARNING: A sub-resource method, public javax.ws.rs.core.Response org.apache.zeppelin.rest.NotebookRestApi.getNoteList() throws java.io.IOException, with URI template, "/", is treated as a resource method ZEPPELIN_CLASSPATH: ::/opt/zeppelin/lib/interpreter/*:/opt/zeppelin/lib/*:/opt/zeppelin/*::/etc/zeppelin Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=512m; support was removed in 8.0 SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/opt/zeppelin-0.7.3/lib/interpreter/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/opt/zeppelin-0.7.3/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] Dec 08, 2017 12:20:37 PM com.sun.jersey.api.core.PackagesResourceConfig init INFO: Scanning for root resource and provider classes in the packages: org.apache.zeppelin.rest De : Jeff Zhang [mailto:zjf...@gmail.com<mailto:zjf...@gmail.com>] Envoyé : mardi 6 février 2018 11:17 À : users@zeppelin.apache.org<mailto:users@zeppelin.apache.org> Objet : Re: How to have a native graphical representation (%sql) of a HiveContext? Could you attach the spark interpreter log ? LINZ, Arnaud <al...@bouyguestelecom.fr<mailto:al...@bouyguestelecom.fr>>于2018年2月6日周二 下午4:49写道: I really tried, but it really does not work… * zeppelin.spark.useHiveContext is set to true * I directly use sqlContext without creating it * I have a link to hive-site.xml in /etc/spark/conf But sqlContext does not see any hive table I don’t see any error log, or any helpful log in zeppelin logs. What can be wrong? De : Jeff Zhang [mailto:zjf...@gmail.com<mailto:zjf...@gmail.com>] Envoyé : lundi 5 février 2018 14:31 À : users@zeppelin.apache.org<mailto:users@zeppelin.apache.org> Objet : Re: How to have a native graphical representation (%sql) of a HiveContext? Sorry I should be more accurate. Just don't create SqlContext/HiveContext in zeppelin as zeppelin will create that for you. If you zeppelin.spark.useHiveContext to true variable sqlContext will be HiveContext, other it is SqlContxt LINZ, Arnaud <al...@bouyguestelecom.fr<mailto:al...@bouyguestelecom.fr>>于2018年2月5日周一 下午9:15写道: Thanks for your answer, but it does not address my problem. * I don’t create sqlContext, I use the one provided by Zeppelin. But sqlContext is not a hive Context and cannot access hive metastore. * Zeppelin can see my hive conf files, and selecting tables through a created hiveContext works. But I cannot visualize them in the %sql graphical interpretor. De : Jeff Zhang [mailto:zjf...@gmail.com<mailto:zjf...@gmail.com>] Envoyé : lundi 5 février 2018 14:01 À : users@zeppelin.apache.org<mailto:users@zeppelin.apache.org> Objet : Re: How to have a native graphical representation (%sql) of a HiveContext? 1. Don't create sqlContext in zeppelin as zeppelin will create that for you, and %sql use the sqlContext created by zeppelin itself. 2. Make sure you have hive-site.xml under SPARK_CONF_DIR if you want to use hiveContext. Otherwise spark will use single user derby instance which is not for production, and will cause conflicts when you create multiple spark interpreter in one zeppelin instance. LINZ, Arnaud <al...@bouyguestelecom.fr<mailto:al...@bouyguestelecom.fr>>于2018年2月5日周一 下午8:33写道: Hello, I’m trying to install Zeppelin (0.7.2) on my CDH cluster, and I am unable to connect the sql + graphical representations of the %sql interpreter with my Hive data, and more surprisingly I really can’t find any good source on the internet (apache zeppelin documentation or stack overflow) that gives a practical answer about how to do this. Most of the time, the data comes from compressed Hive tables and not plain hdfs text files ; so using a hive context is far more convenient than a plain spark sql context. The following : %spark val hc = new org.apache.spark.sql.hive.HiveContext(sc) val result = hc.sql("select * from hivedb.hivetable") result.registerTempTable("myTest") works but no myTest table is available in the following %sql interpreter : %sql select * from myTest org.apache.spark.sql.AnalysisException: Table not found: myTest; However the following : %pyspark result = sqlContext.read.text("hdfs://cluster/test.txt") result.registerTempTable("mySqlTest") works as the %sql interpreter is “plugged” to the sqlContext but result = sqlContext.sql("select * from hivedb.hivetable") does not work as the sqlContext is not a hive context. I have set zeppelin.spark.useHiveContext to true, but it seems to have no effect (btw, it was more of a wild guess since the documentation is not giving much detail on parameters and context configuration) Can you direct me towards how to configure the context used by the %sql interpreter? Best regards, Arnaud PS : %spark and %sql interpreter conf: master yarn-client spark.app.name<http://spark.app.name> Zeppelin spark.cores.max spark.executor.memory 5g zeppelin.R.cmd R zeppelin.R.image.width 100% zeppelin.R.knitr true zeppelin.R.render.options out.format = 'html', comment = NA, echo = FALSE, results = 'asis', message = F, warning = F zeppelin.dep.additionalRemoteRepository spark-packages,http://dl.bintray.com/spark-packages/maven,false; zeppelin.dep.localrepo local-repo zeppelin.interpreter.localRepo /opt/zeppelin/local-repo/2CYVF45A9 zeppelin.interpreter.output.limit 102400 zeppelin.pyspark.python /usr/bin/pyspark zeppelin.spark.concurrentSQL true zeppelin.spark.importImplicit true zeppelin.spark.maxResult 1000 zeppelin.spark.printREPLOutput true zeppelin.spark.sql.stacktrace true zeppelin.spark.useHiveContext true ________________________________ L'intégrité de ce message n'étant pas assurée sur internet, la société expéditrice ne peut être tenue responsable de son contenu ni de ses pièces jointes. Toute utilisation ou diffusion non autorisée est interdite. Si vous n'êtes pas destinataire de ce message, merci de le détruire et d'avertir l'expéditeur. The integrity of this message cannot be guaranteed on the Internet. The company that sent this message cannot therefore be held liable for its content nor attachments. Any unauthorized use or dissemination is prohibited. If you are not the intended recipient of this message, then please delete it and notify the sender.