Ooops, sorry. No zeppelin-interpreter-log, but I did looked at the wrong log
file.
No error on startup, and here is the stack trace when I query a hive table :
INFO [2018-02-06 12:12:18,405] ({qtp1632392469-74}
AbstractValidatingSessionManager.java[enableSessionValidation]:230) - Enabling
session validation scheduler...
WARN [2018-02-06 12:12:18,484] ({qtp1632392469-74}
LoginRestApi.java[postLogin]:119) -
{"status":"OK","message":"","body":{"principal":"alinz","ticket":"f4ebf4ec-d889-4d01-808a-ee8b9800c80e","roles":"[]"}}
INFO [2018-02-06 12:12:18,710] ({qtp1632392469-74}
NotebookServer.java[sendNote]:711) - New operation from 172.24.193.234 : 61997
: alinz : GET_NOTE : 2D1SETNNQ
WARN [2018-02-06 12:12:18,764] ({qtp1632392469-74}
GitNotebookRepo.java[revisionHistory]:157) - No Head found for 2D1SETNNQ, No
HEAD exists and no explicit starting revision was specified
INFO [2018-02-06 12:12:19,102] ({qtp1632392469-71}
InterpreterFactory.java[createInterpretersForNote]:188) - Create interpreter
instance spark for note 2D1SETNNQ
INFO [2018-02-06 12:12:19,107] ({qtp1632392469-71}
InterpreterFactory.java[createInterpretersForNote]:221) - Interpreter
org.apache.zeppelin.spark.SparkInterpreter 1563969204 created
INFO [2018-02-06 12:12:19,109] ({qtp1632392469-71}
InterpreterFactory.java[createInterpretersForNote]:221) - Interpreter
org.apache.zeppelin.spark.SparkSqlInterpreter 119386711 created
INFO [2018-02-06 12:12:19,110] ({qtp1632392469-71}
InterpreterFactory.java[createInterpretersForNote]:221) - Interpreter
org.apache.zeppelin.spark.DepInterpreter 1706768324 created
INFO [2018-02-06 12:12:19,111] ({qtp1632392469-71}
InterpreterFactory.java[createInterpretersForNote]:221) - Interpreter
org.apache.zeppelin.spark.PySparkInterpreter 261340605 created
INFO [2018-02-06 12:12:19,112] ({qtp1632392469-71}
InterpreterFactory.java[createInterpretersForNote]:221) - Interpreter
org.apache.zeppelin.spark.SparkRInterpreter 276911901 created
INFO [2018-02-06 12:12:19,111] ({qtp1632392469-71}
InterpreterFactory.java[createInterpretersForNote]:221) - Interpreter
org.apache.zeppelin.spark.PySparkInterpreter 261340605 created
INFO [2018-02-06 12:12:19,112] ({qtp1632392469-71}
InterpreterFactory.java[createInterpretersForNote]:221) - Interpreter
org.apache.zeppelin.spark.SparkRInterpreter 276911901 created
INFO [2018-02-06 12:13:38,419] ({pool-2-thread-2}
SchedulerFactory.java[jobStarted]:131) - Job paragraph_1513703910594_823567073
started by scheduler
org.apache.zeppelin.interpreter.remote.RemoteInterpretershared_session1977922636
INFO [2018-02-06 12:13:38,420] ({pool-2-thread-2} Paragraph.java[jobRun]:362) -
run paragraph 20171219-181830_1097537317 using pyspark
org.apache.zeppelin.interpreter.LazyOpenInterpreter@f93bdbd<mailto:org.apache.zeppelin.interpreter.LazyOpenInterpreter@f93bdbd>
INFO [2018-02-06 12:13:38,428] ({pool-2-thread-2}
RemoteInterpreterManagedProcess.java[start]:126) - Run interpreter process
[/opt/zeppelin/bin/interpreter.sh, -d, /opt/zeppelin/interpreter/spark, -p,
43118, -u, alinz, -l, /opt/zeppelin/local-repo/2CYVF45A9]
INFO [2018-02-06 12:13:39,474] ({pool-2-thread-2}
RemoteInterpreter.java[init]:221) - Create remote interpreter
org.apache.zeppelin.spark.SparkInterpreter
INFO [2018-02-06 12:13:39,570] ({pool-2-thread-2}
RemoteInterpreter.java[pushAngularObjectRegistryToRemote]:551) - Push local
angular object registry from ZeppelinServer to remote interpreter group
2CYVF45A9:alinz:
INFO [2018-02-06 12:13:39,585] ({pool-2-thread-2}
RemoteInterpreter.java[init]:221) - Create remote interpreter
org.apache.zeppelin.spark.SparkSqlInterpreter
INFO [2018-02-06 12:13:39,587] ({pool-2-thread-2}
RemoteInterpreter.java[init]:221) - Create remote interpreter
org.apache.zeppelin.spark.DepInterpreter
INFO [2018-02-06 12:13:39,592] ({pool-2-thread-2}
RemoteInterpreter.java[init]:221) - Create remote interpreter
org.apache.zeppelin.spark.PySparkInterpreter
INFO [2018-02-06 12:13:39,603] ({pool-2-thread-2}
RemoteInterpreter.java[init]:221) - Create remote interpreter
org.apache.zeppelin.spark.SparkRInterpreter
WARN [2018-02-06 12:13:53,749] ({pool-2-thread-2}
NotebookServer.java[afterStatusChange]:2064) - Job 20171219-181830_1097537317
is finished, status: ERROR, exception: null, result: %text WARNING: Running
python applications through 'pyspark' is deprecated as of Spark 1.0.
se ./bin/spark-submit <python file>
/tmp/zeppelin_pyspark-5465626759760627770.py:151: UserWarning: Unable to load
inline matplotlib backend, falling back to Agg
warnings.warn("Unable to load inline matplotlib backend, "
%text Traceback (most recent call last):
File
"/opt/cloudera/parcels/CDH-5.11.1-1.cdh5.11.1.p0.4/lib/spark/python/lib/pyspark.zip/pyspark/sql/utils.py",
line 45, in deco
return f(*a, **kw)
File
"/opt/cloudera/parcels/CDH-5.11.1-1.cdh5.11.1.p0.4/lib/spark/python/lib/py4j-0.9-src.zip/py4j/protocol.py",
line 308, in get_return_value
format(target_id, ".", name), value)
py4j.protocol.Py4JJavaError: An error occurred while calling o73.sql.
: org.apache.spark.sql.AnalysisException: Table not found:
`hiveDatabase`.`hiveTable`;
at
org.apache.spark.sql.catalyst.analysis.package$AnalysisErrorAt.failAnalysis(package.scala:42)
at
org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$checkAnalysis$1.apply(CheckAnalysis.scala:54)
at
org.apache.spark.sql.catalyst.analysis.CheckAnalysis$$anonfun$checkAnalysis$1.apply(CheckAnalysis.scala:50)
at
org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:121)
at
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$foreachUp$1.apply(TreeNode.scala:120)
at
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$foreachUp$1.apply(TreeNode.scala:120)
at scala.collection.immutable.List.foreach(List.scala:318)
at
org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:120)
at
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$foreachUp$1.apply(TreeNode.scala:120)
at
org.apache.spark.sql.catalyst.trees.TreeNode$$anonfun$foreachUp$1.apply(TreeNode.scala:120)
at scala.collection.immutable.List.foreach(List.scala:318)
at
org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:120)
at
org.apache.spark.sql.catalyst.analysis.CheckAnalysis$class.checkAnalysis(CheckAnalysis.scala:50)
at
org.apache.spark.sql.catalyst.analysis.Analyzer.checkAnalysis(Analyzer.scala:44)
at
org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:35)
at org.apache.spark.sql.DataFrame.<init>(DataFrame.scala:133)
at org.apache.spark.sql.DataFrame$.apply(DataFrame.scala:52)
at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:829)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:231)
at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:381)
at py4j.Gateway.invoke(Gateway.java:259)
at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
at py4j.commands.CallCommand.execute(CallCommand.java:79)
at py4j.GatewayConnection.run(GatewayConnection.java:209)
at java.lang.Thread.run(Thread.java:745)
(…)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/tmp/zeppelin_pyspark-5465626759760627770.py", line 355, in <module>
exec(code, _zcUserQueryNameSpace)
File "<stdin>", line 1, in <module>
File
"/opt/cloudera/parcels/CDH-5.11.1-1.cdh5.11.1.p0.4/lib/spark/python/lib/pyspark.zip/pyspark/sql/context.py",
line 580, in sql
return DataFrame(self._ssql_ctx.sql(sqlQuery), self)
File
"/opt/cloudera/parcels/CDH-5.11.1-1.cdh5.11.1.p0.4/lib/spark/python/lib/py4j-0.9-src.zip/py4j/java_gateway.py",
line 813, in __call__
answer, self.gateway_client, self.target_id, self.name)
File
"/opt/cloudera/parcels/CDH-5.11.1-1.cdh5.11.1.p0.4/lib/spark/python/lib/pyspark.zip/pyspark/sql/utils.py",
line 51, in deco
raise AnalysisException(s.split(': ', 1)[1], stackTrace)
pyspark.sql.utils.AnalysisException: 'Table not found:
`hiveDatabase`.`hiveTable`;'
INFO [2018-02-06 12:13:53,795] ({pool-2-thread-2}
SchedulerFactory.java[jobFinished]:137) - Job paragraph_1513703910594_823567073
finished by scheduler
org.apache.zeppelin.interpreter.remote.RemoteInterpretershared_session1977922636
De : Jeff Zhang [mailto:[email protected]]
Envoyé : mardi 6 février 2018 12:04
À : [email protected]
Objet : Re: How to have a native graphical representation (%sql) of a
HiveContext?
It is weird, there should be a interpreter log file with name
'zeppelin-interpreter-spark*.log'
Are you sure you are looking at the correct log folder ? Because from the log
you pasted, it is the log of 2017-12-08
LINZ, Arnaud
<[email protected]<mailto:[email protected]>>于2018年2月6日周二
下午6:45写道:
I’ve been searching for such a log, but I don’t see anything related to spark
in …/zeppelin/logs
The only logs I have are
zeppelin-$USER-$HOST.log
and
zeppelin-$USER-$HOST.out
but they really don’t contain anything useful.
Log =
INFO [2017-12-08 12:21:36,847] ({main} ZeppelinConfiguration.java[create]:101)
- Load configuration from file:/etc/zeppelin/zeppelin-site.xml
INFO [2017-12-08 12:21:36,896] ({main} ZeppelinConfiguration.java[create]:109)
- Server Host: 0.0.0.0
INFO [2017-12-08 12:21:36,896] ({main} ZeppelinConfiguration.java[create]:113)
- Server SSL Port: 8080
INFO [2017-12-08 12:21:36,896] ({main} ZeppelinConfiguration.java[create]:115)
- Context Path: /
INFO [2017-12-08 12:21:36,900] ({main} ZeppelinConfiguration.java[create]:116)
- Zeppelin Version: 0.7.3
INFO [2017-12-08 12:21:36,917] ({main} Log.java[initialized]:186) - Logging
initialized @275ms
INFO [2017-12-08 12:21:36,979] ({main}
ZeppelinServer.java[setupWebAppContext]:346) - ZeppelinServer Webapp path:
/opt/zeppelin/webapps
INFO [2017-12-08 12:21:37,196] ({main} IniRealm.java[processDefinitions]:188) -
IniRealm defined, but there is no [users] section defined. This realm will not
be populated with any users and it is assumed that they will be populated
programatically. Users must be defined for this Realm instance to be useful.
INFO [2017-12-08 12:21:37,197] ({main}
AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or
cacheManager properties have been set. Authorization cache cannot be obtained.
INFO [2017-12-08 12:21:37,227] ({main} ZeppelinServer.java[main]:187) -
Starting zeppelin server
INFO [2017-12-08 12:21:37,229] ({main} Server.java[doStart]:327) -
jetty-9.2.15.v20160210
INFO [2017-12-08 12:21:38,913] ({main}
StandardDescriptorProcessor.java[visitServlet]:297) - NO JSP Support for /, did
not find org.eclipse.jetty.jsp.JettyJspServlet
INFO [2017-12-08 12:21:38,922] ({main} ContextHandler.java[log]:2052) -
Initializing Shiro environment
INFO [2017-12-08 12:21:38,922] ({main}
EnvironmentLoader.java[initEnvironment]:128) - Starting Shiro environment
initialization.
INFO [2017-12-08 12:21:38,972] ({main} IniRealm.java[processDefinitions]:188) -
IniRealm defined, but there is no [users] section defined. This realm will not
be populated with any users and it is assumed that they will be populated
programatically. Users must be defined for this Realm instance to be useful.
INFO [2017-12-08 12:21:38,972] ({main}
AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or
cacheManager properties have been set. Authorization cache cannot be obtained.
INFO [2017-12-08 12:21:38,977] ({main}
EnvironmentLoader.java[initEnvironment]:141) - Shiro environment initialized in
55 ms.
WARN [2017-12-08 12:21:39,115] ({main} Helium.java[loadConf]:101) -
/etc/zeppelin/helium.json does not exists
WARN [2017-12-08 12:21:39,402] ({main} Interpreter.java[register]:406) - Static
initialization is deprecated for interpreter sql, You should change it to use
interpreter-setting.json in your jar or
interpreter/{interpreter}/interpreter-setting.json
INFO [2017-12-08 12:21:39,403] ({main}
InterpreterSettingManager.java[init]:305) - Interpreter psql.sql found.
class=org.apache.zeppelin.postgresql.PostgreSqlInterpreter
INFO [2017-12-08 12:21:39,497] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name ignite
INFO [2017-12-08 12:21:39,497] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name python
INFO [2017-12-08 12:21:39,497] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name jdbc
INFO [2017-12-08 12:21:39,497] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name psql
INFO [2017-12-08 12:21:39,497] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name lens
INFO [2017-12-08 12:21:39,497] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name pig
INFO [2017-12-08 12:21:39,498] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name flink
INFO [2017-12-08 12:21:39,498] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name angular
INFO [2017-12-08 12:21:39,498] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name livy
INFO [2017-12-08 12:21:39,498] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name file
INFO [2017-12-08 12:21:39,498] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name
elasticsearch
INFO [2017-12-08 12:21:39,498] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name cassandra
INFO [2017-12-08 12:21:39,498] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name sh
INFO [2017-12-08 12:21:39,498] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name spark
INFO [2017-12-08 12:21:39,498] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name md
INFO [2017-12-08 12:21:39,498] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name alluxio
INFO [2017-12-08 12:21:39,499] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name bigquery
INFO [2017-12-08 12:21:39,499] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name hbase
INFO [2017-12-08 12:21:39,499] ({main}
InterpreterSettingManager.java[init]:337) - InterpreterSettingRef name kylin
INFO [2017-12-08 12:21:39,532] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group md :
id=2D18QE9AX, name=md
INFO [2017-12-08 12:21:39,533] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group flink :
id=2CYSDG6AU, name=flink
INFO [2017-12-08 12:21:39,533] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group angular :
id=2CZY6QWUG, name=angular
INFO [2017-12-08 12:21:39,533] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group sh :
id=2CYZFQZXG, name=sh
INFO [2017-12-08 12:21:39,533] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group file :
id=2D1RRAN3P, name=file
INFO [2017-12-08 12:21:39,533] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group python :
id=2D1A42TEJ, name=python
INFO [2017-12-08 12:21:39,533] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group livy :
id=2D26DBPPT, name=livy
INFO [2017-12-08 12:21:39,533] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group psql :
id=2CYQH5RKE, name=psql
INFO [2017-12-08 12:21:39,533] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group kylin :
id=2D1VSHNAX, name=kylin
INFO [2017-12-08 12:21:39,533] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group lens :
id=2D3AQXBDD, name=lens
INFO [2017-12-08 12:21:39,533] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group jdbc :
id=2CZ1RS873, name=jdbc
INFO [2017-12-08 12:21:39,534] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group cassandra
: id=2D338M3RA, name=cassandra
INFO [2017-12-08 12:21:39,534] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group
elasticsearch : id=2D372MTWM, name=elasticsearch
INFO [2017-12-08 12:21:39,534] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group ignite :
id=2D3A12WV4, name=ignite
INFO [2017-12-08 12:21:39,534] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group alluxio :
id=2D1D8DBB6, name=alluxio
INFO [2017-12-08 12:21:39,534] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group hbase :
id=2CZ72SGDR, name=hbase
INFO [2017-12-08 12:21:39,534] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group pig :
id=2CZJA495Z, name=pig
INFO [2017-12-08 12:21:39,534] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group spark :
id=2CYVF45A9, name=spark
INFO [2017-12-08 12:21:39,534] ({main}
InterpreterSettingManager.java[init]:366) - InterpreterSetting group bigquery :
id=2D1KTF7YE, name=bigquery
INFO [2017-12-08 12:21:39,547] ({main} InterpreterFactory.java[<init>]:130) -
shiroEnabled: false
INFO [2017-12-08 12:21:39,587] ({main} VfsLog.java[info]:138) - Using
"/tmp/vfs_cache" as temporary files store.
INFO [2017-12-08 12:21:39,621] ({main} GitNotebookRepo.java[<init>]:63) -
Opening a git repo at '/opt/zeppelin/notebook'
INFO [2017-12-08 12:21:39,840] ({main}
NotebookAuthorization.java[loadFromFile]:96) -
/etc/zeppelin/notebook-authorization.json
INFO [2017-12-08 12:21:39,843] ({main} Credentials.java[loadFromFile]:102) -
/etc/zeppelin/credentials.json
INFO [2017-12-08 12:21:39,867] ({main}
StdSchedulerFactory.java[instantiate]:1184) - Using default implementation for
ThreadExecutor
INFO [2017-12-08 12:21:39,870] ({main} SimpleThreadPool.java[initialize]:268) -
Job execution threads will use class loader of thread: main
INFO [2017-12-08 12:21:39,880] ({main} SchedulerSignalerImpl.java[<init>]:61) -
Initialized Scheduler Signaller of type: class
org.quartz.core.SchedulerSignalerImpl
INFO [2017-12-08 12:21:39,881] ({main} QuartzScheduler.java[<init>]:240) -
Quartz Scheduler v.2.2.1 created.
INFO [2017-12-08 12:21:39,881] ({main} RAMJobStore.java[initialize]:155) -
RAMJobStore initialized.
INFO [2017-12-08 12:21:39,882] ({main} QuartzScheduler.java[initialize]:305) -
Scheduler meta-data: Quartz Scheduler (v2.2.1) 'DefaultQuartzScheduler' with
instanceId 'NON_CLUSTERED'
Scheduler class: 'org.quartz.core.QuartzScheduler' - running locally.
NOT STARTED.
Currently in standby mode.
Number of jobs executed: 0
Using thread pool 'org.quartz.simpl.SimpleThreadPool' - with 10 threads.
Using job-store 'org.quartz.simpl.RAMJobStore' - which does not support
persistence. and is not clustered.
INFO [2017-12-08 12:21:39,882] ({main}
StdSchedulerFactory.java[instantiate]:1339) - Quartz scheduler
'DefaultQuartzScheduler' initialized from default resource file in Quartz
package: 'quartz.properties'
INFO [2017-12-08 12:21:39,882] ({main}
StdSchedulerFactory.java[instantiate]:1343) - Quartz scheduler version: 2.2.1
INFO [2017-12-08 12:21:39,882] ({main} QuartzScheduler.java[start]:575) -
Scheduler DefaultQuartzScheduler_$_NON_CLUSTERED started.
INFO [2017-12-08 12:21:39,982] ({main} FolderView.java[createFolder]:107) -
Create folder Zeppelin Tutorial
INFO [2017-12-08 12:21:39,982] ({main} FolderView.java[createFolder]:107) -
Create folder /
INFO [2017-12-08 12:21:39,983] ({main} Folder.java[setParent]:168) - Set parent
of / to /
INFO [2017-12-08 12:21:39,983] ({main} Folder.java[setParent]:168) - Set parent
of Zeppelin Tutorial to /
INFO [2017-12-08 12:21:39,983] ({main} Folder.java[addNote]:184) - Add note
2C2AUG798 to folder Zeppelin Tutorial
INFO [2017-12-08 12:21:39,994] ({main} Folder.java[addNote]:184) - Add note
2BWJFTXKJ to folder Zeppelin Tutorial
INFO [2017-12-08 12:21:40,000] ({main} Folder.java[addNote]:184) - Add note
2C35YU814 to folder Zeppelin Tutorial
INFO [2017-12-08 12:21:40,002] ({main} Folder.java[addNote]:184) - Add note
2A94M5J1Z to folder Zeppelin Tutorial
INFO [2017-12-08 12:21:40,009] ({main} Folder.java[addNote]:184) - Add note
2BYEZ5EVK to folder Zeppelin Tutorial
INFO [2017-12-08 12:21:40,011] ({main} Folder.java[addNote]:184) - Add note
2C57UKYWR to folder Zeppelin Tutorial
INFO [2017-12-08 12:21:40,011] ({main} Notebook.java[<init>]:127) - Notebook
indexing started...
INFO [2017-12-08 12:21:40,142] ({main} LuceneSearch.java[addIndexDocs]:305) -
Indexing 6 notebooks took 130ms
INFO [2017-12-08 12:21:40,142] ({main} Notebook.java[<init>]:129) - Notebook
indexing finished: 6 indexed in 0s
INFO [2017-12-08 12:21:40,245] ({main} ServerImpl.java[initDestination]:94) -
Setting the server's publish address to be /
INFO [2017-12-08 12:21:40,688] ({main} ContextHandler.java[doStart]:744) -
Started
o.e.j.w.WebAppContext@418e7838{/,file:/opt/zeppelin-0.7.3/webapps/webapp/,AVAILABLE}{/opt/zeppelin/zeppelin-web-0.7.3.war}<mailto:o.e.j.w.WebAppContext@418e7838%7b/,file:/opt/zeppelin-0.7.3/webapps/webapp/,AVAILABLE%7d%7b/opt/zeppelin/zeppelin-web-0.7.3.war%7d>
INFO [2017-12-08 12:21:40,743] ({main} AbstractConnector.java[doStart]:266) -
Started
ServerConnector@2630dbc4{SSL-HTTP/1.1}{0.0.0.0:8080<http://0.0.0.0:8080>}
INFO [2017-12-08 12:21:40,743] ({main} Server.java[doStart]:379) - Started
@4103ms
INFO [2017-12-08 12:21:40,743] ({main} ZeppelinServer.java[main]:197) - Done,
zeppelin server started
INFO [2017-12-08 12:21:42,392] ({qtp1632392469-71}
NotebookServer.java[onOpen]:157) - New connection from 10.136.169.200 : 43470
INFO [2017-12-08 12:21:46,800] ({qtp1632392469-71}
AbstractValidatingSessionManager.java[enableSessionValidation]:230) - Enabling
session validation scheduler...
INFO [2017-12-08 12:21:46,864] ({qtp1632392469-71}
AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or
cacheManager properties have been set. Authorization cache cannot be obtained.
INFO [2017-12-08 12:21:46,896] ({qtp1632392469-71}
AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or
cacheManager properties have been set. Authorization cache cannot be obtained.
INFO [2017-12-08 12:21:46,931] ({qtp1632392469-71}
AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or
cacheManager properties have been set. Authorization cache cannot be obtained.
INFO [2017-12-08 12:21:46,958] ({qtp1632392469-71}
AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or
cacheManager properties have been set. Authorization cache cannot be obtained.
INFO [2017-12-08 12:21:46,992] ({qtp1632392469-71}
AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or
cacheManager properties have been set. Authorization cache cannot be obtained.
INFO [2017-12-08 12:21:47,019] ({qtp1632392469-71}
AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or
cacheManager properties have been set. Authorization cache cannot be obtained.
INFO [2017-12-08 12:21:47,047] ({qtp1632392469-71}
AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or
cacheManager properties have been set. Authorization cache cannot be obtained.
INFO [2017-12-08 12:21:47,076] ({qtp1632392469-71}
AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or
cacheManager properties have been set. Authorization cache cannot be obtained.
WARN [2017-12-08 12:21:47,109] ({qtp1632392469-71}
LoginRestApi.java[postLogin]:119) -
{"status":"OK","message":"","body":{"principal":"zepptest","ticket":"e0c65401-e791-4fd9-bcc0-ec68b47f2b27","roles":"[ipausers]"}}
INFO [2017-12-08 12:22:03,273] ({qtp1632392469-71}
AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or
cacheManager properties have been set. Authorization cache cannot be obtained.
INFO [2017-12-08 12:22:03,290] ({qtp1632392469-77}
AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or
cacheManager properties have been set. Authorization cache cannot be obtained.
INFO [2017-12-08 12:22:03,305] ({qtp1632392469-79}
AuthorizingRealm.java[getAuthorizationCacheLazy]:248) - No cache or
cacheManager properties have been set. Authorization cache cannot be obtained.
OUT
Dec 08, 2017 12:20:03 PM com.sun.jersey.api.core.ScanningResourceConfig
logClasses
INFO: Root resource classes found:
class org.apache.zeppelin.rest.ConfigurationsRestApi
class org.apache.zeppelin.rest.InterpreterRestApi
class org.apache.zeppelin.rest.CredentialRestApi
class org.apache.zeppelin.rest.LoginRestApi
class org.apache.zeppelin.rest.NotebookRestApi
class org.apache.zeppelin.rest.NotebookRepoRestApi
class org.apache.zeppelin.rest.SecurityRestApi
class org.apache.zeppelin.rest.ZeppelinRestApi
class org.apache.zeppelin.rest.HeliumRestApi
Dec 08, 2017 12:20:03 PM com.sun.jersey.api.core.ScanningResourceConfig init
INFO: No provider classes found.
Dec 08, 2017 12:20:03 PM
com.sun.jersey.server.impl.application.WebApplicationImpl _initiate
INFO: Initiating Jersey application, version 'Jersey: 1.13 06/29/2012 05:14 PM'
Dec 08, 2017 12:20:03 PM com.sun.jersey.spi.inject.Errors processErrorMessages
WARNING: The following warnings have been detected with resource and/or
provider classes:
WARNING: A HTTP GET method, public javax.ws.rs.core.Response
org.apache.zeppelin.rest.CredentialRestApi.getCredentials(java.lang.String)
throws java.io.IOException,java.lang.IllegalArgumentException, should not
consume any entity.
WARNING: A HTTP GET method, public javax.ws.rs.core.Response
org.apache.zeppelin.rest.InterpreterRestApi.listInterpreter(java.lang.String),
should not consume any entity.
WARNING: A sub-resource method, public javax.ws.rs.core.Response
org.apache.zeppelin.rest.NotebookRestApi.createNote(java.lang.String) throws
java.io.IOException, with URI template, "/", is treated as a resource method
WARNING: A sub-resource method, public javax.ws.rs.core.Response
org.apache.zeppelin.rest.NotebookRestApi.getNoteList() throws
java.io.IOException, with URI template, "/", is treated as a resource method
ZEPPELIN_CLASSPATH:
::/opt/zeppelin/lib/interpreter/*:/opt/zeppelin/lib/*:/opt/zeppelin/*::/etc/zeppelin
Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=512m;
support was removed in 8.0
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in
[jar:file:/opt/zeppelin-0.7.3/lib/interpreter/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in
[jar:file:/opt/zeppelin-0.7.3/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
Dec 08, 2017 12:20:37 PM com.sun.jersey.api.core.PackagesResourceConfig init
INFO: Scanning for root resource and provider classes in the packages:
org.apache.zeppelin.rest
De : Jeff Zhang [mailto:[email protected]<mailto:[email protected]>]
Envoyé : mardi 6 février 2018 11:17
À : [email protected]<mailto:[email protected]>
Objet : Re: How to
have a native graphical representation (%sql) of a HiveContext?
Could you attach the spark interpreter log ?
LINZ, Arnaud
<[email protected]<mailto:[email protected]>>于2018年2月6日周二
下午4:49写道:
I really tried, but it really does not work…
* zeppelin.spark.useHiveContext is set to true
* I directly use sqlContext without creating it
* I have a link to hive-site.xml in /etc/spark/conf
But sqlContext does not see any hive table
I don’t see any error log, or any helpful log in zeppelin logs. What can be
wrong?
De : Jeff Zhang [mailto:[email protected]<mailto:[email protected]>]
Envoyé : lundi 5 février 2018 14:31
À : [email protected]<mailto:[email protected]>
Objet : Re: How to have a native graphical representation (%sql) of a
HiveContext?
Sorry I should be more accurate.
Just don't create SqlContext/HiveContext in zeppelin as zeppelin will create
that for you. If you zeppelin.spark.useHiveContext to true variable sqlContext
will be HiveContext, other it is SqlContxt
LINZ, Arnaud
<[email protected]<mailto:[email protected]>>于2018年2月5日周一
下午9:15写道:
Thanks for your answer, but it does not address my problem.
* I don’t create sqlContext, I use the one provided by Zeppelin. But
sqlContext is not a hive Context and cannot access hive metastore.
* Zeppelin can see my hive conf files, and selecting tables through a
created hiveContext works. But I cannot visualize them in the %sql graphical
interpretor.
De : Jeff Zhang [mailto:[email protected]<mailto:[email protected]>]
Envoyé : lundi 5 février 2018 14:01
À : [email protected]<mailto:[email protected]>
Objet : Re: How to have a native graphical representation (%sql) of a
HiveContext?
1. Don't create sqlContext in zeppelin as zeppelin will create that for you,
and %sql use the sqlContext created by zeppelin itself.
2. Make sure you have hive-site.xml under SPARK_CONF_DIR if you want to use
hiveContext. Otherwise spark will use single user derby instance which is not
for production, and will cause conflicts when you create multiple spark
interpreter in one zeppelin instance.
LINZ, Arnaud
<[email protected]<mailto:[email protected]>>于2018年2月5日周一
下午8:33写道:
Hello,
I’m trying to install Zeppelin (0.7.2) on my CDH cluster, and I am unable to
connect the sql + graphical representations of the %sql interpreter with my
Hive data, and more surprisingly I really can’t find any good source on the
internet (apache zeppelin documentation or stack overflow) that gives a
practical answer about how to do this.
Most of the time, the data comes from compressed Hive tables and not plain hdfs
text files ; so using a hive context is far more convenient than a plain spark
sql context.
The following :
%spark
val hc = new org.apache.spark.sql.hive.HiveContext(sc)
val result = hc.sql("select * from hivedb.hivetable")
result.registerTempTable("myTest")
works but no myTest table is available in the following %sql interpreter :
%sql
select * from myTest
org.apache.spark.sql.AnalysisException: Table not found: myTest;
However the following :
%pyspark
result = sqlContext.read.text("hdfs://cluster/test.txt")
result.registerTempTable("mySqlTest")
works as the %sql interpreter is “plugged” to the sqlContext
but
result = sqlContext.sql("select * from hivedb.hivetable") does not work as the
sqlContext is not a hive context.
I have set zeppelin.spark.useHiveContext to true, but it seems to have no
effect (btw, it was more of a wild guess since the documentation is not giving
much detail on parameters and context configuration)
Can you direct me towards how to configure the context used by the %sql
interpreter?
Best regards,
Arnaud
PS : %spark and %sql interpreter conf:
master yarn-client
spark.app.name<http://spark.app.name> Zeppelin
spark.cores.max
spark.executor.memory 5g
zeppelin.R.cmd R
zeppelin.R.image.width 100%
zeppelin.R.knitr true
zeppelin.R.render.options out.format = 'html', comment = NA, echo = FALSE,
results = 'asis', message = F, warning = F
zeppelin.dep.additionalRemoteRepository
spark-packages,http://dl.bintray.com/spark-packages/maven,false;
zeppelin.dep.localrepo local-repo
zeppelin.interpreter.localRepo /opt/zeppelin/local-repo/2CYVF45A9
zeppelin.interpreter.output.limit 102400
zeppelin.pyspark.python /usr/bin/pyspark
zeppelin.spark.concurrentSQL true
zeppelin.spark.importImplicit true
zeppelin.spark.maxResult 1000
zeppelin.spark.printREPLOutput true
zeppelin.spark.sql.stacktrace true
zeppelin.spark.useHiveContext true
________________________________
L'intégrité de ce message n'étant pas assurée sur internet, la société
expéditrice ne peut être tenue responsable de son contenu ni de ses pièces
jointes. Toute utilisation ou diffusion non autorisée est interdite. Si vous
n'êtes pas destinataire de ce message, merci de le détruire et d'avertir
l'expéditeur.
The integrity of this message cannot be guaranteed on the Internet. The company
that sent this message cannot therefore be held liable for its content nor
attachments. Any unauthorized use or dissemination is prohibited. If you are
not the intended recipient of this message, then please delete it and notify
the sender.