Re: Query about the high availability of Zeppelin

2017-06-30 Thread Jeff Zhang
Basically Zeppelin HA require 2 major things: 1. Shared storage (storage for notebook, interpreter setting, zeppelin-site.xml, zeppelin-env.sh, shiro.ini, credentials.json) 2. Recover running interpreter.The standby zeppelin instance don't know where are the running interpreter (host:port), so it

Re: Query about the high availability of Zeppelin

2017-06-30 Thread Ruslan Dautkhanov
I think if you have a shared storage for notebooks (for example, NFS mounted from a third server), and a load-balancer that supports sticky sessions (like F5) on top, it should be possible to have HA without any code change in Zeppelin. Am I missing something? -- Ruslan Dautkhanov On Fri, Jun

Re: Query about the high availability of Zeppelin

2017-06-30 Thread Alexander Filipchik
Honestly, HA requires more than just active stand by. It should be able to scale without major surgeries, which is not possible right now. For example, if you start too many interpreters, zeppelin box will simply run out of memory. Alex On Thu, Jun 29, 2017 at 10:59 PM, wenxing zheng wrote: >

Re: Configuring Zeppelin spark interpreter to work with different hadoop clusters

2017-06-30 Thread Jeff Zhang
HADOOP_CONF_DIR in zeppelin-env.sh would affect the whole zeppelin instance, and define it in interpreter setting would affect that interpreter. Jeff Zhang 于2017年7月1日周六 上午7:26写道: > > HADOOP_CONF_DIR would affect the whole zeppelin instance. and define it > interpreter setting would affect that in

Re: Configuring Zeppelin spark interpreter to work with different hadoop clusters

2017-06-30 Thread Jeff Zhang
HADOOP_CONF_DIR would affect the whole zeppelin instance. and define it interpreter setting would affect that interpreter. And all the capitalized property name would be taken as env variable. Serega Sheypak 于2017年7月1日周六 上午3:20写道: > hi, thanks for your reply. How should I set this variable? > I'

Re: Configuring Zeppelin spark interpreter to work with different hadoop clusters

2017-06-30 Thread Serega Sheypak
hi, thanks for your reply. How should I set this variable? I'm looking at Spark interpreter config UI. It doesn't allow me to set env variable. https://zeppelin.apache.org/docs/latest/interpreter/spark.html#1-export-spark_home tells that HADOOP_CONF_DIR should be set once per whole Zeppelin instanc

Re: Configuring Zeppelin spark interpreter to work with different hadoop clusters

2017-06-30 Thread Jeff Zhang
Right, create three spark interpreters for your 3 yarn cluster. Serega Sheypak 于2017年6月30日周五 下午10:33写道: > Hi, thanks for your reply! > What do you mean by that? > I can have only one env variable HADOOP_CONF_DIR... > And how can user pick which env to run? > > Or you mean I have to create three

Re: Configuring Zeppelin spark interpreter to work with different hadoop clusters

2017-06-30 Thread Serega Sheypak
Hi, thanks for your reply! What do you mean by that? I can have only one env variable HADOOP_CONF_DIR... And how can user pick which env to run? Or you mean I have to create three Spark interpreters and each of them would have it's own HADOOP_CONF_DIR pointed to single cluster config? 2017-06-30

Re: Configuring Zeppelin spark interpreter to work with different hadoop clusters

2017-06-30 Thread Jeff Zhang
Try set HADOOP_CONF_DIR for each yarn conf in interpreter setting. Serega Sheypak 于2017年6月30日周五 下午10:11写道: > Hi I have several different hadoop clusters, each of them has it's own > YARN. > Is it possible to configure single Zeppelin instance to work with > different clusters? > I want to run spa

Configuring Zeppelin spark interpreter to work with different hadoop clusters

2017-06-30 Thread Serega Sheypak
Hi I have several different hadoop clusters, each of them has it's own YARN. Is it possible to configure single Zeppelin instance to work with different clusters? I want to run spark on cluster A if data is there. Right now my Zeppelin runs on single cluster and it sucks data from remote clusters w

Re: custom spark (fresh !)

2017-06-30 Thread Mathieu D
Great ! It's available on Zeppelin 0.8 though. Let's upgrade... Le 30 juin 2017 12:19 AM, "Jeff Zhang" a écrit : > > set zeppelin.spark.enableSupportedVersionCheck as false in this custom > spark's interpreter setting. > > > Jun Kim 于2017年6月30日周五 上午1:04写道: > >> Here is a patch that I used for Sp