Basically Zeppelin HA require 2 major things:
1. Shared storage (storage for notebook, interpreter setting,
zeppelin-site.xml, zeppelin-env.sh, shiro.ini, credentials.json)
2. Recover running interpreter.The standby zeppelin instance don't know
where are the running interpreter (host:port), so it
I think if you have a shared storage for notebooks (for example, NFS
mounted from a third server),
and a load-balancer that supports sticky sessions (like F5) on top, it
should be possible to have HA without
any code change in Zeppelin. Am I missing something?
--
Ruslan Dautkhanov
On Fri, Jun
Honestly, HA requires more than just active stand by.
It should be able to scale without major surgeries, which is not possible
right now. For example, if you start too many interpreters, zeppelin box
will simply run out of memory.
Alex
On Thu, Jun 29, 2017 at 10:59 PM, wenxing zheng
wrote:
>
HADOOP_CONF_DIR in zeppelin-env.sh would affect the whole zeppelin
instance, and define it in interpreter setting would affect that
interpreter.
Jeff Zhang 于2017年7月1日周六 上午7:26写道:
>
> HADOOP_CONF_DIR would affect the whole zeppelin instance. and define it
> interpreter setting would affect that in
HADOOP_CONF_DIR would affect the whole zeppelin instance. and define it
interpreter setting would affect that interpreter.
And all the capitalized property name would be taken as env variable.
Serega Sheypak 于2017年7月1日周六 上午3:20写道:
> hi, thanks for your reply. How should I set this variable?
> I'
hi, thanks for your reply. How should I set this variable?
I'm looking at Spark interpreter config UI. It doesn't allow me to set env
variable.
https://zeppelin.apache.org/docs/latest/interpreter/spark.html#1-export-spark_home
tells that HADOOP_CONF_DIR should be set once per whole Zeppelin instanc
Right, create three spark interpreters for your 3 yarn cluster.
Serega Sheypak 于2017年6月30日周五 下午10:33写道:
> Hi, thanks for your reply!
> What do you mean by that?
> I can have only one env variable HADOOP_CONF_DIR...
> And how can user pick which env to run?
>
> Or you mean I have to create three
Hi, thanks for your reply!
What do you mean by that?
I can have only one env variable HADOOP_CONF_DIR...
And how can user pick which env to run?
Or you mean I have to create three Spark interpreters and each of them
would have it's own HADOOP_CONF_DIR pointed to single cluster config?
2017-06-30
Try set HADOOP_CONF_DIR for each yarn conf in interpreter setting.
Serega Sheypak 于2017年6月30日周五 下午10:11写道:
> Hi I have several different hadoop clusters, each of them has it's own
> YARN.
> Is it possible to configure single Zeppelin instance to work with
> different clusters?
> I want to run spa
Hi I have several different hadoop clusters, each of them has it's own
YARN.
Is it possible to configure single Zeppelin instance to work with different
clusters?
I want to run spark on cluster A if data is there. Right now my Zeppelin
runs on single cluster and it sucks data from remote clusters w
Great !
It's available on Zeppelin 0.8 though. Let's upgrade...
Le 30 juin 2017 12:19 AM, "Jeff Zhang" a écrit :
>
> set zeppelin.spark.enableSupportedVersionCheck as false in this custom
> spark's interpreter setting.
>
>
> Jun Kim 于2017年6月30日周五 上午1:04写道:
>
>> Here is a patch that I used for Sp
11 matches
Mail list logo