Github user AhyoungRyu commented on the issue:
https://github.com/apache/zeppelin/pull/1339
### To whom may concern about the breaking current UX with this change
This change has many benefits comparing to current embedded Spark as I
wrote in the PR description (and @tae-jun mentioned in [this
comment](https://github.com/apache/zeppelin/pull/1339#issuecomment-259486249)
as well. Thanks!).
But as always, this kind of big change brings downside as well (e.g.
breaking current UX). So I wanna write down how we can address some major cases
as below. I think it would be better to share my opinion and get more feedback
before merging. :)
1. New Spark/Zeppelin user, running Zeppelin for the first time
: Quite easy to cover and already handled by updating the related docs
pages I guess.
2. Existing Spark/Zeppelin user, running new Zeppelin installation (e.g.
upgrading version)
: Definitely this case is harder to handle than 1. As the user already
has expectation, that local mode will **just works** and surely they won't read
the docs. To resolve this, Iâll update `bin/download-spark.sh` to print sth
like âYou donât have local-spark/, you can download embedded Spark with
`get-spark` option.â When the user run `./bin/zeppelin-daemon.sh start`. And
this sentences can be removed in the future when Zeppelin users can be getting
accustomed with `get-spark` option.
3. Docker user, starting `bin/zeppelin.sh` inside the container
: This one can be also hard to handle because the user might assume that
Spark just works. So I would suggest start applying this change to #1538 as a
first step. Since it can be a Zeppelin-provided official docker script.
4. CI issue
Since @bzz raised some concern about CI issue, let me answer again in here
to make sure :)
The reason I removed `-Ppyspark` in `.travis` is `pyspark` profile is only
existed in `spark-dependencies/pom.xml`. So `pyspark` profile wonât be
anymore after this PR merged. Actually the Pyspark testcase that @astroshim
added recently had some conflict with this change. But we solved by simply
adding `export SPARK_HOME=`pwd`/spark-$SPARK_VER-bin-hadoop$HADOOP_VER` to
`.travis.yml` so that travis can run it before running the script. So there are
no more CI issues especially concerning about removing `spark-dependencies`
related build profiles.
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---