livy doesn't support adding dependency via in note like %spark.dep, you
have to do it in interpreter setting.


Anandha L Ranganathan <analog.s...@gmail.com>于2017年11月23日周四 上午4:37写道:

> Thanks Jeff.
>
> Is that something I can use it in the notebook or in the interpreter? If
> it is in the notebook can you provide me with syntax ? I tried in the
> notebook and it is throwing an error.
>
>
>
>
> On Tue, Nov 21, 2017 at 5:28 PM, Jeff Zhang <zjf...@gmail.com> wrote:
>
>>
>> You can do it via livy interpreter setting.
>>
>> Here's 2 configuration which can help you add external jars and external
>> packages
>>
>> livy.spark.jars
>> livy.spark.jars.packages
>>
>> And this is the configuration for queue name
>>
>> livy.spark.yarn.queue
>>
>>
>> Anandha L Ranganathan <analog.s...@gmail.com>于2017年11月22日周三 上午9:13写道:
>>
>>> We are using Livy interpreter from Zeppelin to connect to Spark.
>>>
>>> In this,  we want to give the users an option to download the external
>>> libraries.
>>> By default we have added some basic libraries in interpreter setting.
>>>
>>> In spark interpreter, an users can download the external libraries they
>>> want using this command.
>>> %spark.dep
>>> z.reset()
>>> z.addRepo("Spark Packages Repo").url("
>>> http://dl.bintray.com/spark-packages/maven";)
>>> z.load("com.databricks:spark-csv_2.11:1.2.0")
>>>
>>>
>>> How can we import the external libraries using livy ?
>>>
>>>
>>> Another question, is there a way to change the yarn queue name at
>>> runtime? Some users want to use different queue rather than default queue
>>> assigned in the interpreter.  If that feature is not available, then what
>>> is the best approach to implement this ?
>>>
>>> Thanks
>>> Anand
>>>
>>>
>>>
>>
>

Reply via email to