Thanks Jeff.

We will add dependencies though livy.spark.jars.packages


Thanks
Anand



On Wed, Nov 22, 2017 at 4:29 PM, Jeff Zhang <zjf...@gmail.com> wrote:

>
> livy doesn't support adding dependency via in note like %spark.dep, you
> have to do it in interpreter setting.
>
>
> Anandha L Ranganathan <analog.s...@gmail.com>于2017年11月23日周四 上午4:37写道:
>
>> Thanks Jeff.
>>
>> Is that something I can use it in the notebook or in the interpreter? If
>> it is in the notebook can you provide me with syntax ? I tried in the
>> notebook and it is throwing an error.
>>
>>
>>
>>
>> On Tue, Nov 21, 2017 at 5:28 PM, Jeff Zhang <zjf...@gmail.com> wrote:
>>
>>>
>>> You can do it via livy interpreter setting.
>>>
>>> Here's 2 configuration which can help you add external jars and external
>>> packages
>>>
>>> livy.spark.jars
>>> livy.spark.jars.packages
>>>
>>> And this is the configuration for queue name
>>>
>>> livy.spark.yarn.queue
>>>
>>>
>>> Anandha L Ranganathan <analog.s...@gmail.com>于2017年11月22日周三 上午9:13写道:
>>>
>>>> We are using Livy interpreter from Zeppelin to connect to Spark.
>>>>
>>>> In this,  we want to give the users an option to download the external
>>>> libraries.
>>>> By default we have added some basic libraries in interpreter setting.
>>>>
>>>> In spark interpreter, an users can download the external libraries they
>>>> want using this command.
>>>> %spark.dep
>>>> z.reset()
>>>> z.addRepo("Spark Packages Repo").url("http://dl.bintray.
>>>> com/spark-packages/maven")
>>>> z.load("com.databricks:spark-csv_2.11:1.2.0")
>>>>
>>>>
>>>> How can we import the external libraries using livy ?
>>>>
>>>>
>>>> Another question, is there a way to change the yarn queue name at
>>>> runtime? Some users want to use different queue rather than default queue
>>>> assigned in the interpreter.  If that feature is not available, then what
>>>> is the best approach to implement this ?
>>>>
>>>> Thanks
>>>> Anand
>>>>
>>>>
>>>>
>>>
>>

Reply via email to