Hello.
I think Wail Alkowaileet's comment is possible.
Balachandar, Could you try to run your application with spark-shell?


2016-01-25 15:45 GMT+09:00 Wail Alkowaileet <wael....@gmail.com>:

> I used z.load in my case and it seems to be working just fine.
> Can you try spark-shell with your jar file? and see what is the error?
>
> I assume the problem that your application requires third-party jars.
> Therefore, you need to build your app with 'assembly'.
>
>
> On Mon, Jan 25, 2016 at 9:39 AM, Balachandar R.A. <
> balachandar...@gmail.com> wrote:
>
>> Hello Hyung,
>>
>> There is nothig I could make out from error log as it is plain
>> straightforward that classNotFoundException
>>
>> On 25 January 2016 at 11:34, Hyung Sung Shim <hss...@nflabs.com> wrote:
>>
>>> It's weird..so Could you send the error log for details?
>>>
>>> 2016-01-25 15:00 GMT+09:00 Balachandar R.A. <balachandar...@gmail.com>:
>>>
>>>> Hi Hyung,
>>>>
>>>> Thanks for the response. This I have tried but did not work.
>>>>
>>>> regards
>>>> Bala
>>>>
>>>> On 25 January 2016 at 11:27, Hyung Sung Shim <hss...@nflabs.com> wrote:
>>>>
>>>>> Hello. Balachandar.
>>>>> In case of third one that you've tried, It must be first executed in
>>>>> the notebook.
>>>>> Could you try restart the zeppelin and run first the "%dep z.load()"
>>>>> paragraph?
>>>>>
>>>>>
>>>>> 2016-01-25 14:39 GMT+09:00 Balachandar R.A. <balachandar...@gmail.com>
>>>>> :
>>>>>
>>>>>> Hi
>>>>>>
>>>>>> Any help would be greatly appreciated :-)
>>>>>>
>>>>>>
>>>>>> ---------- Forwarded message ----------
>>>>>> From: Balachandar R.A. <balachandar...@gmail.com>
>>>>>> Date: 21 January 2016 at 14:11
>>>>>> Subject: Providing third party jar files to spark
>>>>>> To: users@zeppelin.incubator.apache.org
>>>>>>
>>>>>>
>>>>>> Hello
>>>>>>
>>>>>> My spark based map tasks needs to access third party jar files. I
>>>>>> found below options to submit third party jar files to spark interpreter
>>>>>>
>>>>>> 1. export SPARK_SUBMIT_OPTIONS=<all the jar files with comma
>>>>>> seprated> in conf/zeppelin-env.sh
>>>>>>
>>>>>> 2. include the statement spark.jars  <all the jar files with comma
>>>>>> separated> in <spark>?conf/spark-defaults.conf
>>>>>>
>>>>>> 3. use the z.load("the location of jar file in the local filesystem")
>>>>>> in zepelin notebook
>>>>>>
>>>>>> I could test the first two and they both works fine. The third one
>>>>>> does not work. Here is the snippet i use
>>>>>>
>>>>>> %dep
>>>>>> z.reset()
>>>>>>
>>>>>> z.load("file:///home/bala/Projects/pocv8.new/mapreduce/build/libs/mapreduce.jar")
>>>>>>
>>>>>>
>>>>>> Further, the import of class belongs to the above jar file is working
>>>>>> when I use the statement import com.....  in zeppelin notebook. However, 
>>>>>> I
>>>>>> get the class not found exception in the executor for the same class.
>>>>>>
>>>>>> Any clue here would help greatly
>>>>>>
>>>>>>
>>>>>> regards
>>>>>> Bala
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
>
> --
>
> *Regards,*
> Wail Alkowaileet
>

Reply via email to