Hello. I think Wail Alkowaileet's comment is possible. Balachandar, Could you try to run your application with spark-shell?
2016-01-25 15:45 GMT+09:00 Wail Alkowaileet <wael....@gmail.com>: > I used z.load in my case and it seems to be working just fine. > Can you try spark-shell with your jar file? and see what is the error? > > I assume the problem that your application requires third-party jars. > Therefore, you need to build your app with 'assembly'. > > > On Mon, Jan 25, 2016 at 9:39 AM, Balachandar R.A. < > balachandar...@gmail.com> wrote: > >> Hello Hyung, >> >> There is nothig I could make out from error log as it is plain >> straightforward that classNotFoundException >> >> On 25 January 2016 at 11:34, Hyung Sung Shim <hss...@nflabs.com> wrote: >> >>> It's weird..so Could you send the error log for details? >>> >>> 2016-01-25 15:00 GMT+09:00 Balachandar R.A. <balachandar...@gmail.com>: >>> >>>> Hi Hyung, >>>> >>>> Thanks for the response. This I have tried but did not work. >>>> >>>> regards >>>> Bala >>>> >>>> On 25 January 2016 at 11:27, Hyung Sung Shim <hss...@nflabs.com> wrote: >>>> >>>>> Hello. Balachandar. >>>>> In case of third one that you've tried, It must be first executed in >>>>> the notebook. >>>>> Could you try restart the zeppelin and run first the "%dep z.load()" >>>>> paragraph? >>>>> >>>>> >>>>> 2016-01-25 14:39 GMT+09:00 Balachandar R.A. <balachandar...@gmail.com> >>>>> : >>>>> >>>>>> Hi >>>>>> >>>>>> Any help would be greatly appreciated :-) >>>>>> >>>>>> >>>>>> ---------- Forwarded message ---------- >>>>>> From: Balachandar R.A. <balachandar...@gmail.com> >>>>>> Date: 21 January 2016 at 14:11 >>>>>> Subject: Providing third party jar files to spark >>>>>> To: users@zeppelin.incubator.apache.org >>>>>> >>>>>> >>>>>> Hello >>>>>> >>>>>> My spark based map tasks needs to access third party jar files. I >>>>>> found below options to submit third party jar files to spark interpreter >>>>>> >>>>>> 1. export SPARK_SUBMIT_OPTIONS=<all the jar files with comma >>>>>> seprated> in conf/zeppelin-env.sh >>>>>> >>>>>> 2. include the statement spark.jars <all the jar files with comma >>>>>> separated> in <spark>?conf/spark-defaults.conf >>>>>> >>>>>> 3. use the z.load("the location of jar file in the local filesystem") >>>>>> in zepelin notebook >>>>>> >>>>>> I could test the first two and they both works fine. The third one >>>>>> does not work. Here is the snippet i use >>>>>> >>>>>> %dep >>>>>> z.reset() >>>>>> >>>>>> z.load("file:///home/bala/Projects/pocv8.new/mapreduce/build/libs/mapreduce.jar") >>>>>> >>>>>> >>>>>> Further, the import of class belongs to the above jar file is working >>>>>> when I use the statement import com..... in zeppelin notebook. However, >>>>>> I >>>>>> get the class not found exception in the executor for the same class. >>>>>> >>>>>> Any clue here would help greatly >>>>>> >>>>>> >>>>>> regards >>>>>> Bala >>>>>> >>>>>> >>>>>> >>>>>> >>>>> >>>> >>> >> > > > -- > > *Regards,* > Wail Alkowaileet >