Hello, I have run the code in spark-shell successfully but the jar files were all specified in the config files (spark-defaults.conf). However, I will not be able to use z.load() in spark-shell. Isn't? I am sorry but I did not pick up the idea of running using spark-shell. Wail suggestion is to create a fatJar? I will give it a try but still how do i make sure this fatJar is accessible to spark executors? ANyway, I will keep you posted on this
regards Bala On 25 January 2016 at 13:39, Hyung Sung Shim <hss...@nflabs.com> wrote: > Hello. > I think Wail Alkowaileet's comment is possible. > Balachandar, Could you try to run your application with spark-shell? > > > 2016-01-25 15:45 GMT+09:00 Wail Alkowaileet <wael....@gmail.com>: > >> I used z.load in my case and it seems to be working just fine. >> Can you try spark-shell with your jar file? and see what is the error? >> >> I assume the problem that your application requires third-party jars. >> Therefore, you need to build your app with 'assembly'. >> >> >> On Mon, Jan 25, 2016 at 9:39 AM, Balachandar R.A. < >> balachandar...@gmail.com> wrote: >> >>> Hello Hyung, >>> >>> There is nothig I could make out from error log as it is plain >>> straightforward that classNotFoundException >>> >>> On 25 January 2016 at 11:34, Hyung Sung Shim <hss...@nflabs.com> wrote: >>> >>>> It's weird..so Could you send the error log for details? >>>> >>>> 2016-01-25 15:00 GMT+09:00 Balachandar R.A. <balachandar...@gmail.com>: >>>> >>>>> Hi Hyung, >>>>> >>>>> Thanks for the response. This I have tried but did not work. >>>>> >>>>> regards >>>>> Bala >>>>> >>>>> On 25 January 2016 at 11:27, Hyung Sung Shim <hss...@nflabs.com> >>>>> wrote: >>>>> >>>>>> Hello. Balachandar. >>>>>> In case of third one that you've tried, It must be first executed in >>>>>> the notebook. >>>>>> Could you try restart the zeppelin and run first the "%dep z.load()" >>>>>> paragraph? >>>>>> >>>>>> >>>>>> 2016-01-25 14:39 GMT+09:00 Balachandar R.A. <balachandar...@gmail.com >>>>>> >: >>>>>> >>>>>>> Hi >>>>>>> >>>>>>> Any help would be greatly appreciated :-) >>>>>>> >>>>>>> >>>>>>> ---------- Forwarded message ---------- >>>>>>> From: Balachandar R.A. <balachandar...@gmail.com> >>>>>>> Date: 21 January 2016 at 14:11 >>>>>>> Subject: Providing third party jar files to spark >>>>>>> To: users@zeppelin.incubator.apache.org >>>>>>> >>>>>>> >>>>>>> Hello >>>>>>> >>>>>>> My spark based map tasks needs to access third party jar files. I >>>>>>> found below options to submit third party jar files to spark interpreter >>>>>>> >>>>>>> 1. export SPARK_SUBMIT_OPTIONS=<all the jar files with comma >>>>>>> seprated> in conf/zeppelin-env.sh >>>>>>> >>>>>>> 2. include the statement spark.jars <all the jar files with comma >>>>>>> separated> in <spark>?conf/spark-defaults.conf >>>>>>> >>>>>>> 3. use the z.load("the location of jar file in the local >>>>>>> filesystem") in zepelin notebook >>>>>>> >>>>>>> I could test the first two and they both works fine. The third one >>>>>>> does not work. Here is the snippet i use >>>>>>> >>>>>>> %dep >>>>>>> z.reset() >>>>>>> >>>>>>> z.load("file:///home/bala/Projects/pocv8.new/mapreduce/build/libs/mapreduce.jar") >>>>>>> >>>>>>> >>>>>>> Further, the import of class belongs to the above jar file is >>>>>>> working when I use the statement import com..... in zeppelin notebook. >>>>>>> However, I get the class not found exception in the executor for the >>>>>>> same >>>>>>> class. >>>>>>> >>>>>>> Any clue here would help greatly >>>>>>> >>>>>>> >>>>>>> regards >>>>>>> Bala >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >> >> -- >> >> *Regards,* >> Wail Alkowaileet >> > >