How are you adding that to the classpath? Through spark-submit or otherwise?

On Mon, Jun 22, 2015 at 5:02 PM, Murthy Chelankuri <kmurt...@gmail.com>
wrote:

> Yes I have the producer in the class path. And I am using in standalone
> mode.
>
> Sent from my iPhone
>
> On 23-Jun-2015, at 3:31 am, Tathagata Das <t...@databricks.com> wrote:
>
> Do you have Kafka producer in your classpath? If so how are adding that
> library? Are you running on YARN, or Mesos or Standalone or local. These
> details will be very useful.
>
> On Mon, Jun 22, 2015 at 8:34 AM, Murthy Chelankuri <kmurt...@gmail.com>
> wrote:
>
>> I am using spark streaming. what i am trying to do is sending few
>> messages to some kafka topic. where its failing.
>>
>> java.lang.ClassNotFoundException: com.abc.mq.msg.ObjectEncoder
>>     at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
>>     at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>>     at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)
>>     at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>>     at java.lang.Class.forName0(Native Method)
>>     at java.lang.Class.forName(Class.java:264)
>>     at kafka.utils.Utils$.createObject(Utils.scala:438)
>>     at kafka.producer.Producer.<init>(Producer.scala:61)
>>
>> On Mon, Jun 22, 2015 at 8:24 PM, Murthy Chelankuri <kmurt...@gmail.com>
>> wrote:
>>
>>> I have been using the spark from the last 6 months with the version
>>> 1.2.0.
>>>
>>> I am trying to migrate to the 1.3.0 but the same problem i have written
>>> is not wokring.
>>>
>>> Its giving class not found error when i try to load some dependent jars
>>> from the main program.
>>>
>>> This use to work in 1.2.0 when set all the dependent jars array to the
>>> spark context but not working in 1.3.0
>>>
>>>
>>> Please help me how to resolve this.
>>>
>>>
>>> Thanks,
>>> Murthy Chelankuri
>>>
>>
>>
>

Reply via email to