Where are you getting serialization error. Its likely to be a different
problem. Which class is not getting serialized?


Mayur Rustagi
Ph: +1 (760) 203 3257
http://www.sigmoidanalytics.com
@mayur_rustagi <https://twitter.com/mayur_rustagi>



On Thu, Jun 5, 2014 at 6:32 PM, Vibhor Banga <vibhorba...@gmail.com> wrote:

> Any inputs on this will be helpful.
>
> Thanks,
> -Vibhor
>
>
> On Thu, Jun 5, 2014 at 3:41 PM, Vibhor Banga <vibhorba...@gmail.com>
> wrote:
>
>> Hi,
>>
>> I am trying to do something like following in Spark:
>>
>> JavaPairRDD<byte[], MyObject> eventRDD = hBaseRDD.map(new
>> PairFunction<Tuple2<ImmutableBytesWritable, Result>, byte[], MyObject >() {
>>             @Override
>>             public Tuple2<byte[], MyObject >
>> call(Tuple2<ImmutableBytesWritable, Result>
>> immutableBytesWritableResultTuple2) throws Exception {
>>                 return new
>> Tuple2<byte[], MyObject >(immutableBytesWritableResultTuple2._1.get(),
>> MyClass.get(immutableBytesWritableResultTuple2._2));
>>             }
>>         });
>>
>>         eventRDD.foreach(new VoidFunction<Tuple2<byte[], Event>>() {
>>             @Override
>>             public void call(Tuple2<byte[], Event> eventTuple2) throws
>> Exception {
>>
>>                 processForEvent(eventTuple2._2);
>>             }
>>         });
>>
>>
>> processForEvent() function flow contains some processing and ultimately
>> writing to HBase Table. But I am getting serialisation issues with Hadoop
>> and HBase inbuilt classes. How do I solve this ? Does using Kyro
>> Serialisation help in this case ?
>>
>> Thanks,
>> -Vibhor
>>
>
>
>
> --
> Vibhor Banga
> Software Development Engineer
> Flipkart Internet Pvt. Ltd., Bangalore
>
>

Reply via email to