Can you show us the code for loading Hive into hbase ? There shouldn't be 'return' statement in that code.
Cheers > On Jun 20, 2015, at 10:10 PM, Nishant Patel <nishant.k.pa...@gmail.com> wrote: > > Hi, > > I am loading data from Hive table to Hbase after doing some manipulation. > > I am getting error as 'Task not Serializable'. > > My code is as below. > > public class HiveToHbaseLoader implements Serializable { > > public static void main(String[] args) throws Exception { > > String hbaseTableName = args[0]; > String hiveQuery = args[1]; > > SparkConf conf = new SparkConf().setAppName("Hive to Hbase Loader") > .setMaster("????"); > JavaSparkContext sc = new JavaSparkContext(conf); > > HiveContext hiveContext = new HiveContext(sc.sc()); > > hiveContext.setConf("hive.metastore.uris", > "?????"); > > DataFrame dataFrame = hiveContext.sql(hiveQuery); > > dataFrame > .foreachPartition(new > AbstractFunction1<scala.collection.Iterator<Row>, BoxedUnit>() { > > //Logic to load row from hive to Hbase. > > } > }} > > > Getting error as below. > > > Exception in thread "main" org.apache.spark.SparkException: Task not > serializable > at > org.apache.spark.util.ClosureCleaner$.ensureSerializable(ClosureCleaner.scala:166) > at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:158) > at org.apache.spark.SparkContext.clean(SparkContext.scala:1623) > at org.apache.spark.rdd.RDD.foreachPartition(RDD.scala:805) > at org.apache.spark.sql.DataFrame.foreachPartition(DataFrame.scala:875) > at com.philips.bda.HiveToHbaseLoader.main(HiveToHbaseLoader.java:46) > Caused by: java.io.NotSerializableException: > com.philips.bda.HiveToHbaseLoader$1 > Serialization stack: > - object not serializable (class: com.philips.bda.HiveToHbaseLoader$1, > value: <function1>) > at > org.apache.spark.serializer.SerializationDebugger$.improveException(SerializationDebugger.scala:38) > at > org.apache.spark.serializer.JavaSerializationStream.writeObject(JavaSerializer.scala:47) > at > org.apache.spark.serializer.JavaSerializerInstance.serialize(JavaSerializer.scala:80) > at > org.apache.spark.util.ClosureCleaner$.ensureSerializable(ClosureCleaner.scala:164) > ... 5 more > > > -- > Regards, > Nishant >