Thanks for all respnding.
Finally I figured out the way to use bulk load to hbase using scala on
spark.
The sample code is here which others can refer in future:
http://www.openkb.info/2015/01/how-to-use-scala-on-spark-to-load-data.html
Thanks!
On Tue, Jan 27, 2015 at 6:27 PM, Jim Green wrote:
Thanks Sun.
My understanding is , savaAsNewHadoopFile is to save as Hfile on hdfs.
Is it doable to use saveAsNewAPIHadoopDataset to directly loading to hbase?
If so, is there any sample code for that?
Thanks!
On Tue, Jan 27, 2015 at 6:07 PM, fightf...@163.com
wrote:
> Hi, Jim
> Your generated
Hi, Jim
Your generated rdd should be the type of RDD[ImmutableBytesWritable, KeyValue],
while your current type goes to RDD[ImmutableBytesWritable, Put].
You can go like this and the result should be type of
RDD[ImmutableBytesWritable, KeyValue] that can be savaAsNewHadoopFile
val result = n