On Fri, Jun 27, 2014 at 12:22 AM, Guillermo Ortiz
wrote:
> If I have to how me reducers I should have??
Depends. Best if you can have zero. Otherwise, try default partitioning
and go from there?
> as many as number of
> regions?? I have read about HRegionPartitioner, but it has some
>
If I have to how me reducers I should have?? as many as number of
regions?? I have read about HRegionPartitioner, but it has some
limitations, and you have to be sure that any region isn't going to split
while you're putting new data in your table. Is it only for performance?
what could it happ
Be sure to read http://hbase.apache.org/book.html#d3314e5975 Guillermo if
you have not already. Avoid reduce phase if you can.
St.Ack
On Thu, Jun 26, 2014 at 8:24 AM, Guillermo Ortiz
wrote:
> I have a question.
> I want to execute an MapReduce and the output of my reduce it's going to
> store
Hi Guillermo,
You can use the TableOutputFormat as the output format for your job, then on
your reduce, you just need to write Put objects.
On your driver:
Job job = new Job(conf);
…
job.setOutputFormatClass(TableOutputFormatClass);
job.setReducerClass(AverageReducer.class);
job.setOutputForma
Depending on the MapOutputValueClass, you can override corresponding
XXXSortReducer so that your custom logic is added.
Cheers
On Thu, Jun 26, 2014 at 8:24 AM, Guillermo Ortiz
wrote:
> I have a question.
> I want to execute an MapReduce and the output of my reduce it's going to
> store in HBas
I have a question.
I want to execute an MapReduce and the output of my reduce it's going to
store in HBase.
So, it's a MapReduce with an output which it's going to be stored in HBase.
I can do a Map and use HFileOutputFormat.configureIncrementalLoad(pJob,
table); but, I don't know how I could do i