HBase setup problem mvn eclipse:eclipse

2011-09-03 Thread vamshi krishna
Hi all, i am new to hbase world.i am trying to setup hbase on my machine. I am using ubuntu 10.04 machine. eclipse helios IDE.. i am trying to setup , modify hbase source code and rebuild it. For that i am following http://hbase.apache.org/book.html for setting up. when i execute mvn eclipse:eclip

Re: Hbase bulk insert.

2011-09-03 Thread Dhaval Makawana
Hi Sriram, You can use bulk upload utility to insert massive data in HBase. http://archive.cloudera.com/cdh/3/hbase/bulk-loads.html There is an alternative I use in my application where I had to insert large amount of data from map-reduce job. I use async hbase client( https://github.com/stumbleu

Re: HBase Meetup during Hadoop World NYC '11

2011-09-03 Thread Otis Gospodnetic
Todd, +1 Otis > >From: Todd Lipcon >To: user@hbase.apache.org >Sent: Friday, August 26, 2011 6:33 PM >Subject: HBase Meetup during Hadoop World NYC '11 > >Dear HBase users, > >I think this was brought up previously on the list, but I couldn't dig >up the thread,

Re: prevent region splits?

2011-09-03 Thread Jack Levin
Make hbase.hregion.max.filesize to be very large. Then your regions won't split. We use this method when copying 'live' hbase to make a backup. -Jack On Sat, Sep 3, 2011 at 4:32 PM, Geoff Hendrey wrote: > Is there a way to prevent regions from splitting while we are running a > mapreduce job th

prevent region splits?

2011-09-03 Thread Geoff Hendrey
Is there a way to prevent regions from splitting while we are running a mapreduce job that does a lot of Puts? It seems that there is a lot of HDFS activity related to the splitting of regions while my M/R job is doing the puts. Is it sensible to disable splitting during the job that does lots of P

RE: PENDING_CLOSE for too long

2011-09-03 Thread Geoff Hendrey
"Are you having trouble getting to any of your data out in tables?" depends what you mean. We see corruptions from time to time that prevent us from getting data, one way or another. Today's corruption was regions with duplicate start and end rows. We fixed that by deleting the offending regions f