Re: Hbck errors

2011-03-06 Thread Marc Limotte
ri.getStartKey()) # Start key of next matches end key of previous ... Marc On Sun, Mar 6, 2011 at 9:18 AM, Stack wrote: > So, yeah Marc, what are the rows that follow the ones you post below? > Are they the info:splitA and info:splitB or something else? > Thanks, > St.Ack > &g

Hbck errors

2011-03-05 Thread Marc Limotte
We had an issue a day ago with some OOME's on the region servers. The master shutdown ok, but most of the RegionServers didn't and so eventually had to kill -9 them. Brought it all back up and ran a major compaction to change the hbase block size. This seemed to work, but now we have an inconsis

Re: Bulk upload with multiple reducers with hbase-0.90.0

2011-01-25 Thread Marc Limotte
Ashish, I had similar experiences with our data. You do have to explicitly turn on compression for importtsv, it doesn't pick up the config for the family automatically. If you don't do that, then you have to wait for major_compaction to go through an compress everything. For inporttsv, you can

Re: hbase bulk load / table split

2011-01-05 Thread Marc Limotte
Michael, Your solution seems to work. My keys are not evenly dispersed, so I couldn't use the createTable with N regions signature, but I was able to sample one data set and come up with a half-way reasonable set of starter keys. I used those to create a new table with 20 splits. HBase appears

Re: hbase bulk load / table split

2011-01-04 Thread Marc Limotte
Thanks for the suggestion, Michael. I could give that a shot. I'm still wondering what the system is currently doing. Is it trying to split that one region? Why is it taking so long? Any way to check the progress? Marc On Tue, Jan 4, 2011 at 6:03 AM, Marc Limotte wrote: > I'

hbase bulk load / table split

2011-01-04 Thread Marc Limotte
I've made some good progress using the HBase Bulk Load Tool. With HBase 0.89.20100924+28. My initial implementation did not have importtsv do compression, and it ran directly on the hbase cluster's hadoop. It's been working ok for a while (but slow, because of limited resources). My next imple

Re: HBase Bulk Load script

2010-12-27 Thread Marc Limotte
Lars, Todd, Thanks for the info. If I understand correctly, the importtsv command line tool will not compress by default and there is no command line switch for it, but I can modify the source at hbase-0.89.20100924+28/src/main/java/org/apache/hadoop/hbase/mapreduce/ImportTsv.java to call FileOut

HBase Bulk Load script

2010-12-23 Thread Marc Limotte
Hi, I'm using the HBase Bulk Loaderwith 0.89. Very easy to use. I have a few of questions: 1) It seems importtsv will only accept one family at a time. It shows some sort of security access error if I give it a column list with columns fr