lol - i just ran a rowcount via mapreduce, and it took 6 hours for 7.5
million rows...
On Sun, Oct 9, 2011 at 7:50 AM, Rita wrote:
> Hi,
>
> I have been doing a rowcount via mapreduce and its taking about 4-5 hours
> to
> count a 500million rows in a table. I was wondering if there are any map
>
Following a lot of research, and some great input from J-D, I figured my
recent endeavors might prove useful for the rest of the list.
http://tech.tomgoren.com/archives/284
Basically it's hbase export -> hdfs -> distcp via hftp to new cluster ->
hbase import.
Any it can be divided by row time st
OK got it to work.
It seems that the problem was in a non-related network issue.
The CopyTable ran with the syntax I specified earlier.
Thanks!
Tom
On Thu, Sep 8, 2011 at 5:14 AM, Tom Goren wrote:
> J-D, thanks a lot.
> However I was not able to understand the correct usage fro
og/2010/04/10/how-to-export-and-import-an-hbase-table/
more
or less)
Surely there is something better?
Thanks in advance, your help is greatly appreciated!
Tom
On Wed, Sep 7, 2011 at 8:07 PM, Jean-Daniel Cryans wrote:
> Inline.
>
> J-D
>
> On Wed, Sep 7, 2011 at 8:02 PM, Tom
additional feedback.
Thanks!
Tom
On Wed, Sep 7, 2011 at 6:03 PM, Doug Meil wrote:
>
> Hi there-
>
> Have you tried this?
>
> http://hbase.apache.org/book.html#copytable
>
> It's the Java invocation of the copy-table function (without the ruby)
>
>
>
> On
So I have read http://blog.sematext.com/2011/03/11/hbase-backup-options/
The built-in solution, namely the 'Export' and 'Import' classes are
relatively straight forward to use, however, the exports include the data
alone, and nothing in regards to the table description (column families,
versioning