RE: Best practices for loading data into hbase

2013-06-05 Thread David Poisson
nnect from my vm to my vm. I'm at a lost here. Could really use some guidance. Thanks! David From: David Poisson [david.pois...@ca.fujitsu.com] Sent: Friday, May 31, 2013 4:19 PM To: user@hbase.apache.org Subject: Best practices for loading data into

Re: Best practices for loading data into hbase

2013-05-31 Thread Mohammad Tariq
I am sorry to barge in when heavyweights are already involved here. But, just out of curiosity, why don't you use Sqoop to import the data directly from your existing systems into HBase instead of first taking the dump and then doing the import. Sqoop allows us to do incr

Re: Best practices for loading data into hbase

2013-05-31 Thread Ted Yu
bq. Once we process an xml file and we populate our 3 "production" hbase tables, could we bulk load another xml file and append this new data to our 3 tables or would it write over what was written before? You can bulk load another XML file. bq. should we process our input xml file with 3 MapRedu

Re: Best practices for loading data into hbase

2013-05-31 Thread Jean-Daniel Cryans
You cannot use the local job tracker (that is, the one that gets started if you don't have one running) with the TotalOrderPartitioner. You'll need to fully install hadoop on that vmware node. Google that error to find other relevant comments. J-D On Fri, May 31, 2013 at 1:19 PM, David Poisson

Best practices for loading data into hbase

2013-05-31 Thread David Poisson
Hi, We are still very new at all of this hbase/hadoop/mapreduce stuff. We are looking for the best practices that will fit our requirements. We are currently using the latest cloudera vmware's (single node) for our development tests. The problem is as follows: We have multiple sources in