Re: Clarifications on HBase Durability

2014-10-08 Thread lars hofhansl
Correct on all points. I really need to pick up HBASE-5954 again. What prevented me before was that (a) there was little interest (even from myself in the end) and (b) with the variety of Hadoop versions we supported in 0.94 this was reflection hell. In HBase 1.x or 2.x we will likely only suppo

Re: NoSuchMethodError using zipkin with hbase 0.98.5

2014-10-08 Thread Ted Yu
With the following change, I was able to build from the root of htrace workspace: diff --git a/htrace-hbase/pom.xml b/htrace-hbase/pom.xml index f9f81c9..74fb4e6 100644 --- a/htrace-hbase/pom.xml +++ b/htrace-hbase/pom.xml @@ -30,7 +30,7 @@ language governing permissions and limitations under the

Re: NoSuchMethodError using zipkin with hbase 0.98.5

2014-10-08 Thread abhishek1015
I tried with 0.98.5 as you suggested, but htrace-hbase does not compile. Is it possible that it is developed for 2.0.0-SNAPSHOT version? My suspicion is based on the HttpServer2 missing error. I could not find this class in any 0.98.5 jar. ERROR] Failed to execute goal org.apache.maven.plugins:mav

Re: NoSuchMethodError using zipkin with hbase 0.98.5

2014-10-08 Thread Ted Yu
Here is the cause: Type 'org/apache/hadoop/hbase/trace/HBaseHTraceConfiguration' (current frame, stack[1]) is not assignable to 'org/htrace/HTraceConfiguration' Can you try changing the dependency in htrace-hbase/pom.xml to 0.98.5 : 2.0.0-SNAPSHOT Build hbase 0.98.5 locally by using its so

Re: QualifierFilter with Stargate

2014-10-08 Thread Nick Dimiduk
Hi Anil, Stargate has two scanner implementations: "stateful" and "stateless". Which one are you using? The stateful scanner has long supported QualifierFilter. See the ScannerModel class [0], it has a FilterModel component that you'll need to populate when you create the scanner. Stateless scann

Re: NoSuchMethodError using zipkin with hbase 0.98.5

2014-10-08 Thread abhishek1015
Thanks Ted. Earlier I was not running "mvn install -DskipTests" command under hbase workspace as I am using the 0.98.5 binaries and there is no pom.xml in that directory. So, I cloned the hbase and then run the above command in hbase directory, and then run "mvn package -DskipTests" command unde

Re: Loading hbase from parquet files

2014-10-08 Thread Nishanth S
Thank you guys for the information. -cheers Nishan On Wed, Oct 8, 2014 at 12:49 PM, Andrey Stepachev wrote: > For that use case I'd prefer to write new filtered HFiles with map reduce > and then import those data into hbase using bulk import. Keep in mind, that > incremental load tool moves fil

Re: snapshot timeouts

2014-10-08 Thread Brian Jeltema
Thanks for the quick responses. I’ll get back on this later; I discovered that HBase didn’t restart properly after changing the timeouts, so the second ERROR may be a side-effect of that. I also just discovered that the table in question was not pre-split properly, and the region distribution is

Re: snapshot timeouts

2014-10-08 Thread Brian Jeltema
Sorry, I usually include that info. HBase version is 0.98. hbase.rpc.timeout is the default. When the ‘ERROR: Call id….’ occurred, there was no stack trace. That was the entire error output. Before I increased the snapshot timeout parameters, the timeout I was seeing looked like: ERROR: org.a

Re: snapshot timeouts

2014-10-08 Thread Jean-Marc Spaggiari
Also looking for the number of regions for this table, and the number of region servers... 2014-10-08 15:18 GMT-04:00 Ted Yu : > Can you give a bit more information : > > the release of hbase you're using > value for hbase.rpc.timeout (looks like you leave it @ default) > more of the error (pleas

Re: snapshot timeouts

2014-10-08 Thread Ted Yu
Can you give a bit more information : the release of hbase you're using value for hbase.rpc.timeout (looks like you leave it @ default) more of the error (please include stack trace if possible) Cheers On Wed, Oct 8, 2014 at 12:09 PM, Brian Jeltema < brian.jelt...@digitalenvoy.net> wrote: > I’m

snapshot timeouts

2014-10-08 Thread Brian Jeltema
I’m trying to snapshot a moderately large table (3 billion rows, but not a huge amount of data per row). Those snapshots have been timing out, so I set the following parameters to relatively large values: hbase.snapshot.master.timeoutMillis hbase.snapshot.region.timeout hbase.snap

Re: Loading hbase from parquet files

2014-10-08 Thread Andrey Stepachev
For that use case I'd prefer to write new filtered HFiles with map reduce and then import those data into hbase using bulk import. Keep in mind, that incremental load tool moves files, not copies them. So once written you will not do any additional writes (except for those regions which was split w

QualifierFilter with Stargate

2014-10-08 Thread anil gupta
Hi, I need to use QualifierFilter with Stargate. I went through the documentation but it seems like all the filters are not supported. If QualifierFilter is supported then please let me know how to use it? -- Thanks & Regards, Anil Gupta

Re: Loading hbase from parquet files

2014-10-08 Thread Ted Yu
Since storage is your primary concern, take a look at Doug Meil's blog 'The Effect of ColumnFamily, RowKey and KeyValue Design on HFile Size': http://blogs.apache.org/hbase/ Cheers On Wed, Oct 8, 2014 at 9:45 AM, Nishanth S wrote: > Thanks Andrey.In the current system the hbase cfs have a ttl

Re: Loading hbase from parquet files

2014-10-08 Thread Nishanth S
Thanks Andrey.In the current system the hbase cfs have a ttl of 30 days and data gets deleted after this(has snappy compression).Below is something what I am trying to acheive. 1.Export the data from hbase table before it gets deleted. 2.Store it in some format which supports maximum compres

Re: Loading hbase from parquet files

2014-10-08 Thread Andrey Stepachev
Hi Nishanth. Not clear what exactly you are building. Can you share more detailed description of what you are building, how parquet files are supposed to be ingested. Some questions arise: 1. is that online import or bulk load 2. why rules need to be deployed to cluster. Do you suppose to do readi

Re: Loading hbase from parquet files

2014-10-08 Thread Nishanth S
I was thinking of using org.apache.hadoop.hbase.mapreduce.Driver import. I could see that we can pass in filters to this utility but looks less flexible since you need to deploy a new filter every time the rules for processing records change.Is there some way that we could define a rules engine?

Loading hbase from parquet files

2014-10-08 Thread Nishanth S
Hey folks, I am evaluating on loading an hbase table from parquet files based on some rules that would be applied on parquet file records.Could some one help me on what would be the best way to do this?. Thanks, Nishan