Correct on all points.
I really need to pick up HBASE-5954 again. What prevented me before was that
(a) there was little interest (even from myself in the end) and (b) with the
variety of Hadoop versions we supported in 0.94 this was reflection hell. In
HBase 1.x or 2.x we will likely only suppo
With the following change, I was able to build from the root of htrace
workspace:
diff --git a/htrace-hbase/pom.xml b/htrace-hbase/pom.xml
index f9f81c9..74fb4e6 100644
--- a/htrace-hbase/pom.xml
+++ b/htrace-hbase/pom.xml
@@ -30,7 +30,7 @@ language governing permissions and limitations under the
I tried with 0.98.5 as you suggested, but htrace-hbase does not compile. Is
it possible that it is developed for 2.0.0-SNAPSHOT version? My suspicion is
based on the HttpServer2 missing error. I could not find this class in any
0.98.5 jar.
ERROR] Failed to execute goal
org.apache.maven.plugins:mav
Here is the cause:
Type 'org/apache/hadoop/hbase/trace/HBaseHTraceConfiguration' (current
frame, stack[1]) is not assignable to 'org/htrace/HTraceConfiguration'
Can you try changing the dependency in htrace-hbase/pom.xml to 0.98.5 :
2.0.0-SNAPSHOT
Build hbase 0.98.5 locally by using its so
Hi Anil,
Stargate has two scanner implementations: "stateful" and "stateless". Which
one are you using?
The stateful scanner has long supported QualifierFilter. See the
ScannerModel class [0], it has a FilterModel component that you'll need to
populate when you create the scanner. Stateless scann
Thanks Ted.
Earlier I was not running "mvn install -DskipTests" command under hbase
workspace as I am using the 0.98.5 binaries and there is no pom.xml in that
directory.
So, I cloned the hbase and then run the above command in hbase directory,
and then run "mvn package -DskipTests" command unde
Thank you guys for the information.
-cheers
Nishan
On Wed, Oct 8, 2014 at 12:49 PM, Andrey Stepachev wrote:
> For that use case I'd prefer to write new filtered HFiles with map reduce
> and then import those data into hbase using bulk import. Keep in mind, that
> incremental load tool moves fil
Thanks for the quick responses. I’ll get back on this later; I discovered that
HBase didn’t restart properly
after changing the timeouts, so the second ERROR may be a side-effect of that.
I also just discovered that the table in question was not pre-split properly,
and the region distribution
is
Sorry, I usually include that info. HBase version is 0.98. hbase.rpc.timeout is
the default.
When the ‘ERROR: Call id….’ occurred, there was no stack trace. That was the
entire error output.
Before I increased the snapshot timeout parameters, the timeout I was seeing
looked like:
ERROR: org.a
Also looking for the number of regions for this table, and the number of
region servers...
2014-10-08 15:18 GMT-04:00 Ted Yu :
> Can you give a bit more information :
>
> the release of hbase you're using
> value for hbase.rpc.timeout (looks like you leave it @ default)
> more of the error (pleas
Can you give a bit more information :
the release of hbase you're using
value for hbase.rpc.timeout (looks like you leave it @ default)
more of the error (please include stack trace if possible)
Cheers
On Wed, Oct 8, 2014 at 12:09 PM, Brian Jeltema <
brian.jelt...@digitalenvoy.net> wrote:
> I’m
I’m trying to snapshot a moderately large table (3 billion rows, but not a huge
amount of data per row).
Those snapshots have been timing out, so I set the following parameters to
relatively large values:
hbase.snapshot.master.timeoutMillis
hbase.snapshot.region.timeout
hbase.snap
For that use case I'd prefer to write new filtered HFiles with map reduce
and then import those data into hbase using bulk import. Keep in mind, that
incremental load tool moves files, not copies them. So once written you
will not do any additional writes (except for those regions which was split
w
Hi,
I need to use QualifierFilter with Stargate. I went through the
documentation but it seems like all the filters are not supported. If
QualifierFilter is supported then please let me know how to use it?
--
Thanks & Regards,
Anil Gupta
Since storage is your primary concern, take a look at Doug Meil's blog 'The
Effect of ColumnFamily, RowKey and KeyValue Design on HFile Size':
http://blogs.apache.org/hbase/
Cheers
On Wed, Oct 8, 2014 at 9:45 AM, Nishanth S wrote:
> Thanks Andrey.In the current system the hbase cfs have a ttl
Thanks Andrey.In the current system the hbase cfs have a ttl of 30 days
and data gets deleted after this(has snappy compression).Below is something
what I am trying to acheive.
1.Export the data from hbase table before it gets deleted.
2.Store it in some format which supports maximum compres
Hi Nishanth.
Not clear what exactly you are building.
Can you share more detailed description of what you are building, how
parquet files are supposed to be ingested.
Some questions arise:
1. is that online import or bulk load
2. why rules need to be deployed to cluster. Do you suppose to do readi
I was thinking of using org.apache.hadoop.hbase.mapreduce.Driver import. I
could see that we can pass in filters to this utility but looks less
flexible since you need to deploy a new filter every time the rules for
processing records change.Is there some way that we could define a rules
engine?
Hey folks,
I am evaluating on loading an hbase table from parquet files based on
some rules that would be applied on parquet file records.Could some one
help me on what would be the best way to do this?.
Thanks,
Nishan
19 matches
Mail list logo