On Mon, Nov 4, 2013 at 8:22 PM, fateme Abiri wrote:
> hi Dhaval
> thanks for replying...
> my other project which dont use this class has been run successfuly!!!
> can you say me how can i submit the jars for hbase and hadoop with my job
> submission ?? can you give me an exapmle, tanx very much
hi Dhaval
thanks for replying...
my other project which dont use this class has been run successfuly!!!
can you say me how can i submit the jars for hbase and hadoop with my job
submission ?? can you give me an exapmle, tanx very much Dhaval,
On Monday, November 4, 2013 8:18 PM, Dhaval Shah
One more: "hbase.ipc.client.tcpnodelay" set to true. It is worth trying as well.
Best regards,
Vladimir Rodionov
Principal Platform Engineer
Carrier IQ, www.carrieriq.com
e-mail: vrodio...@carrieriq.com
From: lars hofhansl [la...@apache.org]
Sent: Monday,
Here're one more thing to try. By default each HConnection will use a single
TCP connection to multiplex traffic to each region server.
Try setting hbase.client.ipc.pool.size on the client to something > 1.
-- Lars
From: lars hofhansl
To: "user@hbase.apache
No. This is terrible.
If you can, please send a jstack and do some profiling. Is there an easy way to
reproduce this with just a single RegionServer?
If so, I'd offer to do some profiling.
Thanks.
-- Lars
From: "michael.grund...@high5games.com"
To: user@hba
setBatch and setCaching are totally independent from each other. The latter one
controls numbers of rows transferred from server to client in
one RPC call, the former one controls how many cells (key values) read from
underlying storage per one call in HBase InternalScanner implementation.
To av
No problem
Sent from Yahoo Mail on Android
Sweet! Thanks for the tip :)
On Mon, Nov 4, 2013 at 5:10 PM, Dhaval Shah wrote:
> You can use scan.setBatch() to limit the number of columns returned.. Note
> that it will split up a row into multiple rows from a client's perspective
> and client code might need to be modified to make use of the
You can use scan.setBatch() to limit the number of columns returned.. Note that
it will split up a row into multiple rows from a client's perspective and
client code might need to be modified to make use of the setBatch feature
Regards,
Dhaval
From: Patrick S
We have an application where a row can contain anywhere between 1 and
360 cells (there's only 1 column family). In practice, most rows have
under 100 cells.
Now we want to run some mapreduce jobs that touch every cell within a range
(eg count how many cells we have). With scanner caching set
I should mention that I shot an email to Cloudera today, and expect to hear
from them soon. It might be that they're the standard go-to for this sort
of thing, but I'm wondering if there are other options I should be
considering.
On Mon, Nov 4, 2013 at 4:36 PM, Patrick Schless
wrote:
> Our team'
Our team's strength's lie more around application development and devops
than jvm and hadoop tuning. We currently run two CDH4 clusters (without the
cloudera manager), and are interested in establishing a relationship with a
consultancy that can help us tune and maintain things (config, server
spec
I think, setBatch is your OOME protection as since it limits the total amount
of KeyValues
processed in a single call to StoreScanner.next.
Best regards,
Vladimir Rodionov
Principal Platform Engineer
Carrier IQ, www.carrieriq.com
e-mail: vrodio...@carrieriq.com
_
An example query would be the following, say the column qualifier was of
the form
:
where should be an integer value, and msg type is a string. E.g.
1:abc
1000:abc
2: abc
would appear in the above sequence, which is out of order when doing prefix
filtering. Zero padding could fix this:
0001:a
bq. is there an OOME protections which lowers the caching in case ther
server is stressed in memory?
I am not aware of such mechanism.
On Mon, Nov 4, 2013 at 9:55 AM, Asaf Mesika wrote:
> On this note:
> If I hammer a region server with scans that has let's say 100 rows caching
> - is there an
You might try asynchbase Michael.
St.Ack
On Mon, Nov 4, 2013 at 11:00 AM, wrote:
> Not yet, this is just a load test client. It literally does nothing but
> create threads to talk to HBase and run 4 different calls. Nothing else is
> done in the app at all.
>
> To eliminate even more of our cod
It seems that there is a serious contention point inside standard HBase client
API, that is why your application does not scale with multiple threads,
but scales with multiple JVMs. I would start with analyzing thread stack traces
of your client application - you will easily spot excessive lockin
Not yet, this is just a load test client. It literally does nothing but create
threads to talk to HBase and run 4 different calls. Nothing else is done in the
app at all.
To eliminate even more of our code from the loop, we just tried removing our
connection pool entirely and just using a sing
Michael, have you tried jstack on your client application?
Best regards,
Vladimir Rodionov
Principal Platform Engineer
Carrier IQ, www.carrieriq.com
e-mail: vrodio...@carrieriq.com
From: michael.grund...@high5games.com [michael.grund...@high5games.com]
Sen
@Anoop: hmmm, I tried it some days ago and didn't see any improvement. But
maybe I'm wrong. Can anybody confirm this?
kind regards
2013/11/4 Asaf Mesika
> On this note:
> If I hammer a region server with scans that has let's say 100 rows caching
> - is there an OOME protections which lowers th
I've gone through the code in detail - we are using unmanaged connections and
they are not being closed when the table is closed. Thanks!
From: lars hofhansl [mailto:la...@apache.org]
Sent: Monday, November 04, 2013 12:03 PM
To: Michael Grundvig; user@hbase.apache.org
Subject: Re: HBase Client Pe
HConnectionManager.createConnection is a different API creating an "unmanaged"
connection. If you're not using that each HTable.close() might close the
underlying connection.
-- Lars
From: "michael.grund...@high5games.com"
To: user@hbase.apache.org; la...@ap
On this note:
If I hammer a region server with scans that has let's say 100 rows caching
- is there an OOME protections which lowers the caching in case ther server
is stressed in memory? Or will it fail the region server with OOME?
On Monday, November 4, 2013, Anoop John wrote:
> Have u tested t
Have u tested this? AFAIK it happens the way u like.
-Anoop-
On Mon, Nov 4, 2013 at 10:49 PM, John wrote:
> Hi,
>
> If I have a question about the setCaching function. As I know the caching
> value influence how many rows will be cached in the memory of the region
> server. So for example I ha
Hi,
If I have a question about the setCaching function. As I know the caching
value influence how many rows will be cached in the memory of the region
server. So for example I have 10 rows with 20k columns and I set a scanner
with the caching option 5 and the batch option 1k. What will hapen If I
You need to add the Hadoop and HBase libraries to the Hadoop Classpath. You
successfully added it on the classpath of your mainproject but when it submits
the job to Hadoop, the classpath is lost. The easiest way is to modify
hadoop_env.sh. Another way would be to submit the jars for hbase and h
hi all
i'm running a mapreduce job in my hbase project.
my hadoop & hbase are remote and i run my code by this command in my terminal:
$ java -cp myproject.jar:/user/HadoopAndHBaseLibrary/* mainproject
but i get this error:
attempt_201207261322_0002_m_00_0, Status : FAILED
Error: java.lang.
Hi, HBase users.
I'm working with a 2 Nodes HDFS+HBASE cluster in full distributed
mode for 2 months with success now (TABLE with more than
1.000.000.000 rows) .
Everything works fine, I used also PHOENIX code from Saleforces with
success.
BUT. When
Have you taken a look at HBASE-8691 ?
Cheers
On Nov 3, 2013, at 10:33 PM, JunYoung Kim wrote:
> hi, hbase users.
>
> I am under a question to get(read) all region very quickly.
> currently, we are reading a bunch of rows from a start-key to an end-key.
>
> for putting it on the more faster
29 matches
Mail list logo