Could it be data dependent?  For example, if you have a lot of data that has 
passed its TTL you may be scanning across a lot of data to find data that is 
eligible to return.  Similar situations could have to do with visibilities that 
you can’t access…  Or, maybe it’s related to your scan range?  You are scanning 
across a lot of data, but most of the rows do not match your scan criteria?

If you think it could be related to age-off rather than visibilities or scan 
range, can you run a full compaction on the table and see if that improves 
things?  That would eliminate data that has aged off and reduce the amount of 
data that must be scanned.  If you can, use hdfs to determine the directory 
size of the table – it would be under /accumulo/tables/[TABLE-ID] then run the 
compaction (compact -w -t tablename) and when it finishes and the accumulo gc 
runs to remove the “old” files and check the size again.  That should give you 
an idea of how much data was removed by the compaction.

Ed Coleman

From: Christopher <ctubb...@apache.org>
Sent: Monday, October 31, 2022 5:42 PM
To: accumulo-user <user@accumulo.apache.org>
Subject: Re: very-high latency updates through thrift proxy server

That's odd. They should be available immediately. Are you using replication? 
What kind of scan are you doing? Is it an offline scanner, or isolated scanner?
On Mon, Oct 31, 2022, 15:41 Jeff Turner 
<sjtsp2...@gmail.com<mailto:sjtsp2...@gmail.com>> wrote:
any ideas why mutations via thrift proxy server would take 120 to 150
seconds to show up in a scan?

accumulo 1.9.3

the mutations have all been submitted through thrift (from python), the
writer has been closed, and the writing process has exited.

95% of the time, the latency is between 120.1 and 120.5 seconds.
occasionally the latency is 150 seconds.

there don't appear to be many configuration options for the proxy
server.  and other people i have talked to said that they see their
updates through thrift proxy immediately.

updates via java/jython have millisecond latency.  (for a while i had
been trying to blame tservers or the main server (maybe some delay in
processing compactions, ...).  i don't think that's the issue)



Reply via email to