You can disable compaction and enable it later. Use nodetool and 
setcompactionthreshold to 0 0

-Chris

On Dec 18, 2010, at 6:05 PM, Wayne wrote:

> Rereading through everything again I am starting to wonder if the page cache 
> is being affected by compaction. We have been heavily loading data for weeks 
> and compaction is basically running non-stop. The manual compaction should be 
> done some time tomorrow, so when totally caught up I will try again. What 
> changes can be hoped for in 1470 or 1882 in terms of isolating compactions 
> (or writes) affects on read requests?
> 
> Thanks
> 
> 
> 
> On Sat, Dec 18, 2010 at 2:36 PM, Peter Schuller <peter.schul...@infidyne.com> 
> wrote:
> > You are absolutely back to my main concern. Initially we were consistently
> > seeing < 10ms read latency and now we see 25ms (30GB sstable file), 50ms
> > (100GB sstable file) and 65ms (330GB table file) read times for a single
> > read with nothing else going on in the cluster. Concurrency is not our
> > problem/concern (at this point), our problem is slow reads in total
> > isolation. Frankly the concern is that a 2TB node with a 1TB sstable (worst
> > case scenario) will result in > 100ms read latency in total isolation.
> 
> So if you have a single non-concurrent client, along, submitting these
> reads that take 65 ms - are you disk bound (according to the last
> column of iostat -x 1), and how many reads per second (rps column) are
> you seeing relative to client reads? Is the number of disk reads per
> client read consistent with the actual number of sstables at the time?
> 
> The behavior you're describing really does seem indicative of a
> problem, unless the the bottleneck is legitimately reads from disk
> from multiple sstables resulting from rows being spread over said
> sstables.
> 
> --
> / Peter Schuller
> 

Reply via email to