Hi, I'm working on an application using a Cassandra (2.1.0) cluster where
- our entire dataset is around 22GB - each node has 48GB of memory but only a single (mechanical) hard disk - in normal operation we have a low level of writes and no reads - very occasionally we need to read rows very fast (>1.5K rows/second), and only read each row once. When we try and read the rows it takes up to five minutes before Cassandra is able to keep up. The problem seems to be that it takes a while to get the data into the page cache and until then Cassandra can't retrieve the data from disk fast enough (e.g. if I drop the page cache mid-test then Cassandra slows down for the next 5 minutes). Given that the total amount of should fit comfortably in memory I've been trying to find a way to keep the rows cached in memory but there doesn't seem to be a particularly great way to achieve this. I've tried enabling the row cache and pre-populating the test by querying every row before starting the load which gives good performance, but the row cache isn't really intended to be used this way and we'd be fighting the row cache to keep the rows in (e.g. by cyclically reading through all the rows during normal operation). Keeping the page cache warm by running a background task to keep accessing the files for the sstables would be simpler and currently this is the solution we're leaning towards, but we have less control over the page cache, it would be vulnerable to other processes knocking Cassandra's files out, and it generally feels like a bit of a hack. Has anyone had any success with trying to do something similar to this or have any suggestions for possible solutions? Thanks, Thomas