When dealing with large SliceRanges, it better to read all the results in to memory (by setting "count" to the largest value possible), or is it better to divide the query in to smaller SliceRange queries? Large in this case being on the order of millions of rows.
There's a footnote concerning SliceRanges on the main Apache Cassandra project site that reads: ".Thrift will materialize the whole result into memory before returning it to the client, so be aware that you may be better served by iterating through slices by passing the last value of one call in as the start of the next instead of increasing count arbitrarily large." . but it doesn't delve in to the reasons why going about things that way is better. Can someone shed some light on this? And would the same logic apply to large KeyRanges?