When dealing with large SliceRanges, it better to read all the results in to
memory (by setting "count" to the largest value possible), or is it better
to divide the query in to smaller SliceRange queries? Large in this case
being on the order of millions of rows.

 

There's a footnote concerning SliceRanges on the main Apache Cassandra
project site that reads:

 

".Thrift will materialize the whole result into memory before returning it
to the client, so be aware that you may be better served by iterating
through slices by passing the last value of one call in as the start of the
next instead of increasing count arbitrarily large."

 

. but it doesn't delve in to the reasons why going about things that way is
better. 

 

Can someone shed some light on this? And would the same logic apply to large
KeyRanges?

 

Reply via email to