Hello. I have a schema that represents a filesystem and one example of a Super CF is:
CF FilesPerDir: (DIRNAME -> (FILENAME -> (attribute1: value1, attribute2: value2)) And in cases of directory moves, I have to fetch all files of that directory and subdirectories. This implies one cassandra query per dir, or a multiget for all needed dirs. The multiget() is more efficient, but I'm having trouble trying to limit the size of the data returned in order to not crash the cassandra node. I'm using the pycassa client lib, and until now I have been doing per-directory get()s specifiying a column_count. This effectively limits the size of the dataset, but I would like to perform a multiget() to fetch the contents of multiple dirs at a time. The problem is that it seems that the column_count is per-key, and not global per dataset. So if I set column_count = 10000, as I have now, but fetch 1000 dirs (rows) and each one happens to have 10000 files (columns) the dataset is 1000x10000. Is there a better way to query for this data or does multiget deal with this through the "buffer_size"? Thanks, André