Hi all,

I've been messing with sstable2json as a means of mass-exporting some data 
(mainly for backups, but also for some convenience trickery on an individual 
nodes' data). However, I've run into a situation where sstable2json appears to 
be dumping out TONS of duplicate columns for a single row.

For example, for a single key, I did

$CASSANDRA_HOME/bin/sstable2json <sstable> -k <key> > output.file

which ran until I killed it manually. Then I executed
cat output.file | sed 's/]/\n/g'  | wc -l

which gave me 40 million and some change. On the other hand,

cat output.file | sed 's/]\n/g' | sort -n | uniq | wc -l

gave me around 10K (much closer to reality).

For my particular data set, the total size of any given row cannot exceed 80K 
columns. So I'm wondering: Is this normal behavior for sstable2json? Assuming 
that it is, is there any way in which I can massage sstable2json into not 
emitting duplicates? These duplicates eat a great deal of disk space and 
processing power to manipulate, which I'd like to avoid.


Thanks for your help,

Scott


Reply via email to