> Our team decided to use Cassandra as storage solution to a dataset.
> I am very new to the NoSQL world and Cassandra so I am hoping to get
> some help from the community: The dataset is pretty simple, we have
> for each key a number of columns with values. Each day we compute a
> new version of
> After a memtable flush, you see minimum cpu and maximum read
> throughput both in term of disk and cassandra records read.
> As memtable increase in size, cpu goes up and read drops.
> If this is because of memtable or GC performance issue, this is the
> big question.
>
> As each memtable is just
update column family CommentSent with column_type = 'Super' and
comparator=UTF8Type and default_validation_class=UTF8Type and
column_metadata=[{column_name:null,validation_class:UTF8Type}]
When I load 17GB(as nodetool ring shows) data to a Cassandra node which is
clean before data loading, all the files in the data directory can have a size
larger than 100GB. Is it normal?