Background info:

6 node cluster.
24 gigs of ram per machine
8 gigs of ram dedicated to c*
4 4 core cpu's
2 250 gig SSD's raid 0
Running c* 1.2.6

The CF is configured as followed

CREATE TABLE behaviors (
  uid text,
  buid int,
  name text,
  expires text,
  value text,
  PRIMARY KEY (uid, buid, name)
) WITH
  bloom_filter_fp_chance=0.010000 AND
  caching='KEYS_ONLY' AND
  comment='' AND
  dclocal_read_repair_chance=0.000000 AND
  gc_grace_seconds=864000 AND
  read_repair_chance=0.100000 AND
  replicate_on_write='true' AND
  populate_io_cache_on_flush='false' AND
  compaction={'sstable_size_in_mb': '160', 'class':
'LeveledCompactionStrategy'} AND
  compression={'sstable_compression': 'SnappyCompressor'};

I am noticing that the read latency is very high considering when I look at
the output of nodetool cfstats.

This is the example output of one of the nodes:

  Column Family: behaviors
                SSTable count: 2
                SSTables in each level: [1, 1, 0, 0, 0, 0, 0, 0, 0]
                Space used (live): 171496198
                Space used (total): 171496591
                Number of Keys (estimate): 1153664
                Memtable Columns Count: 14445
                Memtable Data Size: 1048576
                Memtable Switch Count: 1
                Read Count: 1894
                Read Latency: 0.497 ms.
                Write Count: 7169
                Write Latency: 0.041 ms.
                Pending Tasks: 0
                Bloom Filter False Positives: 4
                Bloom Filter False Ratio: 0.00862
                Bloom Filter Space Used: 3533152
                Compacted row minimum size: 125
                Compacted row maximum size: 9887
                Compacted row mean size: 365

The write latency is awesome, but the read latency, not so much. The output
of iostat doesn't show anything out of the ordinary. The cpu utilization is
between 1% to 5%.

All read queries are issued with a CL of ONE. We always include "WHERE uid
= '<somevalue>'" for the queries.

If there is any more info I can provide, please let me know. At this point
in time, I am a bit stumped.

Regards,

Eric Plowe

Reply via email to