Hi everybody, State column contains "R" or "D" values. Just a single character. As Rajsekhar said, only difference is the table can contain high number of cell count. In the mean time we made a major compaction and data per node was 5-6 gb.
On Mon, Jul 22, 2019, 10:56 AM Rajsekhar Mallick <raj.mallic...@gmail.com> wrote: > Hello Team, > > The difference in write latencies between both the tables though > significant,but the higher latency being 11.353 ms is still acceptable. > > Overall Writes not being an issue, but write latency for this particular > table on the higher side does point towards data being written to the table. > Few things which I noticed, is the data in cell count column in nodetool > tablehistogram o/p for message_history_state table is scattered. > The partition size histogram for the tables is consistent, but the column > count histogram for the impacted table isn't uniform. > May be we can start thinking on these lines. > > I would also wait for some expert advice here. > > Thanks > > > On Mon,the 22 Jul, 2019, 12:31 PM Ben Slater, <ben.sla...@instaclustr.com> > wrote: > >> Is the size of the data in your “state” column variable? The higher write >> latencies at the 95%+ could line up with large volumes of data for >> particular rows in that column (the one column not in both tables)? >> >> Cheers >> Ben >> >> --- >> >> >> *Ben Slater**Chief Product Officer* >> >> <https://www.instaclustr.com/platform/> >> >> <https://www.facebook.com/instaclustr> >> <https://twitter.com/instaclustr> >> <https://www.linkedin.com/company/instaclustr> >> >> Read our latest technical blog posts here >> <https://www.instaclustr.com/blog/>. >> >> This email has been sent on behalf of Instaclustr Pty. Limited >> (Australia) and Instaclustr Inc (USA). >> >> This email and any attachments may contain confidential and legally >> privileged information. If you are not the intended recipient, do not copy >> or disclose its content, but please reply to this email immediately and >> highlight the error to the sender and then immediately delete the message. >> >> >> On Mon, 22 Jul 2019 at 16:46, CPC <acha...@gmail.com> wrote: >> >>> Hi guys, >>> >>> Any idea? I thought it might be a bug but could not find anything >>> related on jira. >>> >>> On Fri, Jul 19, 2019, 12:45 PM CPC <acha...@gmail.com> wrote: >>> >>>> Hi Rajsekhar, >>>> >>>> Here the details: >>>> >>>> 1) >>>> >>>> [cassadm@bipcas00 ~]$ nodetool tablestats tims.MESSAGE_HISTORY >>>> Total number of tables: 259 >>>> ---------------- >>>> Keyspace : tims >>>> Read Count: 208256144 >>>> Read Latency: 7.655146714749506 ms >>>> Write Count: 2218205275 >>>> Write Latency: 1.7826005103175133 ms >>>> Pending Flushes: 0 >>>> Table: MESSAGE_HISTORY >>>> SSTable count: 41 >>>> Space used (live): 976964101899 >>>> Space used (total): 976964101899 >>>> Space used by snapshots (total): 3070598526780 >>>> Off heap memory used (total): 185828820 >>>> SSTable Compression Ratio: 0.8219217809913125 >>>> Number of partitions (estimate): 8175715 >>>> Memtable cell count: 73124 >>>> Memtable data size: 26543733 >>>> Memtable off heap memory used: 27829672 >>>> Memtable switch count: 1607 >>>> Local read count: 7871917 >>>> Local read latency: 1.187 ms >>>> Local write count: 172220954 >>>> Local write latency: 0.021 ms >>>> Pending flushes: 0 >>>> Percent repaired: 0.0 >>>> Bloom filter false positives: 130 >>>> Bloom filter false ratio: 0.00000 >>>> Bloom filter space used: 10898488 >>>> Bloom filter off heap memory used: 10898160 >>>> Index summary off heap memory used: 2480140 >>>> Compression metadata off heap memory used: 144620848 >>>> Compacted partition minimum bytes: 36 >>>> Compacted partition maximum bytes: 557074610 >>>> Compacted partition mean bytes: 155311 >>>> Average live cells per slice (last five minutes): >>>> 25.56639344262295 >>>> Maximum live cells per slice (last five minutes): 5722 >>>> Average tombstones per slice (last five minutes): >>>> 1.8681948424068768 >>>> Maximum tombstones per slice (last five minutes): 770 >>>> Dropped Mutations: 97812 >>>> >>>> ---------------- >>>> [cassadm@bipcas00 ~]$ nodetool tablestats tims.MESSAGE_HISTORY_STATE >>>> Total number of tables: 259 >>>> ---------------- >>>> Keyspace : tims >>>> Read Count: 208257486 >>>> Read Latency: 7.655137315414438 ms >>>> Write Count: 2218218966 >>>> Write Latency: 1.7825896304427324 ms >>>> Pending Flushes: 0 >>>> Table: MESSAGE_HISTORY_STATE >>>> SSTable count: 5 >>>> Space used (live): 6403033568 >>>> Space used (total): 6403033568 >>>> Space used by snapshots (total): 19086872706 >>>> Off heap memory used (total): 6727565 >>>> SSTable Compression Ratio: 0.271857664111622 >>>> Number of partitions (estimate): 1396462 >>>> Memtable cell count: 77450 >>>> Memtable data size: 620776 >>>> Memtable off heap memory used: 1338914 >>>> Memtable switch count: 1616 >>>> Local read count: 988278 >>>> Local read latency: 0.518 ms >>>> Local write count: 109292691 >>>> Local write latency: 11.353 ms >>>> Pending flushes: 0 >>>> Percent repaired: 0.0 >>>> Bloom filter false positives: 0 >>>> Bloom filter false ratio: 0.00000 >>>> Bloom filter space used: 1876208 >>>> Bloom filter off heap memory used: 1876168 >>>> Index summary off heap memory used: 410747 >>>> Compression metadata off heap memory used: 3101736 >>>> Compacted partition minimum bytes: 36 >>>> Compacted partition maximum bytes: 129557750 >>>> Compacted partition mean bytes: 17937 >>>> Average live cells per slice (last five minutes): >>>> 4.692893401015229 >>>> Maximum live cells per slice (last five minutes): 258 >>>> Average tombstones per slice (last five minutes): 1.0 >>>> Maximum tombstones per slice (last five minutes): 1 >>>> Dropped Mutations: 1344158 >>>> >>>> 2) >>>> >>>> [cassadm@bipcas00 conf]$ nodetool tablehistograms tims MESSAGE_HISTORY >>>> tims/MESSAGE_HISTORY histograms >>>> Percentile SSTables Write Latency Read Latency Partition Size >>>> Cell Count >>>> (micros) (micros) (bytes) >>>> 50% 3.00 20.50 454.83 14237 >>>> 17 >>>> 75% 17.00 24.60 2346.80 88148 >>>> 103 >>>> 95% 17.00 35.43 14530.76 454826 >>>> 924 >>>> 98% 17.00 42.51 20924.30 1131752 >>>> 2299 >>>> 99% 17.00 42.51 30130.99 1955666 >>>> 4768 >>>> Min 0.00 3.97 73.46 36 >>>> 0 >>>> Max 20.00 263.21 74975.55 386857368 >>>> 943127 >>>> >>>> [cassadm@bipcas00 conf]$ nodetool tablehistograms tims >>>> MESSAGE_HISTORY_STATE >>>> tims/MESSAGE_HISTORY_STATE histograms >>>> Percentile SSTables Write Latency Read Latency Partition Size >>>> Cell Count >>>> (micros) (micros) (bytes) >>>> 50% 5.00 20.50 315.85 924 >>>> 1 >>>> 75% 6.00 35.43 379.02 5722 >>>> 7 >>>> 95% 10.00 4055.27 785.94 61214 >>>> 310 >>>> 98% 10.00 74975.55 3379.39 182785 >>>> 924 >>>> 99% 10.00 107964.79 10090.81 315852 >>>> 1916 >>>> Min 0.00 3.31 42.51 36 >>>> 0 >>>> Max 10.00 322381.14 25109.16 129557750 >>>> 1629722 >>>> >>>> 3) RF=3 >>>> >>>> 4)CL QUORUM >>>> >>>> 5) Single insert prepared statement. no LOGGED/UNLOGGED batch or LWT >>>> >>>> >>>> On Thu, 18 Jul 2019 at 20:51, Rajsekhar Mallick < >>>> raj.mallic...@gmail.com> wrote: >>>> >>>>> Hello, >>>>> >>>>> Kindly post below details >>>>> >>>>> 1. Nodetool cfstats for both the tables. >>>>> 2. Nodetool cfhistograms for both the tables. >>>>> 3. Replication factor of the tables. >>>>> 4. Consistency with which write requests are sent >>>>> 5. Also the type of write queries for the table if handy would also >>>>> help (Light weight transactions or Batch writes or Prepared statements) >>>>> >>>>> Thanks >>>>> >>>>> On 2019/07/18 15:48:09, CPC <a...@gmail.com> wrote: >>>>> > Hi all,> >>>>> > >>>>> > Our cassandra cluster consist of two dc and every dc we have 10 >>>>> nodes. We> >>>>> > are using DSE 5.1.12 (cassandra 3.11).We have a high local write >>>>> latency on> >>>>> > a single table. All other tables in our keyspace have normal >>>>> latencies> >>>>> > like 0.02 msec,even tables that have more write tps and more data. >>>>> Below> >>>>> > you can find two table descriptions and their latencies.> >>>>> > message_history_state have high local write latency. This is not >>>>> node> >>>>> > specific every node have this high local write latency for> >>>>> > message_history_state. Have you ever see such a behavior or any clue >>>>> why> >>>>> > this could happen?> >>>>> > >>>>> > CREATE TABLE tims."MESSAGE_HISTORY" (> >>>>> > > username text,> >>>>> > > date_partition text,> >>>>> > > jid text,> >>>>> > > sent_time timestamp,> >>>>> > > message_id text,> >>>>> > > stanza text,> >>>>> > > PRIMARY KEY ((username, date_partition), jid, sent_time, >>>>> message_id)> >>>>> > > ) WITH CLUSTERING ORDER BY (jid ASC, sent_time DESC, message_id >>>>> ASC)> >>>>> > > AND bloom_filter_fp_chance = 0.01> >>>>> > > AND caching = {'keys': 'ALL', 'rows_per_partition': 'NONE'}> >>>>> > > AND comment = ''> >>>>> > > AND compaction = {'bucket_high': '1.5', 'bucket_low': '0.5', >>>>> 'class':> >>>>> > > >>>>> 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy',> >>>>> > > 'enabled': 'true', 'max_threshold': '32', 'min_sstable_size': >>>>> '50',> >>>>> > > 'min_threshold': '4', 'tombstone_compaction_interval': '86400',> >>>>> > > 'tombstone_threshold': '0.2', 'unchecked_tombstone_compaction': >>>>> 'false'}> >>>>> > > AND compression = {'chunk_length_in_kb': '64', 'class':> >>>>> > > 'org.apache.cassandra.io.compress.LZ4Compressor'}> >>>>> > > AND crc_check_chance = 1.0> >>>>> > > AND dclocal_read_repair_chance = 0.0> >>>>> > > AND default_time_to_live = 0> >>>>> > > AND gc_grace_seconds = 86400> >>>>> > > AND max_index_interval = 2048> >>>>> > > AND memtable_flush_period_in_ms = 0> >>>>> > > AND min_index_interval = 128> >>>>> > > AND read_repair_chance = 0.0> >>>>> > > AND speculative_retry = '99PERCENTILE';> >>>>> > >> >>>>> > > CREATE TABLE tims."MESSAGE_HISTORY_STATE" (> >>>>> > > username text,> >>>>> > > date_partition text,> >>>>> > > message_id text,> >>>>> > > jid text,> >>>>> > > state text,> >>>>> > > sent_time timestamp,> >>>>> > > PRIMARY KEY ((username, date_partition), message_id, jid, >>>>> state)> >>>>> > > ) WITH CLUSTERING ORDER BY (message_id ASC, jid ASC, state ASC)> >>>>> > > AND bloom_filter_fp_chance = 0.01> >>>>> > > AND caching = {'keys': 'ALL', 'rows_per_partition': 'NONE'}> >>>>> > > AND comment = ''> >>>>> > > AND compaction = {'class':> >>>>> > > >>>>> 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy',> >>>>> > > 'max_threshold': '32', 'min_threshold': '4'}> >>>>> > > AND compression = {'chunk_length_in_kb': '64', 'class':> >>>>> > > 'org.apache.cassandra.io.compress.LZ4Compressor'}> >>>>> > > AND crc_check_chance = 1.0> >>>>> > > AND dclocal_read_repair_chance = 0.1> >>>>> > > AND default_time_to_live = 0> >>>>> > > AND gc_grace_seconds = 864000> >>>>> > > AND max_index_interval = 2048> >>>>> > > AND memtable_flush_period_in_ms = 0> >>>>> > > AND min_index_interval = 128> >>>>> > > AND read_repair_chance = 0.0> >>>>> > > AND speculative_retry = '99PERCENTILE';> >>>>> > >> >>>>> > >>>>> > message_history Local write latency: 0.021 ms> >>>>> > message_history_state Local write latency: 11.353 ms> >>>>> > >>>>> > Thanks in advance.> >>>>> > >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org >>>>> For additional commands, e-mail: user-h...@cassandra.apache.org >>>>> >>>>>