Hi everybody,

State column contains "R" or "D" values. Just a single character. As Rajsekhar
said, only difference is the table can contain high number of cell count.
In the mean time we made a major compaction and data per node was 5-6 gb.

On Mon, Jul 22, 2019, 10:56 AM Rajsekhar Mallick <raj.mallic...@gmail.com>
wrote:

> Hello Team,
>
> The difference in write latencies between both the tables though
> significant,but the higher latency being 11.353 ms is still acceptable.
>
> Overall Writes not being an issue, but write latency for this particular
> table on the higher side does point towards data being written to the table.
> Few things which I noticed, is the data in cell count column in nodetool
> tablehistogram o/p for message_history_state table is scattered.
> The partition size histogram for the tables is consistent, but the column
> count histogram for the impacted table isn't uniform.
> May be we can start thinking on these lines.
>
> I would also wait for some expert advice here.
>
> Thanks
>
>
> On Mon,the 22 Jul, 2019, 12:31 PM Ben Slater, <ben.sla...@instaclustr.com>
> wrote:
>
>> Is the size of the data in your “state” column variable? The higher write
>> latencies at the 95%+ could line up with large volumes of data for
>> particular rows in that column (the one column not in both tables)?
>>
>> Cheers
>> Ben
>>
>> ---
>>
>>
>> *Ben Slater**Chief Product Officer*
>>
>> <https://www.instaclustr.com/platform/>
>>
>> <https://www.facebook.com/instaclustr>
>> <https://twitter.com/instaclustr>
>> <https://www.linkedin.com/company/instaclustr>
>>
>> Read our latest technical blog posts here
>> <https://www.instaclustr.com/blog/>.
>>
>> This email has been sent on behalf of Instaclustr Pty. Limited
>> (Australia) and Instaclustr Inc (USA).
>>
>> This email and any attachments may contain confidential and legally
>> privileged information.  If you are not the intended recipient, do not copy
>> or disclose its content, but please reply to this email immediately and
>> highlight the error to the sender and then immediately delete the message.
>>
>>
>> On Mon, 22 Jul 2019 at 16:46, CPC <acha...@gmail.com> wrote:
>>
>>> Hi guys,
>>>
>>> Any idea? I thought it might be a bug but could not find anything
>>> related on jira.
>>>
>>> On Fri, Jul 19, 2019, 12:45 PM CPC <acha...@gmail.com> wrote:
>>>
>>>> Hi Rajsekhar,
>>>>
>>>> Here the details:
>>>>
>>>> 1)
>>>>
>>>> [cassadm@bipcas00 ~]$ nodetool tablestats tims.MESSAGE_HISTORY
>>>> Total number of tables: 259
>>>> ----------------
>>>> Keyspace : tims
>>>>         Read Count: 208256144
>>>>         Read Latency: 7.655146714749506 ms
>>>>         Write Count: 2218205275
>>>>         Write Latency: 1.7826005103175133 ms
>>>>         Pending Flushes: 0
>>>>                 Table: MESSAGE_HISTORY
>>>>                 SSTable count: 41
>>>>                 Space used (live): 976964101899
>>>>                 Space used (total): 976964101899
>>>>                 Space used by snapshots (total): 3070598526780
>>>>                 Off heap memory used (total): 185828820
>>>>                 SSTable Compression Ratio: 0.8219217809913125
>>>>                 Number of partitions (estimate): 8175715
>>>>                 Memtable cell count: 73124
>>>>                 Memtable data size: 26543733
>>>>                 Memtable off heap memory used: 27829672
>>>>                 Memtable switch count: 1607
>>>>                 Local read count: 7871917
>>>>                 Local read latency: 1.187 ms
>>>>                 Local write count: 172220954
>>>>                 Local write latency: 0.021 ms
>>>>                 Pending flushes: 0
>>>>                 Percent repaired: 0.0
>>>>                 Bloom filter false positives: 130
>>>>                 Bloom filter false ratio: 0.00000
>>>>                 Bloom filter space used: 10898488
>>>>                 Bloom filter off heap memory used: 10898160
>>>>                 Index summary off heap memory used: 2480140
>>>>                 Compression metadata off heap memory used: 144620848
>>>>                 Compacted partition minimum bytes: 36
>>>>                 Compacted partition maximum bytes: 557074610
>>>>                 Compacted partition mean bytes: 155311
>>>>                 Average live cells per slice (last five minutes): 
>>>> 25.56639344262295
>>>>                 Maximum live cells per slice (last five minutes): 5722
>>>>                 Average tombstones per slice (last five minutes): 
>>>> 1.8681948424068768
>>>>                 Maximum tombstones per slice (last five minutes): 770
>>>>                 Dropped Mutations: 97812
>>>>
>>>> ----------------
>>>> [cassadm@bipcas00 ~]$ nodetool tablestats tims.MESSAGE_HISTORY_STATE
>>>> Total number of tables: 259
>>>> ----------------
>>>> Keyspace : tims
>>>>         Read Count: 208257486
>>>>         Read Latency: 7.655137315414438 ms
>>>>         Write Count: 2218218966
>>>>         Write Latency: 1.7825896304427324 ms
>>>>         Pending Flushes: 0
>>>>                 Table: MESSAGE_HISTORY_STATE
>>>>                 SSTable count: 5
>>>>                 Space used (live): 6403033568
>>>>                 Space used (total): 6403033568
>>>>                 Space used by snapshots (total): 19086872706
>>>>                 Off heap memory used (total): 6727565
>>>>                 SSTable Compression Ratio: 0.271857664111622
>>>>                 Number of partitions (estimate): 1396462
>>>>                 Memtable cell count: 77450
>>>>                 Memtable data size: 620776
>>>>                 Memtable off heap memory used: 1338914
>>>>                 Memtable switch count: 1616
>>>>                 Local read count: 988278
>>>>                 Local read latency: 0.518 ms
>>>>                 Local write count: 109292691
>>>>                 Local write latency: 11.353 ms
>>>>                 Pending flushes: 0
>>>>                 Percent repaired: 0.0
>>>>                 Bloom filter false positives: 0
>>>>                 Bloom filter false ratio: 0.00000
>>>>                 Bloom filter space used: 1876208
>>>>                 Bloom filter off heap memory used: 1876168
>>>>                 Index summary off heap memory used: 410747
>>>>                 Compression metadata off heap memory used: 3101736
>>>>                 Compacted partition minimum bytes: 36
>>>>                 Compacted partition maximum bytes: 129557750
>>>>                 Compacted partition mean bytes: 17937
>>>>                 Average live cells per slice (last five minutes): 
>>>> 4.692893401015229
>>>>                 Maximum live cells per slice (last five minutes): 258
>>>>                 Average tombstones per slice (last five minutes): 1.0
>>>>                 Maximum tombstones per slice (last five minutes): 1
>>>>                 Dropped Mutations: 1344158
>>>>
>>>> 2)
>>>>
>>>> [cassadm@bipcas00 conf]$ nodetool tablehistograms tims MESSAGE_HISTORY
>>>> tims/MESSAGE_HISTORY histograms
>>>> Percentile  SSTables     Write Latency      Read Latency    Partition Size 
>>>>        Cell Count
>>>>                               (micros)          (micros)           (bytes)
>>>> 50%             3.00             20.50            454.83             14237 
>>>>                17
>>>> 75%            17.00             24.60           2346.80             88148 
>>>>               103
>>>> 95%            17.00             35.43          14530.76            454826 
>>>>               924
>>>> 98%            17.00             42.51          20924.30           1131752 
>>>>              2299
>>>> 99%            17.00             42.51          30130.99           1955666 
>>>>              4768
>>>> Min             0.00              3.97             73.46                36 
>>>>                 0
>>>> Max            20.00            263.21          74975.55         386857368 
>>>>            943127
>>>>
>>>> [cassadm@bipcas00 conf]$ nodetool tablehistograms tims 
>>>> MESSAGE_HISTORY_STATE
>>>> tims/MESSAGE_HISTORY_STATE histograms
>>>> Percentile  SSTables     Write Latency      Read Latency    Partition Size 
>>>>        Cell Count
>>>>                               (micros)          (micros)           (bytes)
>>>> 50%             5.00             20.50            315.85               924 
>>>>                 1
>>>> 75%             6.00             35.43            379.02              5722 
>>>>                 7
>>>> 95%            10.00           4055.27            785.94             61214 
>>>>               310
>>>> 98%            10.00          74975.55           3379.39            182785 
>>>>               924
>>>> 99%            10.00         107964.79          10090.81            315852 
>>>>              1916
>>>> Min             0.00              3.31             42.51                36 
>>>>                 0
>>>> Max            10.00         322381.14          25109.16         129557750 
>>>>           1629722
>>>>
>>>> 3) RF=3
>>>>
>>>> 4)CL  QUORUM
>>>>
>>>> 5) Single insert prepared statement. no LOGGED/UNLOGGED batch or LWT
>>>>
>>>>
>>>> On Thu, 18 Jul 2019 at 20:51, Rajsekhar Mallick <
>>>> raj.mallic...@gmail.com> wrote:
>>>>
>>>>> Hello,
>>>>>
>>>>> Kindly post below details
>>>>>
>>>>> 1. Nodetool cfstats for both the tables.
>>>>> 2. Nodetool cfhistograms for both the tables.
>>>>> 3. Replication factor of the tables.
>>>>> 4. Consistency with which write requests are sent
>>>>> 5. Also the type of write queries for the table  if handy would also
>>>>> help (Light weight transactions or Batch writes or Prepared statements)
>>>>>
>>>>> Thanks
>>>>>
>>>>> On 2019/07/18 15:48:09, CPC <a...@gmail.com> wrote:
>>>>> > Hi all,>
>>>>> >
>>>>> > Our cassandra cluster consist of two dc and every dc we have 10
>>>>> nodes. We>
>>>>> > are using DSE 5.1.12 (cassandra 3.11).We have a high local write
>>>>> latency on>
>>>>> > a single table. All other tables in our keyspace  have normal
>>>>> latencies>
>>>>> > like 0.02 msec,even tables that have more write tps and more data.
>>>>> Below>
>>>>> > you can find two table descriptions and their latencies.>
>>>>> > message_history_state have high local write latency. This is not
>>>>> node>
>>>>> > specific every node have this high local write latency for>
>>>>> > message_history_state. Have you ever see such a behavior or any clue
>>>>> why>
>>>>> > this could happen?>
>>>>> >
>>>>> >  CREATE TABLE tims."MESSAGE_HISTORY" (>
>>>>> > >     username text,>
>>>>> > >     date_partition text,>
>>>>> > >     jid text,>
>>>>> > >     sent_time timestamp,>
>>>>> > >     message_id text,>
>>>>> > >     stanza text,>
>>>>> > >     PRIMARY KEY ((username, date_partition), jid, sent_time,
>>>>> message_id)>
>>>>> > > ) WITH CLUSTERING ORDER BY (jid ASC, sent_time DESC, message_id
>>>>> ASC)>
>>>>> > >     AND bloom_filter_fp_chance = 0.01>
>>>>> > >     AND caching = {'keys': 'ALL', 'rows_per_partition': 'NONE'}>
>>>>> > >     AND comment = ''>
>>>>> > >     AND compaction = {'bucket_high': '1.5', 'bucket_low': '0.5',
>>>>> 'class':>
>>>>> > >
>>>>> 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy',>
>>>>> > > 'enabled': 'true', 'max_threshold': '32', 'min_sstable_size':
>>>>> '50',>
>>>>> > > 'min_threshold': '4', 'tombstone_compaction_interval': '86400',>
>>>>> > > 'tombstone_threshold': '0.2', 'unchecked_tombstone_compaction':
>>>>> 'false'}>
>>>>> > >     AND compression = {'chunk_length_in_kb': '64', 'class':>
>>>>> > > 'org.apache.cassandra.io.compress.LZ4Compressor'}>
>>>>> > >     AND crc_check_chance = 1.0>
>>>>> > >     AND dclocal_read_repair_chance = 0.0>
>>>>> > >     AND default_time_to_live = 0>
>>>>> > >     AND gc_grace_seconds = 86400>
>>>>> > >     AND max_index_interval = 2048>
>>>>> > >     AND memtable_flush_period_in_ms = 0>
>>>>> > >     AND min_index_interval = 128>
>>>>> > >     AND read_repair_chance = 0.0>
>>>>> > >     AND speculative_retry = '99PERCENTILE';>
>>>>> > >>
>>>>> > > CREATE TABLE tims."MESSAGE_HISTORY_STATE" (>
>>>>> > >     username text,>
>>>>> > >     date_partition text,>
>>>>> > >     message_id text,>
>>>>> > >     jid text,>
>>>>> > >     state text,>
>>>>> > >     sent_time timestamp,>
>>>>> > >     PRIMARY KEY ((username, date_partition), message_id, jid,
>>>>> state)>
>>>>> > > ) WITH CLUSTERING ORDER BY (message_id ASC, jid ASC, state ASC)>
>>>>> > >     AND bloom_filter_fp_chance = 0.01>
>>>>> > >     AND caching = {'keys': 'ALL', 'rows_per_partition': 'NONE'}>
>>>>> > >     AND comment = ''>
>>>>> > >     AND compaction = {'class':>
>>>>> > >
>>>>> 'org.apache.cassandra.db.compaction.SizeTieredCompactionStrategy',>
>>>>> > > 'max_threshold': '32', 'min_threshold': '4'}>
>>>>> > >     AND compression = {'chunk_length_in_kb': '64', 'class':>
>>>>> > > 'org.apache.cassandra.io.compress.LZ4Compressor'}>
>>>>> > >     AND crc_check_chance = 1.0>
>>>>> > >     AND dclocal_read_repair_chance = 0.1>
>>>>> > >     AND default_time_to_live = 0>
>>>>> > >     AND gc_grace_seconds = 864000>
>>>>> > >     AND max_index_interval = 2048>
>>>>> > >     AND memtable_flush_period_in_ms = 0>
>>>>> > >     AND min_index_interval = 128>
>>>>> > >     AND read_repair_chance = 0.0>
>>>>> > >     AND speculative_retry = '99PERCENTILE';>
>>>>> > >>
>>>>> >
>>>>> > message_history Local write latency: 0.021 ms>
>>>>> > message_history_state Local write latency: 11.353 ms>
>>>>> >
>>>>> > Thanks in advance.>
>>>>> >
>>>>> ---------------------------------------------------------------------
>>>>> To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
>>>>> For additional commands, e-mail: user-h...@cassandra.apache.org
>>>>>
>>>>>

Reply via email to