I solved this problem with a sub-properties of compaction. (unchecked_tombstone_compaction, tombstone_threshold, tombstone_compaction_interval)
It took time. Eventually, two datacenters were again balanced. Thank you. > On 24 Dec 2018, at 3:48 PM, Eunsu Kim <eunsu.bil...@gmail.com> wrote: > > Oh I’m sorry. > It is marked as included in 3.11.1. > It seems to be confused with other comments in the middle. > However, I am not sure what to do with this page.. > >> On 24 Dec 2018, at 3:35 PM, Eunsu Kim <eunsu.bil...@gmail.com >> <mailto:eunsu.bil...@gmail.com>> wrote: >> >> Thank you for your response. >> >> The patch for the issue page you linked to may be not included in 3.11.3. >> >> If I run repair -pr on all nodes, will both datacenter use the same amount >> of disk? >> >>> On 24 Dec 2018, at 2:25 PM, Jeff Jirsa <jji...@gmail.com >>> <mailto:jji...@gmail.com>> wrote: >>> >>> Seems like this is getting asked more and more, that’s unfortunate. Wish I >>> had time to fix this by making flush smarter or TWCS split old data. But I >>> don’t. >>> >>> You can search the list archives for more examples, but what’s probably >>> happening is that you have sstables overlapping which prevents TWCS from >>> dropping them when fully expired >>> >>> The overlaps probably come from either probabilistic read repair or >>> speculative retry read-repairing data into the memtable on the dc that >>> coordinates your reads >>> >>> Cassandra-13418 ( https://issues.apache.org/jira/browse/CASSANDRA-13418 >>> <https://issues.apache.org/jira/browse/CASSANDRA-13418> ) makes it so you >>> can force sstables to be dropped at expiration regardless of overlaps, but >>> you have to set some properties because it’s technically unsafe (if you >>> write to the table with anything other than ttls). >>> >>> >>> >>> -- >>> Jeff Jirsa >>> >>> >>> On Dec 24, 2018, at 12:05 AM, Eunsu Kim <eunsu.bil...@gmail.com >>> <mailto:eunsu.bil...@gmail.com>> wrote: >>> >>>> I’m using TimeWindowCompactionStrategy. >>>> >>>> All consistency level is ONE. >>>> >>>>> On 24 Dec 2018, at 2:01 PM, Jeff Jirsa <jji...@gmail.com >>>>> <mailto:jji...@gmail.com>> wrote: >>>>> >>>>> What compaction strategy are you using ? >>>>> >>>>> What consistency level do you use on writes? Reads? >>>>> >>>>> -- >>>>> Jeff Jirsa >>>>> >>>>> >>>>>> On Dec 23, 2018, at 11:53 PM, Eunsu Kim <eunsu.bil...@gmail.com >>>>>> <mailto:eunsu.bil...@gmail.com>> wrote: >>>>>> >>>>>> Merry Christmas >>>>>> >>>>>> The Cassandra cluster I operate consists of two datacenters. >>>>>> >>>>>> Most data has a TTL of 14 days and stores one data for each data center. >>>>>> (NetworkTopologyStrategy, datacenter1: 1, datacenter2: 1) >>>>>> >>>>>> However, for a few days ago, only the datacenter1 disk usage is >>>>>> increasing rapidly. >>>>>> >>>>>> There is no change in nodetool cleanup on each node of datacenter1. >>>>>> >>>>>> How does this happen? What can I do? >>>>>> >>>>>> I would appreciate your advice. >>>>>> >>>>>> Thank you in advance. >>>>>> >>>>>> <PastedGraphic-2.png> >>>>>> >>>>>> <PastedGraphic-1.png> >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org >>>>> <mailto:user-unsubscr...@cassandra.apache.org> >>>>> For additional commands, e-mail: user-h...@cassandra.apache.org >>>>> <mailto:user-h...@cassandra.apache.org> >>>>> >>>> >>>> >>>> --------------------------------------------------------------------- >>>> To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org >>>> <mailto:user-unsubscr...@cassandra.apache.org> >>>> For additional commands, e-mail: user-h...@cassandra.apache.org >>>> <mailto:user-h...@cassandra.apache.org> >>>> >> >