Oh I’m sorry. It is marked as included in 3.11.1. It seems to be confused with other comments in the middle. However, I am not sure what to do with this page..
> On 24 Dec 2018, at 3:35 PM, Eunsu Kim <eunsu.bil...@gmail.com> wrote: > > Thank you for your response. > > The patch for the issue page you linked to may be not included in 3.11.3. > > If I run repair -pr on all nodes, will both datacenter use the same amount of > disk? > >> On 24 Dec 2018, at 2:25 PM, Jeff Jirsa <jji...@gmail.com >> <mailto:jji...@gmail.com>> wrote: >> >> Seems like this is getting asked more and more, that’s unfortunate. Wish I >> had time to fix this by making flush smarter or TWCS split old data. But I >> don’t. >> >> You can search the list archives for more examples, but what’s probably >> happening is that you have sstables overlapping which prevents TWCS from >> dropping them when fully expired >> >> The overlaps probably come from either probabilistic read repair or >> speculative retry read-repairing data into the memtable on the dc that >> coordinates your reads >> >> Cassandra-13418 ( https://issues.apache.org/jira/browse/CASSANDRA-13418 >> <https://issues.apache.org/jira/browse/CASSANDRA-13418> ) makes it so you >> can force sstables to be dropped at expiration regardless of overlaps, but >> you have to set some properties because it’s technically unsafe (if you >> write to the table with anything other than ttls). >> >> >> >> -- >> Jeff Jirsa >> >> >> On Dec 24, 2018, at 12:05 AM, Eunsu Kim <eunsu.bil...@gmail.com >> <mailto:eunsu.bil...@gmail.com>> wrote: >> >>> I’m using TimeWindowCompactionStrategy. >>> >>> All consistency level is ONE. >>> >>>> On 24 Dec 2018, at 2:01 PM, Jeff Jirsa <jji...@gmail.com >>>> <mailto:jji...@gmail.com>> wrote: >>>> >>>> What compaction strategy are you using ? >>>> >>>> What consistency level do you use on writes? Reads? >>>> >>>> -- >>>> Jeff Jirsa >>>> >>>> >>>>> On Dec 23, 2018, at 11:53 PM, Eunsu Kim <eunsu.bil...@gmail.com >>>>> <mailto:eunsu.bil...@gmail.com>> wrote: >>>>> >>>>> Merry Christmas >>>>> >>>>> The Cassandra cluster I operate consists of two datacenters. >>>>> >>>>> Most data has a TTL of 14 days and stores one data for each data center. >>>>> (NetworkTopologyStrategy, datacenter1: 1, datacenter2: 1) >>>>> >>>>> However, for a few days ago, only the datacenter1 disk usage is >>>>> increasing rapidly. >>>>> >>>>> There is no change in nodetool cleanup on each node of datacenter1. >>>>> >>>>> How does this happen? What can I do? >>>>> >>>>> I would appreciate your advice. >>>>> >>>>> Thank you in advance. >>>>> >>>>> <PastedGraphic-2.png> >>>>> >>>>> <PastedGraphic-1.png> >>>> >>>> --------------------------------------------------------------------- >>>> To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org >>>> <mailto:user-unsubscr...@cassandra.apache.org> >>>> For additional commands, e-mail: user-h...@cassandra.apache.org >>>> <mailto:user-h...@cassandra.apache.org> >>>> >>> >>> >>> --------------------------------------------------------------------- >>> To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org >>> <mailto:user-unsubscr...@cassandra.apache.org> >>> For additional commands, e-mail: user-h...@cassandra.apache.org >>> <mailto:user-h...@cassandra.apache.org> >>> >