If you’re really really advanced you MIGHT be able to use spark + cqlsstablewriter to create a ton of sstables with just tombstones one them representing deletes, then either nodetool refresh or sstableloader them into the cluster
If you create sstables on the right timestamp boundaries to match your twcs windows, each one will compact with the data file or the same window and delete the data. Will be a ton of compaction though. Not as efficient as the deleting strategy. Also not sure if the offline cqlsstablewriter actually supports deletes because I’m on my phone and too lazy to check. If it doesn’t it probably wouldn’t be that hard to add. > On Feb 28, 2020, at 10:50 PM, Oleksandr Shulgin > <oleksandr.shul...@zalando.de> wrote: > > >> On Fri, 28 Feb 2020, 23:02 Erick Ramirez, <erick.rami...@datastax.com> wrote: >> I'm not personally aware of anyone who is using it successfully other than >> ProtectWise where it was a good fit for their narrow use case. My limited >> knowledge of it is that it has some sharp edges which is the reason they >> haven't pushed for it to be added to Cassandra (that's second hand info so >> please don't quote me). >> >> >> In any case, my initial reaction is that Spark would probably be a better >> fit for what you're trying to achieve particularly since you don't know the >> PK up front. Cheers! > > > I should have added that this is a one-time task: we do not intend to run it > constantly > > And we are using TWCS, so after issuing millions of deletes with Spark (or > however else) we would have a hard time compacting the tombstones. > > -- > Alex >>