If you’re really really advanced you MIGHT be able to  use spark + 
cqlsstablewriter to create a ton of sstables with just tombstones one them 
representing deletes, then either nodetool refresh or sstableloader them into 
the cluster 

If you create sstables on the right timestamp boundaries to match your twcs 
windows, each one will compact with the data file or the same window and delete 
the data. 

Will be a ton of compaction though. Not as efficient as the deleting strategy. 
Also not sure if the offline cqlsstablewriter actually supports deletes because 
I’m on my phone and too lazy to check. If it doesn’t it probably wouldn’t be 
that hard to add.


> On Feb 28, 2020, at 10:50 PM, Oleksandr Shulgin 
> <oleksandr.shul...@zalando.de> wrote:
> 
> 
>> On Fri, 28 Feb 2020, 23:02 Erick Ramirez, <erick.rami...@datastax.com> wrote:
>> I'm not personally aware of anyone who is using it successfully other than 
>> ProtectWise where it was a good fit for their narrow use case. My limited 
>> knowledge of it is that it has some sharp edges which is the reason they 
>> haven't pushed for it to be added to Cassandra (that's second hand info so 
>> please don't quote me).
>> 
>> 
>> In any case, my initial reaction is that Spark would probably be a better 
>> fit for what you're trying to achieve particularly since you don't know the 
>> PK up front. Cheers!
> 
> 
> I should have added that this is a one-time task: we do not intend to run it 
> constantly
> 
> And we are using TWCS, so after issuing millions of deletes with Spark (or 
> however else) we would have a hard time compacting the tombstones.
> 
> --
> Alex
>> 

Reply via email to