I realize I'm kind of late to this party, but what about
using the auto-expire feature and letting riak do the deletion of data
for you?

The link is for an older version, but I know the
functionality still exists in riak2.
http://docs.basho.com/riak/latest/community/faqs/developing/#how-can-i-automatically-expire-a-key-from-riak

Warm regards,
Bryce


On Thu, 4 Jun 2015 09:28:04 +0200
Daniel Abrahamsson <daniel.abrahams...@klarna.com> wrote:

> Hi Peter,
> 
> What is "large-scale" in your case? How many keys do you need to
> delete, and how often?
> 
> //Daniel
> 
> On Wed, Jun 3, 2015 at 9:54 PM, Peter Herndon <tphern...@gmail.com>
> wrote:
> 
> > Interesting thought. It might work for us, it might not, I’ll have
> > to check with our CTO to see whether the expense makes sense under
> > our circumstances.
> >
> > Thanks!
> >
> > —Peter
> > > On Jun 3, 2015, at 2:21 PM, Drew Kerrigan <d...@kerrigan.io>
> > > wrote:
> > >
> > > Another idea for a large-scale one-time removal of data, as well
> > > as an
> > opportunity for a fresh start, would be to:
> > >
> > > 1. set up multi-data center replication between 2 clusters
> > > 2. implement a recv/2 hook on the sink which refuses data from the
> > buckets / keys you would like to ignore / delete
> > > 3. trigger a full sync replication
> > > 4. start using the sync as your new source of data sans the
> > > ignored data
> > >
> > > Obviously this is costly, but it should have a fairly minimal
> > > impact to
> > existing production users other than the moment that you switch
> > traffic from the old cluster to the new one.
> > >
> > > Caveats: Not all Riak features are supported with MDC (search
> > > indexes
> > and strong consistency in particular).
> > >
> > > On Wed, Jun 3, 2015 at 2:11 PM Peter Herndon <tphern...@gmail.com>
> > wrote:
> > > Sadly, this is a production cluster already using leveldb as the
> > backend. With that constraint in mind, and rebuilding the cluster
> > not really being an option to enable multi-backends or bitcask,
> > what would our best approach be?
> > >
> > > Thanks!
> > >
> > > —Peter
> > >
> > > > On Jun 3, 2015, at 12:09 PM, Alexander Sicular
> > > > <sicul...@gmail.com>
> > wrote:
> > > >
> > > > We are actively investigating better options for deletion of
> > > > large
> > amounts of keys. As Sargun mentioned, deleting the data dir for an
> > entire backend via an operationalized rolling restart is probably
> > the best approach right now for killing large amounts of keys.
> > > >
> > > > But if your key space can fit in memory the best way to kill
> > > > keys is
> > to use bitcask ttl if that's an option. 1. If you can even use
> > bitcask in your environment due to the memory overhead and 2. If
> > your use case allows for ttls which it may considering you may
> > already be using time bound buckets....
> > > >
> > > > -Alexander
> > > >
> > > > @siculars
> > > > http://siculars.posthaven.com
> > > >
> > > > Sent from my iRotaryPhone
> > > >
> > > > On Jun 3, 2015, at 09:54, Sargun Dhillon <sdhil...@basho.com>
> > > > wrote:
> > > >
> > > >> You could map your keys to a given bucket, and that bucket to
> > > >> a given
> > backend using multi_backend. There is some cost to having lots of
> > backends (memory overhead, FDs, etc...). When you want to do a mass
> > drop, you could down the node, and delete that given backend, and
> > bring it up. Caveat: AAE, MDC, nor mutable data play well with this
> > scenario.
> > > >>
> > > >> On Wed, Jun 3, 2015 at 10:43 AM, Peter Herndon
> > > >> <tphern...@gmail.com>
> > wrote:
> > > >> Hi list,
> > > >>
> > > >> We’re looking for the best way to handle large scale
> > > >> expiration of
> > no-longer-useful data stored in Riak. We asked a while back, and the
> > recommendation was to store the data in time-segmented buckets
> > (bucket per day or per month), query on the current buckets, and
> > use the streaming list keys API to handle slowly deleting the
> > buckets that have aged out.
> > > >>
> > > >> Is that still the best approach for doing this kind of task?
> > > >> Or is
> > there a better approach?
> > > >>
> > > >> Thanks!
> > > >>
> > > >> —Peter Herndon
> > > >> Sr. Application Engineer
> > > >> @Bitly
> > > >> _______________________________________________
> > > >> riak-users mailing list
> > > >> riak-users@lists.basho.com
> > > >> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> > > >>
> > > >> _______________________________________________
> > > >> riak-users mailing list
> > > >> riak-users@lists.basho.com
> > > >> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> > >
> > >
> > > _______________________________________________
> > > riak-users mailing list
> > > riak-users@lists.basho.com
> > > http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> >
> >
> > _______________________________________________
> > riak-users mailing list
> > riak-users@lists.basho.com
> > http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> >


_______________________________________________
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com

Reply via email to