DOH! I should have known that answer wouldn't have been that simple.
Best of luck.

Warm regards,
Bryce


On Thu, 4 Jun 2015 13:46:28 -0400
Peter Herndon <tphern...@gmail.com> wrote:

> Thanks, Bryce, but auto-expire relies on bitcask being the back-end,
> and we’re on leveldb.
> 
> > On Jun 4, 2015, at 1:24 PM, Bryce Verdier <bryceverd...@gmail.com>
> > wrote:
> > 
> > I realize I'm kind of late to this party, but what about
> > using the auto-expire feature and letting riak do the deletion of
> > data for you?
> > 
> > The link is for an older version, but I know the
> > functionality still exists in riak2.
> > http://docs.basho.com/riak/latest/community/faqs/developing/#how-can-i-automatically-expire-a-key-from-riak
> > 
> > Warm regards,
> > Bryce
> > 
> > 
> > On Thu, 4 Jun 2015 09:28:04 +0200
> > Daniel Abrahamsson <daniel.abrahams...@klarna.com> wrote:
> > 
> >> Hi Peter,
> >> 
> >> What is "large-scale" in your case? How many keys do you need to
> >> delete, and how often?
> >> 
> >> //Daniel
> >> 
> >> On Wed, Jun 3, 2015 at 9:54 PM, Peter Herndon <tphern...@gmail.com>
> >> wrote:
> >> 
> >>> Interesting thought. It might work for us, it might not, I’ll have
> >>> to check with our CTO to see whether the expense makes sense under
> >>> our circumstances.
> >>> 
> >>> Thanks!
> >>> 
> >>> —Peter
> >>>> On Jun 3, 2015, at 2:21 PM, Drew Kerrigan <d...@kerrigan.io>
> >>>> wrote:
> >>>> 
> >>>> Another idea for a large-scale one-time removal of data, as well
> >>>> as an
> >>> opportunity for a fresh start, would be to:
> >>>> 
> >>>> 1. set up multi-data center replication between 2 clusters
> >>>> 2. implement a recv/2 hook on the sink which refuses data from
> >>>> the
> >>> buckets / keys you would like to ignore / delete
> >>>> 3. trigger a full sync replication
> >>>> 4. start using the sync as your new source of data sans the
> >>>> ignored data
> >>>> 
> >>>> Obviously this is costly, but it should have a fairly minimal
> >>>> impact to
> >>> existing production users other than the moment that you switch
> >>> traffic from the old cluster to the new one.
> >>>> 
> >>>> Caveats: Not all Riak features are supported with MDC (search
> >>>> indexes
> >>> and strong consistency in particular).
> >>>> 
> >>>> On Wed, Jun 3, 2015 at 2:11 PM Peter Herndon
> >>>> <tphern...@gmail.com>
> >>> wrote:
> >>>> Sadly, this is a production cluster already using leveldb as the
> >>> backend. With that constraint in mind, and rebuilding the cluster
> >>> not really being an option to enable multi-backends or bitcask,
> >>> what would our best approach be?
> >>>> 
> >>>> Thanks!
> >>>> 
> >>>> —Peter
> >>>> 
> >>>>> On Jun 3, 2015, at 12:09 PM, Alexander Sicular
> >>>>> <sicul...@gmail.com>
> >>> wrote:
> >>>>> 
> >>>>> We are actively investigating better options for deletion of
> >>>>> large
> >>> amounts of keys. As Sargun mentioned, deleting the data dir for an
> >>> entire backend via an operationalized rolling restart is probably
> >>> the best approach right now for killing large amounts of keys.
> >>>>> 
> >>>>> But if your key space can fit in memory the best way to kill
> >>>>> keys is
> >>> to use bitcask ttl if that's an option. 1. If you can even use
> >>> bitcask in your environment due to the memory overhead and 2. If
> >>> your use case allows for ttls which it may considering you may
> >>> already be using time bound buckets....
> >>>>> 
> >>>>> -Alexander
> >>>>> 
> >>>>> @siculars
> >>>>> http://siculars.posthaven.com
> >>>>> 
> >>>>> Sent from my iRotaryPhone
> >>>>> 
> >>>>> On Jun 3, 2015, at 09:54, Sargun Dhillon <sdhil...@basho.com>
> >>>>> wrote:
> >>>>> 
> >>>>>> You could map your keys to a given bucket, and that bucket to
> >>>>>> a given
> >>> backend using multi_backend. There is some cost to having lots of
> >>> backends (memory overhead, FDs, etc...). When you want to do a
> >>> mass drop, you could down the node, and delete that given
> >>> backend, and bring it up. Caveat: AAE, MDC, nor mutable data play
> >>> well with this scenario.
> >>>>>> 
> >>>>>> On Wed, Jun 3, 2015 at 10:43 AM, Peter Herndon
> >>>>>> <tphern...@gmail.com>
> >>> wrote:
> >>>>>> Hi list,
> >>>>>> 
> >>>>>> We’re looking for the best way to handle large scale
> >>>>>> expiration of
> >>> no-longer-useful data stored in Riak. We asked a while back, and
> >>> the recommendation was to store the data in time-segmented buckets
> >>> (bucket per day or per month), query on the current buckets, and
> >>> use the streaming list keys API to handle slowly deleting the
> >>> buckets that have aged out.
> >>>>>> 
> >>>>>> Is that still the best approach for doing this kind of task?
> >>>>>> Or is
> >>> there a better approach?
> >>>>>> 
> >>>>>> Thanks!
> >>>>>> 
> >>>>>> —Peter Herndon
> >>>>>> Sr. Application Engineer
> >>>>>> @Bitly
> >>>>>> _______________________________________________
> >>>>>> riak-users mailing list
> >>>>>> riak-users@lists.basho.com
> >>>>>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> >>>>>> 
> >>>>>> _______________________________________________
> >>>>>> riak-users mailing list
> >>>>>> riak-users@lists.basho.com
> >>>>>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> >>>> 
> >>>> 
> >>>> _______________________________________________
> >>>> riak-users mailing list
> >>>> riak-users@lists.basho.com
> >>>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> >>> 
> >>> 
> >>> _______________________________________________
> >>> riak-users mailing list
> >>> riak-users@lists.basho.com
> >>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> >>> 
> > 
> 


_______________________________________________
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com

Reply via email to