The problem I'm working on is very similar to this. I'm working on a reputation system and we keep a fixed number of day buckets for the scores. So when new data comes in you need to find out what bucket is supposed to be used, remove the data in it if you've moved to a new bucket as the data there would be at least n + 1 days old where n is the number of days you are keeping, and then store the value. In our case we are willing to accept that we might occasionally lose a bit of data, as the data tends to trend towards a "good enough" value quite quickly. Still it would be cool to know a way to make sure that we really know that we are safe to "nuke" a bucket shy if always insisting on blocking writes on a "all" read. This can be very painful if you are replicating data across datacentres.
On 04/16/2010 11:50 AM, Chris Shorrock wrote:
I'm attempting to come up with a technique for limiting the number of columns a single key (or super column - doesn't matter too much for the context of this conversation) may contain at any one time. My actual use-case is a little too meaty to try to describe so an alternate use-case of this mechanism could be:

    /Construct a twitter-esque feed which maintains a list N tweets.
    Tweets (in this system - and in reality I suppose) occur at such a
    rate that you want to limit a given users "feed" to N items. You
    do not have the ability to store an infinite number of tweets due
    to the physical constraints of your hardware./


The "/my first idea/" answer is when a tweet is inserted into the the feed of a given person, that you then do a count and delete of any outstanding tweets. In reality you could first count, then (if count >= N) do a batch mutate for the insertion of the new entry and the removal of the old. My issue with this approach is that after a certain point every new entry into the system will incur the removal of an old entry. The count, once a feed has reached N will always be >= N on any subsequent queries. Depending on how you index the tweets you may need to actually do a read instead of count to get the row identifiers.

My second approach was to utilize a "slot" system where you have a record stored somewhere that indicates the next slot for insertion. This can be thought of as a fixed length array where you store the next insertion point in some other column family. When a new tweet occurs you retrieve the current "slot" meta-data, insert into that index, then update the meta-data for the next insertion. My concerns with this relate around synchronization and losing entries due to concurrent operations. I'd rather not have to something like ZooKeeper to synchronize in the application cluster.

I have some other ideas but I'm mostly just spit-balling at this point. So I thought I'd reach out the collective intelligence of the group to see if anyone has implemented something similar. Thanks in advance.

Reply via email to