Les,

This is pretty much Dynamo 101 territory at this point and one of the
tradeoffs with a distributed model.

If you aren't familiar with Dynamo, Andy Gross (from Basho) gives an
AWESOME walkthrough in an episode of TheNoSQLTapes:

http://nosqltapes.com/video/understanding-dynamo-with-andy-gross

Concurrent writes really do behave the same way they would during a
network partition event. You can just stick your riak cluster behind a
load balancer and go to town.

On Tue, Feb 15, 2011 at 3:38 PM, Les Mikesell <lesmikes...@gmail.com> wrote:
> In this scenario, the behavior I would want would be to only see one copy
> for each unique key, but to always have one even if one of the feeds or
> writers fails or the cluster is partitioned, then re-joined. This sounds
> good in theory, but what about the details? Is there likely to be a big
> performance hit from the normally-colliding writes?  Will it take twice the
> disk space, the eventually clean up the duplicates? Would it be reasonable
> to do this to riak-search with something like news stories?
>
>
> On 2/15/2011 2:02 PM, Alexander Sicular wrote:
>>
>> What about siblings? http://wiki.basho.com/REST-API.html (seach for
>> sibling)
>>
>> On Tue, Feb 15, 2011 at 14:57, Dan Reverri<d...@basho.com>  wrote:
>>>
>>> Riak maintains a single value per key and provides mechanisms (vector
>>> clocks) to detect/resolve conflicting values. In the proposed use case
>>> the
>>> multiple copies would overwrite each other and Riak, by default, would
>>> return a single value for a requested key.
>>> Behind the scenes Riak determines the appropriate value per key using
>>> vector
>>> clocks. More information about vector clocks is available here:
>>> http://blog.basho.com/2010/01/29/why-vector-clocks-are-easy/
>>> http://blog.basho.com/2010/04/05/why-vector-clocks-are-hard/
>>> Thanks,
>>> Dan
>>> Daniel Reverri
>>> Developer Advocate
>>> Basho Technologies, Inc.
>>> d...@basho.com
>>>
>>>
>>> On Tue, Feb 15, 2011 at 10:11 AM, Les Mikesell<lesmikes...@gmail.com>
>>> wrote:
>>>>
>>>> Is riak suitable as a very reliable store where you have multiple feeds
>>>> of
>>>> streaming data that are at least theoretically identical?  That is, can
>>>> you
>>>> count on writing multiple copies with the same keys at the same time to
>>>> do
>>>> something reasonable regardless of cluster partitioning?  And is this a
>>>> common usage scenario?
>>>>
>>>> --
>>>>  Les Mikesell
>>>>   lesmikes...@gmail.com
>>>>
>>>> _______________________________________________
>>>> riak-users mailing list
>>>> riak-users@lists.basho.com
>>>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>>>
>>>
>>> _______________________________________________
>>> riak-users mailing list
>>> riak-users@lists.basho.com
>>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>>>
>>>
>
>
> _______________________________________________
> riak-users mailing list
> riak-users@lists.basho.com
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>

_______________________________________________
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com

Reply via email to