Re: Python client performance issue

2011-02-15 Thread Nico Meyer
Hi Mike,

perhaps you can try to upgrade the protocol buffers library to at least
version 2.3.0. This is from the changelog for that version:

Python
  * 10-25 times faster than 2.2.0, still pure-Python.


Cheers,
Nico

Am Montag, den 14.02.2011, 19:35 -0500 schrieb Mike Stoddart:
> Will do when I get time. Would the REST API be any faster?
> 
> Thanks
> Mike
> 
> On Mon, Feb 14, 2011 at 7:01 PM, Thomas Burdick
>  wrote:
> > I would highly recommend looking in to the cProfile and pstat module and
> > profile the code that is going slow. If your using the protocol buffer
> > client it could possibly be related to the fact that python protocol buffers
> > is extraordinarily slow and is well known to be slow. Profile until proven
> > guilty though.
> > Tom Burdick
> >
> > On Mon, Feb 14, 2011 at 7:09 AM, Mike Stoddart  wrote:
> >>
> >> I added some code to my system to test writing data into Riak. I'm
> >> using the Python client library with protocol buffers. I'm writing a
> >> snapshot of my current data, which is one json object containing on
> >> average 60 individual json sub-objects. Each sub object contains about
> >> 22 values.
> >>
> >># Archived entry. ts is a formatted timestamp.
> >>entry = self._bucket.new(ts, data=data)
> >>entry.store()
> >>
> >># Now write the current entry.
> >>entry = self._bucket.new("current", data=data)
> >>entry.store()
> >>
> >> I'm writing the same data twice; the archived copy and the current
> >> copy, which I can easily retrieve later. Performance is lower than
> >> expected; top is showing a constant cpu usage of 10-12%.
> >>
> >> I haven't decided to use Riak; this is to help me decide. But for now
> >> are there any optimisations I can do here? A similiar test with
> >> MongoDB shows a steady cpu usage of 1%. The cpu usages are for my
> >> client, not Riak's own processes. The only differences in my test app
> >> is the code that writes the data to the database. Otherwise all other
> >> code is 100% the same between these two test apps.
> >>
> >> Any suggestions appreciated.
> >> Thanks
> >> Mike
> >>
> >> ___
> >> riak-users mailing list
> >> riak-users@lists.basho.com
> >> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> >
> >
> 
> ___
> riak-users mailing list
> riak-users@lists.basho.com
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com



___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: Python client performance issue

2011-02-15 Thread Andy Gross
python-riak-client already uses version 2.3.0. Adventurous types might want to 
check out https://github.com/Greplin/fast-python-pb, which wraps the C/C++ 
protocol buffers library. 

-- 
Andy Gross
Principal Architect
Basho Technologies, Inc.
On Tuesday, February 15, 2011 at 1:46 AM, Nico Meyer wrote: 
> Hi Mike,
> 
> perhaps you can try to upgrade the protocol buffers library to at least
> version 2.3.0. This is from the changelog for that version:
> 
> Python
>  * 10-25 times faster than 2.2.0, still pure-Python.
> 
> 
> Cheers,
> Nico
> 
> Am Montag, den 14.02.2011, 19:35 -0500 schrieb Mike Stoddart:
> > Will do when I get time. Would the REST API be any faster?
> > 
> > Thanks
> > Mike
> > 
> > On Mon, Feb 14, 2011 at 7:01 PM, Thomas Burdick
> >  wrote:
> > > I would highly recommend looking in to the cProfile and pstat module and
> > > profile the code that is going slow. If your using the protocol buffer
> > > client it could possibly be related to the fact that python protocol 
> > > buffers
> > > is extraordinarily slow and is well known to be slow. Profile until proven
> > > guilty though.
> > > Tom Burdick
> > > 
> > > On Mon, Feb 14, 2011 at 7:09 AM, Mike Stoddart  wrote:
> > > > 
> > > > I added some code to my system to test writing data into Riak. I'm
> > > > using the Python client library with protocol buffers. I'm writing a
> > > > snapshot of my current data, which is one json object containing on
> > > > average 60 individual json sub-objects. Each sub object contains about
> > > > 22 values.
> > > > 
> > > >  # Archived entry. ts is a formatted timestamp.
> > > >  entry = self._bucket.new(ts, data=data)
> > > >  entry.store()
> > > > 
> > > >  # Now write the current entry.
> > > >  entry = self._bucket.new("current", data=data)
> > > >  entry.store()
> > > > 
> > > > I'm writing the same data twice; the archived copy and the current
> > > > copy, which I can easily retrieve later. Performance is lower than
> > > > expected; top is showing a constant cpu usage of 10-12%.
> > > > 
> > > > I haven't decided to use Riak; this is to help me decide. But for now
> > > > are there any optimisations I can do here? A similiar test with
> > > > MongoDB shows a steady cpu usage of 1%. The cpu usages are for my
> > > > client, not Riak's own processes. The only differences in my test app
> > > > is the code that writes the data to the database. Otherwise all other
> > > > code is 100% the same between these two test apps.
> > > > 
> > > > Any suggestions appreciated.
> > > > Thanks
> > > > Mike
> > > > 
> > > > ___
> > > > riak-users mailing list
> > > > riak-users@lists.basho.com
> > > > http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> > 
> > ___
> > riak-users mailing list
> > riak-users@lists.basho.com
> > http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> 
> 
> 
> ___
> riak-users mailing list
> riak-users@lists.basho.com
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> 
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: Python client performance issue

2011-02-15 Thread Nico Meyer
Hi Andy.

I am not quite sure what you mean, is the protobuf library included with
riak-python-client? Or are you talking about the version of the protobuf
compiler that was used to create riakclient_pb2.py from 
riakclient.proto?

Cheers,
Nico

Am Dienstag, den 15.02.2011, 02:23 -0800 schrieb Andy Gross:
> python-riak-client already uses version 2.3.0.  Adventurous types
> might want to check out https://github.com/Greplin/fast-python-pb,
> which wraps the C/C++ protocol buffers library. 
> 
> -- 
> Andy Gross
> Principal Architect
> Basho Technologies, Inc.
> 
> 
> On Tuesday, February 15, 2011 at 1:46 AM, Nico Meyer wrote:
> 
> > Hi Mike,
> > 
> > perhaps you can try to upgrade the protocol buffers library to at
> > least
> > version 2.3.0. This is from the changelog for that version:
> > 
> > Python
> > * 10-25 times faster than 2.2.0, still pure-Python.
> > 
> > 
> > Cheers,
> > Nico
> > 
> > Am Montag, den 14.02.2011, 19:35 -0500 schrieb Mike Stoddart:
> > > Will do when I get time. Would the REST API be any faster?
> > > 
> > > Thanks
> > > Mike
> > > 
> > > On Mon, Feb 14, 2011 at 7:01 PM, Thomas Burdick
> > >  wrote:
> > > > I would highly recommend looking in to the cProfile and pstat
> > > > module and
> > > > profile the code that is going slow. If your using the protocol
> > > > buffer
> > > > client it could possibly be related to the fact that python
> > > > protocol buffers
> > > > is extraordinarily slow and is well known to be slow. Profile
> > > > until proven
> > > > guilty though.
> > > > Tom Burdick
> > > > 
> > > > On Mon, Feb 14, 2011 at 7:09 AM, Mike Stoddart
> > > >  wrote:
> > > > > 
> > > > > I added some code to my system to test writing data into Riak.
> > > > > I'm
> > > > > using the Python client library with protocol buffers. I'm
> > > > > writing a
> > > > > snapshot of my current data, which is one json object
> > > > > containing on
> > > > > average 60 individual json sub-objects. Each sub object
> > > > > contains about
> > > > > 22 values.
> > > > > 
> > > > > # Archived entry. ts is a formatted timestamp.
> > > > > entry = self._bucket.new(ts, data=data)
> > > > > entry.store()
> > > > > 
> > > > > # Now write the current entry.
> > > > > entry = self._bucket.new("current", data=data)
> > > > > entry.store()
> > > > > 
> > > > > I'm writing the same data twice; the archived copy and the
> > > > > current
> > > > > copy, which I can easily retrieve later. Performance is lower
> > > > > than
> > > > > expected; top is showing a constant cpu usage of 10-12%.
> > > > > 
> > > > > I haven't decided to use Riak; this is to help me decide. But
> > > > > for now
> > > > > are there any optimisations I can do here? A similiar test
> > > > > with
> > > > > MongoDB shows a steady cpu usage of 1%. The cpu usages are for
> > > > > my
> > > > > client, not Riak's own processes. The only differences in my
> > > > > test app
> > > > > is the code that writes the data to the database. Otherwise
> > > > > all other
> > > > > code is 100% the same between these two test apps.
> > > > > 
> > > > > Any suggestions appreciated.
> > > > > Thanks
> > > > > Mike
> > > > > 
> > > > > ___
> > > > > riak-users mailing list
> > > > > riak-users@lists.basho.com
> > > > > http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> > > > > 
> > > 
> > > ___
> > > riak-users mailing list
> > > riak-users@lists.basho.com
> > > http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> > > 
> > 
> > 
> > 
> > ___
> > riak-users mailing list
> > riak-users@lists.basho.com
> > http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> > 
> 
> 



___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: Python client performance issue

2011-02-15 Thread Andy Gross


Sorry, I should have been more clear. The Python client depends on 
"protobuf>=2.3.0" in setup.py, so people are already most likely using 
protobufs-2.3.0.

- Andy

On Tuesday, February 15, 2011 at 3:09 AM, Nico Meyer wrote: 
> Hi Andy.
> 
> I am not quite sure what you mean, is the protobuf library included with
> riak-python-client? Or are you talking about the version of the protobuf
> compiler that was used to create riakclient_pb2.py from 
> riakclient.proto?
> 
> Cheers,
> Nico
> 
> Am Dienstag, den 15.02.2011, 02:23 -0800 schrieb Andy Gross:
> > python-riak-client already uses version 2.3.0. Adventurous types
> > might want to check out https://github.com/Greplin/fast-python-pb,
> > which wraps the C/C++ protocol buffers library. 
> > 
> > -- 
> > Andy Gross
> > Principal Architect
> > Basho Technologies, Inc.
> > 
> > 
> > On Tuesday, February 15, 2011 at 1:46 AM, Nico Meyer wrote:
> > 
> > > Hi Mike,
> > > 
> > > perhaps you can try to upgrade the protocol buffers library to at
> > > least
> > > version 2.3.0. This is from the changelog for that version:
> > > 
> > > Python
> > > * 10-25 times faster than 2.2.0, still pure-Python.
> > > 
> > > 
> > > Cheers,
> > > Nico
> > > 
> > > Am Montag, den 14.02.2011, 19:35 -0500 schrieb Mike Stoddart:
> > > > Will do when I get time. Would the REST API be any faster?
> > > > 
> > > > Thanks
> > > > Mike
> > > > 
> > > > On Mon, Feb 14, 2011 at 7:01 PM, Thomas Burdick
> > > >  wrote:
> > > > > I would highly recommend looking in to the cProfile and pstat
> > > > > module and
> > > > > profile the code that is going slow. If your using the protocol
> > > > > buffer
> > > > > client it could possibly be related to the fact that python
> > > > > protocol buffers
> > > > > is extraordinarily slow and is well known to be slow. Profile
> > > > > until proven
> > > > > guilty though.
> > > > > Tom Burdick
> > > > > 
> > > > > On Mon, Feb 14, 2011 at 7:09 AM, Mike Stoddart
> > > > >  wrote:
> > > > > > 
> > > > > > I added some code to my system to test writing data into Riak.
> > > > > > I'm
> > > > > > using the Python client library with protocol buffers. I'm
> > > > > > writing a
> > > > > > snapshot of my current data, which is one json object
> > > > > > containing on
> > > > > > average 60 individual json sub-objects. Each sub object
> > > > > > contains about
> > > > > > 22 values.
> > > > > > 
> > > > > > # Archived entry. ts is a formatted timestamp.
> > > > > > entry = self._bucket.new(ts, data=data)
> > > > > > entry.store()
> > > > > > 
> > > > > > # Now write the current entry.
> > > > > > entry = self._bucket.new("current", data=data)
> > > > > > entry.store()
> > > > > > 
> > > > > > I'm writing the same data twice; the archived copy and the
> > > > > > current
> > > > > > copy, which I can easily retrieve later. Performance is lower
> > > > > > than
> > > > > > expected; top is showing a constant cpu usage of 10-12%.
> > > > > > 
> > > > > > I haven't decided to use Riak; this is to help me decide. But
> > > > > > for now
> > > > > > are there any optimisations I can do here? A similiar test
> > > > > > with
> > > > > > MongoDB shows a steady cpu usage of 1%. The cpu usages are for
> > > > > > my
> > > > > > client, not Riak's own processes. The only differences in my
> > > > > > test app
> > > > > > is the code that writes the data to the database. Otherwise
> > > > > > all other
> > > > > > code is 100% the same between these two test apps.
> > > > > > 
> > > > > > Any suggestions appreciated.
> > > > > > Thanks
> > > > > > Mike
> > > > > > 
> > > > > > ___
> > > > > > riak-users mailing list
> > > > > > riak-users@lists.basho.com
> > > > > > http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> > > > 
> > > > ___
> > > > riak-users mailing list
> > > > riak-users@lists.basho.com
> > > > http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> > > 
> > > 
> > > 
> > > ___
> > > riak-users mailing list
> > > riak-users@lists.basho.com
> > > http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> 
> 
> 
> ___
> riak-users mailing list
> riak-users@lists.basho.com
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> 
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


redundant writers

2011-02-15 Thread Les Mikesell
Is riak suitable as a very reliable store where you have multiple feeds 
of streaming data that are at least theoretically identical?  That is, 
can you count on writing multiple copies with the same keys at the same 
time to do something reasonable regardless of cluster partitioning?  And 
is this a common usage scenario?


--
  Les Mikesell
   lesmikes...@gmail.com

___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: redundant writers

2011-02-15 Thread Dan Reverri
Riak maintains a single value per key and provides mechanisms (vector
clocks) to detect/resolve conflicting values. In the proposed use case the
multiple copies would overwrite each other and Riak, by default, would
return a single value for a requested key.

Behind the scenes Riak determines the appropriate value per key using vector
clocks. More information about vector clocks is available here:
http://blog.basho.com/2010/01/29/why-vector-clocks-are-easy/
http://blog.basho.com/2010/04/05/why-vector-clocks-are-hard/

Thanks,
Dan

Daniel Reverri
Developer Advocate
Basho Technologies, Inc.
d...@basho.com


On Tue, Feb 15, 2011 at 10:11 AM, Les Mikesell wrote:

> Is riak suitable as a very reliable store where you have multiple feeds of
> streaming data that are at least theoretically identical?  That is, can you
> count on writing multiple copies with the same keys at the same time to do
> something reasonable regardless of cluster partitioning?  And is this a
> common usage scenario?
>
> --
>  Les Mikesell
>   lesmikes...@gmail.com
>
> ___
> riak-users mailing list
> riak-users@lists.basho.com
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: redundant writers

2011-02-15 Thread Alexander Sicular
What about siblings? http://wiki.basho.com/REST-API.html (seach for sibling)

On Tue, Feb 15, 2011 at 14:57, Dan Reverri  wrote:
> Riak maintains a single value per key and provides mechanisms (vector
> clocks) to detect/resolve conflicting values. In the proposed use case the
> multiple copies would overwrite each other and Riak, by default, would
> return a single value for a requested key.
> Behind the scenes Riak determines the appropriate value per key using vector
> clocks. More information about vector clocks is available here:
> http://blog.basho.com/2010/01/29/why-vector-clocks-are-easy/
> http://blog.basho.com/2010/04/05/why-vector-clocks-are-hard/
> Thanks,
> Dan
> Daniel Reverri
> Developer Advocate
> Basho Technologies, Inc.
> d...@basho.com
>
>
> On Tue, Feb 15, 2011 at 10:11 AM, Les Mikesell 
> wrote:
>>
>> Is riak suitable as a very reliable store where you have multiple feeds of
>> streaming data that are at least theoretically identical?  That is, can you
>> count on writing multiple copies with the same keys at the same time to do
>> something reasonable regardless of cluster partitioning?  And is this a
>> common usage scenario?
>>
>> --
>>  Les Mikesell
>>   lesmikes...@gmail.com
>>
>> ___
>> riak-users mailing list
>> riak-users@lists.basho.com
>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>
>
> ___
> riak-users mailing list
> riak-users@lists.basho.com
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>
>

___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: redundant writers

2011-02-15 Thread Les Mikesell
In this scenario, the behavior I would want would be to only see one 
copy for each unique key, but to always have one even if one of the 
feeds or writers fails or the cluster is partitioned, then re-joined. 
This sounds good in theory, but what about the details? Is there likely 
to be a big performance hit from the normally-colliding writes?  Will it 
take twice the disk space, the eventually clean up the duplicates? Would 
it be reasonable to do this to riak-search with something like news stories?



On 2/15/2011 2:02 PM, Alexander Sicular wrote:

What about siblings? http://wiki.basho.com/REST-API.html (seach for sibling)

On Tue, Feb 15, 2011 at 14:57, Dan Reverri  wrote:

Riak maintains a single value per key and provides mechanisms (vector
clocks) to detect/resolve conflicting values. In the proposed use case the
multiple copies would overwrite each other and Riak, by default, would
return a single value for a requested key.
Behind the scenes Riak determines the appropriate value per key using vector
clocks. More information about vector clocks is available here:
http://blog.basho.com/2010/01/29/why-vector-clocks-are-easy/
http://blog.basho.com/2010/04/05/why-vector-clocks-are-hard/
Thanks,
Dan
Daniel Reverri
Developer Advocate
Basho Technologies, Inc.
d...@basho.com


On Tue, Feb 15, 2011 at 10:11 AM, Les Mikesell
wrote:


Is riak suitable as a very reliable store where you have multiple feeds of
streaming data that are at least theoretically identical?  That is, can you
count on writing multiple copies with the same keys at the same time to do
something reasonable regardless of cluster partitioning?  And is this a
common usage scenario?

--
  Les Mikesell
   lesmikes...@gmail.com

___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com



___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com





___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: redundant writers

2011-02-15 Thread John E. Vincent
Les,

This is pretty much Dynamo 101 territory at this point and one of the
tradeoffs with a distributed model.

If you aren't familiar with Dynamo, Andy Gross (from Basho) gives an
AWESOME walkthrough in an episode of TheNoSQLTapes:

http://nosqltapes.com/video/understanding-dynamo-with-andy-gross

Concurrent writes really do behave the same way they would during a
network partition event. You can just stick your riak cluster behind a
load balancer and go to town.

On Tue, Feb 15, 2011 at 3:38 PM, Les Mikesell  wrote:
> In this scenario, the behavior I would want would be to only see one copy
> for each unique key, but to always have one even if one of the feeds or
> writers fails or the cluster is partitioned, then re-joined. This sounds
> good in theory, but what about the details? Is there likely to be a big
> performance hit from the normally-colliding writes?  Will it take twice the
> disk space, the eventually clean up the duplicates? Would it be reasonable
> to do this to riak-search with something like news stories?
>
>
> On 2/15/2011 2:02 PM, Alexander Sicular wrote:
>>
>> What about siblings? http://wiki.basho.com/REST-API.html (seach for
>> sibling)
>>
>> On Tue, Feb 15, 2011 at 14:57, Dan Reverri  wrote:
>>>
>>> Riak maintains a single value per key and provides mechanisms (vector
>>> clocks) to detect/resolve conflicting values. In the proposed use case
>>> the
>>> multiple copies would overwrite each other and Riak, by default, would
>>> return a single value for a requested key.
>>> Behind the scenes Riak determines the appropriate value per key using
>>> vector
>>> clocks. More information about vector clocks is available here:
>>> http://blog.basho.com/2010/01/29/why-vector-clocks-are-easy/
>>> http://blog.basho.com/2010/04/05/why-vector-clocks-are-hard/
>>> Thanks,
>>> Dan
>>> Daniel Reverri
>>> Developer Advocate
>>> Basho Technologies, Inc.
>>> d...@basho.com
>>>
>>>
>>> On Tue, Feb 15, 2011 at 10:11 AM, Les Mikesell
>>> wrote:

 Is riak suitable as a very reliable store where you have multiple feeds
 of
 streaming data that are at least theoretically identical?  That is, can
 you
 count on writing multiple copies with the same keys at the same time to
 do
 something reasonable regardless of cluster partitioning?  And is this a
 common usage scenario?

 --
  Les Mikesell
   lesmikes...@gmail.com

 ___
 riak-users mailing list
 riak-users@lists.basho.com
 http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>>>
>>>
>>> ___
>>> riak-users mailing list
>>> riak-users@lists.basho.com
>>> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>>>
>>>
>
>
> ___
> riak-users mailing list
> riak-users@lists.basho.com
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
>

___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Memory usage

2011-02-15 Thread Matthias-Christian Ott
Hi,

does somebody have some figures on Riak memory usage? I consider using
it in a highly memory constrained environment (< 32 MiB), but I'm not
sure whether Riak can handle low workloads under such tight constrains
(it consumed > 20 MiB idle). Performance is not relevant to me, I just
need partition tolerance, availability and persistence.

Regards,
Matthias-Christian

___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Riak Java Client

2011-02-15 Thread Mark Wolfe
Gday

Been messing around on some personal projects which use riak with Java and
was interested to know if their was a roadmap or task list of the java
client.

Just wanted to know if there was any plans to build a new version in the
future as I would be interested in contributing.

I am currently working on a Log4J appender which writes directly to riak
(see http://github.com/wolfeidau/riak-log4j-appender ). Currently using
Netty network library to communicate with riak over HTTP as I don't need a
complete riak client, and wanted to keep my dependencies as lean as
possible.


-- 
Regards,

Mark Wolfe

--
I am not young enough to know everything.
--
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: Memory usage

2011-02-15 Thread Jeremiah Peschka
You'll be constrained in the number of keys that you can store (using the 
bitcask backend). There's a good spreadsheet for capacity planning: 
https://spreadsheets.google.com/ccc?key=0Ak4OBkABJPsxdEowYXc2akxnYU9xNkJmbmZscnhaTFE&hl=en&authkey=CMHw8tYO

I don't know how well Riak would behave otherwise in an environment that tight. 
I'd think you'd be limited to single key lookups, but I have been wrong before.

-- 
Jeremiah Peschka
Microsoft SQL Server MVP
MCITP: Database Developer, DBA
On Tuesday, February 15, 2011 at 4:34 PM, Matthias-Christian Ott wrote: 
> Hi,
> 
> does somebody have some figures on Riak memory usage? I consider using
> it in a highly memory constrained environment (< 32 MiB), but I'm not
> sure whether Riak can handle low workloads under such tight constrains
> (it consumed > 20 MiB idle). Performance is not relevant to me, I just
> need partition tolerance, availability and persistence.
> 
> Regards,
> Matthias-Christian
> 
> ___
> riak-users mailing list
> riak-users@lists.basho.com
> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com
> 
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: Memory usage

2011-02-15 Thread Matthias-Christian Ott
On Tue, Feb 15, 2011 at 04:44:58PM -0800, Jeremiah Peschka wrote:
> You'll be constrained in the number of keys that you can store (using the 
> bitcask backend). There's a good spreadsheet for capacity planning: 
> https://spreadsheets.google.com/ccc?key=0Ak4OBkABJPsxdEowYXc2akxnYU9xNkJmbmZscnhaTFE&hl=en&authkey=CMHw8tYO

The storage seems sufficient for me.

> I don't know how well Riak would behave otherwise in an environment that 
> tight. I'd think you'd be limited to single key lookups, but I have been 
> wrong before.

So you think it might not be possible to run map-reduce functions?

Regards,
Matthias-Christian

___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Re: Memory usage

2011-02-15 Thread Jeremiah Peschka
I'm not sure. I think it would depend on the number of JavaScript MR engines 
you have running, the operations you're performing, and the amount of data that 
is being read into memory.

-- 
Jeremiah Peschka
Microsoft SQL Server MVP
MCITP: Database Developer, DBA
On Tuesday, February 15, 2011 at 5:10 PM, Matthias-Christian Ott wrote: 
> On Tue, Feb 15, 2011 at 04:44:58PM -0800, Jeremiah Peschka wrote:
> > You'll be constrained in the number of keys that you can store (using the 
> > bitcask backend). There's a good spreadsheet for capacity planning: 
> > https://spreadsheets.google.com/ccc?key=0Ak4OBkABJPsxdEowYXc2akxnYU9xNkJmbmZscnhaTFE&hl=en&authkey=CMHw8tYO
> 
> The storage seems sufficient for me.
> 
> > I don't know how well Riak would behave otherwise in an environment that 
> > tight. I'd think you'd be limited to single key lookups, but I have been 
> > wrong before.
> 
> So you think it might not be possible to run map-reduce functions?
> 
> Regards,
> Matthias-Christian
> 
___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com


Enhanced Querying for Ripple

2011-02-15 Thread Sean Cribbs
Lately I've noticed the increased demand for a higher-level, abstract querying 
API in the Ripple document-mapper library -- something analogous to 
ActiveRecord + arel. At the current time, I'm preparing a strawman proposal for 
the feature that we can debate.

I'd like to discuss this feature in-depth on Friday with interested members of 
the community and then set about implementing it.  If you would like to 
participate in the conversation, please enter your email in the form I've 
linked to below. 

HOWEVER, I only ask that if you want to be involved in the conversation, **you 
should be willing to contribute some time to making this API happen** 
preferably by implementing and testing needed features. As much as I like being 
a code-hero, this is too big to do by myself.

I will email the strawman proposal to participants by Thursday so you have time 
to review it. (We will likely meet on GoToMeeting, which requires Windows or 
Mac, sorry Linux users.)

https://spreadsheets.google.com/a/basho.com/viewform?formkey=dDFKU0lNMkY0aVZyMFZ1Y244UXA1RHc6MQ

Cheers!

Sean Cribbs 
Developer Advocate
Basho Technologies, Inc.
http://basho.com/

___
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com