Thanks. This is an extremely efficient way of counting the size of the bucket.
On 22 November 2011 15:23, francisco treacy <francisco.tre...@gmail.com>wrote: > I would probably stream keys to the client and count them there (it's the > most efficient method I can think of) > > If you have node.js installed, do this: > > npm install riak-js@latest > node -e "require('riak-js').getClient({ port: 8098 }).count('bucket');" > > > 2011/11/21 Stephen Bennett <st...@bennettweb.org> > >> I have a bucket which contains images refernced by a key which is made up >> from a guid. I have a number of servers in my cluster and my bucket is set >> up to store 3 versions of every item in the bucket across the servers in >> the cluster. I'd like to understand a little bit more about how my cluster >> is performing in terms of data storage. I can find out how much space each >> bitcask is currently taking up on each server, but I'd like to compare >> reference this against the number of unique keys that are being stored in >> the system. >> >> I've tried to use map-reduce methods using the erlang methods defined in >> the riak_kv_mapreduce, calling them against the HTTP interface but my >> queries are timing out. I've tried to extend the timeout, but it's still >> timing out. >> >> What's the most efficient way to find out how many keys exist in a >> particular bucket? >> _______________________________________________ >> riak-users mailing list >> riak-users@lists.basho.com >> http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com >> >> >
_______________________________________________ riak-users mailing list riak-users@lists.basho.com http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com