Sounds like you want something like http://oss.oetiker.ch/rrdtool/

Assuming you are trying to store computer log data.

Do you have any other data that can spread the data load?  Like a machine name? 
 If so, you can use a hash of that value to place that "machine"  randomly on 
the net, then append the timestamp, this groups each machine on the ring 
(assuming you don't have a massive # of writes about each machine at one 
time).....

-----Original Message-----
From: pres...@gmail.com [mailto:pres...@gmail.com] On Behalf Of Paul Prescod
Sent: Wednesday, April 07, 2010 5:52 AM
To: cassandra user
Subject: OrderPreservingPartitioner limits and workarounds

I have one append-oriented workload and I would like to know if
Cassandra is appropriate for it.

Given:

 * 100 nodes

 * an OrderPreservingPartitioner

 * a replication factor of "3"

 * a write-pattern of "always append"

 * a strong requirement for range queries

My understanding is that there will exist 3 nodes will end up being
responsible for all writes and potentially a disproportionate amount
of reads (in the common case that users care more about recent data
than older data).

Is there some manual way I can fiddle with InitialTokens and
ReplicationFactors to share the load more fairly?

 Paul Prescod

Reply via email to