On Sat, Sep 3, 2011 at 8:53 PM, Jonathan Ellis <jbel...@gmail.com> wrote:

> I strongly suspect that you're optimizing prematurely.  What evidence
> do you have that timestamps are producing unacceptable overhead for
> your workload?


It's possible … this is back of the envelope at the moment as right now it's
a nonstarter.


> You do realize that the sparse data model means that
> we spend a lot more than 8 bytes storing column names in-line with
> each column too, right?
>

Yeah… this can be mitigated if the column names are your data.


>
> If disk space is really the limiting factor for your workload, I would
> recommend testing the compression code in trunk.  That will get you a
> lot farther than adding extra options for a very niche scenario.
>
>
Another thing I've been considering is building a serializer/deserializer in
front of Cassandra and running my own protocol to talk to it which builds
its own encoding per row to avoid using excessive columns.

Kevin

-- 

Founder/CEO Spinn3r.com

Location: *San Francisco, CA*
Skype: *burtonator*

Skype-in: *(415) 871-0687*

Reply via email to