So it would appear that 0.7 will have solved the requirement that a single row 
must be able to fit in memory.  That issue aside, how would one expect the 
read/write performance to be in the scenarios listed below?

________________________________
From: Mason Hale [mailto:ma...@onespot.com]
Sent: Tuesday, July 13, 2010 8:41 AM
To: user@cassandra.apache.org
Subject: Re: Consequences of having many columns

Currently there is a limitation that each row must fit in memory (with some not 
insignificant overhead), thus having lots of columns per row can trigger 
out-of-memory errors. This limitation should be removed in a future release.

Please see:
  - http://wiki.apache.org/cassandra/CassandraLimitations
  - https://issues.apache.org/jira/browse/CASSANDRA-16  (notice this is marked 
as resolved now)

Mason
On Tue, Jul 13, 2010 at 9:38 AM, Kochheiser,Todd W - TOK-DITT-1 
<twkochhei...@bpa.gov<mailto:twkochhei...@bpa.gov>> wrote:
I recently ran across a blog posting with a comment from a Cassandra committer 
that indicated a performance penalty when having a large number of columns per 
row/key.  Unfortunately I didn't bookmark the blog posting and now I can't find 
it.  Regardless, since our current plan and design is to have several thousand 
columns per row/key, it made me question our design and if it might cause 
unintended performance consequences.  As a somewhat concrete example for 
discussion purposes, which of the following scenarios would "potential" perform 
better or worse?

Assume:

 *   Single ColumnFamily
 *   Three node cluster
 *   10 to 1 read/write ratio (10 reads to every write)

Scenario A:


 *   10k rows
 *   5k columns/row
 *   Each column ~ 64kB
 *   Hot spot for writes and reads would be a single column in each row (the 
hot column would change every hour).  We would be accessing every row 
constantly, but in general accessing just a few columns in each.
 *   A low volume of reads accessing ~100 columns per row (range queries would 
work)
 *   Access is generally direct (row key / column key)
 *   Data growth would be horizontal (adding columns) as apposed to vertically 
(adding rows)
 *   This is our current design

Scenario B:


 *   50M rows/keys
 *   1 column/key
 *   Each column ~ 64kB
 *   Hot spot for writes and reads would be the single column in 10k rows, but 
the 10k rows accessed would change every hour.
 *   Access would generally be direct (row key / column key)
 *   Data growth would be vertically (adding rows 10k at a time) as apposed to 
horizontal (adding columns)

Scenario C:


 *   5k rows/keys
 *   10k columns/row
 *   Each column ~64kB
 *   Hot spot for writes and reads would be every column in a single row.  Row 
being access would change every hour
 *   Access is generally direct (row key / column key)
 *   Low volume of queries accessing a single column in many rows
 *   Data growth would be by adding rows, each with 10k column.

In all three scenarios the amount of data is the same but the access pattern in 
different.  From an application coding perspective any of the approaches are 
feasible, although the data is easier to think about in Scenario A (i.e. fewer 
mental gymnastics and fewer composite keys).  In all of the scenarios there are 
10k columns that are constantly accessed (read and write).

Some thoughts: Scenario A has the advantage of evenly distributing reads/writes 
across all cluster nodes (I think).  Scenario B has the potential advantage of 
having one column per row (I think) but *not* necessarily distributing evenly 
reads/writes across all cluster nodes.  I'm not serious about Scenario C, but 
it is an option.  Scenario C would probably cause one node in the cluster to 
take the brunt of all reads/writes so I think this design would be a bad idea.  
And, if having lots of columns is a bad idea then this is even worse than 
scenario A.

Regards,
Todd






Reply via email to