Same type of error, but I'm not currently using TTL's.  I am, however, 
generating a lot of tombstones as I add elements to collections….


On Jul 24, 2013, at 6:42 AM, Fabien Rousseau <fab...@yakaz.com> wrote:

> Hi Paul,
> 
> Concerning large rows which are not compacting, I've probably managed to 
> reproduce your problem.
> I suppose you're using collections, but also TTLs ?
> 
> Anyway, I opened an issue here : 
> https://issues.apache.org/jira/browse/CASSANDRA-5799 
> 
> Hope this helps
> 
> 
> 2013/7/24 Christopher Wirt <chris.w...@struq.com>
> Hi Paul,
> 
>  
> 
> Sorry to hear you’re having a low point.
> 
>  
> 
> We ended up not using the collection features of 1.2.
> 
> Instead storing a compressed string containing the map and handling client 
> side.
> 
>  
> 
> We only have fixed schema short rows so no experience with large row 
> compaction.
> 
>  
> 
> File descriptors have never got that high for us. But, if you only have a 
> couple physical nodes with loads of data and small ss-tables maybe they could 
> get that high?
> 
>  
> 
> Only time I’ve had file descriptors get out of hand was then compaction got 
> slightly confused with a new schema when I dropped and recreated instead of 
> truncating. https://issues.apache.org/jira/browse/CASSANDRA-4857 restarting 
> the node fixed the issue.
> 
>  
> 
>  
> 
> From my limited experience I think Cassandra is a dangerous choice for an 
> young limited funding/experience start-up expecting to scale fast. We are a 
> fairly mature start-up with funding. We’ve just spent 3-5 months moving from 
> Mongo to Cassandra. It’s been expensive and painful getting Cassandra to read 
> like Mongo, but we’ve made it J
> 
>  
> 
>  
> 
>  
> 
>  
> 
> From: Paul Ingalls [mailto:paulinga...@gmail.com] 
> Sent: 24 July 2013 06:01
> To: user@cassandra.apache.org
> Subject: disappointed
> 
>  
> 
> I want to check in.  I'm sad, mad and afraid.  I've been trying to get a 1.2 
> cluster up and working with my data set for three weeks with no success.  
> I've been running a 1.1 cluster for 8 months now with no hiccups, but for me 
> at least 1.2 has been a disaster.  I had high hopes for leveraging the new 
> features of 1.2, specifically vnodes and collections.   But at this point I 
> can't release my system into production, and will probably need to find a new 
> back end.  As a small startup, this could be catastrophic.  I'm mostly mad at 
> myself.  I took a risk moving to the new tech.  I forgot sometimes when you 
> gamble, you lose.
> 
>  
> 
> First, the performance of 1.2.6 was horrible when using collections.  I 
> wasn't able to push through 500k rows before the cluster became unusable.  
> With a lot of digging, and way too much time, I discovered I was hitting a 
> bug that had just been fixed, but was unreleased.  This scared me, because 
> the release was already at 1.2.6 and I would have expected something as 
> https://issues.apache.org/jira/browse/CASSANDRA-5677 would have been 
> addressed long before.  But gamely I grabbed the latest code from the 1.2 
> branch, built it and I was finally able to get past half a million rows.  
> 
>  
> 
> But, then I hit ~4 million rows, and a multitude of problems.  Even with the 
> fix above, I was still seeing a ton of compactions failing, specifically the 
> ones for large rows.  Not a single large row will compact, they all assert 
> with the wrong size.  Worse, and this is what kills the whole thing, I keep 
> hitting a wall with open files, even after dumping the whole DB, dropping 
> vnodes and trying again.  Seriously, 650k open file descriptors?  When it 
> hits this limit, the whole DB craps out and is basically unusable.  This 
> isn't that many rows.  I have close to a half a billion in 1.1…
> 
>  
> 
> I'm now at a standstill.  I figure I have two options unless someone here can 
> help me.  Neither of them involve 1.2.  I can either go back to 1.1 and 
> remove the features that collections added to my service, or I find another 
> data backend that has similar performance characteristics to cassandra but 
> allows collections type behavior in a scalable manner.  Cause as far as I can 
> tell, 1.2 doesn't scale.  Which makes me sad, I was proud of what I 
> accomplished with 1.1….
> 
>  
> 
> Does anyone know why there are so many open file descriptors?  Any ideas on 
> why a large row won't compact?
> 
>  
> 
> Paul
> 
> 
> 
> 
> -- 
> Fabien Rousseau
> 
> 
> www.yakaz.com

Reply via email to