You shouldn't be using beta1 unless your goal is to help shake the bugs out. :)
0.7.5 release is in progress; the artifacts are at http://people.apache.org/~slebresne/ On Tue, Apr 26, 2011 at 12:19 AM, Sanjeev Kulkarni <sanj...@locomatix.com> wrote: > BTW where do i download 0.7.5? I went > to http://www.apache.org/dyn/closer.cgi?path=/cassandra/0.7.5/apache-cassandra-0.7.5-bin.tar.gz > but all the links there are broken. > I was thinking if I just skip 0.7.5 and go with 0.8-beta1, would that be > more advisable? > Thanks! > > On Mon, Apr 25, 2011 at 9:30 PM, Jonathan Ellis <jbel...@gmail.com> wrote: >> >> No. You'll need to run scrub. >> >> On Mon, Apr 25, 2011 at 11:19 PM, Sanjeev Kulkarni >> <sanj...@locomatix.com> wrote: >> > Hi, >> > Thanks for pointing out the fix. My followup question is if I install >> > 0.7.5 >> > will the problem go away with the current data? >> > Thanks! >> > On Mon, Apr 25, 2011 at 8:25 PM, Jonathan Ellis <jbel...@gmail.com> >> > wrote: >> >> >> >> Ah... could be https://issues.apache.org/jira/browse/CASSANDRA-2349 >> >> (fixed for 0.7.5) >> >> >> >> On Mon, Apr 25, 2011 at 9:47 PM, Sanjeev Kulkarni >> >> <sanj...@locomatix.com> >> >> wrote: >> >> > The only other interesting information is that the columns of these >> >> > rows >> >> > all >> >> > had some ttl attached to them. Not sure if that matters. >> >> > Thanks! >> >> > >> >> > On Mon, Apr 25, 2011 at 5:27 PM, Terje Marthinussen >> >> > <tmarthinus...@gmail.com> wrote: >> >> >> >> >> >> First column in the row has offset in the file of 190226525, last >> >> >> valid >> >> >> column is at 380293592, about 181MB from first column to last. >> >> >> in_memory_compaction_limit was 128MB, so almost certainly above the >> >> >> limit. >> >> >> Terje >> >> >> >> >> >> On Tue, Apr 26, 2011 at 8:53 AM, Terje Marthinussen >> >> >> <tmarthinus...@gmail.com> wrote: >> >> >>> >> >> >>> In my case, probably yes. From thw rows I have looked at, I think I >> >> >>> have >> >> >>> only seen this on rows with 1 million plus columns/supercolumns. >> >> >>> >> >> >>> May very well been larger than in memory limit. I think the >> >> >>> compacted >> >> >>> row >> >> >>> I looked closer at was about 200MB and the in memory limit may have >> >> >>> been >> >> >>> 256MB. >> >> >>> >> >> >>> I will see if we still got files around to verify. >> >> >>> >> >> >>> Regards, >> >> >>> Terje >> >> >>> >> >> >>> On 26 Apr 2011, at 02:08, Jonathan Ellis <jbel...@gmail.com> wrote: >> >> >>> >> >> >>> > Was it on a "large" row? (> in_memory_compaction_limit?) >> >> >>> > >> >> >>> > I'm starting to suspect that LazilyCompactedRow is computing row >> >> >>> > size >> >> >>> > incorrectly in some cases. >> >> >>> > >> >> >>> > On Mon, Apr 25, 2011 at 11:47 AM, Terje Marthinussen >> >> >>> > <tmarthinus...@gmail.com> wrote: >> >> >>> >> I have been hunting similar looking corruptions, especially in >> >> >>> >> the >> >> >>> >> hints >> >> >>> >> column family, but I believe it occurs somewhere while >> >> >>> >> compacting. >> >> >>> >> I looked in greater detail on one sstable and the row length was >> >> >>> >> longer than >> >> >>> >> the actual data in the row, and as far as I could see, either >> >> >>> >> the >> >> >>> >> length was >> >> >>> >> wrong or the row was missing data as there was was no extra data >> >> >>> >> in >> >> >>> >> the row >> >> >>> >> after the last column. >> >> >>> >> This was however on a somewhat aging dataset, so suspected it >> >> >>> >> could >> >> >>> >> be >> >> >>> >> related to 2376. >> >> >>> >> >> >> >>> >> Playing around with 0.8 at the moment and not seen it there >> >> >>> >> yet.... >> >> >>> >> (bet it >> >> >>> >> will show up tomorrow once I wrote that.. :)) >> >> >>> >> Terje >> >> >>> >> >> >> >>> >> On Tue, Apr 26, 2011 at 12:44 AM, Sanjeev Kulkarni >> >> >>> >> <sanj...@locomatix.com> >> >> >>> >> wrote: >> >> >>> >>> >> >> >>> >>> Hi Sylvain, >> >> >>> >>> I started it from 0.7.4 with the patch 2376. No upgrade. >> >> >>> >>> Thanks! >> >> >>> >>> >> >> >>> >>> On Mon, Apr 25, 2011 at 7:48 AM, Sylvain Lebresne >> >> >>> >>> <sylv...@datastax.com> >> >> >>> >>> wrote: >> >> >>> >>>> >> >> >>> >>>> Hi Sanjeev, >> >> >>> >>>> >> >> >>> >>>> What's the story of the cluster ? Did you started with 0.7.4, >> >> >>> >>>> or >> >> >>> >>>> is >> >> >>> >>>> it >> >> >>> >>>> upgraded from >> >> >>> >>>> some earlier version ? >> >> >>> >>>> >> >> >>> >>>> On Mon, Apr 25, 2011 at 5:54 AM, Sanjeev Kulkarni >> >> >>> >>>> <sanj...@locomatix.com> >> >> >>> >>>> wrote: >> >> >>> >>>>> Hey guys, >> >> >>> >>>>> Running a one node cassandra server with version 0.7.4 >> >> >>> >>>>> patched >> >> >>> >>>>> with https://issues.apache.org/jira/browse/CASSANDRA-2376 >> >> >>> >>>>> The system was running fine for a couple of days when we >> >> >>> >>>>> started >> >> >>> >>>>> noticing >> >> >>> >>>>> something strange with cassandra. I stopped all applications >> >> >>> >>>>> and >> >> >>> >>>>> restarted >> >> >>> >>>>> cassandra. And then did a scrub. During scrub, I noticed >> >> >>> >>>>> these >> >> >>> >>>>> in >> >> >>> >>>>> the >> >> >>> >>>>> logs >> >> >>> >>>>> WARN [CompactionExecutor:1] 2011-04-24 23:37:07,561 >> >> >>> >>>>> CompactionManager.java >> >> >>> >>>>> (line 607) Non-fatal error reading row (stacktrace follows) >> >> >>> >>>>> java.io.IOError: java.io.IOException: Impossible row size >> >> >>> >>>>> 1516029079813320210 >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.CompactionManager.doScrub(CompactionManager.java:589) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.CompactionManager.access$600(CompactionManager.java:56) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.CompactionManager$3.call(CompactionManager.java:195) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) >> >> >>> >>>>> at >> >> >>> >>>>> java.util.concurrent.FutureTask.run(FutureTask.java:138) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) >> >> >>> >>>>> at java.lang.Thread.run(Thread.java:662) >> >> >>> >>>>> Caused by: java.io.IOException: Impossible row size >> >> >>> >>>>> 1516029079813320210 >> >> >>> >>>>> ... 8 more >> >> >>> >>>>> INFO [CompactionExecutor:1] 2011-04-24 23:37:07,640 >> >> >>> >>>>> CompactionManager.java >> >> >>> >>>>> (line 613) Retrying from row index; data is -1768177699 bytes >> >> >>> >>>>> starting >> >> >>> >>>>> at >> >> >>> >>>>> 2626524914 >> >> >>> >>>>> WARN [CompactionExecutor:1] 2011-04-24 23:37:07,641 >> >> >>> >>>>> CompactionManager.java >> >> >>> >>>>> (line 633) Retry failed too. Skipping to next row (retry's >> >> >>> >>>>> stacktrace >> >> >>> >>>>> follows) >> >> >>> >>>>> java.io.IOError: java.io.EOFException: bloom filter claims to >> >> >>> >>>>> be >> >> >>> >>>>> 1868982636 >> >> >>> >>>>> bytes, longer than entire row size -1768177699 at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.io.sstable.SSTableIdentityIterator.<init>(SSTableIdentityIterator.java:117) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.CompactionManager.doScrub(CompactionManager.java:618) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.CompactionManager.access$600(CompactionManager.java:56) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.CompactionManager$3.call(CompactionManager.java:195) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) >> >> >>> >>>>> at >> >> >>> >>>>> java.util.concurrent.FutureTask.run(FutureTask.java:138) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) >> >> >>> >>>>> at java.lang.Thread.run(Thread.java:662) >> >> >>> >>>>> Caused by: java.io.EOFException: bloom filter claims to be >> >> >>> >>>>> 1868982636 >> >> >>> >>>>> bytes, >> >> >>> >>>>> longer than entire row size -1768177699 at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.io.sstable.IndexHelper.defreezeBloomFilter(IndexHelper.java:116) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.io.sstable.SSTableIdentityIterator.<init>(SSTableIdentityIterator.java:87) >> >> >>> >>>>> ... 8 more >> >> >>> >>>>> WARN [CompactionExecutor:1] 2011-04-24 23:37:16,545 >> >> >>> >>>>> CompactionManager.java >> >> >>> >>>>> (line 607) Non-fatal error reading row (stacktrace follows) >> >> >>> >>>>> java.io.IOError: java.io.EOFException >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.io.sstable.SSTableIdentityIterator.next(SSTableIdentityIterator.java:144) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.io.sstable.SSTableIdentityIterator.next(SSTableIdentityIterator.java:40) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.commons.collections.iterators.CollatingIterator.set(CollatingIterator.java:284) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.commons.collections.iterators.CollatingIterator.least(CollatingIterator.java:326) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.commons.collections.iterators.CollatingIterator.next(CollatingIterator.java:230) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.utils.ReducingIterator.computeNext(ReducingIterator.java:68) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:136) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:131) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> com.google.common.collect.Iterators$7.computeNext(Iterators.java:604) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:136) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:131) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.ColumnIndexer.serializeInternal(ColumnIndexer.java:76) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.ColumnIndexer.serialize(ColumnIndexer.java:50) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.io.LazilyCompactedRow.<init>(LazilyCompactedRow.java:90) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.CompactionManager.getCompactedRow(CompactionManager.java:778) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.CompactionManager.doScrub(CompactionManager.java:591) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.CompactionManager.access$600(CompactionManager.java:56) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.CompactionManager$3.call(CompactionManager.java:195) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303) >> >> >>> >>>>> at >> >> >>> >>>>> java.util.concurrent.FutureTask.run(FutureTask.java:138) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) >> >> >>> >>>>> at java.lang.Thread.run(Thread.java:662) >> >> >>> >>>>> Caused by: java.io.EOFException >> >> >>> >>>>> at >> >> >>> >>>>> java.io.RandomAccessFile.readFully(RandomAccessFile.java:383) >> >> >>> >>>>> at >> >> >>> >>>>> java.io.RandomAccessFile.readFully(RandomAccessFile.java:361) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.io.util.BufferedRandomAccessFile.readBytes(BufferedRandomAccessFile.java:270) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.utils.ByteBufferUtil.read(ByteBufferUtil.java:317) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.utils.ByteBufferUtil.readWithLength(ByteBufferUtil.java:273) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.ColumnSerializer.deserialize(ColumnSerializer.java:94) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.ColumnSerializer.deserialize(ColumnSerializer.java:35) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.io.sstable.SSTableIdentityIterator.next(SSTableIdentityIterator.java:140) >> >> >>> >>>>> ... 22 more >> >> >>> >>>>> INFO [CompactionExecutor:1] 2011-04-24 23:37:16,561 >> >> >>> >>>>> CompactionManager.java >> >> >>> >>>>> (line 613) Retrying from row index; data is 78540539 bytes >> >> >>> >>>>> starting >> >> >>> >>>>> at >> >> >>> >>>>> 2229643127 >> >> >>> >>>>> >> >> >>> >>>>> And then when i restarted the readers, i get the following >> >> >>> >>>>> crash >> >> >>> >>>>> ERROR [ReadStage:24] 2011-04-24 23:43:05,658 >> >> >>> >>>>> AbstractCassandraDaemon.java >> >> >>> >>>>> (line 112) Fatal exception in thread >> >> >>> >>>>> Thread[ReadStage:24,5,main] >> >> >>> >>>>> java.lang.AssertionError: mmap segment underflow; remaining >> >> >>> >>>>> is >> >> >>> >>>>> 791462117 but >> >> >>> >>>>> 1970433058 requested >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.io.util.MappedFileDataInput.readBytes(MappedFileDataInput.java:119) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.utils.ByteBufferUtil.read(ByteBufferUtil.java:317) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.utils.ByteBufferUtil.readWithLength(ByteBufferUtil.java:273) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.ColumnSerializer.deserialize(ColumnSerializer.java:94) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.ColumnSerializer.deserialize(ColumnSerializer.java:35) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.columniterator.IndexedSliceReader$IndexedBlockFetcher.getNextBlock(IndexedSliceReader.java:181) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.columniterator.IndexedSliceReader.computeNext(IndexedSliceReader.java:121) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.columniterator.IndexedSliceReader.computeNext(IndexedSliceReader.java:49) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:136) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:131) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.columniterator.SSTableSliceIterator.hasNext(SSTableSliceIterator.java:108) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.commons.collections.iterators.CollatingIterator.set(CollatingIterator.java:283) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.commons.collections.iterators.CollatingIterator.least(CollatingIterator.java:326) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.commons.collections.iterators.CollatingIterator.next(CollatingIterator.java:230) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.utils.ReducingIterator.computeNext(ReducingIterator.java:68) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:136) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:131) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.filter.SliceQueryFilter.collectReducedColumns(SliceQueryFilter.java:116) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.filter.QueryFilter.collectCollatedColumns(QueryFilter.java:130) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.ColumnFamilyStore.getTopLevelColumns(ColumnFamilyStore.java:1368) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.ColumnFamilyStore.getColumnFamily(ColumnFamilyStore.java:1245) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.ColumnFamilyStore.getColumnFamily(ColumnFamilyStore.java:1173) >> >> >>> >>>>> at >> >> >>> >>>>> org.apache.cassandra.db.Table.getRow(Table.java:333) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.db.SliceFromReadCommand.getRow(SliceFromReadCommand.java:63) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.service.StorageProxy$LocalReadRunnable.runMayThrow(StorageProxy.java:453) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:30) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886) >> >> >>> >>>>> at >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> >> >> >>> >>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908) >> >> >>> >>>>> at java.lang.Thread.run(Thread.java:662) >> >> >>> >>>>> >> >> >>> >>>>> Any ideas? >> >> >>> >>>>> Thanks! >> >> >>> >>> >> >> >>> >> >> >> >>> >> >> >> >>> > >> >> >>> > >> >> >>> > >> >> >>> > -- >> >> >>> > Jonathan Ellis >> >> >>> > Project Chair, Apache Cassandra >> >> >>> > co-founder of DataStax, the source for professional Cassandra >> >> >>> > support >> >> >>> > http://www.datastax.com >> >> >> >> >> > >> >> > >> >> >> >> >> >> >> >> -- >> >> Jonathan Ellis >> >> Project Chair, Apache Cassandra >> >> co-founder of DataStax, the source for professional Cassandra support >> >> http://www.datastax.com >> > >> > >> >> >> >> -- >> Jonathan Ellis >> Project Chair, Apache Cassandra >> co-founder of DataStax, the source for professional Cassandra support >> http://www.datastax.com > > -- Jonathan Ellis Project Chair, Apache Cassandra co-founder of DataStax, the source for professional Cassandra support http://www.datastax.com