Re: get_range_slices OOM on CompressionMetadata.readChunkOffsets(..)

2011-10-31 Thread Jonathan Ellis
Cleanup would have the same effect I think, in exchange for a minor amount of extra CPU used. On Mon, Oct 31, 2011 at 4:08 AM, Sylvain Lebresne wrote: > On Mon, Oct 31, 2011 at 9:07 AM, Mick Semb Wever wrote: >> On Mon, 2011-10-31 at 08:00 +0100, Mick Semb Wever wrote: >>> After an upgrade to ca

Re: get_range_slices OOM on CompressionMetadata.readChunkOffsets(..)

2011-10-31 Thread Sylvain Lebresne
On Mon, Oct 31, 2011 at 11:41 AM, Mick Semb Wever wrote: > On Mon, 2011-10-31 at 10:08 +0100, Sylvain Lebresne wrote: >> you can >> trigger a "user defined compaction" through JMX on each of the sstable >> you want to rebuild. > > May i ask how? > Everything i see from NodeProbe to StorageProxy is

Re: get_range_slices OOM on CompressionMetadata.readChunkOffsets(..)

2011-10-31 Thread Sylvain Lebresne
On Mon, Oct 31, 2011 at 11:35 AM, Mick Semb Wever wrote: > On Mon, 2011-10-31 at 10:08 +0100, Sylvain Lebresne wrote: >> >> I set chunk_length_kb to 16 as my rows are very skinny (typically 100b) >> > >> > >> > I see now this was a bad choice. >> > The read pattern of these rows is always in bulk

Re: get_range_slices OOM on CompressionMetadata.readChunkOffsets(..)

2011-10-31 Thread Mick Semb Wever
On Mon, 2011-10-31 at 10:08 +0100, Sylvain Lebresne wrote: > you can > trigger a "user defined compaction" through JMX on each of the sstable > you want to rebuild. May i ask how? Everything i see from NodeProbe to StorageProxy is ks and cf based. ~mck -- “Anyone who lives within their means s

Re: get_range_slices OOM on CompressionMetadata.readChunkOffsets(..)

2011-10-31 Thread Mick Semb Wever
On Mon, 2011-10-31 at 10:08 +0100, Sylvain Lebresne wrote: > >> I set chunk_length_kb to 16 as my rows are very skinny (typically 100b) > > > > > > I see now this was a bad choice. > > The read pattern of these rows is always in bulk so the chunk_length > > could have been much higher so to reduce

Re: get_range_slices OOM on CompressionMetadata.readChunkOffsets(..)

2011-10-31 Thread Sylvain Lebresne
On Mon, Oct 31, 2011 at 9:07 AM, Mick Semb Wever wrote: > On Mon, 2011-10-31 at 08:00 +0100, Mick Semb Wever wrote: >> After an upgrade to cassandra-1.0 any get_range_slices gives me: >> >> java.lang.OutOfMemoryError: Java heap space >>       at >> org.apache.cassandra.io.compress.CompressionMeta

Re: get_range_slices OOM on CompressionMetadata.readChunkOffsets(..)

2011-10-31 Thread Mick Semb Wever
On Mon, 2011-10-31 at 08:00 +0100, Mick Semb Wever wrote: > After an upgrade to cassandra-1.0 any get_range_slices gives me: > > java.lang.OutOfMemoryError: Java heap space > at > org.apache.cassandra.io.compress.CompressionMetadata.readChunkOffsets(CompressionMetadata.java:93) > at

get_range_slices OOM on CompressionMetadata.readChunkOffsets(..)

2011-10-31 Thread Mick Semb Wever
After an upgrade to cassandra-1.0 any get_range_slices gives me: java.lang.OutOfMemoryError: Java heap space at org.apache.cassandra.io.compress.CompressionMetadata.readChunkOffsets(CompressionMetadata.java:93) at org.apache.cassandra.io.compress.CompressionMetadata.(CompressionM