Hi Vaibhav, thanks for reaching out.Based on my understanding of this exception, this may be due to the index for this partition 
exceeding 2GiB (which is *extremely* large for a partition index component).Reducing the size of the column index below 2GiB may 
resolve this issue. You may be able to do so by increasing the size of "column_index_size_in_kb" in cassandra.yaml (which 
defines the granularity of this index). This may enable the compaction to complete - e.g., by swapping the default of "64" 
with "128".Docs for this param are in cassandra.yaml:# Granularity of the collation index of rows within a partition.# 
Increase if your rows are large, or if you have a very large# number of rows per partition.  The competing goals are these:## - a 
smaller granularity means more index entries are generated#   and looking up rows withing the partition by collation column#   is 
faster# - but, Cassandra will keep the collation index in memory for hot#   rows (as part of the key cache), so a larger granularity 
means#   you can cache more hot rowsHowever, the root of this issue is likely a partition whose size has become extraordinary. Would 
recommend determining what has resulted in an individual partition growing so large that its index has exceeded 2GiB and determining 
if it can be removed or if the data model for the table can be adjusted to avoid such a large number of rows being stored within one 
partition.– ScottOn Jan 13, 2023, at 9:24 PM, vaibhav khedkar <vkhedk...@gmail.com> wrote:Hello All, We are facing an issue 
where few of the nodes are not able to complete compactions. We tried restarting, scrubbing and even rebuilding an entire node but 
nothing seems to work so far. It's a 10 Region installation with close to 150 nodes. Datatax support suggested rebuilding the node 
but that did not help. Any help is appreciated. Following is the logtrace. ERROR [CompactionExecutor:50] 2023-01-14 05:12:20,795 
CassandraDaemon.java:581 - Exception in thread 
Thread[CompactionExecutor:50,1,main]org.apache.cassandra.db.rows.PartitionSerializationException: Failed to serialize partition key 
'<key>'  on table '<table>' in keyspace '<keyspace>'. at 
org.apache.cassandra.io.sstable.format.big.BigTableWriter.append(BigTableWriter.java:240) at 
org.apache.cassandra.io.sstable.SSTableRewriter.append(SSTableRewriter.java:125) at 
org.apache.cassandra.db.compaction.writers.MaxSSTableSizeWriter.realAppend(MaxSSTableSizeWriter.java:84) at 
org.apache.cassandra.db.compaction.writers.CompactionAwareWriter.append(CompactionAwareWriter.java:137) at 
org.apache.cassandra.db.compaction.CompactionTask.runMayThrow(CompactionTask.java:193) at 
org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) at 
org.apache.cassandra.db.compaction.CompactionTask.executeInternal(CompactionTask.java:77) at 
org.apache.cassandra.db.compaction.AbstractCompactionTask.execute(AbstractCompactionTask.java:100) at 
org.apache.cassandra.db.compaction.CompactionManager$BackgroundCompactionCandidate.run(CompactionManager.java:298) at 
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) 
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at 
io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) at java.lang.Thread.run(Thread.java:748)Caused 
by: java.nio.BufferOverflowException: null at 
org.apache.cassandra.io.util.DataOutputBuffer.validateReallocation(DataOutputBuffer.java:136) at 
org.apache.cassandra.io.util.DataOutputBuffer.calculateNewSize(DataOutputBuffer.java:154) at 
org.apache.cassandra.io.util.DataOutputBuffer.expandToFit(DataOutputBuffer.java:161) at 
org.apache.cassandra.io.util.DataOutputBuffer.doFlush(DataOutputBuffer.java:121) at 
org.apache.cassandra.io.util.BufferedDataOutputStreamPlus.write(BufferedDataOutputStreamPlus.java:121) at 
org.apache.cassandra.io.util.BufferedDataOutputStreamPlus.write(BufferedDataOutputStreamPlus.java:93) at 
org.apache.cassandra.db.marshal.ByteArrayAccessor.write(ByteArrayAccessor.java:61) at 
org.apache.cassandra.db.marshal.ByteArrayAccessor.write(ByteArrayAccessor.java:38) at 
org.apache.cassandra.db.marshal.ValueAccessor.writeWithVIntLength(ValueAccessor.java:164) at 
org.apache.cassandra.db.marshal.AbstractType.writeValue(AbstractType.java:451) at 
org.apache.cassandra.db.ClusteringPrefix$Serializer.serializeValuesWithoutSize(ClusteringPrefix.java:397) at 
org.apache.cassandra.db.Clustering$Serializer.serialize(Clustering.java:132) at 
org.apache.cassandra.db.ClusteringPrefix$Serializer.serialize(ClusteringPrefix.java:339) at 
org.apache.cassandra.io.sstable.IndexInfo$Serializer.serialize(IndexInfo.java:110) at 
org.apache.cassandra.io.sstable.IndexInfo$Serializer.serialize(IndexInfo.java:91) at 
org.apache.cassandra.db.ColumnIndex.addIndexBlock(ColumnIndex.java:223) at 
org.apache.cassandra.db.ColumnIndex.add(ColumnIndex.java:271) at 
org.apache.cassandra.db.ColumnIndex.buildRowIndex(ColumnIndex.java:118) at 
org.apache.cassandra.io.sstable.format.big.BigTableWriter.append(BigTableWriter.java:216) ... 14 common frames omittedThanksvaibhav

Reply via email to