Hi Jonathan, You are right I had 1 node 1.0.2 for some reason so I did the upgrade again. I have now a 4 cluster upgraded to 1.0.3 but now I get the following error on 2 nodes on the cluster:
ERROR [HintedHandoff:3] 2011-11-23 06:39:31,250 AbstractCassandraDaemon.java (line 133) Fatal exception in thread Thread[HintedHandoff:3,1,main] java.lang.AssertionError at org.apache.cassandra.db.HintedHandOffManager.deliverHintsToEndpoint(HintedHa ndOffManager.java:301) at org.apache.cassandra.db.HintedHandOffManager.access$100(HintedHandOffManager .java:81) at org.apache.cassandra.db.HintedHandOffManager$2.runMayThrow(HintedHandOffMana ger.java:353) at org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:30) at java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.ja va:886) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:9 08) at java.lang.Thread.run(Thread.java:619) if I did cassandra repair on all the servers on the cluster can I simply truncate the column family HintsColumnFamily? Will I lose any data? Thanks Michael -----Original Message----- From: Jonathan Ellis [mailto:jbel...@gmail.com] Sent: Monday, November 21, 2011 8:40 PM To: user@cassandra.apache.org Cc: cassandra-u...@incubator.apache.org Subject: Re: Upgrade Cassandra Cluster to 1.0.3 Sounds to me like https://issues.apache.org/jira/browse/CASSANDRA-3491, which was present in 1.0.2 and fixed in 1.0.3. It sounds like you're running the wrong version by mistake. On Mon, Nov 21, 2011 at 7:53 AM, Michael Vaknine <micha...@citypath.com> wrote: > Hi, > Any help will be appreciated. > > I am upgrading Cassandra 1.0.0 to 1.0.3 got error > ERROR [CompactionExecutor:3] 2011-11-21 11:10:59,075 > AbstractCassandraDaemon.java (line 133) Fatal exception in thread > Thread[CompactionExecutor: 3,1,main] > 3645 java.lang.StackOverflowError > 3646 at com.google.common.base.Objects.equal(Objects.java:51) > 3647 at org.apache.cassandra.utils.Pair.equals(Pair.java:48) > 3648 at > java.util.concurrent.ConcurrentHashMap$Segment.get(ConcurrentHashMap.java:33 > 8) > 3649 at > java.util.concurrent.ConcurrentHashMap.get(ConcurrentHashMap.java:769) > 3650 at > com.googlecode.concurrentlinkedhashmap.ConcurrentLinkedHashMap.get(Concurren > tLinkedHashMap.java:740) > 3651 at > org.apache.cassandra.cache.ConcurrentLinkedHashCache.get(ConcurrentLinkedHas > hCache.java:81) > 3652 at > org.apache.cassandra.cache.InstrumentingCache.get(InstrumentingCache.java:68 > ) > 3653 at > org.apache.cassandra.io.sstable.SSTableReader.getCachedPosition(SSTableReade > r.java:598) > 3654 at > org.apache.cassandra.io.sstable.SSTableReader.getPosition(SSTableReader.java > :621) > 3655 at > org.apache.cassandra.io.sstable.SSTableReader.getFileDataInput(SSTableReader > .java:786) > 3656 at > org.apache.cassandra.db.columniterator.SSTableNamesIterator.<init>(SSTableNa > mesIterator.java:61) > > I tried to increase Xss to 640M and then got this error > > ERROR [Thread-28] 2011-11-21 12:52:40,808 AbstractCassandraDaemon.java (line > 133) Fatal exception in thread Thread[Thread-28,5,main] > 4940 java.lang.RuntimeException: java.util.concurrent.ExecutionException: > java.lang.Error: Maximum lock count exceeded > 4941 at > org.apache.cassandra.db.index.SecondaryIndexManager.maybeBuildSecondaryIndex > es(SecondaryIndexManager.java:131) > 4942 at > org.apache.cassandra.streaming.StreamInSession.closeIfFinished(StreamInSessi > on.java:151) > 4943 at > org.apache.cassandra.streaming.IncomingStreamReader.read(IncomingStreamReade > r.java:102) > 4944 at > org.apache.cassandra.net.IncomingTcpConnection.stream(IncomingTcpConnection. > java:184) > 4945 at > org.apache.cassandra.net.IncomingTcpConnection.run(IncomingTcpConnection.jav > a:81) > 4946 Caused by: java.util.concurrent.ExecutionException: java.lang.Error: > Maximum lock count exceeded > 4947 at > java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:222) > 4948 at java.util.concurrent.FutureTask.get(FutureTask.java:83) > 4949 at > org.apache.cassandra.db.index.SecondaryIndexManager.maybeBuildSecondaryIndex > es(SecondaryIndexManager.java:122) > 4950 ... 4 more > 4951 Caused by: java.lang.Error: Maximum lock count exceeded > > None of this errors happen on 1.0.0 > > > Thanks > Michael > > -- Jonathan Ellis Project Chair, Apache Cassandra co-founder of DataStax, the source for professional Cassandra support http://www.datastax.com