Hi Jonathan,

You are right I had 1 node 1.0.2 for some reason so I did the upgrade again.
I have now a 4 cluster upgraded to 1.0.3 but now I get the following error
on 2 nodes on the cluster:

ERROR [HintedHandoff:3] 2011-11-23 06:39:31,250 AbstractCassandraDaemon.java
(line 133) Fatal exception in thread Thread[HintedHandoff:3,1,main]
java.lang.AssertionError
        at
org.apache.cassandra.db.HintedHandOffManager.deliverHintsToEndpoint(HintedHa
ndOffManager.java:301)
        at
org.apache.cassandra.db.HintedHandOffManager.access$100(HintedHandOffManager
.java:81)
        at
org.apache.cassandra.db.HintedHandOffManager$2.runMayThrow(HintedHandOffMana
ger.java:353)
        at
org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:30)
        at
java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.ja
va:886)
        at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:9
08)
        at java.lang.Thread.run(Thread.java:619)

if I did cassandra repair on all the servers on the cluster can I simply
truncate the column family HintsColumnFamily?
Will I lose any data?

Thanks
Michael

-----Original Message-----
From: Jonathan Ellis [mailto:jbel...@gmail.com] 
Sent: Monday, November 21, 2011 8:40 PM
To: user@cassandra.apache.org
Cc: cassandra-u...@incubator.apache.org
Subject: Re: Upgrade Cassandra Cluster to 1.0.3

Sounds to me like
https://issues.apache.org/jira/browse/CASSANDRA-3491, which was
present in 1.0.2 and fixed in 1.0.3.  It sounds like you're running
the wrong version by mistake.

On Mon, Nov 21, 2011 at 7:53 AM, Michael Vaknine <micha...@citypath.com>
wrote:
> Hi,
> Any help will be appreciated.
>
> I am upgrading Cassandra 1.0.0 to 1.0.3 got error
> ERROR [CompactionExecutor:3] 2011-11-21 11:10:59,075
> AbstractCassandraDaemon.java (line 133) Fatal exception in thread
> Thread[CompactionExecutor:     3,1,main]
> 3645 java.lang.StackOverflowError
> 3646     at com.google.common.base.Objects.equal(Objects.java:51)
> 3647     at org.apache.cassandra.utils.Pair.equals(Pair.java:48)
> 3648     at
>
java.util.concurrent.ConcurrentHashMap$Segment.get(ConcurrentHashMap.java:33
> 8)
> 3649     at
> java.util.concurrent.ConcurrentHashMap.get(ConcurrentHashMap.java:769)
> 3650     at
>
com.googlecode.concurrentlinkedhashmap.ConcurrentLinkedHashMap.get(Concurren
> tLinkedHashMap.java:740)
> 3651     at
>
org.apache.cassandra.cache.ConcurrentLinkedHashCache.get(ConcurrentLinkedHas
> hCache.java:81)
> 3652     at
>
org.apache.cassandra.cache.InstrumentingCache.get(InstrumentingCache.java:68
> )
> 3653     at
>
org.apache.cassandra.io.sstable.SSTableReader.getCachedPosition(SSTableReade
> r.java:598)
> 3654     at
>
org.apache.cassandra.io.sstable.SSTableReader.getPosition(SSTableReader.java
> :621)
> 3655     at
>
org.apache.cassandra.io.sstable.SSTableReader.getFileDataInput(SSTableReader
> .java:786)
> 3656     at
>
org.apache.cassandra.db.columniterator.SSTableNamesIterator.<init>(SSTableNa
> mesIterator.java:61)
>
> I tried to increase Xss to 640M and then got this error
>
> ERROR [Thread-28] 2011-11-21 12:52:40,808 AbstractCassandraDaemon.java
(line
> 133) Fatal exception in thread Thread[Thread-28,5,main]
> 4940 java.lang.RuntimeException: java.util.concurrent.ExecutionException:
> java.lang.Error: Maximum lock count exceeded
> 4941     at
>
org.apache.cassandra.db.index.SecondaryIndexManager.maybeBuildSecondaryIndex
> es(SecondaryIndexManager.java:131)
> 4942     at
>
org.apache.cassandra.streaming.StreamInSession.closeIfFinished(StreamInSessi
> on.java:151)
> 4943     at
>
org.apache.cassandra.streaming.IncomingStreamReader.read(IncomingStreamReade
> r.java:102)
> 4944     at
>
org.apache.cassandra.net.IncomingTcpConnection.stream(IncomingTcpConnection.
> java:184)
> 4945     at
>
org.apache.cassandra.net.IncomingTcpConnection.run(IncomingTcpConnection.jav
> a:81)
> 4946 Caused by: java.util.concurrent.ExecutionException: java.lang.Error:
> Maximum lock count exceeded
> 4947     at
> java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:222)
> 4948     at java.util.concurrent.FutureTask.get(FutureTask.java:83)
> 4949     at
>
org.apache.cassandra.db.index.SecondaryIndexManager.maybeBuildSecondaryIndex
> es(SecondaryIndexManager.java:122)
> 4950     ... 4 more
> 4951 Caused by: java.lang.Error: Maximum lock count exceeded
>
> None of this errors happen on 1.0.0
>
>
> Thanks
> Michael
>
>



-- 
Jonathan Ellis
Project Chair, Apache Cassandra
co-founder of DataStax, the source for professional Cassandra support
http://www.datastax.com

Reply via email to