This looks like https://issues.apache.org/jira/browse/CASSANDRA-6959, but that was fixed for 2.1.0-rc1.
Is there any chance you can put together a script to reproduce the issue? On Thu, Jul 10, 2014 at 8:51 AM, Pavel Kogan <pavel.ko...@cortica.com> wrote: > It seems that memtable tries to flush itself to SSTable of not existing > keyspace. I don't know why it is happens, but probably running nodetool > flush before drop should prevent this issue. > > Pavel > > > On Thu, Jul 10, 2014 at 4:09 AM, Fabrice Larcher < > fabrice.larc...@level5.fr> wrote: > >> ​Hello, >> >> I am using the 'development' version 2.1-rc2. >> >> With one node (=localhost), I get timeouts trying to connect to C* after >> running a 'DROP KEYSPACE' command. I have following error messages in >> system.log : >> >> INFO [SharedPool-Worker-3] 2014-07-09 16:29:36,578 >> MigrationManager.java:319 - Drop Keyspace 'test_main' >> (...) >> ERROR [MemtableFlushWriter:6] 2014-07-09 16:29:37,178 >> CassandraDaemon.java:166 - Exception in thread >> Thread[MemtableFlushWriter:6,5,main] >> java.lang.RuntimeException: Last written key >> DecoratedKey(91e7f660-076f-11e4-a36d-28d2444c0b1b, >> 52446dde90244ca49789b41671e4ca7c) >= current key >> DecoratedKey(91e7f660-076f-11e4-a36d-28d2444c0b1b, >> 52446dde90244ca49789b41671e4ca7c) writing into >> ./../data/data/test_main/user-911d5360076f11e4812d3d4ba97474ac/test_main-user.user_account-tmp-ka-1-Data.db >> at >> org.apache.cassandra.io.sstable.SSTableWriter.beforeAppend(SSTableWriter.java:172) >> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> org.apache.cassandra.io.sstable.SSTableWriter.append(SSTableWriter.java:215) >> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> org.apache.cassandra.db.Memtable$FlushRunnable.writeSortedContents(Memtable.java:351) >> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> org.apache.cassandra.db.Memtable$FlushRunnable.runWith(Memtable.java:314) >> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> org.apache.cassandra.io.util.DiskAwareRunnable.runMayThrow(DiskAwareRunnable.java:48) >> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) >> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> com.google.common.util.concurrent.MoreExecutors$SameThreadExecutorService.execute(MoreExecutors.java:297) >> ~[guava-16.0.jar:na] >> at >> org.apache.cassandra.db.ColumnFamilyStore$Flush.run(ColumnFamilyStore.java:1054) >> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >> ~[na:1.7.0_55] >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >> ~[na:1.7.0_55] >> at java.lang.Thread.run(Thread.java:744) ~[na:1.7.0_55] >> >> Then, I can not connect to the Cluster anymore from my app (Java Driver >> 2.1-SNAPSHOT) and got in application logs : >> >> com.datastax.driver.core.exceptions.NoHostAvailableException: All host(s) >> tried for query failed (tried: /127.0.0.1:9042 >> (com.datastax.driver.core.exceptions.DriverException: Timeout during read)) >> at >> com.datastax.driver.core.exceptions.NoHostAvailableException.copy(NoHostAvailableException.java:65) >> at >> com.datastax.driver.core.DefaultResultSetFuture.extractCauseFromExecutionException(DefaultResultSetFuture.java:258) >> at >> com.datastax.driver.core.DefaultResultSetFuture.getUninterruptibly(DefaultResultSetFuture.java:174) >> at >> com.datastax.driver.core.AbstractSession.execute(AbstractSession.java:52) >> at >> com.datastax.driver.core.AbstractSession.execute(AbstractSession.java:36) >> (...) >> Caused by: com.datastax.driver.core.exceptions.NoHostAvailableException: >> All host(s) tried for query failed (tried: /127.0.0.1:9042 >> (com.datastax.driver.core.exceptions.DriverException: Timeout during read)) >> at >> com.datastax.driver.core.RequestHandler.sendRequest(RequestHandler.java:103) >> at >> com.datastax.driver.core.RequestHandler$1.run(RequestHandler.java:175) >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >> at java.lang.Thread.run(Thread.java:745) >> >> I can still connect through CQLSH but if I run (again) a "DROP KEYSPACE" >> command from CQLSH, I get the following error : >> errors={}, last_host=127.0.0.1 >> >> Now, on a 2 nodes cluster I also have a similar issue but the error's >> stacktrace is different : >> >> From application logs : >> >> 17971 [Cassandra Java Driver worker-2] WARN >> com.datastax.driver.core.Cluster - No schema agreement from live replicas >> after 10000 ms. The schema may not be up to date on some nodes. >> >> From system.log : >> >> INFO [SharedPool-Worker-2] 2014-07-10 09:04:53,434 >> MigrationManager.java:319 - Drop Keyspace 'test_main' >> (...) >> ERROR [MigrationStage:1] 2014-07-10 09:04:56,553 >> CommitLogSegmentManager.java:304 - Failed waiting for a forced recycle of >> in-use commit log segments >> java.lang.AssertionError: null >> at >> org.apache.cassandra.db.commitlog.CommitLogSegmentManager.forceRecycleAll(CommitLogSegmentManager.java:299) >> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> org.apache.cassandra.db.commitlog.CommitLog.forceRecycleAllSegments(CommitLog.java:160) >> [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> org.apache.cassandra.db.DefsTables.dropColumnFamily(DefsTables.java:516) >> [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> org.apache.cassandra.db.DefsTables.mergeColumnFamilies(DefsTables.java:300) >> [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> org.apache.cassandra.db.DefsTables.mergeSchemaInternal(DefsTables.java:187) >> [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> org.apache.cassandra.db.DefsTables.mergeSchema(DefsTables.java:164) >> [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> org.apache.cassandra.service.MigrationManager$2.runMayThrow(MigrationManager.java:393) >> [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) >> [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >> [na:1.7.0_45] >> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >> [na:1.7.0_45] >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >> [na:1.7.0_45] >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >> [na:1.7.0_45] >> at java.lang.Thread.run(Thread.java:744) [na:1.7.0_45] >> >> >> In both cases, I have to restart the C* server, so that I can connect >> again from my application. The connection from CQLSH is still possible but >> some key-space related commands fail. >> >> NB : I also get sometimes the following error on my nodes (I do not know >> if there is any link with the topic) : >> ERROR [CompactionExecutor:62] 2014-07-10 08:49:02,059 >> CassandraDaemon.java:166 - Exception in thread >> Thread[CompactionExecutor:62,1,main] >> java.lang.NullPointerException: null >> at >> org.apache.cassandra.cache.CounterCacheKey.getPathInfo(CounterCacheKey.java:53) >> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> org.apache.cassandra.cache.AutoSavingCache$Writer.saveCache(AutoSavingCache.java:215) >> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> org.apache.cassandra.db.compaction.CompactionManager$11.run(CompactionManager.java:1052) >> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >> at >> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >> ~[na:1.7.0_55] >> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >> ~[na:1.7.0_55] >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >> ~[na:1.7.0_55] >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >> [na:1.7.0_55] >> at java.lang.Thread.run(Thread.java:744) [na:1.7.0_55] >> >> It seems to be a bug (is it known ?). Or do you have any clue on how I >> can solve that ? >> >> Thanks for your help >> >> Fabrice >> > > -- Tyler Hobbs DataStax <http://datastax.com/>