It would be nice if you can try with 2.1.0-rc5 (there has been quite a bit of bug fixes since rc2). If you can still reproduce that NPE there, please do open a jira ticket with the reproduction steps.
On Thu, Aug 7, 2014 at 11:29 AM, Fabrice Larcher <fabrice.larc...@level5.fr> wrote: > Hello, > > After a 'DROP TABLE' command that returns "errors={}, last_host=127.0.0.1" > (like most DROP commands do) from CQLSH with C* 2.1.0-rc2, I stopped C*. > And I can not start one node. It says : > ERROR 09:18:34 Exception encountered during startup > java.lang.NullPointerException: null > at org.apache.cassandra.db.Directories.<init>(Directories.java:191) > ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] > at > org.apache.cassandra.db.ColumnFamilyStore.removeUnfinishedCompactionLeftovers(ColumnFamilyStore.java:553) > ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] > at > org.apache.cassandra.service.CassandraDaemon.setup(CassandraDaemon.java:245) > [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] > at > org.apache.cassandra.service.CassandraDaemon.activate(CassandraDaemon.java:455) > [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] > at > org.apache.cassandra.service.CassandraDaemon.main(CassandraDaemon.java:544) > [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] > java.lang.NullPointerException > at org.apache.cassandra.db.Directories.<init>(Directories.java:191) > at > org.apache.cassandra.db.ColumnFamilyStore.removeUnfinishedCompactionLeftovers(ColumnFamilyStore.java:553) > at > org.apache.cassandra.service.CassandraDaemon.setup(CassandraDaemon.java:245) > at > org.apache.cassandra.service.CassandraDaemon.activate(CassandraDaemon.java:455) > at > org.apache.cassandra.service.CassandraDaemon.main(CassandraDaemon.java:544) > Exception encountered during startup: null > > I do not now if it can help. > > > Fabrice LARCHER > > > 2014-07-18 7:23 GMT+02:00 Fabrice Larcher <fabrice.larc...@level5.fr>: > > Hello, >> >> I still experience a similar issue after a 'DROP KEYSPACE' command with >> C* 2.1-rc3. Connection to the node may fail after a 'DROP'. >> >> But I did not see this issue with 2.1-rc1 (-> it seems like to be a >> regression brought with 2.1-rc2). >> >> Fabrice LARCHER >> >> >> 2014-07-17 9:19 GMT+02:00 Benedict Elliott Smith < >> belliottsm...@datastax.com>: >> >> Also https://issues.apache.org/jira/browse/CASSANDRA-7437 and >>> https://issues.apache.org/jira/browse/CASSANDRA-7465 for rc3, although >>> the CounterCacheKey assertion looks like an independent (though >>> comparatively benign) bug I will file a ticket for. >>> >>> Can you try this against rc3 to see if the problem persists? You may see >>> the last exception, but it shouldn't affect the stability of the cluster. >>> If either of the other exceptions persist, please file a ticket. >>> >>> >>> On Thu, Jul 17, 2014 at 1:41 AM, Tyler Hobbs <ty...@datastax.com> wrote: >>> >>>> This looks like https://issues.apache.org/jira/browse/CASSANDRA-6959, >>>> but that was fixed for 2.1.0-rc1. >>>> >>>> Is there any chance you can put together a script to reproduce the >>>> issue? >>>> >>>> >>>> On Thu, Jul 10, 2014 at 8:51 AM, Pavel Kogan <pavel.ko...@cortica.com> >>>> wrote: >>>> >>>>> It seems that memtable tries to flush itself to SSTable of not >>>>> existing keyspace. I don't know why it is happens, but probably running >>>>> nodetool flush before drop should prevent this issue. >>>>> >>>>> Pavel >>>>> >>>>> >>>>> On Thu, Jul 10, 2014 at 4:09 AM, Fabrice Larcher < >>>>> fabrice.larc...@level5.fr> wrote: >>>>> >>>>>> ​Hello, >>>>>> >>>>>> I am using the 'development' version 2.1-rc2. >>>>>> >>>>>> With one node (=localhost), I get timeouts trying to connect to C* >>>>>> after running a 'DROP KEYSPACE' command. I have following error messages >>>>>> in >>>>>> system.log : >>>>>> >>>>>> INFO [SharedPool-Worker-3] 2014-07-09 16:29:36,578 >>>>>> MigrationManager.java:319 - Drop Keyspace 'test_main' >>>>>> (...) >>>>>> ERROR [MemtableFlushWriter:6] 2014-07-09 16:29:37,178 >>>>>> CassandraDaemon.java:166 - Exception in thread >>>>>> Thread[MemtableFlushWriter:6,5,main] >>>>>> java.lang.RuntimeException: Last written key >>>>>> DecoratedKey(91e7f660-076f-11e4-a36d-28d2444c0b1b, >>>>>> 52446dde90244ca49789b41671e4ca7c) >= current key >>>>>> DecoratedKey(91e7f660-076f-11e4-a36d-28d2444c0b1b, >>>>>> 52446dde90244ca49789b41671e4ca7c) writing into >>>>>> ./../data/data/test_main/user-911d5360076f11e4812d3d4ba97474ac/test_main-user.user_account-tmp-ka-1-Data.db >>>>>> at >>>>>> org.apache.cassandra.io.sstable.SSTableWriter.beforeAppend(SSTableWriter.java:172) >>>>>> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> org.apache.cassandra.io.sstable.SSTableWriter.append(SSTableWriter.java:215) >>>>>> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> org.apache.cassandra.db.Memtable$FlushRunnable.writeSortedContents(Memtable.java:351) >>>>>> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> org.apache.cassandra.db.Memtable$FlushRunnable.runWith(Memtable.java:314) >>>>>> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> org.apache.cassandra.io.util.DiskAwareRunnable.runMayThrow(DiskAwareRunnable.java:48) >>>>>> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) >>>>>> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> com.google.common.util.concurrent.MoreExecutors$SameThreadExecutorService.execute(MoreExecutors.java:297) >>>>>> ~[guava-16.0.jar:na] >>>>>> at >>>>>> org.apache.cassandra.db.ColumnFamilyStore$Flush.run(ColumnFamilyStore.java:1054) >>>>>> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>>>> ~[na:1.7.0_55] >>>>>> at >>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>>>> ~[na:1.7.0_55] >>>>>> at java.lang.Thread.run(Thread.java:744) ~[na:1.7.0_55] >>>>>> >>>>>> Then, I can not connect to the Cluster anymore from my app (Java >>>>>> Driver 2.1-SNAPSHOT) and got in application logs : >>>>>> >>>>>> com.datastax.driver.core.exceptions.NoHostAvailableException: All >>>>>> host(s) tried for query failed (tried: /127.0.0.1:9042 >>>>>> (com.datastax.driver.core.exceptions.DriverException: Timeout during >>>>>> read)) >>>>>> at >>>>>> com.datastax.driver.core.exceptions.NoHostAvailableException.copy(NoHostAvailableException.java:65) >>>>>> at >>>>>> com.datastax.driver.core.DefaultResultSetFuture.extractCauseFromExecutionException(DefaultResultSetFuture.java:258) >>>>>> at >>>>>> com.datastax.driver.core.DefaultResultSetFuture.getUninterruptibly(DefaultResultSetFuture.java:174) >>>>>> at >>>>>> com.datastax.driver.core.AbstractSession.execute(AbstractSession.java:52) >>>>>> at >>>>>> com.datastax.driver.core.AbstractSession.execute(AbstractSession.java:36) >>>>>> (...) >>>>>> Caused by: >>>>>> com.datastax.driver.core.exceptions.NoHostAvailableException: All host(s) >>>>>> tried for query failed (tried: /127.0.0.1:9042 >>>>>> (com.datastax.driver.core.exceptions.DriverException: Timeout during >>>>>> read)) >>>>>> at >>>>>> com.datastax.driver.core.RequestHandler.sendRequest(RequestHandler.java:103) >>>>>> at >>>>>> com.datastax.driver.core.RequestHandler$1.run(RequestHandler.java:175) >>>>>> at >>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>>>> at >>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>>>> at java.lang.Thread.run(Thread.java:745) >>>>>> >>>>>> I can still connect through CQLSH but if I run (again) a "DROP >>>>>> KEYSPACE" command from CQLSH, I get the following error : >>>>>> errors={}, last_host=127.0.0.1 >>>>>> >>>>>> Now, on a 2 nodes cluster I also have a similar issue but the error's >>>>>> stacktrace is different : >>>>>> >>>>>> From application logs : >>>>>> >>>>>> 17971 [Cassandra Java Driver worker-2] WARN >>>>>> com.datastax.driver.core.Cluster - No schema agreement from live >>>>>> replicas >>>>>> after 10000 ms. The schema may not be up to date on some nodes. >>>>>> >>>>>> From system.log : >>>>>> >>>>>> INFO [SharedPool-Worker-2] 2014-07-10 09:04:53,434 >>>>>> MigrationManager.java:319 - Drop Keyspace 'test_main' >>>>>> (...) >>>>>> ERROR [MigrationStage:1] 2014-07-10 09:04:56,553 >>>>>> CommitLogSegmentManager.java:304 - Failed waiting for a forced recycle of >>>>>> in-use commit log segments >>>>>> java.lang.AssertionError: null >>>>>> at >>>>>> org.apache.cassandra.db.commitlog.CommitLogSegmentManager.forceRecycleAll(CommitLogSegmentManager.java:299) >>>>>> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> org.apache.cassandra.db.commitlog.CommitLog.forceRecycleAllSegments(CommitLog.java:160) >>>>>> [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> org.apache.cassandra.db.DefsTables.dropColumnFamily(DefsTables.java:516) >>>>>> [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> org.apache.cassandra.db.DefsTables.mergeColumnFamilies(DefsTables.java:300) >>>>>> [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> org.apache.cassandra.db.DefsTables.mergeSchemaInternal(DefsTables.java:187) >>>>>> [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> org.apache.cassandra.db.DefsTables.mergeSchema(DefsTables.java:164) >>>>>> [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> org.apache.cassandra.service.MigrationManager$2.runMayThrow(MigrationManager.java:393) >>>>>> [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> org.apache.cassandra.utils.WrappedRunnable.run(WrappedRunnable.java:28) >>>>>> [apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >>>>>> [na:1.7.0_45] >>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>>>> [na:1.7.0_45] >>>>>> at >>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>>>> [na:1.7.0_45] >>>>>> at >>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>>>> [na:1.7.0_45] >>>>>> at java.lang.Thread.run(Thread.java:744) [na:1.7.0_45] >>>>>> >>>>>> >>>>>> In both cases, I have to restart the C* server, so that I can connect >>>>>> again from my application. The connection from CQLSH is still possible >>>>>> but >>>>>> some key-space related commands fail. >>>>>> >>>>>> NB : I also get sometimes the following error on my nodes (I do not >>>>>> know if there is any link with the topic) : >>>>>> ERROR [CompactionExecutor:62] 2014-07-10 08:49:02,059 >>>>>> CassandraDaemon.java:166 - Exception in thread >>>>>> Thread[CompactionExecutor:62,1,main] >>>>>> java.lang.NullPointerException: null >>>>>> at >>>>>> org.apache.cassandra.cache.CounterCacheKey.getPathInfo(CounterCacheKey.java:53) >>>>>> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> org.apache.cassandra.cache.AutoSavingCache$Writer.saveCache(AutoSavingCache.java:215) >>>>>> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> org.apache.cassandra.db.compaction.CompactionManager$11.run(CompactionManager.java:1052) >>>>>> ~[apache-cassandra-2.1.0-rc2.jar:2.1.0-rc2] >>>>>> at >>>>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >>>>>> ~[na:1.7.0_55] >>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>>>> ~[na:1.7.0_55] >>>>>> at >>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>>>> ~[na:1.7.0_55] >>>>>> at >>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>>>> [na:1.7.0_55] >>>>>> at java.lang.Thread.run(Thread.java:744) [na:1.7.0_55] >>>>>> >>>>>> It seems to be a bug (is it known ?). Or do you have any clue on how >>>>>> I can solve that ? >>>>>> >>>>>> Thanks for your help >>>>>> >>>>>> Fabrice >>>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> Tyler Hobbs >>>> DataStax <http://datastax.com/> >>>> >>> >>> >> >