Ah, that sounds like https://issues.apache.org/jira/browse/CASSANDRA-1198. That it happened after removetoken is just that that happened to change your ring topology enough to make your queries start hitting it.
On Tue, Jun 22, 2010 at 10:39 AM, Joost Ouwerkerk <jo...@openplaces.org> wrote: > I don't mind missing data for a few hours, it's the weird behaviour of > get_range_slices that's bothering me. I added some logging to > ColumnFamilyRecordReader to see what's going on: > > Split startToken=67160993471237854630929198835217410155, > endToken=68643623863384825230116928934887817211 > > ... > > Getting batch for range: 67965855060996012099315582648654139032 to > 68643623863384825230116928934887817211 > > Token for last row is: 50448492574454416067449808504057295946 > > Getting batch for range: 50448492574454416067449808504057295946 to > 68643623863384825230116928934887817211 > > ... > > Notice how the get_range_slices response is invalid since it returns an > out-of-range row. This poisons the batching loop and causes the task to > spin out of control. > /joost > > On Tue, Jun 22, 2010 at 9:09 AM, Jonathan Ellis <jbel...@gmail.com> wrote: >> >> What I would expect to have happen is for the removed node to >> disappear from the ring and for nodes that are supposed to get more >> data to start streaming it over. I would expect it to be hours before >> any new data started appearing anywhere when you are anticompacting >> 80+GB prior to the streaming part. >> http://wiki.apache.org/cassandra/Streaming >> >> On Tue, Jun 22, 2010 at 12:57 AM, Joost Ouwerkerk <jo...@openplaces.org> >> wrote: >> > Yes, although "forget" implies that we once knew we were supposed to do >> > so. >> > Given the following before-and-after states, on which nodes are we >> > supposed >> > to run repair? Should the cluster be restarted? Is there anything else >> > we >> > should be doing, or not doing? >> > >> > 1. Node is down due to hardware failure >> > >> > 192.168.1.104 Up 111.75 GB >> > 8954799129498380617457226511362321354 | ^ >> > 192.168.1.106 Up 113.25 GB >> > 17909598258996761234914453022724642708 v | >> > 192.168.1.107 Up 75.65 GB >> > 22386997823745951543643066278405803385 | ^ >> > 192.168.1.108 Down 75.77 GB >> > 26864397388495141852371679534086964062 v | >> > 192.168.1.109 Up 76.14 GB >> > 35819196517993522469828906045449285416 | ^ >> > 192.168.1.110 Up 75.9 GB >> > 40296596082742712778557519301130446093 v | >> > 192.168.1.111 Up 95.21 GB >> > 49251395212241093396014745812492767447 | ^ >> > >> > 2. nodetool removetoken 26864397388495141852371679534086964062 >> > >> > 192.168.1.104 Up 111.75 GB >> > 8954799129498380617457226511362321354 | ^ >> > 192.168.1.106 Up 113.25 GB >> > 17909598258996761234914453022724642708 v | >> > 192.168.1.107 Up 75.65 GB >> > 22386997823745951543643066278405803385 | ^ >> > 192.168.1.109 Up 76.14 GB >> > 35819196517993522469828906045449285416 | ^ >> > 192.168.1.110 Up 75.9 GB >> > 40296596082742712778557519301130446093 v | >> > 192.168.1.111 Up 95.21 GB >> > 49251395212241093396014745812492767447 | ^ >> > >> > At this point we're expecting 192.168.1.107 to pick up the slack for the >> > removed token, and for 192.168.1.109 and/or 192.168.1.110 to start >> > streaming >> > data to 192.168.1.107 since they are holding the replicated data for >> > that >> > range. >> > >> > 3. nodetool repair ? >> > >> > On Tue, Jun 22, 2010 at 12:03 AM, Benjamin Black <b...@b3k.us> wrote: >> >> >> >> Did you forget to run repair? >> >> >> >> On Mon, Jun 21, 2010 at 7:02 PM, Joost Ouwerkerk <jo...@openplaces.org> >> >> wrote: >> >> > I believe we did nodetool removetoken on nodes that were already down >> >> > (due >> >> > to hardware failure), but I will check to make sure. We're running >> >> > Cassandra >> >> > 0.6.2. >> >> > >> >> > On Mon, Jun 21, 2010 at 9:59 PM, Joost Ouwerkerk >> >> > <jo...@openplaces.org> >> >> > wrote: >> >> >> >> >> >> Greg, can you describe the steps we took to decommission the nodes? >> >> >> >> >> >> ---------- Forwarded message ---------- >> >> >> From: Rob Coli <rc...@digg.com> >> >> >> Date: Mon, Jun 21, 2010 at 8:08 PM >> >> >> Subject: Re: get_range_slices confused about token ranges after >> >> >> decommissioning a node >> >> >> To: user@cassandra.apache.org >> >> >> >> >> >> >> >> >> On 6/21/10 4:57 PM, Joost Ouwerkerk wrote: >> >> >>> >> >> >>> We're seeing very strange behaviour after decommissioning a node: >> >> >>> when >> >> >>> requesting a get_range_slices with a KeyRange by token, we are >> >> >>> getting >> >> >>> back tokens that are out of range. >> >> >> >> >> >> What sequence of actions did you take to "decommission" the node? >> >> >> What >> >> >> version of Cassandra are you running? >> >> >> >> >> >> =Rob >> >> >> >> >> > >> >> > >> > >> > >> >> >> >> -- >> Jonathan Ellis >> Project Chair, Apache Cassandra >> co-founder of Riptano, the source for professional Cassandra support >> http://riptano.com > > -- Jonathan Ellis Project Chair, Apache Cassandra co-founder of Riptano, the source for professional Cassandra support http://riptano.com