I should add that the new node does have some data.

On Mon, Apr 3, 2023 at 7:04 AM David Tinker <david.tin...@gmail.com> wrote:

> It is also in the seeds list for the other nodes. Should I remove it from
> those, restart them one at a time, then restart it?
>
> /etc/cassandra # grep -i bootstrap *
> doesn't show anything so I don't think I have auto_bootstrap false.
>
> Thanks very much for the help.
>
>
> On Mon, Apr 3, 2023 at 7:01 AM Carlos Diaz <crdiaz...@gmail.com> wrote:
>
>> Just remove it from the seed list in the cassandra.yaml file and restart
>> the node.  Make sure that auto_bootstrap is set to true first though.
>>
>> On Sun, Apr 2, 2023 at 9:59 PM David Tinker <david.tin...@gmail.com>
>> wrote:
>>
>>> So likely because I made it a seed node when I added it to the cluster
>>> it didn't do the bootstrap process. How can I recover this?
>>>
>>> On Mon, Apr 3, 2023 at 6:41 AM David Tinker <david.tin...@gmail.com>
>>> wrote:
>>>
>>>> Yes replication factor is 3.
>>>>
>>>> I ran nodetool repair -pr on all the nodes (one at a time) and am
>>>> still having issues getting data back from queries.
>>>>
>>>> I did make the new node a seed node.
>>>>
>>>> Re "rack4": I assumed that was just an indication as to the physical
>>>> location of the server for redundancy. This one is separate from the others
>>>> so I used rack4.
>>>>
>>>> On Mon, Apr 3, 2023 at 6:30 AM Carlos Diaz <crdiaz...@gmail.com> wrote:
>>>>
>>>>> I'm assuming that your replication factor is 3.  If that's the case,
>>>>> did you intentionally put this node in rack 4?  Typically, you want to add
>>>>> nodes in multiples of your replication factor in order to keep the "racks"
>>>>> balanced.  In other words, this node should have been added to rack 1, 2 
>>>>> or
>>>>> 3.
>>>>>
>>>>> Having said that, you should be able to easily fix your problem by
>>>>> running a nodetool repair -pr on the new node.
>>>>>
>>>>> On Sun, Apr 2, 2023 at 8:16 PM David Tinker <david.tin...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hi All
>>>>>>
>>>>>> I recently added a node to my 3 node Cassandra 4.0.5 cluster and now
>>>>>> many reads are not returning rows! What do I need to do to fix this? 
>>>>>> There
>>>>>> weren't any errors in the logs or other problems that I could see. I
>>>>>> expected the cluster to balance itself but this hasn't happened (yet?). 
>>>>>> The
>>>>>> nodes are similar so I have num_tokens=256 for each. I am using the
>>>>>> Murmur3Partitioner.
>>>>>>
>>>>>> # nodetool status
>>>>>> Datacenter: dc1
>>>>>> ===============
>>>>>> Status=Up/Down
>>>>>> |/ State=Normal/Leaving/Joining/Moving
>>>>>> --  Address          Load       Tokens  Owns (effective)  Host ID
>>>>>>                           Rack
>>>>>> UN  xxx.xxx.xxx.105  2.65 TiB   256     72.9%
>>>>>> afd02287-3f88-4c6f-8b27-06f7a8192402  rack3
>>>>>> UN  xxx.xxx.xxx.253  2.6 TiB    256     73.9%
>>>>>> e1af72be-e5df-4c6b-a124-c7bc48c6602a  rack2
>>>>>> UN  xxx.xxx.xxx.24   93.82 KiB  256     80.0%
>>>>>> c4e8b4a0-f014-45e6-afb4-648aad4f8500  rack4
>>>>>> UN  xxx.xxx.xxx.107  2.65 TiB   256     73.2%
>>>>>> ab72f017-be96-41d2-9bef-a551dec2c7b5  rack1
>>>>>>
>>>>>> # nodetool netstats
>>>>>> Mode: NORMAL
>>>>>> Not sending any streams.
>>>>>> Read Repair Statistics:
>>>>>> Attempted: 0
>>>>>> Mismatch (Blocking): 0
>>>>>> Mismatch (Background): 0
>>>>>> Pool Name                    Active   Pending      Completed   Dropped
>>>>>> Large messages                  n/a         0          71754         0
>>>>>> Small messages                  n/a         0        8398184        14
>>>>>> Gossip messages                 n/a         0        1303634         0
>>>>>>
>>>>>> # nodetool ring
>>>>>> Datacenter: dc1
>>>>>> ==========
>>>>>> Address               Rack        Status State   Load            Owns
>>>>>>                Token
>>>>>>
>>>>>>                9189523899826545641
>>>>>> xxx.xxx.xxx.24        rack4       Up     Normal  93.82 KiB
>>>>>> 79.95%              -9194674091837769168
>>>>>> xxx.xxx.xxx.107       rack1       Up     Normal  2.65 TiB
>>>>>>  73.25%              -9168781258594813088
>>>>>> xxx.xxx.xxx.253       rack2       Up     Normal  2.6 TiB
>>>>>> 73.92%              -9163037340977721917
>>>>>> xxx.xxx.xxx.105       rack3       Up     Normal  2.65 TiB
>>>>>>  72.88%              -9148860739730046229
>>>>>> xxx.xxx.xxx.107       rack1       Up     Normal  2.65 TiB
>>>>>>  73.25%              -9125240034139323535
>>>>>> xxx.xxx.xxx.253       rack2       Up     Normal  2.6 TiB
>>>>>> 73.92%              -9112518853051755414
>>>>>> xxx.xxx.xxx.105       rack3       Up     Normal  2.65 TiB
>>>>>>  72.88%              -9100516173422432134
>>>>>> ...
>>>>>>
>>>>>> This is causing a serious production issue. Please help if you can.
>>>>>>
>>>>>> Thanks
>>>>>> David
>>>>>>
>>>>>>
>>>>>>
>>>>>>

Reply via email to