Yes replication factor is 3.

I ran nodetool repair -pr on all the nodes (one at a time) and am still
having issues getting data back from queries.

I did make the new node a seed node.

Re "rack4": I assumed that was just an indication as to the physical
location of the server for redundancy. This one is separate from the others
so I used rack4.

On Mon, Apr 3, 2023 at 6:30 AM Carlos Diaz <crdiaz...@gmail.com> wrote:

> I'm assuming that your replication factor is 3.  If that's the case, did
> you intentionally put this node in rack 4?  Typically, you want to add
> nodes in multiples of your replication factor in order to keep the "racks"
> balanced.  In other words, this node should have been added to rack 1, 2 or
> 3.
>
> Having said that, you should be able to easily fix your problem by running
> a nodetool repair -pr on the new node.
>
> On Sun, Apr 2, 2023 at 8:16 PM David Tinker <david.tin...@gmail.com>
> wrote:
>
>> Hi All
>>
>> I recently added a node to my 3 node Cassandra 4.0.5 cluster and now many
>> reads are not returning rows! What do I need to do to fix this? There
>> weren't any errors in the logs or other problems that I could see. I
>> expected the cluster to balance itself but this hasn't happened (yet?). The
>> nodes are similar so I have num_tokens=256 for each. I am using the
>> Murmur3Partitioner.
>>
>> # nodetool status
>> Datacenter: dc1
>> ===============
>> Status=Up/Down
>> |/ State=Normal/Leaving/Joining/Moving
>> --  Address          Load       Tokens  Owns (effective)  Host ID
>>                       Rack
>> UN  xxx.xxx.xxx.105  2.65 TiB   256     72.9%
>> afd02287-3f88-4c6f-8b27-06f7a8192402  rack3
>> UN  xxx.xxx.xxx.253  2.6 TiB    256     73.9%
>> e1af72be-e5df-4c6b-a124-c7bc48c6602a  rack2
>> UN  xxx.xxx.xxx.24   93.82 KiB  256     80.0%
>> c4e8b4a0-f014-45e6-afb4-648aad4f8500  rack4
>> UN  xxx.xxx.xxx.107  2.65 TiB   256     73.2%
>> ab72f017-be96-41d2-9bef-a551dec2c7b5  rack1
>>
>> # nodetool netstats
>> Mode: NORMAL
>> Not sending any streams.
>> Read Repair Statistics:
>> Attempted: 0
>> Mismatch (Blocking): 0
>> Mismatch (Background): 0
>> Pool Name                    Active   Pending      Completed   Dropped
>> Large messages                  n/a         0          71754         0
>> Small messages                  n/a         0        8398184        14
>> Gossip messages                 n/a         0        1303634         0
>>
>> # nodetool ring
>> Datacenter: dc1
>> ==========
>> Address               Rack        Status State   Load            Owns
>>            Token
>>
>>            9189523899826545641
>> xxx.xxx.xxx.24        rack4       Up     Normal  93.82 KiB       79.95%
>>            -9194674091837769168
>> xxx.xxx.xxx.107       rack1       Up     Normal  2.65 TiB        73.25%
>>            -9168781258594813088
>> xxx.xxx.xxx.253       rack2       Up     Normal  2.6 TiB         73.92%
>>            -9163037340977721917
>> xxx.xxx.xxx.105       rack3       Up     Normal  2.65 TiB        72.88%
>>            -9148860739730046229
>> xxx.xxx.xxx.107       rack1       Up     Normal  2.65 TiB        73.25%
>>            -9125240034139323535
>> xxx.xxx.xxx.253       rack2       Up     Normal  2.6 TiB         73.92%
>>            -9112518853051755414
>> xxx.xxx.xxx.105       rack3       Up     Normal  2.65 TiB        72.88%
>>            -9100516173422432134
>> ...
>>
>> This is causing a serious production issue. Please help if you can.
>>
>> Thanks
>> David
>>
>>
>>
>>

Reply via email to