did you tried solr delete -c YOUR_DUPLICATA ? I am noob and I am just asking.
Le mer. 7 mai 2025 à 09:07, Karl Stoney <karl.sto...@autotrader.co.uk.invalid> a écrit : > Ah no worries, thanks for the reply. > I already have a customer operator that watches the statefulsets and does > some admin type stuff, guess ill shim into there the recovery logic for now. > > > From: Jan Høydahl <jan....@cominvent.com> > Date: Tuesday, 6 May 2025 at 17:45 > To: users@solr.apache.org <users@solr.apache.org> > Subject: Re: Recovery from disk failure > > Hi, > > I have seen the same happening myself, and I agree it is somewhat > unexpected. I believe there may be some reason beind the behavior, but > cannot think of any right now. > > So what would work for you right now would be to get rid of the dead > replica in zookeeper (try DELETEREPLICA), and then do an ADDREPLICA on the > new empty box, which will create a new core and start syncing. > Not sure if you are able to remove the replica in that state, but give it > a try. > So, until we decide to build a different default behavior in the "startup > on empty disk but zk says we should have collection-A-shard-1-replica-2" > case, the best way would be to first move all replicas away from the node > that will upgrade disk, and then move them back again. > > Jan > > > 6. mai 2025 kl. 13:55 skrev Karl Stoney <karl.sto...@autotrader.co.uk > .INVALID>: > > > > Hi, > > I run solr cloud on GKE; and I’m trying to move my pods to a new disk > type. In doing so the disk will be brand new. I’ve landed in a position > that I’m unsure how to recover from, where the new node is not syncing data > from the leader. > > > > To explain exactly what’s happening, lets say I have two nodes: > > > > * > > solr-0 > > * > > solr-1 > > > > And both are active and fully replicated. > > I take solr-1 down, and point it at the new disk (which is empty), and > bring it back up. > > The server starts fine, I can access solr-1 via the UI, but it never > recovers, in the “Cloud -> Graph” UI, I can see the shard on solr-1 is > down. > > > > I can see it in the “Cloud -> Nodes” GUI as up, however its collections > have a funny state, for example: "postcodes-006_s1r9_(down): undefined", > vs solr-0 which shows "postcodes-006_s1r11: 847.3Mb”. > > > > I was expecting the node to come up and see its disk was empty, and > resync its data from the leader, but instead it’s just sat doing, nothing…. > > > > The fact I’m moving to new disks is somewhat moot, more broadly this is > showing me that if we lost data on a node for whatever reason, it doesn’t > “fix itself” - which I always (maybe blindly) assumed it would, because > when I bring up brand new nodes (different name) it does. > > > > Could anyone advise what I’ve done wrong here, and what the process > should be to get a node to resend its data entirely? > > > > This is what the API shows: > > > > > > shard1":{ > > > > "range":"80000000-7fffffff", > > > > "replicas":{ > > > > "core_node10":{ > > > > "core":"postcodes-006_shard1_replica_n9", > > > > > "node_name":"solr-1.search-solr-next.svc.cluster.local:80_solr", > > > > "type":"NRT", > > > > "state":"down", > > > > "force_set_state":"false", > > > > "base_url":" > https://eur03.safelinks.protection.outlook.com/?url=http%3A%2F%2Fsolr-1.search-solr-next.svc.cluster.local%2Fsolr&data=05%7C02%7CKarl.Stoney%40autotrader.co.uk%7C320a9d3baf2d456311a208dd8cbd57f6%7C926f3743f3d24b8a816818cfcbe776fe%7C0%7C0%7C638821467062718137%7CUnknown%7CTWFpbGZsb3d8eyJFbXB0eU1hcGkiOnRydWUsIlYiOiIwLjAuMDAwMCIsIlAiOiJXaW4zMiIsIkFOIjoiTWFpbCIsIldUIjoyfQ%3D%3D%7C0%7C%7C%7C&sdata=sAPOw6r%2Bpw0GnNV00bPdMTiQMNYybKlk2rgKfTZIHaw%3D&reserved=0 > " > > > > }, > > > > "core_node12":{ > > > > "core":"postcodes-006_shard1_replica_n11", > > > > > "node_name":"solr-0.search-solr-next.svc.cluster.local:80_solr", > > > > "type":"NRT", > > > > "state":"active", > > > > "leader":"true", > > > > "force_set_state":"false", > > > > "base_url":" > https://eur03.safelinks.protection.outlook.com/?url=http%3A%2F%2Fsolr-0.search-solr-next.svc.cluster.local%2Fsolr&data=05%7C02%7CKarl.Stoney%40autotrader.co.uk%7C320a9d3baf2d456311a208dd8cbd57f6%7C926f3743f3d24b8a816818cfcbe776fe%7C0%7C0%7C638821467062735681%7CUnknown%7CTWFpbGZsb3d8eyJFbXB0eU1hcGkiOnRydWUsIlYiOiIwLjAuMDAwMCIsIlAiOiJXaW4zMiIsIkFOIjoiTWFpbCIsIldUIjoyfQ%3D%3D%7C0%7C%7C%7C&sdata=LqHQ7N0PhNlDjUHq52Kn4AfgzV8uvFf7A2eXC%2BLlfqY%3D&reserved=0 > ", > > > > "property.preferredleader":"true" > > > > } > > > > }, > > > > "state":"active", > > > > "health":"ORANGE" > > > > } > > > > > > > > > > Unless expressly stated otherwise in this email, this e-mail is sent on > behalf of Auto Trader Limited Registered Office: 1 Tony Wilson Place, > Manchester, Lancashire, M15 4FN (Registered in England No. 03909628). Auto > Trader Limited is part of the Auto Trader Group Plc group. This email and > any files transmitted with it are confidential and may be legally > privileged, and intended solely for the use of the individual or entity to > whom they are addressed. If you have received this email in error please > notify the sender. This email message has been swept for the presence of > computer viruses. > > > > Unless expressly stated otherwise in this email, this e-mail is sent on > behalf of Auto Trader Limited Registered Office: 1 Tony Wilson Place, > Manchester, Lancashire, M15 4FN (Registered in England No. 03909628). Auto > Trader Limited is part of the Auto Trader Group Plc group. This email and > any files transmitted with it are confidential and may be legally > privileged, and intended solely for the use of the individual or entity to > whom they are addressed. If you have received this email in error please > notify the sender. This email message has been swept for the presence of > computer viruses. >