This sounds odd. Can you create a ticket in the tracker with all the details you can remember or reconstruct? -Greg
On Wed, Jul 29, 2015 at 8:34 PM Steve Taylor <steve.tay...@storagecraft.com> wrote: > I recently migrated 240 OSDs to new servers this way in a single cluster, > and it worked great. There are two additional items I would note based on > my experience though. > > First, if you're using dmcrypt then of course you need to copy the dmcrypt > keys for the OSDs to the new host(s). I had to do this in my case, but it > was very straightforward. > > Second was an issue I didn't expect, probably just because of my > ignorance. I was not able to migrate existing OSDs from different failure > domains into a new, single failure domain without waiting for full recovery > to HEALTH_OK in between. The very first server I put OSD disks from two > different failure domains into had issues. The OSDs came up and in just > fine, but immediately started flapping and failed to make progress toward > recovery. I removed the disks from one failure domain and left the others, > and recovery progressed as expected. As soon as I saw HEALTH_OK I > re-migrated the OSDs from the other failure domain and again the cluster > recovered as expected. Proceeding via this method allowed me to migrate all > 240 OSDs without any further problems. I was also able to migrate as many > OSDs as I wanted to simultaneously as long as I didn't mix OSDs from > different, old failure domains in a new failure domain without recovering > in between. I understand mixing failure domains li > ke this is risky, but I sort of expected it to work anyway. Maybe it was > better in the end that Ceph forced me to do it more safely. > > Steve Taylor | Senior Software Engineer | StorageCraft Technology > Corporation > 380 Data Drive Suite 300 | Draper | Utah | 84020 > Office: 801.871.2799 | Fax: 801.545.4705 > > If you are not the intended recipient of this message, be advised that any > dissemination or copying of this message is prohibited. > If you received this message erroneously, please notify the sender and > delete it, together with any attachments. > > -----Original Message----- > From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of > Peter Hinman > Sent: Wednesday, July 29, 2015 12:58 PM > To: Robert LeBlanc <rob...@leblancnet.us> > Cc: ceph-users@lists.ceph.com > Subject: Re: [ceph-users] Recovery question > > Thanks for the guidance. I'm working on building a valid ceph.conf right > now. I'm not familiar with the osd-bootstrap key. Is that the standard > filename for it? Is it the keyring that is stored on the osd? > > I'll see if the logs turn up anything I can decipher after I rebuild the > ceph.conf file. > > -- > Peter Hinman > > On 7/29/2015 12:49 PM, Robert LeBlanc wrote: > > -----BEGIN PGP SIGNED MESSAGE----- > > Hash: SHA256 > > > > Did you use ceph-depoy or ceph-disk to create the OSDs? If so, it > > should use udev to start he OSDs. In that case, a new host that has > > the correct ceph.conf and osd-bootstrap key should be able to bring up > > the OSDs into the cluster automatically. Just make sure you have the > > correct journal in the same host with the matching OSD disk, udev > > should do the magic. > > > > The OSD logs are your friend if they don't start properly. > > - ---------------- > > Robert LeBlanc > > PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 > > > > > > On Wed, Jul 29, 2015 at 10:48 AM, Peter Hinman wrote: > >> I've got a situation that seems on the surface like it should be > >> recoverable, but I'm struggling to understand how to do it. > >> > >> I had a cluster of 3 monitors, 3 osd disks, and 3 journal ssds. After > >> multiple hardware failures, I pulled the 3 osd disks and 3 journal > >> ssds and am attempting to bring them back up again on new hardware in a > new cluster. > >> I see plenty of documentation on how to zap and initialize and add "new" > >> osds, but I don't see anything on rebuilding with existing osd disks. > >> > >> Could somebody provide guidance on how to do this? I'm running 94.2 > >> on all machines. > >> > >> Thanks, > >> > >> -- > >> Peter Hinman > >> > >> > >> _______________________________________________ > >> ceph-users mailing list > >> ceph-users@lists.ceph.com > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > -----BEGIN PGP SIGNATURE----- > > Version: Mailvelope v0.13.1 > > Comment: https://www.mailvelope.com > > > > wsFcBAEBCAAQBQJVuSA/CRDmVDuy+mK58QAAfGAQAMq62W7QvCAo2RSDWLli > > 13AJTpAWhk+ilBwcmxFr/gP/Aa9hMN5bV8idDqI56YWBjGO2WPQIUT8CXH5v > > ocBUZZJ0X08gOgHqFQ8x3rSSe6QINy1bQONMql3Jgpy8He/ctLnXROhNT9SU > > l30CI4qKwG48AZU5E4PoWgwQmdbFv0WIuFwCzPOVIU6GvO0umirerw3C7tZQ > > I34+OINURzCjKzLY/OEF4hRvRq3PV0KZAoolQTeBJtEdlyNgAQ/bHOgpfJ/h > > diGwQZyhSzqTvFYOEHWUuh5ZnhZAMNtaLBulwreUEKoI0IcXGxpH6KsC7ag4 > > KJ1kD8U0I18eP4iyTOIXg+DxafUU4wrITlKdomW12XqmlHadi2vYYBCqataI > > uc4KeXHP4/SrA1qoEDtXroAV2iuV6UUNIwsY4HPBJ/CNKXFU5QSdGOey3Kjs > > Mz2zuCpMkTf6fj8B4XJfenfFulRVJwrKJml7JebPFpLTRPFMbsuZ5htUMASn > > UWyCA9IfxLYsC5tPlii79Kkb93mvN3cCdvchkH2CQ38jxkVRZRUqeJlzvtVp > > 2mwinvqPD0irTvr+LvmlKOdtvFSOKJM0XmRSVk1LgLlpoyIZ9BqI02ul01fE > > 7nZ892/17zdv0Nguxr8F8bps0jA7NLFpgRhEsakdmTVTJQLMwSv7z6c9fdP0 > > 7AWQ > > =VJV0 > > -----END PGP SIGNATURE----- > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com