Can you dump your osd map and post it in a tracker ticket? Or if you're not comfortable with that you can upload it with ceph-post-file and only developers will be able to see it. -Greg
On Thu, Dec 7, 2017 at 10:36 AM Graham Allan <[email protected]> wrote: > Just updated a fairly long-lived (originally firefly) cluster from jewel > to luminous 12.2.2. > > One of the issues I see is a new health warning: > > OSD_ORPHAN 3 osds exist in the crush map but not in the osdmap > osd.2 exists in crush map but not in osdmap > osd.14 exists in crush map but not in osdmap > osd.19 exists in crush map but not in osdmap > > Seemed reasonable enough, these low-numbered OSDs were on > long-decommissioned hardware. I thought I had removed them completely > though, and it seems I had: > > > # ceph osd crush ls osd.2 > > Error ENOENT: node 'osd.2' does not exist > > # ceph osd crush remove osd.2 > > device 'osd.2' does not appear in the crush map > > so I wonder where it's getting this warning from, and if it's erroneous, > how can I clear it? > > Graham > -- > Graham Allan > Minnesota Supercomputing Institute - [email protected] > _______________________________________________ > ceph-users mailing list > [email protected] > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
