Paul / All I'm not sure what warning your are referring to, I'm on Nautilus. The point I'm getting at is if you weight out all OSD on a host with a cluster of 3 OSD hosts with 3 OSD each, crush rule = host, then write to the cluster, it *should* imo not just say remapped but undersized / degraded.
See below, 1 out of the 3 OSD hosts has ALL it's OSD marked out and weight = 0. When you write (say using FIO), the PGs *only* have 2 OSD in them (UP set), which is pool min size. I don't understand why it's not saying undersized/degraded, this seems like a bug. Who cares that the Acting Set has the 3 original OSD in it, the actual data is only on 2 OSD, which is a degraded state root@hostadmin:~# ceph -s cluster: id: 33d41932-9df2-40ba-8e16-8dedaa4b3ef6 health: HEALTH_WARN application not enabled on 1 pool(s) services: mon: 1 daemons, quorum hostmonitor1 (age 29m) mgr: hostmonitor1(active, since 31m) osd: 9 osds: 9 up, 6 in; 100 remapped pgs data: pools: 1 pools, 100 pgs objects: 520 objects, 2.0 GiB usage: 15 GiB used, 75 GiB / 90 GiB avail pgs: 520/1560 objects misplaced (33.333%) 100 active+clean+remapped root@hostadmin:~# ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 0.08817 root default -3 0.02939 host hostosd1 0 hdd 0.00980 osd.0 up 1.00000 1.00000 3 hdd 0.00980 osd.3 up 1.00000 1.00000 6 hdd 0.00980 osd.6 up 1.00000 1.00000 -5 0.02939 host hostosd2 1 hdd 0.00980 osd.1 up 0 1.00000 4 hdd 0.00980 osd.4 up 0 1.00000 7 hdd 0.00980 osd.7 up 0 1.00000 -7 0.02939 host hostosd3 2 hdd 0.00980 osd.2 up 1.00000 1.00000 5 hdd 0.00980 osd.5 up 1.00000 1.00000 8 hdd 0.00980 osd.8 up 1.00000 1.00000 root@hostadmin:~# ceph osd df ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META AVAIL %USE VAR PGS STATUS 0 hdd 0.00980 1.00000 10 GiB 1.7 GiB 765 MiB 12 KiB 1024 MiB 8.2 GiB 17.48 1.03 34 up 3 hdd 0.00980 1.00000 10 GiB 1.7 GiB 765 MiB 12 KiB 1024 MiB 8.2 GiB 17.48 1.03 36 up 6 hdd 0.00980 1.00000 10 GiB 1.6 GiB 593 MiB 4 KiB 1024 MiB 8.4 GiB 15.80 0.93 30 up 1 hdd 0.00980 0 0 B 0 B 0 B 0 B 0 B 0 B 0 0 0 up 4 hdd 0.00980 0 0 B 0 B 0 B 0 B 0 B 0 B 0 0 0 up 7 hdd 0.00980 0 0 B 0 B 0 B 0 B 0 B 0 B 0 0 100 up 2 hdd 0.00980 1.00000 10 GiB 1.5 GiB 525 MiB 8 KiB 1024 MiB 8.5 GiB 15.13 0.89 20 up 5 hdd 0.00980 1.00000 10 GiB 1.9 GiB 941 MiB 4 KiB 1024 MiB 8.1 GiB 19.20 1.13 43 up 8 hdd 0.00980 1.00000 10 GiB 1.6 GiB 657 MiB 8 KiB 1024 MiB 8.4 GiB 16.42 0.97 37 up TOTAL 90 GiB 15 GiB 6.2 GiB 61 KiB 9.0 GiB 75 GiB 16.92 MIN/MAX VAR: 0.89/1.13 STDDEV: 1.32 Tarek Zegar Senior SDS Engineer Email tze...@us.ibm.com Mobile 630.974.7172 From: Paul Emmerich <paul.emmer...@croit.io> To: Tarek Zegar <tze...@us.ibm.com> Cc: Ceph Users <ceph-users@lists.ceph.com> Date: 06/07/2019 05:25 AM Subject: [EXTERNAL] Re: [ceph-users] Reweight OSD to 0, why doesn't report degraded if UP set under Pool Size remapped no longer triggers a health warning in nautilus. Your data is still there, it's just on the wrong OSD if that OSD is still up and running. Paul -- Paul Emmerich Looking for help with your Ceph cluster? Contact us at https://croit.io croit GmbH Freseniusstr. 31h 81247 München www.croit.io Tel: +49 89 1896585 90 On Thu, Jun 6, 2019 at 10:48 PM Tarek Zegar <tze...@us.ibm.com> wrote: For testing purposes I set a bunch of OSD to 0 weight, this correctly forces Ceph to not use said OSD. I took enough out such that the UP set only had Pool min size # of OSD (i.e 2 OSD). Two Questions: 1. Why doesn't the acting set eventually match the UP set and simply point to [6,5] only 2. Why are none of the PGs marked as undersized and degraded? The data is only hosted on 2 OSD rather then Pool size (3), I would expect a undersized warning and degraded for PG with data? Example PG: PG 1.4d active+clean+remapped UP= [6,5] Acting = [6,5,4] OSD Tree: ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 0.08817 root default -3 0.02939 host hostosd1 0 hdd 0.00980 osd.0 up 1.00000 1.00000 3 hdd 0.00980 osd.3 up 1.00000 1.00000 6 hdd 0.00980 osd.6 up 1.00000 1.00000 -5 0.02939 host hostosd2 1 hdd 0.00980 osd.1 up 0 1.00000 4 hdd 0.00980 osd.4 up 0 1.00000 7 hdd 0.00980 osd.7 up 0 1.00000 -7 0.02939 host hostosd3 2 hdd 0.00980 osd.2 up 1.00000 1.00000 5 hdd 0.00980 osd.5 up 1.00000 1.00000 8 hdd 0.00980 osd.8 up 0 1.00000 _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com