Have you upgraded any OSD's? On a test cluster I saw the same and as I upgraded / restarted the OSD's the PG's started to show online till it was 100%.
I know it says to not change anything to do with pool's during the upgrade so I am guessing there is a code change that cause this till all is on the same version. On Tue, Mar 26, 2019 at 11:37 PM Stadsnet <jwil...@stads.net> wrote: > > We did a upgrade from luminous to nautilus > > after upgrading the three monitors we got that all our pgs where inactive > > cluster: > id: 5bafad08-31b2-4716-be77-07ad2e2647eb > health: HEALTH_ERR > noout flag(s) set > 1 scrub errors > Reduced data availability: 1429 pgs inactive > 316 pgs not deep-scrubbed in time > 520 pgs not scrubbed in time > 3 monitors have not enabled msgr2 > > services: > mon: 3 daemons, quorum Ceph-Mon1,Ceph-Mon2,Ceph-Mon3 (age 51m) > mgr: Ceph-Mon1(active, since 23m), standbys: Ceph-Mon3, Ceph-Mon2 > osd: 103 osds: 103 up, 103 in > flags noout > rgw: 2 daemons active (S3-Ceph1, S3-Ceph2) > > data: > pools: 26 pools, 3248 pgs > objects: 134.92M objects, 202 TiB > usage: 392 TiB used, 486 TiB / 879 TiB avail > pgs: 100.000% pgs unknown > 3248 unknown > > System seems to keep working. > > Did we loose reference "-1 0 root default" ? > > is there a fix for that ? > > ceph osd tree > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF > -18 16.00000 root ssd > -10 2.00000 host Ceph-Stor1-SSD > 80 nvme 2.00000 osd.80 up 1.00000 1.00000 > -11 2.00000 host Ceph-Stor2-SSD > 81 nvme 2.00000 osd.81 up 1.00000 1.00000 > -12 2.00000 host Ceph-Stor3-SSD > 82 nvme 2.00000 osd.82 up 1.00000 1.00000 > -13 2.00000 host Ceph-Stor4-SSD > 83 nvme 2.00000 osd.83 up 1.00000 1.00000 > -14 2.00000 host Ceph-Stor5-SSD > 84 nvme 2.00000 osd.84 up 1.00000 1.00000 > -15 2.00000 host Ceph-Stor6-SSD > 85 nvme 2.00000 osd.85 up 1.00000 1.00000 > -16 2.00000 host Ceph-Stor7-SSD > 86 nvme 2.00000 osd.86 up 1.00000 1.00000 > -17 2.00000 host Ceph-Stor8-SSD > 87 nvme 2.00000 osd.87 up 1.00000 1.00000 > -1 865.93420 root default > -2 110.96700 host Ceph-Stor1 > 0 hdd 9.09599 osd.0 up 1.00000 1.00000 > 1 hdd 9.09599 osd.1 up 1.00000 1.00000 > 2 hdd 9.09599 osd.2 up 1.00000 1.00000 > 3 hdd 9.09599 osd.3 up 1.00000 1.00000 > 4 hdd 9.09599 osd.4 up 1.00000 1.00000 > 5 hdd 9.09599 osd.5 up 1.00000 1.00000 > 6 hdd 9.09599 osd.6 up 1.00000 1.00000 > 7 hdd 9.09599 osd.7 up 1.00000 1.00000 > 8 hdd 9.09599 osd.8 up 1.00000 1.00000 > 9 hdd 9.09599 osd.9 up 1.00000 1.00000 > 88 hdd 9.09599 osd.88 up 1.00000 1.00000 > 89 hdd 9.09599 osd.89 up 1.00000 1.00000 > -3 109.15189 host Ceph-Stor2 > 10 hdd 9.09599 osd.10 up 1.00000 1.00000 > 11 hdd 9.09599 osd.11 up 1.00000 1.00000 > 12 hdd 9.09599 osd.12 up 1.00000 1.00000 > 13 hdd 9.09599 osd.13 up 1.00000 1.00000 > 14 hdd 9.09599 osd.14 up 1.00000 1.00000 > 15 hdd 9.09599 osd.15 up 1.00000 1.00000 > 16 hdd 9.09599 osd.16 up 1.00000 1.00000 > 17 hdd 9.09599 osd.17 up 1.00000 1.00000 > 18 hdd 9.09599 osd.18 up 1.00000 1.00000 > 19 hdd 9.09599 osd.19 up 1.00000 1.00000 > 90 hdd 9.09598 osd.90 up 1.00000 1.00000 > 91 hdd 9.09598 osd.91 up 1.00000 1.00000 > -4 109.15189 host Ceph-Stor3 > 20 hdd 9.09599 osd.20 up 1.00000 1.00000 > 21 hdd 9.09599 osd.21 up 1.00000 1.00000 > 22 hdd 9.09599 osd.22 up 1.00000 1.00000 > 23 hdd 9.09599 osd.23 up 1.00000 1.00000 > 24 hdd 9.09599 osd.24 up 1.00000 1.00000 > 25 hdd 9.09599 osd.25 up 1.00000 1.00000 > 26 hdd 9.09599 osd.26 up 1.00000 1.00000 > 27 hdd 9.09599 osd.27 up 1.00000 1.00000 > 28 hdd 9.09599 osd.28 up 1.00000 1.00000 > 29 hdd 9.09599 osd.29 up 1.00000 1.00000 > 92 hdd 9.09598 osd.92 up 1.00000 1.00000 > 93 hdd 9.09598 osd.93 up 0.80002 1.00000 > -5 109.15189 host Ceph-Stor4 > 30 hdd 9.09599 osd.30 up 1.00000 1.00000 > 31 hdd 9.09599 osd.31 up 1.00000 1.00000 > 32 hdd 9.09599 osd.32 up 1.00000 1.00000 > 33 hdd 9.09599 osd.33 up 1.00000 1.00000 > 34 hdd 9.09599 osd.34 up 0.90002 1.00000 > 35 hdd 9.09599 osd.35 up 1.00000 1.00000 > 36 hdd 9.09599 osd.36 up 1.00000 1.00000 > 37 hdd 9.09599 osd.37 up 1.00000 1.00000 > 38 hdd 9.09599 osd.38 up 1.00000 1.00000 > 39 hdd 9.09599 osd.39 up 1.00000 1.00000 > 94 hdd 9.09598 osd.94 up 1.00000 1.00000 > 95 hdd 9.09598 osd.95 up 1.00000 1.00000 > -6 109.15189 host Ceph-Stor5 > 40 hdd 9.09599 osd.40 up 1.00000 1.00000 > 41 hdd 9.09599 osd.41 up 1.00000 1.00000 > 42 hdd 9.09599 osd.42 up 1.00000 1.00000 > 43 hdd 9.09599 osd.43 up 1.00000 1.00000 > 44 hdd 9.09599 osd.44 up 1.00000 1.00000 > 45 hdd 9.09599 osd.45 up 1.00000 1.00000 > 46 hdd 9.09599 osd.46 up 1.00000 1.00000 > 47 hdd 9.09599 osd.47 up 1.00000 1.00000 > 48 hdd 9.09599 osd.48 up 1.00000 1.00000 > 49 hdd 9.09599 osd.49 up 1.00000 1.00000 > 96 hdd 9.09598 osd.96 up 1.00000 1.00000 > 97 hdd 9.09598 osd.97 up 1.00000 1.00000 > -7 109.15187 host Ceph-Stor6 > 50 hdd 9.09599 osd.50 up 1.00000 1.00000 > 51 hdd 9.09599 osd.51 up 1.00000 1.00000 > 52 hdd 9.09598 osd.52 up 0.80005 1.00000 > 53 hdd 9.09599 osd.53 up 1.00000 1.00000 > 54 hdd 9.09599 osd.54 up 1.00000 1.00000 > 55 hdd 9.09599 osd.55 up 1.00000 1.00000 > 56 hdd 9.09599 osd.56 up 1.00000 1.00000 > 57 hdd 9.09599 osd.57 up 1.00000 1.00000 > 58 hdd 9.09599 osd.58 up 1.00000 1.00000 > 59 hdd 9.09599 osd.59 up 1.00000 1.00000 > 98 hdd 9.09598 osd.98 up 1.00000 1.00000 > 99 hdd 9.09598 osd.99 up 1.00000 1.00000 > -8 109.15189 host Ceph-Stor7 > 60 hdd 9.09599 osd.60 up 1.00000 1.00000 > 61 hdd 9.09599 osd.61 up 1.00000 1.00000 > 62 hdd 9.09599 osd.62 up 1.00000 1.00000 > 63 hdd 9.09599 osd.63 up 1.00000 1.00000 > 64 hdd 9.09599 osd.64 up 1.00000 1.00000 > 65 hdd 9.09599 osd.65 up 1.00000 1.00000 > 66 hdd 9.09599 osd.66 up 1.00000 1.00000 > 67 hdd 9.09599 osd.67 up 1.00000 1.00000 > 68 hdd 9.09599 osd.68 up 1.00000 1.00000 > 69 hdd 9.09599 osd.69 up 1.00000 1.00000 > 100 hdd 9.09598 osd.100 up 1.00000 1.00000 > 101 hdd 9.09598 osd.101 up 1.00000 1.00000 > -9 100.05589 host Ceph-Stor8 > 70 hdd 9.09599 osd.70 up 0.90002 1.00000 > 71 hdd 9.09599 osd.71 up 1.00000 1.00000 > 72 hdd 9.09599 osd.72 up 1.00000 1.00000 > 73 hdd 9.09599 osd.73 up 0.90002 1.00000 > 74 hdd 9.09599 osd.74 up 1.00000 1.00000 > 75 hdd 9.09599 osd.75 up 1.00000 1.00000 > 76 hdd 9.09599 osd.76 up 1.00000 1.00000 > 77 hdd 9.09599 osd.77 up 0.95000 1.00000 > 78 hdd 9.09598 osd.78 up 0.95000 1.00000 > 79 hdd 9.09599 osd.79 up 1.00000 1.00000 > 102 hdd 9.09598 osd.102 up 1.00000 1.00000 > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com