Hi David, root@or1010051251044:~# ceph df GLOBAL: SIZE AVAIL RAW USED %RAW USED 79793G 56832G 22860G 28.65 POOLS: NAME ID USED %USED MAX AVAIL OBJECTS rbd 0 0 0 14395G 0 compute 1 0 0 14395G 0 volumes 2 7605G 28.60 14395G 1947372 images 4 0 0 14395G 0 root@or1010051251044:~#
pool : 4 0 1 2 | SUM ------------------------------------------------ osd.10 8 10 44 96 | 158 osd.11 14 8 58 100 | 180 osd.12 12 6 50 95 | 163 osd.13 14 4 49 121 | 188 osd.14 9 8 54 86 | 157 osd.15 12 5 55 103 | 175 osd.16 23 5 56 99 | 183 osd.30 6 4 31 47 | 88 osd.17 8 8 50 114 | 180 osd.31 7 1 23 35 | 66 osd.18 15 5 42 94 | 156 osd.32 12 6 24 54 | 96 osd.19 13 5 54 116 | 188 osd.33 4 2 28 49 | 83 osd.34 7 5 18 62 | 92 osd.35 10 2 21 56 | 89 osd.36 5 1 34 35 | 75 osd.37 4 4 24 45 | 77 osd.39 14 8 48 106 | 176 osd.0 12 3 27 67 | 109 osd.1 8 3 27 43 | 81 osd.2 4 5 27 45 | 81 osd.3 4 3 19 50 | 76 osd.4 4 1 23 54 | 82 osd.5 4 2 23 56 | 85 osd.6 1 5 32 50 | 88 osd.7 9 1 32 66 | 108 osd.8 7 4 27 49 | 87 osd.9 6 4 24 55 | 89 osd.20 7 4 43 122 | 176 osd.21 14 5 46 95 | 160 osd.22 13 8 51 107 | 179 osd.23 11 7 54 105 | 177 osd.24 11 6 52 112 | 181 osd.25 16 6 36 98 | 156 osd.26 15 7 59 101 | 182 osd.27 7 9 58 101 | 175 osd.28 16 5 60 89 | 170 osd.29 18 7 53 94 | 172 ------------------------------------------------ SUM : 384 192 1536 3072 root@or1010051251044:~# for i in `rados lspools`; do echo "================="; echo Working on pool: $i; ceph osd pool get $i pg_num; ceph osd pool get $i pgp_num; done ================= Working on pool: rbd pg_num: 64 pgp_num: 64 ================= Working on pool: compute pg_num: 512 pgp_num: 512 ================= Working on pool: volumes pg_num: 1024 pgp_num: 1024 ================= Working on pool: images pg_num: 128 pgp_num: 128 root@or1010051251044:~# Thanks, Pardhiv Karri On Tue, May 22, 2018 at 9:16 AM, David Turner <drakonst...@gmail.com> wrote: > This is all weird. Maybe it just doesn't have any PGs with data on them. > `ceph df`, how many PGs you have in each pool, and which PGs are on osd 38. > > > On Tue, May 22, 2018, 11:19 AM Pardhiv Karri <meher4in...@gmail.com> > wrote: > >> Hi David, >> >> >> >> root@or1010051251044:~# ceph osd tree >> ID WEIGHT TYPE NAME UP/DOWN REWEIGHT >> PRIMARY-AFFINITY >> -1 80.00000 root default >> >> -2 40.00000 rack rack_A1 >> >> -3 20.00000 host or1010051251040 >> >> 0 2.00000 osd.0 up 1.00000 >> 1.00000 >> 1 2.00000 osd.1 up 1.00000 >> 1.00000 >> 2 2.00000 osd.2 up 1.00000 >> 1.00000 >> 3 2.00000 osd.3 up 1.00000 >> 1.00000 >> 4 2.00000 osd.4 up 1.00000 >> 1.00000 >> 5 2.00000 osd.5 up 1.00000 >> 1.00000 >> 6 2.00000 osd.6 up 1.00000 >> 1.00000 >> 7 2.00000 osd.7 up 1.00000 >> 1.00000 >> 8 2.00000 osd.8 up 1.00000 >> 1.00000 >> 9 2.00000 osd.9 up 1.00000 >> 1.00000 >> -8 20.00000 host or1010051251044 >> >> 30 2.00000 osd.30 up 1.00000 >> 1.00000 >> 31 2.00000 osd.31 up 1.00000 >> 1.00000 >> 32 2.00000 osd.32 up 1.00000 >> 1.00000 >> 33 2.00000 osd.33 up 1.00000 >> 1.00000 >> 34 2.00000 osd.34 up 1.00000 >> 1.00000 >> 35 2.00000 osd.35 up 1.00000 >> 1.00000 >> 36 2.00000 osd.36 up 1.00000 >> 1.00000 >> 37 2.00000 osd.37 up 1.00000 >> 1.00000 >> 38 2.00000 osd.38 up 1.00000 >> 1.00000 >> 39 2.00000 osd.39 up 1.00000 >> 1.00000 >> -4 20.00000 rack rack_B1 >> >> -5 20.00000 host or1010051251041 >> >> 10 2.00000 osd.10 up 1.00000 >> 1.00000 >> 11 2.00000 osd.11 up 1.00000 >> 1.00000 >> 12 2.00000 osd.12 up 1.00000 >> 1.00000 >> 13 2.00000 osd.13 up 1.00000 >> 1.00000 >> 14 2.00000 osd.14 up 1.00000 >> 1.00000 >> 15 2.00000 osd.15 up 1.00000 >> 1.00000 >> 16 2.00000 osd.16 up 1.00000 >> 1.00000 >> 17 2.00000 osd.17 up 1.00000 >> 1.00000 >> 18 2.00000 osd.18 up 1.00000 >> 1.00000 >> 19 2.00000 osd.19 up 1.00000 >> 1.00000 >> -9 0 host or1010051251045 >> >> -6 20.00000 rack rack_C1 >> >> -7 20.00000 host or1010051251042 >> >> 20 2.00000 osd.20 up 1.00000 >> 1.00000 >> 21 2.00000 osd.21 up 1.00000 >> 1.00000 >> 22 2.00000 osd.22 up 1.00000 >> 1.00000 >> 23 2.00000 osd.23 up 1.00000 >> 1.00000 >> 24 2.00000 osd.24 up 1.00000 >> 1.00000 >> 25 2.00000 osd.25 up 1.00000 >> 1.00000 >> 26 2.00000 osd.26 up 1.00000 >> 1.00000 >> 27 2.00000 osd.27 up 1.00000 >> 1.00000 >> 28 2.00000 osd.28 up 1.00000 >> 1.00000 >> 29 2.00000 osd.29 up 1.00000 >> 1.00000 >> -10 0 host or1010051251046 >> >> -11 0 host or1010051251023 >> >> root@or1010051251044:~# >> >> >> >> >> >> root@or1010051251044:~# ceph -s >> cluster 6eacac66-087a-464d-94cb-9ca2585b98d5 >> health HEALTH_OK >> monmap e3: 3 mons at {or1010051251037=10.51.251.37: >> 6789/0,or1010051251038=10.51.251.38:6789/0,or1010051251039= >> 10.51.251.39:6789/0} >> election epoch 144, quorum 0,1,2 or1010051251037, >> or1010051251038,or1010051251039 >> osdmap e1814: 40 osds: 40 up, 40 in >> pgmap v446581: 1728 pgs, 4 pools, 7389 GB data, 1847 kobjects >> 22221 GB used, 57472 GB / 79793 GB avail >> 1728 active+clean >> client io 61472 kB/s wr, 30 op/s >> root@or1010051251044:~# >> >> >> Thanks, >> Pardhiv Karri >> >> On Tue, May 22, 2018 at 5:01 AM, David Turner <drakonst...@gmail.com> >> wrote: >> >>> What are your `ceph osd tree` and `ceph status` as well? >>> >>> On Tue, May 22, 2018, 3:05 AM Pardhiv Karri <meher4in...@gmail.com> >>> wrote: >>> >>>> Hi, >>>> >>>> We are using Ceph Hammer 0.94.9. Some of our OSDs never get any data or >>>> PGs even at their full crush weight, up and running. Rest of the OSDs are >>>> at 50% full. Is there a bug in Hammer that is causing this issue? Does >>>> upgrading to Jewel or Luminous fix this issue? >>>> >>>> I tried deleting and recreating this OSD N number of times and still >>>> the same issue. I am seeing this in 3 of our 4 ceph clusters in different >>>> datacenters. We are using HDD as OSD and SSD as Journal drive. >>>> >>>> The below is from our lab and OSD 38 is the one that never fills. >>>> >>>> >>>> ID WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR TYPE NAME >>>> >>>> -1 80.00000 - 0 0 0 0 0 root default >>>> >>>> -2 40.00000 - 39812G 6190G 33521G 15.55 0.68 rack rack_A1 >>>> >>>> -3 20.00000 - 19852G 3718G 16134G 18.73 0.82 host >>>> or1010051251040 >>>> 0 2.00000 1.00000 1861G 450G 1410G 24.21 1.07 osd.0 >>>> >>>> 1 2.00000 1.00000 1999G 325G 1673G 16.29 0.72 osd.1 >>>> >>>> 2 2.00000 1.00000 1999G 336G 1662G 16.85 0.74 osd.2 >>>> >>>> 3 2.00000 1.00000 1999G 386G 1612G 19.35 0.85 osd.3 >>>> >>>> 4 2.00000 1.00000 1999G 385G 1613G 19.30 0.85 osd.4 >>>> >>>> 5 2.00000 1.00000 1999G 364G 1634G 18.21 0.80 osd.5 >>>> >>>> 6 2.00000 1.00000 1999G 319G 1679G 15.99 0.70 osd.6 >>>> >>>> 7 2.00000 1.00000 1999G 434G 1564G 21.73 0.96 osd.7 >>>> >>>> 8 2.00000 1.00000 1999G 352G 1646G 17.63 0.78 osd.8 >>>> >>>> 9 2.00000 1.00000 1999G 362G 1636G 18.12 0.80 osd.9 >>>> >>>> -8 20.00000 - 19959G 2472G 17387G 12.39 0.55 host >>>> or1010051251044 >>>> 30 2.00000 1.00000 1999G 362G 1636G 18.14 0.80 >>>> osd.30 >>>> 31 2.00000 1.00000 1999G 293G 1705G 14.66 0.65 >>>> osd.31 >>>> 32 2.00000 1.00000 1999G 202G 1796G 10.12 0.45 >>>> osd.32 >>>> 33 2.00000 1.00000 1999G 215G 1783G 10.76 0.47 >>>> osd.33 >>>> 34 2.00000 1.00000 1999G 192G 1806G 9.61 0.42 >>>> osd.34 >>>> 35 2.00000 1.00000 1999G 337G 1661G 16.90 0.74 >>>> osd.35 >>>> 36 2.00000 1.00000 1999G 206G 1792G 10.35 0.46 >>>> osd.36 >>>> 37 2.00000 1.00000 1999G 266G 1732G 13.33 0.59 >>>> osd.37 >>>> 38 2.00000 1.00000 1999G 55836k 1998G 0.00 0 >>>> osd.38 >>>> 39 2.00000 1.00000 1968G 396G 1472G 20.12 0.89 >>>> osd.39 >>>> -4 20.00000 - 0 0 0 0 0 rack rack_B1 >>>> >>>> -5 20.00000 - 19990G 5978G 14011G 29.91 1.32 host >>>> or1010051251041 >>>> 10 2.00000 1.00000 1999G 605G 1393G 30.27 1.33 >>>> osd.10 >>>> 11 2.00000 1.00000 1999G 592G 1406G 29.62 1.30 >>>> osd.11 >>>> 12 2.00000 1.00000 1999G 539G 1460G 26.96 1.19 >>>> osd.12 >>>> 13 2.00000 1.00000 1999G 684G 1314G 34.22 1.51 >>>> osd.13 >>>> 14 2.00000 1.00000 1999G 510G 1488G 25.56 1.13 >>>> osd.14 >>>> 15 2.00000 1.00000 1999G 590G 1408G 29.52 1.30 >>>> osd.15 >>>> 16 2.00000 1.00000 1999G 595G 1403G 29.80 1.31 >>>> osd.16 >>>> 17 2.00000 1.00000 1999G 652G 1346G 32.64 1.44 >>>> osd.17 >>>> 18 2.00000 1.00000 1999G 544G 1454G 27.23 1.20 >>>> osd.18 >>>> 19 2.00000 1.00000 1999G 665G 1333G 33.27 1.46 >>>> osd.19 >>>> -9 0 - 0 0 0 0 0 host >>>> or1010051251045 >>>> -6 20.00000 - 0 0 0 0 0 rack rack_C1 >>>> >>>> -7 20.00000 - 19990G 5956G 14033G 29.80 1.31 host >>>> or1010051251042 >>>> 20 2.00000 1.00000 1999G 701G 1297G 35.11 1.55 >>>> osd.20 >>>> 21 2.00000 1.00000 1999G 573G 1425G 28.70 1.26 >>>> osd.21 >>>> 22 2.00000 1.00000 1999G 652G 1346G 32.64 1.44 >>>> osd.22 >>>> 23 2.00000 1.00000 1999G 612G 1386G 30.62 1.35 >>>> osd.23 >>>> 24 2.00000 1.00000 1999G 614G 1384G 30.74 1.35 >>>> osd.24 >>>> 25 2.00000 1.00000 1999G 561G 1437G 28.11 1.24 >>>> osd.25 >>>> 26 2.00000 1.00000 1999G 558G 1440G 27.93 1.23 >>>> osd.26 >>>> 27 2.00000 1.00000 1999G 610G 1388G 30.52 1.34 >>>> osd.27 >>>> 28 2.00000 1.00000 1999G 515G 1483G 25.81 1.14 >>>> osd.28 >>>> 29 2.00000 1.00000 1999G 555G 1443G 27.78 1.22 >>>> osd.29 >>>> -10 0 - 0 0 0 0 0 host >>>> or1010051251046 >>>> -11 0 - 0 0 0 0 0 host >>>> or1010051251023 >>>> TOTAL 79793G 18126G 61566G 22.72 >>>> >>>> MIN/MAX VAR: 0/1.55 STDDEV: 8.26 >>>> >>>> >>>> Thanks >>>> Pardhiv karri >>>> >>>> >>>> _______________________________________________ >>>> ceph-users mailing list >>>> ceph-users@lists.ceph.com >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>> >>> >> >> >> -- >> *Pardhiv Karri* >> "Rise and Rise again until LAMBS become LIONS" >> >> >> -- *Pardhiv Karri* "Rise and Rise again until LAMBS become LIONS"
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com