Glen, It appears you have 447G, 931G, and 558G disks in your cluster, all with a weight of 1.0. This means that although the new disks are bigger, they are not going to be utilized by pgs any more than any other disk.
I would suggest reweighting your other disks (they are smaller), so that you balance your cluster. You should do this gradually over time, preferably during off-peak times, when remapping will not affect operations. I do a little math, first by taking total cluster capacity and dividing it by total capacity of each bucket. I then do the same thing in each bucket, until everything is proportioned appropriately down to the osds. On Fri, Jul 20, 2018 at 8:43 PM, Glen Baars <g...@onsitecomputers.com.au> wrote: > Hello Ceph Users, > > > > We have added more ssd storage to our ceph cluster last night. We added 4 > x 1TB drives and the available space went from 1.6TB to 0.6TB ( in `ceph > df` for the SSD pool ). > > > > I would assume that the weight needs to be changed but I didn’t think I > would need to? Should I change them to 0.75 from 0.9 and hopefully it will > rebalance correctly? > > > > #ceph osd tree | grep -v hdd > > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF > > -1 534.60309 root default > > -19 62.90637 host NAS-AUBUN-RK2-CEPH06 > > 115 ssd 0.43660 osd.115 up 1.00000 1.00000 > > 116 ssd 0.43660 osd.116 up 1.00000 1.00000 > > 117 ssd 0.43660 osd.117 up 1.00000 1.00000 > > 118 ssd 0.43660 osd.118 up 1.00000 1.00000 > > -22 105.51169 host NAS-AUBUN-RK2-CEPH07 > > 138 ssd 0.90970 osd.138 up 1.00000 1.00000 > Added > > 139 ssd 0.90970 osd.139 up 1.00000 1.00000 > Added > > -25 105.51169 host NAS-AUBUN-RK2-CEPH08 > > 140 ssd 0.90970 osd.140 up 1.00000 1.00000 > Added > > 141 ssd 0.90970 osd.141 up 1.00000 1.00000 > Added > > -3 56.32617 host NAS-AUBUN-RK3-CEPH01 > > 60 ssd 0.43660 osd.60 up 1.00000 1.00000 > > 61 ssd 0.43660 osd.61 up 1.00000 1.00000 > > 62 ssd 0.43660 osd.62 up 1.00000 1.00000 > > 63 ssd 0.43660 osd.63 up 1.00000 1.00000 > > -5 56.32617 host NAS-AUBUN-RK3-CEPH02 > > 64 ssd 0.43660 osd.64 up 1.00000 1.00000 > > 65 ssd 0.43660 osd.65 up 1.00000 1.00000 > > 66 ssd 0.43660 osd.66 up 1.00000 1.00000 > > 67 ssd 0.43660 osd.67 up 1.00000 1.00000 > > -7 56.32617 host NAS-AUBUN-RK3-CEPH03 > > 68 ssd 0.43660 osd.68 up 1.00000 1.00000 > > 69 ssd 0.43660 osd.69 up 1.00000 1.00000 > > 70 ssd 0.43660 osd.70 up 1.00000 1.00000 > > 71 ssd 0.43660 osd.71 up 1.00000 1.00000 > > -13 45.84741 host NAS-AUBUN-RK3-CEPH04 > > 72 ssd 0.54579 osd.72 up 1.00000 1.00000 > > 73 ssd 0.54579 osd.73 up 1.00000 1.00000 > > 76 ssd 0.54579 osd.76 up 1.00000 1.00000 > > 77 ssd 0.54579 osd.77 up 1.00000 1.00000 > > -16 45.84741 host NAS-AUBUN-RK3-CEPH05 > > 74 ssd 0.54579 osd.74 up 1.00000 1.00000 > > 75 ssd 0.54579 osd.75 up 1.00000 1.00000 > > 78 ssd 0.54579 osd.78 up 1.00000 1.00000 > > 79 ssd 0.54579 osd.79 up 1.00000 1.00000 > > > > # ceph osd df | grep -v hdd > > ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS > > 115 ssd 0.43660 1.00000 447G 250G 196G 56.00 1.72 103 > > 116 ssd 0.43660 1.00000 447G 191G 255G 42.89 1.32 84 > > 117 ssd 0.43660 1.00000 447G 213G 233G 47.79 1.47 92 > > 118 ssd 0.43660 1.00000 447G 208G 238G 46.61 1.43 85 > > 138 ssd 0.90970 1.00000 931G 820G 111G 88.08 2.71 216 Added > > 139 ssd 0.90970 1.00000 931G 771G 159G 82.85 2.55 207 Added > > 140 ssd 0.90970 1.00000 931G 709G 222G 76.12 2.34 197 Added > > 141 ssd 0.90970 1.00000 931G 664G 267G 71.31 2.19 184 Added > > 60 ssd 0.43660 1.00000 447G 275G 171G 61.62 1.89 100 > > 61 ssd 0.43660 1.00000 447G 237G 209G 53.04 1.63 90 > > 62 ssd 0.43660 1.00000 447G 275G 171G 61.58 1.89 95 > > 63 ssd 0.43660 1.00000 447G 260G 187G 58.15 1.79 97 > > 64 ssd 0.43660 1.00000 447G 232G 214G 52.08 1.60 83 > > 65 ssd 0.43660 1.00000 447G 207G 239G 46.36 1.42 75 > > 66 ssd 0.43660 1.00000 447G 217G 230G 48.54 1.49 84 > > 67 ssd 0.43660 1.00000 447G 252G 195G 56.36 1.73 92 > > 68 ssd 0.43660 1.00000 447G 248G 198G 55.56 1.71 94 > > 69 ssd 0.43660 1.00000 447G 229G 217G 51.25 1.57 84 > > 70 ssd 0.43660 1.00000 447G 259G 187G 58.01 1.78 87 > > 71 ssd 0.43660 1.00000 447G 267G 179G 59.83 1.84 97 > > 72 ssd 0.54579 1.00000 558G 217G 341G 38.96 1.20 100 > > 73 ssd 0.54579 1.00000 558G 283G 275G 50.75 1.56 121 > > 76 ssd 0.54579 1.00000 558G 286G 272G 51.33 1.58 129 > > 77 ssd 0.54579 1.00000 558G 246G 312G 44.07 1.35 104 > > 74 ssd 0.54579 1.00000 558G 273G 285G 48.91 1.50 122 > > 75 ssd 0.54579 1.00000 558G 281G 276G 50.45 1.55 114 > > 78 ssd 0.54579 1.00000 558G 289G 269G 51.80 1.59 133 > > 79 ssd 0.54579 1.00000 558G 276G 282G 49.39 1.52 119 > > Kind regards, > > *Glen Baars* > > BackOnline Manager > > > This e-mail is intended solely for the benefit of the addressee(s) and any > other named recipient. It is confidential and may contain legally > privileged or confidential information. If you are not the recipient, any > use, distribution, disclosure or copying of this e-mail is prohibited. The > confidentiality and legal privilege attached to this communication is not > waived or lost by reason of the mistaken transmission or delivery to you. > If you have received this e-mail in error, please notify us immediately. > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > -- Shawn Iverson, CETL Director of Technology Rush County Schools 765-932-3901 x1171 ivers...@rushville.k12.in.us
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com