You can adjust the Primary Affinity down on the larger drives so they’ll get less read load. In one test I’ve seen this result in a 10-15% increase in read throughout but it depends on your situation.
Optimal settings would require calculations that make my head hurt, maybe someone has a tool but I haven’t seen it. You might want to ensure that the drives are spread evenly across your (unspecified) failure domains so that the extra capacity isn’t wasted, again depending on your topology. > On Mar 31, 2020, at 8:49 AM, Eneko Lacunza <elacu...@binovo.es> wrote: > > Hi Andras, > >> El 31/3/20 a las 16:42, Andras Pataki escribió: >> I'm looking for some advice on what to do about drives of different sizes in >> the same cluster. >> >> We have so far kept the drive sizes consistent on our main ceph cluster >> (using 8TB drives). We're getting some new hardware with larger, 12TB >> drives next, and I'm pondering on how best to configure them. If I just >> simply add them, they will have 1.5x the data (which is less of a problem), >> but will also get 1.5x the iops - so I presume it will slow the whole >> cluster down as a result (these drives will be busy, and the rest will not >> be as much). I'm wondering how people generally handle this. >> >> I'm more concerned about these larger drives being busier than the rest - so >> I'd like to be able to put for example 1/3 drive of less accessed data on >> them in addition to the usual data - to use the extra capacity but not >> increase the load on them. Is there an easy way to accomplish this? One >> possibility is to run two OSDs on the drive (in two crush hierarchies), >> which isn't ideal. Can I just run one OSD somehow and put it into two crush >> roots, or something similar? > You should adjust the weight of the new 12TB disk OSDs, to match the weight > of the current 8TB OSDs. > > That will make the new disks the same as the old disks to Ceph :-) But you'll > lost the extra 4TB space, until you remove the 8TB disks from the cluster... > > Cheers > Eneko > > -- > Zuzendari Teknikoa / Director Técnico > Binovo IT Human Project, S.L. > Telf. 943569206 > Astigarragako bidea 2, 2º izq. oficina 11; 20180 Oiartzun (Gipuzkoa) > www.binovo.es > _______________________________________________ > ceph-users mailing list -- ceph-users@ceph.io > To unsubscribe send an email to ceph-users-le...@ceph.io _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io