I have 2 clusters; both have HDDs and SSDs. Reporting only the HDDs which have 
their own pools:

"rep-cluster": hdd-pool 3-replication,   86 OSDs (16 TiB each), 1024 PGs, 78 
%RAW USED, 100 M objects
"ec-cluster":  hdd-pool erasure k=4 m=2, 58 OSDs (16 TiB each),  256 PGs, 60 
%RAW USED, 450 M objects

Both are Ceph 18.2.1, Bluestore, and have the autoscaler enabled.
As you can see, I have many small objects.

My PGs-copies-per-OSD seem far off from the recommendation of 100 PGs per OSD 
(`mon_target_pg_per_osd`):

rep-cluster: 35 PGs/OSD (= 1024*3/86)
ec-cluster:  26 PGs/OSD (= 256*6/58)

So I'm at least 3x-4x off.
Why?
Should the autoscaler not have increased the PGs here?

`ceph osd pool autoscale-status`:

rep-cluster:
  POOL        SIZE  TARGET SIZE  RATE  RAW CAPACITY   RATIO  TARGET RATIO  
EFFECTIVE RATIO  BIAS  PG_NUM  NEW PG_NUM  AUTOSCALE  BULK
  data      349.4T                3.0         1343T  0.7802                     
             1.0    1024              on         False
ec-cluster:
  data_ec   347.3T                1.5        876.4T  0.5944                     
             1.0     256              on         False


I believe that because of this I suffer some drawbacks:

* On ec-cluster, a PG contains ~2 TiB and ~2 M objects, causing rebalances to 
happen in coarse, slow steps.

Should I take some steps to force the autoscaler to increase PGs, and if yes, 
which approach would be best here?

Thanks for your tips!
Niklas
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to