> 
> Helllo,
> 
> I've reviewed some recent posts in this list and also searched Google for
> info about autoscale and overlapping roots.  In what I have found I do not
> see anything that I can understand regarding how to fix the issue -
> probably because I don't deal with Crush on a regular basis.


Checkout the Note in this section:  
https://docs.ceph.com/en/reef/rados/operations/placement-groups/#viewing-pg-scaling-recommendations

I added that last year I think it was as a result of how Rook was creating 
pools.

> 
> From what I read and looking at 'ceph osd crush rule dump', it looks like
> the 8 replicated pools have
> 
>                    "op": "take",
>                    "item": -1,
>                    "item_name": "default"
> 
> whereas the 2 EC pools have
> 
>                    "op": "take",
>                    "item": -2,
>                    "item_name": "default~hdd"
> 
> To be sure, all of my OSDs are identical - HDD with SSD WAL/DB.
> 
> Please advise on how to fix this.

The subtlety that's easy to miss is that when you specify a device class for 
only *some* pools, the pools/rules that specify a device class effectively act 
on a "shadow" CRUSH root.  My terminology may be inexact there.

So I think if you adjust your CRUSH rules so that they all specify a device 
class -- in your case all the same device class -- your problem (and balancer 
performance perhaps) will improve.


_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to