Some time ago on Luminous I also had to change the crush rules on a all 
hdd cluster to hdd (to prepare for adding ssd's and ssd pools). And pg's 
started migrating while everything already was on hdd's, looks like this 
is still not fixed?





-----Original Message-----
From: Raymond Berg Hansen [mailto:raymon...@gmail.com] 
Sent: dinsdag 1 oktober 2019 14:32
To: ceph-users@ceph.io
Subject: [ceph-users] Re: Nautilus pg autoscale, data lost?

You are absolutly right, I had made a crush rule for device class hdd. 
Did not put this in connection with this problem. When I put the pools 
back in the default crush rule things are starting to fix itself it 
seems.
Have I done something wrong with this crush rule?

# rules
rule replicated_rule {
        id 0
        type replicated
        min_size 1
        max_size 10
        step take default
        step chooseleaf firstn 0 type host
        step emit
}
rule replicated-hdd {
        id 1
        type replicated
        min_size 1
        max_size 10
        step take default class hdd
        step chooseleaf firstn 0 type datacenter
        step emit
}
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an 
email to ceph-users-le...@ceph.io

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to