Hi, Robert, Here is my crush map.
# begin crush map tunable choose_local_tries 0 tunable choose_total_tries 50 tunable chooseleaf_descend_once 1 tunable chooseleaf_vary_r 1 # devices device 0 osd.0 device 1 osd.1 device 2 osd.2 device 3 osd.3 device 4 osd.4 device 5 osd.5 device 6 osd.6 device 7 osd.7 device 8 osd.8 device 9 osd.9 # types type 0 osd type 1 host type 2 zone type 3 storage_group type 4 root # buckets host controller_performance_zone_one { id -1 # do not change unnecessarily alg straw hash 0 # rjenkins1 item osd.2 weight 1.000 } host controller_capacity_zone_one { id -2 # do not change unnecessarily alg straw hash 0 # rjenkins1 item osd.0 weight 1.000 item osd.1 weight 1.000 } host compute2_performance_zone_one { id -3 # do not change unnecessarily alg straw hash 0 # rjenkins1 item osd.5 weight 1.000 } host compute2_capacity_zone_one { id -4 # do not change unnecessarily alg straw hash 0 # rjenkins1 item osd.3 weight 1.000 item osd.4 weight 1.000 } host compute3_performance_zone_one { id -5 # do not change unnecessarily alg straw hash 0 # rjenkins1 } host compute3_capacity_zone_one { id -6 # do not change unnecessarily alg straw hash 0 # rjenkins1 item osd.6 weight 1.000 item osd.7 weight 1.000 } zone zone_one_performance { id -7 # do not change unnecessarily alg straw hash 0 # rjenkins1 item controller_performance_zone_one weight 1.000 item compute2_performance_zone_one weight 1.000 item compute3_performance_zone_one weight 0.100 } host compute4_capacity_zone_one { id -12 # do not change unnecessarily alg straw hash 0 # rjenkins1 item osd.8 weight 1.000 item osd.9 weight 1.000 } zone zone_one_capacity { id -8 # do not change unnecessarily alg straw hash 0 # rjenkins1 item controller_capacity_zone_one weight 2.000 item compute2_capacity_zone_one weight 2.000 item compute3_capacity_zone_one weight 2.000 item compute4_capacity_zone_one weight 2.000 } storage_group performance { id -9 # do not change unnecessarily alg straw hash 0 # rjenkins1 item zone_one_performance weight 2.100 } storage_group capacity { id -10 # do not change unnecessarily alg straw hash 0 # rjenkins1 item zone_one_capacity weight 8.000 } root vsm { id -11 # do not change unnecessarily alg straw hash 0 # rjenkins1 item performance weight 2.100 item capacity weight 8.000 } # rules rule capacity { ruleset 0 type replicated min_size 0 max_size 10 step take capacity step chooseleaf firstn 0 type host step emit } rule performance { ruleset 1 type replicated min_size 0 max_size 10 step take performance step chooseleaf firstn 0 type host step emit } # end crush map ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY -11 10.09999 root vsm -9 2.09999 storage_group performance -7 2.09999 zone zone_one_performance -1 1.00000 host controller_performance_zone_one 2 1.00000 osd.2 up 1.00000 1.00000 -3 1.00000 host compute2_performance_zone_one 5 1.00000 osd.5 up 1.00000 1.00000 -5 0.09999 host compute3_performance_zone_one -10 8.00000 storage_group capacity -8 8.00000 zone zone_one_capacity -2 2.00000 host controller_capacity_zone_one 0 1.00000 osd.0 up 1.00000 1.00000 1 1.00000 osd.1 up 1.00000 1.00000 -4 2.00000 host compute2_capacity_zone_one 3 1.00000 osd.3 up 1.00000 1.00000 4 1.00000 osd.4 up 1.00000 1.00000 -6 2.00000 host compute3_capacity_zone_one 6 1.00000 osd.6 up 1.00000 1.00000 7 1.00000 osd.7 up 1.00000 1.00000 -12 2.00000 host compute4_capacity_zone_one 8 1.00000 osd.8 up 1.00000 1.00000 9 1.00000 osd.9 up 1.00000 1.00000 And here are the stats of the pool i was using for tests: root@iclcompute4:# ceph osd pool get Gold crush_ruleset crush_ruleset: 0 root@iclcompute4:# ceph osd pool get Gold size size: 3 root@iclcompute4:# ceph osd pool get Gold min_size min_size: 1 IO freeze happens whetger when I add or remove host with 2 OSD. I just did it with standard manual Ceph procedure at http://docs.ceph.com/docs/master/rados/operations/add-or-rm-osds/ to be sure that scripting mistakes are not involved. So freeze lasts until cluster says OK, then resumes. Regards, Vasily. On Thu, May 14, 2015 at 6:44 PM, Robert LeBlanc <rob...@leblancnet.us> wrote: > -----BEGIN PGP SIGNED MESSAGE----- > Hash: SHA256 > > Can you provide the output of the CRUSH map and a copy of the script that you > are using to add the OSDs? Can you also provide the pool size and pool > min_size? > -----BEGIN PGP SIGNATURE----- > Version: Mailvelope v0.13.1 > Comment: https://www.mailvelope.com > > wsFcBAEBCAAQBQJVVMLvCRDmVDuy+mK58QAAHVIQALIZ8aOWE5P8DkRe+8pz > XS+rMdA17nPUd2mX6PIqhjBxetrUhIjQUho8HSIswT9JVkjVSIj+QHs5CI1C > 6ArWIPt/U8L78d1hI8NuH/vWwWydYfV32n2L2LExIgUpFAbJA81AnjjDFLvo > T63KLitQ1wz8lyhAWXp4ze15CgAv1u9VbJhazeeWunxZxd8eSGuUS8RTdhLD > sD0pSQnVT4W04TSKYfvbUlpqm68wGY+MApnuQXdpC0jBLcDz0OSu1P+OQC03 > 0vBCERY1er/rSskJ6TRrQGLzXAc/vc3HbPMvegIhp2voeXgONdO5P/qLfSfD > ZwVUoi6EfFe+na3S4rEjOeBU+v2P00komVEcvjOJDQb3IVcE23iVJOezk3p+ > AgJqOz9VLdGvdmZTZnR08PKPZEja80QzrSklRW5f8JyjKlbE8tB5lBoM5mKo > oRcBSDbGSKvXInqygQ3XLdxULHaXbNqNPj+JvPbmfkTU6Iq6pXqcBdUSqG0o > /5Rx16+2Rouz4f8uu5irmDjz0ivKL6QCIzBwZbBTdLIwqhf9vCl1ACDWq4U3 > DMorcafZbMArdOqlkVhQJiMioZEQ8U/ThY2bInkNdhii/2A35CToyOfMKyfq > FLAK5lCiM6gRfCkEBPTwkDR6GNAfgY7khz34adsBRlZPB6a3MeucAGtTjyWt > AJIV > =bcYd > -----END PGP SIGNATURE----- > > > ---------------- > Robert LeBlanc > GPG Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1 > > On Thu, May 14, 2015 at 6:33 AM, Vasiliy Angapov <anga...@gmail.com> > wrote: > >> Thanks, Robert, for sharing so many experience! I feel like I don't >> deserve it :) >> >> I have another but very same situation which I don't understand. >> Last time i tried to hard kill OSD daemons. >> This time i add a new node with 2 OSDs to my cluster and also monitor the >> IO. I wrote a script which adds a node with OSDs fully automatically. And >> seems like when I start the script - an IO is also blocked until the >> cluster shows HEALTH_OK which takes quite an amount of time. After Ceph >> status is OK - copying resumes. >> >> What should I tune this time to avoid long IO interuption? >> >> Thanks in advance again :) >> > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com