Yeah,this setting can not see in asok config.
You just set it in ceph.conf and restart mon and osd service(sorry I forget if these restart is necessary)

what I use this config is when I changed crushmap manually,and I do not want the service init script to rebuild crushmap as default way.

maybe this is not siut for your problem.just have a try.

在 Fri, 08 Jan 2016 21:51:32 +0800,Wade Holler <wade.hol...@gmail.com> 写道:

That is not set as far as I can tell. Actually it is strange that I don't see that setting at all.


[root@cpn00001 ~]# ceph daemon osd.0 config show | grep update | grep crush

[root@cpn00001 ~]# grep update /etc/ceph/ceph.conf

[root@cpn00001 ~]#
On Fri, Jan 8, 2016 at 1:50 AM Mart van Santen <m...@greenhost.nl> wrote:


Hi,

Do you have by any chance disabled automatic crushmap updates in your ceph config?

osd crush update on start = false

If this is the case, and you move disks around hosts, they won't update their position/host in the crushmap, even if the crushmap does not reflect reality.
Regards,

Mart





On 01/08/2016 02:16 AM, Wade Holler wrote:
Sure. Apologies for all the text: We have 12 Nodes for OSDs, 15 OSDs per node, but I will only include a sample:

ceph osd tree | head -35

ID  WEIGHT    TYPE NAME         UP/DOWN REWEIGHT PRIMARY-AFFINITY
-1 130.98450 root default -2 5.82153 host cpn00001 4 0.72769 osd.4 up 1.00000 1.00000
14   0.72769         osd.14         up  1.00000          1.00000
 3   0.72769         osd.3          up  1.00000          1.00000
24   0.72769         osd.24         up  1.00000          1.00000
 5   0.72769         osd.5          up  1.00000          1.00000
 2   0.72769         osd.2          up  1.00000          1.00000
17   0.72769         osd.17         up  1.00000          1.00000
69   0.72769         osd.69         up  1.00000          1.00000
-3 6.54922 host cpn00003 7 0.72769 osd.7 up 1.00000 1.00000
 8   0.72769         osd.8          up  1.00000          1.00000
 9   0.72769         osd.9          up  1.00000          1.00000
 0   0.72769         osd.0          up  1.00000          1.00000
28   0.72769         osd.28         up  1.00000          1.00000
10   0.72769         osd.10         up  1.00000          1.00000
 1   0.72769         osd.1          up  1.00000          1.00000
 6   0.72769         osd.6          up  1.00000          1.00000
29   0.72769         osd.29         up  1.00000          1.00000
-4 2.91077 host cpn00004

Compared with the actual processes that are running:



[root@cpx00001 ~]# ssh cpn00001 ps -ef | grep ceph\-osd

ceph 92638 1 26 16:19 ? 01:00:55 /usr/bin/ceph-osd -f --cluster ceph --id 6 --setuser ceph --setgroup ceph

ceph 92667 1 20 16:19 ? 00:48:04 /usr/bin/ceph-osd -f --cluster ceph --id 0 --setuser ceph --setgroup ceph

ceph 92673 1 18 16:19 ? 00:42:48 /usr/bin/ceph-osd -f --cluster ceph --id 8 --setuser ceph --setgroup ceph

ceph 92681 1 19 16:19 ? 00:45:52 /usr/bin/ceph-osd -f --cluster ceph --id 7 --setuser ceph --setgroup ceph

ceph 92701 1 15 16:19 ? 00:36:05 /usr/bin/ceph-osd -f --cluster ceph --id 12 --setuser ceph --setgroup ceph

ceph 92748 1 14 16:19 ? 00:34:07 /usr/bin/ceph-osd -f --cluster ceph --id 10 --setuser ceph --setgroup ceph

ceph 92756 1 16 16:19 ? 00:38:40 /usr/bin/ceph-osd -f --cluster ceph --id 9 --setuser ceph --setgroup ceph

ceph 92758 1 17 16:19 ? 00:39:28 /usr/bin/ceph-osd -f --cluster ceph --id 13 --setuser ceph --setgroup ceph

ceph 92777 1 19 16:19 ? 00:46:17 /usr/bin/ceph-osd -f --cluster ceph --id 1 --setuser ceph --setgroup ceph

ceph 92988 1 18 16:19 ? 00:42:47 /usr/bin/ceph-osd -f --cluster ceph --id 5 --setuser ceph --setgroup ceph

ceph 93058 1 18 16:19 ? 00:43:18 /usr/bin/ceph-osd -f --cluster ceph --id 11 --setuser ceph --setgroup ceph

ceph 93078 1 17 16:19 ? 00:41:38 /usr/bin/ceph-osd -f --cluster ceph --id 14 --setuser ceph --setgroup ceph

ceph 93127 1 15 16:19 ? 00:36:29 /usr/bin/ceph-osd -f --cluster ceph --id 4 --setuser ceph --setgroup ceph

ceph 93130 1 17 16:19 ? 00:40:44 /usr/bin/ceph-osd -f --cluster ceph --id 2 --setuser ceph --setgroup ceph

ceph 93173 1 21 16:19 ? 00:49:37 /usr/bin/ceph-osd -f --cluster ceph --id 3 --setuser ceph --setgroup ceph

[root@cpx00001 ~]# ssh cpn00003 ps -ef | grep ceph\-osd

ceph 82454 1 18 16:19 ? 00:43:58 /usr/bin/ceph-osd -f --cluster ceph --id 25 --setuser ceph --setgroup ceph

ceph 82464 1 24 16:19 ? 00:55:40 /usr/bin/ceph-osd -f --cluster ceph --id 21 --setuser ceph --setgroup ceph

ceph 82473 1 21 16:19 ? 00:50:14 /usr/bin/ceph-osd -f --cluster ceph --id 17 --setuser ceph --setgroup ceph

ceph 82612 1 19 16:19 ? 00:45:25 /usr/bin/ceph-osd -f --cluster ceph --id 22 --setuser ceph --setgroup ceph

ceph 82629 1 20 16:19 ? 00:48:38 /usr/bin/ceph-osd -f --cluster ceph --id 16 --setuser ceph --setgroup ceph

ceph 82651 1 16 16:19 ? 00:39:24 /usr/bin/ceph-osd -f --cluster ceph --id 20 --setuser ceph --setgroup ceph

ceph 82687 1 17 16:19 ? 00:40:31 /usr/bin/ceph-osd -f --cluster ceph --id 18 --setuser ceph --setgroup ceph

ceph 82697 1 26 16:19 ? 01:02:12 /usr/bin/ceph-osd -f --cluster ceph --id 23 --setuser ceph --setgroup ceph

ceph 82719 1 20 16:19 ? 00:47:15 /usr/bin/ceph-osd -f --cluster ceph --id 15 --setuser ceph --setgroup ceph

ceph 82722 1 14 16:19 ? 00:33:41 /usr/bin/ceph-osd -f --cluster ceph --id 28 --setuser ceph --setgroup ceph

ceph 82725 1 14 16:19 ? 00:33:16 /usr/bin/ceph-osd -f --cluster ceph --id 26 --setuser ceph --setgroup ceph

ceph 82743 1 14 16:19 ? 00:34:17 /usr/bin/ceph-osd -f --cluster ceph --id 29 --setuser ceph --setgroup ceph

ceph 82769 1 19 16:19 ? 00:46:00 /usr/bin/ceph-osd -f --cluster ceph --id 19 --setuser ceph --setgroup ceph

ceph 82816 1 13 16:19 ? 00:30:26 /usr/bin/ceph-osd -f --cluster ceph --id 27 --setuser ceph --setgroup ceph

ceph 82828 1 27 16:19 ? 01:04:38 /usr/bin/ceph-osd -f --cluster ceph --id 24 --setuser ceph --setgroup ceph


[root@cpx00001 ~]#



Looks like the crushmap is bad also:

(Cluster appears to be operating ok but this really concerns me.)

# begin crush map

tunable choose_local_tries 0

tunable choose_local_fallback_tries 0

tunable choose_total_tries 50

tunable chooseleaf_descend_once 1

tunable straw_calc_version 1



# devices

device 0 osd.0

device 1 osd.1

device 2 osd.2

device 3 osd.3

device 4 osd.4

device 5 osd.5

device 6 osd.6

device 7 osd.7

device 8 osd.8

device 9 osd.9

device 10 osd.10

device 11 osd.11

device 12 osd.12

device 13 osd.13

device 14 osd.14

device 15 osd.15

device 16 osd.16

device 17 osd.17

device 18 osd.18

device 19 osd.19

device 20 osd.20

device 21 osd.21

device 22 osd.22

device 23 osd.23

device 24 osd.24

device 25 osd.25

device 26 osd.26

device 27 osd.27

device 28 osd.28

device 29 osd.29

device 30 osd.30

device 31 osd.31

device 32 osd.32

device 33 osd.33

device 34 osd.34

device 35 osd.35

device 36 osd.36


device 37 osd.37

...



# types

type 0 osd

type 1 host

type 2 chassis

type 3 rack

type 4 row

type 5 pdu

type 6 pod

type 7 room

type 8 datacenter

type 9 region

type 10 root



# buckets

host cpn00001 {

       id -2           # do not change unnecessarily

       # weight 5.822

       alg straw

       hash 0  # rjenkins1

       item osd.4 weight 0.728

       item osd.14 weight 0.728

       item osd.3 weight 0.728

       item osd.24 weight 0.728

       item osd.5 weight 0.728

       item osd.2 weight 0.728

       item osd.17 weight 0.728

       item osd.69 weight 0.728

}

host cpn00003 {

       id -3           # do not change unnecessarily

       # weight 6.549

       alg straw

       hash 0  # rjenkins1

       item osd.7 weight 0.728

       item osd.8 weight 0.728

       item osd.9 weight 0.728

       item osd.0 weight 0.728

       item osd.28 weight 0.728

       item osd.10 weight 0.728

       item osd.1 weight 0.728

       item osd.6 weight 0.728

       item osd.29 weight 0.728

}


host cpn00004 {....





Thank you for your review !!!!!

Wade







On Thu, Jan 7, 2016 at 6:03 PM Shinobu Kinjo <ski...@redhat.com> wrote:
Can you share the output with us?

Rgds,
Shinobu

----- Original Message -----
From: "Wade Holler" <wade.hol...@gmail.com>
To: "ceph-users" <ceph-users@lists.ceph.com>
Sent: Friday, January 8, 2016 7:29:07 AM
Subject: [ceph-users] ceph osd tree output

Sometimes my ceph osd tree output is wrong. Ie. Wrong osds on the wrong hosts ?

Anyone else have this issue?

I have seen this at Infernalis and Jewell.

Thanks
Wade

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



--
使用Opera的电子邮件客户端:http://www.opera.com/mail/
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to