That is not set as far as I can tell.  Actually it is strange that I don't
see that setting at all.

[root@cpn00001 ~]# ceph daemon osd.0 config show | grep update | grep crush

[root@cpn00001 ~]# grep update /etc/ceph/ceph.conf

[root@cpn00001 ~]#

On Fri, Jan 8, 2016 at 1:50 AM Mart van Santen <m...@greenhost.nl> wrote:

>
>
> Hi,
>
> Do you have by any chance disabled automatic crushmap updates in your ceph
> config?
>
> osd crush update on start = false
>
> If this is the case, and you move disks around hosts, they won't update
> their position/host in the crushmap, even if the crushmap does not reflect
> reality.
>
> Regards,
>
> Mart
>
>
>
>
>
> On 01/08/2016 02:16 AM, Wade Holler wrote:
>
> Sure.  Apologies for all the text: We have 12 Nodes for OSDs, 15 OSDs per
> node,  but I will only include a sample:
>
> ceph osd tree | head -35
>
> ID  WEIGHT    TYPE NAME         UP/DOWN REWEIGHT PRIMARY-AFFINITY
>
>  -1 130.98450 root default
>
>  -2   5.82153     host cpn00001
>
>   4   0.72769         osd.4          up  1.00000          1.00000
>
>  14   0.72769         osd.14         up  1.00000          1.00000
>
>   3   0.72769         osd.3          up  1.00000          1.00000
>
>  24   0.72769         osd.24         up  1.00000          1.00000
>
>   5   0.72769         osd.5          up  1.00000          1.00000
>
>   2   0.72769         osd.2          up  1.00000          1.00000
>
>  17   0.72769         osd.17         up  1.00000          1.00000
>
>  69   0.72769         osd.69         up  1.00000          1.00000
>
>  -3   6.54922     host cpn00003
>
>   7   0.72769         osd.7          up  1.00000          1.00000
>
>   8   0.72769         osd.8          up  1.00000          1.00000
>
>   9   0.72769         osd.9          up  1.00000          1.00000
>
>   0   0.72769         osd.0          up  1.00000          1.00000
>
>  28   0.72769         osd.28         up  1.00000          1.00000
>
>  10   0.72769         osd.10         up  1.00000          1.00000
>
>   1   0.72769         osd.1          up  1.00000          1.00000
>
>   6   0.72769         osd.6          up  1.00000          1.00000
>
>  29   0.72769         osd.29         up  1.00000          1.00000
>
>  -4   2.91077     host cpn00004
>
>
> Compared with the actual processes that are running:
>
>
> [root@cpx00001 ~]# ssh cpn00001 ps -ef | grep ceph\-osd
>
> ceph       92638       1 26 16:19 ?        01:00:55 /usr/bin/ceph-osd -f
> --cluster ceph --id 6 --setuser ceph --setgroup ceph
>
> ceph       92667       1 20 16:19 ?        00:48:04 /usr/bin/ceph-osd -f
> --cluster ceph --id 0 --setuser ceph --setgroup ceph
>
> ceph       92673       1 18 16:19 ?        00:42:48 /usr/bin/ceph-osd -f
> --cluster ceph --id 8 --setuser ceph --setgroup ceph
>
> ceph       92681       1 19 16:19 ?        00:45:52 /usr/bin/ceph-osd -f
> --cluster ceph --id 7 --setuser ceph --setgroup ceph
>
> ceph       92701       1 15 16:19 ?        00:36:05 /usr/bin/ceph-osd -f
> --cluster ceph --id 12 --setuser ceph --setgroup ceph
>
> ceph       92748       1 14 16:19 ?        00:34:07 /usr/bin/ceph-osd -f
> --cluster ceph --id 10 --setuser ceph --setgroup ceph
>
> ceph       92756       1 16 16:19 ?        00:38:40 /usr/bin/ceph-osd -f
> --cluster ceph --id 9 --setuser ceph --setgroup ceph
>
> ceph       92758       1 17 16:19 ?        00:39:28 /usr/bin/ceph-osd -f
> --cluster ceph --id 13 --setuser ceph --setgroup ceph
>
> ceph       92777       1 19 16:19 ?        00:46:17 /usr/bin/ceph-osd -f
> --cluster ceph --id 1 --setuser ceph --setgroup ceph
>
> ceph       92988       1 18 16:19 ?        00:42:47 /usr/bin/ceph-osd -f
> --cluster ceph --id 5 --setuser ceph --setgroup ceph
>
> ceph       93058       1 18 16:19 ?        00:43:18 /usr/bin/ceph-osd -f
> --cluster ceph --id 11 --setuser ceph --setgroup ceph
>
> ceph       93078       1 17 16:19 ?        00:41:38 /usr/bin/ceph-osd -f
> --cluster ceph --id 14 --setuser ceph --setgroup ceph
>
> ceph       93127       1 15 16:19 ?        00:36:29 /usr/bin/ceph-osd -f
> --cluster ceph --id 4 --setuser ceph --setgroup ceph
>
> ceph       93130       1 17 16:19 ?        00:40:44 /usr/bin/ceph-osd -f
> --cluster ceph --id 2 --setuser ceph --setgroup ceph
>
> ceph       93173       1 21 16:19 ?        00:49:37 /usr/bin/ceph-osd -f
> --cluster ceph --id 3 --setuser ceph --setgroup ceph
>
> [root@cpx00001 ~]# ssh cpn00003 ps -ef | grep ceph\-osd
>
> ceph       82454       1 18 16:19 ?        00:43:58 /usr/bin/ceph-osd -f
> --cluster ceph --id 25 --setuser ceph --setgroup ceph
>
> ceph       82464       1 24 16:19 ?        00:55:40 /usr/bin/ceph-osd -f
> --cluster ceph --id 21 --setuser ceph --setgroup ceph
>
> ceph       82473       1 21 16:19 ?        00:50:14 /usr/bin/ceph-osd -f
> --cluster ceph --id 17 --setuser ceph --setgroup ceph
>
> ceph       82612       1 19 16:19 ?        00:45:25 /usr/bin/ceph-osd -f
> --cluster ceph --id 22 --setuser ceph --setgroup ceph
>
> ceph       82629       1 20 16:19 ?        00:48:38 /usr/bin/ceph-osd -f
> --cluster ceph --id 16 --setuser ceph --setgroup ceph
>
> ceph       82651       1 16 16:19 ?        00:39:24 /usr/bin/ceph-osd -f
> --cluster ceph --id 20 --setuser ceph --setgroup ceph
>
> ceph       82687       1 17 16:19 ?        00:40:31 /usr/bin/ceph-osd -f
> --cluster ceph --id 18 --setuser ceph --setgroup ceph
>
> ceph       82697       1 26 16:19 ?        01:02:12 /usr/bin/ceph-osd -f
> --cluster ceph --id 23 --setuser ceph --setgroup ceph
>
> ceph       82719       1 20 16:19 ?        00:47:15 /usr/bin/ceph-osd -f
> --cluster ceph --id 15 --setuser ceph --setgroup ceph
>
> ceph       82722       1 14 16:19 ?        00:33:41 /usr/bin/ceph-osd -f
> --cluster ceph --id 28 --setuser ceph --setgroup ceph
>
> ceph       82725       1 14 16:19 ?        00:33:16 /usr/bin/ceph-osd -f
> --cluster ceph --id 26 --setuser ceph --setgroup ceph
>
> ceph       82743       1 14 16:19 ?        00:34:17 /usr/bin/ceph-osd -f
> --cluster ceph --id 29 --setuser ceph --setgroup ceph
>
> ceph       82769       1 19 16:19 ?        00:46:00 /usr/bin/ceph-osd -f
> --cluster ceph --id 19 --setuser ceph --setgroup ceph
>
> ceph       82816       1 13 16:19 ?        00:30:26 /usr/bin/ceph-osd -f
> --cluster ceph --id 27 --setuser ceph --setgroup ceph
>
> ceph       82828       1 27 16:19 ?        01:04:38 /usr/bin/ceph-osd -f
> --cluster ceph --id 24 --setuser ceph --setgroup ceph
>
> [root@cpx00001 ~]#
>
>
> Looks like the crushmap is bad also:
>
> (Cluster appears to be operating ok but this really concerns me.)
>
> # begin crush map
>
> tunable choose_local_tries 0
>
> tunable choose_local_fallback_tries 0
>
> tunable choose_total_tries 50
>
> tunable chooseleaf_descend_once 1
>
> tunable straw_calc_version 1
>
>
> # devices
>
> device 0 osd.0
>
> device 1 osd.1
>
> device 2 osd.2
>
> device 3 osd.3
>
> device 4 osd.4
>
> device 5 osd.5
>
> device 6 osd.6
>
> device 7 osd.7
>
> device 8 osd.8
>
> device 9 osd.9
>
> device 10 osd.10
>
> device 11 osd.11
>
> device 12 osd.12
>
> device 13 osd.13
>
> device 14 osd.14
>
> device 15 osd.15
>
> device 16 osd.16
>
> device 17 osd.17
>
> device 18 osd.18
>
> device 19 osd.19
>
> device 20 osd.20
>
> device 21 osd.21
>
> device 22 osd.22
>
> device 23 osd.23
>
> device 24 osd.24
>
> device 25 osd.25
>
> device 26 osd.26
>
> device 27 osd.27
>
> device 28 osd.28
>
> device 29 osd.29
>
> device 30 osd.30
>
> device 31 osd.31
>
> device 32 osd.32
>
> device 33 osd.33
>
> device 34 osd.34
>
> device 35 osd.35
>
> device 36 osd.36
>
> device 37 osd.37
>
> ...
>
>
> # types
>
> type 0 osd
>
> type 1 host
>
> type 2 chassis
>
> type 3 rack
>
> type 4 row
>
> type 5 pdu
>
> type 6 pod
>
> type 7 room
>
> type 8 datacenter
>
> type 9 region
>
> type 10 root
>
>
> # buckets
>
> host cpn00001 {
>
>         id -2           # do not change unnecessarily
>
>         # weight 5.822
>
>         alg straw
>
>         hash 0  # rjenkins1
>
>         item osd.4 weight 0.728
>
>         item osd.14 weight 0.728
>
>         item osd.3 weight 0.728
>
>         item osd.24 weight 0.728
>
>         item osd.5 weight 0.728
>
>         item osd.2 weight 0.728
>
>         item osd.17 weight 0.728
>
>         item osd.69 weight 0.728
>
> }
>
> host cpn00003 {
>
>         id -3           # do not change unnecessarily
>
>         # weight 6.549
>
>         alg straw
>
>         hash 0  # rjenkins1
>
>         item osd.7 weight 0.728
>
>         item osd.8 weight 0.728
>
>         item osd.9 weight 0.728
>
>         item osd.0 weight 0.728
>
>         item osd.28 weight 0.728
>
>         item osd.10 weight 0.728
>
>         item osd.1 weight 0.728
>
>         item osd.6 weight 0.728
>
>         item osd.29 weight 0.728
>
> }
>
> host cpn00004 {....
>
>
>
> Thank you for your review !!!!!
>
> Wade
>
>
>
>
>
> On Thu, Jan 7, 2016 at 6:03 PM Shinobu Kinjo <ski...@redhat.com> wrote:
>
>> Can you share the output with us?
>>
>> Rgds,
>> Shinobu
>>
>> ----- Original Message -----
>> From: "Wade Holler" <wade.hol...@gmail.com>
>> To: "ceph-users" <ceph-users@lists.ceph.com>
>> Sent: Friday, January 8, 2016 7:29:07 AM
>> Subject: [ceph-users] ceph osd tree output
>>
>> Sometimes my ceph osd tree output is wrong. Ie. Wrong osds on the wrong
>> hosts ?
>>
>> Anyone else have this issue?
>>
>> I have seen this at Infernalis and Jewell.
>>
>> Thanks
>> Wade
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
> _______________________________________________
> ceph-users mailing 
> listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to