It is not set in the conf file.  So why do I still have this behavior ?

On Fri, Jan 8, 2016 at 11:08 AM hnuzhoulin <hnuzhoul...@gmail.com> wrote:

> Yeah,this setting can not see in asok config.
> You just set it in ceph.conf and restart mon and osd service(sorry I
> forget if these restart is necessary)
>
> what I use this config is when I changed crushmap manually,and I do not
> want the service init script to rebuild crushmap as default way.
>
> maybe this is not siut for your problem.just have a try.
>
> 在 Fri, 08 Jan 2016 21:51:32 +0800,Wade Holler <wade.hol...@gmail.com> 写道:
>
> That is not set as far as I can tell.  Actually it is strange that I don't
> see that setting at all.
>
> [root@cpn00001 ~]# ceph daemon osd.0 config show | grep update | grep
> crush
>
> [root@cpn00001 ~]# grep update /etc/ceph/ceph.conf
>
> [root@cpn00001 ~]#
>
> On Fri, Jan 8, 2016 at 1:50 AM Mart van Santen <m...@greenhost.nl> wrote:
>
>>
>>
>> Hi,
>>
>> Do you have by any chance disabled automatic crushmap updates in your
>> ceph config?
>>
>> osd crush update on start = false
>>
>> If this is the case, and you move disks around hosts, they won't update
>> their position/host in the crushmap, even if the crushmap does not reflect
>> reality.
>>
>> Regards,
>>
>> Mart
>>
>>
>>
>>
>>
>> On 01/08/2016 02:16 AM, Wade Holler wrote:
>>
>> Sure.  Apologies for all the text: We have 12 Nodes for OSDs, 15 OSDs per
>> node,  but I will only include a sample:
>>
>> ceph osd tree | head -35
>>
>> ID  WEIGHT    TYPE NAME         UP/DOWN REWEIGHT PRIMARY-AFFINITY
>>
>>  -1 130.98450 root default
>>
>>  -2   5.82153     host cpn00001
>>
>>   4   0.72769         osd.4          up  1.00000          1.00000
>>
>>  14   0.72769         osd.14         up  1.00000          1.00000
>>
>>   3   0.72769         osd.3          up  1.00000          1.00000
>>
>>  24   0.72769         osd.24         up  1.00000          1.00000
>>
>>   5   0.72769         osd.5          up  1.00000          1.00000
>>
>>   2   0.72769         osd.2          up  1.00000          1.00000
>>
>>  17   0.72769         osd.17         up  1.00000          1.00000
>>
>>  69   0.72769         osd.69         up  1.00000          1.00000
>>
>>  -3   6.54922     host cpn00003
>>
>>   7   0.72769         osd.7          up  1.00000          1.00000
>>
>>   8   0.72769         osd.8          up  1.00000          1.00000
>>
>>   9   0.72769         osd.9          up  1.00000          1.00000
>>
>>   0   0.72769         osd.0          up  1.00000          1.00000
>>
>>  28   0.72769         osd.28         up  1.00000          1.00000
>>
>>  10   0.72769         osd.10         up  1.00000          1.00000
>>
>>   1   0.72769         osd.1          up  1.00000          1.00000
>>
>>   6   0.72769         osd.6          up  1.00000          1.00000
>>
>>  29   0.72769         osd.29         up  1.00000          1.00000
>>
>>  -4   2.91077     host cpn00004
>>
>>
>> Compared with the actual processes that are running:
>>
>>
>> [root@cpx00001 ~]# ssh cpn00001 ps -ef | grep ceph\-osd
>>
>> ceph       92638       1 26 16:19 ?        01:00:55 /usr/bin/ceph-osd -f
>> --cluster ceph --id 6 --setuser ceph --setgroup ceph
>>
>> ceph       92667       1 20 16:19 ?        00:48:04 /usr/bin/ceph-osd -f
>> --cluster ceph --id 0 --setuser ceph --setgroup ceph
>>
>> ceph       92673       1 18 16:19 ?        00:42:48 /usr/bin/ceph-osd -f
>> --cluster ceph --id 8 --setuser ceph --setgroup ceph
>>
>> ceph       92681       1 19 16:19 ?        00:45:52 /usr/bin/ceph-osd -f
>> --cluster ceph --id 7 --setuser ceph --setgroup ceph
>>
>> ceph       92701       1 15 16:19 ?        00:36:05 /usr/bin/ceph-osd -f
>> --cluster ceph --id 12 --setuser ceph --setgroup ceph
>>
>> ceph       92748       1 14 16:19 ?        00:34:07 /usr/bin/ceph-osd -f
>> --cluster ceph --id 10 --setuser ceph --setgroup ceph
>>
>> ceph       92756       1 16 16:19 ?        00:38:40 /usr/bin/ceph-osd -f
>> --cluster ceph --id 9 --setuser ceph --setgroup ceph
>>
>> ceph       92758       1 17 16:19 ?        00:39:28 /usr/bin/ceph-osd -f
>> --cluster ceph --id 13 --setuser ceph --setgroup ceph
>>
>> ceph       92777       1 19 16:19 ?        00:46:17 /usr/bin/ceph-osd -f
>> --cluster ceph --id 1 --setuser ceph --setgroup ceph
>>
>> ceph       92988       1 18 16:19 ?        00:42:47 /usr/bin/ceph-osd -f
>> --cluster ceph --id 5 --setuser ceph --setgroup ceph
>>
>> ceph       93058       1 18 16:19 ?        00:43:18 /usr/bin/ceph-osd -f
>> --cluster ceph --id 11 --setuser ceph --setgroup ceph
>>
>> ceph       93078       1 17 16:19 ?        00:41:38 /usr/bin/ceph-osd -f
>> --cluster ceph --id 14 --setuser ceph --setgroup ceph
>>
>> ceph       93127       1 15 16:19 ?        00:36:29 /usr/bin/ceph-osd -f
>> --cluster ceph --id 4 --setuser ceph --setgroup ceph
>>
>> ceph       93130       1 17 16:19 ?        00:40:44 /usr/bin/ceph-osd -f
>> --cluster ceph --id 2 --setuser ceph --setgroup ceph
>>
>> ceph       93173       1 21 16:19 ?        00:49:37 /usr/bin/ceph-osd -f
>> --cluster ceph --id 3 --setuser ceph --setgroup ceph
>>
>> [root@cpx00001 ~]# ssh cpn00003 ps -ef | grep ceph\-osd
>>
>> ceph       82454       1 18 16:19 ?        00:43:58 /usr/bin/ceph-osd -f
>> --cluster ceph --id 25 --setuser ceph --setgroup ceph
>>
>> ceph       82464       1 24 16:19 ?        00:55:40 /usr/bin/ceph-osd -f
>> --cluster ceph --id 21 --setuser ceph --setgroup ceph
>>
>> ceph       82473       1 21 16:19 ?        00:50:14 /usr/bin/ceph-osd -f
>> --cluster ceph --id 17 --setuser ceph --setgroup ceph
>>
>> ceph       82612       1 19 16:19 ?        00:45:25 /usr/bin/ceph-osd -f
>> --cluster ceph --id 22 --setuser ceph --setgroup ceph
>>
>> ceph       82629       1 20 16:19 ?        00:48:38 /usr/bin/ceph-osd -f
>> --cluster ceph --id 16 --setuser ceph --setgroup ceph
>>
>> ceph       82651       1 16 16:19 ?        00:39:24 /usr/bin/ceph-osd -f
>> --cluster ceph --id 20 --setuser ceph --setgroup ceph
>>
>> ceph       82687       1 17 16:19 ?        00:40:31 /usr/bin/ceph-osd -f
>> --cluster ceph --id 18 --setuser ceph --setgroup ceph
>>
>> ceph       82697       1 26 16:19 ?        01:02:12 /usr/bin/ceph-osd -f
>> --cluster ceph --id 23 --setuser ceph --setgroup ceph
>>
>> ceph       82719       1 20 16:19 ?        00:47:15 /usr/bin/ceph-osd -f
>> --cluster ceph --id 15 --setuser ceph --setgroup ceph
>>
>> ceph       82722       1 14 16:19 ?        00:33:41 /usr/bin/ceph-osd -f
>> --cluster ceph --id 28 --setuser ceph --setgroup ceph
>>
>> ceph       82725       1 14 16:19 ?        00:33:16 /usr/bin/ceph-osd -f
>> --cluster ceph --id 26 --setuser ceph --setgroup ceph
>>
>> ceph       82743       1 14 16:19 ?        00:34:17 /usr/bin/ceph-osd -f
>> --cluster ceph --id 29 --setuser ceph --setgroup ceph
>>
>> ceph       82769       1 19 16:19 ?        00:46:00 /usr/bin/ceph-osd -f
>> --cluster ceph --id 19 --setuser ceph --setgroup ceph
>>
>> ceph       82816       1 13 16:19 ?        00:30:26 /usr/bin/ceph-osd -f
>> --cluster ceph --id 27 --setuser ceph --setgroup ceph
>>
>> ceph       82828       1 27 16:19 ?        01:04:38 /usr/bin/ceph-osd -f
>> --cluster ceph --id 24 --setuser ceph --setgroup ceph
>>
>> [root@cpx00001 ~]#
>>
>>
>> Looks like the crushmap is bad also:
>>
>> (Cluster appears to be operating ok but this really concerns me.)
>>
>> # begin crush map
>>
>> tunable choose_local_tries 0
>>
>> tunable choose_local_fallback_tries 0
>>
>> tunable choose_total_tries 50
>>
>> tunable chooseleaf_descend_once 1
>>
>> tunable straw_calc_version 1
>>
>>
>> # devices
>>
>> device 0 osd.0
>>
>> device 1 osd.1
>>
>> device 2 osd.2
>>
>> device 3 osd.3
>>
>> device 4 osd.4
>>
>> device 5 osd.5
>>
>> device 6 osd.6
>>
>> device 7 osd.7
>>
>> device 8 osd.8
>>
>> device 9 osd.9
>>
>> device 10 osd.10
>>
>> device 11 osd.11
>>
>> device 12 osd.12
>>
>> device 13 osd.13
>>
>> device 14 osd.14
>>
>> device 15 osd.15
>>
>> device 16 osd.16
>>
>> device 17 osd.17
>>
>> device 18 osd.18
>>
>> device 19 osd.19
>>
>> device 20 osd.20
>>
>> device 21 osd.21
>>
>> device 22 osd.22
>>
>> device 23 osd.23
>>
>> device 24 osd.24
>>
>> device 25 osd.25
>>
>> device 26 osd.26
>>
>> device 27 osd.27
>>
>> device 28 osd.28
>>
>> device 29 osd.29
>>
>> device 30 osd.30
>>
>> device 31 osd.31
>>
>> device 32 osd.32
>>
>> device 33 osd.33
>>
>> device 34 osd.34
>>
>> device 35 osd.35
>>
>> device 36 osd.36
>>
>> device 37 osd.37
>>
>> ...
>>
>>
>> # types
>>
>> type 0 osd
>>
>> type 1 host
>>
>> type 2 chassis
>>
>> type 3 rack
>>
>> type 4 row
>>
>> type 5 pdu
>>
>> type 6 pod
>>
>> type 7 room
>>
>> type 8 datacenter
>>
>> type 9 region
>>
>> type 10 root
>>
>>
>> # buckets
>>
>> host cpn00001 {
>>
>>         id -2           # do not change unnecessarily
>>
>>         # weight 5.822
>>
>>         alg straw
>>
>>         hash 0  # rjenkins1
>>
>>         item osd.4 weight 0.728
>>
>>         item osd.14 weight 0.728
>>
>>         item osd.3 weight 0.728
>>
>>         item osd.24 weight 0.728
>>
>>         item osd.5 weight 0.728
>>
>>         item osd.2 weight 0.728
>>
>>         item osd.17 weight 0.728
>>
>>         item osd.69 weight 0.728
>>
>> }
>>
>> host cpn00003 {
>>
>>         id -3           # do not change unnecessarily
>>
>>         # weight 6.549
>>
>>         alg straw
>>
>>         hash 0  # rjenkins1
>>
>>         item osd.7 weight 0.728
>>
>>         item osd.8 weight 0.728
>>
>>         item osd.9 weight 0.728
>>
>>         item osd.0 weight 0.728
>>
>>         item osd.28 weight 0.728
>>
>>         item osd.10 weight 0.728
>>
>>         item osd.1 weight 0.728
>>
>>         item osd.6 weight 0.728
>>
>>         item osd.29 weight 0.728
>>
>> }
>>
>> host cpn00004 {....
>>
>>
>>
>> Thank you for your review !!!!!
>>
>> Wade
>>
>>
>>
>>
>>
>> On Thu, Jan 7, 2016 at 6:03 PM Shinobu Kinjo <ski...@redhat.com> wrote:
>>
>>> Can you share the output with us?
>>>
>>> Rgds,
>>> Shinobu
>>>
>>> ----- Original Message -----
>>> From: "Wade Holler" <wade.hol...@gmail.com>
>>> To: "ceph-users" <ceph-users@lists.ceph.com>
>>> Sent: Friday, January 8, 2016 7:29:07 AM
>>> Subject: [ceph-users] ceph osd tree output
>>>
>>> Sometimes my ceph osd tree output is wrong. Ie. Wrong osds on the wrong
>>> hosts ?
>>>
>>> Anyone else have this issue?
>>>
>>> I have seen this at Infernalis and Jewell.
>>>
>>> Thanks
>>> Wade
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>
>>
>> _______________________________________________
>> ceph-users mailing 
>> listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
>
>
> --
> 使用Opera的电子邮件客户端:http://www.opera.com/mail/
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to