Does anyone else have any suggestions here? I am increasingly concerned about my config if other folks aren't seeing this.
I could change to a manual crushmap but otherwise have no need to. I emailed the Ceph-dev list but have not had a response yet. Best Regards Wade On Fri, Jan 8, 2016 at 11:12 AM Wade Holler <wade.hol...@gmail.com> wrote: > It is not set in the conf file. So why do I still have this behavior ? > > On Fri, Jan 8, 2016 at 11:08 AM hnuzhoulin <hnuzhoul...@gmail.com> wrote: > >> Yeah,this setting can not see in asok config. >> You just set it in ceph.conf and restart mon and osd service(sorry I >> forget if these restart is necessary) >> >> what I use this config is when I changed crushmap manually,and I do not >> want the service init script to rebuild crushmap as default way. >> >> maybe this is not siut for your problem.just have a try. >> >> 在 Fri, 08 Jan 2016 21:51:32 +0800,Wade Holler <wade.hol...@gmail.com> 写道: >> >> That is not set as far as I can tell. Actually it is strange that I >> don't see that setting at all. >> >> [root@cpn00001 ~]# ceph daemon osd.0 config show | grep update | grep >> crush >> >> [root@cpn00001 ~]# grep update /etc/ceph/ceph.conf >> >> [root@cpn00001 ~]# >> >> On Fri, Jan 8, 2016 at 1:50 AM Mart van Santen <m...@greenhost.nl> wrote: >> >>> >>> >>> Hi, >>> >>> Do you have by any chance disabled automatic crushmap updates in your >>> ceph config? >>> >>> osd crush update on start = false >>> >>> If this is the case, and you move disks around hosts, they won't update >>> their position/host in the crushmap, even if the crushmap does not reflect >>> reality. >>> >>> Regards, >>> >>> Mart >>> >>> >>> >>> >>> >>> On 01/08/2016 02:16 AM, Wade Holler wrote: >>> >>> Sure. Apologies for all the text: We have 12 Nodes for OSDs, 15 OSDs >>> per node, but I will only include a sample: >>> >>> ceph osd tree | head -35 >>> >>> ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY >>> >>> -1 130.98450 root default >>> >>> -2 5.82153 host cpn00001 >>> >>> 4 0.72769 osd.4 up 1.00000 1.00000 >>> >>> 14 0.72769 osd.14 up 1.00000 1.00000 >>> >>> 3 0.72769 osd.3 up 1.00000 1.00000 >>> >>> 24 0.72769 osd.24 up 1.00000 1.00000 >>> >>> 5 0.72769 osd.5 up 1.00000 1.00000 >>> >>> 2 0.72769 osd.2 up 1.00000 1.00000 >>> >>> 17 0.72769 osd.17 up 1.00000 1.00000 >>> >>> 69 0.72769 osd.69 up 1.00000 1.00000 >>> >>> -3 6.54922 host cpn00003 >>> >>> 7 0.72769 osd.7 up 1.00000 1.00000 >>> >>> 8 0.72769 osd.8 up 1.00000 1.00000 >>> >>> 9 0.72769 osd.9 up 1.00000 1.00000 >>> >>> 0 0.72769 osd.0 up 1.00000 1.00000 >>> >>> 28 0.72769 osd.28 up 1.00000 1.00000 >>> >>> 10 0.72769 osd.10 up 1.00000 1.00000 >>> >>> 1 0.72769 osd.1 up 1.00000 1.00000 >>> >>> 6 0.72769 osd.6 up 1.00000 1.00000 >>> >>> 29 0.72769 osd.29 up 1.00000 1.00000 >>> >>> -4 2.91077 host cpn00004 >>> >>> >>> Compared with the actual processes that are running: >>> >>> >>> [root@cpx00001 ~]# ssh cpn00001 ps -ef | grep ceph\-osd >>> >>> ceph 92638 1 26 16:19 ? 01:00:55 /usr/bin/ceph-osd >>> -f --cluster ceph --id 6 --setuser ceph --setgroup ceph >>> >>> ceph 92667 1 20 16:19 ? 00:48:04 /usr/bin/ceph-osd >>> -f --cluster ceph --id 0 --setuser ceph --setgroup ceph >>> >>> ceph 92673 1 18 16:19 ? 00:42:48 /usr/bin/ceph-osd >>> -f --cluster ceph --id 8 --setuser ceph --setgroup ceph >>> >>> ceph 92681 1 19 16:19 ? 00:45:52 /usr/bin/ceph-osd >>> -f --cluster ceph --id 7 --setuser ceph --setgroup ceph >>> >>> ceph 92701 1 15 16:19 ? 00:36:05 /usr/bin/ceph-osd >>> -f --cluster ceph --id 12 --setuser ceph --setgroup ceph >>> >>> ceph 92748 1 14 16:19 ? 00:34:07 /usr/bin/ceph-osd >>> -f --cluster ceph --id 10 --setuser ceph --setgroup ceph >>> >>> ceph 92756 1 16 16:19 ? 00:38:40 /usr/bin/ceph-osd >>> -f --cluster ceph --id 9 --setuser ceph --setgroup ceph >>> >>> ceph 92758 1 17 16:19 ? 00:39:28 /usr/bin/ceph-osd >>> -f --cluster ceph --id 13 --setuser ceph --setgroup ceph >>> >>> ceph 92777 1 19 16:19 ? 00:46:17 /usr/bin/ceph-osd >>> -f --cluster ceph --id 1 --setuser ceph --setgroup ceph >>> >>> ceph 92988 1 18 16:19 ? 00:42:47 /usr/bin/ceph-osd >>> -f --cluster ceph --id 5 --setuser ceph --setgroup ceph >>> >>> ceph 93058 1 18 16:19 ? 00:43:18 /usr/bin/ceph-osd >>> -f --cluster ceph --id 11 --setuser ceph --setgroup ceph >>> >>> ceph 93078 1 17 16:19 ? 00:41:38 /usr/bin/ceph-osd >>> -f --cluster ceph --id 14 --setuser ceph --setgroup ceph >>> >>> ceph 93127 1 15 16:19 ? 00:36:29 /usr/bin/ceph-osd >>> -f --cluster ceph --id 4 --setuser ceph --setgroup ceph >>> >>> ceph 93130 1 17 16:19 ? 00:40:44 /usr/bin/ceph-osd >>> -f --cluster ceph --id 2 --setuser ceph --setgroup ceph >>> >>> ceph 93173 1 21 16:19 ? 00:49:37 /usr/bin/ceph-osd >>> -f --cluster ceph --id 3 --setuser ceph --setgroup ceph >>> >>> [root@cpx00001 ~]# ssh cpn00003 ps -ef | grep ceph\-osd >>> >>> ceph 82454 1 18 16:19 ? 00:43:58 /usr/bin/ceph-osd >>> -f --cluster ceph --id 25 --setuser ceph --setgroup ceph >>> >>> ceph 82464 1 24 16:19 ? 00:55:40 /usr/bin/ceph-osd >>> -f --cluster ceph --id 21 --setuser ceph --setgroup ceph >>> >>> ceph 82473 1 21 16:19 ? 00:50:14 /usr/bin/ceph-osd >>> -f --cluster ceph --id 17 --setuser ceph --setgroup ceph >>> >>> ceph 82612 1 19 16:19 ? 00:45:25 /usr/bin/ceph-osd >>> -f --cluster ceph --id 22 --setuser ceph --setgroup ceph >>> >>> ceph 82629 1 20 16:19 ? 00:48:38 /usr/bin/ceph-osd >>> -f --cluster ceph --id 16 --setuser ceph --setgroup ceph >>> >>> ceph 82651 1 16 16:19 ? 00:39:24 /usr/bin/ceph-osd >>> -f --cluster ceph --id 20 --setuser ceph --setgroup ceph >>> >>> ceph 82687 1 17 16:19 ? 00:40:31 /usr/bin/ceph-osd >>> -f --cluster ceph --id 18 --setuser ceph --setgroup ceph >>> >>> ceph 82697 1 26 16:19 ? 01:02:12 /usr/bin/ceph-osd >>> -f --cluster ceph --id 23 --setuser ceph --setgroup ceph >>> >>> ceph 82719 1 20 16:19 ? 00:47:15 /usr/bin/ceph-osd >>> -f --cluster ceph --id 15 --setuser ceph --setgroup ceph >>> >>> ceph 82722 1 14 16:19 ? 00:33:41 /usr/bin/ceph-osd >>> -f --cluster ceph --id 28 --setuser ceph --setgroup ceph >>> >>> ceph 82725 1 14 16:19 ? 00:33:16 /usr/bin/ceph-osd >>> -f --cluster ceph --id 26 --setuser ceph --setgroup ceph >>> >>> ceph 82743 1 14 16:19 ? 00:34:17 /usr/bin/ceph-osd >>> -f --cluster ceph --id 29 --setuser ceph --setgroup ceph >>> >>> ceph 82769 1 19 16:19 ? 00:46:00 /usr/bin/ceph-osd >>> -f --cluster ceph --id 19 --setuser ceph --setgroup ceph >>> >>> ceph 82816 1 13 16:19 ? 00:30:26 /usr/bin/ceph-osd >>> -f --cluster ceph --id 27 --setuser ceph --setgroup ceph >>> >>> ceph 82828 1 27 16:19 ? 01:04:38 /usr/bin/ceph-osd >>> -f --cluster ceph --id 24 --setuser ceph --setgroup ceph >>> >>> [root@cpx00001 ~]# >>> >>> >>> Looks like the crushmap is bad also: >>> >>> (Cluster appears to be operating ok but this really concerns me.) >>> >>> # begin crush map >>> >>> tunable choose_local_tries 0 >>> >>> tunable choose_local_fallback_tries 0 >>> >>> tunable choose_total_tries 50 >>> >>> tunable chooseleaf_descend_once 1 >>> >>> tunable straw_calc_version 1 >>> >>> >>> # devices >>> >>> device 0 osd.0 >>> >>> device 1 osd.1 >>> >>> device 2 osd.2 >>> >>> device 3 osd.3 >>> >>> device 4 osd.4 >>> >>> device 5 osd.5 >>> >>> device 6 osd.6 >>> >>> device 7 osd.7 >>> >>> device 8 osd.8 >>> >>> device 9 osd.9 >>> >>> device 10 osd.10 >>> >>> device 11 osd.11 >>> >>> device 12 osd.12 >>> >>> device 13 osd.13 >>> >>> device 14 osd.14 >>> >>> device 15 osd.15 >>> >>> device 16 osd.16 >>> >>> device 17 osd.17 >>> >>> device 18 osd.18 >>> >>> device 19 osd.19 >>> >>> device 20 osd.20 >>> >>> device 21 osd.21 >>> >>> device 22 osd.22 >>> >>> device 23 osd.23 >>> >>> device 24 osd.24 >>> >>> device 25 osd.25 >>> >>> device 26 osd.26 >>> >>> device 27 osd.27 >>> >>> device 28 osd.28 >>> >>> device 29 osd.29 >>> >>> device 30 osd.30 >>> >>> device 31 osd.31 >>> >>> device 32 osd.32 >>> >>> device 33 osd.33 >>> >>> device 34 osd.34 >>> >>> device 35 osd.35 >>> >>> device 36 osd.36 >>> >>> device 37 osd.37 >>> >>> ... >>> >>> >>> # types >>> >>> type 0 osd >>> >>> type 1 host >>> >>> type 2 chassis >>> >>> type 3 rack >>> >>> type 4 row >>> >>> type 5 pdu >>> >>> type 6 pod >>> >>> type 7 room >>> >>> type 8 datacenter >>> >>> type 9 region >>> >>> type 10 root >>> >>> >>> # buckets >>> >>> host cpn00001 { >>> >>> id -2 # do not change unnecessarily >>> >>> # weight 5.822 >>> >>> alg straw >>> >>> hash 0 # rjenkins1 >>> >>> item osd.4 weight 0.728 >>> >>> item osd.14 weight 0.728 >>> >>> item osd.3 weight 0.728 >>> >>> item osd.24 weight 0.728 >>> >>> item osd.5 weight 0.728 >>> >>> item osd.2 weight 0.728 >>> >>> item osd.17 weight 0.728 >>> >>> item osd.69 weight 0.728 >>> >>> } >>> >>> host cpn00003 { >>> >>> id -3 # do not change unnecessarily >>> >>> # weight 6.549 >>> >>> alg straw >>> >>> hash 0 # rjenkins1 >>> >>> item osd.7 weight 0.728 >>> >>> item osd.8 weight 0.728 >>> >>> item osd.9 weight 0.728 >>> >>> item osd.0 weight 0.728 >>> >>> item osd.28 weight 0.728 >>> >>> item osd.10 weight 0.728 >>> >>> item osd.1 weight 0.728 >>> >>> item osd.6 weight 0.728 >>> >>> item osd.29 weight 0.728 >>> >>> } >>> >>> host cpn00004 {.... >>> >>> >>> >>> Thank you for your review !!!!! >>> >>> Wade >>> >>> >>> >>> >>> >>> On Thu, Jan 7, 2016 at 6:03 PM Shinobu Kinjo <ski...@redhat.com> wrote: >>> >>>> Can you share the output with us? >>>> >>>> Rgds, >>>> Shinobu >>>> >>>> ----- Original Message ----- >>>> From: "Wade Holler" <wade.hol...@gmail.com> >>>> To: "ceph-users" <ceph-users@lists.ceph.com> >>>> Sent: Friday, January 8, 2016 7:29:07 AM >>>> Subject: [ceph-users] ceph osd tree output >>>> >>>> Sometimes my ceph osd tree output is wrong. Ie. Wrong osds on the wrong >>>> hosts ? >>>> >>>> Anyone else have this issue? >>>> >>>> I have seen this at Infernalis and Jewell. >>>> >>>> Thanks >>>> Wade >>>> >>>> _______________________________________________ >>>> ceph-users mailing list >>>> ceph-users@lists.ceph.com >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>> >>> >>> >>> _______________________________________________ >>> ceph-users mailing >>> listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >>> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >> >> >> >> -- >> 使用Opera的电子邮件客户端:http://www.opera.com/mail/ >> >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com