Deployment method: ceph-deploy
Centos 7.2, systemctl
Infernalis.  This also happened when I was testing @ Jewell.

I am restarting ( or stop/start ) the ceph old processes ( after they die
or something ), with:

systemctl stop|start|restart ceph.target

Is there another way that it is more appropriate ?


Thank you ahead of time for your help!


Best Regards,
Wade

On Mon, Jan 11, 2016 at 5:43 PM John Spray <jsp...@redhat.com> wrote:

> On Mon, Jan 11, 2016 at 10:32 PM, Wade Holler <wade.hol...@gmail.com>
> wrote:
> > Does anyone else have any suggestions here? I am increasingly concerned
> > about my config if other folks aren't seeing this.
> >
> > I could change to a manual crushmap but otherwise have no need to.
>
> What did you use to deploy ceph?  What init system are you using to
> start the ceph daemons?  Also, usual questions, what distro, what ceph
> version?
>
> The script that updates the OSD host locations in the crush map is
> meant to be invoked by the init system before the ceph-osd process,
> but if you're starting it in some unconventional way then that might
> not be happening.
>
> John
>
> >
> > I emailed the Ceph-dev list but have not had a response yet.
> >
> > Best Regards
> > Wade
> >
> > On Fri, Jan 8, 2016 at 11:12 AM Wade Holler <wade.hol...@gmail.com>
> wrote:
> >>
> >> It is not set in the conf file.  So why do I still have this behavior ?
> >>
> >> On Fri, Jan 8, 2016 at 11:08 AM hnuzhoulin <hnuzhoul...@gmail.com>
> wrote:
> >>>
> >>> Yeah,this setting can not see in asok config.
> >>> You just set it in ceph.conf and restart mon and osd service(sorry I
> >>> forget if these restart is necessary)
> >>>
> >>> what I use this config is when I changed crushmap manually,and I do not
> >>> want the service init script to rebuild crushmap as default way.
> >>>
> >>> maybe this is not siut for your problem.just have a try.
> >>>
> >>> 在 Fri, 08 Jan 2016 21:51:32 +0800,Wade Holler <wade.hol...@gmail.com>
> 写道:
> >>>
> >>> That is not set as far as I can tell.  Actually it is strange that I
> >>> don't see that setting at all.
> >>>
> >>> [root@cpn00001 ~]# ceph daemon osd.0 config show | grep update | grep
> >>> crush
> >>>
> >>> [root@cpn00001 ~]# grep update /etc/ceph/ceph.conf
> >>>
> >>> [root@cpn00001 ~]#
> >>>
> >>>
> >>> On Fri, Jan 8, 2016 at 1:50 AM Mart van Santen <m...@greenhost.nl>
> wrote:
> >>>>
> >>>>
> >>>>
> >>>> Hi,
> >>>>
> >>>> Do you have by any chance disabled automatic crushmap updates in your
> >>>> ceph config?
> >>>>
> >>>> osd crush update on start = false
> >>>>
> >>>> If this is the case, and you move disks around hosts, they won't
> update
> >>>> their position/host in the crushmap, even if the crushmap does not
> reflect
> >>>> reality.
> >>>>
> >>>> Regards,
> >>>>
> >>>> Mart
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>> On 01/08/2016 02:16 AM, Wade Holler wrote:
> >>>>
> >>>> Sure.  Apologies for all the text: We have 12 Nodes for OSDs, 15 OSDs
> >>>> per node,  but I will only include a sample:
> >>>>
> >>>> ceph osd tree | head -35
> >>>>
> >>>> ID  WEIGHT    TYPE NAME         UP/DOWN REWEIGHT PRIMARY-AFFINITY
> >>>>
> >>>>  -1 130.98450 root default
> >>>>
> >>>>  -2   5.82153     host cpn00001
> >>>>
> >>>>   4   0.72769         osd.4          up  1.00000          1.00000
> >>>>
> >>>>  14   0.72769         osd.14         up  1.00000          1.00000
> >>>>
> >>>>   3   0.72769         osd.3          up  1.00000          1.00000
> >>>>
> >>>>  24   0.72769         osd.24         up  1.00000          1.00000
> >>>>
> >>>>   5   0.72769         osd.5          up  1.00000          1.00000
> >>>>
> >>>>   2   0.72769         osd.2          up  1.00000          1.00000
> >>>>
> >>>>  17   0.72769         osd.17         up  1.00000          1.00000
> >>>>
> >>>>  69   0.72769         osd.69         up  1.00000          1.00000
> >>>>
> >>>>  -3   6.54922     host cpn00003
> >>>>
> >>>>   7   0.72769         osd.7          up  1.00000          1.00000
> >>>>
> >>>>   8   0.72769         osd.8          up  1.00000          1.00000
> >>>>
> >>>>   9   0.72769         osd.9          up  1.00000          1.00000
> >>>>
> >>>>   0   0.72769         osd.0          up  1.00000          1.00000
> >>>>
> >>>>  28   0.72769         osd.28         up  1.00000          1.00000
> >>>>
> >>>>  10   0.72769         osd.10         up  1.00000          1.00000
> >>>>
> >>>>   1   0.72769         osd.1          up  1.00000          1.00000
> >>>>
> >>>>   6   0.72769         osd.6          up  1.00000          1.00000
> >>>>
> >>>>  29   0.72769         osd.29         up  1.00000          1.00000
> >>>>
> >>>>  -4   2.91077     host cpn00004
> >>>>
> >>>>
> >>>> Compared with the actual processes that are running:
> >>>>
> >>>>
> >>>> [root@cpx00001 ~]# ssh cpn00001 ps -ef | grep ceph\-osd
> >>>>
> >>>> ceph       92638       1 26 16:19 ?        01:00:55 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 6 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       92667       1 20 16:19 ?        00:48:04 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 0 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       92673       1 18 16:19 ?        00:42:48 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 8 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       92681       1 19 16:19 ?        00:45:52 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 7 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       92701       1 15 16:19 ?        00:36:05 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 12 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       92748       1 14 16:19 ?        00:34:07 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 10 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       92756       1 16 16:19 ?        00:38:40 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 9 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       92758       1 17 16:19 ?        00:39:28 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 13 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       92777       1 19 16:19 ?        00:46:17 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 1 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       92988       1 18 16:19 ?        00:42:47 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 5 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       93058       1 18 16:19 ?        00:43:18 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 11 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       93078       1 17 16:19 ?        00:41:38 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 14 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       93127       1 15 16:19 ?        00:36:29 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 4 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       93130       1 17 16:19 ?        00:40:44 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 2 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       93173       1 21 16:19 ?        00:49:37 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 3 --setuser ceph --setgroup ceph
> >>>>
> >>>> [root@cpx00001 ~]# ssh cpn00003 ps -ef | grep ceph\-osd
> >>>>
> >>>> ceph       82454       1 18 16:19 ?        00:43:58 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 25 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       82464       1 24 16:19 ?        00:55:40 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 21 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       82473       1 21 16:19 ?        00:50:14 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 17 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       82612       1 19 16:19 ?        00:45:25 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 22 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       82629       1 20 16:19 ?        00:48:38 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 16 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       82651       1 16 16:19 ?        00:39:24 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 20 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       82687       1 17 16:19 ?        00:40:31 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 18 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       82697       1 26 16:19 ?        01:02:12 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 23 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       82719       1 20 16:19 ?        00:47:15 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 15 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       82722       1 14 16:19 ?        00:33:41 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 28 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       82725       1 14 16:19 ?        00:33:16 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 26 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       82743       1 14 16:19 ?        00:34:17 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 29 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       82769       1 19 16:19 ?        00:46:00 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 19 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       82816       1 13 16:19 ?        00:30:26 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 27 --setuser ceph --setgroup ceph
> >>>>
> >>>> ceph       82828       1 27 16:19 ?        01:04:38 /usr/bin/ceph-osd
> -f
> >>>> --cluster ceph --id 24 --setuser ceph --setgroup ceph
> >>>>
> >>>> [root@cpx00001 ~]#
> >>>>
> >>>>
> >>>> Looks like the crushmap is bad also:
> >>>>
> >>>> (Cluster appears to be operating ok but this really concerns me.)
> >>>>
> >>>> # begin crush map
> >>>>
> >>>> tunable choose_local_tries 0
> >>>>
> >>>> tunable choose_local_fallback_tries 0
> >>>>
> >>>> tunable choose_total_tries 50
> >>>>
> >>>> tunable chooseleaf_descend_once 1
> >>>>
> >>>> tunable straw_calc_version 1
> >>>>
> >>>>
> >>>> # devices
> >>>>
> >>>> device 0 osd.0
> >>>>
> >>>> device 1 osd.1
> >>>>
> >>>> device 2 osd.2
> >>>>
> >>>> device 3 osd.3
> >>>>
> >>>> device 4 osd.4
> >>>>
> >>>> device 5 osd.5
> >>>>
> >>>> device 6 osd.6
> >>>>
> >>>> device 7 osd.7
> >>>>
> >>>> device 8 osd.8
> >>>>
> >>>> device 9 osd.9
> >>>>
> >>>> device 10 osd.10
> >>>>
> >>>> device 11 osd.11
> >>>>
> >>>> device 12 osd.12
> >>>>
> >>>> device 13 osd.13
> >>>>
> >>>> device 14 osd.14
> >>>>
> >>>> device 15 osd.15
> >>>>
> >>>> device 16 osd.16
> >>>>
> >>>> device 17 osd.17
> >>>>
> >>>> device 18 osd.18
> >>>>
> >>>> device 19 osd.19
> >>>>
> >>>> device 20 osd.20
> >>>>
> >>>> device 21 osd.21
> >>>>
> >>>> device 22 osd.22
> >>>>
> >>>> device 23 osd.23
> >>>>
> >>>> device 24 osd.24
> >>>>
> >>>> device 25 osd.25
> >>>>
> >>>> device 26 osd.26
> >>>>
> >>>> device 27 osd.27
> >>>>
> >>>> device 28 osd.28
> >>>>
> >>>> device 29 osd.29
> >>>>
> >>>> device 30 osd.30
> >>>>
> >>>> device 31 osd.31
> >>>>
> >>>> device 32 osd.32
> >>>>
> >>>> device 33 osd.33
> >>>>
> >>>> device 34 osd.34
> >>>>
> >>>> device 35 osd.35
> >>>>
> >>>> device 36 osd.36
> >>>>
> >>>> device 37 osd.37
> >>>>
> >>>> ...
> >>>>
> >>>>
> >>>> # types
> >>>>
> >>>> type 0 osd
> >>>>
> >>>> type 1 host
> >>>>
> >>>> type 2 chassis
> >>>>
> >>>> type 3 rack
> >>>>
> >>>> type 4 row
> >>>>
> >>>> type 5 pdu
> >>>>
> >>>> type 6 pod
> >>>>
> >>>> type 7 room
> >>>>
> >>>> type 8 datacenter
> >>>>
> >>>> type 9 region
> >>>>
> >>>> type 10 root
> >>>>
> >>>>
> >>>> # buckets
> >>>>
> >>>> host cpn00001 {
> >>>>
> >>>>         id -2           # do not change unnecessarily
> >>>>
> >>>>         # weight 5.822
> >>>>
> >>>>         alg straw
> >>>>
> >>>>         hash 0  # rjenkins1
> >>>>
> >>>>         item osd.4 weight 0.728
> >>>>
> >>>>         item osd.14 weight 0.728
> >>>>
> >>>>         item osd.3 weight 0.728
> >>>>
> >>>>         item osd.24 weight 0.728
> >>>>
> >>>>         item osd.5 weight 0.728
> >>>>
> >>>>         item osd.2 weight 0.728
> >>>>
> >>>>         item osd.17 weight 0.728
> >>>>
> >>>>         item osd.69 weight 0.728
> >>>>
> >>>> }
> >>>>
> >>>> host cpn00003 {
> >>>>
> >>>>         id -3           # do not change unnecessarily
> >>>>
> >>>>         # weight 6.549
> >>>>
> >>>>         alg straw
> >>>>
> >>>>         hash 0  # rjenkins1
> >>>>
> >>>>         item osd.7 weight 0.728
> >>>>
> >>>>         item osd.8 weight 0.728
> >>>>
> >>>>         item osd.9 weight 0.728
> >>>>
> >>>>         item osd.0 weight 0.728
> >>>>
> >>>>         item osd.28 weight 0.728
> >>>>
> >>>>         item osd.10 weight 0.728
> >>>>
> >>>>         item osd.1 weight 0.728
> >>>>
> >>>>         item osd.6 weight 0.728
> >>>>
> >>>>         item osd.29 weight 0.728
> >>>>
> >>>> }
> >>>>
> >>>> host cpn00004 {....
> >>>>
> >>>>
> >>>>
> >>>> Thank you for your review !!!!!
> >>>>
> >>>> Wade
> >>>>
> >>>>
> >>>>
> >>>>
> >>>>
> >>>> On Thu, Jan 7, 2016 at 6:03 PM Shinobu Kinjo <ski...@redhat.com>
> wrote:
> >>>>>
> >>>>> Can you share the output with us?
> >>>>>
> >>>>> Rgds,
> >>>>> Shinobu
> >>>>>
> >>>>> ----- Original Message -----
> >>>>> From: "Wade Holler" <wade.hol...@gmail.com>
> >>>>> To: "ceph-users" <ceph-users@lists.ceph.com>
> >>>>> Sent: Friday, January 8, 2016 7:29:07 AM
> >>>>> Subject: [ceph-users] ceph osd tree output
> >>>>>
> >>>>> Sometimes my ceph osd tree output is wrong. Ie. Wrong osds on the
> wrong
> >>>>> hosts ?
> >>>>>
> >>>>> Anyone else have this issue?
> >>>>>
> >>>>> I have seen this at Infernalis and Jewell.
> >>>>>
> >>>>> Thanks
> >>>>> Wade
> >>>>>
> >>>>> _______________________________________________
> >>>>> ceph-users mailing list
> >>>>> ceph-users@lists.ceph.com
> >>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >>>>
> >>>>
> >>>>
> >>>> _______________________________________________
> >>>> ceph-users mailing list
> >>>> ceph-users@lists.ceph.com
> >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >>>>
> >>>>
> >>>> _______________________________________________
> >>>> ceph-users mailing list
> >>>> ceph-users@lists.ceph.com
> >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >>>
> >>>
> >>>
> >>>
> >>> --
> >>> 使用Opera的电子邮件客户端:http://www.opera.com/mail/
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to