On Tue, Jan 10, 2017 at 12:59 PM, Samuel Just <sj...@redhat.com> wrote: > Mm, maybe the tag didn't get pushed. Alfredo, is there supposed to be > a v11.1.1 tag?
Yep. You can see there is one here: https://github.com/ceph/ceph/releases Specifically: https://github.com/ceph/ceph/releases/tag/v11.1.1 which points to https://github.com/ceph/ceph/commit/87597971b371d7f497d7eabad3545d72d18dd755 > -Sam > > On Tue, Jan 10, 2017 at 9:57 AM, Stillwell, Bryan J > <bryan.stillw...@charter.com> wrote: >> That's strange, I installed that version using packages from here: >> >> http://download.ceph.com/debian-kraken/pool/main/c/ceph/ >> >> >> Bryan >> >> On 1/10/17, 10:51 AM, "Samuel Just" <sj...@redhat.com> wrote: >> >>>Can you push that branch somewhere? I don't have a v11.1.1 or that sha1. >>>-Sam >>> >>>On Tue, Jan 10, 2017 at 9:41 AM, Stillwell, Bryan J >>><bryan.stillw...@charter.com> wrote: >>>> This is from: >>>> >>>> ceph version 11.1.1 (87597971b371d7f497d7eabad3545d72d18dd755) >>>> >>>> On 1/10/17, 10:23 AM, "Samuel Just" <sj...@redhat.com> wrote: >>>> >>>>>What ceph sha1 is that? Does it include >>>>>6c3d015c6854a12cda40673848813d968ff6afae which fixed the messenger >>>>>spin? >>>>>-Sam >>>>> >>>>>On Tue, Jan 10, 2017 at 9:00 AM, Stillwell, Bryan J >>>>><bryan.stillw...@charter.com> wrote: >>>>>> On 1/10/17, 5:35 AM, "John Spray" <jsp...@redhat.com> wrote: >>>>>> >>>>>>>On Mon, Jan 9, 2017 at 11:46 PM, Stillwell, Bryan J >>>>>>><bryan.stillw...@charter.com> wrote: >>>>>>>> Last week I decided to play around with Kraken (11.1.1-1xenial) on a >>>>>>>> single node, two OSD cluster, and after a while I noticed that the >>>>>>>>new >>>>>>>> ceph-mgr daemon is frequently using a lot of the CPU: >>>>>>>> >>>>>>>> 17519 ceph 20 0 850044 168104 208 S 102.7 4.3 1278:27 >>>>>>>> ceph-mgr >>>>>>>> >>>>>>>> Restarting it with 'systemctl restart ceph-mgr*' seems to get its >>>>>>>>CPU >>>>>>>> usage down to < 1%, but after a while it climbs back up to > 100%. >>>>>>>>Has >>>>>>>> anyone else seen this? >>>>>>> >>>>>>>Definitely worth investigating, could you set "debug mgr = 20" on the >>>>>>>daemon to see if it's obviously spinning in a particular place? >>>>>> >>>>>> I've injected that option to the ceps-mgr process, and now I'm just >>>>>> waiting for it to go out of control again. >>>>>> >>>>>> However, I've noticed quite a few messages like this in the logs >>>>>>already: >>>>>> >>>>>> 2017-01-10 09:56:07.441678 7f70f4562700 0 -- 172.24.88.207:6800/4104 >>>>>>>> >>>>>> 172.24.88.207:0/4168225878 conn(0x563c7e0bc000 :6800 s=STATE_OPEN >>>>>>pgs=2 >>>>>> cs=1 l=0).fault initiating reconnect >>>>>> 2017-01-10 09:56:07.442044 7f70f4562700 0 -- 172.24.88.207:6800/4104 >>>>>>>> >>>>>> 172.24.88.207:0/4168225878 conn(0x563c7dfea800 :6800 >>>>>> s=STATE_ACCEPTING_WAIT_CONNECT_MSG_AUTH pgs=0 cs=0 >>>>>>l=0).handle_connect_msg >>>>>> accept connect_seq 0 vs existing csq=2 existing_state=STATE_CONNECTING >>>>>> 2017-01-10 09:56:07.442067 7f70f4562700 0 -- 172.24.88.207:6800/4104 >>>>>>>> >>>>>> 172.24.88.207:0/4168225878 conn(0x563c7dfea800 :6800 >>>>>> s=STATE_ACCEPTING_WAIT_CONNECT_MSG_AUTH pgs=0 cs=0 >>>>>>l=0).handle_connect_msg >>>>>> accept peer reset, then tried to connect to us, replacing >>>>>> 2017-01-10 09:56:07.443026 7f70f4562700 0 -- 172.24.88.207:6800/4104 >>>>>>>> >>>>>> 172.24.88.207:0/4168225878 conn(0x563c7e0bc000 :6800 >>>>>> s=STATE_ACCEPTING_WAIT_CONNECT_MSG pgs=2 cs=0 l=0).fault with nothing >>>>>>to >>>>>> send and in the half accept state just closed >>>>>> >>>>>> >>>>>> What's weird about that is that this is a single node cluster with >>>>>> ceph-mgr, ceph-mon, and the ceph-osd processes all running on the same >>>>>> host. So none of the communication should be leaving the node. >>>>>> >>>>>> Bryan >>>>>> >>>>>> E-MAIL CONFIDENTIALITY NOTICE: >>>>>> The contents of this e-mail message and any attachments are intended >>>>>>solely for the addressee(s) and may contain confidential and/or legally >>>>>>privileged information. If you are not the intended recipient of this >>>>>>message or if this message has been addressed to you in error, please >>>>>>immediately alert the sender by reply e-mail and then delete this >>>>>>message and any attachments. If you are not the intended recipient, you >>>>>>are notified that any use, dissemination, distribution, copying, or >>>>>>storage of this message or any attachment is strictly prohibited. >>>>>> >>>>>> _______________________________________________ >>>>>> ceph-users mailing list >>>>>> ceph-users@lists.ceph.com >>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>> >>>> E-MAIL CONFIDENTIALITY NOTICE: >>>> The contents of this e-mail message and any attachments are intended >>>>solely for the addressee(s) and may contain confidential and/or legally >>>>privileged information. If you are not the intended recipient of this >>>>message or if this message has been addressed to you in error, please >>>>immediately alert the sender by reply e-mail and then delete this >>>>message and any attachments. If you are not the intended recipient, you >>>>are notified that any use, dissemination, distribution, copying, or >>>>storage of this message or any attachment is strictly prohibited. >>>> >> >> E-MAIL CONFIDENTIALITY NOTICE: >> The contents of this e-mail message and any attachments are intended solely >> for the addressee(s) and may contain confidential and/or legally privileged >> information. If you are not the intended recipient of this message or if >> this message has been addressed to you in error, please immediately alert >> the sender by reply e-mail and then delete this message and any attachments. >> If you are not the intended recipient, you are notified that any use, >> dissemination, distribution, copying, or storage of this message or any >> attachment is strictly prohibited. >> _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com