On Sat, 13 Sep 2014, Alexandre DERUMIER wrote: > Hi, > as ceph user, It could be wonderfull to have it for Giant, > optracker performance impact is really huge (See my ssd benchmark on ceph > user mailing)
Definitely. More importantly, it resolves a few crashes we've observed. It's going through some testing right now, but once that's done it'll go into giant. sage > > Regards, > > Alexandre Derumier > > ----- Mail original ----- > > De: "Somnath Roy" <somnath....@sandisk.com> > ?: "Samuel Just" <sam.j...@inktank.com> > Cc: "Sage Weil" <sw...@redhat.com>, ceph-de...@vger.kernel.org, > ceph-users@lists.ceph.com > Envoy?: Samedi 13 Septembre 2014 10:03:52 > Objet: Re: [ceph-users] OpTracker optimization > > Sam/Sage, > I saw Giant is forked off today. We need the pull request > (https://github.com/ceph/ceph/pull/2440) to be in Giant. So, could you please > merge this into Giant when it will be ready ? > > Thanks & Regards > Somnath > > -----Original Message----- > From: Samuel Just [mailto:sam.j...@inktank.com] > Sent: Thursday, September 11, 2014 11:31 AM > To: Somnath Roy > Cc: Sage Weil; ceph-de...@vger.kernel.org; ceph-users@lists.ceph.com > Subject: Re: OpTracker optimization > > Just added it to wip-sam-testing. > -Sam > > On Thu, Sep 11, 2014 at 11:30 AM, Somnath Roy <somnath....@sandisk.com> > wrote: > > Sam/Sage, > > I have addressed all of your comments and pushed the changes to the same > > pull request. > > > > https://github.com/ceph/ceph/pull/2440 > > > > Thanks & Regards > > Somnath > > > > -----Original Message----- > > From: Sage Weil [mailto:sw...@redhat.com] > > Sent: Wednesday, September 10, 2014 8:33 PM > > To: Somnath Roy > > Cc: Samuel Just; ceph-de...@vger.kernel.org; ceph-users@lists.ceph.com > > Subject: RE: OpTracker optimization > > > > I had two substantiative comments on the first patch and then some trivial > > whitespace nits. Otherwise looks good! > > > > tahnks- > > sage > > > > On Thu, 11 Sep 2014, Somnath Roy wrote: > > > >> Sam/Sage, > >> I have incorporated all of your comments. Please have a look at the same > >> pull request. > >> > >> https://github.com/ceph/ceph/pull/2440 > >> > >> Thanks & Regards > >> Somnath > >> > >> -----Original Message----- > >> From: Samuel Just [mailto:sam.j...@inktank.com] > >> Sent: Wednesday, September 10, 2014 3:25 PM > >> To: Somnath Roy > >> Cc: Sage Weil (sw...@redhat.com); ceph-de...@vger.kernel.org; > >> ceph-users@lists.ceph.com > >> Subject: Re: OpTracker optimization > >> > >> Oh, I changed my mind, your approach is fine. I was unclear. > >> Currently, I just need you to address the other comments. > >> -Sam > >> > >> On Wed, Sep 10, 2014 at 3:13 PM, Somnath Roy <somnath....@sandisk.com> > >> wrote: > >> > As I understand, you want me to implement the following. > >> > > >> > 1. Keep this implementation one sharded optracker for the ios going > >> > through ms_dispatch path. > >> > > >> > 2. Additionally, for ios going through ms_fast_dispatch, you want > >> > me to implement optracker (without internal shard) per opwq shard > >> > > >> > Am I right ? > >> > > >> > Thanks & Regards > >> > Somnath > >> > > >> > -----Original Message----- > >> > From: Samuel Just [mailto:sam.j...@inktank.com] > >> > Sent: Wednesday, September 10, 2014 3:08 PM > >> > To: Somnath Roy > >> > Cc: Sage Weil (sw...@redhat.com); ceph-de...@vger.kernel.org; > >> > ceph-users@lists.ceph.com > >> > Subject: Re: OpTracker optimization > >> > > >> > I don't quite understand. > >> > -Sam > >> > > >> > On Wed, Sep 10, 2014 at 2:38 PM, Somnath Roy <somnath....@sandisk.com> > >> > wrote: > >> >> Thanks Sam. > >> >> So, you want me to go with optracker/shadedopWq , right ? > >> >> > >> >> Regards > >> >> Somnath > >> >> > >> >> -----Original Message----- > >> >> From: Samuel Just [mailto:sam.j...@inktank.com] > >> >> Sent: Wednesday, September 10, 2014 2:36 PM > >> >> To: Somnath Roy > >> >> Cc: Sage Weil (sw...@redhat.com); ceph-de...@vger.kernel.org; > >> >> ceph-users@lists.ceph.com > >> >> Subject: Re: OpTracker optimization > >> >> > >> >> Responded with cosmetic nonsense. Once you've got that and the other > >> >> comments addressed, I can put it in wip-sam-testing. > >> >> -Sam > >> >> > >> >> On Wed, Sep 10, 2014 at 1:30 PM, Somnath Roy <somnath....@sandisk.com> > >> >> wrote: > >> >>> Thanks Sam..I responded back :-) > >> >>> > >> >>> -----Original Message----- > >> >>> From: ceph-devel-ow...@vger.kernel.org > >> >>> [mailto:ceph-devel-ow...@vger.kernel.org] On Behalf Of Samuel > >> >>> Just > >> >>> Sent: Wednesday, September 10, 2014 11:17 AM > >> >>> To: Somnath Roy > >> >>> Cc: Sage Weil (sw...@redhat.com); ceph-de...@vger.kernel.org; > >> >>> ceph-users@lists.ceph.com > >> >>> Subject: Re: OpTracker optimization > >> >>> > >> >>> Added a comment about the approach. > >> >>> -Sam > >> >>> > >> >>> On Tue, Sep 9, 2014 at 1:33 PM, Somnath Roy <somnath....@sandisk.com> > >> >>> wrote: > >> >>>> Hi Sam/Sage, > >> >>>> > >> >>>> As we discussed earlier, enabling the present OpTracker code > >> >>>> degrading performance severely. For example, in my setup a > >> >>>> single OSD node with > >> >>>> 10 clients is reaching ~103K read iops with io served from > >> >>>> memory while optracking is disabled but enabling optracker it is > >> >>>> reduced to ~39K iops. > >> >>>> Probably, running OSD without enabling OpTracker is not an > >> >>>> option for many of Ceph users. > >> >>>> > >> >>>> Now, by sharding the Optracker:: ops_in_flight_lock (thus xlist > >> >>>> ops_in_flight) and removing some other bottlenecks I am able to > >> >>>> match the performance of OpTracking enabled OSD with OpTracking > >> >>>> disabled, but with the expense of ~1 extra cpu core. > >> >>>> > >> >>>> In this process I have also fixed the following tracker. > >> >>>> > >> >>>> > >> >>>> > >> >>>> http://tracker.ceph.com/issues/9384 > >> >>>> > >> >>>> > >> >>>> > >> >>>> and probably http://tracker.ceph.com/issues/8885 too. > >> >>>> > >> >>>> > >> >>>> > >> >>>> I have created following pull request for the same. Please review it. > >> >>>> > >> >>>> > >> >>>> > >> >>>> https://github.com/ceph/ceph/pull/2440 > >> >>>> > >> >>>> > >> >>>> > >> >>>> Thanks & Regards > >> >>>> > >> >>>> Somnath > >> >>>> > >> >>>> > >> >>>> > >> >>>> > >> >>>> ________________________________ > >> >>>> > >> >>>> PLEASE NOTE: The information contained in this electronic mail > >> >>>> message is intended only for the use of the designated > >> >>>> recipient(s) named above. If the reader of this message is not > >> >>>> the intended recipient, you are hereby notified that you have > >> >>>> received this message in error and that any review, > >> >>>> dissemination, distribution, or copying of this message is > >> >>>> strictly prohibited. If you have received this communication in > >> >>>> error, please notify the sender by telephone or e-mail (as shown > >> >>>> above) immediately and destroy any and all copies of this message in > >> >>>> your possession (whether hard copies or electronically stored > >> >>>> copies). > >> >>>> > >> >>> -- > >> >>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" > >> >>> in the body of a message to majord...@vger.kernel.org More > >> >>> majordomo info at http://vger.kernel.org/majordomo-info.html > >> >>> > >> >>> ________________________________ > >> >>> > >> >>> PLEASE NOTE: The information contained in this electronic mail message > >> >>> is intended only for the use of the designated recipient(s) named > >> >>> above. If the reader of this message is not the intended recipient, > >> >>> you are hereby notified that you have received this message in error > >> >>> and that any review, dissemination, distribution, or copying of this > >> >>> message is strictly prohibited. If you have received this > >> >>> communication in error, please notify the sender by telephone or > >> >>> e-mail (as shown above) immediately and destroy any and all copies of > >> >>> this message in your possession (whether hard copies or electronically > >> >>> stored copies). > >> >>> > >> > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majord...@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html > > _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com