Just added it to wip-sam-testing. -Sam
On Thu, Sep 11, 2014 at 11:30 AM, Somnath Roy <somnath....@sandisk.com> wrote: > Sam/Sage, > I have addressed all of your comments and pushed the changes to the same pull > request. > > https://github.com/ceph/ceph/pull/2440 > > Thanks & Regards > Somnath > > -----Original Message----- > From: Sage Weil [mailto:sw...@redhat.com] > Sent: Wednesday, September 10, 2014 8:33 PM > To: Somnath Roy > Cc: Samuel Just; ceph-de...@vger.kernel.org; ceph-users@lists.ceph.com > Subject: RE: OpTracker optimization > > I had two substantiative comments on the first patch and then some trivial > whitespace nits. Otherwise looks good! > > tahnks- > sage > > On Thu, 11 Sep 2014, Somnath Roy wrote: > >> Sam/Sage, >> I have incorporated all of your comments. Please have a look at the same >> pull request. >> >> https://github.com/ceph/ceph/pull/2440 >> >> Thanks & Regards >> Somnath >> >> -----Original Message----- >> From: Samuel Just [mailto:sam.j...@inktank.com] >> Sent: Wednesday, September 10, 2014 3:25 PM >> To: Somnath Roy >> Cc: Sage Weil (sw...@redhat.com); ceph-de...@vger.kernel.org; >> ceph-users@lists.ceph.com >> Subject: Re: OpTracker optimization >> >> Oh, I changed my mind, your approach is fine. I was unclear. >> Currently, I just need you to address the other comments. >> -Sam >> >> On Wed, Sep 10, 2014 at 3:13 PM, Somnath Roy <somnath....@sandisk.com> wrote: >> > As I understand, you want me to implement the following. >> > >> > 1. Keep this implementation one sharded optracker for the ios going >> > through ms_dispatch path. >> > >> > 2. Additionally, for ios going through ms_fast_dispatch, you want me >> > to implement optracker (without internal shard) per opwq shard >> > >> > Am I right ? >> > >> > Thanks & Regards >> > Somnath >> > >> > -----Original Message----- >> > From: Samuel Just [mailto:sam.j...@inktank.com] >> > Sent: Wednesday, September 10, 2014 3:08 PM >> > To: Somnath Roy >> > Cc: Sage Weil (sw...@redhat.com); ceph-de...@vger.kernel.org; >> > ceph-users@lists.ceph.com >> > Subject: Re: OpTracker optimization >> > >> > I don't quite understand. >> > -Sam >> > >> > On Wed, Sep 10, 2014 at 2:38 PM, Somnath Roy <somnath....@sandisk.com> >> > wrote: >> >> Thanks Sam. >> >> So, you want me to go with optracker/shadedopWq , right ? >> >> >> >> Regards >> >> Somnath >> >> >> >> -----Original Message----- >> >> From: Samuel Just [mailto:sam.j...@inktank.com] >> >> Sent: Wednesday, September 10, 2014 2:36 PM >> >> To: Somnath Roy >> >> Cc: Sage Weil (sw...@redhat.com); ceph-de...@vger.kernel.org; >> >> ceph-users@lists.ceph.com >> >> Subject: Re: OpTracker optimization >> >> >> >> Responded with cosmetic nonsense. Once you've got that and the other >> >> comments addressed, I can put it in wip-sam-testing. >> >> -Sam >> >> >> >> On Wed, Sep 10, 2014 at 1:30 PM, Somnath Roy <somnath....@sandisk.com> >> >> wrote: >> >>> Thanks Sam..I responded back :-) >> >>> >> >>> -----Original Message----- >> >>> From: ceph-devel-ow...@vger.kernel.org >> >>> [mailto:ceph-devel-ow...@vger.kernel.org] On Behalf Of Samuel Just >> >>> Sent: Wednesday, September 10, 2014 11:17 AM >> >>> To: Somnath Roy >> >>> Cc: Sage Weil (sw...@redhat.com); ceph-de...@vger.kernel.org; >> >>> ceph-users@lists.ceph.com >> >>> Subject: Re: OpTracker optimization >> >>> >> >>> Added a comment about the approach. >> >>> -Sam >> >>> >> >>> On Tue, Sep 9, 2014 at 1:33 PM, Somnath Roy <somnath....@sandisk.com> >> >>> wrote: >> >>>> Hi Sam/Sage, >> >>>> >> >>>> As we discussed earlier, enabling the present OpTracker code >> >>>> degrading performance severely. For example, in my setup a single >> >>>> OSD node with >> >>>> 10 clients is reaching ~103K read iops with io served from memory >> >>>> while optracking is disabled but enabling optracker it is reduced to >> >>>> ~39K iops. >> >>>> Probably, running OSD without enabling OpTracker is not an option >> >>>> for many of Ceph users. >> >>>> >> >>>> Now, by sharding the Optracker:: ops_in_flight_lock (thus xlist >> >>>> ops_in_flight) and removing some other bottlenecks I am able to >> >>>> match the performance of OpTracking enabled OSD with OpTracking >> >>>> disabled, but with the expense of ~1 extra cpu core. >> >>>> >> >>>> In this process I have also fixed the following tracker. >> >>>> >> >>>> >> >>>> >> >>>> http://tracker.ceph.com/issues/9384 >> >>>> >> >>>> >> >>>> >> >>>> and probably http://tracker.ceph.com/issues/8885 too. >> >>>> >> >>>> >> >>>> >> >>>> I have created following pull request for the same. Please review it. >> >>>> >> >>>> >> >>>> >> >>>> https://github.com/ceph/ceph/pull/2440 >> >>>> >> >>>> >> >>>> >> >>>> Thanks & Regards >> >>>> >> >>>> Somnath >> >>>> >> >>>> >> >>>> >> >>>> >> >>>> ________________________________ >> >>>> >> >>>> PLEASE NOTE: The information contained in this electronic mail >> >>>> message is intended only for the use of the designated >> >>>> recipient(s) named above. If the reader of this message is not >> >>>> the intended recipient, you are hereby notified that you have >> >>>> received this message in error and that any review, >> >>>> dissemination, distribution, or copying of this message is >> >>>> strictly prohibited. If you have received this communication in >> >>>> error, please notify the sender by telephone or e-mail (as shown >> >>>> above) immediately and destroy any and all copies of this message in >> >>>> your possession (whether hard copies or electronically stored copies). >> >>>> >> >>> -- >> >>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" >> >>> in the body of a message to majord...@vger.kernel.org More >> >>> majordomo info at http://vger.kernel.org/majordomo-info.html >> >>> >> >>> ________________________________ >> >>> >> >>> PLEASE NOTE: The information contained in this electronic mail message >> >>> is intended only for the use of the designated recipient(s) named above. >> >>> If the reader of this message is not the intended recipient, you are >> >>> hereby notified that you have received this message in error and that >> >>> any review, dissemination, distribution, or copying of this message is >> >>> strictly prohibited. If you have received this communication in error, >> >>> please notify the sender by telephone or e-mail (as shown above) >> >>> immediately and destroy any and all copies of this message in your >> >>> possession (whether hard copies or electronically stored copies). >> >>> >> _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com