Sam/Sage,
I have addressed all of your comments and pushed the changes to the same pull 
request.

https://github.com/ceph/ceph/pull/2440

Thanks & Regards
Somnath

-----Original Message-----
From: Sage Weil [mailto:sw...@redhat.com] 
Sent: Wednesday, September 10, 2014 8:33 PM
To: Somnath Roy
Cc: Samuel Just; ceph-de...@vger.kernel.org; ceph-users@lists.ceph.com
Subject: RE: OpTracker optimization

I had two substantiative comments on the first patch and then some trivial 
whitespace nits.    Otherwise looks good!

tahnks-
sage

On Thu, 11 Sep 2014, Somnath Roy wrote:

> Sam/Sage,
> I have incorporated all of your comments. Please have a look at the same pull 
> request.
> 
> https://github.com/ceph/ceph/pull/2440
> 
> Thanks & Regards
> Somnath
> 
> -----Original Message-----
> From: Samuel Just [mailto:sam.j...@inktank.com]
> Sent: Wednesday, September 10, 2014 3:25 PM
> To: Somnath Roy
> Cc: Sage Weil (sw...@redhat.com); ceph-de...@vger.kernel.org; 
> ceph-users@lists.ceph.com
> Subject: Re: OpTracker optimization
> 
> Oh, I changed my mind, your approach is fine.  I was unclear.
> Currently, I just need you to address the other comments.
> -Sam
> 
> On Wed, Sep 10, 2014 at 3:13 PM, Somnath Roy <somnath....@sandisk.com> wrote:
> > As I understand, you want me to implement the following.
> >
> > 1.  Keep this implementation one sharded optracker for the ios going 
> > through ms_dispatch path.
> >
> > 2. Additionally, for ios going through ms_fast_dispatch, you want me 
> > to implement optracker (without internal shard) per opwq shard
> >
> > Am I right ?
> >
> > Thanks & Regards
> > Somnath
> >
> > -----Original Message-----
> > From: Samuel Just [mailto:sam.j...@inktank.com]
> > Sent: Wednesday, September 10, 2014 3:08 PM
> > To: Somnath Roy
> > Cc: Sage Weil (sw...@redhat.com); ceph-de...@vger.kernel.org; 
> > ceph-users@lists.ceph.com
> > Subject: Re: OpTracker optimization
> >
> > I don't quite understand.
> > -Sam
> >
> > On Wed, Sep 10, 2014 at 2:38 PM, Somnath Roy <somnath....@sandisk.com> 
> > wrote:
> >> Thanks Sam.
> >> So, you want me to go with optracker/shadedopWq , right ?
> >>
> >> Regards
> >> Somnath
> >>
> >> -----Original Message-----
> >> From: Samuel Just [mailto:sam.j...@inktank.com]
> >> Sent: Wednesday, September 10, 2014 2:36 PM
> >> To: Somnath Roy
> >> Cc: Sage Weil (sw...@redhat.com); ceph-de...@vger.kernel.org; 
> >> ceph-users@lists.ceph.com
> >> Subject: Re: OpTracker optimization
> >>
> >> Responded with cosmetic nonsense.  Once you've got that and the other 
> >> comments addressed, I can put it in wip-sam-testing.
> >> -Sam
> >>
> >> On Wed, Sep 10, 2014 at 1:30 PM, Somnath Roy <somnath....@sandisk.com> 
> >> wrote:
> >>> Thanks Sam..I responded back :-)
> >>>
> >>> -----Original Message-----
> >>> From: ceph-devel-ow...@vger.kernel.org 
> >>> [mailto:ceph-devel-ow...@vger.kernel.org] On Behalf Of Samuel Just
> >>> Sent: Wednesday, September 10, 2014 11:17 AM
> >>> To: Somnath Roy
> >>> Cc: Sage Weil (sw...@redhat.com); ceph-de...@vger.kernel.org; 
> >>> ceph-users@lists.ceph.com
> >>> Subject: Re: OpTracker optimization
> >>>
> >>> Added a comment about the approach.
> >>> -Sam
> >>>
> >>> On Tue, Sep 9, 2014 at 1:33 PM, Somnath Roy <somnath....@sandisk.com> 
> >>> wrote:
> >>>> Hi Sam/Sage,
> >>>>
> >>>> As we discussed earlier, enabling the present OpTracker code 
> >>>> degrading performance severely. For example, in my setup a single 
> >>>> OSD node with
> >>>> 10 clients is reaching ~103K read iops with io served from memory 
> >>>> while optracking is disabled but enabling optracker it is reduced to 
> >>>> ~39K iops.
> >>>> Probably, running OSD without enabling OpTracker is not an option 
> >>>> for many of Ceph users.
> >>>>
> >>>> Now, by sharding the Optracker:: ops_in_flight_lock (thus xlist
> >>>> ops_in_flight) and removing some other bottlenecks I am able to 
> >>>> match the performance of OpTracking enabled OSD with OpTracking 
> >>>> disabled, but with the expense of ~1 extra cpu core.
> >>>>
> >>>> In this process I have also fixed the following tracker.
> >>>>
> >>>>
> >>>>
> >>>> http://tracker.ceph.com/issues/9384
> >>>>
> >>>>
> >>>>
> >>>> and probably http://tracker.ceph.com/issues/8885 too.
> >>>>
> >>>>
> >>>>
> >>>> I have created following pull request for the same. Please review it.
> >>>>
> >>>>
> >>>>
> >>>> https://github.com/ceph/ceph/pull/2440
> >>>>
> >>>>
> >>>>
> >>>> Thanks & Regards
> >>>>
> >>>> Somnath
> >>>>
> >>>>
> >>>>
> >>>>
> >>>> ________________________________
> >>>>
> >>>> PLEASE NOTE: The information contained in this electronic mail 
> >>>> message is intended only for the use of the designated 
> >>>> recipient(s) named above. If the reader of this message is not 
> >>>> the intended recipient, you are hereby notified that you have 
> >>>> received this message in error and that any review, 
> >>>> dissemination, distribution, or copying of this message is 
> >>>> strictly prohibited. If you have received this communication in 
> >>>> error, please notify the sender by telephone or e-mail (as shown 
> >>>> above) immediately and destroy any and all copies of this message in 
> >>>> your possession (whether hard copies or electronically stored copies).
> >>>>
> >>> --
> >>> To unsubscribe from this list: send the line "unsubscribe ceph-devel"
> >>> in the body of a message to majord...@vger.kernel.org More 
> >>> majordomo info at  http://vger.kernel.org/majordomo-info.html
> >>>
> >>> ________________________________
> >>>
> >>> PLEASE NOTE: The information contained in this electronic mail message is 
> >>> intended only for the use of the designated recipient(s) named above. If 
> >>> the reader of this message is not the intended recipient, you are hereby 
> >>> notified that you have received this message in error and that any 
> >>> review, dissemination, distribution, or copying of this message is 
> >>> strictly prohibited. If you have received this communication in error, 
> >>> please notify the sender by telephone or e-mail (as shown above) 
> >>> immediately and destroy any and all copies of this message in your 
> >>> possession (whether hard copies or electronically stored copies).
> >>>
> 
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to