Thanks John.  I'll give that a try as soon as I fix an issue with my MDS
servers that cropped up today.

On Mon, Aug 10, 2015 at 2:58 AM, John Spray <jsp...@redhat.com> wrote:

> On Fri, Aug 7, 2015 at 1:36 AM, Bob Ababurko <b...@ababurko.net> wrote:
> > @John,
> >
> > Can you clarify which values would suggest that my metadata pool is too
> > slow?   I have added a link that includes values for the "op_active" &
> > "handle_client_request"....gathered in a crude fashion but should
> hopefully
> > give enough data to paint a picture of what is happening.
> >
> > http://pastebin.com/5zAG8VXT
>
> Dividing by the first 20s of the second sample period, you're seeing
> ~750 client metadata operations handled per second, which is kind of a
> baseline level of performance (a little better than what I get running
> a ceph cluster locally on my workstation).  That's probably
> corresponding to roughly the same number of file creates per second --
> your workload is very much a small file one, where "files per second"
> is a much more meaningful measure than IOPS or MB/s.
>
> It does look like the kind of pattern where you've got a large clutch
> of several thousand metadata pool rados ops coming out every few
> seconds, then draining out over a few seconds.  Your metadata pool
> isn't pathologically slow (it's completing at least hundreds of ops
> per second), but it is noticeable that during some periods where
> op_active is draining, handle_client_request is not incrementing --
> i.e. client metadata ops are stalling while the MDS waits for its
> RADOS operations to complete.
>
> I can't say a massive amount beyond that, other than what you'd
> already figured out -- it would be worth trying to put some faster
> storage in for your metadata pool.
>
> John
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to