On Thu, Mar 7, 2019 at 2:38 PM Patrick Donnelly <pdonn...@redhat.com> wrote:

> Is this with one active MDS and one standby-replay? The graph is odd
> to me because the session count shows sessions on fs-b and fs-d but
> not fs-c. Or maybe max_mds=2 and fs-d has no activity and fs-c is
> standby-replay?
>

The graphs were taken when we were running with 2 active MDS and 2
standby-replay. Currently we're running with 1 active and 1 standby-replay.


> Three OSDs are going to really struggle with the client load you're
> putting on it. It doesn't surprise me you are getting slow requests
> warning on the MDS for this reason. When you were running Luminous
> 12.2.9+ or Mimic 13.2.2+, were you seeing slow metadata I/O warnings?
> Even if you did not, it possible that the MDS is delayed issuing caps
> to clients because it's waiting for another client to flush writes and
> release conflicting caps.
>

We didn't see any slow metadata I/O warnings, but this is the sort of thing
that I've been suspecting is the underlying issue. However, one thing to
note is that my current test only walks a directory and reads all the files
in it, and it's running on a dev cluster that only I'm using, so I'm not
sure which client would be generating writes that the MDS would be waiting
on.


> Generally we recommend that the metadata pool be located on OSDs with
> fast devices separate from the data pool. This avoids priority
> inversion of MDS metadata I/O with data I/O. See [1] to configure the
> metadata pool on a separate set of OSDs.
>
> Also, you're not going to saturate a 1.9TB NVMe SSD with one OSD. You
> must partition it and setup multiple OSDs. This ends up being positive
> for you so that you can put the metadata pool on its own set of OSDs.
>
> [1] https://ceph.com/community/new-luminous-crush-device-classes/


This is excellent info, especially the second point as it lets me split out
the metadata pool onto separate OSDs without spinning up any additional
resources or using EBS volumes. I will look into partitioning the NVMe
drives, split out the metadata pool, and check how this impacts performance.

Thanks,
Zack
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to