I have installed diamond(built by ksingh found at
https://github.com/ksingh7/ceph-calamari-packages) on the MDS node and I am
not seeing the mds_server.handle_client_request OR objecter.op_active
metrics being sent to graphite.  Mind you, this is not the graphite that is
part of the calamari install but our own internal graphite cluster.
Perhaps that is the reason?  I could not get calamari working correctly on
hammerhead/centos7.1 so I put it on pause for now to concentrate on the
cluster itself.

Ultimately, I need to find a way to get a hold of these metrics to
determine the health of my MDS so I can justify moving forward on a SSD
based cephfs metadata pool.

On Wed, Aug 5, 2015 at 4:05 PM, Bob Ababurko <b...@ababurko.net> wrote:

> Hi John,
>
> You are correct in that my expectations may be incongruent with what is
> possible with ceph(fs).  I'm currently copying many small files(images)
> from a netapp to the cluster...~35k sized files to be exact and the number
> of objects/files copied thus far is fairly significant(below in bold):
>
> [bababurko@cephmon01 ceph]$ sudo rados df
> pool name                 KB      objects       clones     degraded
>  unfound           rd        rd KB           wr        wr KB
> cephfs_data       3289284749    *163993660*            0            0
>       0            0            0    328097038   3369847354
> cephfs_metadata       133364       524363            0            0
>     0      3600023   5264453980     95600004   1361554516
> rbd                        0            0            0            0
>     0            0            0            0            0
>   total used      9297615196    164518023
>   total avail    19990923044
>   total space    29288538240
>
> Yes, that looks like ~164 million objects copied to the cluster.  I would
> assume this will potentially be a burden to the MDS but I have yet to
> confirm with the ceph daemontool mds.<id>.  I cannot seem to run it on the
> mds host as it doesn't seem to know about that command:
>
> [bababurko@cephmds01]$ sudo ceph daemonperf mds.cephmds01
> no valid command found; 10 closest matches:
> osd lost <int[0-]> {--yes-i-really-mean-it}
> osd create {<uuid>}
> osd primary-temp <pgid> <id>
> osd primary-affinity <osdname (id|osd.id)> <float[0.0-1.0]>
> osd reweight <int[0-]> <float[0.0-1.0]>
> osd pg-temp <pgid> {<id> [<id>...]}
> osd in <ids> [<ids>...]
> osd rm <ids> [<ids>...]
> osd down <ids> [<ids>...]
> osd out <ids> [<ids>...]
> Error EINVAL: invalid command
>
> This fails in a similar manner on all the hosts in the cluster.  I'm very
> green w/ ceph and i'm probably missing something obvious.  Is there
> something I need to install to get access to the 'ceph daemonperf' command
> in hammerhead?
>
> thanks,
> Bob
>
> On Wed, Aug 5, 2015 at 2:43 AM, John Spray <jsp...@redhat.com> wrote:
>
>> On Tue, Aug 4, 2015 at 10:36 PM, Bob Ababurko <b...@ababurko.net> wrote:
>> > My writes are not going as I would expect wrt to IOPS(50-1000 IOPs) &
>> write
>> > throughput( ~25MB/s max).  I'm interested in understanding what it
>> takes to
>> > create a SSD pool that I can then migrate the current Cephfs_metadata
>> pool
>> > to.  I suspect that the spinning disk metadata pool is a bottleneck and
>> I
>> > want to try to get the max performance out of this cluster to prove
>> that we
>> > would build out a larger version.  One caveat is that I have copied
>> about 4
>> > TB of data to the cluster via cephfs and dont want to lose the data so I
>> > obviously need to keep the metadata intact.
>>
>> I'm a bit suspicious of this: your IOPS expectations sort of imply
>> doing big files, but you're then suggesting that metadata is the
>> bottleneck (i.e. small file workload).
>>
>> There are lots of statistics that come out of the MDS, you may be
>> particular interested in mds_server.handle_client_request,
>> objecter.op_active, to work out if there really are lots of RADOS
>> operations getting backed up on the MDS (which would be the symptom of
>> a too-slow metadata pool).  "ceph daemonperf mds.<id>" may be some
>> help if you don't already have graphite or similar set up.
>>
>> > If anyone has done this OR understands how this can be done, I would
>> > appreciate the advice.
>>
>> You could potentially do this in a two-phase process where you
>> initially set a crush rule that includes both SSDs and spinners, and
>> then finally set a crush rule that just points to SSDs.  Obviously
>> that'll do lots of data movement, but your metadata is probably a fair
>> bit smaller than your data so that might be acceptable.
>>
>> John
>>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to