> On Nov 17, 2014, at 15:52, Sage Weil <s...@newdream.net> wrote:
> 
> On Mon, 17 Nov 2014, Kevin Sumner wrote:
>> I?ve got a test cluster together with a ~500 OSDs and, 5 MON, and 1 MDS.  All
>> the OSDs also mount CephFS at /ceph.  I?ve got Graphite pointing at a space
>> under /ceph.  Over the weekend, I drove almost 2 million metrics, each of
>> which creates a ~3MB file in a hierarchical path, each sending a datapoint
>> into the metric file once a minute.  CephFS seemed to handle the writes ok
>> while I was driving load.  All files containing each metric are at paths
>> like this:
>> /ceph/whisper/sandbox/cephtest-osd0013/2/3/4/5.wsp
>> 
>> Today, however, with the load generator still running, reading metadata of
>> files (e.g. directory entries and stat(2) info) in the filesystem
>> (presumably MDS-managed data) seems nearly impossible, especially deeper
>> into the tree.  For example, in a shell cd seems to work but ls hangs,
>> seemingly indefinitely.  After turning off the load generator and allowing a
>> while for things to settle down, everything seems to behave better.
>> 
>> ceph status and ceph health both return good statuses the entire time.
>>  During load generation, the ceph-mds process seems pegged at between 100%
>> and 150%, but with load generation turned off, the process has some high
>> variability from near-idle up to similar 100-150% CPU.
>> 
>> Hopefully, I?ve missed something in the CephFS tuning.  However, I?m looking 
>> for
>> direction on figuring out if it is, indeed, a tuning problem or if this
>> behavior is a symptom of the ?not ready for production? banner in the
>> documentation.
> 
> My first guess is that the MDS cache is just too small and it is 
> thrashing.  Try
> 
> ceph mds tell 0 injectargs '--mds-cache-size 1000000'
> 
> That's 10x bigger than the default, tho be aware that it will eat up 10x 
> as much RAM too.
> 
> We've also seen teh cache behave in a non-optimal way when evicting 
> things, making it thrash more often than it should.  I'm hoping we can 
> implement something like MQ instead of our two-level LRU, but it isn't 
> high on the priority list right now.
> 
> sage


Thanks!  I’ll pursue mds cache size tuning.  Is there any guidance on setting 
the cache and other mds tunables correctly, or is it an adjust-and-test sort of 
thing?  Cursory searching doesn’t return any relevant documentation for 
ceph.com <http://ceph.com/>.  I’m plowing through some other list posts now.
--
Kevin Sumner
ke...@sumner.io

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to