Hi Sean,

Isn't there any downsides to increasing the mds cache size?
My colleague mentioned that he tested it previously and then the cluster
didn't recover during a failover..

On Thu, Jun 9, 2016 at 12:41 PM, Sean Crosby <richardnixonsh...@gmail.com>
wrote:

> Hi Elias,
>
> When we have received the same warning, our solution has been to increase
> the inode cache on the MDS.
>
> We have added
>
> mds cache size = 2000000
>
>
> to the [global] section of ceph.conf on the MDS server. We have to restart
> MDS for the changes to be applied.
>
> Sean
>
>
> On 9 June 2016 at 19:55, Elias Abacioglu <
> elias.abacio...@deltaprojects.com> wrote:
>
>> Hi,
>>
>> I know this have been asked here a couple of times, but couldn't find
>> anything concrete.
>>
>> I have the following warning in our ceph cluster.
>> mds0: Client web01:cephfs.web01 failing to respond to cache pressure
>>
>> In previous Ceph versions this might have been a bug. But now we are
>> running Jewel.
>> So is there a way to fix this warning?
>> Do I need to tune some values? Boost the cluster? Boost the client?
>>
>> Here are some details:
>> Client kernel is 4.4.0.
>> Ceph 10.2.1
>>
>> # ceph mds dump
>> dumped fsmap epoch 5755
>> fs_name    cephfs
>> epoch    5755
>> flags    0
>> created    2015-12-03 11:21:28.128193
>> modified    2016-05-16 06:48:47.969430
>> tableserver    0
>> root    0
>> session_timeout    60
>> session_autoclose    300
>> max_file_size    1099511627776
>> last_failure    4900
>> last_failure_osd_epoch    5884
>> compat    compat={},rocompat={},incompat={1=base v0.20,2=client writeable
>> ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds
>> uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table}
>> max_mds    1
>> in    0
>> up    {0=574261}
>> failed
>> damaged
>> stopped
>> data_pools    2
>> metadata_pool    3
>> inline_data    disabled
>> 574261:    10.3.215.5:6801/62035 'ceph-mds03' mds.0.5609 up:active seq
>> 515014
>> 594257:    10.3.215.10:6800/1386 'ceph-mds04' mds.0.0 up:standby-replay
>> seq 1
>>
>> Thanks,
>> Elias
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to