Hi Elias, When we have received the same warning, our solution has been to increase the inode cache on the MDS.
We have added mds cache size = 2000000 to the [global] section of ceph.conf on the MDS server. We have to restart MDS for the changes to be applied. Sean On 9 June 2016 at 19:55, Elias Abacioglu <elias.abacio...@deltaprojects.com> wrote: > Hi, > > I know this have been asked here a couple of times, but couldn't find > anything concrete. > > I have the following warning in our ceph cluster. > mds0: Client web01:cephfs.web01 failing to respond to cache pressure > > In previous Ceph versions this might have been a bug. But now we are > running Jewel. > So is there a way to fix this warning? > Do I need to tune some values? Boost the cluster? Boost the client? > > Here are some details: > Client kernel is 4.4.0. > Ceph 10.2.1 > > # ceph mds dump > dumped fsmap epoch 5755 > fs_name cephfs > epoch 5755 > flags 0 > created 2015-12-03 11:21:28.128193 > modified 2016-05-16 06:48:47.969430 > tableserver 0 > root 0 > session_timeout 60 > session_autoclose 300 > max_file_size 1099511627776 > last_failure 4900 > last_failure_osd_epoch 5884 > compat compat={},rocompat={},incompat={1=base v0.20,2=client writeable > ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds > uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table} > max_mds 1 > in 0 > up {0=574261} > failed > damaged > stopped > data_pools 2 > metadata_pool 3 > inline_data disabled > 574261: 10.3.215.5:6801/62035 'ceph-mds03' mds.0.5609 up:active seq > 515014 > 594257: 10.3.215.10:6800/1386 'ceph-mds04' mds.0.0 up:standby-replay > seq 1 > > Thanks, > Elias > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com