The output of status command of fuse daemon: "dentry_count": 128966, "dentry_pinned_count": 128965, "inode_count": 409696, I saw the pinned dentry is nearly the same as dentry. So I enabled debug log(debug client = 20/20) and read Client.cc source code in general. I found that an entry will not be trimed if it is pinned. But how can I unpin dentrys?
On Wed, Apr 29, 2015 at 12:19 PM Dexter Xiong <dxtxi...@gmail.com> wrote: > I tried set client cache size = 100, but it doesn't solve the problem. > I tested ceph-fuse with kernel version 3.13.0-24 3.13.0-49 and 3.16.0-34. > > > > On Tue, Apr 28, 2015 at 7:39 PM John Spray <john.sp...@redhat.com> wrote: > >> >> >> On 28/04/2015 06:55, Dexter Xiong wrote: >> > Hi, >> > I've deployed a small hammer cluster 0.94.1. And I mount it via >> > ceph-fuse on Ubuntu 14.04. After several hours I found that the >> > ceph-fuse process crashed. The end is the crash log from >> > /var/log/ceph/ceph-client.admin.log. The memory cost of ceph-fuse >> > process was huge(more than 4GB) when it crashed. >> > Then I did some test and found these actions will increase memory >> > cost of ceph-fuse rapidly and the memory cost never seem to decrease: >> > >> > * rsync command to sync small files(rsync -a /mnt/some_small >> /srv/ceph) >> > * chown command/ chmod command(chmod 775 /srv/ceph -R) >> > >> > But chown/chmod command on accessed files will not increase the memory >> > cost. >> > It seems that ceph-fuse caches the file nodes but never releases them. >> > I don't know if there is an option to control the cache size. I >> > set mds cache size = 2147483647 option to improve the performance of >> > mds, and I tried to set mds cache size = 1000 at client side but it >> > doesn't effect the result. >> >> The setting for client-side cache limit is "client cache size", default >> is 16384 >> >> What kernel version are you using on the client? There have been some >> issues with cache trimming vs. fuse in recent kernels, but we thought we >> had workarounds in place... >> >> Cheers, >> John >> >>
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com