Hey Kenneth,

We encountered this when the number of strays (unlinked files yet to be
purged) reached 1 million, which is a result of many many file removals
happening on the fs repeatedly. It can also happen when there are more than
100k files in a dir with default settings.

You can tune it via 'mds_bal_fragment_size_max' setting on the mds either
temporarily to rm files or permanently. Beware setting it too high.

Check num strays in mds cache by running `ceph daemon mds.{mds name} perf
dump` and inspecting the mds cache section for num_strays. The 1 million
limit is a multiple/function of the mds bal fragment size (10x).

Raf


On Fri, May 10, 2019, 9:03 PM Kenneth Waegeman <kenneth.waege...@ugent.be>
wrote:

> Hi all,
>
> I am seeing issues on cephfs running 13.2.5 when deleting files:
>
> [root@osd006 ~]# rm /mnt/ceph/backups/osd006.gigalith.os-2b5a3740.1326700
> rm: remove regular empty file
> ‘/mnt/ceph/backups/osd006.gigalith.os-2b5a3740.1326700’? y
> rm: cannot remove
> ‘/mnt/ceph/backups/osd006.gigalith.os-2b5a3740.1326700’: No space left
> on device
>
> few minutes later, I can remove it without problem. This happens
> especially when there are a lot of files deleted somewhere on the
> filesystem around the same time.
>
> We already have tuned our mds config:
>
> [mds]
> mds_cache_memory_limit=10737418240
> mds_log_max_expiring=200
> mds_log_max_segments=200
> mds_max_purge_files=2560
> mds_max_purge_ops=327600
> mds_max_purge_ops_per_pg=20
>
> ceph -s is reporting everything clean, and the file system space usage
> is less than 50%, also no full osds or anything.
>
> Is there a way to further debug what the bottleneck is when removing
> files that gives this 'no space left on device' error?
>
>
> Thank you very much!
>
> Kenneth
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to