On Wed, Aug 1, 2018 at 6:43 AM Kamble, Nitin A <nitin.kam...@teradata.com> wrote:
> Hi John, > > > > I am running ceph Luminous 12.2.1 release on the storage nodes with > v4.4.114 kernel on the cephfs clients. > > > > 3 client nodes are running 3 instances of a test program. > > The test program is doing this repeatedly in a loop: > > - sequentially write a 256GB file on cephfs > - delete the file > > do the clients write to the same file? I mean same file name in a directory. > > > ‘ceph df’ shows that after delete the space is not getting freed from > cephfs and and cephfs space utilization (number of objects, space used and > % utilization) keeps growing up continuously. > > > > I double checked, and no process is holding an open handle to the closed > files. > > > > When the test program is stopped, the writing workload stops and then the > cephfs space utilization starts going down as expected. > > > > Looks like the cephfs write load is not giving enough opportunity to > actually perform the delete file operations from clients. It is a > consistent behavior, and easy to reproduce. > > > > I tried playing with these advanced MDS config parameters: > > - mds_max_purge_files > - mds_max_purge_ops > - mds_max_purge_ops_per_pg > - mds_purge_queue_busy_flush_period > > > > But it is not helping with the workload. > > > > Is this a known issue? And is there a workaround to give more priority to > the objects purging operations? > > > > Thanks in advance, > > Nitin > > > > *From: *ceph-users <ceph-users-boun...@lists.ceph.com> on behalf of > Alexander Ryabov <arya...@intermedia.net> > *Date: *Thursday, July 19, 2018 at 8:09 AM > *To: *John Spray <jsp...@redhat.com> > *Cc: *"ceph-users@lists.ceph.com" <ceph-users@lists.ceph.com> > *Subject: *Re: [ceph-users] Force cephfs delayed deletion > > > > >Also, since I see this is a log directory, check that you don't have > some processes that are holding their log files open even after they're > unlinked. > > Thank you very much - that was the case. > > lsof /mnt/logs | grep deleted > > > > After dealing with these, space was reclaimed in about 2-3min. > > > > > ------------------------------ > > *From:* John Spray <jsp...@redhat.com> > *Sent:* Thursday, July 19, 2018 17:24 > *To:* Alexander Ryabov > *Cc:* ceph-users@lists.ceph.com > *Subject:* Re: [ceph-users] Force cephfs delayed deletion > > > > On Thu, Jul 19, 2018 at 1:58 PM Alexander Ryabov <arya...@intermedia.net> > wrote: > > Hello, > > I see that free space is not released after files are removed on CephFS. > > I'm using Luminous with replica=3 without any snapshots etc and with > default settings. > > > > From client side: > > $ du -sh /mnt/logs/ > > 4.1G /mnt/logs/ > > $ df -h /mnt/logs/ > > Filesystem Size Used Avail Use% Mounted on > > h1,h2:/logs 125G 87G 39G 70% /mnt/logs > > > > These stats are after couple of large files were removed in /mnt/logs dir, > but that only dropped Useв space a little. > > > > Check what version of the client you're using -- some older clients had > bugs that would hold references to deleted files and prevent them from > being purged. If you find that the space starts getting freed when you > unmount the client, this is likely to be because of a client bug. > > > > Also, since I see this is a log directory, check that you don't have some > processes that are holding their log files open even after they're unlinked. > > > > John > > > > > > > > Doing 'sync' command also changes nothing. > > > > From server side: > > # ceph df > > GLOBAL: > > SIZE AVAIL RAW USED %RAW USED > > 124G 39226M 88723M 69.34 > > POOLS: > > NAME ID USED %USED MAX AVAIL OBJECTS > > cephfs_data 1 28804M 76.80 8703M 7256 > > cephfs_metadata 2 236M 2.65 8703M 101 > > > > Why there are such a large difference between 'du' and 'USED'? > > I've found that it could be due to 'delayed delete' > http://docs.ceph.com/docs/luminous/dev/delayed-delete/ > <https://imsva91-ctp.trendmicro.com:443/wis/clicktime/v1/query?url=https%3a%2f%2furl.emailprotection.link%2f%3fa%2dzj7k72CS0gb415itDM1eU90VEm5BhojI3q4cHQrsilpaYjPTTNWfFTqC14bd5j2XtNq%2dUFuEZul7eZHtnh%2d5g%7e%7e&umid=A4F1CB59-715B-8E05-AA77-B7AD0CCE487F&auth=5e584526fc71bf85011d6d2e8a81aa05f4bd018d-f875139c2b2492612900e3cbf4d3611ffd4146fd> > > And previously it seems could be tuned by adjusting the "mds max purge > files" and "mds max purge ops" > > > http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-October/013679.html > <https://imsva91-ctp.trendmicro.com:443/wis/clicktime/v1/query?url=https%3a%2f%2furl.emailprotection.link%2f%3fahP6OBo0IWu9FuToCxmcjXNlOi%5fXi08d0Sp5qC%2dIarmsg%2dWUfPMfZGva3XnsWI0H5JwptWRr%5fO01mBq7%5fSTKpTpSIHa813nlawWz6bWtlwV0p%5fVIDvbWClfITb4PwNDx3&umid=A4F1CB59-715B-8E05-AA77-B7AD0CCE487F&auth=5e584526fc71bf85011d6d2e8a81aa05f4bd018d-d910bc9758f99a5deeee06f7ee9186b56cbfc3ba> > > But there is no more of such options in > http://docs.ceph.com/docs/luminous/cephfs/mds-config-ref/ > <https://imsva91-ctp.trendmicro.com:443/wis/clicktime/v1/query?url=https%3a%2f%2furl.emailprotection.link%2f%3fa%2dzj7k72CS0gb415itDM1eU90VEm5BhojI3q4cHQrsimLce%2dzmpj%2dukHA75X0Jqtb8s9Rl8x3JX2E68Wl%2d6DuGg%7e%7e&umid=A4F1CB59-715B-8E05-AA77-B7AD0CCE487F&auth=5e584526fc71bf85011d6d2e8a81aa05f4bd018d-ee5880c281daff327c5800bae0292c5dc74ba0b7> > > > > So the question is - how to purge deleted data and reclaim free space? > > Thank you. > > > > ______________________________________________ > > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > <https://imsva91-ctp.trendmicro.com:443/wis/clicktime/v1/query?url=https%3a%2f%2furl.emailprotection.link%2f%3fahP6OBo0IWu9FuToCxmcjXCl2cKOQOsBa372K2gMLXSslcSDG0hOfqsfgDj2933TkMvnibtT%5f5Wj7G0DNATrJ5g%7e%7e&umid=A4F1CB59-715B-8E05-AA77-B7AD0CCE487F&auth=5e584526fc71bf85011d6d2e8a81aa05f4bd018d-6997b2c56f1e5034a20a9557401fcadcf9cad590> > > > Intermedia > 10th floor, 2, Alexander Nevsky Sq. > Saint-Petersburg, Russia 191167 > > www.intermedia.net > > [image: J.D. Power] > > J.D. Power certifies Intermedia for technical support excellence two years > running, a first among cloud application providers > > > > > ------------------------------ > > > This message is intended only for the person(s) to which it is addressed > and may contain Intermedia.net. Inc. privileged, confidential and/or > proprietary information. If you have received this communication in error, > please notify us immediately by replying to the message and deleting it > from your computer. Any disclosure, copying, distribution, or the taking of > any action concerning the contents of this message and any attachment(s) by > anyone other than the named recipient(s) is strictly prohibited. > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com