2018-06-13 23:53 GMT+02:00 <c...@elchaka.de>: > Hi yao, > > IIRC there is a *sleep* Option which is usefull when delete Operation is > being done from ceph.... sleep_trim or something like that. >
you are thinking of "osd_snap_trim_sleep" which is indeed a very helpful option - but not for deletions. It rate limites snapshot deletion only. Paul > > - Mehmet > > Am 7. Juni 2018 04:11:11 MESZ schrieb Yao Guotao <yaoguo_...@163.com>: >> >> Hi Jason, >> >> Thank you very much for your reply. >> I think the RBD trash is a good way. But, the QoS in Ceph is a better >> solution. >> I am looking forward to the backend QoS of Ceph. >> >> Thanks. >> >> >> At 2018-06-06 21:53:23, "Jason Dillaman" <jdill...@redhat.com> wrote: >> >The 'rbd_concurrent_management_ops' setting controls how many >> >concurrent, in-flight RADOS object delete operations are possible per >> >image removal. The default is only 10, so given ten 10 images being >> >deleted concurrently, I am actually surprised that blocked all IO from >> >your VMs. >> > >> >Adding support for limiting the maximum number of concurrent image >> >deletions would definitely be an OpenStack enhancement. There is an >> >open blueprint for optionally utilizing the RBD trash instead of >> >having Cinder delete the images [1], which would allow you to defer >> >deletions to whenever is convenient. Additionally, once Ceph adds >> >support for backend QoS (fingers crossed in Nautilus), we can change >> >librbd to flag all IO for maintenance activities to background (best >> >effort) priority, which might be the best long-term solution. >> > >> >[1] >> >https://blueprints.launchpad.net/cinder/+spec/rbd-deferred-volume-deletion >> > >> >On Wed, Jun 6, 2018 at 6:40 AM, Yao Guotao <yaoguo_...@163.com> wrote: >> >> Hi Cephers, >> >> >> >> We use Ceph with Openstack by librbd library. >> >> >> >> Last week, my colleague delete 10 volumes from Openstack dashboard at the >> >> same time, each volume has about 1T used. >> >> During this time, the disk of OSDs are busy, and there have no I/O for >> >> normal vm. >> >> >> >> So, I want to konw if there are any parameters that can be set to >> >> throttle? >> >> >> >> I find a parameter about rbd op is 'rbd_concurrent_management_ops'. >> >> I am trying to figure out how it works in code, and I find the parameter >> >> can >> >> only control the asyncchronous deletion of all objects of an image. >> >> >> >> Besides, Should it be controlled at Openstack Nova or Cinder layer? >> >> >> >> Thanks, >> >> Yao Guotao >> >> >> >> >> >> >> >> >> >> >> >> _______________________________________________ >> >> ceph-users mailing list >> >> ceph-users@lists.ceph.com >> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> >> > >> > >> > >> >-- >> >Jason >> >> >> >> >> > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > -- Paul Emmerich Looking for help with your Ceph cluster? Contact us at https://croit.io croit GmbH Freseniusstr. 31h 81247 München www.croit.io Tel: +49 89 1896585 90
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com