What # of objects do you have? After all, such large footprint can be just a bug in your build if you do not have ultimate high object count(>~1e8) or any extraordinary configuration parameter.
On Mon, Apr 28, 2014 at 12:26 AM, Gandalf Corvotempesta <gandalf.corvotempe...@gmail.com> wrote: > So, are you suggesting to lower the pg count ? > Actually i'm using the suggested number of OSD*100/Replicas > and I have just 2 OSDs per server. > > > 2014-04-24 19:34 GMT+02:00 Andrey Korolyov <and...@xdel.ru>: >> On 04/24/2014 08:14 PM, Gandalf Corvotempesta wrote: >>> During a recovery, I'm hitting oom-killer for ceph-osd because it's >>> using more than 90% of avaialble ram (8GB) >>> >>> How can I decrease the memory footprint during a recovery ? >> >> You can reduce pg count per OSD for example, it scales down well enough. >> OSD memory footprint (during recovery or normal operations) depends of >> number of objects, e.g. commited data and total count of PGs per OSD. >> Because deleting some data is not an option, I may suggest only one >> remaining way :) >> >> I had raised related question a long ago, it was about post-recovery >> memory footprint patterns - OSD shrinks memory usage after successful >> recovery in a relatively long period, up to some days and by couple of >> fast 'leaps'. Heap has nothing to do with this bug I had not profiled >> the daemon itself yet. >> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >> _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com