Sorry if I wasn't clear. Going from 2GB to 8GB is not normal, although some slight bloating is expected. In your case it just got much worse than usual for reasons yet unknown.
Jan > On 09 Sep 2015, at 12:40, Mariusz Gronczewski > <mariusz.gronczew...@efigence.com> wrote: > > > well I was going by > http://ceph.com/docs/master/start/hardware-recommendations/ and planning for > 2GB per OSD so that was a suprise.... maybe there should be warning somewhere > ? > > > On Wed, 9 Sep 2015 12:21:15 +0200, Jan Schermer <j...@schermer.cz> wrote: > >> The memory gets used for additional PGs on the OSD. >> If you were to "swap" PGs between two OSDs, you'll get memory wasted on both >> of them because tcmalloc doesn't release it.* >> It usually gets stable after few days even during backfills, so it does get >> reused if needed. >> If for some reason your OSDs get to 8GB RSS then I recommend you just get >> more memory, or try disabling tcmalloc which can either help or make it even >> worse :-) >> >> * E.g. if you do something silly like "ceph osd crush reweight osd.1 10000" >> you will see the RSS of osd.28 skyrocket. Reweighting it back down will not >> release the memory until you do "heap release". >> >> Jan >> >> >>> On 09 Sep 2015, at 12:05, Mariusz Gronczewski >>> <mariusz.gronczew...@efigence.com> wrote: >>> >>> On Tue, 08 Sep 2015 16:14:15 -0500, Chad William Seys >>> <cws...@physics.wisc.edu> wrote: >>> >>>> Does 'ceph tell osd.* heap release' help with OSD RAM usage? >>>> >>>> From >>>> http://lists.ceph.com/pipermail/ceph-users-ceph.com/2015-August/003932.html >>>> >>>> Chad. >>> >>> it did help now, but cluster is in clean state at the moment. But I >>> didnt know that one, thanks. >>> >>> High memory usage stopped once cluster rebuilt, but I've planned >>> cluster to have 2GB per OSD so I needed to add ram to even get to the >>> point of ceph starting to rebuild, as some OSD ate up to 8 GBs during >>> recover >>> >>> -- >>> Mariusz Gronczewski, Administrator >>> >>> Efigence S. A. >>> ul. Wołoska 9a, 02-583 Warszawa >>> T: [+48] 22 380 13 13 >>> F: [+48] 22 380 13 14 >>> E: mariusz.gronczew...@efigence.com >>> <mailto:mariusz.gronczew...@efigence.com> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > > > > -- > Mariusz Gronczewski, Administrator > > Efigence S. A. > ul. Wołoska 9a, 02-583 Warszawa > T: [+48] 22 380 13 13 > F: [+48] 22 380 13 14 > E: mariusz.gronczew...@efigence.com > <mailto:mariusz.gronczew...@efigence.com> _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com