Thanks Craig. That’s exactly what I was looking for. —Jiten
On Sep 16, 2014, at 2:42 PM, Craig Lewis <cle...@centraldesktop.com> wrote: > > > On Fri, Sep 12, 2014 at 4:35 PM, JIten Shah <jshah2...@me.com> wrote: > > 1. If we need to modify those numbers, do we need to update the values in > ceph.conf and restart every OSD or we can run a command on MON, that will > overwrite it? > > That will work. You can also update the values without a restart using: > ceph tell mon.\* injectargs '--mon_osd_nearfull_ratio 0.85' > > > You might also need to look at mon_osd_full_ratio, osd_backfill_full_ratio, > osd_failsafe_full_ratio, and osd_failsafe_nearfull_ratio. > > Variables that start with mon should be sent to all the monitors (ceph tell > mon.\* ...), variables that start with osd should be send to the osds (ceph > tell osd.\* ...). > > > > 2. What is the best way to get the OSD’s to work again, if we reach the full > ration amount? You can’t delete the data because read/write is blocked. > > Add more OSDs. Preferably before they become full, but it'll work if they're > toofull. It may take a while though, Ceph doesn't seem to weight which > backfills should be done first, so it might take a while to get to the OSDs > that are toofull. > > Since not everybody has nodes and disks laying around, you can stop all of > your writes, and bump the nearfull and full ratios. I've bumped them while I > was using ceph osd reweight, and had some toofull disks that wanted to > exchange PGs. Keep in mind that Ceph stops when the percentage is > than > toofull, so don't set full_ratio to 0.99. You really don't want to fill up > your disks. > > If all else fails (or you get a disk down to 0 kB free) you can manually > delete some PGs on disk. This is fairly risky, and prone to human error > causing data loss. You'll have to figure out the best ones to delete, and > you'll want to make sure you don't delete every replica of the PG. You'll > want to disable backfilling (ceph osd set nobackfill), otherwise Ceph will > repair things back to toofull. > > > > 3. If we add new OSD’s, will it start rebalancing the OSD’s or do I need to > trigger it manually and how? > > Adding and starting the OSDs will start rebalancing. The expected location > will change as soon as you add the OSD to the crushmap. Shortly after the > OSD starts, it will begin updating to make reality match expectations. For > most people, that happens in a single step, with ceph-deploy or a Config > Management tool. > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com