Does rebalancing across multiple new OSDs at once like that affect cluster performance more or less than one at a time?
On Mar 7, 2014, at 2:04 PM, Michael J. Kidd <michael.k...@inktank.com> wrote: > Hello Sid, > You may try setting the 'noup' flag (instead of the 'noout' flag). This > would prevent new OSDs from being set 'up' and therefore, the data rebalance > shouldn't occur. Once you add all OSDs, then unset the 'noup' flag and > ensure they're set 'up' automatically... if not, use 'ceph osd up <osdid>' to > bring them up manually. > > Hope this helps! > > Michael J. Kidd > Sr. Storage Consultant > Inktank Professional Services > > > On Fri, Mar 7, 2014 at 3:06 PM, Sidharta Mukerjee <smukerje...@gmail.com> > wrote: > When I use ceph-deploy to add a bunch of new OSDs (from a new machine), the > ceph cluster starts rebalancing immediately; as a result, the first couple > OSDs are started properly; but the last few can't start because I keep > getting a "timeout problem", as shown here: > > [root@ia6 ia_scripts]# service ceph start osd.24 > === osd.24 === > failed: 'timeout 10 /usr/bin/ceph --name=osd.24 > --keyring=/var/lib/ceph/osd/ceph-24/keyring osd crush > create-or-move -- 24 > 1.82 root=default host=ia6 > > Is there a way I can pause the "recovery" so that the overall system behaves > way faster and I can then start all the OSDs, make sure they're up and they > look "normal" (via ceph osd tree) , and then unpause recovery? > > -Sid > _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com