2016-12-16 10:19 GMT+01:00 Wido den Hollander <w...@42on.com>:

>
> > Op 16 december 2016 om 9:49 schreef Alessandro Brega <
> alessandro.bre...@gmail.com>:
> >
> >
> > 2016-12-16 9:33 GMT+01:00 Wido den Hollander <w...@42on.com>:
> >
> > >
> > > > Op 16 december 2016 om 9:26 schreef Alessandro Brega <
> > > alessandro.bre...@gmail.com>:
> > > >
> > > >
> > > > Hi guys,
> > > >
> > > > I'm running a ceph cluster using 0.94.9-1trusty release on XFS for
> RBD
> > > > only. I'd like to replace some SSDs because they are close to their
> TBW.
> > > >
> > > > I know I can simply shutdown the OSD, replace the SSD, restart the
> OSD
> > > and
> > > > ceph will take care of the rest. However I don't want to do it this
> way,
> > > > because it leaves my cluster for the time of the rebalance/
> backfilling
> > > in
> > > > a degraded state.
> > > >
> > > > I'm thinking about this process:
> > > > 1. keep old OSD running
> > > > 2. copy all data from current OSD folder to new OSD folder (using
> rsync)
> > > > 3. shutdown old OSD
> > > > 4. redo step 3 to update to the latest changes
> > > > 5. restart OSD with new folder
> > > >
> > > > Are there any issues with this approach? Do I need any special rsync
> > > flags
> > > > (rsync -avPHAX --delete-during)?
> > > >
> > >
> > > Indeed X for transferring xattrs, but also make sure that the
> partitions
> > > are GPT with the proper GUIDs.
> > >
> > > I would never go for this approach in a running setup. Since it's a SSD
> > > cluster I wouldn't worry about the rebalance and just have Ceph do the
> work
> > > for you.
> > >
> > >
> > Why not - if it's completely safe. It's much faster (local copy), doesn't
> > put load on the network (local copy), much safer (2-3 minutes instead of
> > 1-2 hours degraded time (2TB SSD)), and it's really simple (2 rsync
> > commands). Thank you.
> >
>
> I wouldn't say it is completely safe, hence my remark. If you copy, indeed
> make sure you copy all the xattrs, but also make sure the partitions tables
> match.
>
> That way it should work, but it's not a 100% guarantee.
>
>
Ok, thanks!  Can a ceph dev confirm? I do not want to loose any data ;)

Alessandro
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to