Given that the MDS knows everything, it seems trivial to add a ceph 'mv' command to do this. I looked at using tiering to try and do the move, but I don't know how to tell cephfs that the data is now on the new pool instead of the old pool name. Since we can't take a long enough downtime to move hundreds of Terabytes, we need something that can be done online, and if it has a minute or two of downtime would be okay. ---------------- Robert LeBlanc PGP Fingerprint 79A2 9CA4 6CC4 45DD A904 C70E E654 3BB2 FA62 B9F1
On Fri, Jun 28, 2019 at 9:02 AM Marc Roos <m.r...@f1-outsourcing.eu> wrote: > > > 1. > change data pool for a folder on the file system: > setfattr -n ceph.dir.layout.pool -v fs_data.ec21 foldername > > 2. > cp /oldlocation /foldername > Remember that you preferably want to use mv, but this leaves (meta) data > on the old pool, that is not what you want when you want to delete that > pool. > > 3. When everything is copied-removed, you should end up with an empty > datapool with zero objects. > > 4. Verify here with others, if you can just remove this one. > > I think this is a reliable technique to switch, because you use the > basic cephfs functionality that supposed to work. I prefer that the ceph > guys implement a mv that does what you expect from it. Now it acts more > or less like a linking. > > > > > -----Original Message----- > From: Jorge Garcia [mailto:jgar...@soe.ucsc.edu] > Sent: vrijdag 28 juni 2019 17:52 > To: Marc Roos; ceph-users > Subject: Re: [ceph-users] Migrating a cephfs data pool > > Are you talking about adding the new data pool to the current > filesystem? Like: > > $ ceph fs add_data_pool my_ceph_fs new_ec_pool > > I have done that, and now the filesystem shows up as having two data > pools: > > $ ceph fs ls > name: my_ceph_fs, metadata pool: cephfs_meta, data pools: > [cephfs_data new_ec_pool ] > > but then I run into two issues: > > 1. How do I actually copy/move/migrate the data from the old pool to the > new pool? > 2. When I'm done moving the data, how do I get rid of the old data pool? > > I know there's a rm_data_pool option, but I have read on the mailing > list that you can't remove the original data pool from a cephfs > filesystem. > > The other option is to create a whole new cephfs with a new metadata > pool and the new data pool, but creating multiple filesystems is still > experimental and not allowed by default... > > On 6/28/19 8:28 AM, Marc Roos wrote: > > > > What about adding the new data pool, mounting it and then moving the > > files? (read copy because move between data pools does not what you > > expect it do) > > > > > > -----Original Message----- > > From: Jorge Garcia [mailto:jgar...@soe.ucsc.edu] > > Sent: vrijdag 28 juni 2019 17:26 > > To: ceph-users > > Subject: *****SPAM***** [ceph-users] Migrating a cephfs data pool > > > > This seems to be an issue that gets brought up repeatedly, but I > > haven't seen a definitive answer yet. So, at the risk of repeating a > > question that has already been asked: > > > > How do you migrate a cephfs data pool to a new data pool? The obvious > > case would be somebody that has set up a replicated pool for their > > cephfs data and then wants to convert it to an erasure code pool. Is > > there a simple way to do this, other than creating a whole new ceph > > cluster and copying the data using rsync? > > > > Thanks for any clues > > > > Jorge > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com