John, Thanks for your answers. I have a clarification on my questions see below inline. Bruce
From: John Spray <jsp...@redhat.com> Date: Thursday, June 8, 2017 at 1:45 AM To: "McFarland, Bruce" <bruce.mcfarl...@teradata.com> Cc: "ceph-users@lists.ceph.com" <ceph-users@lists.ceph.com> Subject: Re: [ceph-users] CephFS Snapshot questions On Wed, Jun 7, 2017 at 11:46 PM, McFarland, Bruce <bruce.mcfarl...@teradata.com<mailto:bruce.mcfarl...@teradata.com>> wrote: I have a couple of CephFS snapshot questions - Is there any functionality similar to rbd clone/flatten such that the snapshot can be made writable? Or is that as simple as copying the .snap/<dirname> to another cluster? No, there's no cloning. You don't need another cluster though -- you can "cp -r" your snapshot anywhere on any filesystem, and you'll end up with fresh files that you can write to. - If the first object write since the snapid was created is a user error how is that object recovered if it isn’t added to the snapid until it’s 1st write after snapid creation? Don't understand the question at all. "user error"? I think I’ve answered this for myself. The case would be a user’s first write to an object after the snap is created being an error they wanted to “fix” by restoring the object from the objects clone. So when the user writes the “error” to the object it is copied to the snap while it is also being written. The object can then be restored from the clone in this case where the first write to it is in error and it can be recovered from its clone which hadn’t been populated with that object until that write. - If I want to clone the .snap/<dirname>/ and not all objects have been written since .snap/<dirname>/ was created how do I know if or get all objects into the snap if I wanted to move the snap to another cluster? There's no concept of moving a snapshot between clusters. If you're just talking about doing a "cp -r" of the snapshot, then the MDS should do the right thing in terms of blocking your reads on files that have dirty data in client caches -- when we make a snapshot then clients doing buffered writes are asked to flush those buffers. There are 2 cases I’m wondering about here that I didn’t accurately describe. 1 would be data migration between clusters which might not be possible and 2 would be storing clones on a second cluster. 1. Is it possible to snap a directory tree on its source cluster and then copy it to a new/different destination cluster? Would that be prohibited due to the snaps MDS being on the source cluster? I can see that being useful for migrating data/users between clusters, but that it might not be possible. 2. I would expect this to be possible where a snap is created, it’s then compressed into a tarball, and that tarball is stored on a second cluster for any future DR at which point it’s copied back to the source cluster, extracted restoring directory tree to state at time of snap creation. John I might not be making complete sense yet and am in the process of testing to see how CephFS snapshots behave. _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com