On Wed, Jun 5, 2019 at 12:59 PM Jordan Share <readm...@krotus.com> wrote: > > One thing to keep in mind when pipelining rbd export/import is that the > default is just a raw image dump. > > So if you have a large, but not very full, RBD, you will dump all those > zeroes into the pipeline. > > In our case, it was actually faster to write to a (sparse) temp file and > read it in again afterwards than to pipeline. > > However, we are not using --export-format 2, which I now suspect would > mitigate this.
It's supposed to help since it's only using diffs -- never the full image export. > Jordan > > > On 6/5/2019 8:30 AM, CUZA Frédéric wrote: > > Hi, > > > > Thank you all for you quick answer. > > I think that will solve our problem. > > > > This is what we came up with this : > > rbd -c /etc/ceph/Oceph.conf --keyring /etc/ceph/Oceph.client.admin.keyring > > export rbd/disk_test - | rbd -c /etc/ceph/Nceph.conf --keyring > > /etc/ceph/Nceph.client.admin.keyring import - rbd/disk_test > > > > This rbd image is a test with only 5Gb of datas inside of it. > > > > Unfortunately the command seems to be stuck and nothing happens, both ports > > 7800 / 6789 / 22. > > > > We can't find no logs on any monitors. > > > > Thanks ! > > > > -----Message d'origine----- > > De : ceph-users <ceph-users-boun...@lists.ceph.com> De la part de Jason > > Dillaman > > Envoyé : 04 June 2019 14:11 > > À : Burkhard Linke <burkhard.li...@computational.bio.uni-giessen.de> > > Cc : ceph-users <ceph-users@lists.ceph.com> > > Objet : Re: [ceph-users] Multiple rbd images from different clusters > > > > On Tue, Jun 4, 2019 at 8:07 AM Jason Dillaman <jdill...@redhat.com> wrote: > >> > >> On Tue, Jun 4, 2019 at 4:45 AM Burkhard Linke > >> <burkhard.li...@computational.bio.uni-giessen.de> wrote: > >>> > >>> Hi, > >>> > >>> On 6/4/19 10:12 AM, CUZA Frédéric wrote: > >>> > >>> Hi everyone, > >>> > >>> > >>> > >>> We want to migrate datas from one cluster (Hammer) to a new one (Mimic). > >>> We do not wish to upgrade the actual cluster as all the hardware is EOS > >>> and we upgrade the configuration of the servers. > >>> > >>> We can’t find a “proper” way to mount two rbd images from two different > >>> cluster on the same host. > >>> > >>> Does anyone know what is the “good” procedure to achieve this ? > >> > >> Copy your "/etc/ceph/ceph.conf" and associated keyrings for both > >> clusters to a single machine (preferably running a Mimic "rbd" client) > >> under "/etc/ceph/<cluster-name>.conf" and > >> "/etc/ceph/<cluster-name>.client.<id>.keyring". > >> > >> You can then use "rbd -c <old cluster name> export --export-format 2 > >> <image-spec> - | rbd -c <new cluster name> import --export-format=2 - > >> <image-spec>". The "--export-format=2" option will also copy all > >> associated snapshots with the images. If you don't want/need the > >> snapshots, just drop that optional. > > > > That "-c" should be "--cluster" if specifying by name, otherwise with "-c" > > it's the full path to the two different conf files. > > > >>> > >>> Just my 2 ct: > >>> > >>> the 'rbd' commands allows specifying a configuration file (-c). You need > >>> to setup two configuration files, one for each cluster. You can also use > >>> two different cluster names (--cluster option). AFAIK the name is only > >>> used to locate the configuration file. I'm not sure how well the kernel > >>> works with mapping RBDs from two different cluster..... > >>> > >>> > >>> If you only want to transfer RBDs from one cluster to another, you do not > >>> need to map and mount them; the 'rbd' command has the sub commands > >>> 'export' and 'import'. You can pipe them to avoid writing data to a local > >>> disk. This should be the fastest way to transfer the RBDs. > >>> > >>> > >>> Regards, > >>> > >>> Burkhard > >>> > >>> -- > >>> Dr. rer. nat. Burkhard Linke > >>> Bioinformatics and Systems Biology > >>> Justus-Liebig-University Giessen > >>> 35392 Giessen, Germany > >>> Phone: (+49) (0)641 9935810 > >>> > >>> _______________________________________________ > >>> ceph-users mailing list > >>> ceph-users@lists.ceph.com > >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >> > >> > >> > >> -- > >> Jason > > > > > > > > -- > > Jason > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Jason _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com