On Tue, Aug 8, 2023 at 9:16 AM Adiga, Anantha <anantha.ad...@intel.com> wrote: > > Hi Venky, > > Is this correct? > (copied ceph.conf from secondary cluster to /etc/ce/ph/crsite directory in > primary cluster, copied ceph.mon.keyring from secondary as > ceph.client.crsite.mon.keyring in /etc/ceph on primary) > root@fl31ca104ja0201:/etc/ceph# ls > ceph.client.admin.keyring ceph.client.crsite.admin.keyring > ceph.client.mirror_remote.keying crsite fio-fs.test fs-mnt rbdmap > ceph.client.crash.keyring ceph.client.crsite.mon.keyring ceph.conf > fio-bsd.test fio-nfs.test nfs-mnt remote_ceph.conf > root@fl31ca104ja0201:/etc/ceph# ls crsite > ceph.conf ceph.mon.keyring > > root@fl31ca104ja0201:/etc/ceph/crsite# ceph -c ceph.conf --id=crsite.mon > --cluster=ceph --verbose > parsed_args: Namespace(admin_socket=None, block=False, cephconf='ceph.conf', > client_id='crsite.mon', client_name=None, cluster='ceph', > cluster_timeout=None, completion=False, help=False, input_file=None, > output_file=None, output_format=None, period=1, setgroup=None, setuser=None, > status=False, verbose=True, version=False, watch=False, watch_channel=None, > watch_debug=False, watch_error=False, watch_info=False, watch_sec=False, > watch_warn=False), childargs: [] > ^CCluster connection aborted > > Not sure if the --id (CLIENT_ID) is correct.. not able to connect
use `remote_ceph.conf` and id as `mirror_remote` (since I guess these are the secondary clusters' conf given the names). > > Thank you, > Anantha > > -----Original Message----- > From: Venky Shankar <vshan...@redhat.com> > Sent: Monday, August 7, 2023 7:05 PM > To: Adiga, Anantha <anantha.ad...@intel.com> > Cc: ceph-users@ceph.io > Subject: Re: [ceph-users] Re: cephfs snapshot mirror peer_bootstrap import > hung > > Hi Anantha, > > On Tue, Aug 8, 2023 at 6:29 AM Adiga, Anantha <anantha.ad...@intel.com> wrote: > > > > Hi Venky, > > > > The primary and secondary clusters both have the same cluster name "ceph" > > and both have a single filesystem by name "cephfs". > > That's not an issue. > > > How do I check the connection from primary to secondary using mon addr and > > key? What is command line > > A quick way to check this would be to place the secondary cluster ceph config > file and the user key on one of the primary node (preferably, the ceph-mgr > host, just for tests - so purge these when done) and then running > > ceph -c /path/to/secondary/ceph.conf --id <> status > > If that runs all fine, then the mirror daemon is probably hitting some bug. > > > These two clusters are configured for rgw multisite and is functional. > > > > Thank you, > > Anantha > > > > -----Original Message----- > > From: Venky Shankar <vshan...@redhat.com> > > Sent: Monday, August 7, 2023 5:46 PM > > To: Adiga, Anantha <anantha.ad...@intel.com> > > Cc: ceph-users@ceph.io > > Subject: Re: [ceph-users] Re: cephfs snapshot mirror peer_bootstrap > > import hung > > > > Hi Anantha, > > > > On Mon, Aug 7, 2023 at 11:52 PM Adiga, Anantha <anantha.ad...@intel.com> > > wrote: > > > > > > Hi Venky, > > > > > > > > > > > > I tried on another secondary Quincy cluster and it is the same problem. > > > The peer_bootstrap mport command hangs. > > > > A pacific cluster generated peer token should be importable in a quincy > > source cluster. Looking at the logs, I suspect that the perceived hang is > > the mirroring module blocked on connecting to the secondary cluster (to set > > mirror info xattr). Are you able to connect to the secondary cluster from > > the host running ceph-mgr on the primary cluster using its monitor address > > (and a key)? > > > > The primary and secondary clusters both have the same cluster name "ceph" > > and both have a single filesystem by name "cephfs". How do I check that > > connection from primary to secondary using mon addr and key? > > These two clusters are configured for rgw multisite and is functional. > > > > > > > > > > > > > > > > > > > > root@fl31ca104ja0201:/# ceph fs snapshot mirror peer_bootstrap > > > import cephfs > > > eyJmc2lkIjogIjJlYWMwZWEwLTYwNDgtNDQ0Zi04NGIyLThjZWVmZWQyN2E1YiIsICJm > > > aW > > > xlc3lzdGVtIjogImNlcGhmcyIsICJ1c2VyIjogImNsaWVudC5taXJyb3JfcmVtb3RlIi > > > wg > > > InNpdGVfbmFtZSI6ICJzaGdSLXNpdGUiLCAia2V5IjogIkFRQ0lGdEZrSStTTE5oQUFX > > > bW > > > V6MkRKcEg5ZUdyYnhBOWVmZG9BPT0iLCAibW9uX2hvc3QiOiAiW3YyOjEwLjIzOS4xNT > > > Uu > > > MTg6MzMwMC8wLHYxOjEwLjIzOS4xNTUuMTg6Njc4OS8wXSBbdjI6MTAuMjM5LjE1NS4x > > > OT > > > ozMzAwLzAsdjE6MTAuMjM5LjE1NS4xOTo2Nzg5LzBdIFt2MjoxMC4yMzkuMTU1LjIwOj > > > Mz MDAvMCx2MToxMC4yMzkuMTU1LjIwOjY3ODkvMF0ifQ== > > > > > > …… > > > > > > ……. > > > > > > ..command does not complete..waits here > > > > > > ^C to exit. > > > > > > Thereafter some commands do not complete… > > > > > > root@fl31ca104ja0201:/# ceph -s > > > > > > cluster: > > > > > > id: d0a3b6e0-d2c3-11ed-be05-a7a3a1d7a87e > > > > > > health: HEALTH_OK > > > > > > > > > > > > services: > > > > > > mon: 3 daemons, quorum > > > fl31ca104ja0202,fl31ca104ja0203,fl31ca104ja0201 (age 2d) > > > > > > mgr: fl31ca104ja0201.kkoono(active, since 3d), standbys: > > > fl31ca104ja0202, fl31ca104ja0203 > > > > > > mds: 1/1 daemons up, 2 standby > > > > > > osd: 44 osds: 44 up (since 2d), 44 in (since 5w) > > > > > > cephfs-mirror: 1 daemon active (1 hosts) > > > > > > rgw: 3 daemons active (3 hosts, 1 zones) > > > > > > > > > > > > data: > > > > > > volumes: 1/1 healthy > > > > > > pools: 25 pools, 769 pgs > > > > > > objects: 614.40k objects, 1.9 TiB > > > > > > usage: 2.9 TiB used, 292 TiB / 295 TiB avail > > > > > > pgs: 769 active+clean > > > > > > > > > > > > io: > > > > > > client: 32 KiB/s rd, 0 B/s wr, 33 op/s rd, 1 op/s wr > > > > > > > > > > > > root@fl31ca104ja0201:/# > > > > > > root@fl31ca104ja0201:/# ceph fs status cephfs > > > > > > This command also waits. …… > > > > > > > > > > > > I have attached the mgr log > > > > > > root@fl31ca104ja0201:/# ceph service status > > > > > > { > > > > > > "cephfs-mirror": { > > > > > > "5306346": { > > > > > > "status_stamp": "2023-08-07T17:35:56.884907+0000", > > > > > > "last_beacon": "2023-08-07T17:45:01.903540+0000", > > > > > > "status": { > > > > > > "status_json": > > > "{\"1\":{\"name\":\"cephfs\",\"directory_count\":0,\"peers\":{}}}" > > > > > > } > > > > > > } > > > > > > > > > > > > Quincy secondary cluster > > > > > > > > > > > > root@a001s008-zz14l47008:/# ceph mgr module enable mirroring > > > > > > root@a001s008-zz14l47008:/# ceph fs authorize cephfs > > > client.mirror_remote / rwps > > > > > > [client.mirror_remote] > > > > > > key = AQCIFtFkI+SLNhAAWmez2DJpH9eGrbxA9efdoA== > > > > > > root@a001s008-zz14l47008:/# ceph auth get client.mirror_remote > > > > > > [client.mirror_remote] > > > > > > key = AQCIFtFkI+SLNhAAWmez2DJpH9eGrbxA9efdoA== > > > > > > caps mds = "allow rwps fsname=cephfs" > > > > > > caps mon = "allow r fsname=cephfs" > > > > > > caps osd = "allow rw tag cephfs data=cephfs" > > > > > > root@a001s008-zz14l47008:/# > > > > > > root@a001s008-zz14l47008:/# ceph fs snapshot mirror peer_bootstrap > > > create cephfs client.mirror_remote shgR-site > > > > > > {"token": > > > "eyJmc2lkIjogIjJlYWMwZWEwLTYwNDgtNDQ0Zi04NGIyLThjZWVmZWQyN2E1YiIsICJ > > > ma > > > Wxlc3lzdGVtIjogImNlcGhmcyIsICJ1c2VyIjogImNsaWVudC5taXJyb3JfcmVtb3RlI > > > iw > > > gInNpdGVfbmFtZSI6ICJzaGdSLXNpdGUiLCAia2V5IjogIkFRQ0lGdEZrSStTTE5oQUF > > > Xb > > > WV6MkRKcEg5ZUdyYnhBOWVmZG9BPT0iLCAibW9uX2hvc3QiOiAiW3YyOjEwLjIzOS4xN > > > TU > > > uMTg6MzMwMC8wLHYxOjEwLjIzOS4xNTUuMTg6Njc4OS8wXSBbdjI6MTAuMjM5LjE1NS4 > > > xO > > > TozMzAwLzAsdjE6MTAuMjM5LjE1NS4xOTo2Nzg5LzBdIFt2MjoxMC4yMzkuMTU1LjIwO > > > jM zMDAvMCx2MToxMC4yMzkuMTU1LjIwOjY3ODkvMF0ifQ=="} > > > > > > root@a001s008-zz14l47008:/# > > > > > > > > > > > > Thank you, > > > > > > Anantha > > > > > > > > > > > > From: Adiga, Anantha > > > Sent: Friday, August 4, 2023 11:55 AM > > > To: Venky Shankar <vshan...@redhat.com>; ceph-users@ceph.io > > > Subject: RE: [ceph-users] Re: cephfs snapshot mirror peer_bootstrap > > > import hung > > > > > > > > > > > > Hi Venky, > > > > > > > > > > > > Thank you so much for the guidance. Attached is the mgr log. > > > > > > > > > > > > Note: the 4th node in the primary cluster has smaller capacity drives, > > > the other 3 nodes have the larger capacity drives. > > > > > > 32 ssd 6.98630 1.00000 7.0 TiB 44 GiB 44 GiB 183 KiB 148 > > > MiB 6.9 TiB 0.62 0.64 40 up osd.32 > > > > > > -7 76.84927 - 77 TiB 652 GiB 648 GiB 20 MiB 3.0 > > > GiB 76 TiB 0.83 0.86 - host fl31ca104ja0203 > > > > > > 1 ssd 6.98630 1.00000 7.0 TiB 73 GiB 73 GiB 8.0 MiB 333 > > > MiB 6.9 TiB 1.02 1.06 54 up osd.1 > > > > > > 4 ssd 6.98630 1.00000 7.0 TiB 77 GiB 77 GiB 1.1 MiB 174 > > > MiB 6.9 TiB 1.07 1.11 55 up osd.4 > > > > > > 7 ssd 6.98630 1.00000 7.0 TiB 47 GiB 47 GiB 140 KiB 288 > > > MiB 6.9 TiB 0.66 0.68 51 up osd.7 > > > > > > 10 ssd 6.98630 1.00000 7.0 TiB 75 GiB 75 GiB 299 KiB 278 > > > MiB 6.9 TiB 1.05 1.09 44 up osd.10 > > > > > > 13 ssd 6.98630 1.00000 7.0 TiB 94 GiB 94 GiB 1018 KiB 291 > > > MiB 6.9 TiB 1.31 1.36 72 up osd.13 > > > > > > 16 ssd 6.98630 1.00000 7.0 TiB 31 GiB 31 GiB 163 KiB 267 > > > MiB 7.0 TiB 0.43 0.45 49 up osd.16 > > > > > > 19 ssd 6.98630 1.00000 7.0 TiB 14 GiB 14 GiB 756 KiB 333 > > > MiB 7.0 TiB 0.20 0.21 50 up osd.19 > > > > > > 22 ssd 6.98630 1.00000 7.0 TiB 105 GiB 104 GiB 1.3 MiB 313 > > > MiB 6.9 TiB 1.46 1.51 48 up osd.22 > > > > > > 25 ssd 6.98630 1.00000 7.0 TiB 17 GiB 16 GiB 257 KiB 272 > > > MiB 7.0 TiB 0.23 0.24 45 up osd.25 > > > > > > 28 ssd 6.98630 1.00000 7.0 TiB 72 GiB 72 GiB 6.1 MiB 180 > > > MiB 6.9 TiB 1.01 1.05 43 up osd.28 > > > > > > 31 ssd 6.98630 1.00000 7.0 TiB 47 GiB 46 GiB 592 KiB 358 > > > MiB 6.9 TiB 0.65 0.68 56 up osd.31 > > > > > > -9 64.04089 - 64 TiB 728 GiB 726 GiB 17 MiB 1.8 > > > GiB 63 TiB 1.11 1.15 - host fl31ca104ja0302 > > > > > > 33 ssd 5.82190 1.00000 5.8 TiB 65 GiB 65 GiB 245 KiB 144 > > > MiB 5.8 TiB 1.09 1.13 47 up osd.33 > > > > > > 34 ssd 5.82190 1.00000 5.8 TiB 14 GiB 14 GiB 815 KiB 83 > > > MiB 5.8 TiB 0.24 0.25 55 up osd.34 > > > > > > 35 ssd 5.82190 1.00000 5.8 TiB 77 GiB 77 GiB 224 KiB 213 > > > MiB 5.7 TiB 1.30 1.34 44 up osd.35 > > > > > > 36 ssd 5.82190 1.00000 5.8 TiB 117 GiB 117 GiB 8.5 MiB 284 > > > MiB 5.7 TiB 1.96 2.03 52 up osd.36 > > > > > > 37 ssd 5.82190 1.00000 5.8 TiB 58 GiB 58 GiB 501 KiB 132 > > > MiB 5.8 TiB 0.98 1.01 40 up osd.37 > > > > > > 38 ssd 5.82190 1.00000 5.8 TiB 123 GiB 123 GiB 691 KiB 266 > > > MiB 5.7 TiB 2.07 2.14 73 up osd.38 > > > > > > 39 ssd 5.82190 1.00000 5.8 TiB 77 GiB 77 GiB 609 KiB 193 > > > MiB 5.7 TiB 1.30 1.34 62 up osd.39 > > > > > > 40 ssd 5.82190 1.00000 5.8 TiB 77 GiB 77 GiB 262 KiB 148 > > > MiB 5.7 TiB 1.29 1.34 55 up osd.40 > > > > > > 41 ssd 5.82190 1.00000 5.8 TiB 44 GiB 44 GiB 4.4 MiB 140 > > > MiB 5.8 TiB 0.75 0.77 44 up osd.41 > > > > > > 42 ssd 5.82190 1.00000 5.8 TiB 45 GiB 45 GiB 886 KiB 135 > > > MiB 5.8 TiB 0.75 0.78 47 up osd.42 > > > > > > 43 ssd 5.82190 1.00000 5.8 TiB 28 GiB 28 GiB 187 KiB 104 > > > MiB 5.8 TiB 0.48 0.49 58 up osd.43 > > > > > > > > > > > > [Also: Yesterday I had two cfs-mirror running one on fl31ca104ja0201 > > > and fl31ca104ja0302. The cfs-mirror on fl31ca104ja0201 was stopped. > > > When the import token was run on fl31ca104ja0302, the cfs-mirror > > > log was active. Just in case it is useful I have attached that log > > > (cfsmirror-container.log) as well. ] > > > > > > > > > > > > How can I list the token on the target cluster after running the create > > > peer_bootstrap command? > > > > > > > > > > > > Here is today’s status with your suggestion: > > > > > > There is only one cfs-mirror daemon running now. It is on fl31ca104ja0201 > > > node. > > > > > > > > > > > > root@fl31ca104ja0201:/# ceph -s > > > > > > cluster: > > > > > > id: d0a3b6e0-d2c3-11ed-be05-a7a3a1d7a87e > > > > > > health: HEALTH_OK > > > > > > > > > > > > services: > > > > > > mon: 3 daemons, quorum > > > fl31ca104ja0202,fl31ca104ja0203,fl31ca104ja0201 (age 7m) > > > > > > mgr: fl31ca104ja0201.kkoono(active, since 13m), standbys: > > > fl31ca104ja0202, fl31ca104ja0203 > > > > > > mds: 1/1 daemons up, 2 standby > > > > > > osd: 44 osds: 44 up (since 7m), 44 in (since 4w) > > > > > > cephfs-mirror: 1 daemon active (1 hosts) > > > > > > rgw: 3 daemons active (3 hosts, 1 zones) > > > > > > > > > > > > data: > > > > > > volumes: 1/1 healthy > > > > > > pools: 25 pools, 769 pgs > > > > > > objects: 614.40k objects, 1.9 TiB > > > > > > usage: 2.8 TiB used, 292 TiB / 295 TiB avail > > > > > > pgs: 769 active+clean > > > > > > > > > > > > io: > > > > > > client: 32 MiB/s rd, 0 B/s wr, 57 op/s rd, 1 op/s wr > > > > > > > > > > > > root@fl31ca104ja0201:/# > > > > > > root@fl31ca104ja0201:/# > > > > > > root@fl31ca104ja0201:/# ceph tell mgr.fl31ca104ja0201.kkoono config > > > set debug_mgr 20 > > > > > > { > > > > > > "success": "" > > > > > > } > > > > > > root@fl31ca104ja0201:/# ceph fs snapshot mirror peer_bootstrap > > > import cephfs > > > eyJmc2lkIjogImE2ZjUyNTk4LWU1Y2QtNGEwOC04NDIyLTdiNmZkYjFkNWRiZSIsICJm > > > aW > > > xlc3lzdGVtIjogImNlcGhmcyIsICJ1c2VyIjogImNsaWVudC5taXJyb3JfcmVtb3RlIi > > > wg > > > InNpdGVfbmFtZSI6ICJmbGV4Mi1zaXRlIiwgImtleSI6ICJBUUNmd01sa005MHBMQkFB > > > d1 > > > h0dnBwOGowNEl2Qzh0cXBBRzliQT09IiwgIm1vbl9ob3N0IjogIlt2MjoxNzIuMTguNT > > > Uu > > > NzE6MzMwMC8wLHYxOjE3Mi4xOC41NS43MTo2Nzg5LzBdIFt2MjoxNzIuMTguNTUuNzM6 > > > Mz MwMC8wLHYxOjE3Mi4xOC41NS43Mzo2Nzg5LzBdIn0= > > > > > > ^CInterrupted > > > > > > > > > > > > Ctrl-C after 15 min. Once the command is run, the health status goes to > > > WARN . > > > > > > > > > > > > root@fl31ca104ja0201:/# ceph -s > > > > > > cluster: > > > > > > id: d0a3b6e0-d2c3-11ed-be05-a7a3a1d7a87e > > > > > > health: HEALTH_WARN > > > > > > 6 slow ops, oldest one blocked for 1095 sec, > > > mon.fl31ca104ja0203 has slow ops > > > > > > > > > > > > services: > > > > > > mon: 3 daemons, quorum > > > fl31ca104ja0202,fl31ca104ja0203,fl31ca104ja0201 (age 30m) > > > > > > mgr: fl31ca104ja0201.kkoono(active, since 35m), standbys: > > > fl31ca104ja0202, fl31ca104ja0203 > > > > > > mds: 1/1 daemons up, 2 standby > > > > > > osd: 44 osds: 44 up (since 29m), 44 in (since 4w) > > > > > > cephfs-mirror: 1 daemon active (1 hosts) > > > > > > rgw: 3 daemons active (3 hosts, 1 zones) > > > > > > > > > > > > data: > > > > > > volumes: 1/1 healthy > > > > > > pools: 25 pools, 769 pgs > > > > > > objects: 614.40k objects, 1.9 TiB > > > > > > usage: 2.8 TiB used, 292 TiB / 295 TiB avail > > > > > > pgs: 769 active+clean > > > > > > > > > > > > io: > > > > > > client: 67 KiB/s rd, 0 B/s wr, 68 op/s rd, 21 op/s wr > > > > > > > > > > > > > > > > > > -----Original Message----- > > > From: Venky Shankar <vshan...@redhat.com> > > > Sent: Thursday, August 3, 2023 11:03 PM > > > To: Adiga, Anantha <anantha.ad...@intel.com> > > > Cc: ceph-users@ceph.io > > > Subject: [ceph-users] Re: cephfs snapshot mirror peer_bootstrap > > > import hung > > > > > > > > > > > > Hi Anantha, > > > > > > > > > > > > On Fri, Aug 4, 2023 at 2:27 AM Adiga, Anantha <anantha.ad...@intel.com> > > > wrote: > > > > > > > > > > > > > > Hi > > > > > > > > > > > > > > Could you please provide guidance on how to diagnose this issue: > > > > > > > > > > > > > > In this case, there are two Ceph clusters: cluster A, 4 nodes and > > > > cluster B, 3 node, in different locations. Both are already running > > > > RGW multi-site, A is master. > > > > > > > > > > > > > > Cephfs snapshot mirroring is being configured on the clusters. Cluster > > > > A is the primary, cluster B is the peer. Cephfs snapshot mirroring is > > > > being configured. The bootstrap import step on the primary node hangs. > > > > > > > > > > > > > > On the target cluster : > > > > > > > --------------------------- > > > > > > > "version": "16.2.5", > > > > > > > "release": "pacific", > > > > > > > "release_type": "stable" > > > > > > > > > > > > > > root@cr21meg16ba0101:/# ceph fs snapshot mirror peer_bootstrap > > > > create > > > > > > > cephfs client.mirror_remote flex2-site > > > > > > > {"token": > > > > > > > "eyJmc2lkIjogImE2ZjUyNTk4LWU1Y2QtNGEwOC04NDIyLTdiNmZkYjFkNWRiZSIsI > > > > CJ > > > > ma > > > > > > > Wxlc3lzdGVtIjogImNlcGhmcyIsICJ1c2VyIjogImNsaWVudC5taXJyb3JfcmVtb3R > > > > lI > > > > iw > > > > > > > gInNpdGVfbmFtZSI6ICJmbGV4Mi1zaXRlIiwgImtleSI6ICJBUUNmd01sa005MHBMQ > > > > kF > > > > Bd > > > > > > > 1h0dnBwOGowNEl2Qzh0cXBBRzliQT09IiwgIm1vbl9ob3N0IjogIlt2MjoxNzIuMTg > > > > uN > > > > TU > > > > > > > uNzE6MzMwMC8wLHYxOjE3Mi4xOC41NS43MTo2Nzg5LzBdIFt2MjoxNzIuMTguNTUuN > > > > zM > > > > 6M > > > > > > > zMwMC8wLHYxOjE3Mi4xOC41NS43Mzo2Nzg5LzBdIn0="} > > > > > > > > > > > > Seems fine uptil here. > > > > > > > > > > > > > root@cr21meg16ba0101:/var/run/ceph# > > > > > > > > > > > > > > On the source cluster: > > > > > > > ---------------------------- > > > > > > > "version": "17.2.6", > > > > > > > "release": "quincy", > > > > > > > "release_type": "stable" > > > > > > > > > > > > > > root@fl31ca104ja0201:/# ceph -s > > > > > > > cluster: > > > > > > > id: d0a3b6e0-d2c3-11ed-be05-a7a3a1d7a87e > > > > > > > health: HEALTH_OK > > > > > > > > > > > > > > services: > > > > > > > mon: 3 daemons, quorum > > > > fl31ca104ja0202,fl31ca104ja0203,fl31ca104ja0201 (age 111m) > > > > > > > mgr: fl31ca104ja0201.nwpqlh(active, since 11h), standbys: > > > > fl31ca104ja0203, fl31ca104ja0202 > > > > > > > mds: 1/1 daemons up, 2 standby > > > > > > > osd: 44 osds: 44 up (since 111m), 44 in (since 4w) > > > > > > > cephfs-mirror: 1 daemon active (1 hosts) > > > > > > > rgw: 3 daemons active (3 hosts, 1 zones) > > > > > > > > > > > > > > data: > > > > > > > volumes: 1/1 healthy > > > > > > > pools: 25 pools, 769 pgs > > > > > > > objects: 614.40k objects, 1.9 TiB > > > > > > > usage: 2.8 TiB used, 292 TiB / 295 TiB avail > > > > > > > pgs: 769 active+clean > > > > > > > > > > > > > > root@fl31ca104ja0302:/# ceph mgr module enable mirroring module > > > > > > > 'mirroring' is already enabled root@fl31ca104ja0302:/# ceph fs > > > > > > > snapshot mirror peer_bootstrap import cephfs > > > > > > > eyJmc2lkIjogImE2ZjUyNTk4LWU1Y2QtNGEwOC04NDIyLTdiNmZkYjFkNWRiZSIsIC > > > > Jm > > > > aW > > > > > > > xlc3lzdGVtIjogImNlcGhmcyIsICJ1c2VyIjogImNsaWVudC5taXJyb3JfcmVtb3Rl > > > > Ii > > > > wg > > > > > > > InNpdGVfbmFtZSI6ICJmbGV4Mi1zaXRlIiwgImtleSI6ICJBUUNmd01sa005MHBMQk > > > > FB > > > > d1 > > > > > > > h0dnBwOGowNEl2Qzh0cXBBRzliQT09IiwgIm1vbl9ob3N0IjogIlt2MjoxNzIuMTgu > > > > NT > > > > Uu > > > > > > > NzE6MzMwMC8wLHYxOjE3Mi4xOC41NS43MTo2Nzg5LzBdIFt2MjoxNzIuMTguNTUuNz > > > > M6 > > > > Mz > > > > > > > MwMC8wLHYxOjE3Mi4xOC41NS43Mzo2Nzg5LzBdIn0= > > > > > > > > > > > > Going by your description, I'm guessing this is the command that > > > hangs? If that's the case, set `debug_mgr=20`, repeat the token > > > import step and share the ceph-mgr log. Also note that you can check > > > the mirror daemon status as detailed in > > > > > > > > > > > > > > > https://docs.ceph.com/en/latest/dev/cephfs-mirroring/#mirror-daemon- > > > st > > > atus > > > > > > > > > > > > > > > > > > > > > > > > > > > root@fl31ca104ja0302:/var/run/ceph# ceph --admin-daemon > > > > > > > /var/run/ceph/ceph-client.cephfs-mirror.fl31ca104ja0302.sypagt.7.9 > > > > 40 > > > > 83135960976.asok status { > > > > > > > "metadata": { > > > > > > > "ceph_sha1": "d7ff0d10654d2280e08f1ab989c7cdf3064446a5", > > > > > > > "ceph_version": "ceph version 17.2.6 > > > > (d7ff0d10654d2280e08f1ab989c7cdf3064446a5) quincy (stable)", > > > > > > > "entity_id": "cephfs-mirror.fl31ca104ja0302.sypagt", > > > > > > > "hostname": "fl31ca104ja0302", > > > > > > > "pid": "7", > > > > > > > "root": "/" > > > > > > > }, > > > > > > > "dentry_count": 0, > > > > > > > "dentry_pinned_count": 0, > > > > > > > "id": 5194553, > > > > > > > "inst": { > > > > > > > "name": { > > > > > > > "type": "client", > > > > > > > "num": 5194553 > > > > > > > }, > > > > > > > "addr": { > > > > > > > "type": "v1", > > > > > > > "addr": "10.45.129.5:0", > > > > > > > "nonce": 2497002034 > > > > > > > } > > > > > > > }, > > > > > > > "addr": { > > > > > > > "type": "v1", > > > > > > > "addr": "10.45.129.5:0", > > > > > > > "nonce": 2497002034 > > > > > > > }, > > > > > > > "inst_str": "client.5194553 10.45.129.5:0/2497002034", > > > > > > > "addr_str": "10.45.129.5:0/2497002034", > > > > > > > "inode_count": 1, > > > > > > > "mds_epoch": 118, > > > > > > > "osd_epoch": 6266, > > > > > > > "osd_epoch_barrier": 0, > > > > > > > "blocklisted": false, > > > > > > > "fs_name": "cephfs" > > > > > > > } > > > > > > > > > > > > > > root@fl31ca104ja0302:/home/general# docker logs > > > > > > > ceph-d0a3b6e0-d2c3-11ed-be05-a7a3a1d7a87e-cephfs-mirror-fl31ca104j > > > > a0 > > > > 30 > > > > > > > 2-sypagt --tail 10 debug 2023-08-03T05:24:27.413+0000 > > > > 7f8eb6fc0280 > > > > 0 > > > > > > > ceph version 17.2.6 (d7ff0d10654d2280e08f1ab989c7cdf3064446a5) > > > > quincy > > > > > > > (stable), process cephfs-mirror, pid 7 debug > > > > > > > 2023-08-03T05:24:27.413+0000 7f8eb6fc0280 0 pidfile_write: ignore > > > > > > > empty --pid-file debug 2023-08-03T05:24:27.445+0000 7f8eb6fc0280 > > > > 1 > > > > > > > mgrc service_daemon_register cephfs-mirror.5184622 metadata > > > > > > > {arch=x86_64,ceph_release=quincy,ceph_version=ceph version 17.2.6 > > > > > > > (d7ff0d10654d2280e08f1ab989c7cdf3064446a5) quincy > > > > > > > (stable),ceph_version_short=17.2.6,container_hostname=fl31ca104ja0 > > > > 30 > > > > 2, > > > > > > > container_image=quay.io/ceph/ceph@sha256:af79fedafc42237b7612fe2d1 > > > > 8a > > > > 9c > > > > > > > 64ca62a0b38ab362e614ad671efa4a0547e,cpu=Intel(R) Xeon(R) Gold 6252 > > > > CPU > > > > > > > @ 2.10GHz,distro=centos,distro_description=CentOS Stream > > > > > > > 8,distro_version=8,hostname=fl31ca104ja0302,id=fl31ca104ja0302.syp > > > > ag > > > > t, > > > > > > > instance_id=5184622,kernel_description=#82-Ubuntu SMP Tue Jun 6 > > > > > > > 23:10:23 UTC > > > > > > > 2023,kernel_version=5.15.0-75-generic,mem_swap_kb=8388604,mem_tota > > > > l_ > > > > kb > > > > > > > =527946928,os=Linux} debug 2023-08-03T05:27:10.419+0000 > > > > 7f8ea1b2c700 > > > > > > > 0 client.5194553 ms_handle_reset on v2:10.45.128.141:3300/0 debug > > > > > > > 2023-08-03T05:50:10.917+0000 7f8ea1b2c700 0 client.5194553 > > > > > > > ms_handle_reset on v2:10.45.128.139:3300/0 > > > > > > > > > > > > > > Thank you, > > > > > > > Anantha > > > > > > > _______________________________________________ > > > > > > > ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send > > > > an > > > > > > > email to ceph-users-le...@ceph.io > > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > > > Cheers, > > > > > > Venky > > > > > > _______________________________________________ > > > > > > ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an > > > email to ceph-users-le...@ceph.io > > > > > > > > -- > > Cheers, > > Venky > > > > > -- > Cheers, > Venky > -- Cheers, Venky _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io