On Tue, Mar 12, 2019 at 11:09 PM Vikas Rana <vr...@vtiersys.com> wrote: > > Hi there, > > > > We are replicating a RBD image from Primary to DR site using RBD mirroring. > > On Primary, we were using 10.2.10.
Just a note that Jewel is end-of-life upstream. > DR site is luminous and we promoted the DR copy to test the failure. > Everything checked out good. > > > > Now we are trying to restart the replication and we did the demote and then > resync the image but it stuck in “starting_replay” state for last 3 days. > It’s a 200TB RBD image You would need to run "rbd --cluster <primary-site> mirror image resync nfs/dir_research" and wait for that to complete *before* demoting the primary image on cluster "cephdr". Without a primary image, there is nothing to resync against. > > > :~# rbd --cluster cephdr mirror pool status nfs --verbose > > health: WARNING > > images: 1 total > > 1 starting_replay > > > > dir_research: > > global_id: 3ad67d0c-e06b-406a-9469-4e5faedd09a4 > > state: down+unknown > > description: status not found > > last_update: > > > > > > #rbd info nfs/dir_research > > rbd image 'dir_research': > > size 200TiB in 52428800 objects > > order 22 (4MiB objects) > > block_name_prefix: rbd_data.652186b8b4567 > > format: 2 > > features: layering, exclusive-lock, journaling > > flags: > > create_timestamp: Thu Feb 7 11:53:36 2019 > > journal: 652186b8b4567 > > mirroring state: disabling > > mirroring global id: 3ad67d0c-e06b-406a-9469-4e5faedd09a4 > > mirroring primary: false > > > So the question is, how do we know the progress of the replay and how much > its already completed and any ETA estimation on when it will go back to OK > state? > > > > > Thanks, > > -Vikas > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Jason _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com