On Mon, Jan 14, 2019 at 11:09 AM Kenneth Van Alstyne
<kvanalst...@knightpoint.com> wrote:
>
> In this case, I’m imagining Clusters A/B both having write access to a third 
> “Cluster C”.  So A/B -> C rather than A -> C -> B / B -> C -> A / A -> B-> C. 
>  I admit, in the event that I need to replicate back to either primary 
> cluster, there may be challenges.

While this is possible, in addition to the failback question, you
would also need to use unique pool names in clusters A and B since on
cluster C you are currently prevented from adding more than a single
peer per pool.

> Thanks,
>
> --
> Kenneth Van Alstyne
> Systems Architect
> Knight Point Systems, LLC
> Service-Disabled Veteran-Owned Business
> 1775 Wiehle Avenue Suite 101 | Reston, VA 20190
> c: 228-547-8045 f: 571-266-3106
> www.knightpoint.com
> DHS EAGLE II Prime Contractor: FC1 SDVOSB Track
> GSA Schedule 70 SDVOSB: GS-35F-0646S
> GSA MOBIS Schedule: GS-10F-0404Y
> ISO 9001 / ISO 20000 / ISO 27001 / CMMI Level 3
>
> Notice: This e-mail message, including any attachments, is for the sole use 
> of the intended recipient(s) and may contain confidential and privileged 
> information. Any unauthorized review, copy, use, disclosure, or distribution 
> is STRICTLY prohibited. If you are not the intended recipient, please contact 
> the sender by reply e-mail and destroy all copies of the original message.
>
> On Jan 14, 2019, at 9:50 AM, Jason Dillaman <jdill...@redhat.com> wrote:
>
> On Mon, Jan 14, 2019 at 10:10 AM Kenneth Van Alstyne
> <kvanalst...@knightpoint.com> wrote:
>
>
> Thanks for the reply Jason — I was actually thinking of emailing you 
> directly, but thought it may be beneficial to keep the conversation to the 
> list so that everyone can see the thread.   Can you think of a reason why 
> one-way RBD mirroring would not work to a shared tertiary cluster?  I need to 
> build out a test lab to see how that would work for us.
>
>
> I guess I don't understand what the tertiary cluster is doing? If the
> goal is to replicate from cluster A -> cluster B -> cluster C, that is
> not currently supported since (by design choice) we don't currently
> re-write the RBD image journal entries from the source cluster to the
> destination cluster but instead just directly apply the journal
> entries to the destination image (to save IOPS).
>
> Thanks,
>
> --
> Kenneth Van Alstyne
> Systems Architect
> Knight Point Systems, LLC
> Service-Disabled Veteran-Owned Business
> 1775 Wiehle Avenue Suite 101 | Reston, VA 20190
> c: 228-547-8045 f: 571-266-3106
> www.knightpoint.com
> DHS EAGLE II Prime Contractor: FC1 SDVOSB Track
> GSA Schedule 70 SDVOSB: GS-35F-0646S
> GSA MOBIS Schedule: GS-10F-0404Y
> ISO 9001 / ISO 20000 / ISO 27001 / CMMI Level 3
>
> Notice: This e-mail message, including any attachments, is for the sole use 
> of the intended recipient(s) and may contain confidential and privileged 
> information. Any unauthorized review, copy, use, disclosure, or distribution 
> is STRICTLY prohibited. If you are not the intended recipient, please contact 
> the sender by reply e-mail and destroy all copies of the original message.
>
> On Jan 12, 2019, at 4:01 PM, Jason Dillaman <jdill...@redhat.com> wrote:
>
> On Fri, Jan 11, 2019 at 2:09 PM Kenneth Van Alstyne
> <kvanalst...@knightpoint.com> wrote:
>
>
> Hello all (and maybe this would be better suited for the ceph devel mailing 
> list):
> I’d like to use RBD mirroring between two sites (to each other), but I have 
> the following limitations:
> - The clusters use the same name (“ceph”)
>
>
> That's actually not an issue. The "ceph" name is used to locate
> configuration files for RBD mirroring (a la
> /etc/ceph/<cluster-name>.conf and
> /etc/ceph/<cluster-name>.client.<id>.keyring). You just need to map
> that cluster config file name to the remote cluster name in the RBD
> mirroring configuration. Additionally, starting with Nautilus, the
> configuration details for connecting to a remote cluster can now be
> stored in the monitor (via the rbd CLI and dashbaord), so there won't
> be any need to fiddle with configuration files for remote clusters
> anymore.
>
> - The clusters share IP address space on a private, non-routed storage network
>
>
> Unfortunately, that is an issue since the rbd-mirror daemon needs to
> be able to connect to both clusters. If the two clusters are at least
> on different subnets and your management servers can talk to each
> side, you might be able to run the rbd-mirror daemon there.
>
>
> There are management servers on each side that can talk to the respective 
> storage networks, but the storage networks cannot talk directly to each 
> other.  I recall reading, some years back, of possibly adding support for an 
> RBD mirror proxy, which would potentially solve my issues.  Has anything been 
> done in this regard?
>
>
> No, I haven't really seen much demand for such support so it's never
> bubbled up as a priority yet.
>
> If not, is my best bet perhaps a tertiary clusters that both can reach and do 
> one-way replication to?
>
> Thanks,
>
> --
> Kenneth Van Alstyne
> Systems Architect
> Knight Point Systems, LLC
> Service-Disabled Veteran-Owned Business
> 1775 Wiehle Avenue Suite 101 | Reston, VA 20190
> c: 228-547-8045 f: 571-266-3106
> www.knightpoint.com
> DHS EAGLE II Prime Contractor: FC1 SDVOSB Track
> GSA Schedule 70 SDVOSB: GS-35F-0646S
> GSA MOBIS Schedule: GS-10F-0404Y
> ISO 9001 / ISO 20000 / ISO 27001 / CMMI Level 3
>
> Notice: This e-mail message, including any attachments, is for the sole use 
> of the intended recipient(s) and may contain confidential and privileged 
> information. Any unauthorized review, copy, use, disclosure, or distribution 
> is STRICTLY prohibited. If you are not the intended recipient, please contact 
> the sender by reply e-mail and destroy all copies of the original message.
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
>
> --
> Jason
>
>
>
>
> --
> Jason
>
>


-- 
Jason
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to