for example, to have one cluster with four instances, two on each data center sharing the same storage cluster, so if data center #1 let's said goes down along with instances 1 & 2, you can still run the cluster with instances 3 & 4 on the data center #2 with the same storage cluster pool and no data loss or impact (just performance)
Best, *German* 2016-10-21 15:01 GMT-03:00 Wes Dillingham <wes_dilling...@harvard.edu>: > What is the use case that requires you to have it in two datacenters? > In addition to RBD mirroring already mentioned by others, you can do > RBD snapshots and ship those snapshots to a remote location (separate > cluster or separate pool). Similar to RBD mirroring, in this situation > your client writes are not subject to that latency. > > On Thu, Oct 20, 2016 at 1:51 PM, German Anders <gand...@despegar.com> > wrote: > > Thanks, that's too far actually lol. And how things going with rbd > > mirroring? > > > > German > > > > 2016-10-20 14:49 GMT-03:00 yan cui <ccuiy...@gmail.com>: > >> > >> The two data centers are actually cross US. One is in the west, and the > >> other in the east. > >> We try to sync rdb images using RDB mirroring. > >> > >> 2016-10-20 9:54 GMT-07:00 German Anders <gand...@despegar.com>: > >>> > >>> from curiosity I wanted to ask you what kind of network topology are > you > >>> trying to use across the cluster? In this type of scenario you really > need > >>> an ultra low latency network, how far from each other? > >>> > >>> Best, > >>> > >>> German > >>> > >>> 2016-10-18 16:22 GMT-03:00 Sean Redmond <sean.redmo...@gmail.com>: > >>>> > >>>> Maybe this would be an option for you: > >>>> > >>>> http://docs.ceph.com/docs/jewel/rbd/rbd-mirroring/ > >>>> > >>>> > >>>> On Tue, Oct 18, 2016 at 8:18 PM, yan cui <ccuiy...@gmail.com> wrote: > >>>>> > >>>>> Hi Guys, > >>>>> > >>>>> Our company has a use case which needs the support of Ceph across > >>>>> two data centers (one data center is far away from the other). The > >>>>> experience of using one data center is good. We did some > benchmarking on two > >>>>> data centers, and the performance is bad because of the > synchronization > >>>>> feature in Ceph and large latency between data centers. So, are there > >>>>> setting ups like data center aware features in Ceph, so that we have > good > >>>>> locality? Usually, we use rbd to create volume and snapshot. But we > want the > >>>>> volume is high available with acceptable performance in case one > data center > >>>>> is down. Our current setting ups does not consider data center > difference. > >>>>> Any ideas? > >>>>> > >>>>> > >>>>> Thanks, Yan > >>>>> > >>>>> -- > >>>>> Think big; Dream impossible; Make it happen. > >>>>> > >>>>> _______________________________________________ > >>>>> ceph-users mailing list > >>>>> ceph-users@lists.ceph.com > >>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >>>>> > >>>> > >>>> > >>>> _______________________________________________ > >>>> ceph-users mailing list > >>>> ceph-users@lists.ceph.com > >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >>>> > >>> > >> > >> > >> > >> -- > >> Think big; Dream impossible; Make it happen. > > > > > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > -- > Respectfully, > > Wes Dillingham > wes_dilling...@harvard.edu > Research Computing | Infrastructure Engineer > Harvard University | 38 Oxford Street, Cambridge, Ma 02138 | Room 210 >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com