ok, thanks. but it seems to me that having pool replicas spread over sites is a bit too risky performance wise. how about ganesha? will it work with cephfs and multi site setup?
I was previously reading about rgw with ganesha and it was full of limitations. with cephfs - there is only one and one I can live with. Will it work? On Mon, May 21, 2018 at 10:57 AM, Adrian Saul <adrian.s...@tpgtelecom.com.au > wrote: > > > We run CephFS in a limited fashion in a stretched cluster of about 40km > with redundant 10G fibre between sites – link latency is in the order of > 1-2ms. Performance is reasonable for our usage but is noticeably slower > than comparable local ceph based RBD shares. > > > > Essentially we just setup the ceph pools behind cephFS to have replicas on > each site. To export it we are simply using Linux kernel NFS and it gets > exported from 4 hosts that act as CephFS clients. Those 4 hosts are then > setup in an DNS record that resolves to all 4 IPs, and we then use > automount to do automatic mounting and host failover on the NFS clients. > Automount takes care of finding the quickest and available NFS server. > > > > I stress this is a limited setup that we use for some fairly light duty, > but we are looking to move things like user home directories onto this. > YMMV. > > > > > > *From:* ceph-users [mailto:ceph-users-boun...@lists.ceph.com] *On Behalf > Of *Up Safe > *Sent:* Monday, 21 May 2018 5:36 PM > *To:* David Turner <drakonst...@gmail.com> > *Cc:* ceph-users <ceph-users@lists.ceph.com> > *Subject:* Re: [ceph-users] multi site with cephfs > > > > Hi, > > can you be a bit more specific? > > I need to understand whether this is doable at all. > > Other options would be using ganesha, but I understand it's very limited > on NFS; > > or start looking at gluster. > > > > Basically, I need the multi site option, i.e. active-active read-write. > > > > Thanks > > > > On Wed, May 16, 2018 at 5:50 PM, David Turner <drakonst...@gmail.com> > wrote: > > Object storage multi-site is very specific to using object storage. It > uses the RGW API's to sync s3 uploads between each site. For CephFS you > might be able to do a sync of the rados pools, but I don't think that's > actually a thing yet. RBD mirror is also a layer on top of things to sync > between sites. Basically I think you need to do something on top of the > Filesystem as opposed to within Ceph to sync it between sites. > > > > On Wed, May 16, 2018 at 9:51 AM Up Safe <upands...@gmail.com> wrote: > > But this is not the question here. > > The question is whether I can configure multi site for CephFS. > > Will I be able to do so by following the guide to set up the multi site > for object storage? > > > > Thanks > > > > On Wed, May 16, 2018, 16:45 John Hearns <hear...@googlemail.com> wrote: > > The answer given at the seminar yesterday was that a practical limit was > around 60km. > > I don't think 100km is that much longer. I defer to the experts here. > > > > > > > > > > > > > > On 16 May 2018 at 15:24, Up Safe <upands...@gmail.com> wrote: > > Hi, > > > > About a 100 km. > > I have a 2-4ms latency between them. > > > > Leon > > > > On Wed, May 16, 2018, 16:13 John Hearns <hear...@googlemail.com> wrote: > > Leon, > > I was at a Lenovo/SuSE seminar yesterday and asked a similar question > regarding separated sites. > > How far apart are these two geographical locations? It does matter. > > > > On 16 May 2018 at 15:07, Up Safe <upands...@gmail.com> wrote: > > Hi, > > I'm trying to build a multi site setup. > > But the only guides I've found on the net were about building it with > object storage or rbd. > > What I need is cephfs. > > I.e. I need to have 2 synced file storages at 2 geographical locations. > > Is this possible? > > Also, if I understand correctly - cephfs is just a component on top of the > object storage. > > Following this logic - it should be possible, right? > > Or am I totally off here? > > Thanks, > > Leon > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > Confidentiality: This email and any attachments are confidential and may > be subject to copyright, legal or some other professional privilege. They > are intended solely for the attention and use of the named addressee(s). > They may only be copied, distributed or disclosed with the consent of the > copyright owner. If you have received this email by mistake or by breach of > the confidentiality clause, please notify the sender immediately by return > email and delete or destroy all copies of the email. Any confidentiality, > privilege or copyright is not waived or lost because this email has been > sent to you by mistake. >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com