Active-passive sounds not what I want. But maybe I misunderstand. Does rbd mirror replicate both ways? And how do I do it with nfs?
Thanks On Mon, May 21, 2018, 17:42 Paul Emmerich <paul.emmer...@croit.io> wrote: > For active/passive and async replication with a POSIX filesystem: > Maybe two Ceph clusters with RBD mirror and re-exporting the RBD(s) via > NFS? > > > Paul > > 2018-05-21 16:33 GMT+02:00 Up Safe <upands...@gmail.com>: > >> I'll explain. >> Right now we have 2 sites (racks) with several dozens of servers at each >> accessing a NAS (let's call it a NAS, although it's an IBM v7000 Unified >> that serves the files via NFS). >> >> The biggest problem is that it works active-passive, i.e. we always >> access one of the storages for read/write >> and the other one is replicated once every few hours, so it's more for >> backup needs. >> >> In this setup once the power goes down in our main site - we're stuck >> with a bit (several hours) outdated files >> and we need to remount all of the servers and what not. >> >> The multi site ceph was supposed to solve this problem for us. This way >> we would have only local mounts, i.e. >> each server would only access the filesystem that is in the same site. >> And if one of the sited go down - no pain. >> >> The files are rather small, pdfs and xml of 50-300KB mostly. >> The total size is about 25 TB right now. >> >> We're a low budget company, so your advise about developing is not going >> to happen as we have no such skills or resources for this. >> Plus, I want to make this transparent for the devs and everyone - just an >> infrastructure replacement that will buy me all of the ceph benefits and >> allow the company to survive the power outages or storage crashes. >> >> >> >> On Mon, May 21, 2018 at 5:12 PM, David Turner <drakonst...@gmail.com> >> wrote: >> >>> Not a lot of people use object storage multi-site. I doubt anyone is >>> using this like you are. In theory it would work, but even if somebody has >>> this setup running, it's almost impossible to tell if it would work for >>> your needs and use case. You really should try it out for yourself to see >>> if it works to your needs. And if you feel so inclined, report back here >>> with how it worked. >>> >>> If you're asking for advice, why do you need a networked posix >>> filesystem? Unless you are using proprietary software with this >>> requirement, it's generally lazy coding that requires a mounted filesystem >>> like this and you should aim towards using object storage instead without >>> any sort of NFS layer. It's a little more work for the developers, but is >>> drastically simpler to support and manage. >>> >>> On Mon, May 21, 2018 at 10:06 AM Up Safe <upands...@gmail.com> wrote: >>> >>>> guys, >>>> please tell me if I'm in the right direction. >>>> If ceph object storage can be set up in multi site configuration, >>>> and I add ganesha (which to my understanding is an "adapter" >>>> that serves s3 objects via nfs to clients) - >>>> won't this work as active-active? >>>> >>>> >>>> Thanks >>>> >>>> On Mon, May 21, 2018 at 11:48 AM, Up Safe <upands...@gmail.com> wrote: >>>> >>>>> ok, thanks. >>>>> but it seems to me that having pool replicas spread over sites is a >>>>> bit too risky performance wise. >>>>> how about ganesha? will it work with cephfs and multi site setup? >>>>> >>>>> I was previously reading about rgw with ganesha and it was full of >>>>> limitations. >>>>> with cephfs - there is only one and one I can live with. >>>>> >>>>> Will it work? >>>>> >>>>> >>>>> On Mon, May 21, 2018 at 10:57 AM, Adrian Saul < >>>>> adrian.s...@tpgtelecom.com.au> wrote: >>>>> >>>>>> >>>>>> >>>>>> We run CephFS in a limited fashion in a stretched cluster of about >>>>>> 40km with redundant 10G fibre between sites – link latency is in the >>>>>> order >>>>>> of 1-2ms. Performance is reasonable for our usage but is noticeably >>>>>> slower >>>>>> than comparable local ceph based RBD shares. >>>>>> >>>>>> >>>>>> >>>>>> Essentially we just setup the ceph pools behind cephFS to have >>>>>> replicas on each site. To export it we are simply using Linux kernel NFS >>>>>> and it gets exported from 4 hosts that act as CephFS clients. Those 4 >>>>>> hosts are then setup in an DNS record that resolves to all 4 IPs, and we >>>>>> then use automount to do automatic mounting and host failover on the NFS >>>>>> clients. Automount takes care of finding the quickest and available NFS >>>>>> server. >>>>>> >>>>>> >>>>>> >>>>>> I stress this is a limited setup that we use for some fairly light >>>>>> duty, but we are looking to move things like user home directories onto >>>>>> this. YMMV. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> *From:* ceph-users [mailto:ceph-users-boun...@lists.ceph.com] *On >>>>>> Behalf Of *Up Safe >>>>>> *Sent:* Monday, 21 May 2018 5:36 PM >>>>>> *To:* David Turner <drakonst...@gmail.com> >>>>>> *Cc:* ceph-users <ceph-users@lists.ceph.com> >>>>>> *Subject:* Re: [ceph-users] multi site with cephfs >>>>>> >>>>>> >>>>>> >>>>>> Hi, >>>>>> >>>>>> can you be a bit more specific? >>>>>> >>>>>> I need to understand whether this is doable at all. >>>>>> >>>>>> Other options would be using ganesha, but I understand it's very >>>>>> limited on NFS; >>>>>> >>>>>> or start looking at gluster. >>>>>> >>>>>> >>>>>> >>>>>> Basically, I need the multi site option, i.e. active-active >>>>>> read-write. >>>>>> >>>>>> >>>>>> >>>>>> Thanks >>>>>> >>>>>> >>>>>> >>>>>> On Wed, May 16, 2018 at 5:50 PM, David Turner <drakonst...@gmail.com> >>>>>> wrote: >>>>>> >>>>>> Object storage multi-site is very specific to using object storage. >>>>>> It uses the RGW API's to sync s3 uploads between each site. For CephFS >>>>>> you >>>>>> might be able to do a sync of the rados pools, but I don't think that's >>>>>> actually a thing yet. RBD mirror is also a layer on top of things to >>>>>> sync >>>>>> between sites. Basically I think you need to do something on top of the >>>>>> Filesystem as opposed to within Ceph to sync it between sites. >>>>>> >>>>>> >>>>>> >>>>>> On Wed, May 16, 2018 at 9:51 AM Up Safe <upands...@gmail.com> wrote: >>>>>> >>>>>> But this is not the question here. >>>>>> >>>>>> The question is whether I can configure multi site for CephFS. >>>>>> >>>>>> Will I be able to do so by following the guide to set up the multi >>>>>> site for object storage? >>>>>> >>>>>> >>>>>> >>>>>> Thanks >>>>>> >>>>>> >>>>>> >>>>>> On Wed, May 16, 2018, 16:45 John Hearns <hear...@googlemail.com> >>>>>> wrote: >>>>>> >>>>>> The answer given at the seminar yesterday was that a practical limit >>>>>> was around 60km. >>>>>> >>>>>> I don't think 100km is that much longer. I defer to the experts here. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On 16 May 2018 at 15:24, Up Safe <upands...@gmail.com> wrote: >>>>>> >>>>>> Hi, >>>>>> >>>>>> >>>>>> >>>>>> About a 100 km. >>>>>> >>>>>> I have a 2-4ms latency between them. >>>>>> >>>>>> >>>>>> >>>>>> Leon >>>>>> >>>>>> >>>>>> >>>>>> On Wed, May 16, 2018, 16:13 John Hearns <hear...@googlemail.com> >>>>>> wrote: >>>>>> >>>>>> Leon, >>>>>> >>>>>> I was at a Lenovo/SuSE seminar yesterday and asked a similar question >>>>>> regarding separated sites. >>>>>> >>>>>> How far apart are these two geographical locations? It does matter. >>>>>> >>>>>> >>>>>> >>>>>> On 16 May 2018 at 15:07, Up Safe <upands...@gmail.com> wrote: >>>>>> >>>>>> Hi, >>>>>> >>>>>> I'm trying to build a multi site setup. >>>>>> >>>>>> But the only guides I've found on the net were about building it with >>>>>> object storage or rbd. >>>>>> >>>>>> What I need is cephfs. >>>>>> >>>>>> I.e. I need to have 2 synced file storages at 2 geographical >>>>>> locations. >>>>>> >>>>>> Is this possible? >>>>>> >>>>>> Also, if I understand correctly - cephfs is just a component on top >>>>>> of the object storage. >>>>>> >>>>>> Following this logic - it should be possible, right? >>>>>> >>>>>> Or am I totally off here? >>>>>> >>>>>> Thanks, >>>>>> >>>>>> Leon >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> ceph-users mailing list >>>>>> ceph-users@lists.ceph.com >>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> ceph-users mailing list >>>>>> ceph-users@lists.ceph.com >>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> ceph-users mailing list >>>>>> ceph-users@lists.ceph.com >>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>>> >>>>>> _______________________________________________ >>>>>> ceph-users mailing list >>>>>> ceph-users@lists.ceph.com >>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>>> >>>>>> >>>>>> Confidentiality: This email and any attachments are confidential and >>>>>> may be subject to copyright, legal or some other professional privilege. >>>>>> They are intended solely for the attention and use of the named >>>>>> addressee(s). They may only be copied, distributed or disclosed with the >>>>>> consent of the copyright owner. If you have received this email by >>>>>> mistake >>>>>> or by breach of the confidentiality clause, please notify the sender >>>>>> immediately by return email and delete or destroy all copies of the >>>>>> email. >>>>>> Any confidentiality, privilege or copyright is not waived or lost because >>>>>> this email has been sent to you by mistake. >>>>>> >>>>> >>>>> >>>> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> > > > -- > -- > Paul Emmerich > > Looking for help with your Ceph cluster? Contact us at https://croit.io > > croit GmbH > Freseniusstr. 31h > 81247 München > www.croit.io > Tel: +49 89 1896585 90 >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com