What is your expected behavior for when Client A writes to File B in
Datacenter 1 and Client C writes to File B in Datacenter 2 at the exact
same time?

I don't think you can perfectly achieve what you are requesting with Ceph
or many other storage solutions.

On Mon, May 21, 2018 at 9:33 AM, Up Safe <upands...@gmail.com> wrote:

> I'll explain.
> Right now we have 2 sites (racks) with several dozens of servers at each
> accessing a NAS (let's call it a NAS, although it's an IBM v7000 Unified
> that serves the files via NFS).
>
> The biggest problem is that it works active-passive, i.e. we always access
> one of the storages for read/write
> and the other one is replicated once every few hours, so it's more for
> backup needs.
>
> In this setup once the power goes down in our main site - we're stuck with
> a bit (several hours) outdated files
> and we need to remount all of the servers and what not.
>
> The multi site ceph was supposed to solve this problem for us. This way we
> would have only local mounts, i.e.
> each server would only access the filesystem that is in the same site. And
> if one of the sited go down - no pain.
>
> The files are rather small, pdfs and xml of 50-300KB mostly.
> The total size is about 25 TB right now.
>
> We're a low budget company, so your advise about developing is not going
> to happen as we have no such skills or resources for this.
> Plus, I want to make this transparent for the devs and everyone - just an
> infrastructure replacement that will buy me all of the ceph benefits and
> allow the company to survive the power outages or storage crashes.
>
>
>
> On Mon, May 21, 2018 at 5:12 PM, David Turner <drakonst...@gmail.com>
> wrote:
>
>> Not a lot of people use object storage multi-site.  I doubt anyone is
>> using this like you are.  In theory it would work, but even if somebody has
>> this setup running, it's almost impossible to tell if it would work for
>> your needs and use case.  You really should try it out for yourself to see
>> if it works to your needs.  And if you feel so inclined, report back here
>> with how it worked.
>>
>> If you're asking for advice, why do you need a networked posix
>> filesystem?  Unless you are using proprietary software with this
>> requirement, it's generally lazy coding that requires a mounted filesystem
>> like this and you should aim towards using object storage instead without
>> any sort of NFS layer.  It's a little more work for the developers, but is
>> drastically simpler to support and manage.
>>
>> On Mon, May 21, 2018 at 10:06 AM Up Safe <upands...@gmail.com> wrote:
>>
>>> guys,
>>> please tell me if I'm in the right direction.
>>> If ceph object storage can be set up in multi site configuration,
>>> and I add ganesha (which to my understanding is an "adapter"
>>> that serves s3 objects via nfs to clients) -
>>> won't this work as active-active?
>>>
>>>
>>> Thanks
>>>
>>> On Mon, May 21, 2018 at 11:48 AM, Up Safe <upands...@gmail.com> wrote:
>>>
>>>> ok, thanks.
>>>> but it seems to me that having pool replicas spread over sites is a bit
>>>> too risky performance wise.
>>>> how about ganesha? will it work with cephfs and multi site setup?
>>>>
>>>> I was previously reading about rgw with ganesha and it was full of
>>>> limitations.
>>>> with cephfs - there is only one and one I can live with.
>>>>
>>>> Will it work?
>>>>
>>>>
>>>> On Mon, May 21, 2018 at 10:57 AM, Adrian Saul <
>>>> adrian.s...@tpgtelecom.com.au> wrote:
>>>>
>>>>>
>>>>>
>>>>> We run CephFS in a limited fashion in a stretched cluster of about
>>>>> 40km with redundant 10G fibre between sites – link latency is in the order
>>>>> of 1-2ms.  Performance is reasonable for our usage but is noticeably 
>>>>> slower
>>>>> than comparable local ceph based RBD shares.
>>>>>
>>>>>
>>>>>
>>>>> Essentially we just setup the ceph pools behind cephFS to have
>>>>> replicas on each site.  To export it we are simply using Linux kernel NFS
>>>>> and it gets exported from 4 hosts that act as CephFS clients.  Those 4
>>>>> hosts are then setup in an DNS record that resolves to all 4 IPs, and we
>>>>> then use automount to do automatic mounting and host failover on the NFS
>>>>> clients.  Automount takes care of finding the quickest and available NFS
>>>>> server.
>>>>>
>>>>>
>>>>>
>>>>> I stress this is a limited setup that we use for some fairly light
>>>>> duty, but we are looking to move things like user home directories onto
>>>>> this.  YMMV.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> *From:* ceph-users [mailto:ceph-users-boun...@lists.ceph.com] *On
>>>>> Behalf Of *Up Safe
>>>>> *Sent:* Monday, 21 May 2018 5:36 PM
>>>>> *To:* David Turner <drakonst...@gmail.com>
>>>>> *Cc:* ceph-users <ceph-users@lists.ceph.com>
>>>>> *Subject:* Re: [ceph-users] multi site with cephfs
>>>>>
>>>>>
>>>>>
>>>>> Hi,
>>>>>
>>>>> can you be a bit more specific?
>>>>>
>>>>> I need to understand whether this is doable at all.
>>>>>
>>>>> Other options would be using ganesha, but I understand it's very
>>>>> limited on NFS;
>>>>>
>>>>> or start looking at gluster.
>>>>>
>>>>>
>>>>>
>>>>> Basically, I need the multi site option, i.e. active-active read-write.
>>>>>
>>>>>
>>>>>
>>>>> Thanks
>>>>>
>>>>>
>>>>>
>>>>> On Wed, May 16, 2018 at 5:50 PM, David Turner <drakonst...@gmail.com>
>>>>> wrote:
>>>>>
>>>>> Object storage multi-site is very specific to using object storage.
>>>>> It uses the RGW API's to sync s3 uploads between each site.  For CephFS 
>>>>> you
>>>>> might be able to do a sync of the rados pools, but I don't think that's
>>>>> actually a thing yet.  RBD mirror is also a layer on top of things to sync
>>>>> between sites.  Basically I think you need to do something on top of the
>>>>> Filesystem as opposed to within Ceph  to sync it between sites.
>>>>>
>>>>>
>>>>>
>>>>> On Wed, May 16, 2018 at 9:51 AM Up Safe <upands...@gmail.com> wrote:
>>>>>
>>>>> But this is not the question here.
>>>>>
>>>>> The question is whether I can configure multi site for CephFS.
>>>>>
>>>>> Will I be able to do so by following the guide to set up the multi
>>>>> site for object storage?
>>>>>
>>>>>
>>>>>
>>>>> Thanks
>>>>>
>>>>>
>>>>>
>>>>> On Wed, May 16, 2018, 16:45 John Hearns <hear...@googlemail.com>
>>>>> wrote:
>>>>>
>>>>> The answer given at the seminar yesterday was that a practical limit
>>>>> was around 60km.
>>>>>
>>>>> I don't think 100km is that much longer.  I defer to the experts here.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On 16 May 2018 at 15:24, Up Safe <upands...@gmail.com> wrote:
>>>>>
>>>>> Hi,
>>>>>
>>>>>
>>>>>
>>>>> About a 100 km.
>>>>>
>>>>> I have a 2-4ms latency between them.
>>>>>
>>>>>
>>>>>
>>>>> Leon
>>>>>
>>>>>
>>>>>
>>>>> On Wed, May 16, 2018, 16:13 John Hearns <hear...@googlemail.com>
>>>>> wrote:
>>>>>
>>>>> Leon,
>>>>>
>>>>> I was at a Lenovo/SuSE seminar yesterday and asked a similar question
>>>>> regarding separated sites.
>>>>>
>>>>> How far apart are these two geographical locations?   It does matter.
>>>>>
>>>>>
>>>>>
>>>>> On 16 May 2018 at 15:07, Up Safe <upands...@gmail.com> wrote:
>>>>>
>>>>> Hi,
>>>>>
>>>>> I'm trying to build a multi site setup.
>>>>>
>>>>> But the only guides I've found on the net were about building it with
>>>>> object storage or rbd.
>>>>>
>>>>> What I need is cephfs.
>>>>>
>>>>> I.e. I need to have 2 synced file storages at 2 geographical locations.
>>>>>
>>>>> Is this possible?
>>>>>
>>>>> Also, if I understand correctly - cephfs is just a component on top of
>>>>> the object storage.
>>>>>
>>>>> Following this logic - it should be possible, right?
>>>>>
>>>>> Or am I totally off here?
>>>>>
>>>>> Thanks,
>>>>>
>>>>> Leon
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> ceph-users@lists.ceph.com
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> ceph-users@lists.ceph.com
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> ceph-users@lists.ceph.com
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>
>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> ceph-users@lists.ceph.com
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>
>>>>>
>>>>> Confidentiality: This email and any attachments are confidential and
>>>>> may be subject to copyright, legal or some other professional privilege.
>>>>> They are intended solely for the attention and use of the named
>>>>> addressee(s). They may only be copied, distributed or disclosed with the
>>>>> consent of the copyright owner. If you have received this email by mistake
>>>>> or by breach of the confidentiality clause, please notify the sender
>>>>> immediately by return email and delete or destroy all copies of the email.
>>>>> Any confidentiality, privilege or copyright is not waived or lost because
>>>>> this email has been sent to you by mistake.
>>>>>
>>>>
>>>>
>>>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to