Yes, Ceph is probably a good fit for what you are planning.

The documentation should answer your questions:
http://docs.ceph.com/docs/master/
Look for erasure coding, crush rules, and CephFS-specific pages in
particular.



Paul


2018-07-10 18:40 GMT+02:00 Jones de Andrade <johanne...@gmail.com>:

> Hi all.
>
> I'm looking for some information on several distributed filesystems for
> our application.
>
> It looks like it finally came down to two candidates, Ceph being one of
> them. But there are still a few questions about ir that I would really like
> to clarify, if possible.
>
> Our plan, initially on 6 workstations, is to have it hosting a distributed
> file system that can withstand two simultaneous computers failures without
> data loss (something that can remember a raid 6, but over the network).
> This file system will also need to be also remotely mounted (NFS server
> with fallbacks) by other 5+ computers. Students will be working on all 11+
> computers at the same time (different requisites from different softwares:
> some use many small files, other a few really big, 100s gb, files), and
> absolutely no hardware modifications are allowed. This initial test bed is
> for undergraduate students usage, but if successful will be employed also
> for our small clusters. The connection is a simple GbE.
>
> Our actual concerns are:
> 1) Data Resilience: It seems that double copy of each block is the
> standard setting, is it correct? As such, it will strip-parity data among
> three computers for each block?
>
> 2) Metadata Resilience: We seen that we can now have more than a single
> Metadata Server (which was a show-stopper on previous versions). However,
> do they have to be dedicated boxes, or they can share boxes with the Data
> Servers? Can it be configured in such a way that even if two metadata
> server computers fail the whole system data will still be accessible from
> the remaining computers, without interruptions, or they share different
> data aiming only for performance?
>
> 3) Other softwares compability: We seen that there is NFS incompability,
> is it correct? Also, any posix issues?
>
> 4) No single (or double) point of failure: every single possible stance
> has to be able to endure a *double* failure (yes, things can get time to be
> fixed here). Does Ceph need s single master server for any of its
> activities? Can it endure double failure? How long would it take to any
> sort of "fallback" to be completed, users would need to wait to regain
> access?
>
> I think that covers the initial questions we have. Sorry if this is the
> wrong list, however.
>
> Looking forward for any answer or suggestion,
>
> Regards,
>
> Jones
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>


-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to