Thanks, I'll look into these.
On Thu, Dec 18, 2014 at 5:12 PM, Craig Lewis <cle...@centraldesktop.com> wrote: > > I think this is it: > https://engage.redhat.com/inktank-ceph-reference-architecture-s-201409080939 > > You can also check out a presentation on Cern's Ceph cluster: > http://www.slideshare.net/Inktank_Ceph/scaling-ceph-at-cern > > > At large scale, the biggest problem will likely be network I/O on the > inter-switch links. > > > > On Thu, Dec 18, 2014 at 3:29 PM, Robert LeBlanc <rob...@leblancnet.us> > wrote: >> >> I'm interested to know if there is a reference to this reference >> architecture. It would help alleviate some of the fears we have about >> scaling this thing to a massive scale (10,000's OSDs). >> >> Thanks, >> Robert LeBlanc >> >> On Thu, Dec 18, 2014 at 3:43 PM, Craig Lewis <cle...@centraldesktop.com> >> wrote: >> >>> >>> >>> On Thu, Dec 18, 2014 at 5:16 AM, Patrick McGarry <patr...@inktank.com> >>> wrote: >>>> >>>> >>>> > 2. What should be the minimum hardware requirement of the server (CPU, >>>> > Memory, NIC etc) >>>> >>>> There is no real "minimum" to run Ceph, it's all about what your >>>> workload will look like and what kind of performance you need. We have >>>> seen Ceph run on Raspberry Pis. >>> >>> >>> Technically, the smallest cluster is a single node with a 10 GiB disk. >>> Anything smaller won't work. >>> >>> That said, Ceph was envisioned to run on large clusters. IIRC, the >>> reference architecture has 7 rows, each row having 10 racks, all full. >>> >>> Those of us running small clusters (less than 10 nodes) are noticing >>> that it doesn't work quite as well. We have to significantly scale back >>> the amount of backfilling and recovery that is allowed. I try to keep all >>> backfill/recovery operations touching less than 20% of my OSDs. In the >>> reference architecture, it could lose a whole row, and still keep under >>> that limit. My 5 nodes cluster is noticeably better better than the 3 node >>> cluster. It's faster, has lower latency, and latency doesn't increase as >>> much during recovery operations. >>> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >>>
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com