On 05/20/2013 08:01 AM, Bjorn Mork wrote:
Hi Team,
This is my first post to this community.
I have some basic queries to start with CEPH software...I found that
http://www.supermicro.com.tw/products/system/2U/6027/SSG-6027R-E1R12T.cfm is
being recommend as a start of storage server.
This is a reasonable server for a basic Ceph POC using spinning disks
with no SSD journals. Since it's using on-board ethernet and RAID, it
should be relatively inexpensive, but if any of the on-board components
fail the whole motherboard has to be replaced. It's a good starting
point though.
As my target is to start with 12 TB solution (production environment,
high performance) having three copies of my data. I am confused, that
1. How many servers will be required i.e OSD, MON, MDS (above
mentioned chassis).
For production you should have at least 3 MONs. You only need an MDS if
you plan to use CephFS. We tend to recommend 1 OSD per disk for most
configurations.
2. Should I separate role to each server? or single server will be
good enough?
You want each MON on a different server, and for a production deployment
I really don't like seeing less than 5 servers for OSDs. You can
technically run a single mon and all of your OSDs on 1 server, but it's
not really what Ceph was designed for.
3. How many raid-cards in each server will be required?
3.1 I mean separate for read and write can be configured or not? I
need best performance and throughput.
There's a lot of different ways you can configure ceph servers with
various trade-offs. A general rule of thumb is that you want at least
3-5 servers for OSDs (and preferably more), and for high performance SSD
journals or at the very least a controller with WB cache and 1 OSD per disk.
You may be interested in some of our performance comparison tests:
http://ceph.com/community/ceph-performance-part-1-disk-controller-write-throughput/
http://ceph.com/community/ceph-performance-part-2-write-throughput-without-ssd-journals/
http://ceph.com/uncategorized/argonaut-vs-bobtail-performance-preview/
Mark
Can anyone suggest? Thanks in advance...
B~Mork
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com