On 05/20/2013 08:01 AM, Bjorn Mork wrote:
Hi Team,

This is my first post to this community.

I have some basic queries to start with CEPH software...I found that
http://www.supermicro.com.tw/products/system/2U/6027/SSG-6027R-E1R12T.cfm is
being recommend as a start of storage server.

This is a reasonable server for a basic Ceph POC using spinning disks with no SSD journals. Since it's using on-board ethernet and RAID, it should be relatively inexpensive, but if any of the on-board components fail the whole motherboard has to be replaced. It's a good starting point though.


As my target is to start with 12 TB solution (production environment,
high performance) having three copies of my data. I am confused, that

1.   How many servers will be required i.e OSD, MON, MDS (above
mentioned chassis).

For production you should have at least 3 MONs. You only need an MDS if you plan to use CephFS. We tend to recommend 1 OSD per disk for most configurations.


2.   Should I separate role to each server? or single server will be
good enough?

You want each MON on a different server, and for a production deployment I really don't like seeing less than 5 servers for OSDs. You can technically run a single mon and all of your OSDs on 1 server, but it's not really what Ceph was designed for.


3.     How many raid-cards in each server will be required?
3.1   I mean separate for read and write can be configured or not? I
need best performance and throughput.

There's a lot of different ways you can configure ceph servers with various trade-offs. A general rule of thumb is that you want at least 3-5 servers for OSDs (and preferably more), and for high performance SSD journals or at the very least a controller with WB cache and 1 OSD per disk.

You may be interested in some of our performance comparison tests:

http://ceph.com/community/ceph-performance-part-1-disk-controller-write-throughput/
http://ceph.com/community/ceph-performance-part-2-write-throughput-without-ssd-journals/
http://ceph.com/uncategorized/argonaut-vs-bobtail-performance-preview/

Mark


Can anyone suggest? Thanks in advance...

B~Mork


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to