Hi,

I am building my first Ceph cluster from hardware leftover from a previous
project. I have been reading a lot of Ceph documentation but need some help
to make sure I going the right way.
To set the stage below is what I have

Rack-1

1 x HP DL360 G9 with
   - 256 GB Memory
   - 5 x 300GB HDD
   - 2 x HBA SAS
   - 4 x 10GBe Networking Card

1 x SuperMicro chassis with 17 x HP Enterprise 400GB SSD and 17 x HP
Enterprise 1.7TB HDD
Chassis and HP server are connected with 2 x SAS HBA for redundancy.


Rack-2 (Same as Rack-1)

1 x HP DL360 G9 with
   - 256 GB Memory
   - 5 x 300GB HDD
   - 2 x HBA SAS
   - 4 x 10GBe Networking Card

1 x SuperMicro chassis with 17 x HP Enterprise 400GB SSD and 17 x HP
Enterprise 1.7TB HDD
Chassis and HP server are connected with 2 x SAS HBA for redundancy.


Rack-3

5 x HP DL360 G8 with
   - 128 GB Memory
   - 2 x 400GB HP Enterprise SSD
   - 3 x 1.7TB Enterprise HDD

Requirements
- To serve storage to around 200 VMware VMs via iSCSI. VMs use disks
moderately.
- To serve storage to some docker containers using ceph volume driver
- To serve storage to some legacy apps using NFS

Plan

- Create a ceph cluster with all machines
- Use Bluestore as osd backing ( 3 x SSD for DB and WAL in SuperMicro
Chassis and 1 x SSD for DB and WAL in Rack 3 G8s)
- Use remaining SSDs ( 14 x in SuperMicro and 1 x Rack 3 G8s ) for Rados
Cache Tier
- Update CRUSH map to make Rack as minimum failure domain. So almost all
data is replicated across racks and in case one of the host dies the
storage still works.
- Single bonded network (4x10GBe) connected to ToR switches.
- Same public and cluster network

Questions

- First of all, is this kind of setup workable.
- I have seen that Ceph uses /dev/sdx names in guides, is it a good
approach considering the disks die and can come up with different /dev/sdx
identifier on reboot.
- What should be the approx size of WAL and DB partitions for my kind of
setup?
- Can i install ceph in a VM and use other VMs on these hosts. Is Ceph too
CPU demanding?

Thanks,
Shantur
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to