Hmm change 40Gbps to 100Gbps networking.

40Gbps technology its just a bond of 4x10 Links with some latency due link 
aggregation.
100 Gbps and 25Gbps got less latency and Good performance. In ceph a 50% of the 
latency comes from Network commits and the other 50% from disk commits.

A fast graph : 
https://blog.mellanox.com/wp-content/uploads/John-Kim-030416-Fig-3a-1024x747.jpg
Article: 
https://blog.mellanox.com/2016/03/25-is-the-new-10-50-is-the-new-40-100-is-the-new-amazing/

Micron got their own Whitepaper for CEPH and looks like performs fine.
https://www.micron.com/-/media/client/global/documents/products/other-documents/micron_9200_max_ceph_12,-d-,2,-d-,8_luminous_bluestore_reference_architecture.pdf?la=en


AS your Budget is high, please buy 3 x 1.5K $ nodes for your monitors and you 
Will sleep better. They just need 4 cores / 16GB RAM and 2x128GB SSD or NVME M2 
.

-----Mensaje original-----
De: Adam Boyhan <[email protected]> 
Enviado el: viernes, 31 de enero de 2020 13:59
Para: ceph-users <[email protected]>
Asunto: [ceph-users] Micron SSD/Basic Config

Looking to role out a all flash Ceph cluster. Wanted to see if anyone else was 
using Micron drives along with some basic input on my design so far? 

Basic Config
Ceph OSD Nodes
8x Supermicro A+ Server 2113S-WTRT
- AMD EPYC 7601 32 Core 2.2Ghz
- 256G Ram
- AOC-S3008L-L8e HBA
- 10GB SFP+ for client network
- 40GB QSFP+ for ceph cluster network 

OSD
10x Micron 5300 PRO 7.68TB in each ceph node
- 80 total drives across the 8 nodes 

WAL/DB
5x Micron 7300 MAX NVMe 800GB per Ceph Node
- Plan on dedicating 1 for each 2 OSD's 

Still thinking out a external monitor node as I have a lot of options, but this 
is a pretty good start. Open to suggestions as well! 

_______________________________________________
ceph-users mailing list -- [email protected] To unsubscribe send an email to 
[email protected]
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to