I'm going to echo what Stefan said.

I would ditch the 2x SATA drives to free up your slots.
Replace with an M.2 or SATADOM.

I would also recommend moving from the 2x X710-DA2 cards to 1x X710-DA4 card.
It can't saturate the x8 slot, and it frees up a PCIe slot for possibly another 
NVMe card or something else if you need it down the line.

The only other thing I would say to consider is making sure you know the 
endurance of the 4510 is enough for your workload long term.

Reed

> On Mar 10, 2021, at 1:12 PM, Stefan Kooman <ste...@bit.nl> wrote:
> 
> On 3/10/21 5:43 PM, Ignazio Cassano wrote:
>> Hello, what do you think about of ceph cluster made up of 6 nodes each one
>> with the following configuration ?
>> A+ Server 1113S-WN10RT
>> Barebone
>> Supermicro A+ Server 1113S-WN10RT - 1U - 10x U.2 NVMe - 2x M.2 - Dual
>> 10-Gigabit LAN - 750W Redundant
>> Processor
>> AMD EPYC™ 7272 Processor 12-core 2.90GHz 64MB Cache (120W)
>> Memory
>> 8 x 8GB PC4-25600 3200MHz DDR4 ECC RDIMM
> 
> ^^ I would double that amount of RAM, especially (see below) if you plan on 
> adding more NVMe drives.
> 
>> U.2/U.3 NVMe Drive
>> 5 x 8.0TB Intel® SSD DC P4510 Series U.2 PCIe 3.1 x4 NVMe Solid State Drive
>> Hard Drive
> 
> ^^ Why 5 * 8.0 TB instead of 10 * 4.0 TB? Are you planning on upgrading 
> later? Ceph likes more OSDs better than fewer larger ones. Recovery will be 
> faster as well, and the impact of one NVMe dying will be lower.
> 
>> 2 x 240GB Intel® SSD D3-S4610 Series 2.5" SATA 6.0Gb/s Solid State Drive
> 
> ^^ Do you sacrifce two NVMe ports for two SATA OS disks? If so, I would 
> advise for getting (redundant, optional) U.2 NVMe or SATADOM or similar.
> 
>> Network Card
>> 2 x Intel® 10-Gigabit Ethernet Converged Network Adapter X710-DA2 (2x SFP+)
>> Server Management
> 
> ^ Why two? One for "public" and one for "cluster"? Than most probably you 
> won't need that, and one bond would suffice (see current Ceph best 
> practices). If you need 40 Gb/s in one LACP trunk: perfectly fine.
> 
> Gr. Stefan
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to