> 
> ----- Le 14 Mar 25, à 8:40, joachim kraftmayer joachim.kraftma...@clyso.com a 
> écrit :
> 
>> Hi Thomas & Anthony,
>> 
>> Anthony provided great recommendations.

Danke!

>> ssd read performance:
>> I find the total number of pg per ssd osd too low; it can be twice as high.

They’re short on RAM, though, so I hesitate to suggest that.  YMMV.

>> 
>> hdd read performance
>> What makes me a little suspicious is that the maximum throughput of about
>> 120 MB/s is exactly the maximum of a 1 Gbit/s connection.

Modulo framing bits and collisions, but good observation.  I’ve seen a K8s 
deployment where some workers anomalously had a /32 PREFIX in interface 
configurations that sent intra-subnet traffic up to the router for no good 
reason.

>> (I have seen this in the past if the routing is not correct and if you use
>> VMs for testing the network could be limited.)

Now that I’m not typing on my phone … this informs checking the networking on 
all nodes:

Assuming that your links are
bonded for only a public network:

* inspect /proc/net/bonding/bond0 on every host, maybe ethtool on each physical 
interface as well.
* Are both links active?
* Are their link speeds as expected? 
* Use iftop or other tool to see if both links are getting very roughly 
comparable traffic.
* Does netstat -i show a high number of errors?
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to