I kind of doubt this will provide much of an advantage, I think recovery is
the only time you might have some chance of speedup, but i'm not sure
network throughput is always the bottleneck.  There was some discussion a
while back about this, client IO is still going to be impacted by recovery.


On Wed, Sep 25, 2019 at 6:36 AM Adrien Georget <adrien.geor...@cc.in2p3.fr>
wrote:

> Hi,
>
> I need your advice about the following setup.
> Currently, we have a Ceph nautilus cluster used by Openstack Cinder with
> single NIC in 10Gbps on OSD hosts.
> We will upgrade the cluster by adding 7 new hosts dedicated to
> Nova/Glance and we would like to add a cluster network to isolate
> replication and recovery traffic.
> For now, it's not possible to add a second NIC and FC so we are thinking
> about enabling DELL NPAR [1] which allows splitting a single physical
> NIC in 2 logical NICs (1 for public network and 1 for Cluster network).
> We can set max and min bandwidth and implement dynamic bandwidth
> balancing for NPAR to get the appropriate bandwidth when Ceph need it
> (default alloc is 66% for cluster network and 34% for public network).
> Any experiences with this kind of configuration? Do you see any
> disadvantages doing this?
>
> And one question, if we put this in production, adding cluster network
> value in ceph.conf and restarting each OSD is enough for Ceph?
>
> Best,
>
> Adrien
>
> [1]
>
> https://www.dell.com/support/article/fr/fr/frbsdt1/how12596/how-npar-works?lang=en
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
>
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to