This bug is missing log files that will aid in diagnosing the problem.
While running an Ubuntu kernel (not a mainline or third-party kernel)
please enter the following command in a terminal window:
apport-collect 1960826
and then change the status of the bug to 'Confirmed'.
If, due to the nature of the issue you have encountered, you are unable
to run this command, please add a comment stating that fact and change
the bug status to 'Confirmed'.
This change has been made by an automated script, maintained by the
Ubuntu Kernel Team.
** Changed in: linux (Ubuntu)
Status: New => Incomplete
--
You received this bug notification because you are a member of Kernel
Packages, which is subscribed to linux in Ubuntu.
https://bugs.launchpad.net/bugs/1960826
Title:
NFSv4 performance problem with newer kernels
Status in linux package in Ubuntu:
Incomplete
Bug description:
I am seeing an issue with NFSv4 performance on Ubuntu 20.04.3 LTS on both
clients and server (tested kernels 5.4 to 5.16), where the server is connected
by 10Gbit Ethernet and with multiple clients connected by 1Gbit. I am reading a
large file from an NFS mount via "dd" sending output to /dev/null with bs=4096.
Using default sysctl and mount options I am seeing speeds maxing out below
1Gbit/sec. If I force NFSv3 I see speeds close to 10Gbit/sec with sufficient
clients connected. I also see no issue with Ubuntu 16.04 (when used for both
server and clients) in conjunction with NFSv4. I have attached the output from
two iftop's which shows the status when using NFSv4 and when using NFSv3, in
the NFSv4 you can clearly see one client reading at max speed and all the
others apparently throttling back to practically nothing.
I have additionally tested a range of mount options, which made no
difference, BBR congestion control which made no difference and the following
kernel settings which also made no difference:
net.core.netdev_max_backlog=250000
net.core.rmem_max=4194304
net.core.wmem_max=4194304
net.core.rmem_default=4194304
net.core.wmem_default=4194304
net.core.optmem_max=4194304
net.ipv4.tcp_rmem=4096 87380 4194304
net.ipv4.tcp_wmem=4096 65536 4194304
net.ipv4.tcp_mem=786432 1048576 26777216
net.ipv4.udp_mem=1529892 2039859 3059784
net.ipv4.udp_rmem_min=16384
net.ipv4.udp_wmem_min=16384
The problem is seen on dissimilar hardware, ie this problem exists
when testing with an HP DL380 G10 with Mellanox 10Gbit Ethernet
connected to Cisco switch, and also on a Dell R430 with Broadcom
10Gbit Ethernet connected to a Netgear switch (just to name two of
several configurations that have been tested). The clients vary in
each test case also, but are desktop PCs and laptops.
To manage notifications about this bug go to:
https://bugs.launchpad.net/ubuntu/+source/linux/+bug/1960826/+subscriptions
--
Mailing list: https://launchpad.net/~kernel-packages
Post to : [email protected]
Unsubscribe : https://launchpad.net/~kernel-packages
More help : https://help.launchpad.net/ListHelp