Hi, We collected some data comparing the USB throughput we're getting now on NetBSD against the throughput on Linux. For those who are interested in the current performance on NetBSD, I've included a summary. The full set of measurements taken (along with the summary) is available at: http://acert.ir.bbn.com/viewvc/adroitgrdevel/adroitgrdevel/radio_test/usb/test-results?view=co
Summary ======= The following USB throughput results were collected on two systems with the same hardware, running NetBSD-current with our ugen changes and SuSE Linux. The ugen changes allow specifying the length of the transfer to request from the host controller, and here the fusb_netbsd testing code was recompiled with the different sizes. The fusb_linux code uses 16k requests (and says that this is the largest request possible). In both cases the USRP library's default buffer size of 2 MB was used. The ugen driver could also be changed to avoid a copy to the buffer in the driver, and these tests investigate how much performance is improved in that case. For reference, here is how interpolation/decimation relates to the intended data rate: data rate | decimation | interpolation -------------------------------------- 16 MB/s 16 32 18.3 MB/s 14 28 21.3 MB/s 12 24 25.6 MB/s 10 20 32 MB/s 8 16 42.6 MB/s 6 12 benchmark_usb.py (bidirectional test) driver | xfer size | maximum (read+write) ---------------------------------- NetBSD 16k 32 MB/s Linux 16k 36.57 MB/s NetBSD 64k 32 MB/s (usually gets 36.57) NetBSD 128k 32 MB/s NetBSD -copy 16k 32 MB/s NetBSD -copy 64k 42.6 MB/s NetBSD -copy 128k 42.6 MB/s test_standard_usrp_rx driver | xfer size | maximum ---------------------------------- NetBSD 16k 21.3 Linux 16k 32 NetBSD 64k 25.6 NetBSD 128k 21.3 NetBSD -copy 16k 25.6 NetBSD -copy 64k 25.6 NetBSD -copy 128k 25.6 test_standard_usrp_tx driver | xfer size | maximum ---------------------------------- NetBSD 16k 21.3 Linux 16k 32 NetBSD 64k 25.6 NetBSD 128k 21.3 NetBSD -copy 16k 21.3 NetBSD -copy 64k 25.6 NetBSD -copy 128k 25.6 The Linux numbers suggest that there is about 36 MB/s bandwidth available total (maybe more but less than 42), and it must be divided between transmit and receive. So 32 can be done one-way, but as soon as one needs bidirectional traffic, neither direction can do 32. Probably the USRP could be set up to use, say, 25.6 and 8 between read and write instead of 16 and 16, but not 25.6 and 16. This follows fairly well from the implementation. On Linux, USRP reads and writes are all done via a generic request mechanism funneled through the control endpoint. So the sum of reads and writes in aggregate seems to be constrained by how fast data can be pushed through this system. With our NetBSD implementation, unless the transactions go in lock-step and thus one of read and write has to wait while the other's completion interrupt is being handled, read and write are handled independently all the way down until you get to the host controller driver. Therefore the bidirectional numbers are more related to the sum of the two unidirectional numbers, instead of bidirectional being essentially equal to unidirectional as we're seeing with Linux. The NetBSD numbers demonstrate that 128k transfers perform worse than 64k. As would be expected, 128k transfers aren't worse with the extra copy removed but they also aren't notably better. So while there is clearly too much cost copying 128k at a time vs. copying 64k, there is still a lot of cost that's not in the copy at all, because the numbers don't get vastly better when the copy is removed. The latter cost is what's preventing us from getting unidirectional rates comparable to Linux. Copying to/from user space is not showing to be the bottleneck; the kernel debug logs clearly show that user space consumes and writes faster than the bus in these tests. Choosing a Good Buffer Size =========================== The previous results are all using a buffer size of 2 MB (which is 2 MB for each of read and write with fusb_netbsd). Also, all reads and writes from user space were 16k. The following tests indicated the read and write length does not matter very much. However, reducing the buffer size from 2 MB demonstrably helps with the bidirectional throughput. Because the highest rate reached is not always the same, these results include several runs of benchmark_usb.py. The maximum rate is based on what benchmark_usb.py claimed for five runs, trying to take into account that all the higher transfer rates report underruns or overruns occasionally. driver | xfer | buffer | maximum | size | size | rate -------------------------------------- NetBSD 16k 2M 32 NetBSD 64k 2M 32 NetBSD 128k 2M 32 NetBSD 16k 1M 32 NetBSD 32k 1M 36.57 NetBSD 64k 1M 36.57 NetBSD 128k 1M 32 NetBSD 32k 256k 36.57 NetBSD 64k 256k 42.6 NetBSD 32k 128k 36.57 NetBSD 64k 128k 42.6 NetBSD 32k 64k 36.57 NetBSD 64k 64k 36.57 NetBSD 16k 64k 32 NetBSD 4k 64k 32 NetBSD 4k 32k 32 It appears that the best performance for these tests is 64k transfers and a 256k buffer. The same is true with the copy removed, although larger buffer and transfer sizes show an improvement: driver | xfer | buffer | maximum | size | size | rate -------------------------------------- NetBSD -copy 16k 2M 32 NetBSD -copy 64k 2M 42.6 NetBSD -copy 128k 2M 42.6 NetBSD -copy 64k 1M 42.6 NetBSD -copy 128k 1M 42.6 NetBSD -copy 32k 256k 42.6 NetBSD -copy 64k 256k 42.6 NetBSD -copy 32k 128k 36.57 NetBSD -copy 64k 128k 42.6 _______________________________________________ Discuss-gnuradio mailing list Discuss-gnuradio@gnu.org http://lists.gnu.org/mailman/listinfo/discuss-gnuradio