[design idea]
The vhost-blk uses two kernel threads to handle the guests' requests.
One is tosubmit them via Linux kernel's internal AIO structs, and the other is
signal the guests the completion of the IO requests.
The current qemu-kvm's native AIO in the user mode acctually just uses
one io-thread to submitting and signalling. One more nuance is that qemu-kvm
AIO signals the completion of the requests one by one.
Like vhost-net, the in-kernel vhost-blk module reduces the number of
the system calls during the requests handling and the code path is shorter than
the implementation of the qemu-kvm.
[performance]
Currently, the fio benchmarking number is rather promising. The seq
read is imporved as much as 16% for throughput and the latency is dropped up to
14%. For seq write, 13.5% and 13% respectively.
sequential read:
+-------------+-------------+---------------+---------------+
| iodepth | 1 | 2 | 3 |
+-------------+-------------+---------------+----------------
| virtio-blk | 4116(214) | 7814(222) | 8867(306) |
+-------------+-------------+---------------+---------------+
| vhost-blk | 4755(183) | 8645(202) | 10084(266) |
+-------------+-------------+---------------+---------------+
4116(214) means 4116 IOPS/s, the it is completion latency is 214 us.
seqeuential write:
+-------------+-------------+----------------+--------------+
| iodepth | 1 | 2 | 3 |
+-------------+-------------+----------------+--------------+
| virtio-blk | 3848(228) | 6505(275) | 9335(291) |
+-------------+-------------+----------------+--------------+
| vhost-blk | 4370(198) | 7009(249) | 9938(264) |
+-------------+-------------+----------------+--------------+
the fio command for sequential read:
sudo fio -name iops -readonly -rw=read -runtime=120 -iodepth 1 -filename
/dev/vda -ioengine libaio -direct=1 -bs=512
and config file for sequential write is:
dev@taobao:~$ cat rw.fio
-------------------------
[test]
rw=rw
size=200M
directory=/home/dev/data
ioengine=libaio
iodepth=1
direct=1
bs=512
-------------------------
These numbers are collected on my laptop with Intel Core i5 CPU,
2.67GHz, SATA harddisk with 7200 RPM. Both guest and host use Linux 3.0-rc6
kernel with ext4 filesystem.
I setup the Guest by:
sudo x86_64-softmmu/qemu-system-x86_64 -cpu host -m 512 -drive
file=/dev/sda6,if=virtio,cache=none,aio=native -nographic
The patchset is very primitive and need much further improvement for
both funtionality and performance.
Inputs and suggestions are more than welcome.
Yuan
--
drivers/vhost/Makefile | 3 +
drivers/vhost/blk.c | 568 ++++++++++++++++++++++++++++++++++++++++++++++++
drivers/vhost/vhost.h | 11 +
fs/aio.c | 44 ++---
fs/eventfd.c | 1 +
include/linux/aio.h | 31 +++
6 files changed, 631 insertions(+), 27 deletions(-)
--
Makefile.target | 2 +-
hw/vhost_blk.c | 84 +++++++++++++++++++++++++++++++++++++++++++++++++++++++
hw/vhost_blk.h | 44 ++++++++++++++++++++++++++++
hw/virtio-blk.c | 74 ++++++++++++++++++++++++++++++++++++++----------
hw/virtio-blk.h | 15 ++++++++++
hw/virtio-pci.c | 12 ++++++-
6 files changed, 213 insertions(+), 18 deletions(-)
In-Reply-To:
--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html