Hi all,

I can confirm that when using libgfapi with oVirt + Gluster replica 3 
(Hyperconverged) read and write performance under a VM was 4 to 5 times better 
than when using fuse.

--------------------------------------------------------------------------------------------------
Tested with a VM CentOS 6 and 7 under  the hyperconverged cluster HW:
--------------------------------------------------------------------------------------------------
ovirt 4.3.10 hypervisors with replica 3
- 256Gb Ram
- 32 total cores with hyperthreading
- RAID 1 (2 HDDs) for OS
- RAID 6 (9 SSDs) for Gluster , also tested with RAID 10, JBOD, all provided 
similar improvements with libgfapi (4 to 5 times better), replica 3 volumes.
- 10Gbe NICs, 1 for ovirtmgmnt and 1 for Gluster
- Ran tests using fio

-------------------------------------------------------------------------------
Test results using fuse (1500 MTU) (Took about 4~5 min):
-------------------------------------------------------------------------------

[root@test3 mail]# fio --randrepeat=1 --ioengine=libaio --direct=1 
--gtod_reduce=1 --name=test --filename=test --bs=4k --iodepth=64 --size=4G 
--readwrite=randrw --rwmixread=75
test: (g=0): rw=randrw, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio, iodepth=64
fio-2.0.13
Starting 1 process
test: Laying out IO file(s) (1 file(s) / 4096MB)
Jobs: 1 (f=1): [m] [100.0% done] [11984K/4079K/0K /s] [2996 /1019 /0  iops] 
[eta 00m:00s]
test: (groupid=0, jobs=1): err= 0: pid=8894: Mon Mar 29 10:05:35 2021
  read : io=3070.5MB, bw=12286KB/s, iops=3071 , runt=255918msec  
<------------------
  write: io=1025.6MB, bw=4103.5KB/s, iops=1025 , runt=255918msec  
<------------------
  cpu          : usr=1.84%, sys=10.50%, ctx=859129, majf=0, minf=19
  IO depths    : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=100.0%
     submit    : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
     complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0%
     issued    : total=r=786043/w=262533/d=0, short=r=0/w=0/d=0

Run status group 0 (all jobs):
   READ: io=3070.5MB, aggrb=12285KB/s, minb=12285KB/s, maxb=12285KB/s, 
mint=255918msec, maxt=255918msec
  WRITE: io=1025.6MB, aggrb=4103KB/s, minb=4103KB/s, maxb=4103KB/s, 
mint=255918msec, maxt=255918msec

Disk stats (read/write):
    dm-3: ios=785305/262494, merge=0/0, ticks=492833/15794537, 
in_queue=16289356, util=100.00%, aggrios=786024/262789, aggrmerge=19/45, 
aggrticks=492419/15811831, aggrin_queue=16303803, aggrutil=100.00%
  sda: ios=786024/262789, merge=19/45, ticks=492419/15811831, 
in_queue=16303803, util=100.00%


--------------------------------------------------------------------------------------------------------------------------
Test results using fuse (9000 MTU) // Did not see much of a difference (Took 
about 4~5 min):
--------------------------------------------------------------------------------------------------------------------------
[root@test3 mail]# fio --randrepeat=1 --ioengine=libaio --direct=1 
--gtod_reduce=1 --name=test --filename=test --bs=4k --iodepth=64 --size=4G 
--readwrite=randrw --rwmixread=75
test: (g=0): rw=randrw, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio, iodepth=64
fio-2.0.13
Starting 1 process
Jobs: 1 (f=1): [m] [100.0% done] [14956K/4596K/0K /s] [3739 /1149 /0  iops] 
[eta 00m:00s]
test: (groupid=0, jobs=1): err= 0: pid=2193: Mon Mar 29 10:22:44 2021
  read : io=3070.8MB, bw=12882KB/s, iops=3220 , runt=244095msec   
<------------------
  write: io=1025.3MB, bw=4300.1KB/s, iops=1075 , runt=244095msec   
<------------------
  cpu          : usr=1.85%, sys=10.43%, ctx=849742, majf=0, minf=21
  IO depths    : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=100.0%
     submit    : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
     complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0%
     issued    : total=r=786117/w=262459/d=0, short=r=0/w=0/d=0

Run status group 0 (all jobs):
   READ: io=3070.8MB, aggrb=12882KB/s, minb=12882KB/s, maxb=12882KB/s, 
mint=244095msec, maxt=244095msec
  WRITE: io=1025.3MB, aggrb=4300KB/s, minb=4300KB/s, maxb=4300KB/s, 
mint=244095msec, maxt=244095msec

Disk stats (read/write):
    dm-3: ios=785805/262493, merge=0/0, ticks=511951/15009580, 
in_queue=15523355, util=100.00%, aggrios=786105/262713, aggrmerge=18/19, 
aggrticks=511235/15026104, aggrin_queue=15536995, aggrutil=100.00%
  sda: ios=786105/262713, merge=18/19, ticks=511235/15026104, 
in_queue=15536995, util=100.00%

--------------------------------------------------------------------------------------
Test results using LIBGFAPI (9000 MTU), took about 38 seconds
--------------------------------------------------------------------------------------
[root@vmm04 ~]# ping -I glusternet -M do -s 8972 192.168.1.6
PING 192.168.1.6 (192.168.1.6) from 192.168.1.4 glusternet: 8972(9000) bytes of 
data.
8980 bytes from 192.168.1.6: icmp_seq=1 ttl=64 time=0.300 ms

[root@vmm04 ~]# ping -I ovirtmgmt -M do -s 8972 192.168.0.6
PING 192.168.0.6 (192.168.0.6) from 192.168.0.4 ovirtmgmt: 8972(9000) bytes of 
data.
8980 bytes from 192.168.0.6: icmp_seq=1 ttl=64 time=0.171 ms

[root@test3 mail]# fio --randrepeat=1 --ioengine=libaio --direct=1 
--gtod_reduce=1 --name=test --filename=test --bs=4k --iodepth=64 --size=4G 
--readwrite=randrw --rwmixread=75
test: (g=0): rw=randrw, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio, iodepth=64
fio-2.0.13
Starting 1 process
Jobs: 1 (f=1): [m] [100.0% done] [25878K/8599K/0K /s] [6469 /2149 /0  iops] 
[eta 00m:00s] 
test: (groupid=0, jobs=1): err= 0: pid=2188: Mon Mar 29 10:43:00 2021
  read : io=3071.2MB, bw=80703KB/s, iops=20175 , runt= 38969msec   
<------------------
  write: io=1024.9MB, bw=26929KB/s, iops=6732 , runt= 38969msec     
<------------------
  cpu          : usr=8.00%, sys=41.19%, ctx=374931, majf=0, minf=20
  IO depths    : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=100.0%
     submit    : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
     complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.1%, >=64=0.0%
     issued    : total=r=786224/w=262352/d=0, short=r=0/w=0/d=0

Run status group 0 (all jobs):
   READ: io=3071.2MB, aggrb=80702KB/s, minb=80702KB/s, maxb=80702KB/s, 
mint=38969msec, maxt=38969msec
  WRITE: io=1024.9MB, aggrb=26929KB/s, minb=26929KB/s, maxb=26929KB/s, 
mint=38969msec, maxt=38969msec

Disk stats (read/write):
    dm-3: ios=784858/261925, merge=0/0, ticks=1403884/1028357, 
in_queue=2433435, util=99.88%, aggrios=786155/262410, aggrmerge=70/51, 
aggrticks=1409868/1039790, aggrin_queue=2449280, aggrutil=99.82%
  sda: ios=786155/262410, merge=70/51, ticks=1409868/1039790, in_queue=2449280, 
util=99.82%


So I do agree with Guillaume, it be worth to re-evaluate the situation :) 


Regards,

Adrian
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/XSOW6SM3CQBGRR4FNPNX3ZN6KEHEBGVC/

Reply via email to