I’ve tried various methods to improve gluster performance on similar
hardware and never had much luck. Small file workloads were particularly
troublesome. I ended up switching high performance vms to nfs storage and
performance with nfs improved greatly in my use case.

On Sun, Jun 28, 2020 at 6:42 PM shadow emy <[email protected]> wrote:

> > Hello ,
>
> Hello and thank you for the reply.Bellow are the answers to your questions.
> >
> > Let me ask some questions:
> > 1. What is the scheduler for your PV ?
>
>
> On the Raid Controller device where the SSD disks are in Raid 0 (device
> sda) it is set to "deadline". But on the lvm volume logical volume dm-7,
> where the logical block is set for "data" volunr it is set to none.(i think
> this is ok )
>
>
> [root@host1 ~]# ls -al /dev/mapper/gluster_vg_sd
> v_data ter_l
> lrwxrwxrwx. 1 root root 7 Jun 28 14:14 /dev/mapper/gluster_v
> g_sda3-gluster_lv_data -> ../dm-7
> [root@host1 ~]# cat /sys/block/dm-7/queue/scheduler
> none
> root@host1:~[root@host1 ~]# cat /sys/block/dm-7/queue/schedu
> [root@host1 ~]# cat /sys/block/sda/queue/scheduler
> noop [deadline] cfq
>
>
>
> > 2. Have you aligned your PV during the setup 'pvcreate
> --dataalignment alignment_value
> > device'
>
>
> I did not made other alignment then the default.Bellow are the partitions
> on /dev/sda.
> Can i enable partition alignment now, if yes how ?
>
> sfdisk -d /dev/sda
> # partition table of /dev/sda
> unit: sectors
>
> /dev/sda1 : start=     2048, size=   487424, Id=83, bootable
> /dev/sda2 : start=   489472, size= 95731712, Id=8e
> /dev/sda3 : start= 96221184, size=3808675840, Id=83
> /dev/sda4 : start=        0, size=        0, Id= 0
>
>
>
> > 3. What is your tuned profile ? Do you use rhgs-random-io from
> > the
> ftp://ftp.redhat.com/redhat/linux/enterprise/7Server/en/RHS/SRPMS/red...
> > ?
>
> My tuned active profile is virtual-host
>
> Current active profile: virtual-host
>
>  No i dont use any of the rhgs-random-io profiles
>
> > 4. What is the output of "xfs_info /path/to/your/gluster/brick" ?
>
> xfs_info /gluster_bricks/data
> meta-data=/dev/mapper/gluster_vg_sda3-gluster_lv_data isize=
> 512    agcount=32, agsize=6553600 blks
>          =                       sectsz=512   attr=2, projid
> 32bit=1
>          =                       crc=1        finobt=0 spino
> des=0
> data     =                       bsize=4096   blocks=2097152
> 00, imaxpct=25
>          =                       sunit=64     swidth=64 blks
> naming   =version 2              bsize=8192   ascii-ci=0 fty
> pe=1
> log      =internal               bsize=4096   blocks=102400,
>  version=2
>          =                       sectsz=512   sunit=64 blks,
>  lazy-count=1
> realtime =none                   extsz=4096   blocks=0, rtex
> tents=0
>
> > 5. Are you using Jumbo Frames ? Does your infra support them?
> > Usually MTU of 9k is standard, but some switches and NICs support up to
> 16k.
> >
>
> Unfortunately  I can not enable MTU to 9000 and Jumbo Frames on these
> Cisco SG350X switches to specific ports.The switches  dont suport Jumbo
> Frames enable  to a single port, only on all ports .
> I have others devices connected to the switches on the remaining 48 ports
> that have  1Gb/s.
>
> > All the options for "optimize for virt...." are located
> > at /var/lib/glusterd/groups/virt on each gluster node.
>
> I have already looked  previously at that file, but not all the volume
> settings  that are set by "Optime for Virt Store" are stored there.
> For example  "Optimize for Virt Store " sets network.remote.dio   to
> disable and in the glusterd/groups/virt is set to enabled.Or
> cluster.granular-entry-heal: enable is not present there, bit it is set by
> "Optimize for Virt Store"
>
> >
> > Best Regards,
> > Strahil Nikolov
> >
> >
> >
> >
> > В неделя, 28 юни 2020 г., 22:13:09 Гринуич+3, jury cat <shadow.emy1(a)
> gmail.com&gt;
> > написа:
> >
> >
> >
> >
> >
> > Hello all,
> >
> > I am using Ovirt 4.3.10 on Centos 7.8 with glusterfs 6.9 .
> > My Gluster setup is of 3 hosts in replica 3 (2 hosts + 1 arbiter).
> > All the 3 hosts are Dell R720  with Perc Raid Controller H710 mini(that
> has maximim
> > throughtout 6Gbs)  and  with 2×1TB samsumg SSD in RAID 0. The volume is
> partitioned using
> > LVM thin provision and formated XFS.
> > The hosts have separate 10GE network cards for storage traffic.
> > The Gluster Network is connected to this 10GE network cards and is
> mounted using Fuse
> > Glusterfs(NFS is disabled).Also Migration Network is activated on the
> same storage
> > network.
> >
> >
> > The problem is that the 10GE network is not used at full potential by
> the Gluster.
> > If i do live Migration of Vms i can see speeds of 7GB/s ~ 9GB/s.
> > The same network tests using iperf3 reported 9.9GB/s ,  these exluding
> the network setup
> > as a bottleneck(i will not paste all the iperf3 tests here for now).
> > I did not enable all the Volume options  from "Optimize for Virt Store",
> because
> > of the bug that cant set volume  cluster.granural-heal to enable(this
> was fixed in vdsm-4
> > 40, but that is working only on Centos 8 with ovirt 4.4 ) .
> > i whould be happy to know what are all these "Optimize for Virt Store"
> options,
> > so i can set them manually.
> >
> >
> > The speed on the disk inside the host using dd is b etween 1GB/s to
> 700Mbs.
> >
> >
> > [root@host1 ~]# dd if=/dev/zero of=test bs=100M count=40 cou nt=80
> status=progress
> > 8074035200 bytes (8.1 GB) copied, 11.059372 s, 730 MB/s 80+0 records in
> 80+0 records out
> > 8388608000 bytes (8.4 GB) copied, 11.9928 s, 699 MB/s
> >
> >
> > The dd  write test on the gluster volme inside the host is poor only  ~
> 120MB/s .
> > During the dd test, if i look at Networks->Gluster network ->Hosts at Tx
> and Rx the
> > network speed barerly reaches over  1Gbs  (~1073 Mbs) out of maximum of
> 10000 Mbs.
> >
> >
> >  dd if=/dev/zero of=/rhev/data-center/mnt/glu
> sterSD/gluster1.domain.local\: _data/test
> > bs=100M count=80 status=progress 8283750400 bytes (8.3 GB) copied,
> 71.297942 s, 116 MB/s
> > 80+0 records in 80+0 records out 8388608000 bytes (8.4 GB) copied,
> 71.9545 s, 117 MB/s
> >
> >
> > I have attached  my Gluster volume settings and mount options.
> >
> > Thanks,
> > Emy
> >
> >
> > _______________________________________________
> > Users mailing list -- users(a)ovirt.org
> > To unsubscribe send an email to users-leave(a)ovirt.org
> > Privacy Statement: https://www.ovirt.org/privacy-policy.html
> > oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> > List Archives:
> >
> https://lists.ovirt.org/archives/list/[email protected]/message/7BR6TZQ4EXS.
> ..
> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
> Privacy Statement: https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/[email protected]/message/R5XPLZ7QARJG2TYRTOK6BADJDNONPOPQ/
>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/XSBM4D4PS6R6FNYP3MZDNAI7X4PKIMQT/

Reply via email to