Hello I am trying to run a Hyperconverged setup "COnfigure gluster storage and
ovirt hosted engine", however I get the following error
__________________________________________________________________________________________________
PLAY [gluster_servers]
*********************************************************
TASK [Create LVs with specified size for the VGs] ******************************
failed: [ovirt01.grupokino.com] (item={u'lv': u'gluster_thinpool_sdb', u'size':
u'45GB', u'extent': u'100%FREE', u'vg': u'gluster_vg_sdb'}) => {"changed":
false, "item": {"extent": "100%FREE", "lv": "gluster_thinpool_sdb", "size":
"45GB", "vg": "gluster_vg_sdb"}, "msg": "lvcreate: metadata/pv_map.c:198:
consume_pv_area: Assertion `to_go <= pva->count' failed.\n", "rc": -6}
to retry, use: --limit @/tmp/tmpwo4SNB/lvcreate.retry
PLAY RECAP *********************************************************************
ovirt01.grupokino.com : ok=0 changed=0 unreachable=0 failed=1
__________________________________________________________________________________________________
I know that oVirt Hosted Engine Setup GUI for "gluster wizard (gluster
deployment) does not populate the geodeployConfig.conf file properly (Generated
Gdeploy configuration :
/var/lib/ovirt-hosted-engine-setup/gdeploy/gdeployConfig.conf) so I have tried
to modify it to fit our needs but keep getting the above error everytime.
Any ideas or comments are welcome...Thanks!
My servers are setup with 4x50GB disks, 1 for the OS and the rest for Gluster
Hyperconverged setup.
__________________________________________________________________________________________________
my gdeployConfig.conf file:
__________________________________________________________________________________________________
#gdeploy configuration generated by cockpit-gluster plugin
[hosts]
ovirt01.mydomain.com
ovirt02.mydomain.com
ovirt03.mydomain.com
[script1:ovirt01.mydomain.com]
action=execute
ignore_script_errors=no
file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb,sdc,sdd -h
ovirt01.mydomain.com, ovirt02.mydomain.com, ovirt03.mydomain.com
[script1:ovirt02.mydomain.com]
action=execute
ignore_script_errors=no
file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb,sdc,sdd -h
ovirt01.mydomain.com, ovirt02.mydomain.com, ovirt03.mydomain.com
[script1:ovirt03.mydomain.com]
action=execute
ignore_script_errors=no
file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb,sdc,sdd -h
ovirt01.mydomain.com, ovirt02.mydomain.com, ovirt03.mydomain.com
[disktype]
jbod
[diskcount]
3
[stripesize]
256
[service1]
action=enable
service=chronyd
[service2]
action=restart
service=chronyd
[shell2]
action=execute
command=vdsm-tool configure --force
[script3]
action=execute
file=/usr/share/gdeploy/scripts/blacklist_all_disks.sh
ignore_script_errors=no
[pv1:ovirt01.mydomain.com]
action=create
devices=sdb
ignore_pv_errors=no
[pv1:ovirt02.mydomain.com]
action=create
devices=sdb
ignore_pv_errors=no
[pv1:ovirt03.mydomain.com]
action=create
devices=sdb
ignore_pv_errors=no
[pv2:ovirt01.mydomain.com]
action=create
devices=sdc
ignore_pv_errors=no
[pv2:ovirt02.mydomain.com]
action=create
devices=sdc
ignore_pv_errors=no
[pv2:ovirt03.mydomain.com]
action=create
devices=sdc
ignore_pv_errors=no
[pv3:ovirt01.mydomain.com]
action=create
devices=sdd
ignore_pv_errors=no
[pv3:ovirt02.mydomain.com]
action=create
devices=sdd
ignore_pv_errors=no
[pv3:ovirt03.mydomain.com]
action=create
devices=sdd
ignore_pv_errors=no
[vg1:ovirt01.mydomain.com]
action=create
vgname=gluster_vg_sdb
pvname=sdb
ignore_vg_errors=no
[vg1:ovirt02.mydomain.com]
action=create
vgname=gluster_vg_sdb
pvname=sdb
ignore_vg_errors=no
[vg1:ovirt03.mydomain.com]
action=create
vgname=gluster_vg_sdb
pvname=sdb
ignore_vg_errors=no
[vg2:ovirt01.mydomain.com]
action=create
vgname=gluster_vg_sdc
pvname=sdc
ignore_vg_errors=no
[vg2:ovirt02.mydomain.com]
action=create
vgname=gluster_vg_sdc
pvname=sdc
ignore_vg_errors=no
[vg2:ovirt03.mydomain.com]
action=create
vgname=gluster_vg_sdc
pvname=sdc
ignore_vg_errors=no
[vg3:ovirt01.mydomain.com]
action=create
vgname=gluster_vg_sdd
pvname=sdd
ignore_vg_errors=no
[vg3:ovirt02.mydomain.com]
action=create
vgname=gluster_vg_sdd
pvname=sdd
ignore_vg_errors=no
[vg3:ovirt03.mydomain.com]
action=create
vgname=gluster_vg_sdd
pvname=sdd
ignore_vg_errors=no
[lv1:ovirt01.mydomain.com]
action=create
poolname=gluster_thinpool_sdb
ignore_lv_errors=no
vgname=gluster_vg_sdb
lvtype=thinpool
size=45GB
poolmetadatasize=3GB
[lv2:ovirt02.mydomain.com]
action=create
poolname=gluster_thinpool_sdc
ignore_lv_errors=no
vgname=gluster_vg_sdc
lvtype=thinpool
size=45GB
poolmetadatasize=3GB
[lv3:ovirt03.mydomain.com]
action=create
poolname=gluster_thinpool_sdd
ignore_lv_errors=no
vgname=gluster_vg_sdd
lvtype=thinpool
size=45GB
poolmetadatasize=3GB
[lv4:ovirt01.mydomain.com]
action=create
lvname=gluster_lv_engine
ignore_lv_errors=no
vgname=gluster_vg_sdb
mount=/gluster_bricks/engine
size=42GB
lvtype=thick
[lv5:ovirt01.mydomain.com]
action=create
lvname=gluster_lv_data
ignore_lv_errors=no
vgname=gluster_vg_sdc
mount=/gluster_bricks/data
lvtype=thinlv
poolname=gluster_thinpool_sdc
virtualsize=42GB
[lv6:ovirt01.mydomain.com]
action=create
lvname=gluster_lv_vmstore
ignore_lv_errors=no
vgname=gluster_vg_sdd
mount=/gluster_bricks/vmstore
lvtype=thinlv
poolname=gluster_thinpool_sdd
virtualsize=42GB
[lv7:ovirt02.mydomain.com]
action=create
lvname=gluster_lv_engine
ignore_lv_errors=no
vgname=gluster_vg_sdb
mount=/gluster_bricks/engine
size=42GB
lvtype=thick
[lv8:ovirt02.mydomain.com]
action=create
lvname=gluster_lv_data
ignore_lv_errors=no
vgname=gluster_vg_sdc
mount=/gluster_bricks/data
lvtype=thinlv
poolname=gluster_thinpool_sdc
virtualsize=42GB
[lv9:ovirt02.mydomain.com]
action=create
lvname=gluster_lv_vmstore
ignore_lv_errors=no
vgname=gluster_vg_sdd
mount=/gluster_bricks/vmstore
lvtype=thinlv
poolname=gluster_thinpool_sdd
virtualsize=42GB
[lv10:ovirt03.mydomain.com]
action=create
lvname=gluster_lv_engine
ignore_lv_errors=no
vgname=gluster_vg_sdb
mount=/gluster_bricks/engine
size=42GB
lvtype=thick
[lv11:ovirt03.mydomain.com]
action=create
lvname=gluster_lv_data
ignore_lv_errors=no
vgname=gluster_vg_sdc
mount=/gluster_bricks/data
lvtype=thinlv
poolname=gluster_thinpool_sdc
virtualsize=42GB
[lv12:ovirt03.mydomain.com]
action=create
lvname=gluster_lv_vmstore
ignore_lv_errors=no
vgname=gluster_vg_sdd
mount=/gluster_bricks/vmstore
lvtype=thinlv
poolname=gluster_thinpool_sdd
virtualsize=42GB
[selinux]
yes
[service3]
action=restart
service=glusterd
slice_setup=yes
[firewalld]
action=add
ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,54322/tcp
services=glusterfs
[script2]
action=execute
file=/usr/share/gdeploy/scripts/disable-gluster-hooks.sh
[shell3]
action=execute
command=usermod -a -G gluster qemu
[volume1]
action=create
volname=engine
transport=tcp
replica=yes
replica_count=3
key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal
value=virt,36,36,30,on,off,enable
brick_dirs=ovirt01.mydomain.com:/gluster_bricks/engine/engine,ovirt02.mydomain.com:/gluster_bricks/engine/engine,ovirt03.mydomain.com:/gluster_bricks/engine/engine
ignore_volume_errors=no
[volume2]
action=create
volname=data
transport=tcp
replica=yes
replica_count=3
key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal
value=virt,36,36,30,on,off,enable
brick_dirs=ovirt01.mydomain.com:/gluster_bricks/data/data,ovirt02.mydomain.com:/gluster_bricks/data/data,ovirt03.mydomain.com:/gluster_bricks/data/data
ignore_volume_errors=no
[volume3]
action=create
volname=vmstore
transport=tcp
replica=yes
replica_count=3
key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal
value=virt,36,36,30,on,off,enable
brick_dirs=ovirt01.mydomain.com:/gluster_bricks/vmstore/vmstore,ovirt02.mydomain.com:/gluster_bricks/vmstore/vmstore,ovirt03.mydomain.com:/gluster_bricks/vmstore/vmstore
ignore_volume_errors=no
---------------------------------------------------------------------------------------------------
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/[email protected]/message/4DCUGOSWOK6KNT26ROWYOITD5JQMQ3NE/