adding sahina Il ven 22 feb 2019, 14:01 matteo fedeli <[email protected]> ha scritto:
> Hi considering that the deploy with 4.2.7.8 failed I try to reinstall > ovirt to version 4.2.8 and there are appened two strange things. > During the volume step if i choose jbod mode in the deploy conf remain > raid6 type... Why? To solve I have only tried to editing manually the file > at line about volume type and the deploy stuck on creating physical > volume... > > this is my conf file: (I used 3 HDDs 500GB each, node,engine + vmstore and > data) > > #gdeploy configuration generated by cockpit-gluster plugin > [hosts] > kansas.planet.bn > germany.planet.bn > singapore.planet.bn > > [script1:kansas.planet.bn] > action=execute > ignore_script_errors=no > file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb -h > kansas.planet.bn, germany.planet.bn, singapore.planet.bn > > [script1:germany.planet.bn] > action=execute > ignore_script_errors=no > file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb -h > kansas.planet.bn, germany.planet.bn, singapore.planet.bn > > [script1:singapore.planet.bn] > action=execute > ignore_script_errors=no > file=/usr/share/gdeploy/scripts/grafton-sanity-check.sh -d sdb -h > kansas.planet.bn, germany.planet.bn, singapore.planet.bn > > [disktype] > jbod > > [diskcount] > 12 > > [stripesize] > 256 > > [service1] > action=enable > service=chronyd > > [service2] > action=restart > service=chronyd > > [shell2] > action=execute > command=vdsm-tool configure --force > > [script3] > action=execute > file=/usr/share/gdeploy/scripts/blacklist_all_disks.sh > ignore_script_errors=no > > [pv1:kansas.planet.bn] > action=create > devices=sdb > ignore_pv_errors=no > > [pv1:germany.planet.bn] > action=create > devices=sdb > ignore_pv_errors=no > > [pv1:singapore.planet.bn] > action=create > devices=sdb > ignore_pv_errors=no > > [vg1:kansas.planet.bn] > action=create > vgname=gluster_vg_sdb > pvname=sdb > ignore_vg_errors=no > > [vg1:germany.planet.bn] > action=create > vgname=gluster_vg_sdb > pvname=sdb > ignore_vg_errors=no > > [vg1:singapore.planet.bn] > action=create > vgname=gluster_vg_sdb > pvname=sdb > ignore_vg_errors=no > > [lv1:kansas.planet.bn] > action=create > poolname=gluster_thinpool_sdb > ignore_lv_errors=no > vgname=gluster_vg_sdb > lvtype=thinpool > size=1005GB > poolmetadatasize=5GB > > [lv2:germany.planet.bn] > action=create > poolname=gluster_thinpool_sdb > ignore_lv_errors=no > vgname=gluster_vg_sdb > lvtype=thinpool > size=1005GB > poolmetadatasize=5GB > > [lv3:singapore.planet.bn] > action=create > poolname=gluster_thinpool_sdb > ignore_lv_errors=no > vgname=gluster_vg_sdb > lvtype=thinpool > size=1005GB > poolmetadatasize=5GB > > [lv4:kansas.planet.bn] > action=create > lvname=gluster_lv_engine > ignore_lv_errors=no > vgname=gluster_vg_sdb > mount=/gluster_bricks/engine > size=100GB > lvtype=thick > > [lv5:kansas.planet.bn] > action=create > lvname=gluster_lv_data > ignore_lv_errors=no > vgname=gluster_vg_sdb > mount=/gluster_bricks/data > lvtype=thinlv > poolname=gluster_thinpool_sdb > virtualsize=500GB > > [lv6:kansas.planet.bn] > action=create > lvname=gluster_lv_vmstore > ignore_lv_errors=no > vgname=gluster_vg_sdb > mount=/gluster_bricks/vmstore > lvtype=thinlv > poolname=gluster_thinpool_sdb > virtualsize=500GB > > [lv7:germany.planet.bn] > action=create > lvname=gluster_lv_engine > ignore_lv_errors=no > vgname=gluster_vg_sdb > mount=/gluster_bricks/engine > size=100GB > lvtype=thick > > [lv8:germany.planet.bn] > action=create > lvname=gluster_lv_data > ignore_lv_errors=no > vgname=gluster_vg_sdb > mount=/gluster_bricks/data > lvtype=thinlv > poolname=gluster_thinpool_sdb > virtualsize=500GB > > [lv9:germany.planet.bn] > action=create > lvname=gluster_lv_vmstore > ignore_lv_errors=no > vgname=gluster_vg_sdb > mount=/gluster_bricks/vmstore > lvtype=thinlv > poolname=gluster_thinpool_sdb > virtualsize=500GB > > [lv10:singapore.planet.bn] > action=create > lvname=gluster_lv_engine > ignore_lv_errors=no > vgname=gluster_vg_sdb > mount=/gluster_bricks/engine > size=100GB > lvtype=thick > > [lv11:singapore.planet.bn] > action=create > lvname=gluster_lv_data > ignore_lv_errors=no > vgname=gluster_vg_sdb > mount=/gluster_bricks/data > lvtype=thinlv > poolname=gluster_thinpool_sdb > virtualsize=500GB > > [lv12:singapore.planet.bn] > action=create > lvname=gluster_lv_vmstore > ignore_lv_errors=no > vgname=gluster_vg_sdb > mount=/gluster_bricks/vmstore > lvtype=thinlv > poolname=gluster_thinpool_sdb > virtualsize=500GB > > [selinux] > yes > > [service3] > action=restart > service=glusterd > slice_setup=yes > > [firewalld] > action=add > > ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp,54322/tcp > services=glusterfs > > [script2] > action=execute > file=/usr/share/gdeploy/scripts/disable-gluster-hooks.sh > > [shell3] > action=execute > command=usermod -a -G gluster qemu > > [volume1] > action=create > volname=engine > transport=tcp > replica=yes > replica_count=3 > > key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal > value=virt,36,36,30,on,off,enable > brick_dirs=kansas.planet.bn: > /gluster_bricks/engine/engine,germany.planet.bn: > /gluster_bricks/engine/engine,singapore.planet.bn: > /gluster_bricks/engine/engine > ignore_volume_errors=no > > [volume2] > action=create > volname=data > transport=tcp > replica=yes > replica_count=3 > > key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal > value=virt,36,36,30,on,off,enable > brick_dirs=kansas.planet.bn:/gluster_bricks/data/data,germany.planet.bn: > /gluster_bricks/data/data,singapore.planet.bn:/gluster_bricks/data/data > ignore_volume_errors=no > > [volume3] > action=create > volname=vmstore > transport=tcp > replica=yes > replica_count=3 > > key=group,storage.owner-uid,storage.owner-gid,network.ping-timeout,performance.strict-o-direct,network.remote-dio,cluster.granular-entry-heal > value=virt,36,36,30,on,off,enable > brick_dirs=kansas.planet.bn: > /gluster_bricks/vmstore/vmstore,germany.planet.bn: > /gluster_bricks/vmstore/vmstore,singapore.planet.bn: > /gluster_bricks/vmstore/vmstore > ignore_volume_errors=no > _______________________________________________ > Users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/GS6YGVEQNIQ63RHH4WNGUSQIMOJJNE3B/ >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/7Z5WCGIGSLQHUXQHJLGA3PBCR54WUG5I/

