On Thu, Jun 13, 2019 at 7:11 AM <[email protected]> wrote:

> While trying to do a hyperconverged setup and trying to use "configure LV
> Cache" /dev/sdf the deployment fails. If I dont use the LV cache SSD Disk
> the setup succeds, thought you mighg want to know, for now I retested with
> 4.3.3 and all worked fine, so reverting to 4.3.3 unless you know of a
> workaround?
>
> Error:
> TASK [gluster.infra/roles/backend_setup : Extend volume group]
> *****************
> failed: [vmm11.mydomain.com] (item={u'vgname': u'gluster_vg_sdb',
> u'cachethinpoolname': u'gluster_thinpool_gluster_vg_sdb', u'cachelvname':
> u'cachelv_gluster_thinpool_gluster_vg_sdb', u'cachedisk': u'/dev/sdf',
> u'cachemetalvname': u'cache_gluster_thinpool_gluster_vg_sdb', u'cachemode':
> u'writethrough', u'cachemetalvsize': u'0.1G', u'cachelvsize': u'0.9G'}) =>
> {"ansible_loop_var": "item", "changed": false, "err": "  Physical volume
> \"/dev/sdb\" still in use\n", "item": {"cachedisk": "/dev/sdf",
> "cachelvname": "cachelv_gluster_thinpool_gluster_vg_sdb", "cachelvsize":
> "0.9G", "cachemetalvname": "cache_gluster_thinpool_gluster_vg_sdb",
> "cachemetalvsize": "0.1G", "cachemode": "writethrough",
> "cachethinpoolname": "gluster_thinpool_gluster_vg_sdb", "vgname":


The variable file does not seem to be right.
You have mentioned cachethinpoolname: gluster_thinpool_gluster_vg_sdb but
you are not creating it anywhere.
So, the Ansible module is trying to shrink the volume group.

Also why is cachelvsize is 0.9G and chachemetasize 0.1G? Isn't it too less?

Please refer:
https://github.com/gluster/gluster-ansible/blob/master/playbooks/hc-ansible-deployment/gluster_inventory.yml
for example.

-sac

>
> "gluster_vg_sdb"}, "msg": "Unable to reduce gluster_vg_sdb by /dev/sdb.",
> "rc": 5}
>
> failed: [vmm12.mydomain.com] (item={u'vgname': u'gluster_vg_sdb',
> u'cachethinpoolname': u'gluster_thinpool_gluster_vg_sdb', u'cachelvname':
> u'cachelv_gluster_thinpool_gluster_vg_sdb', u'cachedisk': u'/dev/sdf',
> u'cachemetalvname': u'cache_gluster_thinpool_gluster_vg_sdb', u'cachemode':
> u'writethrough', u'cachemetalvsize': u'0.1G', u'cachelvsize': u'0.9G'}) =>
> {"ansible_loop_var": "item", "changed": false, "err": "  Physical volume
> \"/dev/sdb\" still in use\n", "item": {"cachedisk": "/dev/sdf",
> "cachelvname": "cachelv_gluster_thinpool_gluster_vg_sdb", "cachelvsize":
> "0.9G", "cachemetalvname": "cache_gluster_thinpool_gluster_vg_sdb",
> "cachemetalvsize": "0.1G", "cachemode": "writethrough",
> "cachethinpoolname": "gluster_thinpool_gluster_vg_sdb", "vgname":
> "gluster_vg_sdb"}, "msg": "Unable to reduce gluster_vg_sdb by /dev/sdb.",
> "rc": 5}
>
> failed: [vmm10.mydomain.com] (item={u'vgname': u'gluster_vg_sdb',
> u'cachethinpoolname': u'gluster_thinpool_gluster_vg_sdb', u'cachelvname':
> u'cachelv_gluster_thinpool_gluster_vg_sdb', u'cachedisk': u'/dev/sdf',
> u'cachemetalvname': u'cache_gluster_thinpool_gluster_vg_sdb', u'cachemode':
> u'writethrough', u'cachemetalvsize': u'30G', u'cachelvsize': u'270G'}) =>
> {"ansible_loop_var": "item", "changed": false, "err": "  Physical volume
> \"/dev/sdb\" still in use\n", "item": {"cachedisk": "/dev/sdf",
> "cachelvname": "cachelv_gluster_thinpool_gluster_vg_sdb", "cachelvsize":
> "270G", "cachemetalvname": "cache_gluster_thinpool_gluster_vg_sdb",
> "cachemetalvsize": "30G", "cachemode": "writethrough", "cachethinpoolname":
> "gluster_thinpool_gluster_vg_sdb", "vgname": "gluster_vg_sdb"}, "msg":
> "Unable to reduce gluster_vg_sdb by /dev/sdb.", "rc": 5}
>
> PLAY RECAP
> *********************************************************************
> vmm10.mydomain.com           : ok=13   changed=4    unreachable=0
> failed=1    skipped=10   rescued=0    ignored=0
> vmm11.mydomain.com           : ok=13   changed=4    unreachable=0
> failed=1    skipped=10   rescued=0    ignored=0
> vmm12.mydomain.com           : ok=13   changed=4    unreachable=0
> failed=1    skipped=10   rescued=0    ignored=0
>
>
>
>
> ---------------------------------------------------------------------------------------------------------------------
> #cat /etc/ansible/hc_wizard_inventory.yml
>
> ---------------------------------------------------------------------------------------------------------------------
> hc_nodes:
>   hosts:
>     vmm10.mydomain.com:
>       gluster_infra_volume_groups:
>         - vgname: gluster_vg_sdb
>           pvname: /dev/sdb
>         - vgname: gluster_vg_sdc
>           pvname: /dev/sdc
>         - vgname: gluster_vg_sdd
>           pvname: /dev/sdd
>         - vgname: gluster_vg_sde
>           pvname: /dev/sde
>       gluster_infra_mount_devices:
>         - path: /gluster_bricks/engine
>           lvname: gluster_lv_engine
>           vgname: gluster_vg_sdb
>         - path: /gluster_bricks/vmstore1
>           lvname: gluster_lv_vmstore1
>           vgname: gluster_vg_sdc
>         - path: /gluster_bricks/data1
>           lvname: gluster_lv_data1
>           vgname: gluster_vg_sdd
>         - path: /gluster_bricks/data2
>           lvname: gluster_lv_data2
>           vgname: gluster_vg_sde
>       gluster_infra_cache_vars:
>         - vgname: gluster_vg_sdb
>           cachedisk: /dev/sdf
>           cachelvname: cachelv_gluster_thinpool_gluster_vg_sdb
>           cachethinpoolname: gluster_thinpool_gluster_vg_sdb
>           cachelvsize: 270G
>           cachemetalvsize: 30G
>           cachemetalvname: cache_gluster_thinpool_gluster_vg_sdb
>           cachemode: writethrough
>       gluster_infra_thick_lvs:
>         - vgname: gluster_vg_sdb
>           lvname: gluster_lv_engine
>           size: 100G
>       gluster_infra_thinpools:
>         - vgname: gluster_vg_sdc
>           thinpoolname: gluster_thinpool_gluster_vg_sdc
>           poolmetadatasize: 14G
>         - vgname: gluster_vg_sdd
>           thinpoolname: gluster_thinpool_gluster_vg_sdd
>           poolmetadatasize: 14G
>         - vgname: gluster_vg_sde
>           thinpoolname: gluster_thinpool_gluster_vg_sde
>           poolmetadatasize: 14G
>       gluster_infra_lv_logicalvols:
>         - vgname: gluster_vg_sdc
>           thinpool: gluster_thinpool_gluster_vg_sdc
>           lvname: gluster_lv_vmstore1
>           lvsize: 2700G
>         - vgname: gluster_vg_sdd
>           thinpool: gluster_thinpool_gluster_vg_sdd
>           lvname: gluster_lv_data1
>           lvsize: 2700G
>         - vgname: gluster_vg_sde
>           thinpool: gluster_thinpool_gluster_vg_sde
>           lvname: gluster_lv_data2
>           lvsize: 2700G
>     vmm11.mydomain.com:
>       gluster_infra_volume_groups:
>         - vgname: gluster_vg_sdb
>           pvname: /dev/sdb
>         - vgname: gluster_vg_sdc
>           pvname: /dev/sdc
>         - vgname: gluster_vg_sdd
>           pvname: /dev/sdd
>         - vgname: gluster_vg_sde
>           pvname: /dev/sde
>       gluster_infra_mount_devices:
>         - path: /gluster_bricks/engine
>           lvname: gluster_lv_engine
>           vgname: gluster_vg_sdb
>         - path: /gluster_bricks/vmstore1
>           lvname: gluster_lv_vmstore1
>           vgname: gluster_vg_sdc
>         - path: /gluster_bricks/data1
>           lvname: gluster_lv_data1
>           vgname: gluster_vg_sdd
>         - path: /gluster_bricks/data2
>           lvname: gluster_lv_data2
>           vgname: gluster_vg_sde
>       gluster_infra_cache_vars:
>         - vgname: gluster_vg_sdb
>           cachedisk: /dev/sdf
>           cachelvname: cachelv_gluster_thinpool_gluster_vg_sdb
>           cachethinpoolname: gluster_thinpool_gluster_vg_sdb
>           cachelvsize: 0.9G
>           cachemetalvsize: 0.1G
>           cachemetalvname: cache_gluster_thinpool_gluster_vg_sdb
>           cachemode: writethrough
>       gluster_infra_thick_lvs:
>         - vgname: gluster_vg_sdb
>           lvname: gluster_lv_engine
>           size: 100G
>       gluster_infra_thinpools:
>         - vgname: gluster_vg_sdc
>           thinpoolname: gluster_thinpool_gluster_vg_sdc
>           poolmetadatasize: 14G
>         - vgname: gluster_vg_sdd
>           thinpoolname: gluster_thinpool_gluster_vg_sdd
>           poolmetadatasize: 14G
>         - vgname: gluster_vg_sde
>           thinpoolname: gluster_thinpool_gluster_vg_sde
>           poolmetadatasize: 14G
>       gluster_infra_lv_logicalvols:
>         - vgname: gluster_vg_sdc
>           thinpool: gluster_thinpool_gluster_vg_sdc
>           lvname: gluster_lv_vmstore1
>           lvsize: 2700G
>         - vgname: gluster_vg_sdd
>           thinpool: gluster_thinpool_gluster_vg_sdd
>           lvname: gluster_lv_data1
>           lvsize: 2700G
>         - vgname: gluster_vg_sde
>           thinpool: gluster_thinpool_gluster_vg_sde
>           lvname: gluster_lv_data2
>           lvsize: 2700G
>     vmm12.mydomain.com:
>       gluster_infra_volume_groups:
>         - vgname: gluster_vg_sdb
>           pvname: /dev/sdb
>         - vgname: gluster_vg_sdc
>           pvname: /dev/sdc
>         - vgname: gluster_vg_sdd
>           pvname: /dev/sdd
>         - vgname: gluster_vg_sde
>           pvname: /dev/sde
>       gluster_infra_mount_devices:
>         - path: /gluster_bricks/engine
>           lvname: gluster_lv_engine
>           vgname: gluster_vg_sdb
>         - path: /gluster_bricks/vmstore1
>           lvname: gluster_lv_vmstore1
>           vgname: gluster_vg_sdc
>         - path: /gluster_bricks/data1
>           lvname: gluster_lv_data1
>           vgname: gluster_vg_sdd
>         - path: /gluster_bricks/data2
>           lvname: gluster_lv_data2
>           vgname: gluster_vg_sde
>       gluster_infra_cache_vars:
>         - vgname: gluster_vg_sdb
>           cachedisk: /dev/sdf
>           cachelvname: cachelv_gluster_thinpool_gluster_vg_sdb
>           cachethinpoolname: gluster_thinpool_gluster_vg_sdb
>           cachelvsize: 0.9G
>           cachemetalvsize: 0.1G
>           cachemetalvname: cache_gluster_thinpool_gluster_vg_sdb
>           cachemode: writethrough
>       gluster_infra_thick_lvs:
>         - vgname: gluster_vg_sdb
>           lvname: gluster_lv_engine
>           size: 100G
>       gluster_infra_thinpools:
>         - vgname: gluster_vg_sdc
>           thinpoolname: gluster_thinpool_gluster_vg_sdc
>           poolmetadatasize: 14G
>         - vgname: gluster_vg_sdd
>           thinpoolname: gluster_thinpool_gluster_vg_sdd
>           poolmetadatasize: 14G
>         - vgname: gluster_vg_sde
>           thinpoolname: gluster_thinpool_gluster_vg_sde
>           poolmetadatasize: 14G
>       gluster_infra_lv_logicalvols:
>         - vgname: gluster_vg_sdc
>           thinpool: gluster_thinpool_gluster_vg_sdc
>           lvname: gluster_lv_vmstore1
>           lvsize: 2700G
>         - vgname: gluster_vg_sdd
>           thinpool: gluster_thinpool_gluster_vg_sdd
>           lvname: gluster_lv_data1
>           lvsize: 2700G
>         - vgname: gluster_vg_sde
>           thinpool: gluster_thinpool_gluster_vg_sde
>           lvname: gluster_lv_data2
>           lvsize: 2700G
>   vars:
>     gluster_infra_disktype: JBOD
>     gluster_set_selinux_labels: true
>     gluster_infra_fw_ports:
>       - 2049/tcp
>       - 54321/tcp
>       - 5900/tcp
>       - 5900-6923/tcp
>       - 5666/tcp
>       - 16514/tcp
>     gluster_infra_fw_permanent: true
>     gluster_infra_fw_state: enabled
>     gluster_infra_fw_zone: public
>     gluster_infra_fw_services:
>       - glusterfs
>     gluster_features_force_varlogsizecheck: false
>     cluster_nodes:
>       - vmm10.mydomain.com
>       - vmm11.mydomain.com
>       - vmm12.mydomain.com
>     gluster_features_hci_cluster: '{{ cluster_nodes }}'
>     gluster_features_hci_volumes:
>       - volname: engine
>         brick: /gluster_bricks/engine/engine
>         arbiter: 0
>       - volname: vmstore1
>         brick: /gluster_bricks/vmstore1/vmstore1
>         arbiter: 0
>       - volname: data1
>         brick: /gluster_bricks/data1/data1
>         arbiter: 0
>       - volname: data2
>         brick: /gluster_bricks/data2/data2
>         arbiter: false
>
> ---------------------------------------------------------------------------------------------------------------------
>
>
>
> Thanks
> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/[email protected]/message/YYBK7FRRZJMHHGNCIB6UDU6FOXDHT74J/
>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/VOWVDSWP3NCCBN2HKJF4QWDCDTIVHS44/

Reply via email to