I would recommend to do cleanup from cockpit or if you are using cli based deployment then use "/etc/ansible/roles/gluster.ansible/playbooks/hc-ansible-deployment/tasks/gluster_cleanup.yml" with your inventory. Then try to deploy again. Cleanup takes care everything.
On Thu, Apr 30, 2020 at 9:59 PM Strahil Nikolov <[email protected]> wrote: > On April 30, 2020 12:31:59 PM GMT+03:00, Shareef Jalloq < > [email protected]> wrote: > >Changing to /dev/mapper names seems to work but if anyone can tell me > >why > >the /dev/sd* naming is filtered that would help my understanding. > > > >On Thu, Apr 30, 2020 at 10:13 AM Shareef Jalloq <[email protected]> > >wrote: > > > >> Having no luck here. I've had a read on the LVM config usage and > >there > >> were no filters enabled in lvm.conf. I enabled debug logging and can > >see > >> the default global filter being applied. I then manually forced the > >'all' > >> fiter and 'pvcreate /dev/sdb' still tells me it is excluded by a > >filter. > >> The device section from 'lvmconfig' follows. What's wrong? > >> > >> devices { > >> > >> dir="/dev" > >> > >> scan="/dev" > >> > >> obtain_device_list_from_udev=1 > >> > >> external_device_info_source="none" > >> > >> preferred_names=["^/dev/mpath/","^/dev/mapper/mpath","^/dev/[hs]d"] > >> > >> filter="a|.*/|" > >> > >> cache_dir="/etc/lvm/cache" > >> > >> cache_file_prefix="" > >> > >> write_cache_state=1 > >> > >> sysfs_scan=1 > >> > >> scan_lvs=0 > >> > >> multipath_component_detection=1 > >> > >> md_component_detection=1 > >> > >> fw_raid_component_detection=0 > >> > >> md_chunk_alignment=1 > >> > >> data_alignment_detection=1 > >> > >> data_alignment=0 > >> > >> data_alignment_offset_detection=1 > >> > >> ignore_suspended_devices=0 > >> > >> ignore_lvm_mirrors=1 > >> > >> disable_after_error_count=0 > >> > >> require_restorefile_with_uuid=1 > >> > >> pv_min_size=2048 > >> > >> issue_discards=0 > >> > >> allow_changes_with_duplicate_pvs=1 > >> > >> } > >> > >> On Wed, Apr 29, 2020 at 6:21 PM Shareef Jalloq <[email protected]> > >> wrote: > >> > >>> Actually, now I've fixed that, indeed, the deployment now fails with > >an > >>> lvm filter error. I'm not familiar with filters but there aren't > >any > >>> uncommented instances of 'filter' in /etc/lvm/lvm.conf. > >>> > >>> > >>> > >>> On Wed, Apr 29, 2020 at 5:42 PM Shareef Jalloq > ><[email protected]> > >>> wrote: > >>> > >>>> Ah of course. I was assuming something had gone wrong with the > >>>> deployment and it couldn't clean up its own mess. I'll raise a bug > >on the > >>>> documentation. > >>>> > >>>> Strahil, what are the other options to using /dev/sdxxx? > >>>> > >>>> On Wed, Apr 29, 2020 at 10:17 AM Strahil Nikolov > ><[email protected]> > >>>> wrote: > >>>> > >>>>> On April 29, 2020 2:39:05 AM GMT+03:00, Jayme <[email protected]> > >wrote: > >>>>> >Has the drive been used before, it might have existing > >>>>> >partition/filesystem > >>>>> >on it? If you are sure it's fine to overwrite try running wipefs > >-a > >>>>> >/dev/sdb on all hosts. Also make sure there aren't any filters > >setup in > >>>>> >lvm.conf (there shouldn't be on fresh install, but worth > >checking). > >>>>> > > >>>>> >On Tue, Apr 28, 2020 at 8:22 PM Shareef Jalloq > ><[email protected]> > >>>>> >wrote: > >>>>> > > >>>>> >> Hi, > >>>>> >> > >>>>> >> I'm running the gluster deployment flow and am trying to use a > >second > >>>>> >> drive as the gluster volume. It's /dev/sdb on each node and > >I'm > >>>>> >using the > >>>>> >> JBOD mode. > >>>>> >> > >>>>> >> I'm seeing the following gluster ansible task fail and a google > >>>>> >search > >>>>> >> doesn't bring up much. > >>>>> >> > >>>>> >> TASK [gluster.infra/roles/backend_setup : Create volume groups] > >>>>> >> **************** > >>>>> >> > >>>>> >> failed: [ovirt-gluster-01.jalloq.co.uk] (item={u'vgname': > >>>>> >> u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) => > >{"ansible_loop_var": > >>>>> >"item", > >>>>> >> "changed": false, "err": " Couldn't find device with uuid > >>>>> >> Y8FVs8-LP6w-R6CR-Yosh-c40j-17XP-ttP3Np.\n Couldn't find device > >with > >>>>> >uuid > >>>>> >> tA4lpO-hM9f-S8ci-BdPh-lTve-0Rh1-3Bcsfy.\n Couldn't find device > >with > >>>>> >uuid > >>>>> >> RG3w6j-yrxn-2iMw-ngd0-HgMS-i5dP-CGjaRk.\n Couldn't find device > >with > >>>>> >uuid > >>>>> >> lQV02e-TUZE-PXCd-GWEd-eGqe-c2xC-pauHG7.\n Device /dev/sdb > >excluded > >>>>> >by a > >>>>> >> filter.\n", "item": {"pvname": "/dev/sdb", "vgname": > >>>>> >"gluster_vg_sdb"}, > >>>>> >> "msg": "Creating physical volume '/dev/sdb' failed", "rc": 5} > >>>>> >> _______________________________________________ > >>>>> >> Users mailing list -- [email protected] > >>>>> >> To unsubscribe send an email to [email protected] > >>>>> >> Privacy Statement: https://www.ovirt.org/privacy-policy.html > >>>>> >> oVirt Code of Conduct: > >>>>> >> https://www.ovirt.org/community/about/community-guidelines/ > >>>>> >> List Archives: > >>>>> >> > >>>>> > > >>>>> > > > https://lists.ovirt.org/archives/list/[email protected]/message/5U3K3IPYCFOLUFJ56FGJI3TYWT6NOLAZ/ > >>>>> >> > >>>>> > >>>>> Actually best practice is not to use /dev/sdxxx as they can > >change. > >>>>> > >>>>> In your case most peobably the LUN is not fresh, so wipe it with > >>>>> dd/blktrim so any remnants of old FS signature is gone. > >>>>> > >>>>> Best Regards, > >>>>> Strahil Nikolov > >>>>> > >>>> > > Do you use multipath or VDO ? > In the Linux stack you should always use the top layer device. > > For example you might have something like this (formatting is hard to be > done over e-mail) > LV > Lool LV > VG > PV > /dev/mapper/VDO1 > /dev/md0 > /dev/mapper/mpatha > /dev/sdb > > As you can see , in that case sdb is a path for mpatha , which is a > raid device and that raid device is VDO device. > > In your case it could be something else. > > Best Regards, > Strahil Nikolov > _______________________________________________ > Users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > Privacy Statement: https://www.ovirt.org/privacy-policy.html > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/NO5XINMD4YLU5CE4MLJ5MC6EVTWGPVCL/ > -- Thanks, Gobinda
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/2EMBJJL7IDS7236TDL43MLPYDOE6YCZL/

