On Wed, Sep 5, 2018 at 12:01 AM Matt Simonsen <[email protected]> wrote:
> Hello all,
>
> Following this report below, I did a reboot. Now I have a real question.
>
> I added the VG, LV and mount point to this node using the port 9090 web
> interface.
>
> Now the volume group isn't active and will not mount, causing the boot
> to hang.
>
> I am able to do "vgchange -ay data" and then a manual mount in rescue mode.
>
> Any feedback on the best way to add a new volume group to an empty
> partition (sdb) would be appreciated. Prior to using the web interface,
> I was having failures using the manual tools to /dev/sdb with an error
> "device /dev/sdb excluded by filter" which I suspect is related.
>
Maybe you have lvm filter set, which is highly recommend for an oVirt
hypervisor.
To add /dev/sdb, you need to add it to the lvm filter in /etc/lvm/lvm.conf.
After you configure the device properly, you can generate lvm filter
for the current setup using:
vdsm-tool config-lvm-filter
Here is example run on unconfigued oVirt host:
# vdsm-tool config-lvm-filter
Analyzing host...
Found these mounted logical volumes on this host:
logical volume: /dev/mapper/fedora_voodoo1-root
mountpoint: /
devices: /dev/vda2
logical volume: /dev/mapper/fedora_voodoo1-swap
mountpoint: [SWAP]
devices: /dev/vda2
This is the recommended LVM filter for this host:
filter = [ "a|^/dev/vda2$|", "r|.*|" ]
This filter allows LVM to access the local devices used by the
hypervisor, but not shared storage owned by Vdsm. If you add a new
device to the volume group, you will need to edit the filter manually.
Nir
On 09/04/2018 01:23 PM, Matt Simonsen wrote:
> > Hello,
> >
> > I'm running oVirt with several data centers, some with NFS storage and
> > some with local storage.
> >
> > I had problems in the past with a large pool and local storage. The
> > problem was nodectl showed the pool being too full (I think >80%), but
> > it was only the images that made the pool "full" -- and this storage
> > was carefully setup such that there was no chance it would actually
> > fill. The LVs for oVirt itself were all under 20%, yet nodectl still
> > reported the pool was too full.
> >
> > My solution so far has been to use our RAID card tools, so that sda is
> > the oVirt node install, and sdb is for images. There are probably
> > other good reasons for me to handle it this way, for example being
> > able to use different RAID levels, but I'm hoping someone can confirm
> > my partitioning below doesn't have some risk I'm now yet aware of.
> >
> > I setup a new volume group for images, as below:
> >
> >
> > [root@node4-g8-h4 multipath]# pvs
> > PV VG Fmt Attr PSize
> > PFree
> > /dev/mapper/3600508b1001c7e172160824d7b204c3b2 onn_node4-g8-h4 lvm2
> > a-- <119.00g <22.85g
> > /dev/sdb1 data lvm2 a-- 1.13t
> > <361.30g
> >
> > [root@node4-g8-h4 multipath]# vgs
> > VG #PV #LV #SN Attr VSize VFree
> > data 1 1 0 wz--n- 1.13t <361.30g
> > onn_node4-g8-h4 1 13 0 wz--n- <119.00g <22.85g
> >
> > [root@node4-g8-h4 multipath]# lvs
> > LV VG Attr LSize
> > Pool Origin Data% Meta% Move Log
> > Cpy%Sync Convert
> > images_main data -wi-ao---- 800.00g
> > home onn_node4-g8-h4 Vwi-aotz--
> > 1.00g pool00 4.79
> > ovirt-node-ng-4.2.5.1-0.20180816.0 onn_node4-g8-h4 Vwi---tz-k
> > 64.10g pool00 root
> > ovirt-node-ng-4.2.5.1-0.20180816.0+1 onn_node4-g8-h4 Vwi---tz--
> > 64.10g pool00 ovirt-node-ng-4.2.5.1-0.20180816.0
> > ovirt-node-ng-4.2.6-0.20180903.0 onn_node4-g8-h4 Vri---tz-k
> > 64.10g pool00
> > ovirt-node-ng-4.2.6-0.20180903.0+1 onn_node4-g8-h4 Vwi-aotz--
> > 64.10g pool00 ovirt-node-ng-4.2.6-0.20180903.0 4.83
> > pool00 onn_node4-g8-h4 twi-aotz--
> > 91.10g 8.94 0.49
> > root onn_node4-g8-h4 Vwi---tz--
> > 64.10g pool00
> > swap onn_node4-g8-h4 -wi-ao---- 4.00g
> > tmp onn_node4-g8-h4 Vwi-aotz--
> > 1.00g pool00 4.87
> > var onn_node4-g8-h4 Vwi-aotz--
> > 15.00g pool00 3.31
> > var_crash onn_node4-g8-h4 Vwi-aotz--
> > 10.00g pool00 2.86
> > var_log onn_node4-g8-h4 Vwi-aotz--
> > 8.00g pool00 3.57
> > var_log_audit onn_node4-g8-h4 Vwi-aotz--
> > 2.00g pool00 4.89
> >
> >
> >
> > The images_main is setup as "Block device for filesystems" with ext4.
> > Is there any reason I should consider pool for thinly provisioned
> > volumes? I don't need to over-allocate storage and it seems to me
> > like a fixed partition is ideal. Please confirm or let me know if
> > there's anything else I should consider.
> >
> >
> > Thanks
> >
> > Matt
> > _______________________________________________
> > Users mailing list -- [email protected]
> > To unsubscribe send an email to [email protected]
> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> > oVirt Code of Conduct:
> > https://www.ovirt.org/community/about/community-guidelines/
> > List Archives:
> >
> https://lists.ovirt.org/archives/list/[email protected]/message/I7N547X6DC7KHHVCDGKXQGNJV6TG7E3U/
> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/[email protected]/message/LJINANK6PAGVV22H5OTYTJ3M4WIWPTMV/
>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/[email protected]/message/N7M57G7HC46NBQTX6T3KSVHEYV3IDDIP/