Hi David, Thanks for reply, I have Ovirt env which is in very bad shape, and as a new employer I have to fix it :). I have 5 hosts in whole env. There is broken HA for hosted engine, so engine can only works on host1. I can't add another host because it shows me that it was deployed with version 3.5 (If I'm not wrong it's fixed in 4.0). Also I can't update/upgrade ovirt-engine because there is 500MB of free space (after cleaning up) without LVM, so I'm afraid that I'll run out off space during update. Because of that I decided to add completely new server and migrate hosted-engine on fixed HE (with LVM) and properly configured HA on new host.
Below short summary: Hosted engine: CentOS Linux release 7.2.1511 (Core) ovirt-engine-3.6.7.5-1.el7.centos.noarch Host0: with running hosted Hosted-Engine which I need to update/upgrade CentOS Linux release 7.2.1511 (Core) ovirt-vmconsole-1.0.0-1.el7.centos.noarch ovirt-hosted-engine-ha-1.3.5.1-1.el7.centos.noarch ovirt-host-deploy-1.4.1-1.el7.centos.noarch ovirt-release36-007-1.noarch ovirt-engine-sdk-python-3.6.3.0-1.el7.centos.noarch ovirt-image-uploader-3.6.0-1.el7.centos.noarch ovirt-setup-lib-1.0.1-1.el7.centos.noarch libgovirt-0.3.3-1.el7_2.1.x86_64 ovirt-vmconsole-host-1.0.0-1.el7.centos.noarch ovirt-hosted-engine-setup-1.3.4.0-1.el7.centos.noarch ovirt-engine-appliance-3.6-20160301.1.el7.centos.noarch vdsm-jsonrpc-4.17.23.2-0.el7.centos.noarch vdsm-yajsonrpc-4.17.23.2-0.el7.centos.noarch vdsm-4.17.23.2-0.el7.centos.noarch vdsm-python-4.17.23.2-0.el7.centos.noarch vdsm-infra-4.17.23.2-0.el7.centos.noarch vdsm-hook-vmfex-dev-4.17.23.2-0.el7.centos.noarch vdsm-xmlrpc-4.17.23.2-0.el7.centos.noarch vdsm-cli-4.17.23.2-0.el7.centos.noarch Output from hosted-engine --vm-status from host1 --== Host 1 status ==-- Status up-to-date : True Hostname : dev-ovirtnode0.example.com Host ID : 1 Engine status : {"health": "good", "vm": "up", "detail": "up"} Score : 3400 stopped : False Local maintenance : False crc32 : d7fdf8b6 Host timestamp : 1243846 --== Host 2 status ==-- <- it's garbage because there is not installed and configured HA on host2 Status up-to-date : False Hostname : dev-ovirtnode1.example.com Host ID : 2 Engine status : unknown stale-data Score : 0 stopped : True Local maintenance : False crc32 : fb5f379e Host timestamp : 563 Remaining Hosts1 - 4 are updated and configured in the same way (it was done by me). I had to replace network cards and now there is LACP on 4x10G cards (before was only 1G card). Because there is CentOS 7.4 I decided to install vdsm in version 4.17.43 (from repo) to fix bugs. I aware that 3.6 is only supported with verision 7.2, but I want to update whole env to 3.6.X then to 4.0. then to 4.1 to be up to date. vdsm-jsonrpc-4.17.43-1.el7.centos.noarch vdsm-xmlrpc-4.17.43-1.el7.centos.noarch vdsm-4.17.43-1.el7.centos.noarch vdsm-infra-4.17.43-1.el7.centos.noarch vdsm-yajsonrpc-4.17.43-1.el7.centos.noarch vdsm-cli-4.17.43-1.el7.centos.noarch vdsm-python-4.17.43-1.el7.centos.noarch vdsm-hook-vmfex-dev-4.17.43-1.el7.centos.noarch On those host1-4 I have around 400 vm's for used by developers, and I need to shorten downtime as much as possible (the best option is without downtime, but I'm not sure if it's possible). I decided to restore HE on completely new host because I believe that in my case it's the easiest way to update then upgrade the whole env :) Many thanks for all advises Regards Krzysztof 2017-11-14 8:50 GMT+01:00 Yedidyah Bar David <d...@redhat.com>: > On Mon, Nov 13, 2017 at 11:58 PM, Krzysztof Wajda <vajdov...@gmail.com> > wrote: > > Hello, > > > > I have to restore Hosted Engine on another host (completely new > hardware). > > Based on this > > https://www.ovirt.org/documentation/self-hosted/ > chap-Backing_up_and_Restoring_an_EL-Based_Self-Hosted_Environment/ > > is not clear for me if vm's will be rebooted during synchronization hosts > > with engine ? > > They should not be rebooted automatically, but you might need to do > this yourself, see below. > > > > > I have 5 hosts + 1 completely fresh. On host1 I have HE and there is no > vm's > > on other 4 (host1-4) there are around 400 vm which can't be rebooted. > Host5 > > for restore HE. > > Please provide more details about your backup/restore flow. > What died (storage? hosts? data?), what are you going to restore, > how, etc. > > Which hosts are hosted-engine hosts. Do they have running VMs. > > We are working on updating the documentation, but it will take some time. > > For now, you should assume that the safest way is to pass during restore, > to engine-backup, '--he-remove-storage-vm' and '--he-remove-hosts'. This > will remove from the engine all the hosted-engine hosts and storage. So > when you add the hosts back, you'll have to somehow power off the VMs > there - the engine will refuse to add them with running VMs. If you do > not want to use these options, you should plan carefully and test. > > See also: > > https://bugzilla.redhat.com/show_bug.cgi?id=1235200 > https://bugzilla.redhat.com/show_bug.cgi?id=1240466 > https://bugzilla.redhat.com/show_bug.cgi?id=1441322 > > Best regards, > -- > Didi >
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users