On Sun, Dec 24, 2017 at 12:41 AM, Gabriel Stein <[email protected]> wrote:
> Well, I finally upgraded all to 4.2. Unfortunately I broke a Server in the > upgrade process and I needed much more time than expected(the Host was with > Hosted Engine and Gluster). I will not go further on that problems because > I interrupted the upgrade process and try to fix it and I ended with a > kernel panic. > > I recommend to use tmux or screen for the upgrade. > > My experience: > > I used this tutorial for the upgrade process: https://ovirt.org/ > documentation/how-to/hosted-engine/#upgrade-hosted-engine > > You can use it for the Hosted Engine VM and the Hosts. Be patient, don't > interrupt the yum update process, and follow the instructions. > If you have a locale different than US-UTF8, please change it for US-UTF8 > before the upgrade process. Where? /etc/locale.conf on CentOS. If you are > using e.g. puppet, please deactivate it on upgrade process, to avoid the > locale change(if you have something on puppet that changes it). > > Problems: > > I still having problems with glusterfs(Peer Rejected) and unstable, but it > probably happens because I copied the UUID from the broke server from > glusterfs and added the new installed server with same ip back to gluster. > Are your problems with GlusterFS resolved. If not please reach out to us and [email protected] for help > > IMHO: > Please check that you have the engine backup done. Save it somewhere, NFS, > rsync-it to another server.... > When running engine-setup after the yum update on ovirt-engine VM: don't > to the in place upgrade from postgresql. It's really nice to have it, but > you you can avoid risks, why do so? > Keep the Backup from Postgresql. > > Previous Version: 4.1.x > Updated to: 4.2 > > Setup: > > CentOS 7.4.108 > 4 Servers, 3 with Gluster for engine. > > If you have questions.... > > Best Regards, > > Gabriel > > > Gabriel Stein > ------------------------------ > Gabriel Ferraz Stein > Tel.: +49 (0) 170 2881531 > > 2017-12-22 12:19 GMT+01:00 Gianluca Cecchi <[email protected]>: > >> On Fri, Dec 22, 2017 at 12:06 PM, Martin Perina <[email protected]> >> wrote: >> >>> >>> >>>>> >>>>> Only problem I registered was the first start of engine on the first >>>>> upgraded host (that should be step 7 in link above), where I got error on >>>>> engine (not able to access web admin portal); I see this in server.log >>>>> (see full attach here https://drive.google.com/file/ >>>>> d/1UQAllZfjueVGkXDsBs09S7THGDFn9YPa/view?usp=sharing ) >>>>> >>>>> 2017-12-22 00:40:17,674+01 INFO [org.quartz.core.QuartzScheduler] >>>>> (ServerService Thread Pool -- 63) Scheduler >>>>> DefaultQuartzScheduler_$_NON_CLUSTERED >>>>> started. >>>>> 2017-12-22 00:40:17,682+01 INFO [org.jboss.as.clustering.infinispan] >>>>> (ServerService Thread Pool -- 63) WFLYCLINF0002: Started timeout-base >>>>> cache >>>>> from ovirt-engine container >>>>> 2017-12-22 00:44:28,611+01 ERROR >>>>> [org.jboss.as.controller.management-operation] >>>>> (Controller Boot Thread) WFLYCTL0348: Timeout after [300] seconds waiting >>>>> for service container stability. Operation will roll back. Step that first >>>>> updated the service container was 'add' at address '[ >>>>> ("core-service" => "management"), >>>>> ("management-interface" => "native-interface") >>>>> ]' >>>>> >>>> >>>> Adding Vaclav, maybe something in Wildfly? Martin, any hint on engine >>>> side? >>>> >>> >>> ​Yeah, I've already seen such error a few times, it usually happens when >>> access to storage is really slow or the host itself is overloaded and >>> WildFly is not able to startup properly until default 300 seconds interval​ >>> is over. >>> >>> If this is going to happen often, we will have to raise that timeout for >>> all installations. >>> >>> >> Ok, thanks for the confirmation of what I suspected. >> Actually this particular environment is based on a single NUC where I >> have ESXi 6.0U2 and the 3 hosts are actually VMs of this vSphere >> environment, so that the hosted engine is an L2 VM >> And the replica 3 (with arbiter) of the hosts insists at the end on a >> single physical SSD disk below... >> All in all it is already a great success that this kind of infrastructure >> has been able to update from 4.1 to 4.2... I use it basically as a >> functional testing and btw on vSphere there is also another CentOS 7 VM >> running ;-) >> >> Thanks, >> Gianluca >> >> _______________________________________________ >> Users mailing list >> [email protected] >> http://lists.ovirt.org/mailman/listinfo/users >> >> > > _______________________________________________ > Users mailing list > [email protected] > http://lists.ovirt.org/mailman/listinfo/users > >
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

