btw, this happened on an aborted hosted-engine install. So, the ha-agents hadn't even started up.. just the VM running.
On Sat, Jul 19, 2014 at 11:24 PM, Andrew Lau <[email protected]> wrote: > Hi, > > Did anyone find much luck tracking this down? I rebooted one of our > servers and hit this issue again, conveniently, the dell remote access card > has borked as well.. so a 50 minute trip to the DC.. > > > On Thu, Jun 19, 2014 at 10:10 AM, Bob Doolittle <[email protected]> > wrote: > >> Specifically, if do the following: >> >> - Enter global maintenance (hosted-engine --set-maintenance-mode >> --mode=global) >> - init 0 the engine >> - systemctl stop ovirt-ha-agent ovirt-ha-broker libvirtd vdmsd >> >> >> and then run "sanlock client status" I see: >> >> # sanlock client status >> daemon c715b5de-fd98-4146-a0b1-e9801179c768.xion2.smar >> p -1 helper >> p -1 listener >> p -1 status >> s >> 003510e8-966a-47e6-a5eb-3b5c8a6070a9:1:/rhev/data-center/mnt/xion2.smartcity.net\:_export_VM__NewDataDomain/003510e8-966a-47e6-a5eb-3b5c8a6070a9/dom_md/ids:0 >> s >> 18eeab54-e482-497f-b096-11f8a43f94f4:1:/rhev/data-center/mnt/xion2\:_export_vm_he1/18eeab54-e482-497f-b096-11f8a43f94f4/dom_md/ids:0 >> s >> hosted-engine:1:/rhev/data-center/mnt/xion2\:_export_vm_he1/18eeab54-e482-497f-b096-11f8a43f94f4/ha_agent/hosted-engine.lockspace:0 >> >> >> Waiting a few minutes does not change this state. >> >> The earlier data I shared which showed HostedEngine was with a different >> test scenario. >> >> -Bob >> >> >> On 06/18/2014 07:53 AM, Bob Doolittle wrote: >> >> I see I have a very unfortunate typo in my previous mail. As supported by >> the vm-status output I attached, I had set --mode=global (not none) in step >> 1. >> >> I am not the only one experiencing this. I can reproduce it easily. It >> appears that shutting down vdsm causes the HA services to incorrectly think >> the system has come out of Global Maintenance and restart the engine. >> >> -Bob >> On Jun 18, 2014 5:06 AM, "Federico Simoncelli" <[email protected]> >> wrote: >> >>> ----- Original Message ----- >>> > From: "Bob Doolittle" <[email protected]> >>> > To: "Doron Fediuck" <[email protected]>, "Andrew Lau" < >>> [email protected]> >>> > Cc: "users" <[email protected]>, "Federico Simoncelli" < >>> [email protected]> >>> > Sent: Saturday, June 14, 2014 1:29:54 AM >>> > Subject: Re: [ovirt-users] Can HA Agent control NFS Mount? >>> > >>> > >>> > But there may be more going on. Even if I stop vdsmd, the HA services, >>> > and libvirtd, and sleep 60 seconds, I still see a lock held on the >>> > Engine VM storage: >>> > >>> > daemon 6f3af037-d05e-4ad8-a53c-61627e0c2464.xion2.smar >>> > p -1 helper >>> > p -1 listener >>> > p -1 status >>> > s 003510e8-966a-47e6-a5eb-3b5c8a6070a9:1:/rhev/data-center/mnt/ >>> xion2.smartcity.net >>> \:_export_VM__NewDataDomain/003510e8-966a-47e6-a5eb-3b5c8a6070a9/dom_md/ids:0 >>> > s >>> hosted-engine:1:/rhev/data-center/mnt/xion2\:_export_vm_he1/18eeab54-e482-497f-b096-11f8a43f94f4/ha_agent/hosted-engine.lockspace:0 >>> >>> This output shows that the lockspaces are still acquired. When you put >>> hosted-engine >>> in maintenance they must be released. >>> One by directly using rem_lockspace (since it's the hosted-engine one) >>> and the other >>> one by stopMonitoringDomain. >>> >>> I quickly looked at the ovirt-hosted-engine* projects and I haven't >>> found anything >>> related to that. >>> >>> -- >>> Federico >>> >> >> >
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

