Shabbat Shalom,

On Wed, Aug 11, 2021 at 10:03 AM Yedidyah Bar David <[email protected]> wrote:

> On Tue, Aug 10, 2021 at 9:20 PM Gilboa Davara <[email protected]> wrote:
> >
> > Hello,
> >
> > Many thanks again for taking the time to try and help me recover this
> machine (even though it would have been far easier to simply redeploy it...)
> >
> >> >
> >> >
> >> > Sadly enough, it seems that --clean-metadata requires an active agent.
> >> > E.g.
> >> > $ hosted-engine --clean-metadata
> >> > The hosted engine configuration has not been retrieved from shared
> storage. Please ensure that ovirt-ha-agent
> >> > is running and the storage server is reachable.
> >>
> >> Did you try to search the net/list archives?
> >
> >
> > Yes. All of them seem to repeat the same clean-metadata command (which
> fails).
>
> I suppose we need better documentation. Sorry. Perhaps open a
> bug/issue about that.
>

Done.
https://bugzilla.redhat.com/show_bug.cgi?id=1993575


>
> >
> >>
> >>
> >> >
> >> > Can I manually delete the metadata state files?
> >>
> >> Yes, see e.g.:
> >>
> >> https://lists.ovirt.org/pipermail/users/2016-April/072676.html
> >>
> >> As an alternative to the 'find' command there, you can also find the
> IDs with:
> >>
> >> $ grep metadata /etc/ovirt-hosted-engine/hosted-engine.conf
> >>
> >> Best regards,
> >> --
> >> Didi
> >
> >
> > Yippie! Success (At least it seems that way...)
> >
> > Following https://lists.ovirt.org/pipermail/users/2016-April/072676.html
> ,
> > I stopped the broker and agent services, archived the existing hosted
> metadata files, created an empty 1GB metadata file using dd, (dd
> if=/dev/zero of=/run/vdsm/storage/<uuid>/<uuid> bs=1M count=1024), making
> double sure permissions (0660 / 0644), owner (vdsm:kvm) and SELinux labels
> (restorecon, just incase) stay the same.
> > Let everything settle down.
> > Restarted the services....
> > ... and everything is up again :)
> >
> > I plan to let the engine run overnight with zero VMs (making sure all
> backups are fully up-to-date).
> > Once done, I'll return to normal (until I replace this setup with a
> normal multi-node setup).
> >
> > Many thanks again!
>
> Glad to hear that, welcome, thanks for the report!
>
> More tests you might want to do before starting your real VMs:
>
> - Set and later clear global maintenance from each hosts, see that this
> propagates to the others (both 'hosted-engine --vm-status' and agent.log)
>
> - Migrate the engine VM between the hosts and see this propagates
>
> - Shutdown the engine VM without global maint and see that it's started
> automatically.
>
> But I do not think all of this is mandatory, if 'hosted-engine --vm-status'
> looks ok on all hosts.
>
> I'd still be careful with other things that might have been corrupted,
> though - obviously can't tell you what/where...
>
>
Host is back to normal.
The log looks clean (minus some odd smtp errors in the log).

Either way, I'm already in the process of replacing this setup with a real
3 host + gluster setup, so I just need this machine to survive the next
couple of weeks :)

- Gilboa
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/N4WIQXDW2ITLT2KDDH4LI4RTZRP2XWTY/

Reply via email to