Thank you for help Strahil,

But although there where 4 images with a status 4 in the database, and did
the update query on them, same bloody message, and vms wont start.
Eventually, ive decided to delete the vms, and do a from scratch
installation. Persistance openshift vms are still ok, so i should be able
to reuse the volumes somehow.
This is why sometimes a subscription is good, when a lack of knowledge on
my side is happening. Production systems should not rely on upstreams
unless there is a strong understanding of the product.
Again, thank you so much for trying helping me out !
Cheers,

Leo

On Tue, Oct 15, 2019, 07:00 Leo David <[email protected]> wrote:

> Thank you Strahil,
> I'll proceed with these steps and come back to you.
> Cheers,
>
> Leo
>
> On Tue, Oct 15, 2019, 06:45 Strahil <[email protected]> wrote:
>
>> Have you checked this thread :
>> https://lists.ovirt.org/pipermail/users/2016-April/039277.html
>>
>> You can switch to postgre user, then 'source
>> /opt/rhn/postgresql10/enable' & then 'psql engine'.
>>
>> As per the thread you can find illegal snapshots via '*select
>> image_group_id,imagestatus from images where imagestatus =4;*'
>>
>> And then update them via '*update images set imagestatus =1 where
>> imagestatus = 4 and <other **criteria>;** commit'*
>>
>> Best Regards,
>> Strahil Nikolov
>>
>> On Oct 13, 2019 15:45, Leo David <[email protected]> wrote:
>>
>> >
>> > Hi Everyone,
>> > Im still not being able to start the vms... Could anyone give me an
>> advice on sorign this out ?
>> > Still having th "Bad volume specification" error,  although the disk is
>> present on the storage.
>> > This issue would force me to reinstall a 10 nodes Openshift cluster
>> from scratch,  which would not be so funny..
>> > Thanks,
>> >
>> > Leo.
>> >
>> > On Fri, Oct 11, 2019 at 7:12 AM Strahil <[email protected]> wrote:
>>
>> >>
>> >> Nah...
>> >> It's done directly on the DB and I wouldn't recommend such action for
>> Production Cluster.
>> >> I've done it only once and it was based on some old mailing lists.
>> >>
>> >> Maybe someone from the dev can assist?
>> >>
>> >> On Oct 10, 2019 13:31, Leo David <[email protected]> wrote:
>>
>> >>>
>> >>> Thank you Strahil,
>> >>> Could you tell me what do you mean by changing status ? Is this
>> something to be done in the UI ?
>> >>>
>> >>> Thanks,
>> >>>
>> >>> Leo
>> >>>
>> >>> On Thu, Oct 10, 2019, 09:55 Strahil <[email protected]> wrote:
>>
>> >>>>
>> >>>> Maybe you can change the status of the VM in order the engine to
>> know that it has to blockcommit the snapshots.
>> >>>>
>> >>>> Best Regards,
>> >>>> Strahil Nikolov
>> >>>>
>> >>>> On Oct 9, 2019 09:02, Leo David <[email protected]> wrote:
>>
>> >>>>>
>> >>>>> Hi Everyone,
>> >>>>> Please let me know if any thoughts or recommandations that could
>> help me solve this issue..
>> >>>>> The real bad luck in this outage is that these 5 vms are part on an
>> Openshift deployment,  and now we are not able to start it up...
>> >>>>> Before trying to sort this at ocp platform level by replacing the
>> failed nodes with new vms, I would rather prefer to do it at the oVirt
>> level and have the vms starting since the disks are still present on
>> gluster.
>> >>>>> Thank you so much !
>> >>>>>
>> >>>>>
>> >>>>> Leo
>>
>> >
>> >
>> >
>> > --
>> > Best regards, Leo David
>>
>>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/WYE2EO4AOCTWK4EWGMDQ7KSTF3M6JR6Q/

Reply via email to