On Fri, Jan 18, 2019 at 11:16 AM Sandro Bonazzola <[email protected]>
wrote:

>
>
> Il giorno mer 16 gen 2019 alle ore 23:27 Gianluca Cecchi <
> [email protected]> ha scritto:
>
>> Just installed a single host HCI with gluster, with only the engine vm
>> running.
>> Is it expected this situation below?
>>
>> # virsh -r list
>>  Id    Name                           State
>> ----------------------------------------------------
>>  2     HostedEngine                   running
>>
>> and
>> # virsh -r dumpxml 2
>> . . .
>>     <disk type='file' device='disk' snapshot='no'>
>>       <driver name='qemu' type='raw' cache='none' error_policy='stop'
>> io='native' iothread='1'/>
>>       <source
>> file='/var/run/vdsm/storage/e4eb6832-e0f6-40ee-902f-f301e5a3a643/fc34d770-9318-4539-9233-bfb1c5d68d14/b151557e-f1a2-45cb-b5c9-12c1f470467e'>
>>         <seclabel model='dac' relabel='no'/>
>>       </source>
>>       <backingStore/>
>>       <target dev='vda' bus='virtio'/>
>>       <serial>fc34d770-9318-4539-9233-bfb1c5d68d14</serial>
>>       <alias name='ua-fc34d770-9318-4539-9233-bfb1c5d68d14'/>
>>       <address type='pci' domain='0x0000' bus='0x00' slot='0x07'
>> function='0x0'/>
>>     </disk>
>> . . .
>>
>> where
>> # ll /var/run/vdsm/storage/e4eb6832-e0f6-40ee-902f-f301e5a3a643/
>> total 24
>> lrwxrwxrwx. 1 vdsm kvm 133 Jan 16 14:44
>> 39df7b45-4932-4bfe-b69e-4fb2f8872f4f ->
>> /rhev/data-center/mnt/glusterSD/10.10.10.216:
>> _engine/e4eb6832-e0f6-40ee-902f-f301e5a3a643/images/39df7b45-4932-4bfe-b69e-4fb2f8872f4f
>> lrwxrwxrwx. 1 vdsm kvm 133 Jan 16 14:44
>> 5ba6cd9e-b78d-4de4-9b7f-9688365128bf ->
>> /rhev/data-center/mnt/glusterSD/10.10.10.216:
>> _engine/e4eb6832-e0f6-40ee-902f-f301e5a3a643/images/5ba6cd9e-b78d-4de4-9b7f-9688365128bf
>> lrwxrwxrwx. 1 vdsm kvm 133 Jan 16 15:56
>> 8b8e41e0-a875-4204-8ab1-c10214a49f5c ->
>> /rhev/data-center/mnt/glusterSD/10.10.10.216:
>> _engine/e4eb6832-e0f6-40ee-902f-f301e5a3a643/images/8b8e41e0-a875-4204-8ab1-c10214a49f5c
>> lrwxrwxrwx. 1 vdsm kvm 133 Jan 16 15:56
>> c21a62ba-73d2-4914-940f-cee6a67a1b08 ->
>> /rhev/data-center/mnt/glusterSD/10.10.10.216:
>> _engine/e4eb6832-e0f6-40ee-902f-f301e5a3a643/images/c21a62ba-73d2-4914-940f-cee6a67a1b08
>> lrwxrwxrwx. 1 vdsm kvm 133 Jan 16 14:44
>> fc34d770-9318-4539-9233-bfb1c5d68d14 ->
>> /rhev/data-center/mnt/glusterSD/10.10.10.216:
>> _engine/e4eb6832-e0f6-40ee-902f-f301e5a3a643/images/fc34d770-9318-4539-9233-bfb1c5d68d14
>> lrwxrwxrwx. 1 vdsm kvm 133 Jan 16 14:44
>> fd73354d-699b-478e-893c-e2a0bd1e6cbb ->
>> /rhev/data-center/mnt/glusterSD/10.10.10.216:
>> _engine/e4eb6832-e0f6-40ee-902f-f301e5a3a643/images/fd73354d-699b-478e-893c-e2a0bd1e6cbb
>>
>> so hosted engine not using libgfapi?
>> Also, on hosted engine
>>
>> [root@hciengine ~]# engine-config -g LibgfApiSupported
>> LibgfApiSupported: false version: 4.1
>> LibgfApiSupported: false version: 4.2
>> LibgfApiSupported: false version: 4.3
>> [root@hciengine ~]#
>>
>> So that if I import a CentOS7 Atomic host image from glance repo as
>> template and create a new vm from it,when running this VM I get
>>
>> # virsh -r dumpxml 3
>> . . .
>>     <disk type='file' device='disk' snapshot='no'>
>>       <driver name='qemu' type='qcow2' cache='none' error_policy='stop'
>> io='native' iothread='1'/>
>>       <source file='/rhev/data-center/mnt/glusterSD/10.10.10.216:
>> _data/601d725a-1622-4dc8-a24d-2dba72ddf6ae/images/e4f92226-0f56-4822-a622-d1ebff41df9f/c6b2e076-1519-433e-9b37-2005c9ce6d2e'>
>>         <seclabel model='dac' relabel='no'/>
>>       </source>
>>       <backingStore/>
>>       <target dev='vda' bus='virtio'/>
>>       <serial>e4f92226-0f56-4822-a622-d1ebff41df9f</serial>
>>       <boot order='1'/>
>>       <alias name='ua-e4f92226-0f56-4822-a622-d1ebff41df9f'/>
>>       <address type='pci' domain='0x0000' bus='0x00' slot='0x06'
>> function='0x0'/>
>>     </disk>
>> . . .
>>
>> I remember there was an "old" bug opened causing this default of not
>> enabling libgfapi
>> Does this mean it was not solved yet?
>> If I remember correctly the bugzilla was this one related to HA:
>> https://bugzilla.redhat.com/show_bug.cgi?id=1484227
>> that is still in new status.... since almost 2 years
>>
>> Is this the only one open?
>>
>
> Thanks Gianluca for the deep testing and analysis! Sahina, Simone, can you
> please check this?
>

Yes, AFAIK we are still not ready to support libgfapi.


>
>
>
>>
>>
>> Thanks,
>> Gianluca
>> _______________________________________________
>> Users mailing list -- [email protected]
>> To unsubscribe send an email to [email protected]
>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>> oVirt Code of Conduct:
>> https://www.ovirt.org/community/about/community-guidelines/
>> List Archives:
>> https://lists.ovirt.org/archives/list/[email protected]/message/KHWQA72JZVGLVXCIQPU2FDFRS73BW4LY/
>>
>
>
> --
>
> SANDRO BONAZZOLA
>
> MANAGER, SOFTWARE ENGINEERING, EMEA R&D RHV
>
> Red Hat EMEA <https://www.redhat.com/>
>
> [email protected]
> <https://red.ht/sig>
>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/UOB3WQOZBYJMKJO67O2PJ437JU73KKGG/

Reply via email to