pod deletion should invoke unpublish the PVC which detaches it from
the node which is seen in the engine log:
2021-08-20 17:40:35,664+09 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand]
(default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] START,
HotUnPlugDiskVDSCommand(HostName = host,
HotPlugDiskVDSParameters:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094',
vmId='59a7461c-72fe-4e01-86a7-c70243f31596',
diskId='63a64445-1659-4d5f-8847-e7266e64b09e'}), log id: 506ff4a4
2021-08-20 17:40:35,678+09 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand]
(default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] Disk
hot-unplug: <?xml version="1.0" encoding="UTF-8"?><hotunplug>
<devices>
<disk>
<alias name="ua-63a64445-1659-4d5f-8847-e7266e64b09e"/>
</disk>
</devices>
</hotunplug>
2021-08-20 17:40:35,749+09 INFO
[org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand]
(default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] FINISH,
HotUnPlugDiskVDSCommand, return: , log id: 506ff4a4
2021-08-20 17:40:35,842+09 INFO
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] EVENT_ID:
USER_DETACH_DISK_FROM_VM(2,018), Disk
pvc-9845a0ff-e94c-497c-8c65-fc6a1e26db20 was successfully detached
from VM centos by admin@internal-authz.
I suspect something keeps the volume busy, can run:
$ rbd status <pool_name>/volume-63a64445-1659-4d5f-8847-e7266e64b09e
On Mon, Aug 23, 2021 at 3:56 AM <[email protected]> wrote:
>
> Hi all,
>
> I deployed ovirt-csi in the k8s by applying yaml manually. I used the latest
> version of the container image.
> (https://github.com/openshift/ovirt-csi-driver-operator/tree/master/assets)
>
> After successfully creating pvc and pod, I tried to delete it.
> And the pod is deleted, but the pvc is not deleted. This is because deleting
> a pod does not unmap /dev/rbd0 attached to the ovirt vm.
>
> How can I delete the pvc successfully?
>
> oVirt engine version is 4.4.7.6-1.el8.
> Here is the engine log when deleting the pod:
>
> 2021-08-20 17:40:35,385+09 INFO
> [org.ovirt.engine.core.sso.service.AuthenticationService] (default task-149)
> [] User admin@internal-authz with profile [internal] successfully logged in
> with scopes: ovirt-app-api ovirt-ext=token-info:authz-search
> ovirt-ext=token-info:public-authz-search ovirt-ext=token-info:validate
> ovirt-ext=token:password-access
> 2021-08-20 17:40:35,403+09 INFO
> [org.ovirt.engine.core.bll.aaa.CreateUserSessionCommand] (default task-149)
> [68ee3182] Running command: CreateUserSessionCommand internal: false.
> 2021-08-20 17:40:35,517+09 INFO
> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> (default task-149) [68ee3182] EVENT_ID: USER_VDC_LOGIN(30), User
> admin@internal-authz connecting from '192.168.7.169' using session
> 'XfDgNkmAGnPiZahK5itLhHQTCNHZ3JwXMMzOiZrYL3C32+1TTys3xcjrAmCIKPu02hgN1sdVpfZXWd0FznaPCQ=='
> logged in.
> 2021-08-20 17:40:35,520+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
> return null
> 2021-08-20 17:40:35,520+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
> return null
> 2021-08-20 17:40:35,520+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
> return null
> 2021-08-20 17:40:35,520+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
> return null
> 2021-08-20 17:40:35,520+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
> return null
> 2021-08-20 17:40:35,520+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
> return null
> 2021-08-20 17:40:35,663+09 INFO
> [org.ovirt.engine.core.bll.storage.disk.DetachDiskFromVmCommand] (default
> task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] Running command:
> DetachDiskFromVmCommand internal: false. Entities affected : ID:
> 59a7461c-72fe-4e01-86a7-c70243f31596 Type: VMAction group
> CONFIGURE_VM_STORAGE with role type USER
> 2021-08-20 17:40:35,664+09 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default
> task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] START,
> HotUnPlugDiskVDSCommand(HostName = host,
> HotPlugDiskVDSParameters:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094',
> vmId='59a7461c-72fe-4e01-86a7-c70243f31596',
> diskId='63a64445-1659-4d5f-8847-e7266e64b09e'}), log id: 506ff4a4
> 2021-08-20 17:40:35,678+09 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default
> task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] Disk hot-unplug: <?xml
> version="1.0" encoding="UTF-8"?><hotunplug>
> <devices>
> <disk>
> <alias name="ua-63a64445-1659-4d5f-8847-e7266e64b09e"/>
> </disk>
> </devices>
> </hotunplug>
>
> 2021-08-20 17:40:35,749+09 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default
> task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] FINISH,
> HotUnPlugDiskVDSCommand, return: , log id: 506ff4a4
> 2021-08-20 17:40:35,842+09 INFO
> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> (default task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] EVENT_ID:
> USER_DETACH_DISK_FROM_VM(2,018), Disk
> pvc-9845a0ff-e94c-497c-8c65-fc6a1e26db20 was successfully detached from VM
> centos by admin@internal-authz.
> 2021-08-20 17:40:35,916+09 ERROR
> [org.ovirt.engine.core.sso.service.SsoService] (default task-150) []
> OAuthException invalid_grant: The provided authorization grant for the auth
> code has expired.
> 2021-08-20 17:40:35,917+09 ERROR
> [org.ovirt.engine.core.aaa.filters.SsoRestApiAuthFilter] (default task-149)
> [] Cannot authenticate using authentication Headers: invalid_grant: The
> provided authorization grant for the auth code has expired.
> 2021-08-20 17:40:36,029+09 INFO
> [org.ovirt.engine.core.sso.service.AuthenticationService] (default task-149)
> [] User admin@internal-authz with profile [internal] successfully logged in
> with scopes: ovirt-app-api ovirt-ext=token-info:authz-search
> ovirt-ext=token-info:public-authz-search ovirt-ext=token-info:validate
> ovirt-ext=token:password-access
> 2021-08-20 17:40:36,046+09 INFO
> [org.ovirt.engine.core.bll.aaa.CreateUserSessionCommand] (default task-149)
> [4c4bf441] Running command: CreateUserSessionCommand internal: false.
> 2021-08-20 17:40:36,114+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
> return null
> 2021-08-20 17:40:36,114+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
> return null
> 2021-08-20 17:40:36,114+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
> return null
> 2021-08-20 17:40:36,114+09 WARN
> [org.ovirt.engine.api.restapi.util.LinkHelper] (default task-149) [] Can't
> find relative path for class
> "org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will
> return null
> 2021-08-20 17:40:49,241+09 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.DumpXmlsVDSCommand]
> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-27) []
> START, DumpXmlsVDSCommand(HostName = host,
> Params:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094',
> vmIds='[59a7461c-72fe-4e01-86a7-c70243f31596]'}), log id: 7eb54202
> 2021-08-20 17:40:49,244+09 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.DumpXmlsVDSCommand]
> (EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-27) []
> FINISH, DumpXmlsVDSCommand, return:
> {59a7461c-72fe-4e01-86a7-c70243f31596=<domain type='kvm' id='1'
> xmlns:qemu='http://libvirt.org/schemas/domain/qemu/1.0'>
> <name>centos</name>
> <uuid>59a7461c-72fe-4e01-86a7-c70243f31596</uuid>
> <metadata xmlns:ns1="http://ovirt.org/vm/tune/1.0"
> xmlns:ovirt-vm="http://ovirt.org/vm/1.0">
> <ns1:qos/>
> <ovirt-vm:vm xmlns:ovirt-vm="http://ovirt.org/vm/1.0">
> <ovirt-vm:balloonTarget type="int">4194304</ovirt-vm:balloonTarget>
> <ovirt-vm:ballooningEnabled>true</ovirt-vm:ballooningEnabled>
> <ovirt-vm:clusterVersion>4.6</ovirt-vm:clusterVersion>
> <ovirt-vm:destroy_on_reboot type="bool">False</ovirt-vm:destroy_on_reboot>
> <ovirt-vm:launchPaused>false</ovirt-vm:launchPaused>
> <ovirt-vm:memGuaranteedSize type="int">4096</ovirt-vm:memGuaranteedSize>
> <ovirt-vm:minGuaranteedMemoryMb
> type="int">4096</ovirt-vm:minGuaranteedMemoryMb>
> <ovirt-vm:resumeBehavior>auto_resume</ovirt-vm:resumeBehavior>
> <ovirt-vm:startTime type="float">1628558564.8754532</ovirt-vm:startTime>
> <ovirt-vm:device alias="ua-7c9f38e9-8889-46c8-83bb-92efb9272de9"
> mac_address="56:6f:16:a8:00:07">
> <ovirt-vm:network>ovirtmgmt</ovirt-vm:network>
> <ovirt-vm:custom>
> <ovirt-vm:queues>2</ovirt-vm:queues>
> </ovirt-vm:custom>
> </ovirt-vm:device>
> <ovirt-vm:device devtype="disk" name="sda">
>
> <ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
> <ovirt-vm:guestName>/dev/sda</ovirt-vm:guestName>
>
> <ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
> <ovirt-vm:managed type="bool">False</ovirt-vm:managed>
>
> <ovirt-vm:poolID>4ca6e0e8-e3a4-11eb-8830-480fcf63834f</ovirt-vm:poolID>
>
> <ovirt-vm:volumeID>a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:volumeID>
> <ovirt-vm:volumeChain>
> <ovirt-vm:volumeChainNode>
>
> <ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
>
> <ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
> <ovirt-vm:leaseOffset type="int">0</ovirt-vm:leaseOffset>
>
> <ovirt-vm:leasePath>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003.lease</ovirt-vm:leasePath>
>
> <ovirt-vm:path>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003</ovirt-vm:path>
>
> <ovirt-vm:volumeID>4425404b-17f0-4519-812f-b09a952a9003</ovirt-vm:volumeID>
> </ovirt-vm:volumeChainNode>
> <ovirt-vm:volumeChainNode>
>
> <ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
>
> <ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
> <ovirt-vm:leaseOffset type="int">0</ovirt-vm:leaseOffset>
>
> <ovirt-vm:leasePath>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4.lease</ovirt-vm:leasePath>
>
> <ovirt-vm:path>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:path>
>
> <ovirt-vm:volumeID>a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:volumeID>
> </ovirt-vm:volumeChainNode>
> </ovirt-vm:volumeChain>
> </ovirt-vm:device>
> <ovirt-vm:device devtype="disk" name="sdc">
> <ovirt-vm:managed type="bool">False</ovirt-vm:managed>
> </ovirt-vm:device>
> </ovirt-vm:vm>
> </metadata>
> <maxMemory slots='16' unit='KiB'>16777216</maxMemory>
> <memory unit='KiB'>4194304</memory>
> <currentMemory unit='KiB'>4194304</currentMemory>
> <vcpu placement='static' current='2'>16</vcpu>
> <resource>
> <partition>/machine</partition>
> </resource>
> <sysinfo type='smbios'>
> <system>
> <entry name='manufacturer'>oVirt</entry>
> <entry name='product'>RHEL</entry>
> <entry name='version'>8.4-1.2105.el8</entry>
> <entry name='serial'>83e66af8-0500-11e6-9c43-bc00007c0000</entry>
> <entry name='uuid'>59a7461c-72fe-4e01-86a7-c70243f31596</entry>
> <entry name='family'>oVirt</entry>
> </system>
> </sysinfo>
> <os>
> <type arch='x86_64' machine='pc-q35-rhel8.4.0'>hvm</type>
> <boot dev='hd'/>
> <smbios mode='sysinfo'/>
> </os>
> <features>
> <acpi/>
> </features>
> <cpu mode='custom' match='exact' check='full'>
> <model fallback='forbid'>Skylake-Client-noTSX-IBRS</model>
> <topology sockets='16' dies='1' cores='1' threads='1'/>
> <feature policy='require' name='ssbd'/>
> <feature policy='require' name='md-clear'/>
> <feature policy='disable' name='mpx'/>
> <feature policy='require' name='hypervisor'/>
> <numa>
> <cell id='0' cpus='0-15' memory='4194304' unit='KiB'/>
> </numa>
> </cpu>
> <clock offset='variable' adjustment='0' basis='utc'>
> <timer name='rtc' tickpolicy='catchup'/>
> <timer name='pit' tickpolicy='delay'/>
> <timer name='hpet' present='no'/>
> </clock>
> <on_poweroff>destroy</on_poweroff>
> <on_reboot>restart</on_reboot>
> <on_crash>destroy</on_crash>
> <pm>
> <suspend-to-mem enabled='no'/>
> <suspend-to-disk enabled='no'/>
> </pm>
> <devices>
> <emulator>/usr/libexec/qemu-kvm</emulator>
> <disk type='file' device='cdrom'>
> <driver name='qemu' error_policy='report'/>
> <source startupPolicy='optional'/>
> <target dev='sdc' bus='sata'/>
> <readonly/>
> <alias name='ua-df0ac774-3623-4868-8bd3-45c8f2aa3dc4'/>
> <address type='drive' controller='0' bus='0' target='0' unit='2'/>
> </disk>
> <disk type='file' device='disk' snapshot='no'>
> <driver name='qemu' type='qcow2' cache='none' error_policy='stop'
> io='threads'/>
> <source
> file='/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4'
> index='8'>
> <seclabel model='dac' relabel='no'/>
> </source>
> <backingStore type='file' index='1'>
> <format type='qcow2'/>
> <source
> file='/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003'>
> <seclabel model='dac' relabel='no'/>
> </source>
> <backingStore/>
> </backingStore>
> <target dev='sda' bus='scsi'/>
> <serial>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</serial>
> <alias name='ua-bee44276-234f-4ed7-8a8a-d90a5e3cb5b3'/>
> <address type='drive' controller='0' bus='0' target='0' unit='0'/>
> </disk>
> <controller type='pci' index='1' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='1' port='0x10'/>
> <alias name='pci.1'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x0' multifunction='on'/>
> </controller>
> <controller type='pci' index='2' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='2' port='0x11'/>
> <alias name='pci.2'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x1'/>
> </controller>
> <controller type='pci' index='3' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='3' port='0x12'/>
> <alias name='pci.3'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x2'/>
> </controller>
> <controller type='pci' index='4' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='4' port='0x13'/>
> <alias name='pci.4'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x3'/>
> </controller>
> <controller type='pci' index='5' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='5' port='0x14'/>
> <alias name='pci.5'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x4'/>
> </controller>
> <controller type='pci' index='6' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='6' port='0x15'/>
> <alias name='pci.6'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x5'/>
> </controller>
> <controller type='pci' index='7' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='7' port='0x16'/>
> <alias name='pci.7'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x6'/>
> </controller>
> <controller type='pci' index='8' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='8' port='0x17'/>
> <alias name='pci.8'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x02'
> function='0x7'/>
> </controller>
> <controller type='pci' index='9' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='9' port='0x18'/>
> <alias name='pci.9'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x0' multifunction='on'/>
> </controller>
> <controller type='pci' index='10' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='10' port='0x19'/>
> <alias name='pci.10'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x1'/>
> </controller>
> <controller type='pci' index='11' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='11' port='0x1a'/>
> <alias name='pci.11'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x2'/>
> </controller>
> <controller type='pci' index='12' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='12' port='0x1b'/>
> <alias name='pci.12'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x3'/>
> </controller>
> <controller type='pci' index='13' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='13' port='0x1c'/>
> <alias name='pci.13'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x4'/>
> </controller>
> <controller type='pci' index='14' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='14' port='0x1d'/>
> <alias name='pci.14'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x5'/>
> </controller>
> <controller type='pci' index='15' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='15' port='0x1e'/>
> <alias name='pci.15'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x6'/>
> </controller>
> <controller type='pci' index='16' model='pcie-root-port'>
> <model name='pcie-root-port'/>
> <target chassis='16' port='0x1f'/>
> <alias name='pci.16'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x03'
> function='0x7'/>
> </controller>
> <controller type='pci' index='0' model='pcie-root'>
> <alias name='pcie.0'/>
> </controller>
> <controller type='sata' index='0'>
> <alias name='ide'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x1f'
> function='0x2'/>
> </controller>
> <controller type='virtio-serial' index='0' ports='16'>
> <alias name='virtio-serial0'/>
> <address type='pci' domain='0x0000' bus='0x04' slot='0x00'
> function='0x0'/>
> </controller>
> <controller type='scsi' index='0' model='virtio-scsi'>
> <alias name='ua-82c49f93-c4e8-460b-bb7d-95db0e9d87a0'/>
> <address type='pci' domain='0x0000' bus='0x02' slot='0x00'
> function='0x0'/>
> </controller>
> <controller type='usb' index='0' model='qemu-xhci' ports='8'>
> <alias name='ua-ad56daea-edb1-45c7-a1ab-2a7db3aaeee2'/>
> <address type='pci' domain='0x0000' bus='0x03' slot='0x00'
> function='0x0'/>
> </controller>
> <interface type='bridge'>
> <mac address='56:6f:16:a8:00:07'/>
> <source bridge='ovirtmgmt'/>
> <target dev='vnet0'/>
> <model type='virtio'/>
> <driver name='vhost' queues='2'/>
> <filterref filter='vdsm-no-mac-spoofing'/>
> <link state='up'/>
> <mtu size='1500'/>
> <alias name='ua-7c9f38e9-8889-46c8-83bb-92efb9272de9'/>
> <address type='pci' domain='0x0000' bus='0x01' slot='0x00'
> function='0x0'/>
> </interface>
> <channel type='unix'>
> <source mode='bind'
> path='/var/lib/libvirt/qemu/channels/59a7461c-72fe-4e01-86a7-c70243f31596.ovirt-guest-agent.0'/>
> <target type='virtio' name='ovirt-guest-agent.0' state='disconnected'/>
> <alias name='channel0'/>
> <address type='virtio-serial' controller='0' bus='0' port='1'/>
> </channel>
> <channel type='unix'>
> <source mode='bind'
> path='/var/lib/libvirt/qemu/channels/59a7461c-72fe-4e01-86a7-c70243f31596.org.qemu.guest_agent.0'/>
> <target type='virtio' name='org.qemu.guest_agent.0' state='connected'/>
> <alias name='channel1'/>
> <address type='virtio-serial' controller='0' bus='0' port='2'/>
> </channel>
> <channel type='spicevmc'>
> <target type='virtio' name='com.redhat.spice.0' state='disconnected'/>
> <alias name='channel2'/>
> <address type='virtio-serial' controller='0' bus='0' port='3'/>
> </channel>
> <input type='tablet' bus='usb'>
> <alias name='input0'/>
> <address type='usb' bus='0' port='1'/>
> </input>
> <input type='mouse' bus='ps2'>
> <alias name='input1'/>
> </input>
> <input type='keyboard' bus='ps2'>
> <alias name='input2'/>
> </input>
> <graphics type='vnc' port='5900' autoport='yes' listen='192.168.7.18'
> keymap='en-us' passwdValidTo='1970-01-01T00:00:01'>
> <listen type='network' address='192.168.7.18' network='vdsm-ovirtmgmt'/>
> </graphics>
> <graphics type='spice' port='5901' tlsPort='5902' autoport='yes'
> listen='192.168.7.18' passwdValidTo='1970-01-01T00:00:01'>
> <listen type='network' address='192.168.7.18' network='vdsm-ovirtmgmt'/>
> <channel name='main' mode='secure'/>
> <channel name='display' mode='secure'/>
> <channel name='inputs' mode='secure'/>
> <channel name='cursor' mode='secure'/>
> <channel name='playback' mode='secure'/>
> <channel name='record' mode='secure'/>
> <channel name='smartcard' mode='secure'/>
> <channel name='usbredir' mode='secure'/>
> </graphics>
> <video>
> <model type='qxl' ram='65536' vram='8192' vgamem='16384' heads='1'
> primary='yes'/>
> <alias name='ua-799f065a-b2b9-4e37-a502-f86c7cc8dc51'/>
> <address type='pci' domain='0x0000' bus='0x00' slot='0x01'
> function='0x0'/>
> </video>
> <memballoon model='virtio'>
> <stats period='5'/>
> <alias name='ua-c2bfe0b9-065a-46b7-9b0b-ef7e0f699611'/>
> <address type='pci' domain='0x0000' bus='0x05' slot='0x00'
> function='0x0'/>
> </memballoon>
> <rng model='virtio'>
> <backend model='random'>/dev/urandom</backend>
> <alias name='ua-fa110e6b-5eed-4b4b-93d8-0ac5de08aa2e'/>
> <address type='pci' domain='0x0000' bus='0x06' slot='0x00'
> function='0x0'/>
> </rng>
> </devices>
> <seclabel type='dynamic' model='selinux' relabel='yes'>
> <label>system_u:system_r:svirt_t:s0:c437,c650</label>
> <imagelabel>system_u:object_r:svirt_image_t:s0:c437,c650</imagelabel>
> </seclabel>
> <seclabel type='dynamic' model='dac' relabel='yes'>
> <label>+107:+107</label>
> <imagelabel>+107:+107</imagelabel>
> </seclabel>
> <qemu:capabilities>
> <qemu:add capability='blockdev'/>
> <qemu:add capability='incremental-backup'/>
> </qemu:capabilities>
> </domain>
> }, log id: 7eb54202
>
>
>
> Here is the engine log when deleting the pvc:
>
> 2021-08-20 17:43:12,964 - cinder.volume.driver - INFO - Driver hasn't
> implemented _init_vendor_properties()
> 2021-08-20 17:43:12,990 - cinderlib-client - INFO - Deleting volume
> '63a64445-1659-4d5f-8847-e7266e64b09e' [feefc62f-e7cb-435d-ae21-4b52b53fbdfa]
> 2021-08-20 17:43:28,856 - cinder.volume.drivers.rbd - WARNING - ImageBusy
> error raised while deleting rbd volume. This may have been caused by a
> connection from a client that has crashed and, if so, may be resolved by
> retrying the delete after 30 seconds has elapsed.
> 2021-08-20 17:43:28,900 - cinderlib-client - ERROR - Failure occurred when
> trying to run command 'delete_volume': ImageBusy error raised while deleting
> rbd volume. This may have been caused by a connection from a client that has
> crashed and, if so, may be resolved by retrying the delete after 30 seconds
> has elapsed. [feefc62f-e7cb-435d-ae21-4b52b53fbdfa]
> 2021-08-20 17:43:28,901 - cinder - CRITICAL - Unhandled error
> Traceback (most recent call last):
> File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line
> 1195, in delete_volume
> _try_remove_volume(client, volume_name)
> File "/usr/lib/python3.6/site-packages/cinder/utils.py", line 683, in
> _wrapper
> return r.call(f, *args, **kwargs)
> File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 409, in
> call
> do = self.iter(retry_state=retry_state)
> File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 368, in
> iter
> raise retry_exc.reraise()
> File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 186, in
> reraise
> raise self.last_attempt.result()
> File "/usr/lib64/python3.6/concurrent/futures/_base.py", line 425, in result
> return self.__get_result()
> File "/usr/lib64/python3.6/concurrent/futures/_base.py", line 384, in
> __get_result
> raise self._exception
> File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 412, in
> call
> result = fn(*args, **kwargs)
> File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line
> 1190, in _try_remove_volume
> self.RBDProxy().remove(client.ioctx, volume_name)
> File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 190, in doit
> result = proxy_call(self._autowrap, f, *args, **kwargs)
> File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 148, in
> proxy_call
> rv = execute(f, *args, **kwargs)
> File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 129, in
> execute
> six.reraise(c, e, tb)
> File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise
> raise value
> File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 83, in
> tworker
> rv = meth(*args, **kwargs)
> File "rbd.pyx", line 767, in rbd.RBD.remove
> rbd.ImageBusy: [errno 16] RBD image is busy (error removing image)
>
> During handling of the above exception, another exception occurred:
>
> Traceback (most recent call last):
> File "./cinderlib-client.py", line 170, in main
> args.command(args)
> File "./cinderlib-client.py", line 218, in delete_volume
> vol.delete()
> File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line
> 494, in delete
> self._raise_with_resource()
> File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line
> 222, in _raise_with_resource
> six.reraise(*exc_info)
> File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise
> raise value
> File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line
> 487, in delete
> self.backend.driver.delete_volume(self._ovo)
> File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line
> 1205, in delete_volume
> raise exception.VolumeIsBusy(msg, volume_name=volume_name)
> cinder.exception.VolumeIsBusy: ImageBusy error raised while deleting rbd
> volume. This may have been caused by a connection from a client that has
> crashed and, if so, may be resolved by retrying the delete after 30 seconds
> has elapsed.
>
> During handling of the above exception, another exception occurred:
>
> Traceback (most recent call last):
> File "./cinderlib-client.py", line 390, in <module>
> sys.exit(main(sys.argv[1:]))
> File "./cinderlib-client.py", line 176, in main
> sys.stderr.write(traceback.format_exc(e))
> File "/usr/lib64/python3.6/traceback.py", line 167, in format_exc
> return "".join(format_exception(*sys.exc_info(), limit=limit,
> chain=chain))
> File "/usr/lib64/python3.6/traceback.py", line 121, in format_exception
> type(value), value, tb, limit=limit).format(chain=chain))
> File "/usr/lib64/python3.6/traceback.py", line 498, in __init__
> _seen=_seen)
> File "/usr/lib64/python3.6/traceback.py", line 509, in __init__
> capture_locals=capture_locals)
> File "/usr/lib64/python3.6/traceback.py", line 338, in extract
> if limit >= 0:
> TypeError: '>=' not supported between instances of 'VolumeIsBusy' and 'int'
> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
> Privacy Statement: https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/[email protected]/message/AFYZX2HJZ3RBJS5P7ENIQ4UIAJGP5MGG/
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-guidelines/
List Archives:
https://lists.ovirt.org/archives/list/[email protected]/message/22GGSKX527KGLJ3RM3E4UIPWXD5VHG3G/