Hi all,

I deployed ovirt-csi in the k8s by applying yaml manually. I used the latest 
version of the container image.
(https://github.com/openshift/ovirt-csi-driver-operator/tree/master/assets)

After successfully creating pvc and pod, I tried to delete it.
And the pod is deleted, but the pvc is not deleted. This is because deleting a 
pod does not unmap /dev/rbd0 attached to the ovirt vm.

How can I delete the pvc successfully?

oVirt engine version is 4.4.7.6-1.el8.
Here is the engine log when deleting the pod:

2021-08-20 17:40:35,385+09 INFO  
[org.ovirt.engine.core.sso.service.AuthenticationService] (default task-149) [] 
User admin@internal-authz with profile [internal] successfully logged in with 
scopes: ovirt-app-api ovirt-ext=token-info:authz-search 
ovirt-ext=token-info:public-authz-search ovirt-ext=token-info:validate 
ovirt-ext=token:password-access
2021-08-20 17:40:35,403+09 INFO  
[org.ovirt.engine.core.bll.aaa.CreateUserSessionCommand] (default task-149) 
[68ee3182] Running command: CreateUserSessionCommand internal: false.
2021-08-20 17:40:35,517+09 INFO  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (default 
task-149) [68ee3182] EVENT_ID: USER_VDC_LOGIN(30), User admin@internal-authz 
connecting from '192.168.7.169' using session 
'XfDgNkmAGnPiZahK5itLhHQTCNHZ3JwXMMzOiZrYL3C32+1TTys3xcjrAmCIKPu02hgN1sdVpfZXWd0FznaPCQ=='
 logged in.
2021-08-20 17:40:35,520+09 WARN  [org.ovirt.engine.api.restapi.util.LinkHelper] 
(default task-149) [] Can't find relative path for class 
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will 
return null
2021-08-20 17:40:35,520+09 WARN  [org.ovirt.engine.api.restapi.util.LinkHelper] 
(default task-149) [] Can't find relative path for class 
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will 
return null
2021-08-20 17:40:35,520+09 WARN  [org.ovirt.engine.api.restapi.util.LinkHelper] 
(default task-149) [] Can't find relative path for class 
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will 
return null
2021-08-20 17:40:35,520+09 WARN  [org.ovirt.engine.api.restapi.util.LinkHelper] 
(default task-149) [] Can't find relative path for class 
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will 
return null
2021-08-20 17:40:35,520+09 WARN  [org.ovirt.engine.api.restapi.util.LinkHelper] 
(default task-149) [] Can't find relative path for class 
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will 
return null
2021-08-20 17:40:35,520+09 WARN  [org.ovirt.engine.api.restapi.util.LinkHelper] 
(default task-149) [] Can't find relative path for class 
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will 
return null
2021-08-20 17:40:35,663+09 INFO  
[org.ovirt.engine.core.bll.storage.disk.DetachDiskFromVmCommand] (default 
task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] Running command: 
DetachDiskFromVmCommand internal: false. Entities affected :  ID: 
59a7461c-72fe-4e01-86a7-c70243f31596 Type: VMAction group CONFIGURE_VM_STORAGE 
with role type USER
2021-08-20 17:40:35,664+09 INFO  
[org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default 
task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] START, 
HotUnPlugDiskVDSCommand(HostName = host, 
HotPlugDiskVDSParameters:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094', 
vmId='59a7461c-72fe-4e01-86a7-c70243f31596', 
diskId='63a64445-1659-4d5f-8847-e7266e64b09e'}), log id: 506ff4a4
2021-08-20 17:40:35,678+09 INFO  
[org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default 
task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] Disk hot-unplug: <?xml 
version="1.0" encoding="UTF-8"?><hotunplug>
  <devices>
    <disk>
      <alias name="ua-63a64445-1659-4d5f-8847-e7266e64b09e"/>
    </disk>
  </devices>
</hotunplug>

2021-08-20 17:40:35,749+09 INFO  
[org.ovirt.engine.core.vdsbroker.vdsbroker.HotUnPlugDiskVDSCommand] (default 
task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] FINISH, 
HotUnPlugDiskVDSCommand, return: , log id: 506ff4a4
2021-08-20 17:40:35,842+09 INFO  
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (default 
task-149) [198e2dc9-b908-474e-a395-0fe682c29af0] EVENT_ID: 
USER_DETACH_DISK_FROM_VM(2,018), Disk pvc-9845a0ff-e94c-497c-8c65-fc6a1e26db20 
was successfully detached from VM centos by admin@internal-authz.
2021-08-20 17:40:35,916+09 ERROR [org.ovirt.engine.core.sso.service.SsoService] 
(default task-150) [] OAuthException invalid_grant: The provided authorization 
grant for the auth code has expired.
2021-08-20 17:40:35,917+09 ERROR 
[org.ovirt.engine.core.aaa.filters.SsoRestApiAuthFilter] (default task-149) [] 
Cannot authenticate using authentication Headers: invalid_grant: The provided 
authorization grant for the auth code has expired.
2021-08-20 17:40:36,029+09 INFO  
[org.ovirt.engine.core.sso.service.AuthenticationService] (default task-149) [] 
User admin@internal-authz with profile [internal] successfully logged in with 
scopes: ovirt-app-api ovirt-ext=token-info:authz-search 
ovirt-ext=token-info:public-authz-search ovirt-ext=token-info:validate 
ovirt-ext=token:password-access
2021-08-20 17:40:36,046+09 INFO  
[org.ovirt.engine.core.bll.aaa.CreateUserSessionCommand] (default task-149) 
[4c4bf441] Running command: CreateUserSessionCommand internal: false.
2021-08-20 17:40:36,114+09 WARN  [org.ovirt.engine.api.restapi.util.LinkHelper] 
(default task-149) [] Can't find relative path for class 
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will 
return null
2021-08-20 17:40:36,114+09 WARN  [org.ovirt.engine.api.restapi.util.LinkHelper] 
(default task-149) [] Can't find relative path for class 
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will 
return null
2021-08-20 17:40:36,114+09 WARN  [org.ovirt.engine.api.restapi.util.LinkHelper] 
(default task-149) [] Can't find relative path for class 
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will 
return null
2021-08-20 17:40:36,114+09 WARN  [org.ovirt.engine.api.restapi.util.LinkHelper] 
(default task-149) [] Can't find relative path for class 
"org.ovirt.engine.api.resource.StorageDomainVmDiskAttachmentsResource", will 
return null
2021-08-20 17:40:49,241+09 INFO  
[org.ovirt.engine.core.vdsbroker.vdsbroker.DumpXmlsVDSCommand] 
(EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-27) [] 
START, DumpXmlsVDSCommand(HostName = host, 
Params:{hostId='29dc5d53-7ec5-4a38-aaf1-c6eaf32b0094', 
vmIds='[59a7461c-72fe-4e01-86a7-c70243f31596]'}), log id: 7eb54202
2021-08-20 17:40:49,244+09 INFO  
[org.ovirt.engine.core.vdsbroker.vdsbroker.DumpXmlsVDSCommand] 
(EE-ManagedScheduledExecutorService-engineScheduledThreadPool-Thread-27) [] 
FINISH, DumpXmlsVDSCommand, return: 
{59a7461c-72fe-4e01-86a7-c70243f31596=<domain type='kvm' id='1' 
xmlns:qemu='http://libvirt.org/schemas/domain/qemu/1.0'>
  <name>centos</name>
  <uuid>59a7461c-72fe-4e01-86a7-c70243f31596</uuid>
  <metadata xmlns:ns1="http://ovirt.org/vm/tune/1.0"; 
xmlns:ovirt-vm="http://ovirt.org/vm/1.0";>
    <ns1:qos/>
    <ovirt-vm:vm xmlns:ovirt-vm="http://ovirt.org/vm/1.0";>
    <ovirt-vm:balloonTarget type="int">4194304</ovirt-vm:balloonTarget>
    <ovirt-vm:ballooningEnabled>true</ovirt-vm:ballooningEnabled>
    <ovirt-vm:clusterVersion>4.6</ovirt-vm:clusterVersion>
    <ovirt-vm:destroy_on_reboot type="bool">False</ovirt-vm:destroy_on_reboot>
    <ovirt-vm:launchPaused>false</ovirt-vm:launchPaused>
    <ovirt-vm:memGuaranteedSize type="int">4096</ovirt-vm:memGuaranteedSize>
    <ovirt-vm:minGuaranteedMemoryMb 
type="int">4096</ovirt-vm:minGuaranteedMemoryMb>
    <ovirt-vm:resumeBehavior>auto_resume</ovirt-vm:resumeBehavior>
    <ovirt-vm:startTime type="float">1628558564.8754532</ovirt-vm:startTime>
    <ovirt-vm:device alias="ua-7c9f38e9-8889-46c8-83bb-92efb9272de9" 
mac_address="56:6f:16:a8:00:07">
        <ovirt-vm:network>ovirtmgmt</ovirt-vm:network>
        <ovirt-vm:custom>
            <ovirt-vm:queues>2</ovirt-vm:queues>
        </ovirt-vm:custom>
    </ovirt-vm:device>
    <ovirt-vm:device devtype="disk" name="sda">
        
<ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
        <ovirt-vm:guestName>/dev/sda</ovirt-vm:guestName>
        
<ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
        <ovirt-vm:managed type="bool">False</ovirt-vm:managed>
        <ovirt-vm:poolID>4ca6e0e8-e3a4-11eb-8830-480fcf63834f</ovirt-vm:poolID>
        
<ovirt-vm:volumeID>a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:volumeID>
        <ovirt-vm:volumeChain>
            <ovirt-vm:volumeChainNode>
                
<ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
                
<ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
                <ovirt-vm:leaseOffset type="int">0</ovirt-vm:leaseOffset>
                
<ovirt-vm:leasePath>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003.lease</ovirt-vm:leasePath>
                
<ovirt-vm:path>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003</ovirt-vm:path>
                
<ovirt-vm:volumeID>4425404b-17f0-4519-812f-b09a952a9003</ovirt-vm:volumeID>
            </ovirt-vm:volumeChainNode>
            <ovirt-vm:volumeChainNode>
                
<ovirt-vm:domainID>6ce3b498-532e-4dc0-9e22-15d0bb24166a</ovirt-vm:domainID>
                
<ovirt-vm:imageID>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</ovirt-vm:imageID>
                <ovirt-vm:leaseOffset type="int">0</ovirt-vm:leaseOffset>
                
<ovirt-vm:leasePath>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4.lease</ovirt-vm:leasePath>
                
<ovirt-vm:path>/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:path>
                
<ovirt-vm:volumeID>a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4</ovirt-vm:volumeID>
            </ovirt-vm:volumeChainNode>
        </ovirt-vm:volumeChain>
    </ovirt-vm:device>
    <ovirt-vm:device devtype="disk" name="sdc">
        <ovirt-vm:managed type="bool">False</ovirt-vm:managed>
    </ovirt-vm:device>
</ovirt-vm:vm>
  </metadata>
  <maxMemory slots='16' unit='KiB'>16777216</maxMemory>
  <memory unit='KiB'>4194304</memory>
  <currentMemory unit='KiB'>4194304</currentMemory>
  <vcpu placement='static' current='2'>16</vcpu>
  <resource>
    <partition>/machine</partition>
  </resource>
  <sysinfo type='smbios'>
    <system>
      <entry name='manufacturer'>oVirt</entry>
      <entry name='product'>RHEL</entry>
      <entry name='version'>8.4-1.2105.el8</entry>
      <entry name='serial'>83e66af8-0500-11e6-9c43-bc00007c0000</entry>
      <entry name='uuid'>59a7461c-72fe-4e01-86a7-c70243f31596</entry>
      <entry name='family'>oVirt</entry>
    </system>
  </sysinfo>
  <os>
    <type arch='x86_64' machine='pc-q35-rhel8.4.0'>hvm</type>
    <boot dev='hd'/>
    <smbios mode='sysinfo'/>
  </os>
  <features>
    <acpi/>
  </features>
  <cpu mode='custom' match='exact' check='full'>
    <model fallback='forbid'>Skylake-Client-noTSX-IBRS</model>
    <topology sockets='16' dies='1' cores='1' threads='1'/>
    <feature policy='require' name='ssbd'/>
    <feature policy='require' name='md-clear'/>
    <feature policy='disable' name='mpx'/>
    <feature policy='require' name='hypervisor'/>
    <numa>
      <cell id='0' cpus='0-15' memory='4194304' unit='KiB'/>
    </numa>
  </cpu>
  <clock offset='variable' adjustment='0' basis='utc'>
    <timer name='rtc' tickpolicy='catchup'/>
    <timer name='pit' tickpolicy='delay'/>
    <timer name='hpet' present='no'/>
  </clock>
  <on_poweroff>destroy</on_poweroff>
  <on_reboot>restart</on_reboot>
  <on_crash>destroy</on_crash>
  <pm>
    <suspend-to-mem enabled='no'/>
    <suspend-to-disk enabled='no'/>
  </pm>
  <devices>
    <emulator>/usr/libexec/qemu-kvm</emulator>
    <disk type='file' device='cdrom'>
      <driver name='qemu' error_policy='report'/>
      <source startupPolicy='optional'/>
      <target dev='sdc' bus='sata'/>
      <readonly/>
      <alias name='ua-df0ac774-3623-4868-8bd3-45c8f2aa3dc4'/>
      <address type='drive' controller='0' bus='0' target='0' unit='2'/>
    </disk>
    <disk type='file' device='disk' snapshot='no'>
      <driver name='qemu' type='qcow2' cache='none' error_policy='stop' 
io='threads'/>
      <source 
file='/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/a93c6e11-6c7c-4efb-8c93-4ed5712ce7e4'
 index='8'>
        <seclabel model='dac' relabel='no'/>
      </source>
      <backingStore type='file' index='1'>
        <format type='qcow2'/>
        <source 
file='/rhev/data-center/mnt/192.168.7.18:_home_tmax_nfs/6ce3b498-532e-4dc0-9e22-15d0bb24166a/images/bee44276-234f-4ed7-8a8a-d90a5e3cb5b3/4425404b-17f0-4519-812f-b09a952a9003'>
          <seclabel model='dac' relabel='no'/>
        </source>
        <backingStore/>
      </backingStore>
      <target dev='sda' bus='scsi'/>
      <serial>bee44276-234f-4ed7-8a8a-d90a5e3cb5b3</serial>
      <alias name='ua-bee44276-234f-4ed7-8a8a-d90a5e3cb5b3'/>
      <address type='drive' controller='0' bus='0' target='0' unit='0'/>
    </disk>
    <controller type='pci' index='1' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='1' port='0x10'/>
      <alias name='pci.1'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x02' function='0x0' 
multifunction='on'/>
    </controller>
    <controller type='pci' index='2' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='2' port='0x11'/>
      <alias name='pci.2'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x02' 
function='0x1'/>
    </controller>
    <controller type='pci' index='3' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='3' port='0x12'/>
      <alias name='pci.3'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x02' 
function='0x2'/>
    </controller>
    <controller type='pci' index='4' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='4' port='0x13'/>
      <alias name='pci.4'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x02' 
function='0x3'/>
    </controller>
    <controller type='pci' index='5' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='5' port='0x14'/>
      <alias name='pci.5'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x02' 
function='0x4'/>
    </controller>
    <controller type='pci' index='6' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='6' port='0x15'/>
      <alias name='pci.6'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x02' 
function='0x5'/>
    </controller>
    <controller type='pci' index='7' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='7' port='0x16'/>
      <alias name='pci.7'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x02' 
function='0x6'/>
    </controller>
    <controller type='pci' index='8' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='8' port='0x17'/>
      <alias name='pci.8'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x02' 
function='0x7'/>
    </controller>
    <controller type='pci' index='9' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='9' port='0x18'/>
      <alias name='pci.9'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x03' function='0x0' 
multifunction='on'/>
    </controller>
    <controller type='pci' index='10' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='10' port='0x19'/>
      <alias name='pci.10'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x03' 
function='0x1'/>
    </controller>
    <controller type='pci' index='11' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='11' port='0x1a'/>
      <alias name='pci.11'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x03' 
function='0x2'/>
    </controller>
    <controller type='pci' index='12' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='12' port='0x1b'/>
      <alias name='pci.12'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x03' 
function='0x3'/>
    </controller>
    <controller type='pci' index='13' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='13' port='0x1c'/>
      <alias name='pci.13'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x03' 
function='0x4'/>
    </controller>
    <controller type='pci' index='14' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='14' port='0x1d'/>
      <alias name='pci.14'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x03' 
function='0x5'/>
    </controller>
    <controller type='pci' index='15' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='15' port='0x1e'/>
      <alias name='pci.15'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x03' 
function='0x6'/>
    </controller>
    <controller type='pci' index='16' model='pcie-root-port'>
      <model name='pcie-root-port'/>
      <target chassis='16' port='0x1f'/>
      <alias name='pci.16'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x03' 
function='0x7'/>
    </controller>
    <controller type='pci' index='0' model='pcie-root'>
      <alias name='pcie.0'/>
    </controller>
    <controller type='sata' index='0'>
      <alias name='ide'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x1f' 
function='0x2'/>
    </controller>
    <controller type='virtio-serial' index='0' ports='16'>
      <alias name='virtio-serial0'/>
      <address type='pci' domain='0x0000' bus='0x04' slot='0x00' 
function='0x0'/>
    </controller>
    <controller type='scsi' index='0' model='virtio-scsi'>
      <alias name='ua-82c49f93-c4e8-460b-bb7d-95db0e9d87a0'/>
      <address type='pci' domain='0x0000' bus='0x02' slot='0x00' 
function='0x0'/>
    </controller>
    <controller type='usb' index='0' model='qemu-xhci' ports='8'>
      <alias name='ua-ad56daea-edb1-45c7-a1ab-2a7db3aaeee2'/>
      <address type='pci' domain='0x0000' bus='0x03' slot='0x00' 
function='0x0'/>
    </controller>
    <interface type='bridge'>
      <mac address='56:6f:16:a8:00:07'/>
      <source bridge='ovirtmgmt'/>
      <target dev='vnet0'/>
      <model type='virtio'/>
      <driver name='vhost' queues='2'/>
      <filterref filter='vdsm-no-mac-spoofing'/>
      <link state='up'/>
      <mtu size='1500'/>
      <alias name='ua-7c9f38e9-8889-46c8-83bb-92efb9272de9'/>
      <address type='pci' domain='0x0000' bus='0x01' slot='0x00' 
function='0x0'/>
    </interface>
    <channel type='unix'>
      <source mode='bind' 
path='/var/lib/libvirt/qemu/channels/59a7461c-72fe-4e01-86a7-c70243f31596.ovirt-guest-agent.0'/>
      <target type='virtio' name='ovirt-guest-agent.0' state='disconnected'/>
      <alias name='channel0'/>
      <address type='virtio-serial' controller='0' bus='0' port='1'/>
    </channel>
    <channel type='unix'>
      <source mode='bind' 
path='/var/lib/libvirt/qemu/channels/59a7461c-72fe-4e01-86a7-c70243f31596.org.qemu.guest_agent.0'/>
      <target type='virtio' name='org.qemu.guest_agent.0' state='connected'/>
      <alias name='channel1'/>
      <address type='virtio-serial' controller='0' bus='0' port='2'/>
    </channel>
    <channel type='spicevmc'>
      <target type='virtio' name='com.redhat.spice.0' state='disconnected'/>
      <alias name='channel2'/>
      <address type='virtio-serial' controller='0' bus='0' port='3'/>
    </channel>
    <input type='tablet' bus='usb'>
      <alias name='input0'/>
      <address type='usb' bus='0' port='1'/>
    </input>
    <input type='mouse' bus='ps2'>
      <alias name='input1'/>
    </input>
    <input type='keyboard' bus='ps2'>
      <alias name='input2'/>
    </input>
    <graphics type='vnc' port='5900' autoport='yes' listen='192.168.7.18' 
keymap='en-us' passwdValidTo='1970-01-01T00:00:01'>
      <listen type='network' address='192.168.7.18' network='vdsm-ovirtmgmt'/>
    </graphics>
    <graphics type='spice' port='5901' tlsPort='5902' autoport='yes' 
listen='192.168.7.18' passwdValidTo='1970-01-01T00:00:01'>
      <listen type='network' address='192.168.7.18' network='vdsm-ovirtmgmt'/>
      <channel name='main' mode='secure'/>
      <channel name='display' mode='secure'/>
      <channel name='inputs' mode='secure'/>
      <channel name='cursor' mode='secure'/>
      <channel name='playback' mode='secure'/>
      <channel name='record' mode='secure'/>
      <channel name='smartcard' mode='secure'/>
      <channel name='usbredir' mode='secure'/>
    </graphics>
    <video>
      <model type='qxl' ram='65536' vram='8192' vgamem='16384' heads='1' 
primary='yes'/>
      <alias name='ua-799f065a-b2b9-4e37-a502-f86c7cc8dc51'/>
      <address type='pci' domain='0x0000' bus='0x00' slot='0x01' 
function='0x0'/>
    </video>
    <memballoon model='virtio'>
      <stats period='5'/>
      <alias name='ua-c2bfe0b9-065a-46b7-9b0b-ef7e0f699611'/>
      <address type='pci' domain='0x0000' bus='0x05' slot='0x00' 
function='0x0'/>
    </memballoon>
    <rng model='virtio'>
      <backend model='random'>/dev/urandom</backend>
      <alias name='ua-fa110e6b-5eed-4b4b-93d8-0ac5de08aa2e'/>
      <address type='pci' domain='0x0000' bus='0x06' slot='0x00' 
function='0x0'/>
    </rng>
  </devices>
  <seclabel type='dynamic' model='selinux' relabel='yes'>
    <label>system_u:system_r:svirt_t:s0:c437,c650</label>
    <imagelabel>system_u:object_r:svirt_image_t:s0:c437,c650</imagelabel>
  </seclabel>
  <seclabel type='dynamic' model='dac' relabel='yes'>
    <label>+107:+107</label>
    <imagelabel>+107:+107</imagelabel>
  </seclabel>
  <qemu:capabilities>
    <qemu:add capability='blockdev'/>
    <qemu:add capability='incremental-backup'/>
  </qemu:capabilities>
</domain>
}, log id: 7eb54202



Here is the engine log when deleting the pvc:

2021-08-20 17:43:12,964 - cinder.volume.driver - INFO - Driver hasn't 
implemented _init_vendor_properties()
2021-08-20 17:43:12,990 - cinderlib-client - INFO - Deleting volume 
'63a64445-1659-4d5f-8847-e7266e64b09e' [feefc62f-e7cb-435d-ae21-4b52b53fbdfa]
2021-08-20 17:43:28,856 - cinder.volume.drivers.rbd - WARNING - ImageBusy error 
raised while deleting rbd volume. This may have been caused by a connection 
from a client that has crashed and, if so, may be resolved by retrying the 
delete after 30 seconds has elapsed.
2021-08-20 17:43:28,900 - cinderlib-client - ERROR - Failure occurred when 
trying to run command 'delete_volume': ImageBusy error raised while deleting 
rbd volume. This may have been caused by a connection from a client that has 
crashed and, if so, may be resolved by retrying the delete after 30 seconds has 
elapsed. [feefc62f-e7cb-435d-ae21-4b52b53fbdfa]
2021-08-20 17:43:28,901 - cinder - CRITICAL - Unhandled error
Traceback (most recent call last):
  File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line 
1195, in delete_volume
    _try_remove_volume(client, volume_name)
  File "/usr/lib/python3.6/site-packages/cinder/utils.py", line 683, in _wrapper
    return r.call(f, *args, **kwargs)
  File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 409, in 
call
    do = self.iter(retry_state=retry_state)
  File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 368, in 
iter
    raise retry_exc.reraise()
  File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 186, in 
reraise
    raise self.last_attempt.result()
  File "/usr/lib64/python3.6/concurrent/futures/_base.py", line 425, in result
    return self.__get_result()
  File "/usr/lib64/python3.6/concurrent/futures/_base.py", line 384, in 
__get_result
    raise self._exception
  File "/usr/lib/python3.6/site-packages/tenacity/__init__.py", line 412, in 
call
    result = fn(*args, **kwargs)
  File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line 
1190, in _try_remove_volume
    self.RBDProxy().remove(client.ioctx, volume_name)
  File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 190, in doit
    result = proxy_call(self._autowrap, f, *args, **kwargs)
  File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 148, in 
proxy_call
    rv = execute(f, *args, **kwargs)
  File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 129, in 
execute
    six.reraise(c, e, tb)
  File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise
    raise value
  File "/usr/lib/python3.6/site-packages/eventlet/tpool.py", line 83, in tworker
    rv = meth(*args, **kwargs)
  File "rbd.pyx", line 767, in rbd.RBD.remove
rbd.ImageBusy: [errno 16] RBD image is busy (error removing image)

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "./cinderlib-client.py", line 170, in main
    args.command(args)
  File "./cinderlib-client.py", line 218, in delete_volume
    vol.delete()
  File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line 494, 
in delete
    self._raise_with_resource()
  File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line 222, 
in _raise_with_resource
    six.reraise(*exc_info)
  File "/usr/lib/python3.6/site-packages/six.py", line 703, in reraise
    raise value
  File "/usr/local/lib/python3.6/site-packages/cinderlib/objects.py", line 487, 
in delete
    self.backend.driver.delete_volume(self._ovo)
  File "/usr/lib/python3.6/site-packages/cinder/volume/drivers/rbd.py", line 
1205, in delete_volume
    raise exception.VolumeIsBusy(msg, volume_name=volume_name)
cinder.exception.VolumeIsBusy: ImageBusy error raised while deleting rbd 
volume. This may have been caused by a connection from a client that has 
crashed and, if so, may be resolved by retrying the delete after 30 seconds has 
elapsed.

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "./cinderlib-client.py", line 390, in <module>
    sys.exit(main(sys.argv[1:]))
  File "./cinderlib-client.py", line 176, in main
    sys.stderr.write(traceback.format_exc(e))
  File "/usr/lib64/python3.6/traceback.py", line 167, in format_exc
    return "".join(format_exception(*sys.exc_info(), limit=limit, chain=chain))
  File "/usr/lib64/python3.6/traceback.py", line 121, in format_exception
    type(value), value, tb, limit=limit).format(chain=chain))
  File "/usr/lib64/python3.6/traceback.py", line 498, in __init__
    _seen=_seen)
  File "/usr/lib64/python3.6/traceback.py", line 509, in __init__
    capture_locals=capture_locals)
  File "/usr/lib64/python3.6/traceback.py", line 338, in extract
    if limit >= 0:
TypeError: '>=' not supported between instances of 'VolumeIsBusy' and 'int'
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/AFYZX2HJZ3RBJS5P7ENIQ4UIAJGP5MGG/

Reply via email to