Public bug reported: Test failure during vm destroy as:- 2025-01-31 11:57:58,135 120175 INFO [tempest.common.waiters] State transition "RESIZE/resize_prep" ==> "RESIZE/resize_migrating" after 8 second wait
2025-01-31 11:58:01,827 120175 INFO [tempest.common.waiters] State transition "RESIZE/resize_migrating" ==> "RESIZE/resize_migrated" after 11 second wait 2025-01-31 11:58:05,489 120175 INFO [tempest.common.waiters] State transition "RESIZE/resize_migrated" ==> "RESIZE/resize_finish" after 15 second wait 2025-01-31 11:58:15,869 120175 INFO [tempest.common.waiters] State transition "RESIZE/resize_finish" ==> "VERIFY_RESIZE/None" after 25 second wait Response - Headers: {'date': 'Fri, 31 Jan 2025 11:59:10 GMT', 'server': 'Apache/2.4.58 (Ubuntu)', 'content-length': '1718', 'content-type': 'application/json', 'openstack-api-version': 'compute 2.72', 'x-openstack-nova-api-version': '2.72', 'vary': 'OpenStack-API-Version,X-OpenStack-Nova-API-Version', 'x-openstack-request-id': 'req-d15679f5-e820-42ac-b93e-76024fa41940', 'x-compute-request-id': 'req-d15679f5-e820-42ac-b93e-76024fa41940', 'connection': 'close', 'status': '200', 'content-location': 'https://158.69.67.245/compute/v2.1/servers/007a11e4-32d4-4b1c-9199-56a664fe86ed'} Body: b'{"server": {"id": "007a11e4-32d4-4b1c-9199-56a664fe86ed", "name": "tempest-QoSBandwidthAndPacketRateTests-server-1599599327", "status": "ERROR", "tenant_id": "fab8f5903cec400ab103cf64a8530b57", "user_id": "35b1e65e2c354a48b77f2f7247070fde", "metadata": {}, "hostId": "8ac31065c7b89285e06e31bd99c60023f629010753c70c6f324d93ea", "image": {"id": "acb24912-db2b-4ee1-8a71-9ce02d0abf0f", "links": [{"rel": "bookmark", "href": "https://158.69.67.245/compute/images/acb24912-db2b-4ee1-8a71-9ce02d0abf0f"}]}, "flavor": {"vcpus": 1, "ram": 192, "disk": 2, "ephemeral": 0, "swap": 0, "original_name": "m1.nanoextra-391323443", "extra_specs": {}}, "created": "2025-01-31T11:57:38Z", "updated": "2025-01-31T11:59:09Z", "addresses": {"tempest-network-smoke--673944778": [{"version": 4, "addr": "10.1.0.22", "OS-EXT-IPS:type": "fixed", "OS-EXT-IPS-MAC:mac_addr": "fa:16:3e:60:17:c9"}]}, "accessIPv4": "", "accessIPv6": "", "links": [{"rel": "self", "href": "https://158.69.67.245/compute/v2.1/ser vers/007a11e4-32d4-4b1c-9199-56a664fe86ed"}, {"rel": "bookmark", "href": "https://158.69.67.245/compute/servers/007a11e4-32d4-4b1c-9199-56a664fe86ed"}], "OS-DCF:diskConfig": "MANUAL", "fault": {"code": 500, "created": "2025-01-31T11:59:09Z", "message": "libvirtError"}, "OS-EXT-AZ:availability_zone": "nova", "config_drive": "", "key_name": null, "OS-SRV-USG:launched_at": "2025-01-31T11:58:13.000000", "OS-SRV-USG:terminated_at": null, "security_groups": [{"name": "default"}], "OS-EXT-STS:task_state": "resize_reverting", "OS-EXT-STS:vm_state": "error", "OS-EXT-STS:power_state": 1, "os-extended-volumes:volumes_attached": [], "locked": false, "description": null, "tags": [], "trusted_image_certificates": null, "server_groups": []}}' 2025-01-31 11:59:10,617 120175 INFO [tempest.common.waiters] State transition "REVERT_RESIZE/resize_reverting" ==> "ERROR/resize_reverting" after 54 second wait Traceback (most recent call last): File "/opt/stack/tempest/tempest/common/waiters.py", line 136, in wait_for_server_termination raise lib_exc.DeleteErrorException(details, server_id=server_id) tempest.lib.exceptions.DeleteErrorException: Resource %(resource_id)s failed to delete and is in ERROR status Details: Server 007a11e4-32d4-4b1c-9199-56a664fe86ed failed to delete and is in ERROR status. Fault: {'code': 500, 'created': '2025-01-31T11:59:09Z', 'message': 'libvirtError'}. }}} Traceback (most recent call last): File "/opt/stack/tempest/tempest/common/utils/__init__.py", line 65, in wrapper return f(*func_args, **func_kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/stack/tempest/tempest/scenario/test_network_qos_placement.py", line 851, in test_server_resize_revert waiters.wait_for_server_status( File "/opt/stack/tempest/tempest/common/waiters.py", line 80, in wait_for_server_status raise exceptions.BuildErrorException(details, server_id=server_id) tempest.exceptions.BuildErrorException: Server 007a11e4-32d4-4b1c-9199-56a664fe86ed failed to build and is in ERROR status Details: Fault: {'code': 500, 'created': '2025-01-31T11:59:09Z', 'message': 'libvirtError'}. Nova compute logs:- Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: WARNING nova.virt.libvirt.driver [None req-5e44a4a9-7f73-4019-aec0-86bec9372149 tempest-QoSBandwidthAndPacketRateTests-1787381104 tempest-QoSBandwidthAndPacketRateTests-1787381104-project-member] [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] Cannot destroy instance, general system call failure: libvirt.libvirtError: Failed to terminate process 53286 with SIGKILL: Device or resource busy Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [None req-5e44a4a9-7f73-4019-aec0-86bec9372149 tempest-QoSBandwidthAndPacketRateTests-1787381104 tempest-QoSBandwidthAndPacketRateTests-1787381104-project-member] [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] Setting instance vm_state to ERROR: libvirt.libvirtError: Failed to terminate process 53286 with SIGKILL: Device or resource busy Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] Traceback (most recent call last): Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/nova/nova/compute/manager.py", line 11382, in _error_out_instance_on_exception Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] yield Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/nova/nova/compute/manager.py", line 5914, in revert_resize Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] self.driver.destroy(context, instance, network_info, Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/nova/nova/virt/libvirt/driver.py", line 1622, in destroy Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] self._destroy(instance) Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/nova/nova/virt/libvirt/driver.py", line 1567, in _destroy Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] with excutils.save_and_reraise_exception(): Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/data/venv/lib/python3.12/site-packages/oslo_utils/excutils.py", line 227, in __exit__ Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] self.force_reraise() Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/data/venv/lib/python3.12/site-packages/oslo_utils/excutils.py", line 200, in force_reraise Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] raise self.value Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/nova/nova/virt/libvirt/driver.py", line 1535, in _destroy Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] guest.poweroff() Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/nova/nova/virt/libvirt/guest.py", line 173, in poweroff Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] self._domain.destroy() Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/data/venv/lib/python3.12/site-packages/eventlet/tpool.py", line 186, in doit Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] result = proxy_call(self._autowrap, f, *args, **kwargs) Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/data/venv/lib/python3.12/site-packages/eventlet/tpool.py", line 144, in proxy_call Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] rv = execute(f, *args, **kwargs) Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] ^^^^^^^^^^^^^^^^^^^^^^^^^^^ Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/data/venv/lib/python3.12/site-packages/eventlet/tpool.py", line 125, in execute Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] raise e.with_traceback(tb) Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/data/venv/lib/python3.12/site-packages/eventlet/tpool.py", line 82, in tworker Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] rv = meth(*args, **kwargs) Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] ^^^^^^^^^^^^^^^^^^^^^ Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/usr/lib/python3/dist-packages/libvirt.py", line 1461, in destroy Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] raise libvirtError('virDomainDestroy() failed') Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] libvirt.libvirtError: Failed to terminate process 53286 with SIGKILL: Device or resource busy Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] Seen twice, ex builds:- https://0c63ab9652170854bf26-a09d1a3317eb4b9b558e42ad19c25861.ssl.cf2.rackcdn.com/940474/1/gate/neutron-ovs-tempest-multinode-full/5184c7f/testr_results.html https://97ad4d1320a89f3380ef-01f5fc3a5734547a13a0f54725d40b32.ssl.cf5.rackcdn.com/936364/4/gate/nova-multi-cell/3003ee6/testr_results.html Opensearch(creds: openstack/openstack) https://opensearch.logs.openstack.org/_dashboards/app/data- explorer/discover?security_tenant=global#?_a=(discover:(columns:!(_source),isDirty:!f,sort:!()),metadata:(indexPattern:'94869730-aea8-11ec-9e6a-83741af3fdcd',view:discover))&_g=(filters:!(),refreshInterval:(pause:!t,value:0),time:(from:now-30d,to:now))&_q=(filters:!(),query:(language:kuery,query:'message:%22libvirt.libvirtError:%20Failed%20to%20terminate%20process%22')) Found an old bug https://bugs.launchpad.net/nova/+bug/1353939 under which a workaround was applied to retry 6 times before giving up but that was targetting libvirt < 4.7.0 with https://review.opendev.org/c/openstack/nova/+/639091. And then it was removed with support removal of old libvirt versions https://review.opendev.org/c/openstack/nova/+/746984. This bug is to track what exactly triggering this behavior now and if some workaround flag can be added in nova to have some attempts on temporary failure during destroy as before. Also some discussion on nova channel https://meetings.opendev.org/irclogs/%23openstack-nova/%23openstack- nova.2025-02-06.log.html#t2025-02-06T15:04:13 ** Affects: nova Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/2097586 Title: Random test failures with libvirt.libvirtError: Failed to terminate process 53286 with SIGKILL: Device or resource busy Status in OpenStack Compute (nova): New Bug description: Test failure during vm destroy as:- 2025-01-31 11:57:58,135 120175 INFO [tempest.common.waiters] State transition "RESIZE/resize_prep" ==> "RESIZE/resize_migrating" after 8 second wait 2025-01-31 11:58:01,827 120175 INFO [tempest.common.waiters] State transition "RESIZE/resize_migrating" ==> "RESIZE/resize_migrated" after 11 second wait 2025-01-31 11:58:05,489 120175 INFO [tempest.common.waiters] State transition "RESIZE/resize_migrated" ==> "RESIZE/resize_finish" after 15 second wait 2025-01-31 11:58:15,869 120175 INFO [tempest.common.waiters] State transition "RESIZE/resize_finish" ==> "VERIFY_RESIZE/None" after 25 second wait Response - Headers: {'date': 'Fri, 31 Jan 2025 11:59:10 GMT', 'server': 'Apache/2.4.58 (Ubuntu)', 'content-length': '1718', 'content-type': 'application/json', 'openstack-api-version': 'compute 2.72', 'x-openstack-nova-api-version': '2.72', 'vary': 'OpenStack-API-Version,X-OpenStack-Nova-API-Version', 'x-openstack-request-id': 'req-d15679f5-e820-42ac-b93e-76024fa41940', 'x-compute-request-id': 'req-d15679f5-e820-42ac-b93e-76024fa41940', 'connection': 'close', 'status': '200', 'content-location': 'https://158.69.67.245/compute/v2.1/servers/007a11e4-32d4-4b1c-9199-56a664fe86ed'} Body: b'{"server": {"id": "007a11e4-32d4-4b1c-9199-56a664fe86ed", "name": "tempest-QoSBandwidthAndPacketRateTests-server-1599599327", "status": "ERROR", "tenant_id": "fab8f5903cec400ab103cf64a8530b57", "user_id": "35b1e65e2c354a48b77f2f7247070fde", "metadata": {}, "hostId": "8ac31065c7b89285e06e31bd99c60023f629010753c70c6f324d93ea", "image": {"id": "acb24912-db2b-4ee1-8a71-9ce02d0abf0f", "links": [{"rel": "bookmark", "href": "https://158.69.67.245/compute/images/acb24912-db2b-4ee1-8a71-9ce02d0abf0f"}]}, "flavor": {"vcpus": 1, "ram": 192, "disk": 2, "ephemeral": 0, "swap": 0, "original_name": "m1.nanoextra-391323443", "extra_specs": {}}, "created": "2025-01-31T11:57:38Z", "updated": "2025-01-31T11:59:09Z", "addresses": {"tempest-network-smoke--673944778": [{"version": 4, "addr": "10.1.0.22", "OS-EXT-IPS:type": "fixed", "OS-EXT-IPS-MAC:mac_addr": "fa:16:3e:60:17:c9"}]}, "accessIPv4": "", "accessIPv6": "", "links": [{"rel": "self", "href": "https://158.69.67.245/compute/v2.1/s ervers/007a11e4-32d4-4b1c-9199-56a664fe86ed"}, {"rel": "bookmark", "href": "https://158.69.67.245/compute/servers/007a11e4-32d4-4b1c-9199-56a664fe86ed"}], "OS-DCF:diskConfig": "MANUAL", "fault": {"code": 500, "created": "2025-01-31T11:59:09Z", "message": "libvirtError"}, "OS-EXT-AZ:availability_zone": "nova", "config_drive": "", "key_name": null, "OS-SRV-USG:launched_at": "2025-01-31T11:58:13.000000", "OS-SRV-USG:terminated_at": null, "security_groups": [{"name": "default"}], "OS-EXT-STS:task_state": "resize_reverting", "OS-EXT-STS:vm_state": "error", "OS-EXT-STS:power_state": 1, "os-extended-volumes:volumes_attached": [], "locked": false, "description": null, "tags": [], "trusted_image_certificates": null, "server_groups": []}}' 2025-01-31 11:59:10,617 120175 INFO [tempest.common.waiters] State transition "REVERT_RESIZE/resize_reverting" ==> "ERROR/resize_reverting" after 54 second wait Traceback (most recent call last): File "/opt/stack/tempest/tempest/common/waiters.py", line 136, in wait_for_server_termination raise lib_exc.DeleteErrorException(details, server_id=server_id) tempest.lib.exceptions.DeleteErrorException: Resource %(resource_id)s failed to delete and is in ERROR status Details: Server 007a11e4-32d4-4b1c-9199-56a664fe86ed failed to delete and is in ERROR status. Fault: {'code': 500, 'created': '2025-01-31T11:59:09Z', 'message': 'libvirtError'}. }}} Traceback (most recent call last): File "/opt/stack/tempest/tempest/common/utils/__init__.py", line 65, in wrapper return f(*func_args, **func_kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/opt/stack/tempest/tempest/scenario/test_network_qos_placement.py", line 851, in test_server_resize_revert waiters.wait_for_server_status( File "/opt/stack/tempest/tempest/common/waiters.py", line 80, in wait_for_server_status raise exceptions.BuildErrorException(details, server_id=server_id) tempest.exceptions.BuildErrorException: Server 007a11e4-32d4-4b1c-9199-56a664fe86ed failed to build and is in ERROR status Details: Fault: {'code': 500, 'created': '2025-01-31T11:59:09Z', 'message': 'libvirtError'}. Nova compute logs:- Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: WARNING nova.virt.libvirt.driver [None req-5e44a4a9-7f73-4019-aec0-86bec9372149 tempest-QoSBandwidthAndPacketRateTests-1787381104 tempest-QoSBandwidthAndPacketRateTests-1787381104-project-member] [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] Cannot destroy instance, general system call failure: libvirt.libvirtError: Failed to terminate process 53286 with SIGKILL: Device or resource busy Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [None req-5e44a4a9-7f73-4019-aec0-86bec9372149 tempest-QoSBandwidthAndPacketRateTests-1787381104 tempest-QoSBandwidthAndPacketRateTests-1787381104-project-member] [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] Setting instance vm_state to ERROR: libvirt.libvirtError: Failed to terminate process 53286 with SIGKILL: Device or resource busy Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] Traceback (most recent call last): Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/nova/nova/compute/manager.py", line 11382, in _error_out_instance_on_exception Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] yield Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/nova/nova/compute/manager.py", line 5914, in revert_resize Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] self.driver.destroy(context, instance, network_info, Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/nova/nova/virt/libvirt/driver.py", line 1622, in destroy Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] self._destroy(instance) Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/nova/nova/virt/libvirt/driver.py", line 1567, in _destroy Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] with excutils.save_and_reraise_exception(): Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/data/venv/lib/python3.12/site-packages/oslo_utils/excutils.py", line 227, in __exit__ Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] self.force_reraise() Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/data/venv/lib/python3.12/site-packages/oslo_utils/excutils.py", line 200, in force_reraise Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] raise self.value Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/nova/nova/virt/libvirt/driver.py", line 1535, in _destroy Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] guest.poweroff() Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/nova/nova/virt/libvirt/guest.py", line 173, in poweroff Jan 31 11:59:04.005181 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] self._domain.destroy() Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/data/venv/lib/python3.12/site-packages/eventlet/tpool.py", line 186, in doit Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] result = proxy_call(self._autowrap, f, *args, **kwargs) Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/data/venv/lib/python3.12/site-packages/eventlet/tpool.py", line 144, in proxy_call Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] rv = execute(f, *args, **kwargs) Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] ^^^^^^^^^^^^^^^^^^^^^^^^^^^ Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/data/venv/lib/python3.12/site-packages/eventlet/tpool.py", line 125, in execute Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] raise e.with_traceback(tb) Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/opt/stack/data/venv/lib/python3.12/site-packages/eventlet/tpool.py", line 82, in tworker Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] rv = meth(*args, **kwargs) Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] ^^^^^^^^^^^^^^^^^^^^^ Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] File "/usr/lib/python3/dist-packages/libvirt.py", line 1461, in destroy Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] raise libvirtError('virDomainDestroy() failed') Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] libvirt.libvirtError: Failed to terminate process 53286 with SIGKILL: Device or resource busy Jan 31 11:59:04.212284 np0039731693 nova-compute[31800]: ERROR nova.compute.manager [instance: 007a11e4-32d4-4b1c-9199-56a664fe86ed] Seen twice, ex builds:- https://0c63ab9652170854bf26-a09d1a3317eb4b9b558e42ad19c25861.ssl.cf2.rackcdn.com/940474/1/gate/neutron-ovs-tempest-multinode-full/5184c7f/testr_results.html https://97ad4d1320a89f3380ef-01f5fc3a5734547a13a0f54725d40b32.ssl.cf5.rackcdn.com/936364/4/gate/nova-multi-cell/3003ee6/testr_results.html Opensearch(creds: openstack/openstack) https://opensearch.logs.openstack.org/_dashboards/app/data- explorer/discover?security_tenant=global#?_a=(discover:(columns:!(_source),isDirty:!f,sort:!()),metadata:(indexPattern:'94869730-aea8-11ec-9e6a-83741af3fdcd',view:discover))&_g=(filters:!(),refreshInterval:(pause:!t,value:0),time:(from:now-30d,to:now))&_q=(filters:!(),query:(language:kuery,query:'message:%22libvirt.libvirtError:%20Failed%20to%20terminate%20process%22')) Found an old bug https://bugs.launchpad.net/nova/+bug/1353939 under which a workaround was applied to retry 6 times before giving up but that was targetting libvirt < 4.7.0 with https://review.opendev.org/c/openstack/nova/+/639091. And then it was removed with support removal of old libvirt versions https://review.opendev.org/c/openstack/nova/+/746984. This bug is to track what exactly triggering this behavior now and if some workaround flag can be added in nova to have some attempts on temporary failure during destroy as before. Also some discussion on nova channel https://meetings.opendev.org/irclogs/%23openstack-nova/%23openstack- nova.2025-02-06.log.html#t2025-02-06T15:04:13 To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/2097586/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp