@nova team: https://docs.openstack.org/nova/latest/reference/api-
microversion-history.html#maximum-in-2023-1-antelope-and-2023-2-bobcat
If compute service is down in source node and user try to stop instance,
instance gets stuck at powering-off, hence evacuation fails with msg:
Cannot ‘evacuate’ instance <instance-id> while it is in task_state
powering-off. It is now possible for evacuation to ignore the vm task
state. For more details see: bug 1978983
** Also affects: nova
Importance: Undecided
Status: New
** Changed in: nova
Status: New => In Progress
** Summary changed:
- [Caracal][Offline][Masakari] - Instance-HA partially working
+ [Caracal][Offline][Masakari/Nova] - Instance-HA partially working
--
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/2119126
Title:
[Caracal][Offline][Masakari/Nova] - Instance-HA partially working
Status in masakari:
In Progress
Status in OpenStack Compute (nova):
In Progress
Bug description:
++++++++++++
ENV Details:
++++++++++++
OSA Version: OFFLINE Caracal 2024.1
OS: Ubuntu-22.04
Tool: OpenStack-Ansible
Virtual setup
++++++
Issue:
++++++
* Masakari is installed and running.
* Created a instance and enabled ``HA_Enabled=True` properties.
* Instance is tunning on the source node `cmpt001` destination node is
``offline20241``
* On source node Started instance-ha operation by :- ### or if there is a
better way let me know
```
# Source compute
root@cmpt001:~# virsh list --all
Id Name State
-----------------------------------
4 instance-0000001e running
root@cmpt001:~#
```
systemctl stop nova-compute.service
systemctl stop pacemaker.service
kill -9 $(ps -eaf | grep instance-0000001e | awk '{print $2}' | head -n 1)
systemctl stop corosync.service
```
```
root@cmpt001:~#
root@cmpt001:~# virsh list --all
Id Name State
------------------------------------
- instance-0000001e shut off
root@cmpt001:~#
```
* Evacuation started. it took 15 sec to migrate the instance.
* Instance evacuation is happeneing, however the instance on the destination
node is showing the Status as ShutOff
#Source compute logs
```
Jul3014: 50: 25cmpt001.ct.lanmasakari-instancemonitor[
1344
]: 2025-07-3014: 50:
25.5551344INFOmasakarimonitors.instancemonitor.libvirt_handler.callback[
-
]LibvirtEvent: type=VM,
hostname=cmpt001.ct.lan,
uuid=969ca417-9111-42de-836c-eb883e52f131,
time=2025-07-3014: 50: 25.552570,
event_id=LIFECYCLE,
detail=STOPPED_FAILED)```
```
```
Jul3014: 50: 25cmpt001.ct.lanmasakari-instancemonitor[
1344
]: 2025-07-3014: 50: 25.5571344INFOmasakarimonitors.ha.masakari[
-
]Sendanotification.{
'notification': {
'type': 'VM',
'hostname': 'cmpt001.ct.lan',
'generated_time': datetime.datetime(2025,
7,
30,
14,
50,
25,
552570),
'payload': {
'event': 'LIFECYCLE',
'instance_uuid': '969ca417-9111-42de-836c-eb883e52f131',
'vir_domain_event': 'STOPPED_FAILED'
}
}
}
```
Jul3014: 50: 25cmpt001.ct.lanmasakari-instancemonitor[
1344
]: 2025-07-3014: 50: 25.7861344INFOmasakarimonitors.ha.masakari[
-
]Response: openstack.instance_ha.v1.notification.Notification(type=VM,
hostname=cmpt001.ct.lan,
generated_time=2025-07-30T14: 50: 25.552570,
payload={
'event': 'LIFECYCLE',
'instance_uuid': '969ca417-9111-42de-836c-eb883e52f131',
'vir_domain_event': 'STOPPED_FAILED'
},
id=1,
notification_uuid=3be2b8e5-ed78-4085-8898-54b3fd5a9f78,
source_host_uuid=36cd2bdb-29e4-4cc4-9b10-a933e2608edc,
status=new,
created_at=2025-07-30T14: 50: 25.000000,
updated_at=None,
location=Munch({
'cloud': '192.168.131.200',
'region_name': 'RegionOne',
'zone': None,
'project': Munch({
'id': 'a5aebb0fbfc64ac49e3ea028e4f740dc',
'name': None,
'domain_id': None,
'domain_name': None
})
}))
```
# Destination compute + controller logs
```
root@offline20241:~# openstack server show
969ca417-9111-42de-836c-eb883e52f131 --fit
+-------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Field | Value
|
+-------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| OS-DCF:diskConfig | MANUAL
|
| OS-EXT-AZ:availability_zone | nova
|
| OS-EXT-SRV-ATTR:host | offline20241.ct.lan
|
| OS-EXT-SRV-ATTR:hostname | nc-masak-002
|
| OS-EXT-SRV-ATTR:hypervisor_hostname | offline20241.ct.lan
|
| OS-EXT-SRV-ATTR:instance_name | instance-0000001e
|
| OS-EXT-SRV-ATTR:kernel_id |
|
| OS-EXT-SRV-ATTR:launch_index | 0
|
| OS-EXT-SRV-ATTR:ramdisk_id |
|
| OS-EXT-SRV-ATTR:reservation_id | r-01be9wmi
|
| OS-EXT-SRV-ATTR:root_device_name | /dev/vda
|
| OS-EXT-SRV-ATTR:user_data | None
|
| OS-EXT-STS:power_state | Running
|
| OS-EXT-STS:task_state | None
|
| OS-EXT-STS:vm_state | active
|
| OS-SRV-USG:launched_at | 2025-07-30T14:52:30.000000
|
| OS-SRV-USG:terminated_at | None
|
| accessIPv4 |
|
| accessIPv6 |
|
| addresses | provider141=192.168.141.91
|
| config_drive |
|
| created | 2025-07-30T14:46:27Z
|
| description | nc-masak-002
|
| flavor | description=, disk='0',
ephemeral='0', , id='m1.tiny', is_disabled=, is_public='True', location=,
name='m1.tiny', original_name='m1.tiny', ram='512', rxtx_factor=, swap='0',
vcpus='1' |
| hostId |
1cb881e0cf1cb53f01ef64ad3b04badf5e418abc5a42566593853d4a
|
| host_status | UP
|
| id | 969ca417-9111-42de-836c-eb883e52f131
|
| image | N/A (booted from volume)
|
| key_name | None
|
| locked | False
|
| locked_reason | None
|
| name | nc-masak-002
|
| progress | 0
|
| project_id | 2f5a2a06638942cbaaeeb466b2e17e10
|
| properties | HA_Enabled='True'
|
| security_groups | name='secgroup1'
|
| server_groups | []
|
| status | ACTIVE
|
| tags |
|
| trusted_image_certificates | None
|
| updated | 2025-07-30T15:05:57Z
|
| user_id | cfc72fc0bc3d4ba09ef9756ea2fb6395
|
| volumes_attached | delete_on_termination='False',
id='b82238fb-fe80-41ac-b426-33de21eb6756'
|
+-------------------------------------+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
root@offline20241:~# virsh list --all
Id Name State
-----------------------------------
16 instance-0000001e running
root@offline20241:~#
```
### On the source node the vm was showing in virsh list as shutoff, after
reboot source compoute node, virsh stale entry got removed, however the
instance on the destination went to the stutoff state. both in server list and
virsh list.
```
Jul 30 15:04:30 offline20241.ct.lan nova-compute[2108454]: 2025-07-30
15:04:30.670 2108454 DEBUG oslo_concurrency.lockutils [None
req-a919b840-363e-42b5-8ff3-e2ecd871c94b - - - - - -] Acquiring lock
"969ca417-9111-42de-836c-eb883e52f131" by
"nova.compute.manager.ComputeManager._sync_power_states.<locals>._sync.<locals>.query_driver_power_state_and_sync"
inner
/openstack/venvs/nova-29.2.3.dev1/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402
Jul 30 15:04:30 offline20241.ct.lan nova-compute[2108454]: 2025-07-30
15:04:30.671 2108454 DEBUG oslo_concurrency.lockutils [None
req-a919b840-363e-42b5-8ff3-e2ecd871c94b - - - - - -] Lock
"969ca417-9111-42de-836c-eb883e52f131" acquired by
"nova.compute.manager.ComputeManager._sync_power_states.<locals>._sync.<locals>.query_driver_power_state_and_sync"
:: waited 0.001s inner
/openstack/venvs/nova-29.2.3.dev1/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407
Jul 30 15:04:30 offline20241.ct.lan nova-compute[2108454]: 2025-07-30
15:04:30.714 2108454 INFO nova.compute.manager [None
req-a919b840-363e-42b5-8ff3-e2ecd871c94b - - - - - -] [instance:
969ca417-9111-42de-836c-eb883e52f131] During _sync_instance_power_state the DB
power_state (0) does not match the vm_power_state from the hypervisor (1).
Updating power_state in the DB to match the hypervisor.
Jul 30 15:04:30 offline20241.ct.lan neutron-server[3870]: 2025-07-30
15:04:30.795 3870 WARNING neutron.db.agents_db [None
req-c0ba4060-e268-4940-9e06-6366966d1a48 - - - - - -] Agent healthcheck: found
4 dead agents out of 9:
Type Last heartbeat host
DHCP
agent 2025-07-23 12:31:59 offline20241
L3
agent 2025-07-23 12:32:16 offline20241
Open vSwitch
agent 2025-07-23 12:32:16 offline20241
Metadata
agent 2025-07-23 12:31:59 offline20241
Jul 30 15:04:30 offline20241.ct.lan nova-compute[2108454]: 2025-07-30
15:04:30.817 2108454 WARNING nova.compute.manager [None
req-a919b840-363e-42b5-8ff3-e2ecd871c94b - - - - - -] [instance:
969ca417-9111-42de-836c-eb883e52f131] Instance is not stopped. Calling the stop
API. Current vm_state: stopped, current task_state: None, original DB
power_state: 0, current VM power_state: 1
Jul 30 15:04:30 offline20241.ct.lan nova-compute[2108454]: 2025-07-30
15:04:30.818 2108454 DEBUG nova.compute.api [None
req-a919b840-363e-42b5-8ff3-e2ecd871c94b - - - - - -] [instance:
969ca417-9111-42de-836c-eb883e52f131] Going to try to stop instance force_stop
/openstack/venvs/nova-29.2.3.dev1/lib/python3.10/site-packages/nova/compute/api.py:2768
Jul 30 15:04:30 offline20241.ct.lan apache2[1957602]: 192.168.131.60 - -
[30/Jul/2025:15:04:30 +0000] "POST /v3/auth/tokens HTTP/1.1" 201 9818 "-"
"openstacksdk/3.0.0 keystoneauth1/5.6.1 python-requests/2.31.0 CPython/3.10.12"
Jul 30 15:04:30 offline20241.ct.lan haproxy[370326]: 192.168.131.200:55974
[30/Jul/2025:15:04:30.373] keystone_service-front-2
keystone_service-back/offline20241 0/0/0/550/550 201 9770 - - ---- 129/1/0/0/0
0/0 "POST /v3/auth/tokens HTTP/1.1"
Jul 30 15:04:30 offline20241.ct.lan nova-compute[2108454]: 2025-07-30
15:04:30.945 2108454 DEBUG oslo_concurrency.lockutils [None
req-a919b840-363e-42b5-8ff3-e2ecd871c94b - - - - - -] Lock
"969ca417-9111-42de-836c-eb883e52f131" "released" by
"nova.compute.manager.ComputeManager._sync_power_states.<locals>._sync.<locals>.query_driver_power_state_and_sync"
:: held 0.274s inner
/openstack/venvs/nova-29.2.3.dev1/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421
Jul 30 15:04:30 offline20241.ct.lan nova-compute[2108454]: 2025-07-30
15:04:30.975 2108454 DEBUG oslo_concurrency.lockutils [None
req-f69d5ab0-a27a-4006-8473-b4a9f670753d - - - - - -] Acquiring lock
"969ca417-9111-42de-836c-eb883e52f131" by
"nova.compute.manager.ComputeManager.stop_instance.<locals>.do_stop_instance"
inner
/openstack/venvs/nova-29.2.3.dev1/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402
Jul 30 15:04:30 offline20241.ct.lan nova-compute[2108454]: 2025-07-30
15:04:30.976 2108454 DEBUG oslo_concurrency.lockutils [None
req-f69d5ab0-a27a-4006-8473-b4a9f670753d - - - - - -] Lock
"969ca417-9111-42de-836c-eb883e52f131" acquired by
"nova.compute.manager.ComputeManager.stop_instance.<locals>.do_stop_instance"
:: waited 0.001s inner
/openstack/venvs/nova-29.2.3.dev1/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407
Jul 30 15:04:30 offline20241.ct.lan nova-compute[2108454]: 2025-07-30
15:04:30.977 2108454 DEBUG nova.compute.manager [None
req-f69d5ab0-a27a-4006-8473-b4a9f670753d - - - - - -] [instance:
969ca417-9111-42de-836c-eb883e52f131] Checking state _get_power_state
/openstack/venvs/nova-29.2.3.dev1/lib/python3.10/site-packages/nova/compute/manager.py:1782
Jul 30 15:04:30 offline20241.ct.lan nova-compute[2108454]: 2025-07-30
15:04:30.983 2108454 DEBUG nova.compute.manager [None
req-f69d5ab0-a27a-4006-8473-b4a9f670753d - - - - - -] [instance:
969ca417-9111-42de-836c-eb883e52f131] Stopping instance; current vm_state:
stopped, current task_state: powering-off, current DB power_state: 1, current
VM power_state: 1 do_stop_instance
/openstack/venvs/nova-29.2.3.dev1/lib/python3.10/site-packages/nova/compute/manager.py:3359
```
To manage notifications about this bug go to:
https://bugs.launchpad.net/masakari/+bug/2119126/+subscriptions
--
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : [email protected]
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help : https://help.launchpad.net/ListHelp