[Yahoo-eng-team] [Bug 2081847] Re: [RFE] Implement a new DHCP backend for Ironic and Neutron

2024-09-24 Thread Afonne-CID
** Also affects: neutron
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2081847

Title:
  [RFE] Implement a new DHCP backend for Ironic and Neutron

Status in Ironic:
  New
Status in neutron:
  New

Bug description:
  dnsmasq has become unreliable, causing issues while performing DHCP
  operations.

  Here's an RFE proposal to implement Kea DHCP as a new backend and
  alternative DHCP provider option in Ironic and Neutron.

  Kea is actively maintained by ISC, extensible with hooks modules and
  provides a good balance of modern features and compatibility with
  existing systems.

  My Findings about Kea DHCP: https://etherpad.opendev.org/p/research_cid
  RFE spec: 
  Proposed Change Draft:

To manage notifications about this bug go to:
https://bugs.launchpad.net/ironic/+bug/2081847/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2077195] Re: RFE: Add support for "igb" VIF model

2024-09-24 Thread Balazs Gibizer
Being tracked in https://blueprints.launchpad.net/nova/+spec/igb-vif-
model

** Changed in: nova
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/2077195

Title:
  RFE: Add support for "igb" VIF model

Status in OpenStack Compute (nova):
  Invalid

Bug description:
  qemu now supports a network driver called "igb" that simulates SR-IOV
  VFs[1]. Current versions of most OS distributions now include a
  version of qemu with this driver included.

  We would like to use that in the OpenStack gates to test SR-IOV
  functionality in Octavia.

  Currently nova does not allow the "igb" driver to be specified using 
"hw_vif_model=igb" because it is not on the list of valid VIF_MODELS here:
  
https://opendev.org/openstack/nova/src/branch/master/nova/network/model.py#L156

  This RFE is a request to add support for the igb driver to nova.

  I have also registered a blueprint for this here:
  https://blueprints.launchpad.net/nova/+spec/igb-vif-model

  [1] https://www.qemu.org/docs/master/system/devices/igb.html

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/2077195/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2081870] [NEW] neutron.tests.functional.agent.linux.test_process_monitor.TestProcessMonitor.test_respawn_handler failed with "RuntimeError: Not all children (re)spawned."

2024-09-24 Thread Ihar Hrachyshka
Public bug reported:

This happened in functional job here:
https://fc7da22d605cc199ea32-a05951a0ad29f0907f5578ce9b824040.ssl.cf1.rackcdn.com/929793/3/check/neutron-
functional-with-uwsgi/ba361f4/testr_results.html

Traceback:

```
ft1.1: 
neutron.tests.functional.agent.linux.test_process_monitor.TestProcessMonitor.test_respawn_handlertesttools.testresult.real._StringException:
 Traceback (most recent call last):
  File "/home/zuul/src/opendev.org/openstack/neutron/neutron/common/utils.py", 
line 740, in wait_until_true
eventlet.sleep(sleep)
  File 
"/home/zuul/src/opendev.org/openstack/neutron/.tox/dsvm-functional-gate/lib/python3.10/site-packages/eventlet/greenthread.py",
 line 38, in sleep
hub.switch()
  File 
"/home/zuul/src/opendev.org/openstack/neutron/.tox/dsvm-functional-gate/lib/python3.10/site-packages/eventlet/hubs/hub.py",
 line 310, in switch
return self.greenlet.switch()
eventlet.timeout.Timeout: 6 seconds

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "/home/zuul/src/opendev.org/openstack/neutron/neutron/tests/base.py", 
line 178, in func
return f(self, *args, **kwargs)
  File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/tests/functional/agent/linux/test_process_monitor.py",
 line 110, in test_respawn_handler
self.wait_for_all_children_spawned()
  File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/tests/functional/agent/linux/test_process_monitor.py",
 line 92, in wait_for_all_children_spawned
utils.wait_until_true(
  File "/home/zuul/src/opendev.org/openstack/neutron/neutron/common/utils.py", 
line 744, in wait_until_true
raise exception
RuntimeError: Not all children (re)spawned.
```

** Affects: neutron
 Importance: Undecided
 Status: New


** Tags: functional-tests gate-failure

** Tags added: functional-tests gate-failure

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2081870

Title:
  
neutron.tests.functional.agent.linux.test_process_monitor.TestProcessMonitor.test_respawn_handler
  failed with "RuntimeError: Not all children (re)spawned."

Status in neutron:
  New

Bug description:
  This happened in functional job here:
  
https://fc7da22d605cc199ea32-a05951a0ad29f0907f5578ce9b824040.ssl.cf1.rackcdn.com/929793/3/check/neutron-
  functional-with-uwsgi/ba361f4/testr_results.html

  Traceback:

  ```
  ft1.1: 
neutron.tests.functional.agent.linux.test_process_monitor.TestProcessMonitor.test_respawn_handlertesttools.testresult.real._StringException:
 Traceback (most recent call last):
File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/common/utils.py", line 
740, in wait_until_true
  eventlet.sleep(sleep)
File 
"/home/zuul/src/opendev.org/openstack/neutron/.tox/dsvm-functional-gate/lib/python3.10/site-packages/eventlet/greenthread.py",
 line 38, in sleep
  hub.switch()
File 
"/home/zuul/src/opendev.org/openstack/neutron/.tox/dsvm-functional-gate/lib/python3.10/site-packages/eventlet/hubs/hub.py",
 line 310, in switch
  return self.greenlet.switch()
  eventlet.timeout.Timeout: 6 seconds

  During handling of the above exception, another exception occurred:

  Traceback (most recent call last):
File "/home/zuul/src/opendev.org/openstack/neutron/neutron/tests/base.py", 
line 178, in func
  return f(self, *args, **kwargs)
File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/tests/functional/agent/linux/test_process_monitor.py",
 line 110, in test_respawn_handler
  self.wait_for_all_children_spawned()
File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/tests/functional/agent/linux/test_process_monitor.py",
 line 92, in wait_for_all_children_spawned
  utils.wait_until_true(
File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/common/utils.py", line 
744, in wait_until_true
  raise exception
  RuntimeError: Not all children (re)spawned.
  ```

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2081870/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2081868] [NEW] neutron.tests.unit.services.logapi.common.test_db_api.LoggingDBApiTestCase.test_get_logs_bound_sg fails with "webob.exc.HTTPClientError: The server could not compl

2024-09-24 Thread Ihar Hrachyshka
Public bug reported:

This failed here in gate on a patch that doesn't even touch this code:
https://5d6811448796d11c6047-40ec7889760df9011e4354628420e100.ssl.cf5.rackcdn.com/929793/3/check/openstack-
tox-py39/9d559dc/testr_results.html

There is nothing of interest in the full logs either:
https://5d6811448796d11c6047-40ec7889760df9011e4354628420e100.ssl.cf5.rackcdn.com/929793/3/check/openstack-
tox-py39/9d559dc/tox/py39/3-commands%5B1%5D.log

I don't know how frequent it is, but clearly it's some instability issue
in unit tests, which is not very common for this type of tests.

** Affects: neutron
 Importance: Undecided
 Status: New


** Tags: gate-failure unittest

** Tags added: gate-failure unittest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2081868

Title:
  
neutron.tests.unit.services.logapi.common.test_db_api.LoggingDBApiTestCase.test_get_logs_bound_sg
  fails with "webob.exc.HTTPClientError: The server could not comply
  with the request since it is either malformed or otherwise incorrect."

Status in neutron:
  New

Bug description:
  This failed here in gate on a patch that doesn't even touch this code:
  
https://5d6811448796d11c6047-40ec7889760df9011e4354628420e100.ssl.cf5.rackcdn.com/929793/3/check/openstack-
  tox-py39/9d559dc/testr_results.html

  There is nothing of interest in the full logs either:
  
https://5d6811448796d11c6047-40ec7889760df9011e4354628420e100.ssl.cf5.rackcdn.com/929793/3/check/openstack-
  tox-py39/9d559dc/tox/py39/3-commands%5B1%5D.log

  I don't know how frequent it is, but clearly it's some instability
  issue in unit tests, which is not very common for this type of tests.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2081868/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2081871] [NEW] Fullstack neutron.tests.fullstack.test_l3_agent.TestHAL3Agent.test_ha_router failed with "eventlet.timeout.Timeout: 90 seconds" while waiting on self._is_ha_router

2024-09-24 Thread Ihar Hrachyshka
Public bug reported:

This failed here:
https://zuul.opendev.org/t/openstack/build/1a8c094b185d40a4b0df78f4afe25d2e

Traceback:

```

  File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/common/utils.py", line 
740, in wait_until_true
eventlet.sleep(sleep)

  File 
"/home/zuul/src/opendev.org/openstack/neutron/.tox/dsvm-fullstack-gate/lib/python3.10/site-packages/eventlet/greenthread.py",
 line 38, in sleep
hub.switch()

  File 
"/home/zuul/src/opendev.org/openstack/neutron/.tox/dsvm-fullstack-gate/lib/python3.10/site-packages/eventlet/hubs/hub.py",
 line 310, in switch
return self.greenlet.switch()

eventlet.timeout.Timeout: 90 seconds


During handling of the above exception, another exception occurred:


Traceback (most recent call last):

  File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/tests/base.py", line 178, 
in func
return f(self, *args, **kwargs)

  File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/tests/fullstack/test_l3_agent.py",
 line 420, in test_ha_router
common_utils.wait_until_true(

  File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/common/utils.py", line 
745, in wait_until_true
raise WaitTimeout(_("Timed out after %d seconds") % timeout)

neutron.common.utils.WaitTimeout: Timed out after 90 seconds
```

The output of the failure is not very helpful. One idea here to improve
the log message could be to make wait_until_true and/or the passed
predicate function to report additional information on timeout. Then the
next time it hits, we may have more information.

** Affects: neutron
 Importance: Undecided
 Status: New


** Tags: fullstack gate-failure l3-ha

** Tags added: fullstack gate-failure

** Tags added: l3-ha

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2081871

Title:
  Fullstack
  neutron.tests.fullstack.test_l3_agent.TestHAL3Agent.test_ha_router
  failed with "eventlet.timeout.Timeout: 90 seconds" while waiting on
  self._is_ha_router_active_on_one_agent

Status in neutron:
  New

Bug description:
  This failed here:
  https://zuul.opendev.org/t/openstack/build/1a8c094b185d40a4b0df78f4afe25d2e

  Traceback:

  ```

File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/common/utils.py", line 
740, in wait_until_true
  eventlet.sleep(sleep)

File 
"/home/zuul/src/opendev.org/openstack/neutron/.tox/dsvm-fullstack-gate/lib/python3.10/site-packages/eventlet/greenthread.py",
 line 38, in sleep
  hub.switch()

File 
"/home/zuul/src/opendev.org/openstack/neutron/.tox/dsvm-fullstack-gate/lib/python3.10/site-packages/eventlet/hubs/hub.py",
 line 310, in switch
  return self.greenlet.switch()

  eventlet.timeout.Timeout: 90 seconds

  
  During handling of the above exception, another exception occurred:

  
  Traceback (most recent call last):

File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/tests/base.py", line 178, 
in func
  return f(self, *args, **kwargs)

File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/tests/fullstack/test_l3_agent.py",
 line 420, in test_ha_router
  common_utils.wait_until_true(

File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/common/utils.py", line 
745, in wait_until_true
  raise WaitTimeout(_("Timed out after %d seconds") % timeout)

  neutron.common.utils.WaitTimeout: Timed out after 90 seconds
  ```

  The output of the failure is not very helpful. One idea here to
  improve the log message could be to make wait_until_true and/or the
  passed predicate function to report additional information on timeout.
  Then the next time it hits, we may have more information.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2081871/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1483639] Re: Nova and Horizon allow inappropriate actions to be performed on baremetal nodes

2024-09-24 Thread Jay Faulkner
** Changed in: ironic
   Status: Incomplete => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1483639

Title:
  Nova and Horizon allow inappropriate actions to be performed on
  baremetal nodes

Status in OpenStack Dashboard (Horizon):
  Confirmed
Status in Ironic:
  Invalid
Status in OpenStack Compute (nova):
  Confirmed

Bug description:
  Ironic baremetal nodes do not support all variety of operations that
  nova virtual instances do.

  But Nova and Horizon still offers to perform actions with ironic
  baremetal nodes that can be applied to virtual instances only.

  Examples of steps:
  root@node-1:~# nova pause NEW1
  root@node-1:~# nova suspend NEW1
  As result Nova silently accepts commands without any warning or error 
messages. Same actions can be performed via Horizon with green "Success" popup.
  Also see list of actions over baremetal node on screenshot.

  One more example:
  Backup to image baremetal instance: 
  root@node-1:~# nova image-create --poll --show NEW1 IMAGENEW1
  Server snapshotting... 0% complete   

  and process stalls showing 0%  in console infinitely. 
  Expected that nova will not try do this with baremetal node at all. 

  Currently baremetal nodes do not support following actions:

  a) Create Snapshot
  b) Pause 
  c) Suspend
  d) Migrate
  e) Live Migrate
  f) Only one kind or reboot should be supported (Hard reboot?)
  g) Resize 

  These actions should be disabled for baremetal machines in Nova and Horizon.
  Currently there are no destructive aftermaths detected, therefore this bug 
affects user by confusing him when using Horizon and Nova.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1483639/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2081859] [NEW] Nova not initializing os-brick

2024-09-24 Thread Gorka Eguileor
Public bug reported:

In the Zed release os-brick started needing to be initialized by calling
a `setup` method before the library could be used.

At that time there was only 1 feature that depended on it and it was
possible to introduce a failsafe for that instance so things wouldn't
break.

In the Antelope release that failsafe should have been removed from os-
brick and all projects should have been calling the `setup` method.

Currently nova is not initializing os-brick, so if os-brick removes the
failsafe the behavior in os-brick locks will break backward
compatibility.

Related os-brick patch: https://review.opendev.org/c/openstack/os-
brick/+/849324

** Affects: nova
 Importance: Undecided
 Status: In Progress

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/2081859

Title:
  Nova not initializing os-brick

Status in OpenStack Compute (nova):
  In Progress

Bug description:
  In the Zed release os-brick started needing to be initialized by
  calling a `setup` method before the library could be used.

  At that time there was only 1 feature that depended on it and it was
  possible to introduce a failsafe for that instance so things wouldn't
  break.

  In the Antelope release that failsafe should have been removed from
  os-brick and all projects should have been calling the `setup` method.

  Currently nova is not initializing os-brick, so if os-brick removes
  the failsafe the behavior in os-brick locks will break backward
  compatibility.

  Related os-brick patch: https://review.opendev.org/c/openstack/os-
  brick/+/849324

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/2081859/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2081853] [NEW] Booting two VMs with anti-affinity in parallel to the same host results in both failing

2024-09-24 Thread Balazs Gibizer
Public bug reported:

The compute manager late anti-affinity policy check rejects both
parallel VM boot requests even though one of them could be accepted to
the host.

To reproduce:
* create server group with anti-affinity policy
* select a single compute and disable the rest of your computes
* boot two VMs in parallel

Expected:
One of the two VMs succeeds to boot the other VM fails with NoValidHost.

Actual:
If you are (un)lucky then both VMs will fail with 
nova.exception.GroupAffinityViolation

```
❯ journalctl -D 
sosreport-compute-1-2024-09-17-tzgxrpu/var/log/journal/730eba01f47f493698df59515d1c213a
  -u edpm_nova_compute | grep 9d115f6b-bb02-4390-a161-15fb8f83c0cc | grep 
nova.exception.GroupAffinityViolation:
Sep 17 02:05:36 compute-1 nova_compute[84038]: 2024-09-17 00:05:36.406 2 ERROR 
nova.compute.manager [None req-a5316266-aca0-4d11-90f9-631e26d058ab 
188fff18565b4e46b0c04391ec532b3e b698d1d3bfeb4a75bf32b7a80d19dd46 - - default 
default] [instance: 9d115f6b-bb02-4390-a161-15fb8f83c0cc] Failed to build and 
run instance: nova.exception.GroupAffinityViolation: Anti-affinity instance 
group policy was violated
Sep 17 02:05:36 compute-1 nova_compute[84038]: 2024-09-17 00:05:36.406 2 ERROR 
nova.compute.manager [instance: 9d115f6b-bb02-4390-a161-15fb8f83c0cc] 
nova.exception.GroupAffinityViolation: Anti-affinity instance group policy was 
violated

❯ journalctl -D 
sosreport-compute-1-2024-09-17-tzgxrpu/var/log/journal/730eba01f47f493698df59515d1c213a
  -u edpm_nova_compute | grep ea192e6a-4685-45ae-839b-315dfd36697d | grep 
nova.exception.GroupAffinityViolation
Sep 17 02:05:36 compute-1 nova_compute[84038]: 2024-09-17 00:05:36.132 2 ERROR 
nova.compute.manager [None req-b37d5098-75bf-4a3c-a85d-6f2ccdf0104f 
188fff18565b4e46b0c04391ec532b3e b698d1d3bfeb4a75bf32b7a80d19dd46 - - default 
default] [instance: ea192e6a-4685-45ae-839b-315dfd36697d] Failed to build and 
run instance: nova.exception.GroupAffinityViolation: Anti-affinity instance 
group policy was violated
Sep 17 02:05:36 compute-1 nova_compute[84038]: 2024-09-17 00:05:36.132 2 ERROR 
nova.compute.manager [instance: ea192e6a-4685-45ae-839b-315dfd36697d] 
nova.exception.GroupAffinityViolation: Anti-affinity instance group policy was 
violated
```

There is a functional reproduce pushed in
https://review.opendev.org/c/openstack/nova/+/930326

** Affects: nova
 Importance: Undecided
 Status: New


** Tags: compute scheduler

** Tags added: compute scheduler

** Description changed:

  The compute manager late anti-affinity policy check rejects both
  parallel VM boot requests even though one of them could be accepted to
  the host.
  
  To reproduce:
  * create server group with anti-affinity policy
  * select a single compute and disable the rest of your computes
  * boot two VMs in parallel
  
  Expected:
  One of the two VMs succeeds to boot the other VM fails with NoValidHost.
  
  Actual:
- If you are (un)lucky then both VM will fail with 
nova.exception.GroupAffinityViolation
+ If you are (un)lucky then both VMs will fail with 
nova.exception.GroupAffinityViolation
  
  ```
  ❯ journalctl -D 
sosreport-compute-1-2024-09-17-tzgxrpu/var/log/journal/730eba01f47f493698df59515d1c213a
  -u edpm_nova_compute | grep 9d115f6b-bb02-4390-a161-15fb8f83c0cc | grep 
nova.exception.GroupAffinityViolation:
  Sep 17 02:05:36 compute-1 nova_compute[84038]: 2024-09-17 00:05:36.406 2 
ERROR nova.compute.manager [None req-a5316266-aca0-4d11-90f9-631e26d058ab 
188fff18565b4e46b0c04391ec532b3e b698d1d3bfeb4a75bf32b7a80d19dd46 - - default 
default] [instance: 9d115f6b-bb02-4390-a161-15fb8f83c0cc] Failed to build and 
run instance: nova.exception.GroupAffinityViolation: Anti-affinity instance 
group policy was violated
  Sep 17 02:05:36 compute-1 nova_compute[84038]: 2024-09-17 00:05:36.406 2 
ERROR nova.compute.manager [instance: 9d115f6b-bb02-4390-a161-15fb8f83c0cc] 
nova.exception.GroupAffinityViolation: Anti-affinity instance group policy was 
violated
  
  ❯ journalctl -D 
sosreport-compute-1-2024-09-17-tzgxrpu/var/log/journal/730eba01f47f493698df59515d1c213a
  -u edpm_nova_compute | grep ea192e6a-4685-45ae-839b-315dfd36697d | grep 
nova.exception.GroupAffinityViolation
  Sep 17 02:05:36 compute-1 nova_compute[84038]: 2024-09-17 00:05:36.132 2 
ERROR nova.compute.manager [None req-b37d5098-75bf-4a3c-a85d-6f2ccdf0104f 
188fff18565b4e46b0c04391ec532b3e b698d1d3bfeb4a75bf32b7a80d19dd46 - - default 
default] [instance: ea192e6a-4685-45ae-839b-315dfd36697d] Failed to build and 
run instance: nova.exception.GroupAffinityViolation: Anti-affinity instance 
group policy was violated
  Sep 17 02:05:36 compute-1 nova_compute[84038]: 2024-09-17 00:05:36.132 2 
ERROR nova.compute.manager [instance: ea192e6a-4685-45ae-839b-315dfd36697d] 
nova.exception.GroupAffinityViolation: Anti-affinity instance group policy was 
violated
  ```
  
  There is a functional reproduce pushed in
  https://review.opendev.org/c/openstack/nova/+/930326

-- 
You received this bug n

[Yahoo-eng-team] [Bug 2081879] [NEW] Allow the default for "Delete Volume on Instance Delete" instance creation option to be configurable

2024-09-24 Thread Tim Bishop
Public bug reported:

When creating an instance with the "Create New Volume" option selected
(the default), there's an option to "Delete Volume on Instance Delete"
which defaults to "No". This default is hard-coded, and is a sensible
default, but there are situations when an operator may wish to change
the default to "Yes". I'm in that position where we have a more
controlled use case and don't want users accidentally leaving orphaned
volumes behind.

A simple change to allow this is:

--- 
horizon-24.0.0.orig/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/launch-instance-model.service.js
+++ 
horizon-24.0.0/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/launch-instance-model.service.js
@@ -309,6 +309,9 @@
   if ('default_availability_zone' in defaults) {
 model.default_availability_zone = defaults.default_availability_zone;
   }
+  if ('vol_delete_on_instance_delete' in defaults) {
+model.newInstanceSpec.vol_delete_on_instance_delete = 
defaults.vol_delete_on_instance_delete;
+  }
 }

 /**
--- 
horizon-24.0.0.orig/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/source/source.controller.js
+++ 
horizon-24.0.0/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/source/source.controller.js
@@ -471,7 +471,7 @@
   } else {
 $scope.model.newInstanceSpec.vol_create = false;
   }
-  $scope.model.newInstanceSpec.vol_delete_on_instance_delete = false;
+  //$scope.model.newInstanceSpec.vol_delete_on_instance_delete = false;
   changeBootSource(selectedSource, preSelection);
   validateBootSourceType();
 }


This doesn't alter the existing behaviour by default, but does allow 
local_settings.py to optionally contain:

LAUNCH_INSTANCE_DEFAULTS.update({
'vol_delete_on_instance_delete': True,
})

Which then changes the default for "Delete Volume on Instance Delete" to
"Yes" in the web interface.

Change tested against 24.0.0.

** Affects: horizon
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/2081879

Title:
  Allow the default for "Delete Volume on Instance Delete" instance
  creation option to be configurable

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  When creating an instance with the "Create New Volume" option selected
  (the default), there's an option to "Delete Volume on Instance Delete"
  which defaults to "No". This default is hard-coded, and is a sensible
  default, but there are situations when an operator may wish to change
  the default to "Yes". I'm in that position where we have a more
  controlled use case and don't want users accidentally leaving orphaned
  volumes behind.

  A simple change to allow this is:

  --- 
horizon-24.0.0.orig/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/launch-instance-model.service.js
  +++ 
horizon-24.0.0/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/launch-instance-model.service.js
  @@ -309,6 +309,9 @@
 if ('default_availability_zone' in defaults) {
   model.default_availability_zone = defaults.default_availability_zone;
 }
  +  if ('vol_delete_on_instance_delete' in defaults) {
  +model.newInstanceSpec.vol_delete_on_instance_delete = 
defaults.vol_delete_on_instance_delete;
  +  }
   }

   /**
  --- 
horizon-24.0.0.orig/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/source/source.controller.js
  +++ 
horizon-24.0.0/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/source/source.controller.js
  @@ -471,7 +471,7 @@
 } else {
   $scope.model.newInstanceSpec.vol_create = false;
 }
  -  $scope.model.newInstanceSpec.vol_delete_on_instance_delete = false;
  +  //$scope.model.newInstanceSpec.vol_delete_on_instance_delete = false;
 changeBootSource(selectedSource, preSelection);
 validateBootSourceType();
   }

  
  This doesn't alter the existing behaviour by default, but does allow 
local_settings.py to optionally contain:

  LAUNCH_INSTANCE_DEFAULTS.update({
  'vol_delete_on_instance_delete': True,
  })

  Which then changes the default for "Delete Volume on Instance Delete"
  to "Yes" in the web interface.

  Change tested against 24.0.0.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/2081879/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2081878] [NEW] Allow the default for "Delete Volume on Instance Delete" instance creation option to be configurable

2024-09-24 Thread Tim Bishop
Public bug reported:

When creating an instance with the "Create New Volume" option selected
(the default), there's an option to "Delete Volume on Instance Delete"
which defaults to "No". This default is hard-coded, and is a sensible
default, but there are situations when an operator may wish to change
the default to "Yes". I'm in that position where we have a more
controlled use case and don't want users accidentally leaving orphaned
volumes behind.

A simple change to allow this is:

--- 
horizon-24.0.0.orig/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/launch-instance-model.service.js
+++ 
horizon-24.0.0/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/launch-instance-model.service.js
@@ -309,6 +309,9 @@
   if ('default_availability_zone' in defaults) {
 model.default_availability_zone = defaults.default_availability_zone;
   }
+  if ('vol_delete_on_instance_delete' in defaults) {
+model.newInstanceSpec.vol_delete_on_instance_delete = 
defaults.vol_delete_on_instance_delete;
+  }
 }

 /**
--- 
horizon-24.0.0.orig/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/source/source.controller.js
+++ 
horizon-24.0.0/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/source/source.controller.js
@@ -471,7 +471,7 @@
   } else {
 $scope.model.newInstanceSpec.vol_create = false;
   }
-  $scope.model.newInstanceSpec.vol_delete_on_instance_delete = false;
+  //$scope.model.newInstanceSpec.vol_delete_on_instance_delete = false;
   changeBootSource(selectedSource, preSelection);
   validateBootSourceType();
 }


This doesn't alter the existing behaviour by default, but does allow 
local_settings.py to optionally contain:

LAUNCH_INSTANCE_DEFAULTS.update({
'vol_delete_on_instance_delete': True,
})

Which then changes the default for "Delete Volume on Instance Delete" to
"Yes" in the web interface.

Change tested against 24.0.0.

** Affects: horizon
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/2081878

Title:
  Allow the default for "Delete Volume on Instance Delete" instance
  creation option to be configurable

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  When creating an instance with the "Create New Volume" option selected
  (the default), there's an option to "Delete Volume on Instance Delete"
  which defaults to "No". This default is hard-coded, and is a sensible
  default, but there are situations when an operator may wish to change
  the default to "Yes". I'm in that position where we have a more
  controlled use case and don't want users accidentally leaving orphaned
  volumes behind.

  A simple change to allow this is:

  --- 
horizon-24.0.0.orig/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/launch-instance-model.service.js
  +++ 
horizon-24.0.0/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/launch-instance-model.service.js
  @@ -309,6 +309,9 @@
 if ('default_availability_zone' in defaults) {
   model.default_availability_zone = defaults.default_availability_zone;
 }
  +  if ('vol_delete_on_instance_delete' in defaults) {
  +model.newInstanceSpec.vol_delete_on_instance_delete = 
defaults.vol_delete_on_instance_delete;
  +  }
   }

   /**
  --- 
horizon-24.0.0.orig/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/source/source.controller.js
  +++ 
horizon-24.0.0/openstack_dashboard/dashboards/project/static/dashboard/project/workflow/launch-instance/source/source.controller.js
  @@ -471,7 +471,7 @@
 } else {
   $scope.model.newInstanceSpec.vol_create = false;
 }
  -  $scope.model.newInstanceSpec.vol_delete_on_instance_delete = false;
  +  //$scope.model.newInstanceSpec.vol_delete_on_instance_delete = false;
 changeBootSource(selectedSource, preSelection);
 validateBootSourceType();
   }

  
  This doesn't alter the existing behaviour by default, but does allow 
local_settings.py to optionally contain:

  LAUNCH_INSTANCE_DEFAULTS.update({
  'vol_delete_on_instance_delete': True,
  })

  Which then changes the default for "Delete Volume on Instance Delete"
  to "Yes" in the web interface.

  Change tested against 24.0.0.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/2081878/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2079850] Re: Ephemeral with vfat format fails inspection

2024-09-24 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/c/openstack/oslo.utils/+/928448
Committed: 
https://opendev.org/openstack/oslo.utils/commit/3c33e37d64e44addc9a818bd556f5919ed2e9002
Submitter: "Zuul (22348)"
Branch:master

commit 3c33e37d64e44addc9a818bd556f5919ed2e9002
Author: Dan Smith 
Date:   Fri Sep 6 07:51:13 2024 -0700

Avoid detecting FAT VBR as an MBR

The 1980s FAT filesystem has a VBR in the first sector, which looks
almost exactly like an MBR with zero partitions. To avoid detecting
these as MBRs, look for some extra attributes that indicate that the
structure is a VBR and avoid matching it as a GPT/MBR in that case.

We can add an inspector for this as a separate thing, but at the
moment we don't have that immediate need.

Closes-Bug: #2079850
Change-Id: Ibad87743b5a3b6469bd708d4caafe7911b045855


** Changed in: oslo.utils
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/2079850

Title:
  Ephemeral with vfat format fails inspection

Status in OpenStack Compute (nova):
  Fix Released
Status in oslo.utils:
  Fix Released

Bug description:
  When configured to format ephemerals as vfat, we get this failure:

  Sep 03 17:34:28 compute-2 nova_compute[133243]: 2024-09-03 17:34:28.358 2 
DEBUG oslo_utils.imageutils.format_inspector [None 
req-fcf3a278-3417-4a6d-8b10-66e91ca1677d 60ed4d3e522640b6ad19633b28c5b5bb 
ae43aec9c3c242a785c8256abdda1747 - - default default] Format inspector failed, 
aborting: Signature KDMV not found: b'\xebX\x90m' _process_chunk 
/usr/lib/python3.9/site-packages/oslo_utils/imageutils/format_inspector.py:1302
  Sep 03 17:34:28 compute-2 nova_compute[133243]: 2024-09-03 17:34:28.365 2 
DEBUG oslo_utils.imageutils.format_inspector [None 
req-fcf3a278-3417-4a6d-8b10-66e91ca1677d 60ed4d3e522640b6ad19633b28c5b5bb 
ae43aec9c3c242a785c8256abdda1747 - - default default] Format inspector failed, 
aborting: Region signature not found at 3 _process_chunk 
/usr/lib/python3.9/site-packages/oslo_utils/imageutils/format_inspector.py:1302
  Sep 03 17:34:28 compute-2 nova_compute[133243]: 2024-09-03 17:34:28.366 2 
WARNING oslo_utils.imageutils.format_inspector [None 
req-fcf3a278-3417-4a6d-8b10-66e91ca1677d 60ed4d3e522640b6ad19633b28c5b5bb 
ae43aec9c3c242a785c8256abdda1747 - - default default] Safety check mbr on gpt 
failed because GPT MBR has no partitions defined: 
oslo_utils.imageutils.format_inspector.SafetyViolation: GPT MBR has no 
partitions defined
  Sep 03 17:34:28 compute-2 nova_compute[133243]: 2024-09-03 17:34:28.366 2 
WARNING nova.virt.libvirt.imagebackend [None 
req-fcf3a278-3417-4a6d-8b10-66e91ca1677d 60ed4d3e522640b6ad19633b28c5b5bb 
ae43aec9c3c242a785c8256abdda1747 - - default default] Base image 
/var/lib/nova/instances/_base/ephemeral_1_0706d66 failed safety check: Safety 
checks failed: mbr: oslo_utils.imageutils.format_inspector.SafetyCheckFailed: 
Safety checks failed: mbr
  Sep 03 17:34:28 compute-2 nova_compute[133243]: 2024-09-03 17:34:28.367 2 
ERROR nova.compute.manager [None req-fcf3a278-3417-4a6d-8b10-66e91ca1677d 
60ed4d3e522640b6ad19633b28c5b5bb ae43aec9c3c242a785c8256abdda1747 - - default 
default] [instance: 263ccd01-10b1-46a6-9f81-a6fc27c7177a] Instance failed to 
spawn: nova.exception.InvalidDiskInfo: Disk info file is invalid: Base image 
failed safety check
  Sep 03 17:34:28 compute-2 nova_compute[133243]: 2024-09-03 17:34:28.367 2 
ERROR nova.compute.manager [instance: 263ccd01-10b1-46a6-9f81-a6fc27c7177a] 
Traceback (most recent call last):
  Sep 03 17:34:28 compute-2 nova_compute[133243]: 2024-09-03 17:34:28.367 2 
ERROR nova.compute.manager [instance: 263ccd01-10b1-46a6-9f81-a6fc27c7177a]   
File "/usr/lib/python3.9/site-packages/nova/virt/libvirt/imagebackend.py", line 
685, in create_image
  Sep 03 17:34:28 compute-2 nova_compute[133243]: 2024-09-03 17:34:28.367 2 
ERROR nova.compute.manager [instance: 263ccd01-10b1-46a6-9f81-a6fc27c7177a] 
inspector.safety_check()
  Sep 03 17:34:28 compute-2 nova_compute[133243]: 2024-09-03 17:34:28.367 2 
ERROR nova.compute.manager [instance: 263ccd01-10b1-46a6-9f81-a6fc27c7177a]   
File 
"/usr/lib/python3.9/site-packages/oslo_utils/imageutils/format_inspector.py", 
line 430, in safety_check
  Sep 03 17:34:28 compute-2 nova_compute[133243]: 2024-09-03 17:34:28.367 2 
ERROR nova.compute.manager [instance: 263ccd01-10b1-46a6-9f81-a6fc27c7177a] 
raise SafetyCheckFailed(failures)
  Sep 03 17:34:28 compute-2 nova_compute[133243]: 2024-09-03 17:34:28.367 2 
ERROR nova.compute.manager [instance: 263ccd01-10b1-46a6-9f81-a6fc27c7177a] 
oslo_utils.imageutils.format_inspector.SafetyCheckFailed: Safety checks failed: 
mbr
  Sep 03 17:34:28 compute-2 nova_compute[133243]: 2024-09-03 17:34:28.367 2 
ERROR nova.compute.manager [instance: 263ccd01-10b1-46a6-9f81-a6fc27c7177a] 
  Sep 03 17:34:28 compute-2 nova_c

[Yahoo-eng-team] [Bug 2081945] [NEW] [postgresql] AttributeError: 'NoneType' object has no attribute 'id'

2024-09-24 Thread Rodolfo Alonso
Public bug reported:

Neutron API failing with the following error:
```
AttributeError: 'NoneType' object has no attribute 'id'
```

Log: https://e22ff80d6617e0aebdbb-
fab3512704550cb902a3eea3d4491f2b.ssl.cf1.rackcdn.com/periodic/opendev.org/openstack/neutron/master/neutron-
ovn-tempest-postgres-full/6b355b7/controller/logs/screen-q-svc.txt

Snippet: https://paste.opendev.org/show/bx6fuJBVC2RvjVHDH7Ug/

This bug is related to https://bugs.launchpad.net/neutron/+bug/2078787.

** Affects: neutron
 Importance: Undecided
 Assignee: Rodolfo Alonso (rodolfo-alonso-hernandez)
 Status: New

** Changed in: neutron
 Assignee: (unassigned) => Rodolfo Alonso (rodolfo-alonso-hernandez)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2081945

Title:
  [postgresql] AttributeError: 'NoneType' object has no attribute 'id'

Status in neutron:
  New

Bug description:
  Neutron API failing with the following error:
  ```
  AttributeError: 'NoneType' object has no attribute 'id'
  ```

  Log: https://e22ff80d6617e0aebdbb-
  
fab3512704550cb902a3eea3d4491f2b.ssl.cf1.rackcdn.com/periodic/opendev.org/openstack/neutron/master/neutron-
  ovn-tempest-postgres-full/6b355b7/controller/logs/screen-q-svc.txt

  Snippet: https://paste.opendev.org/show/bx6fuJBVC2RvjVHDH7Ug/

  This bug is related to
  https://bugs.launchpad.net/neutron/+bug/2078787.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2081945/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp