[Yahoo-eng-team] [Bug 1260224] [NEW] agent model migration script does not contain ml2 plugin
Public bug reported: + mysql -uroot -proot -h127.0.0.1 -e 'DROP DATABASE IF EXISTS neutron_ml2;' + mysql -uroot -proot -h127.0.0.1 -e 'CREATE DATABASE neutron_ml2 CHARACTER SET utf8;' + /usr/local/bin/neutron-db-manage --config-file /etc/neutron/neutron.conf --config-file /etc/neutron/plugins/ml2/ml2_conf.ini upgrade head No handlers could be found for logger "neutron.common.legacy" INFO [alembic.migration] Context impl MySQLImpl. INFO [alembic.migration] Will assume non-transactional DDL. INFO [alembic.migration] Running upgrade None -> folsom, folsom initial database INFO [alembic.migration] Running upgrade folsom -> 2c4af419145b, l3_support INFO [alembic.migration] Running upgrade 2c4af419145b -> 5a875d0e5c, ryu INFO [alembic.migration] Running upgrade 5a875d0e5c -> 48b6f43f7471, DB support for service types INFO [alembic.migration] Running upgrade 48b6f43f7471 -> 3cb5d900c5de, security_groups INFO [alembic.migration] Running upgrade 3cb5d900c5de -> 1d76643bcec4, nvp_netbinding INFO [alembic.migration] Running upgrade 1d76643bcec4 -> 2a6d0b51f4bb, cisco plugin cleanup INFO [alembic.migration] Running upgrade 2a6d0b51f4bb -> 1b693c095aa3, Quota ext support added in Grizzly INFO [alembic.migration] Running upgrade 1b693c095aa3 -> 1149d7de0cfa, inital port security INFO [alembic.migration] Running upgrade 1149d7de0cfa -> 49332180ca96, ryu plugin update INFO [alembic.migration] Running upgrade 49332180ca96 -> 38335592a0dc, nvp_portmap INFO [alembic.migration] Running upgrade 38335592a0dc -> 54c2c487e913, 'DB support for load balancing service INFO [alembic.migration] Running upgrade 54c2c487e913 -> 45680af419f9, nvp_qos INFO [alembic.migration] Running upgrade 45680af419f9 -> 1c33fa3cd1a1, Support routing table configuration on Router INFO [alembic.migration] Running upgrade 1c33fa3cd1a1 -> 363468ac592c, nvp_network_gw INFO [alembic.migration] Running upgrade 363468ac592c -> 511471cc46b, Add agent management extension model support INFO [alembic.migration] Running upgrade 511471cc46b -> 3b54bf9e29f7, NEC plugin sharednet INFO [alembic.migration] Running upgrade 3b54bf9e29f7 -> 4692d074d587, agent scheduler INFO [alembic.migration] Running upgrade 4692d074d587 -> 1341ed32cc1e, nvp_net_binding INFO [alembic.migration] Running upgrade 1341ed32cc1e -> grizzly, grizzly INFO [alembic.migration] Running upgrade grizzly -> f489cf14a79c, DB support for load balancing service (havana) INFO [alembic.migration] Running upgrade f489cf14a79c -> 176a85fc7d79, Add portbindings db INFO [alembic.migration] Running upgrade 176a85fc7d79 -> 32b517556ec9, remove TunnelIP model INFO [alembic.migration] Running upgrade 32b517556ec9 -> 128e042a2b68, ext_gw_mode INFO [alembic.migration] Running upgrade 128e042a2b68 -> 5ac71e65402c, ml2_initial INFO [alembic.migration] Running upgrade 5ac71e65402c -> 3cbf70257c28, nvp_mac_learning INFO [alembic.migration] Running upgrade 3cbf70257c28 -> 5918cbddab04, add tables for router rules support INFO [alembic.migration] Running upgrade 5918cbddab04 -> 3cabb850f4a5, Table to track port to host associations INFO [alembic.migration] Running upgrade 3cabb850f4a5 -> b7a8863760e, Remove cisco_vlan_bindings table INFO [alembic.migration] Running upgrade b7a8863760e -> 13de305df56e, nec_add_pf_name INFO [alembic.migration] Running upgrade 13de305df56e -> 20ae61555e95, DB Migration for ML2 GRE Type Driver INFO [alembic.migration] Running upgrade 20ae61555e95 -> 477a4488d3f4, DB Migration for ML2 VXLAN Type Driver INFO [alembic.migration] Running upgrade 477a4488d3f4 -> 2032abe8edac, LBaaS add status description INFO [alembic.migration] Running upgrade 2032abe8edac -> 52c5e4a18807, LBaaS Pool scheduler INFO [alembic.migration] Running upgrade 52c5e4a18807 -> 557edfc53098, New service types framework (service providers) INFO [alembic.migration] Running upgrade 557edfc53098 -> e6b16a30d97, Add cisco_provider_networks table INFO [alembic.migration] Running upgrade e6b16a30d97 -> 39cf3f799352, FWaaS Havana-2 model INFO [alembic.migration] Running upgrade 39cf3f799352 -> 52ff27f7567a, Support for VPNaaS INFO [alembic.migration] Running upgrade 52ff27f7567a -> 11c6e18605c8, Pool Monitor status field INFO [alembic.migration] Running upgrade 11c6e18605c8 -> 35c7c198ddea, remove status from HealthMonitor INFO [alembic.migration] Running upgrade 35c7c198ddea -> 263772d65691, Cisco plugin db cleanup part II INFO [alembic.migration] Running upgrade 263772d65691 -> c88b6b5fea3, Cisco N1KV tables INFO [alembic.migration] Running upgrade c88b6b5fea3 -> f9263d6df56, remove_dhcp_lease INFO [alembic.migration] Running upgrade f9263d6df56 -> 569e98a8132b, metering INFO [alembic.migration] Running upgrade 569e98a8132b -> 86cf4d88bd3, remove bigswitch port tracking table INFO [alembic.migration] Running upgrade 86cf4d88bd3 -> 3c6e57a23db4, add multiprovider INFO [alembic.migration] Running upgrade 3c6e57a23db4 -> 63afba73813, Add u
[Yahoo-eng-team] [Bug 1260227] [NEW] Pre-select the network when there's only one.
Public bug reported: When launching instance with neutron, network field is required. If there is only one, it should be pre-selected. ** Affects: horizon Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1260227 Title: Pre-select the network when there's only one. Status in OpenStack Dashboard (Horizon): New Bug description: When launching instance with neutron, network field is required. If there is only one, it should be pre-selected. To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1260227/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260228] [NEW] Pre-select the network when there's only one.
*** This bug is a duplicate of bug 1260227 *** https://bugs.launchpad.net/bugs/1260227 Public bug reported: When launching instance with neutron, network field is required. If there is only one, it should be pre-selected. ** Affects: horizon Importance: Undecided Status: New ** This bug has been marked a duplicate of bug 1260227 Pre-select the network when there's only one. -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1260228 Title: Pre-select the network when there's only one. Status in OpenStack Dashboard (Horizon): New Bug description: When launching instance with neutron, network field is required. If there is only one, it should be pre-selected. To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1260228/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260232] [NEW] db migration on unique constraint should not be applied to all plugins
Public bug reported: There is an assumption that all "neutron plugins" creates the plumgrid_neutron.agents table, which is not the case. I just tested big switch and plumgrid plugins and they are failing: INFO [alembic.migration] Running upgrade havana -> e197124d4b9, add unique constraint to members INFO [alembic.migration] Running upgrade e197124d4b9 -> 1fcfc149aca4, Add a unique constraint on (agent_type, host) columns to prevent a race condition when an agent entry is 'upserted'. Traceback (most recent call last): File "/usr/local/bin/neutron-db-manage", line 10, in sys.exit(main()) File "/opt/stack/neutron/neutron/db/migration/cli.py", line 143, in main CONF.command.func(config, CONF.command.name) File "/opt/stack/neutron/neutron/db/migration/cli.py", line 80, in do_upgrade_downgrade do_alembic_command(config, cmd, revision, sql=CONF.command.sql) File "/opt/stack/neutron/neutron/db/migration/cli.py", line 59, in do_alembic_command getattr(alembic_command, cmd)(config, *args, **kwargs) File "/usr/local/lib/python2.7/dist-packages/alembic/command.py", line 124, in upgrade script.run_env() File "/usr/local/lib/python2.7/dist-packages/alembic/script.py", line 193, in run_env util.load_python_file(self.dir, 'env.py') File "/usr/local/lib/python2.7/dist-packages/alembic/util.py", line 177, in load_python_file module = load_module(module_id, path) File "/usr/local/lib/python2.7/dist-packages/alembic/compat.py", line 39, in load_module return imp.load_source(module_id, path, fp) File "/opt/stack/neutron/neutron/db/migration/alembic_migrations/env.py", line 105, in run_migrations_online() File "/opt/stack/neutron/neutron/db/migration/alembic_migrations/env.py", line 89, in run_migrations_online options=build_options()) File "", line 7, in run_migrations File "/usr/local/lib/python2.7/dist-packages/alembic/environment.py", line 652, in run_migrations self.get_context().run_migrations(**kw) File "/usr/local/lib/python2.7/dist-packages/alembic/migration.py", line 224, in run_migrations change(**kw) File "/opt/stack/neutron/neutron/db/migration/alembic_migrations/versions/1fcfc149aca4_agents_unique_by_type_and_host.py", line 50, in upgrade local_cols=['agent_type', 'host'] File "", line 7, in create_unique_constraint File "/usr/local/lib/python2.7/dist-packages/alembic/operations.py", line 539, in create_unique_constraint schema=schema, **kw) File "/usr/local/lib/python2.7/dist-packages/alembic/ddl/impl.py", line 135, in add_constraint self._exec(schema.AddConstraint(const)) File "/usr/local/lib/python2.7/dist-packages/alembic/ddl/impl.py", line 76, in _exec conn.execute(construct, *multiparams, **params) File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 1449, in execute params) File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 1542, in _execute_ddl compiled File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 1698, in _execute_context context) File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 1691, in _execute_context context) File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/default.py", line 331, in do_execute cursor.execute(statement, parameters) File "/usr/lib/python2.7/dist-packages/MySQLdb/cursors.py", line 174, in execute self.errorhandler(self, exc, value) File "/usr/lib/python2.7/dist-packages/MySQLdb/connections.py", line 36, in defaulterrorhandler raise errorclass, errorvalue sqlalchemy.exc.ProgrammingError: (ProgrammingError) (1146, "Table 'plumgrid_neutron.agents' doesn't exist") 'ALTER TABLE agents ADD CONSTRAINT uniq_agents0agent_type0host UNIQUE (agent_type, host)' () ++ failed ++ local r=1 +++ jobs -p ++ kill ++ set +o xtrace ** Affects: neutron Importance: Undecided Assignee: Edgar Magana (emagana) Status: In Progress ** Changed in: neutron Assignee: (unassigned) => Edgar Magana (emagana) ** Changed in: neutron Status: New => In Progress -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1260232 Title: db migration on unique constraint should not be applied to all plugins Status in OpenStack Neutron (virtual network service): In Progress Bug description: There is an assumption that all "neutron plugins" creates the plumgrid_neutron.agents table, which is not the case. I just tested big switch and plumgrid plugins and they are failing: INFO [alembic.migration] Running upgrade havana -> e197124d4b9, add unique constraint to members INFO [alembic.migration] Running upgrade e197124d4b9 -> 1fcfc149aca4, Add a unique constraint on (agent_type, host) columns to prevent a race condition when an agent entry is 'upserted'. Traceback (most recent call l
[Yahoo-eng-team] [Bug 1260233] [NEW] db migration (agents constraint) fails when using ryu plugin
Public bug reported: Ryu plugin does not support agent extension yet. Therefore, 511471cc46b_agent_ext_model_supp.py does not contain ryu plugin, and agents table is not created. However, 1fcfc149aca4_agents_unique_by_type_and_host.py does not consider this case. I think that migration_for_plugins of 1fcfc149aca4 should be the same as 511471cc46b's. 2013-12-12 01:08:44 INFO [alembic.migration] Running upgrade e197124d4b9 -> 1fcfc149aca4, Add a unique constraint on (agent_type, host) columns to prevent a race 2013-12-12 01:08:44 condition when an agent entry is 'upserted'. 2013-12-12 01:08:44 Traceback (most recent call last): 2013-12-12 01:08:44 File "/usr/local/bin/neutron-db-manage", line 10, in 2013-12-12 01:08:44 sys.exit(main()) 2013-12-12 01:08:44 File "/opt/stack/neutron/neutron/db/migration/cli.py", line 143, in main 2013-12-12 01:08:44 CONF.command.func(config, CONF.command.name) 2013-12-12 01:08:44 File "/opt/stack/neutron/neutron/db/migration/cli.py", line 80, in do_upgrade_downgrade 2013-12-12 01:08:44 do_alembic_command(config, cmd, revision, sql=CONF.command.sql) 2013-12-12 01:08:44 File "/opt/stack/neutron/neutron/db/migration/cli.py", line 59, in do_alembic_command 2013-12-12 01:08:44 getattr(alembic_command, cmd)(config, *args, **kwargs) 2013-12-12 01:08:44 File "/usr/local/lib/python2.7/dist-packages/alembic/command.py", line 124, in upgrade 2013-12-12 01:08:44 script.run_env() 2013-12-12 01:08:44 File "/usr/local/lib/python2.7/dist-packages/alembic/script.py", line 193, in run_env 2013-12-12 01:08:44 util.load_python_file(self.dir, 'env.py') 2013-12-12 01:08:44 File "/usr/local/lib/python2.7/dist-packages/alembic/util.py", line 177, in load_python_file 2013-12-12 01:08:44 module = load_module(module_id, path) 2013-12-12 01:08:44 File "/usr/local/lib/python2.7/dist-packages/alembic/compat.py", line 39, in load_module 2013-12-12 01:08:44 return imp.load_source(module_id, path, fp) 2013-12-12 01:08:44 File "/opt/stack/neutron/neutron/db/migration/alembic_migrations/env.py", line 105, in 2013-12-12 01:08:44 run_migrations_online() 2013-12-12 01:08:44 File "/opt/stack/neutron/neutron/db/migration/alembic_migrations/env.py", line 89, in run_migrations_online 2013-12-12 01:08:44 options=build_options()) 2013-12-12 01:08:44 File "", line 7, in run_migrations 2013-12-12 01:08:44 File "/usr/local/lib/python2.7/dist-packages/alembic/environment.py", line 652, in run_migrations 2013-12-12 01:08:45 self.get_context().run_migrations(**kw) 2013-12-12 01:08:45 File "/usr/local/lib/python2.7/dist-packages/alembic/migration.py", line 224, in run_migrations 2013-12-12 01:08:45 change(**kw) 2013-12-12 01:08:45 File "/opt/stack/neutron/neutron/db/migration/alembic_migrations/versions/1fcfc149aca4_agents_unique_by_type_and_host.py", line 50, in upgrade 2013-12-12 01:08:45 local_cols=['agent_type', 'host'] 2013-12-12 01:08:45 File "", line 7, in create_unique_constraint 2013-12-12 01:08:45 File "/usr/local/lib/python2.7/dist-packages/alembic/operations.py", line 539, in create_unique_constraint 2013-12-12 01:08:45 schema=schema, **kw) 2013-12-12 01:08:45 File "/usr/local/lib/python2.7/dist-packages/alembic/ddl/impl.py", line 135, in add_constraint 2013-12-12 01:08:45 self._exec(schema.AddConstraint(const)) 2013-12-12 01:08:45 File "/usr/local/lib/python2.7/dist-packages/alembic/ddl/impl.py", line 76, in _exec 2013-12-12 01:08:45 conn.execute(construct, *multiparams, **params) 2013-12-12 01:08:45 File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 1449, in execute 2013-12-12 01:08:45 params) 2013-12-12 01:08:45 File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 1542, in _execute_ddl 2013-12-12 01:08:45 compiled 2013-12-12 01:08:45 File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 1698, in _execute_context 2013-12-12 01:08:45 context) 2013-12-12 01:08:45 File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 1691, in _execute_context 2013-12-12 01:08:45 context) 2013-12-12 01:08:45 File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/default.py", line 331, in do_execute 2013-12-12 01:08:45 cursor.execute(statement, parameters) 2013-12-12 01:08:45 File "/usr/lib/python2.7/dist-packages/MySQLdb/cursors.py", line 174, in execute 2013-12-12 01:08:45 self.errorhandler(self, exc, value) 2013-12-12 01:08:45 File "/usr/lib/python2.7/dist-packages/MySQLdb/connections.py", line 36, in defaulterrorhandler 2013-12-12 01:08:45 raise errorclass, errorvalue 2013-12-12 01:08:45 sqlalchemy.exc.ProgrammingError: (ProgrammingError) (1146, "Table 'ovs_neutron.agents' doesn't exist") 'ALTER TABLE agents ADD CONSTRAINT uniq_agents0agent_type0host UNIQUE (agent_type, host)' () ** Affects: neutron Importance: Undecided
[Yahoo-eng-team] [Bug 1260262] [NEW] ml2 gre/vxlan type driver can lose segment_id allocation
Public bug reported: With ML2 gre/vxlan type driver and multiple segment is specified, the type driver can leak segment_id ** Affects: neutron Importance: Undecided Assignee: Isaku Yamahata (yamahata) Status: In Progress -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1260262 Title: ml2 gre/vxlan type driver can lose segment_id allocation Status in OpenStack Neutron (virtual network service): In Progress Bug description: With ML2 gre/vxlan type driver and multiple segment is specified, the type driver can leak segment_id To manage notifications about this bug go to: https://bugs.launchpad.net/neutron/+bug/1260262/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1252603] Re: PMTUD needs to be disabled for tunneling to work in many Grizzly environments.
Limiting scope to cloud-archive; raring actually has a sufficiently recent version of openvswitch to not have this issue. ** Also affects: cloud-archive Importance: Undecided Status: New ** Changed in: quantum (Ubuntu) Status: New => Invalid ** Changed in: cloud-archive Status: New => Triaged ** Changed in: cloud-archive Importance: Undecided => Low -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1252603 Title: PMTUD needs to be disabled for tunneling to work in many Grizzly environments. Status in Ubuntu Cloud Archive: Triaged Status in OpenStack Neutron (virtual network service): Invalid Status in “quantum” package in Ubuntu: Invalid Bug description: In Grizzly, the version of OVS is lower than 1.9.0. As a result, tunnel path MTU discovery default value is set to 'enabled'. But internet-wide path MTU discovery rarely works, we need to add a configuration option to disable tunnel path MTU Discovery. Discussion about this issue: Connectivity issue from within the Instances. http://lists.openstack.org/pipermail/openstack/2013-August/000293.html Blog about this issue: Path MTU discovery and GRE. http://techbackground.blogspot.com/2013/06/path-mtu-discovery-and-gre.html Tunnel Path MTU Discovery default value was set to 'disabled' in OVS 1.9.0. Both inheritance of the Don't Fragment bit in IP tunnels (df_inherit) and path MTU discovery are no longer supported in OVS 1.10.0. So Linux distributions that ship with OVS >= 1.9.0 for Havana have no such issue. To manage notifications about this bug go to: https://bugs.launchpad.net/cloud-archive/+bug/1252603/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260249] Re: migration-list: 'unicode' object has no attribute 'iteritems'
** Also affects: nova Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1260249 Title: migration-list: 'unicode' object has no attribute 'iteritems' Status in OpenStack Compute (Nova): New Status in Python client library for Nova: New Bug description: There is an AttributeError when we try to use the command "nova migration-list" Traceback (most recent call last): File "/opt/stack/python-novaclient/novaclient/shell.py", line 721, in main OpenStackComputeShell().main(map(strutils.safe_decode, sys.argv[1:])) File "/opt/stack/python-novaclient/novaclient/shell.py", line 657, in main args.func(self.cs, args) File "/opt/stack/python-novaclient/novaclient/v1_1/contrib/migrations.py", line 71, in do_migration_list args.cell_name)) File "/opt/stack/python-novaclient/novaclient/v1_1/contrib/migrations.py", line 53, in list return self._list("/os-migrations%s" % query_string, "migrations") File "/opt/stack/python-novaclient/novaclient/base.py", line 80, in _list for res in data if res] File "/opt/stack/python-novaclient/novaclient/base.py", line 426, in __init__ self._add_details(info) File "/opt/stack/python-novaclient/novaclient/base.py", line 449, in _add_details for (k, v) in six.iteritems(info): File "/usr/local/lib/python2.7/dist-packages/six.py", line 439, in iteritems return iter(getattr(d, _iteritems)(**kw)) AttributeError: 'unicode' object has no attribute 'iteritems' ERROR: 'unicode' object has no attribute 'iteritems' To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1260249/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260265] [NEW] BaremetalHostManager cannot distinguish baremetal hosts from other hosts
Public bug reported: BaremetalHostManager cloud distingush baremetal hosts by checking "baremetal_driver" exists in capabilities or not. However, now BaremetalHostManager cannot, because capabilities are not reported to scheduler and BaremetalHostManager always receives empty capabilities. As a result, BaremetalHostManager just does the same thing as the original HostManager. ** Affects: nova Importance: Undecided Status: New ** Tags: baremetal -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1260265 Title: BaremetalHostManager cannot distinguish baremetal hosts from other hosts Status in OpenStack Compute (Nova): New Bug description: BaremetalHostManager cloud distingush baremetal hosts by checking "baremetal_driver" exists in capabilities or not. However, now BaremetalHostManager cannot, because capabilities are not reported to scheduler and BaremetalHostManager always receives empty capabilities. As a result, BaremetalHostManager just does the same thing as the original HostManager. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1260265/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260274] [NEW] NoVNC Console not showing in Internet Explorer
Public bug reported: When accessing the NoVNC console through Internet Explorer (tested with IE9 and IE10) the HTML5 Canvas never renders, instead showing 'Canvas not supported'. Environment: - OS: Ubuntu 12.04 LTS - Platform: Openstack Grizzly - Packages: nova-novncproxy 1:2013.1.3-0ubuntu1~cloud0 novnc 2012.2~20120906+dfsg-0ubuntu4~cloud0 python-novnc 2012.2~20120906+dfsg-0ubuntu4~cloud0 According to https://github.com/kanaka/noVNC/wiki/Browser-support NoVNC should work with IE9 and above. ** Affects: nova Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1260274 Title: NoVNC Console not showing in Internet Explorer Status in OpenStack Compute (Nova): New Bug description: When accessing the NoVNC console through Internet Explorer (tested with IE9 and IE10) the HTML5 Canvas never renders, instead showing 'Canvas not supported'. Environment: - OS: Ubuntu 12.04 LTS - Platform: Openstack Grizzly - Packages: nova-novncproxy 1:2013.1.3-0ubuntu1~cloud0 novnc 2012.2~20120906+dfsg-0ubuntu4~cloud0 python-novnc 2012.2~20120906+dfsg-0ubuntu4~cloud0 According to https://github.com/kanaka/noVNC/wiki/Browser-support NoVNC should work with IE9 and above. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1260274/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260281] [NEW] Rendering of dashboard is broken in Internet Explorer
Public bug reported: In Internet Explorer (tested with IE9 and IE10) the rendering of various dashboard components is broken. - Content section is shown below the left hand navigation menu most often, unless you have a super-wide screen - Network Topology network names do not display inside the vertical network bar - The rounded edges do not render - The buttons look funny While I realise that some of these are due to differences in the way that IE renders CSS we do feel that it's important to ensure that using IE for Openstack End-Users and Administrators gives a reasonable experience. ** Affects: horizon Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1260281 Title: Rendering of dashboard is broken in Internet Explorer Status in OpenStack Dashboard (Horizon): New Bug description: In Internet Explorer (tested with IE9 and IE10) the rendering of various dashboard components is broken. - Content section is shown below the left hand navigation menu most often, unless you have a super-wide screen - Network Topology network names do not display inside the vertical network bar - The rounded edges do not render - The buttons look funny While I realise that some of these are due to differences in the way that IE renders CSS we do feel that it's important to ensure that using IE for Openstack End-Users and Administrators gives a reasonable experience. To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1260281/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260249] Re: migration-list: 'unicode' object has no attribute 'iteritems'
this looks like a novaclient bug only ** Changed in: nova Status: New => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1260249 Title: migration-list: 'unicode' object has no attribute 'iteritems' Status in OpenStack Compute (Nova): Invalid Status in Python client library for Nova: New Bug description: There is an AttributeError when we try to use the command "nova migration-list" Traceback (most recent call last): File "/opt/stack/python-novaclient/novaclient/shell.py", line 721, in main OpenStackComputeShell().main(map(strutils.safe_decode, sys.argv[1:])) File "/opt/stack/python-novaclient/novaclient/shell.py", line 657, in main args.func(self.cs, args) File "/opt/stack/python-novaclient/novaclient/v1_1/contrib/migrations.py", line 71, in do_migration_list args.cell_name)) File "/opt/stack/python-novaclient/novaclient/v1_1/contrib/migrations.py", line 53, in list return self._list("/os-migrations%s" % query_string, "migrations") File "/opt/stack/python-novaclient/novaclient/base.py", line 80, in _list for res in data if res] File "/opt/stack/python-novaclient/novaclient/base.py", line 426, in __init__ self._add_details(info) File "/opt/stack/python-novaclient/novaclient/base.py", line 449, in _add_details for (k, v) in six.iteritems(info): File "/usr/local/lib/python2.7/dist-packages/six.py", line 439, in iteritems return iter(getattr(d, _iteritems)(**kw)) AttributeError: 'unicode' object has no attribute 'iteritems' ERROR: 'unicode' object has no attribute 'iteritems' To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1260249/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260301] [NEW] Expired vnc token doesn't fail fast
Public bug reported: If an expired vnc token is used, I'd expect an appropriate HTTP response. However, currently novnc just hangs at "Starting VNC handshake". To reproduce 1. nova get-vnc-console novnc => returns a URI with the token 2. point browser to the returned URI after the expiry 3. page hangs at "Starting VNC handshake" Expected response An appropriate HTTP response ** Affects: nova Importance: Undecided Status: New ** Summary changed: - Expired vnc token doesn't result in an HTTP 401 + Expired vnc token doesn't fail fast ** Description changed: - If an expired vnc token is used, I'd expect an HTTP 401. However, - currently novnc just hangs at "Starting VNC handshake". + If an expired vnc token is used, I'd expect an appropriate HTTP + response. However, currently novnc just hangs at "Starting VNC + handshake". To reproduce 1. nova get-vnc-console novnc => returns a URI with the token 2. point browser to the returned URI after the expiry 3. page hangs at "Starting VNC handshake" Expected response - HTTP 401 + An appropriate HTTP response -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1260301 Title: Expired vnc token doesn't fail fast Status in OpenStack Compute (Nova): New Bug description: If an expired vnc token is used, I'd expect an appropriate HTTP response. However, currently novnc just hangs at "Starting VNC handshake". To reproduce 1. nova get-vnc-console novnc => returns a URI with the token 2. point browser to the returned URI after the expiry 3. page hangs at "Starting VNC handshake" Expected response An appropriate HTTP response To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1260301/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1253896] Re: test_minimum_basic_scenario fails with SSHException: Error reading SSH protocol banner
This bug is not invalid for neutron, it's unfortunate that jog0 said that earlier, as neutron is exposing this more often than nova-network. ** Changed in: neutron Importance: High => Critical ** Changed in: neutron Status: Invalid => Confirmed -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1253896 Title: test_minimum_basic_scenario fails with SSHException: Error reading SSH protocol banner Status in OpenStack Neutron (virtual network service): Confirmed Status in OpenStack Compute (Nova): Triaged Status in Tempest: Confirmed Bug description: An example of this can be found at http://logs.openstack.org/74/57774/2/gate/gate-tempest-devstack-vm- full/e592961/console.html. This test failing seems to cause the tearDownClass failure and the process exit code failure. Judging by the logs below, the VM is coming up, and the test is connecting to the SSH server (dropbear) running in the VM, but the authentication is failing. It appears that authentication is attempted several times before paramiko gives up causing the test to fail. I think this indicates there isn't a network or compute problem, instead is possible the client doesn't have the correct key or the authorized keys aren't configured properly on the server side. But these are just guesses, I haven't been able to get any concrete data that would support these theories. 2013-11-22 05:36:33.980 | 2013-11-22 05:32:17,029 Adding to shared resources of TestMinimumBasicScenario 2013-11-22 05:36:33.980 | 2013-11-22 05:32:34,226 starting thread (client mode): 0x52e7e50L 2013-11-22 05:36:33.980 | 2013-11-22 05:32:34,232 Connected (version 2.0, client dropbear_2012.55) 2013-11-22 05:36:33.981 | 2013-11-22 05:32:34,237 kex algos:['diffie-hellman-group1-sha1', 'diffie-hellman-group14-sha1'] server key:['ssh-rsa', 'ssh-dss'] client encrypt:['aes128-ctr', '3des-ctr', 'aes256-ctr', 'aes128-cbc', '3des-cbc', 'aes256-cbc', 'twofish256-cbc', 'twofish-cbc', 'twofish128-cbc'] server encrypt:['aes128-ctr', '3des-ctr', 'aes256-ctr', 'aes128-cbc', '3des-cbc', 'aes256-cbc', 'twofish256-cbc', 'twofish-cbc', 'twofish128-cbc'] client mac:['hmac-sha1-96', 'hmac-sha1', 'hmac-md5'] server mac:['hmac-sha1-96', 'hmac-sha1', 'hmac-md5'] client compress:['none'] server compress:['none'] client lang:[''] server lang:[''] kex follows?False 2013-11-22 05:36:33.981 | 2013-11-22 05:32:34,238 Ciphers agreed: local=aes128-ctr, remote=aes128-ctr 2013-11-22 05:36:33.981 | 2013-11-22 05:32:34,238 using kex diffie-hellman-group1-sha1; server key type ssh-rsa; cipher: local aes128-ctr, remote aes128-ctr; mac: local hmac-sha1, remote hmac-sha1; compression: local none, remote none 2013-11-22 05:36:33.981 | 2013-11-22 05:32:34,433 Switch to new keys ... 2013-11-22 05:36:33.982 | 2013-11-22 05:32:34,434 Adding ssh-rsa host key for 172.24.4.227: 189c16acb93fe44ae975e1c653f1856c 2013-11-22 05:36:33.982 | 2013-11-22 05:32:34,434 Trying SSH key 9a9afe52a9485c15495a59b94ebca6b6 2013-11-22 05:36:33.982 | 2013-11-22 05:32:34,437 userauth is OK 2013-11-22 05:36:33.982 | 2013-11-22 05:32:35,104 Authentication (publickey) failed. 2013-11-22 05:36:33.982 | 2013-11-22 05:32:36,693 starting thread (client mode): 0x52f9190L 2013-11-22 05:36:33.982 | 2013-11-22 05:32:36,697 Connected (version 2.0, client dropbear_2012.55) 2013-11-22 05:36:33.983 | 2013-11-22 05:32:36,699 kex algos:['diffie-hellman-group1-sha1', 'diffie-hellman-group14-sha1'] server key:['ssh-rsa', 'ssh-dss'] client encrypt:['aes128-ctr', '3des-ctr', 'aes256-ctr', 'aes128-cbc', '3des-cbc', 'aes256-cbc', 'twofish256-cbc', 'twofish-cbc', 'twofish128-cbc'] server encrypt:['aes128-ctr', '3des-ctr', 'aes256-ctr', 'aes128-cbc', '3des-cbc', 'aes256-cbc', 'twofish256-cbc', 'twofish-cbc', 'twofish128-cbc'] client mac:['hmac-sha1-96', 'hmac-sha1', 'hmac-md5'] server mac:['hmac-sha1-96', 'hmac-sha1', 'hmac-md5'] client compress:['none'] server compress:['none'] client lang:[''] server lang:[''] kex follows?False 2013-11-22 05:36:33.983 | 2013-11-22 05:32:36,699 Ciphers agreed: local=aes128-ctr, remote=aes128-ctr 2013-11-22 05:36:33.983 | 2013-11-22 05:32:36,699 using kex diffie-hellman-group1-sha1; server key type ssh-rsa; cipher: local aes128-ctr, remote aes128-ctr; mac: local hmac-sha1, remote hmac-sha1; compression: local none, remote none 2013-11-22 05:36:33.983 | 2013-11-22 05:32:36,903 Switch to new keys ... 2013-11-22 05:36:33.983 | 2013-11-22 05:32:36,904 Trying SSH key 9a9afe52a9485c15495a59b94ebca6b6 2013-11-22 05:36:33.984 | 2013-11-22 05:32:36,906 userauth is OK 2013-11-22 05:36:33.984 | 2013-11-22 05:32:37,438 Authentication (publickey) failed. 2013-11-22 05:36:33.984 | 2013-11-22 05:32:39,035 starting thread (client mode): 0x24c62d0L 2013-11-22 05:36:33.984 | 2013-11-22 05:32:39,043 Connected (version 2.0,
[Yahoo-eng-team] [Bug 1260310] [NEW] gate-tempest-dsvm-full failure with "An error occurred while enabling hairpin mode on domain with xml"
Public bug reported: Kibana search http://logstash.openstack.org/#eyJzZWFyY2giOiJtZXNzYWdlOiBcIkFuIGVycm9yIG9jY3VycmVkIHdoaWxlIGVuYWJsaW5nIGhhaXJwaW4gbW9kZSBvbiBkb21haW4gd2l0aCB4bWxcIiIsImZpZWxkcyI6W10sIm9mZnNldCI6MCwidGltZWZyYW1lIjoiNjA0ODAwIiwiZ3JhcGhtb2RlIjoiY291bnQiLCJ0aW1lIjp7InVzZXJfaW50ZXJ2YWwiOjB9LCJzdGFtcCI6MTM4Njg1MjgzMTk3M30= This is pretty infrequent ** Affects: nova Importance: High Status: Triaged -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1260310 Title: gate-tempest-dsvm-full failure with "An error occurred while enabling hairpin mode on domain with xml" Status in OpenStack Compute (Nova): Triaged Bug description: Kibana search http://logstash.openstack.org/#eyJzZWFyY2giOiJtZXNzYWdlOiBcIkFuIGVycm9yIG9jY3VycmVkIHdoaWxlIGVuYWJsaW5nIGhhaXJwaW4gbW9kZSBvbiBkb21haW4gd2l0aCB4bWxcIiIsImZpZWxkcyI6W10sIm9mZnNldCI6MCwidGltZWZyYW1lIjoiNjA0ODAwIiwiZ3JhcGhtb2RlIjoiY291bnQiLCJ0aW1lIjp7InVzZXJfaW50ZXJ2YWwiOjB9LCJzdGFtcCI6MTM4Njg1MjgzMTk3M30= This is pretty infrequent To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1260310/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260314] [NEW] glance image-create with invalid store fails but still creates image
Public bug reported: glance checks whether or not a specified store is valid, but if it is invalid the image has already been created. I pulled the latest devstack code and then ran these commands after sourcing openrc: ubuntu@devstack-glance:/mnt/devstack$ glance index ID Name Disk Format Container Format Size -- -- 6792e9a7-f4f8-48cb-b407-80e360b8a773 cirros-0.3.1-x86_64-uecami ami25165824 7808c034-3fdd-4975-af26-e7d5a15d2113 cirros-0.3.1-x86_64-uec-ramdis ari ari 3714968 4efcddb2-9f20-413f-86a3-3bf69455e09b cirros-0.3.1-x86_64-uec-kernel aki aki 4955792 ubuntu@devstack-glance:/mnt/devstack$ ubuntu@devstack-glance:/mnt/devstack$ glance -d image-create --store s3e --disk-format raw --container-format bare --name complete_gibberish ' -H 'Content-Type: application/octet-stream' -H 'x-image-meta-disk_format: raw' -H 'x-image-meta-name: complete_gibberish' -d '', mode 'r' at 0x7f16181b6150>' http://10.4.36.1:9292/v1/images HTTP/1.1 400 Bad Request date: Thu, 12 Dec 2013 12:47:37 GMT content-length: 52 content-type: text/plain; charset=UTF-8 x-openstack-request-id: req-c9bad6ee-d79c-41f3-bd96-d3929afd742c 400 Bad Request Store for scheme s3e not found Request returned failure status. 400 Bad Request Store for scheme s3e not found (HTTP 400) ubuntu@devstack-glance:/mnt/devstack$ glance index ID Name Disk Format Container Format Size -- -- b26c03e4-7cdf-44fe-9187-7de315c9b38b complete_gibberish raw bare221 6792e9a7-f4f8-48cb-b407-80e360b8a773 cirros-0.3.1-x86_64-uecami ami25165824 7808c034-3fdd-4975-af26-e7d5a15d2113 cirros-0.3.1-x86_64-uec-ramdis ari ari 3714968 4efcddb2-9f20-413f-86a3-3bf69455e09b cirros-0.3.1-x86_64-uec-kernel aki aki 4955792 This problem occurs using the v1 API. If using the V2 API the '--store' option does not seem to be present. ** Affects: glance Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to Glance. https://bugs.launchpad.net/bugs/1260314 Title: glance image-create with invalid store fails but still creates image Status in OpenStack Image Registry and Delivery Service (Glance): New Bug description: glance checks whether or not a specified store is valid, but if it is invalid the image has already been created. I pulled the latest devstack code and then ran these commands after sourcing openrc: ubuntu@devstack-glance:/mnt/devstack$ glance index ID Name Disk Format Container Format Size -- -- 6792e9a7-f4f8-48cb-b407-80e360b8a773 cirros-0.3.1-x86_64-uecami ami25165824 7808c034-3fdd-4975-af26-e7d5a15d2113 cirros-0.3.1-x86_64-uec-ramdis ari ari 3714968 4efcddb2-9f20-413f-86a3-3bf69455e09b cirros-0.3.1-x86_64-uec-kernel aki aki 4955792 ubuntu@devstack-glance:/mnt/devstack$ ubuntu@devstack-glance:/mnt/devstack$ glance -d image-create --store s3e --disk-format raw --container-format bare --name complete_gibberish ' -H 'Content-Type: application/octet-stream' -H 'x-image-meta-disk_format: raw' -H 'x-image-meta-name: complete_gibberish' -d '', mode 'r' at 0x7f16181b6150>' http://10.4.36.1:9292/v1/images HTTP/1.1 400 Bad Request date: Thu, 12 Dec 2013 12:47:37 GMT content-length: 52 content-type: text/plain; charset=UTF-8 x-openstack-request-id: req-c9bad6ee-d79c-41f3-bd96-d3929afd742c 400 Bad Request Store for scheme s3e not found Request returned failure status. 400 Bad Request Store for scheme s3e not found (HTTP 400) ubuntu@devstack-glance:/mnt/devstack$ glance index ID Name Disk Format Container Format Size -- -- b26c03e4-7cdf-44fe-9187-7de315c9b38b complete_gibberish raw bare
[Yahoo-eng-team] [Bug 1259894] Re: tempest.api.compute.v3.servers.test_server_rescue.ServerRescueV3TestJSON.test_rescue_unrescue_instance FAILED
It looks like a nova issue, adding nova to the afflicted projects. ** Also affects: nova Importance: Undecided Status: New ** Changed in: tempest Status: New => Incomplete -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1259894 Title: tempest.api.compute.v3.servers.test_server_rescue.ServerRescueV3TestJSON.test_rescue_unrescue_instance FAILED Status in OpenStack Compute (Nova): New Status in Tempest: Incomplete Bug description: Traceback (most recent call last): File "tempest/api/compute/v3/servers/test_server_rescue.py", line 98, in test_rescue_unrescue_instance self.servers_client.wait_for_server_status(self.server_id, 'RESCUE') File "tempest/services/compute/v3/json/servers_client.py", line 167, in wait_for_server_status extra_timeout=extra_timeout) File "tempest/common/waiters.py", line 82, in wait_for_server_status raise exceptions.TimeoutException(message) TimeoutException: Request timed out Details: Server 9c38fcf3-50ab-4a19-86e3-5c0fab501fca failed to reach RESCUE status within the required time (196 s). Current status: ACTIVE. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1259894/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1257460] Re: test_rescued_vm_detach_volume Volume test_detach failed to reach in-use status within the required time (196 s).
Is tempest did anything incorrectly ? ** Also affects: nova Importance: Undecided Status: New ** Also affects: cinder Importance: Undecided Status: New ** Changed in: tempest Status: New => Incomplete -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1257460 Title: test_rescued_vm_detach_volume Volume test_detach failed to reach in- use status within the required time (196 s). Status in Cinder: New Status in OpenStack Compute (Nova): New Status in Tempest: Incomplete Bug description: http://logs.openstack.org/76/59576/3/check/check-tempest-dsvm- postgres-full/af0e4e2/console.html 2013-12-03 01:57:24.699 | == 2013-12-03 01:57:24.699 | FAIL: tempest.api.compute.servers.test_server_rescue.ServerRescueTestJSON.test_rescued_vm_detach_volume[gate,negative] 2013-12-03 01:57:24.699 | tempest.api.compute.servers.test_server_rescue.ServerRescueTestJSON.test_rescued_vm_detach_volume[gate,negative] 2013-12-03 01:57:24.699 | -- 2013-12-03 01:57:24.700 | _StringException: Empty attachments: 2013-12-03 01:57:24.700 | stderr 2013-12-03 01:57:24.700 | stdout 2013-12-03 01:57:24.701 | 2013-12-03 01:57:24.701 | pythonlogging:'': {{{ 2013-12-03 01:57:24.701 | 2013-12-03 01:34:31,546 Request: POST http://127.0.0.1:8774/v2/742879cfcd384dffb721c692c81376be/servers/ab993087-a194-4798-ac63-3b5495690fb6/os-volume_attachments 2013-12-03 01:57:24.701 | 2013-12-03 01:34:31,546 Request Headers: {'Content-Type': 'application/json', 'Accept': 'application/json', 'X-Auth-Token': ''} 2013-12-03 01:57:24.701 | 2013-12-03 01:34:31,546 Request Body: {"volumeAttachment": {"device": "/dev/vdf", "volumeId": "67bc3c74-1d66-4107-9d6d-f66b6a678e52"}} 2013-12-03 01:57:24.701 | 2013-12-03 01:34:33,730 Response Status: 200 2013-12-03 01:57:24.702 | 2013-12-03 01:34:33,730 Nova request id: req-4314d02e-0522-4410-9e6c-909528fb2314 2013-12-03 01:57:24.702 | 2013-12-03 01:34:33,731 Response Headers: {'content-length': '194', 'date': 'Tue, 03 Dec 2013 01:34:33 GMT', 'content-type': 'application/json', 'connection': 'close'} 2013-12-03 01:57:24.702 | 2013-12-03 01:34:33,731 Response Body: {"volumeAttachment": {"device": "/dev/vdf", "serverId": "ab993087-a194-4798-ac63-3b5495690fb6", "id": "67bc3c74-1d66-4107-9d6d-f66b6a678e52", "volumeId": "67bc3c74-1d66-4107-9d6d-f66b6a678e52"}} 2013-12-03 01:57:24.702 | 2013-12-03 01:34:33,731 Request: GET http://127.0.0.1:8774/v2/742879cfcd384dffb721c692c81376be/os-volumes/67bc3c74-1d66-4107-9d6d-f66b6a678e52 2013-12-03 01:57:24.702 | 2013-12-03 01:34:33,732 Request Headers: {'X-Auth-Token': ''} 2013-12-03 01:57:24.702 | 2013-12-03 01:34:34,025 Response Status: 200 2013-12-03 01:57:24.702 | 2013-12-03 01:34:34,026 Nova request id: req-4a9aac2a-55cb-4287-8f4c-392d3498a304 2013-12-03 01:57:24.871 | Traceback (most recent call last): 2013-12-03 01:57:24.871 | File "tempest/api/compute/servers/test_server_rescue.py", line 173, in test_rescued_vm_detach_volume 2013-12-03 01:57:24.871 | self.volume_to_detach['id'], 'in-use') 2013-12-03 01:57:24.872 | File "tempest/services/compute/json/volumes_extensions_client.py", line 104, in wait_for_volume_status 2013-12-03 01:57:24.872 | raise exceptions.TimeoutException(message) 2013-12-03 01:57:24.872 | TimeoutException: Request timed out 2013-12-03 01:57:24.872 | Details: Volume test_detach failed to reach in-use status within the required time (196 s). To manage notifications about this bug go to: https://bugs.launchpad.net/cinder/+bug/1257460/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260138] Re: VMWARE: Unable to spawn instances from sparse/ide images
Duplicate of #1260139 ** Changed in: nova Status: New => Won't Fix -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1260138 Title: VMWARE: Unable to spawn instances from sparse/ide images Status in OpenStack Compute (Nova): Won't Fix Bug description: Branch: stable/havana Traceback: http://paste.openstack.org/show/54855/ Steps to reprodude: Upload a ide/sparse type image to glance. Spawn an instance from that image Actual Result: Failed to spawn an image To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1260138/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1182883] Re: List servers matching a regex fails with Quantum
** Changed in: tempest Status: Confirmed => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1182883 Title: List servers matching a regex fails with Quantum Status in OpenStack Neutron (virtual network service): Invalid Status in OpenStack Compute (Nova): Invalid Status in Tempest: Invalid Bug description: The test tempest.api.compute.servers.test_list_server_filters:ListServerFiltersTestXML.test_list_servers_filtered_by_ip_regex tries to search a server with only a fragment of its IP (GET http://XX/v2/$Tenant/servers?ip=10.0.) which calls the following Quantum request : http://XX/v2.0/ports.json?fixed_ips=ip_address%3D10.0. But it seems this regex search is not supporter by Quantum. Thus the tempest test fauls. To manage notifications about this bug go to: https://bugs.launchpad.net/neutron/+bug/1182883/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1195473] Re: Need some docs about Glance notifications with examples
** No longer affects: glance -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to Glance. https://bugs.launchpad.net/bugs/1195473 Title: Need some docs about Glance notifications with examples Status in OpenStack Manuals: Confirmed Bug description: The notifications emitted by Glance were enhanced for Grizzly. The developer docs have been updated, but we could probably use a listing of what notifications are emitted and what they look like for the operator manual. To manage notifications about this bug go to: https://bugs.launchpad.net/openstack-manuals/+bug/1195473/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1244055] Re: six has no attribute 'add_metaclass'
** Changed in: tempest Status: Confirmed => Fix Released -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1244055 Title: six has no attribute 'add_metaclass' Status in OpenStack Compute (Nova): Invalid Status in Tempest: Fix Released Bug description: I have a patch failing in gate with traces containing the following: 2013-10-23 22:27:54.336 | File "/opt/stack/new/python-novaclient/novaclient/base.py", line 166, in 2013-10-23 22:27:54.336 | @six.add_metaclass(abc.ABCMeta) 2013-10-23 22:27:54.337 | AttributeError: 'module' object has no attribute 'add_metaclass' For full logs, see the failing patch: https://review.openstack.org/#/c/52876/ It looks like this was caused by this recent commit: https://review.openstack.org/#/c/52255/ To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1244055/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1124674] Re: add-fixed-ip causes traceback
** Changed in: tempest Status: Confirmed => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1124674 Title: add-fixed-ip causes traceback Status in OpenStack Compute (Nova): Fix Released Status in Tempest: Invalid Bug description: nova add-fixed-ip leads to a traceback in nova-network 2013-02-13 16:10:57.711 ERROR nova.openstack.common.rpc.amqp [req-434b8722-566c-4345-933a-5646277cd6ef demo demo] Exception during message handling2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp Traceback (most recent call last):2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/nova/nova/openstack/common/rpc/amqp.py", line 276, in _process_data2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp rval = self.proxy.dispatch(ctxt, version, method, **args)2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/nova/nova/openstack/common/rpc/dispatcher.py", line 133, in dispatch2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp return getattr(proxyobj, method)(ctxt, **kwargs)2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/nova/nova/network/manager.py", line 756, in add_fixed_ip_to_instance2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp self._alloc ate_fixed_ips(context, instance_id, host, [network])2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/nova/nova/network/manager.py", line 212, in _allocate_fixed_ips 2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp vpn=vpn, address=address) 2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/nova/nova/network/manager.py", line 801, in allocate_fixed_ip 2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp instance_ref = self.db.instance_get(context, instance_id) 2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/nova/nova/db/api.py", line 593, in instance_get 2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp return IMPL.instance_get(context, instance_id) 2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/nova/nova/db/sqlalchemy/api.py", line 137, in wrapper 2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp return f(*args, **kwargs) 2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/nova/nova/db/sqlalchemy/api.py", line 1520, in instance_get 2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp raise exception.InstanceNotFound(instance_id=instance_id) 2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp InstanceNotFound: Instance f96d5c44-6c17-46cb-8b5c-057717ce076d could not be found. there seems to be quite a bit of mixing of instance['id'] with instance['uuid'] To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1124674/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1213212] Re: test_resize_server_confirm server failed to build
removing this as a tempest issue, as I don't think it actually is a bug in tempest, it's a nova state bug ** Changed in: nova Status: New => Confirmed ** Changed in: nova Importance: Undecided => Medium ** No longer affects: tempest -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1213212 Title: test_resize_server_confirm server failed to build Status in OpenStack Compute (Nova): Confirmed Bug description: When running tempest in parallel occasionally test_resize_server_confirm fails to build the server and goes into an error state see: 2013-08-16 14:08:33.607 | == 2013-08-16 14:08:33.607 | FAIL: tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_resize_server_confirm[gate,smoke] 2013-08-16 14:08:33.607 | tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_resize_server_confirm[gate,smoke] 2013-08-16 14:08:33.608 | -- 2013-08-16 14:08:33.608 | _StringException: Empty attachments: 2013-08-16 14:08:33.608 | stderr 2013-08-16 14:08:33.609 | stdout 2013-08-16 14:08:33.609 | 2013-08-16 14:08:33.609 | Traceback (most recent call last): 2013-08-16 14:08:33.609 | File "tempest/api/compute/servers/test_server_actions.py", line 161, in test_resize_server_confirm 2013-08-16 14:08:33.609 | self.client.wait_for_server_status(self.server_id, 'VERIFY_RESIZE') 2013-08-16 14:08:33.609 | File "tempest/services/compute/json/servers_client.py", line 165, in wait_for_server_status 2013-08-16 14:08:33.609 | raise exceptions.BuildErrorException(server_id=server_id) 2013-08-16 14:08:33.610 | BuildErrorException: Server ed3c7212-f4b6-4365-91b8-bc9e1a60 failed to build and is in ERROR status 2013-08-16 14:08:33.610 | 2013-08-16 14:08:33.610 | 2013-08-16 14:08:33.611 | == A set of logs for this failure can be found here: http://logs.openstack.org/63/42063/1/gate/gate-tempest-devstack-vm-testr-full/fa32f42/ To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1213212/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1213209] Re: test_list_image_filters.py setUpClass created image never becomes active
*** This bug is a duplicate of bug 1258635 *** https://bugs.launchpad.net/bugs/1258635 ** This bug has been marked a duplicate of bug 1258635 Race with changing image status when snapshotting -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1213209 Title: test_list_image_filters.py setUpClass created image never becomes active Status in OpenStack Compute (Nova): New Status in Tempest: Confirmed Bug description: When running in parallel occassionally the tests in test_list_image_filters fail while waiting for one of the created images to become active. See the logs here: http://logs.openstack.org/42/40342/2/gate/gate-tempest-devstack-vm- testr-full/dad876b/ From the tempest log: 2013-08-16 05:51:36.930 368 ERROR tempest.api.compute.images.test_list_image_filters [-] Request timed out 2013-08-16 05:51:36.930 368 TRACE tempest.api.compute.images.test_list_image_filters Traceback (most recent call last): 2013-08-16 05:51:36.930 368 TRACE tempest.api.compute.images.test_list_image_filters File "tempest/api/compute/images/test_list_image_filters.py", line 67, in setUpClass 2013-08-16 05:51:36.930 368 TRACE tempest.api.compute.images.test_list_image_filters cls.client.wait_for_image_status(cls.image2_id, 'ACTIVE') 2013-08-16 05:51:36.930 368 TRACE tempest.api.compute.images.test_list_image_filters File "tempest/services/compute/json/images_client.py", line 110, in wait_for_image_status 2013-08-16 05:51:36.930 368 TRACE tempest.api.compute.images.test_list_image_filters raise exceptions.TimeoutException 2013-08-16 05:51:36.930 368 TRACE tempest.api.compute.images.test_list_image_filters TimeoutException: Request timed out 2013-08-16 05:51:36.930 368 TRACE tempest.api.compute.images.test_list_image_filters To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1213209/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1132879] Re: server reboot hard and rebuild are flaky in tempest when ssh is enabled
** Changed in: tempest Status: Confirmed => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1132879 Title: server reboot hard and rebuild are flaky in tempest when ssh is enabled Status in OpenStack Compute (Nova): Confirmed Status in Tempest: Invalid Bug description: Working on enabling back ssh access to VMs in tempest tests: https://review.openstack.org/#/c/22415/ https://blueprints.launchpad.net/tempest/+spec/ssh-auth-strategy On the gate devstack with nova networking the hard reboot and rebuild test are sometimes passing and sometimes not. On the gate devstack with quantum networking the hard reboot and rebuild tests are systematically not passing, and blocking the overall blueprint implementation. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1132879/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260333] [NEW] Malformed property protection rules return error to end user
Public bug reported: Using a property protections file such as: [.*] create = @,! read = @ update = @ delete = @ The create operation has an invalid rule, duplicate values are not allowed. This should probably result in the service refusing to start, however currently the service will start and operations touching this value will return: 500 Internal Server Error Malformed property protection rule 'some_property': '@' and '!' are mutually exclusive (HTTP 500) to the end user. My feeling is that the end user should not receive any information about the cause of the error, just the 500 status. ** Affects: glance Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to Glance. https://bugs.launchpad.net/bugs/1260333 Title: Malformed property protection rules return error to end user Status in OpenStack Image Registry and Delivery Service (Glance): New Bug description: Using a property protections file such as: [.*] create = @,! read = @ update = @ delete = @ The create operation has an invalid rule, duplicate values are not allowed. This should probably result in the service refusing to start, however currently the service will start and operations touching this value will return: 500 Internal Server Error Malformed property protection rule 'some_property': '@' and '!' are mutually exclusive (HTTP 500) to the end user. My feeling is that the end user should not receive any information about the cause of the error, just the 500 status. To manage notifications about this bug go to: https://bugs.launchpad.net/glance/+bug/1260333/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1254752] Re: test_volume_boot_pattern: SSH timed out
Is there any sign of tempest did something incorrectly ? ** Also affects: neutron Importance: Undecided Status: New ** Changed in: tempest Status: New => Incomplete -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1254752 Title: test_volume_boot_pattern: SSH timed out Status in OpenStack Neutron (virtual network service): New Status in Tempest: Incomplete Bug description: The test_volume_boot_pattern tests fails sporadically: http://logs.openstack.org/97/57797/2/check/check-tempest-devstack-vm- neutron/38fdc5a/console.html.gz 2013-11-22 21:53:07.670 | Traceback (most recent call last): 2013-11-22 21:53:07.670 | File "tempest/scenario/test_volume_boot_pattern.py", line 156, in test_volume_boot_pattern 2013-11-22 21:53:07.670 | ssh_client = self._ssh_to_server(instance_from_snapshot, keypair) 2013-11-22 21:53:07.670 | File "tempest/scenario/test_volume_boot_pattern.py", line 100, in _ssh_to_server 2013-11-22 21:53:07.670 | private_key=keypair.private_key) 2013-11-22 21:53:07.670 | File "tempest/scenario/manager.py", line 475, in get_remote_client 2013-11-22 21:53:07.671 | return RemoteClient(ip, username, pkey=private_key) 2013-11-22 21:53:07.671 | File "tempest/common/utils/linux/remote_client.py", line 47, in __init__ 2013-11-22 21:53:07.671 | if not self.ssh_client.test_connection_auth(): 2013-11-22 21:53:07.671 | File "tempest/common/ssh.py", line 148, in test_connection_auth 2013-11-22 21:53:07.671 | connection = self._get_ssh_connection() 2013-11-22 21:53:07.672 | File "tempest/common/ssh.py", line 76, in _get_ssh_connection 2013-11-22 21:53:07.672 | password=self.password) 2013-11-22 21:53:07.672 | SSHTimeout: Connection to the 172.24.4.230 via SSH timed out. 2013-11-22 21:53:07.672 | User: cirros, Password: None To manage notifications about this bug go to: https://bugs.launchpad.net/neutron/+bug/1254752/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1251920] Re: Tempest failures due to failure to return console logs from an instance
Since this is not hitting on tempest anymore, moving to Fix Released ** Changed in: tempest Status: Fix Committed => Fix Released -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1251920 Title: Tempest failures due to failure to return console logs from an instance Status in OpenStack Compute (Nova): Invalid Status in OpenStack Compute (nova) havana series: Fix Committed Status in Tempest: Fix Released Bug description: Logstash search: http://logstash.openstack.org/#eyJzZWFyY2giOiJmaWxlbmFtZTpjb25zb2xlLmh0bWwgQU5EIG1lc3NhZ2U6XCJhc3NlcnRpb25lcnJvcjogY29uc29sZSBvdXRwdXQgd2FzIGVtcHR5XCIiLCJmaWVsZHMiOltdLCJvZmZzZXQiOjAsInRpbWVmcmFtZSI6IjYwNDgwMCIsImdyYXBobW9kZSI6ImNvdW50IiwidGltZSI6eyJ1c2VyX2ludGVydmFsIjowfSwic3RhbXAiOjEzODQ2NDEwNzIxODl9 An example failure is http://logs.openstack.org/92/55492/8/check /check-tempest-devstack-vm-full/ef3a4a4/console.html console.html === 2013-11-16 21:54:27.998 | 2013-11-16 21:41:20,775 Request: POST http://127.0.0.1:8774/v2/3f6934d9aabf467aa8bc51397ccfa782/servers/10aace14-23c1-4cec-9bfd-2c873df1fbee/action 2013-11-16 21:54:27.998 | 2013-11-16 21:41:20,776 Request Headers: {'Content-Type': 'application/json', 'Accept': 'application/json', 'X-Auth-Token': ''} 2013-11-16 21:54:27.998 | 2013-11-16 21:41:20,776 Request Body: {"os-getConsoleOutput": {"length": 10}} 2013-11-16 21:54:27.998 | 2013-11-16 21:41:21,000 Response Status: 200 2013-11-16 21:54:27.999 | 2013-11-16 21:41:21,001 Nova request id: req-7a2ee0ab-c977-4957-abb5-1d84191bf30c 2013-11-16 21:54:27.999 | 2013-11-16 21:41:21,001 Response Headers: {'content-length': '14', 'date': 'Sat, 16 Nov 2013 21:41:20 GMT', 'content-type': 'application/json', 'connection': 'close'} 2013-11-16 21:54:27.999 | 2013-11-16 21:41:21,001 Response Body: {"output": ""} 2013-11-16 21:54:27.999 | }}} 2013-11-16 21:54:27.999 | 2013-11-16 21:54:27.999 | Traceback (most recent call last): 2013-11-16 21:54:27.999 | File "tempest/api/compute/servers/test_server_actions.py", line 281, in test_get_console_output 2013-11-16 21:54:28.000 | self.wait_for(get_output) 2013-11-16 21:54:28.000 | File "tempest/api/compute/base.py", line 133, in wait_for 2013-11-16 21:54:28.000 | condition() 2013-11-16 21:54:28.000 | File "tempest/api/compute/servers/test_server_actions.py", line 278, in get_output 2013-11-16 21:54:28.000 | self.assertTrue(output, "Console output was empty.") 2013-11-16 21:54:28.000 | File "/usr/lib/python2.7/unittest/case.py", line 420, in assertTrue 2013-11-16 21:54:28.000 | raise self.failureException(msg) 2013-11-16 21:54:28.001 | AssertionError: Console output was empty. n-api 2013-11-16 21:41:20.782 DEBUG nova.api.openstack.wsgi [req-7a2ee0ab-c977-4957-abb5-1d84191bf30c ServerActionsTestJSON-tempest-2102529866-user ServerActionsTestJSON-tempest-2102529866-tenant] Action: 'action', body: {"os-getConsoleOutput": {"length": 10}} _process_stack /opt/stack/new/nova/nova/api/openstack/wsgi.py:963 2013-11-16 21:41:20.782 DEBUG nova.api.openstack.wsgi [req-7a2ee0ab-c977-4957-abb5-1d84191bf30c ServerActionsTestJSON-tempest-2102529866-user ServerActionsTestJSON-tempest-2102529866-tenant] Calling method > _process_stack /opt/stack/new/nova/nova/api/openstack/wsgi.py:964 2013-11-16 21:41:20.865 DEBUG nova.openstack.common.rpc.amqp [req-7a2ee0ab-c977-4957-abb5-1d84191bf30c ServerActionsTestJSON-tempest-2102529866-user ServerActionsTestJSON-tempest-2102529866-tenant] Making synchronous call on compute.devstack-precise-hpcloud-az2-663635 ... multicall /opt/stack/new/nova/nova/openstack/common/rpc/amqp.py:553 2013-11-16 21:41:20.866 DEBUG nova.openstack.common.rpc.amqp [req-7a2ee0ab-c977-4957-abb5-1d84191bf30c ServerActionsTestJSON-tempest-2102529866-user ServerActionsTestJSON-tempest-2102529866-tenant] MSG_ID is a93dceabf6a441eb850b5fbb012d661f multicall /opt/stack/new/nova/nova/openstack/common/rpc/amqp.py:556 2013-11-16 21:41:20.866 DEBUG nova.openstack.common.rpc.amqp [req-7a2ee0ab-c977-4957-abb5-1d84191bf30c ServerActionsTestJSON-tempest-2102529866-user ServerActionsTestJSON-tempest-2102529866-tenant] UNIQUE_ID is 706ab69dc066440fbe1bd7766b73d953. _add_unique_id /opt/stack/new/nova/nova/openstack/common/rpc/amqp.py:341 2013-11-16 21:41:20.869 22679 DEBUG amqp [-] Closed channel #1 _do_close /usr/local/lib/python2.7/dist-packages/amqp/channel.py:95 2013-11-16 21:41:20.869 22679 DEBUG amqp [-] using channel_id: 1 __init__ /usr/local/lib/python2.7/dist-packages/amqp/channel.py:71 2013-11-16 21:41:20.870 22679 DEBUG amqp [-] Channel open _open_ok /usr/local/lib/python2.7/dist-packages/amqp/channel.py:429 2013-11-16 21:41:20.999 INFO nova.osapi_compute.wsgi.server [req-7a2ee0ab-c977-4957-abb5-1d84191bf30c ServerActionsTestJSON-tempest-2102529866-user Se
[Yahoo-eng-team] [Bug 1213215] Re: ServerRescueTest tearDownClass fails with volume status being in-use
** Also affects: cinder Importance: Undecided Status: New ** Changed in: cinder Importance: Undecided => High ** Changed in: nova Importance: Undecided => High ** Changed in: cinder Status: New => Confirmed ** Changed in: nova Status: New => Confirmed -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1213215 Title: ServerRescueTest tearDownClass fails with volume status being in-use Status in Cinder: Confirmed Status in OpenStack Compute (Nova): Confirmed Status in Tempest: Confirmed Bug description: Occasionally running tempest in parallel will fail several tests with timeout errors. The only nontimeout failure message is that the ServerRescueTest failed to delete a volume because it was still marked as in use. My guess is that the leftover volume is somehow interfering with the other tests causing them to timeout. But, I haven't looked at the logs in detail so it's just a wild guess. 2013-08-16 14:11:42.074 | == 2013-08-16 14:11:42.075 | FAIL: tempest.api.compute.servers.test_disk_config.ServerDiskConfigTestJSON.test_rebuild_server_with_auto_disk_config[gate] 2013-08-16 14:11:42.075 | tempest.api.compute.servers.test_disk_config.ServerDiskConfigTestJSON.test_rebuild_server_with_auto_disk_config[gate] 2013-08-16 14:11:42.075 | -- 2013-08-16 14:11:42.075 | _StringException: Empty attachments: 2013-08-16 14:11:42.075 | stderr 2013-08-16 14:11:42.076 | stdout 2013-08-16 14:11:42.076 | 2013-08-16 14:11:42.076 | Traceback (most recent call last): 2013-08-16 14:11:42.076 | File "tempest/api/compute/servers/test_disk_config.py", line 64, in test_rebuild_server_with_auto_disk_config 2013-08-16 14:11:42.076 | wait_until='ACTIVE') 2013-08-16 14:11:42.076 | File "tempest/api/compute/base.py", line 140, in create_server 2013-08-16 14:11:42.076 | server['id'], kwargs['wait_until']) 2013-08-16 14:11:42.077 | File "tempest/services/compute/json/servers_client.py", line 160, in wait_for_server_status 2013-08-16 14:11:42.077 | time.sleep(self.build_interval) 2013-08-16 14:11:42.077 | File "/usr/local/lib/python2.7/dist-packages/fixtures/_fixtures/timeout.py", line 52, in signal_handler 2013-08-16 14:11:42.077 | raise TimeoutException() 2013-08-16 14:11:42.077 | TimeoutException 2013-08-16 14:11:42.077 | 2013-08-16 14:11:42.077 | 2013-08-16 14:11:42.078 | == 2013-08-16 14:11:42.078 | FAIL: setUpClass (tempest.api.compute.images.test_image_metadata.ImagesMetadataTestXML) 2013-08-16 14:11:42.078 | setUpClass (tempest.api.compute.images.test_image_metadata.ImagesMetadataTestXML) 2013-08-16 14:11:42.078 | -- 2013-08-16 14:11:42.078 | _StringException: Traceback (most recent call last): 2013-08-16 14:11:42.078 | File "tempest/api/compute/images/test_image_metadata.py", line 46, in setUpClass 2013-08-16 14:11:42.078 | cls.client.wait_for_image_status(cls.image_id, 'ACTIVE') 2013-08-16 14:11:42.079 | File "tempest/services/compute/xml/images_client.py", line 167, in wait_for_image_status 2013-08-16 14:11:42.079 | raise exceptions.TimeoutException 2013-08-16 14:11:42.079 | TimeoutException: Request timed out 2013-08-16 14:11:42.079 | 2013-08-16 14:11:42.079 | 2013-08-16 14:11:42.079 | == 2013-08-16 14:11:42.079 | FAIL: tempest.api.compute.servers.test_server_rescue.ServerRescueTestJSON.test_rescued_vm_detach_volume[gate,negative] 2013-08-16 14:11:42.080 | tempest.api.compute.servers.test_server_rescue.ServerRescueTestJSON.test_rescued_vm_detach_volume[gate,negative] 2013-08-16 14:11:42.080 | -- 2013-08-16 14:11:42.080 | _StringException: Empty attachments: 2013-08-16 14:11:42.080 | stderr 2013-08-16 14:11:42.080 | stdout 2013-08-16 14:11:42.080 | 2013-08-16 14:11:42.081 | Traceback (most recent call last): 2013-08-16 14:11:42.081 | File "tempest/api/compute/servers/test_server_rescue.py", line 184, in test_rescued_vm_detach_volume 2013-08-16 14:11:42.081 | self.servers_client.wait_for_server_status(self.server_id, 'RESCUE') 2013-08-16 14:11:42.081 | File "tempest/services/compute/json/servers_client.py", line 160, in wait_for_server_status 2013-08-16 14:11:42.081 | time.sleep(self.build_interval) 2013-08-16 14:11:42.081 | File "/usr/local/lib/python2.7/dist-packages/fixtures/_fixtures/timeout.py", line 52, in signal_handler 2013-08-16 14:11:42.081 | raise Timeo
[Yahoo-eng-team] [Bug 1257070] Re: test_glance_timeout flakey fail
Glance configured with one worker, is it possible it is too busy because of another orations? Do we need to increase the timeout and/or the workers ? ** Changed in: tempest Status: New => Incomplete ** Also affects: glance Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to Glance. https://bugs.launchpad.net/bugs/1257070 Title: test_glance_timeout flakey fail Status in OpenStack Image Registry and Delivery Service (Glance): New Status in Tempest: Incomplete Bug description: Transient fail for tempest.cli.simple_read_only.test_glance.SimpleReadOnlyGlanceClientTest test_glance_timeout http://logs.openstack.org/66/55766/3/gate/gate-tempest-devstack-vm- postgres-full/a807434/testr_results.html.gz ft254.6: tempest.cli.simple_read_only.test_glance.SimpleReadOnlyGlanceClientTest.test_glance_timeout_StringException: Empty attachments: stderr stdout pythonlogging:'': {{{ 2013-11-29 07:57:20,345 running: '/usr/local/bin/glance --os-username admin --os-tenant-name admin --os-password secret --os-auth-url http://127.0.0.1:5000/v2.0/ --timeout 15 image-list ' 2013-11-29 07:57:37,633 output of /usr/local/bin/glance --os-username admin --os-tenant-name admin --os-password secret --os-auth-url http://127.0.0.1:5000/v2.0/ --timeout 15 image-list : 2013-11-29 07:57:37,635 error output of /usr/local/bin/glance --os-username admin --os-tenant-name admin --os-password secret --os-auth-url http://127.0.0.1:5000/v2.0/ --timeout 15 image-list : Error communicating with http://127.0.0.1:9292 timed out }}} Traceback (most recent call last): File "tempest/cli/simple_read_only/test_glance.py", line 89, in test_glance_timeout self.glance('image-list', flags='--timeout %d' % CONF.cli.timeout) File "tempest/cli/__init__.py", line 81, in glance 'glance', action, flags, params, admin, fail_ok) File "tempest/cli/__init__.py", line 110, in cmd_with_auth return self.cmd(cmd, action, flags, params, fail_ok) File "tempest/cli/__init__.py", line 132, in cmd stderr=result_err) CommandFailed: Command '['/usr/local/bin/glance', '--os-username', 'admin', '--os-tenant-name', 'admin', '--os-password', 'secret', '--os-auth-url', 'http://127.0.0.1:5000/v2.0/', '--timeout', '15', 'image-list']' returned non-zero exit status 1 To manage notifications about this bug go to: https://bugs.launchpad.net/glance/+bug/1257070/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1221899] Re: test_resize_server_from_auto_to_manual: server failed to reach VERIFY_RESIZE status within the required time
** Changed in: tempest Status: New => Incomplete ** Also affects: nova Importance: Undecided Status: New ** Changed in: tempest Status: Incomplete => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1221899 Title: test_resize_server_from_auto_to_manual: server failed to reach VERIFY_RESIZE status within the required time Status in OpenStack Compute (Nova): New Status in Tempest: Invalid Bug description: 2013-09-06 19:24:49.608 | Traceback (most recent call last): 2013-09-06 19:24:49.608 | File "tempest/api/compute/servers/test_disk_config.py", line 114, in test_resize_server_from_auto_to_manual 2013-09-06 19:24:49.609 | self.client.wait_for_server_status(server['id'], 'VERIFY_RESIZE') 2013-09-06 19:24:49.609 | File "tempest/services/compute/xml/servers_client.py", line 331, in wait_for_server_status 2013-09-06 19:24:49.609 | raise exceptions.TimeoutException(message) 2013-09-06 19:24:49.609 | TimeoutException: Request timed out 2013-09-06 19:24:49.609 | Details: Server dabbdc8d-3194-4e88-bc9c-c897a1fe5f78 failed to reach VERIFY_RESIZE status within the required time (400 s). Current status: RESIZE. http://logs.openstack.org/48/45248/2/check/gate-tempest-devstack-vm-full/66d555c/ To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1221899/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1218582] Re: ServerActionsTestJSON.test_pause_unpause_server fails with a timeout, other failures as a side effect
This is actually also a nova bug. Nova didn't move this into a paused state correctly. ** Also affects: nova Importance: Undecided Status: New ** Changed in: tempest Importance: Undecided => Medium ** Changed in: tempest Status: New => Confirmed -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1218582 Title: ServerActionsTestJSON.test_pause_unpause_server fails with a timeout, other failures as a side effect Status in OpenStack Compute (Nova): New Status in Tempest: Confirmed Bug description: In this case, it looks like a test for pausing an instance got stuck (reported as running the longest at 250 seconds). A number of other tests failed as a side effect it seems. They all report that they can't do what they needed to do because the instance is still in a pausing task state. http://logs.openstack.org/69/42769/7/gate/gate-tempest-devstack-vm-full/b2879b7 http://logs.openstack.org/69/42769/7/gate/gate-tempest-devstack-vm-full/b2879b7/console.html https://review.openstack.org/#/c/42769/ To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1218582/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1259553] Re: ListServersNegativeTestXML flakey ERROR generation
This isn't a tempest bug, because there wouldnt' be a code fix in tempest to fix it ** Changed in: tempest Status: New => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1259553 Title: ListServersNegativeTestXML flakey ERROR generation Status in OpenStack Compute (Nova): New Status in Tempest: Invalid Bug description: Looks like ListServersNegativeTestXML can trigger a race in Nova which causes an ERROR to be logged: 2013-12-10 12:41:34.628 ERROR nova.network.manager [req-7430295d-9c23-4a46-a755-f1e93aa53c6f ListServersNegativeTestXML-tempest-1313593245-user ListServersNegativeTestXML-tempest-1313593245-tenant] Unable to release 10.1.0.10 because vif doesn't exist. http://logs.openstack.org/12/61012/2/check/check-tempest-dsvm- full/c92bea6/logs/screen-n-net.txt.gz?level=ERROR Probably related to https://code.launchpad.net/bugs/968457 which is meant to be fixed, but the comment in the nova code suggests it might not be. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1259553/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260249] Re: migration-list: 'unicode' object has no attribute 'iteritems'
** Changed in: nova Status: Invalid => New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1260249 Title: migration-list: 'unicode' object has no attribute 'iteritems' Status in OpenStack Compute (Nova): New Status in Python client library for Nova: In Progress Bug description: There is an AttributeError when we try to use the command "nova migration-list" Traceback (most recent call last): File "/opt/stack/python-novaclient/novaclient/shell.py", line 721, in main OpenStackComputeShell().main(map(strutils.safe_decode, sys.argv[1:])) File "/opt/stack/python-novaclient/novaclient/shell.py", line 657, in main args.func(self.cs, args) File "/opt/stack/python-novaclient/novaclient/v1_1/contrib/migrations.py", line 71, in do_migration_list args.cell_name)) File "/opt/stack/python-novaclient/novaclient/v1_1/contrib/migrations.py", line 53, in list return self._list("/os-migrations%s" % query_string, "migrations") File "/opt/stack/python-novaclient/novaclient/base.py", line 80, in _list for res in data if res] File "/opt/stack/python-novaclient/novaclient/base.py", line 426, in __init__ self._add_details(info) File "/opt/stack/python-novaclient/novaclient/base.py", line 449, in _add_details for (k, v) in six.iteritems(info): File "/usr/local/lib/python2.7/dist-packages/six.py", line 439, in iteritems return iter(getattr(d, _iteritems)(**kw)) AttributeError: 'unicode' object has no attribute 'iteritems' ERROR: 'unicode' object has no attribute 'iteritems' To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1260249/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1258682] Re: timeout causing gate-tempest-dsvm-full to fail
If the fix is increasing the timeout in the gate, it's not a tempest bug. It looks like in this case libvirt went off the rails, so nova is probably a good bug choice ** Also affects: openstack-ci Importance: Undecided Status: New ** Changed in: tempest Status: New => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1258682 Title: timeout causing gate-tempest-dsvm-full to fail Status in OpenStack Compute (Nova): New Status in OpenStack Core Infrastructure: New Status in Tempest: Invalid Bug description: This has happened several times. A recent example is in https://jenkins02.openstack.org/job/gate-tempest-dsvm-full/775/console There are several mentions of FAIL in the logs, but since the job timed out, no console logs were saved. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1258682/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260359] [NEW] Adding a member to a LBaaS pool when there are no servers available it shows a "Success: Added member(s)." message.
Public bug reported: Steps to reproduce: 1) Make sure there are no running instances 2) Go to "Load Balancers" tab / "Add Member" 3) In the "Add Member" window click on "Add" Issue: No members are added and a "Success: Added member(s)." is shown. ** Affects: horizon Importance: Undecided Status: New ** Tags: lbaas -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1260359 Title: Adding a member to a LBaaS pool when there are no servers available it shows a "Success: Added member(s)." message. Status in OpenStack Dashboard (Horizon): New Bug description: Steps to reproduce: 1) Make sure there are no running instances 2) Go to "Load Balancers" tab / "Add Member" 3) In the "Add Member" window click on "Add" Issue: No members are added and a "Success: Added member(s)." is shown. To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1260359/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1258601] Re: nova.network.manager: Unable to release because vif doesn't exist.
*** This bug is a duplicate of bug 1258848 *** https://bugs.launchpad.net/bugs/1258848 Please include a pointer to the log file for such reports. According to logstash this has hit 48 times in the last two weeks which is a very low failure rate. Ideally flaky bugs like this would be fixed. If the nova team wants to silence this a patch can be submitted to the whitelist in tempest. ** Changed in: tempest Status: New => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1258601 Title: nova.network.manager: Unable to release because vif doesn't exist. Status in OpenStack Compute (Nova): New Status in Tempest: Invalid Bug description: This error shows un in nova-network log. Not sure if it needs to be whitelisted. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1258601/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1153926] Re: flavor show shouldn't read deleted flavors.
** No longer affects: tempest -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1153926 Title: flavor show shouldn't read deleted flavors. Status in OpenStack Dashboard (Horizon): In Progress Status in OpenStack Compute (Nova): In Progress Status in Python client library for Nova: In Progress Bug description: An instance type is created by: return db.instance_type_create(context.get_admin_context(), kwargs) which uses the read_deleted="no" from the admin context. This means, as seen in nova/tests/test_instance_types.py: def test_read_deleted_false_converting_flavorid(self): """ Ensure deleted instance types are not returned when not needed (for example when creating a server and attempting to translate from flavorid to instance_type_id. """ instance_types.create("instance_type1", 256, 1, 120, 100, "test1") instance_types.destroy("instance_type1") instance_types.create("instance_type1_redo", 256, 1, 120, 100, "test1") instance_type = instance_types.get_instance_type_by_flavor_id( "test1", read_deleted="no") self.assertEqual("instance_type1_redo", instance_type["name"]) flavors with colliding ids can exist in the database. From the test we see this looks intended, however it results in undesirable results if we consider the following scenario. For 'show' in the flavors api, it uses read_deleted="yes". The reason for this is if a vm was created in the past with a now-deleted flavor, 'nova show' can still show the flavor name that was specified for that vm creation. The flavor name is retrieved using the flavor id stored with the instance. Well, if there are colliding flavor ids in the database, the first of the duplicates will be picked, and it may not be the correct flavor for the vm. This leads me to believe that maybe at flavor create time, colliding ids should not be allowed, i.e. use return db.instance_type_create(context.get_admin_context(read_deleted="yes"), kwargs) to prevent the possibility of colliding flavor ids. To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1153926/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1224518] Re: test_reboot_server_hard fails sporadically in swift check jobs
10 results in logstash with "Current status: HARD_REBOOT" . ** Also affects: nova Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1224518 Title: test_reboot_server_hard fails sporadically in swift check jobs Status in OpenStack Compute (Nova): New Status in Tempest: New Bug description: See: http://logs.openstack.org/46/46146/2/check/gate-tempest-devstack- vm-postgres-full/b2712f1/console.html 2013-09-12 04:43:17.625 | == 2013-09-12 04:43:17.649 | FAIL: tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_reboot_server_hard[gate,smoke] 2013-09-12 04:43:17.651 | tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_reboot_server_hard[gate,smoke] 2013-09-12 04:43:17.652 | -- 2013-09-12 04:43:17.652 | _StringException: Empty attachments: 2013-09-12 04:43:17.652 | stderr 2013-09-12 04:43:17.652 | stdout 2013-09-12 04:43:17.653 | 2013-09-12 04:43:17.653 | pythonlogging:'': {{{ 2013-09-12 04:43:17.653 | 2013-09-12 04:16:55,739 Request: GET http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf 2013-09-12 04:43:17.654 | 2013-09-12 04:16:55,806 Response Status: 200 2013-09-12 04:43:17.654 | 2013-09-12 04:16:55,806 Nova request id: req-cdc6b1fc-bcf2-4e9c-bea1-8bf935993cbd 2013-09-12 04:43:17.654 | 2013-09-12 04:16:55,807 Request: POST http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf/action 2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,917 Response Status: 202 2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,917 Nova request id: req-3af37dd3-0ddc-4daa-aa6f-6958a5073cc4 2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,918 Request: GET http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf 2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,986 Response Status: 200 2013-09-12 04:43:17.656 | 2013-09-12 04:16:55,986 Nova request id: req-a7298d3e-167c-4c8f-9506-6064ba811e5b . . . 2013-09-12 04:43:17.976 | 2013-09-12 04:23:35,773 Request: GET http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf 2013-09-12 04:43:17.976 | 2013-09-12 04:23:35,822 Response Status: 200 2013-09-12 04:43:17.976 | 2013-09-12 04:23:35,823 Nova request id: req-a122aded-b49b-4847-9920-b2b8b09bc0ca 2013-09-12 04:43:17.976 | }}} 2013-09-12 04:43:17.977 | 2013-09-12 04:43:17.977 | Traceback (most recent call last): 2013-09-12 04:43:17.978 | File "tempest/api/compute/servers/test_server_actions.py", line 81, in test_reboot_server_hard 2013-09-12 04:43:17.978 | self.client.wait_for_server_status(self.server_id, 'ACTIVE') 2013-09-12 04:43:17.979 | File "tempest/services/compute/json/servers_client.py", line 176, in wait_for_server_status 2013-09-12 04:43:17.979 | raise exceptions.TimeoutException(message) 2013-09-12 04:43:17.979 | TimeoutException: Request timed out 2013-09-12 04:43:17.980 | Details: Server 8ad0ad9a-3975-486f-94b4-af1c89b51aaf failed to reach ACTIVE status within the required time (400 s). Current status: HARD_REBOOT. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1224518/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1259542] Re: Send OS distribution on API headers
Moved to oslo as there's nothing keystone-specific about this. ** Changed in: keystone Importance: Undecided => Wishlist ** Project changed: keystone => oslo -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to Keystone. https://bugs.launchpad.net/bugs/1259542 Title: Send OS distribution on API headers Status in Oslo - a Library of Common OpenStack Code: In Progress Bug description: It should be interesting to send OS distribution in API headers, to be able to detect the OS that is serving API calls and collect that on stats. To manage notifications about this bug go to: https://bugs.launchpad.net/oslo/+bug/1259542/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1152623] Re: RFC2616 section 9.7 status code vs. nova server delete
Until this is changed in nova, it's not actually appropriate to have a tempest issue ** No longer affects: tempest -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1152623 Title: RFC2616 section 9.7 status code vs. nova server delete Status in OpenStack Compute (Nova): In Progress Bug description: In REST client implementation is common good practice, when: - request causes an synchronous and asynchronous effect , and - the synchronous operation has any immediately visible effect ie. immediate subsequent request showing any change, we should emphasize the synchronous behavior in the responses (Status code) (Or responding in way which does not distinguish the two cases). However if the HTTP method is DELETE, the rule is the opposite! If the resource on the request URL does not deleted the service MUST NOT response with 204. " A successful response SHOULD be 200 (OK) if the response includes an entity describing the status, 202 (Accepted) if the action has not yet been enacted, or 204 (No Content) if the action has been enacted but the response does not include an entity. " by RFC2616 section 9.7 It means if a DELETE request responded with 204 status code, I MUST get 404 in an immediate subsequent request, unless concurrent operation recreated the resource. $ nova --debug delete ab0ebda6-2c21-4258-8934-1005b970fee5 ; nova --debug show ab0ebda6-2c21-4258-8934-1005b970fee5 Part of the output in the received order: - REQ: curl -i http://10.34.69.149:8774/v2/89a38fe6d3194864995ab0872905a65e/servers/ab0ebda6-2c21-4258-8934-1005b970fee5 -X DELETE -H "X-Auth-Project-Id: admin" -H "User-Agent: python-novaclient" -H "Accept: application/json" -H "X-Auth-Token: c35f5783528d4131bf100604b2fabd6c" send: u'DELETE /v2/89a38fe6d3194864995ab0872905a65e/servers/ab0ebda6-2c21-4258-8934-1005b970fee5 HTTP/1.1\r\nHost: 10.34.69.149:8774\r\nx-auth-project-id: admin\r\nx-auth-token: c35f5783528d4131bf100604b2fabd6c\r\naccept-encoding: gzip, deflate\r\naccept: application/json\r\nuser-agent: python-novaclient\r\n\r\n' reply: 'HTTP/1.1 204 No Content\r\n' header: Content-Length: 0 header: X-Compute-Request-Id: req-53e3503a-8d73-4ffc-ba43-4bd5659a9e22 header: Content-Type: application/json header: Date: Sat, 02 Mar 2013 18:26:21 GMT RESP:{'date': 'Sat, 02 Mar 2013 18:26:21 GMT', 'status': '204', 'content-length': '0', 'content-type': 'application/json', 'x-compute-request-id': 'req-53e3503a-8d73-4ffc-ba43-4bd5659a9e22'} - REQ: curl -i http://10.34.69.149:8774/v2/89a38fe6d3194864995ab0872905a65e/servers/ab0ebda6-2c21-4258-8934-1005b970fee5 -X GET -H "X-Auth-Project-Id: admin" -H "User-Agent: python-novaclient" -H "Accept: application/json" -H "X-Auth-Token: f74d6c7226c14915a26a81b540d43f3b" connect: (10.34.69.149, 8774) send: u'GET /v2/89a38fe6d3194864995ab0872905a65e/servers/ab0ebda6-2c21-4258-8934-1005b970fee5 HTTP/1.1\r\nHost: 10.34.69.149:8774\r\nx-auth-project-id: admin\r\nx-auth-token: f74d6c7226c14915a26a81b540d43f3b\r\naccept-encoding: gzip, deflate\r\naccept: application/json\r\nuser-agent: python-novaclient\r\n\r\n' reply: 'HTTP/1.1 200 OK\r\n' header: X-Compute-Request-Id: req-80c97c68-0b44-4650-b027-84a85ee04b86 header: Content-Type: application/json header: Content-Length: 1502 header: Date: Sat, 02 Mar 2013 18:26:21 GMT RESP:{'status': '200', 'content-length': '1502', 'content-location': u'http://10.34.69.149:8774/v2/89a38fe6d3194864995ab0872905a65e/servers/ab0ebda6-2c21-4258-8934-1005b970fee5', 'x-compute-request-id': 'req-80c97c68-0b44-4650-b027-84a85ee04b86', 'date': 'Sat, 02 Mar 2013 18:26:21 GMT', 'content-type': 'application/json'} {"server": {"status": "ACTIVE", "updated": "2013-03-02T18:26:21Z", "hostId": "31bdffcdffd5b869b87c9be3cdd700e29c4a08286d6d306622b4815a", "OS-EXT-SRV-ATTR:host": "new32.lithium.rhev.lab.eng.brq.redhat.com", "addresses": {"novanetwork": [{"version": 4, "addr": "192.168.32.2"}]}, "links": [{"href": "http://10.34.69.149:8774/v2/89a38fe6d3194864995ab0872905a65e/servers/ab0ebda6-2c21-4258-8934-1005b970fee5";, "rel": "self"}, {"href": "http://10.34.69.149:8774/89a38fe6d3194864995ab0872905a65e/servers/ab0ebda6-2c21-4258-8934-1005b970fee5";, "rel": "bookmark"}], "key_name": null, "image": {"id": "12e9c131-aaf4-4f73-9659-ed2da9759cd2", "links": [{"href": "http://10.34.69.149: 8774/89a38fe6d3194864995ab0872905a65e/images/12e9c131-aaf4-4f73-9659-ed2da9759cd2", "rel": "bookmark"}]}, "OS-EXT-STS:task_state": "deleting", "OS-EXT-STS:vm_state": "active", "OS-EXT-SRV-ATTR:instance_name": "instance-0003", "OS-EXT-SRV-ATTR:hypervisor_hostname": "new32.lithium.rhev.lab.eng.brq.redhat.com", "flavor": {"id": "1", "links": [{"href": "http://10.34.69
[Yahoo-eng-team] [Bug 1006725] Re: Incorrect error returned during Create Image and multi byte characters used for Image name
Definitely not fixed on the nova side. Attempts to unskip the bug generated issues. ** Changed in: nova Status: Invalid => Confirmed ** Changed in: nova Importance: Low => High -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1006725 Title: Incorrect error returned during Create Image and multi byte characters used for Image name Status in OpenStack Compute (Nova): Confirmed Status in Tempest: In Progress Bug description: Our tempest tests that checks for 400 Bad Request return code fails with a ComputeFault instead. Pass multi-byte character image name during Create Image Actual Response Code: ComputeFault, 500 Expected Response Code: 400 Bad Request Return an error if the server name has a multi-byte character ... FAIL == FAIL: Return an error if the server name has a multi-byte character -- Traceback (most recent call last): File "/opt/stack/tempest/tests/test_images.py", line 251, in test_create_image_specify_multibyte_character_server_name self.fail("Should return 400 Bad Request if multi byte characters" AssertionError: Should return 400 Bad Request if multi byte characters are used for image name >> begin captured logging << tempest.config: INFO: Using tempest config file /opt/stack/tempest/etc/tempest.conf tempest.common.rest_client: ERROR: Request URL: http://10.2.3.164:8774/v2/1aeac1cfbfdd43c2845b2cb3a4f15790/images/24ceff93-1af3-41ab-802f-9fc4d8b90b69 tempest.common.rest_client: ERROR: Request Body: None tempest.common.rest_client: ERROR: Response Headers: {'date': 'Thu, 31 May 2012 06:02:33 GMT', 'status': '404', 'content-length': '62', 'content-type': 'application/json; charset=UTF-8', 'x-compute-request-id': 'req-7a15d284-e934-47a1-87f4-7746e949c7a2'} tempest.common.rest_client: ERROR: Response Body: {"itemNotFound": {"message": "Image not found.", "code": 404}} tempest.common.rest_client: ERROR: Request URL: http://10.2.3.164:8774/v2/1aeac1cfbfdd43c2845b2cb3a4f15790/servers/ecb51dfb-493d-4ef8-9178-1adc3d96a04d/action tempest.common.rest_client: ERROR: Request Body: {"createImage": {"name": "\ufeff43802479847"}} tempest.common.rest_client: ERROR: Response Headers: {'date': 'Thu, 31 May 2012 06:02:44 GMT', 'status': '500', 'content-length': '128', 'content-type': 'application/json; charset=UTF-8', 'x-compute-request-id': 'req-1a9505f5-4dfc-44e7-b04a-f8daec0f956e'} tempest.common.rest_client: ERROR: Response Body: {u'computeFault': {u'message': u'The server has either erred or is incapable of performing the requested operation.', u'code': 500}} - >> end captured logging << - To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1006725/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1226943] Re: Need Use built-in print() function instead of print statement
** No longer affects: tempest -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1226943 Title: Need Use built-in print() function instead of print statement Status in OpenStack Telemetry (Ceilometer): Fix Released Status in Cinder: Fix Released Status in OpenStack Image Registry and Delivery Service (Glance): Fix Released Status in Orchestration API (Heat): Fix Released Status in OpenStack Identity (Keystone): In Progress Status in OpenStack Neutron (virtual network service): Fix Released Status in OpenStack Compute (Nova): Fix Released Bug description: In python 3 print statement is not supported, so we should use only print() functions. built-in function was introduce in python 2.6 http://www.python.org/dev/peps/pep-3105/ Note :This function is not normally available as a built-in since the name print is recognized as the print statement. To disable the statement and use the print() function, use this future statement at the top of your module: from __future__ import print_function http://docs.python.org/2/library/functions.html#print To manage notifications about this bug go to: https://bugs.launchpad.net/ceilometer/+bug/1226943/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1216903] Re: logical_resource_id disappeared in favor of resource_name
** Changed in: tempest Status: Fix Committed => Fix Released -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1216903 Title: logical_resource_id disappeared in favor of resource_name Status in OpenStack Dashboard (Horizon): Fix Released Status in Python client library for heat: Fix Released Status in Tempest: Fix Released Bug description: impact from Heat side : https://review.openstack.org/#/c/43391/ To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1216903/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1239891] [NEW] tempest.api.object_storage.test_account_services.AccountTest fails under neutron-pg-isolated
You have been subscribed to a public bug: http://logs.openstack.org/38/51738/1/check/check-tempest-devstack-vm- neutron-pg-isolated/73aad7a/console.html 2013-10-15 00:19:04.556 | Error in atexit._run_exitfuncs: 2013-10-15 00:19:04.556 | Traceback (most recent call last): 2013-10-15 00:19:04.556 | File "/usr/lib/python2.7/atexit.py", line 24, in _run_exitfuncs 2013-10-15 00:19:04.557 | func(*targs, **kargs) 2013-10-15 00:19:04.558 | File "tempest/test.py", line 167, in validate_tearDownClass 2013-10-15 00:19:04.558 | + str(at_exit_set)) 2013-10-15 00:19:04.558 | RuntimeError: tearDownClass does not calls the super's tearDownClass in these classes: set([]) 2013-10-15 00:19:04.559 | Error in sys.exitfunc: 2013-10-15 00:19:04.663 | 2013-10-15 00:19:04.664 | process-returncode 2013-10-15 00:19:04.664 | process-returncode ... FAIL 2013-10-15 00:19:04.980 | 2013-10-15 00:19:04.981 | == 2013-10-15 00:19:04.981 | FAIL: tearDownClass (tempest.api.object_storage.test_account_services.AccountTest) 2013-10-15 00:19:04.981 | tearDownClass (tempest.api.object_storage.test_account_services.AccountTest) 2013-10-15 00:19:04.982 | -- 2013-10-15 00:19:04.982 | _StringException: Traceback (most recent call last): 2013-10-15 00:19:04.982 | File "tempest/api/object_storage/test_account_services.py", line 41, in tearDownClass 2013-10-15 00:19:04.983 | super(AccountTest, cls).tearDownClass() 2013-10-15 00:19:04.983 | File "tempest/api/object_storage/base.py", line 77, in tearDownClass 2013-10-15 00:19:04.983 | cls.isolated_creds.clear_isolated_creds() 2013-10-15 00:19:04.984 | File "tempest/common/isolated_creds.py", line 453, in clear_isolated_creds 2013-10-15 00:19:04.984 | self._clear_isolated_net_resources() 2013-10-15 00:19:04.984 | File "tempest/common/isolated_creds.py", line 445, in _clear_isolated_net_resources 2013-10-15 00:19:04.985 | self._clear_isolated_network(network['id'], network['name']) 2013-10-15 00:19:04.985 | File "tempest/common/isolated_creds.py", line 399, in _clear_isolated_network 2013-10-15 00:19:04.985 | net_client.delete_network(network_id) 2013-10-15 00:19:04.985 | File "tempest/services/network/json/network_client.py", line 76, in delete_network 2013-10-15 00:19:04.986 | resp, body = self.delete(uri, self.headers) 2013-10-15 00:19:04.986 | File "tempest/common/rest_client.py", line 308, in delete 2013-10-15 00:19:04.986 | return self.request('DELETE', url, headers) 2013-10-15 00:19:04.987 | File "tempest/common/rest_client.py", line 436, in request 2013-10-15 00:19:04.987 | resp, resp_body) 2013-10-15 00:19:04.987 | File "tempest/common/rest_client.py", line 522, in _error_checker 2013-10-15 00:19:04.988 | raise exceptions.ComputeFault(message) 2013-10-15 00:19:04.988 | ComputeFault: Got compute fault 2013-10-15 00:19:04.988 | Details: {"NeutronError": "Request Failed: internal server error while processing your request."} 2013-10-15 00:19:04.988 | 2013-10-15 00:19:04.989 | 2013-10-15 00:19:04.989 | == 2013-10-15 00:19:04.989 | FAIL: process-returncode 2013-10-15 00:19:04.990 | process-returncode 2013-10-15 00:19:04.990 | -- 2013-10-15 00:19:04.990 | _StringException: Binary content: 2013-10-15 00:19:04.991 | traceback (test/plain; charset="utf8") 2013-10-15 00:19:04.991 | 2013-10-15 00:19:04.991 | 2013-10-15 00:19:04.991 | -- ** Affects: neutron Importance: Undecided Status: New -- tempest.api.object_storage.test_account_services.AccountTest fails under neutron-pg-isolated https://bugs.launchpad.net/bugs/1239891 You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1232971] Re: tempest gating error: test_run_stop_terminate_instance_with_tags
** Changed in: tempest Status: Confirmed => Fix Released -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1232971 Title: tempest gating error: test_run_stop_terminate_instance_with_tags Status in OpenStack Neutron (virtual network service): Confirmed Status in Tempest: Fix Released Bug description: 2013-09-29 21:56:27.035 | == 2013-09-29 21:56:27.035 | FAIL: tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance_with_tags[gate,smoke] 2013-09-29 21:56:27.035 | tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance_with_tags[gate,smoke] 2013-09-29 21:56:27.035 | -- 2013-09-29 21:56:27.036 | _StringException: Empty attachments: 2013-09-29 21:56:27.036 | stderr 2013-09-29 21:56:27.036 | stdout 2013-09-29 21:56:27.036 | 2013-09-29 21:56:27.036 | pythonlogging:'': {{{2013-09-29 21:42:26,545 state: pending}}} 2013-09-29 21:56:27.036 | 2013-09-29 21:56:27.037 | Traceback (most recent call last): 2013-09-29 21:56:27.037 | File "tempest/thirdparty/boto/test_ec2_instance_run.py", line 175, in test_run_stop_terminate_instance_with_tags 2013-09-29 21:56:27.037 | self.assertInstanceStateWait(instance, "running") 2013-09-29 21:56:27.037 | File "tempest/thirdparty/boto/test.py", line 356, in assertInstanceStateWait 2013-09-29 21:56:27.037 | state = self.waitInstanceState(lfunction, wait_for) 2013-09-29 21:56:27.037 | File "tempest/thirdparty/boto/test.py", line 341, in waitInstanceState 2013-09-29 21:56:27.037 | self.valid_instance_state) 2013-09-29 21:56:27.038 | File "tempest/thirdparty/boto/test.py", line 331, in state_wait_gone 2013-09-29 21:56:27.038 | state = state_wait(lfunction, final_set, valid_set) 2013-09-29 21:56:27.038 | File "tempest/thirdparty/boto/utils/wait.py", line 57, in state_wait 2013-09-29 21:56:27.038 | (dtime, final_set, status)) 2013-09-29 21:56:27.038 | AssertionError: State change timeout exceeded!(400s) While waitingfor set(['running', '_GONE']) at "pending" To manage notifications about this bug go to: https://bugs.launchpad.net/neutron/+bug/1232971/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1232303] Re: FAIL: tempest test_large_ops_scenario - failed to get to expected status. In ERROR state.
** No longer affects: tempest -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1232303 Title: FAIL: tempest test_large_ops_scenario - failed to get to expected status. In ERROR state. Status in OpenStack Neutron (virtual network service): New Status in OpenStack Compute (Nova): New Bug description: http://logs.openstack.org/13/48513/2/check/check-tempest-devstack-vm- large-ops/8ce3344/console.html 2013-09-27 21:08:58.498 | == 2013-09-27 21:08:58.498 | FAIL: tempest.scenario.test_large_ops.TestLargeOpsScenario.test_large_ops_scenario[compute,image] 2013-09-27 21:08:58.498 | tags: worker-0 2013-09-27 21:08:58.498 | -- 2013-09-27 21:08:58.498 | Empty attachments: 2013-09-27 21:08:58.498 | stderr 2013-09-27 21:08:58.499 | stdout 2013-09-27 21:08:58.499 | 2013-09-27 21:08:58.499 | pythonlogging:'': {{{ 2013-09-27 21:08:58.499 | 2013-09-27 21:04:54,743 Starting new HTTP connection (1): 127.0.0.1 2013-09-27 21:08:58.499 | 2013-09-27 21:04:54,864 Starting new HTTP connection (1): 127.0.0.1 2013-09-27 21:08:58.499 | }}} 2013-09-27 21:08:58.500 | 2013-09-27 21:08:58.500 | Traceback (most recent call last): 2013-09-27 21:08:58.500 | File "tempest/scenario/test_large_ops.py", line 105, in test_large_ops_scenario 2013-09-27 21:08:58.500 | self.nova_boot() 2013-09-27 21:08:58.500 | File "tempest/scenario/test_large_ops.py", line 98, in nova_boot 2013-09-27 21:08:58.500 | self._wait_for_server_status('ACTIVE') 2013-09-27 21:08:58.501 | File "tempest/scenario/test_large_ops.py", line 42, in _wait_for_server_status 2013-09-27 21:08:58.501 | self.compute_client.servers, server.id, status) 2013-09-27 21:08:58.501 | File "tempest/scenario/manager.py", line 290, in status_timeout 2013-09-27 21:08:58.501 | self._status_timeout(things, thing_id, expected_status=expected_status) 2013-09-27 21:08:58.501 | File "tempest/scenario/manager.py", line 338, in _status_timeout 2013-09-27 21:08:58.501 | self.config.compute.build_interval): 2013-09-27 21:08:58.502 | File "tempest/test.py", line 237, in call_until_true 2013-09-27 21:08:58.502 | if func(): 2013-09-27 21:08:58.502 | File "tempest/scenario/manager.py", line 329, in check_status 2013-09-27 21:08:58.502 | raise exceptions.BuildErrorException(message) 2013-09-27 21:08:58.502 | BuildErrorException: Server %(server_id)s failed to build and is in ERROR status 2013-09-27 21:08:58.503 | Details: failed to get to expected status. In ERROR state. To manage notifications about this bug go to: https://bugs.launchpad.net/neutron/+bug/1232303/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1230407] Re: VMs can't progress through state changes because Neutron is deadlocking on it's database queries, and thus leaving networks in inconsistent states
** No longer affects: tempest ** No longer affects: nova -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1230407 Title: VMs can't progress through state changes because Neutron is deadlocking on it's database queries, and thus leaving networks in inconsistent states Status in OpenStack Neutron (virtual network service): Confirmed Bug description: This is most often seen with the "State change timeout exceeded" in the tempest logs. 2013-09-25 16:03:28.319 | FAIL: tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance_with_tags[gate,smoke] 2013-09-25 16:03:28.319 | tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance_with_tags[gate,smoke] 2013-09-25 16:03:28.319 | -- 2013-09-25 16:03:28.319 | _StringException: Empty attachments: 2013-09-25 16:03:28.319 | stderr 2013-09-25 16:03:28.320 | stdout 2013-09-25 16:03:28.320 | 2013-09-25 16:03:28.320 | pythonlogging:'': {{{2013-09-25 15:49:34,792 state: pending}}} 2013-09-25 16:03:28.320 | 2013-09-25 16:03:28.320 | Traceback (most recent call last): 2013-09-25 16:03:28.320 | File "tempest/thirdparty/boto/test_ec2_instance_run.py", line 175, in test_run_stop_terminate_instance_with_tags 2013-09-25 16:03:28.320 | self.assertInstanceStateWait(instance, "running") 2013-09-25 16:03:28.321 | File "tempest/thirdparty/boto/test.py", line 356, in assertInstanceStateWait 2013-09-25 16:03:28.321 | state = self.waitInstanceState(lfunction, wait_for) 2013-09-25 16:03:28.321 | File "tempest/thirdparty/boto/test.py", line 341, in waitInstanceState 2013-09-25 16:03:28.321 | self.valid_instance_state) 2013-09-25 16:03:28.321 | File "tempest/thirdparty/boto/test.py", line 331, in state_wait_gone 2013-09-25 16:03:28.321 | state = state_wait(lfunction, final_set, valid_set) 2013-09-25 16:03:28.322 | File "tempest/thirdparty/boto/utils/wait.py", line 57, in state_wait 2013-09-25 16:03:28.322 | (dtime, final_set, status)) 2013-09-25 16:03:28.322 | AssertionError: State change timeout exceeded!(400s) While waitingfor set(['running', '_GONE']) at "pending" full log: http://logs.openstack.org/38/47438/1/gate/gate-tempest- devstack-vm-neutron/93db162/ To manage notifications about this bug go to: https://bugs.launchpad.net/neutron/+bug/1230407/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1230354] Re: tempest.scenario.test_minimum_basic.TestMinimumBasicScenario.test_minimum_basic_scenario fails sporadically
This bug is basically useless. It doesn't get far enough to get to any root causes. I'm marking this as invalid and the narrower bugs can be refiled. ** No longer affects: cinder ** No longer affects: nova ** Changed in: tempest Status: Confirmed => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1230354 Title: tempest.scenario.test_minimum_basic.TestMinimumBasicScenario.test_minimum_basic_scenario fails sporadically Status in Tempest: Invalid Bug description: See: http://logs.openstack.org/97/44097/18/check/gate-tempest- devstack-vm-postgres-full/270e611/console.html 2013-09-25 15:15:37.447 | == 2013-09-25 15:15:37.463 | FAIL: tempest.scenario.test_minimum_basic.TestMinimumBasicScenario.test_minimum_basic_scenario[compute,image,network,volume] 2013-09-25 15:15:37.464 | tempest.scenario.test_minimum_basic.TestMinimumBasicScenario.test_minimum_basic_scenario[compute,image,network,volume] 2013-09-25 15:15:37.464 | -- 2013-09-25 15:15:37.464 | _StringException: Empty attachments: 2013-09-25 15:15:37.464 | stderr 2013-09-25 15:15:37.464 | stdout 2013-09-25 15:15:37.464 | 2013-09-25 15:15:37.465 | pythonlogging:'': {{{ 2013-09-25 15:15:37.465 | 2013-09-25 15:07:17,364 Starting new HTTP connection (1): 127.0.0.1 2013-09-25 15:15:37.465 | 2013-09-25 15:07:17,628 Starting new HTTP connection (1): 127.0.0.1 2013-09-25 15:15:37.465 | 2013-09-25 15:07:29,212 Starting new HTTP connection (1): 127.0.0.1 . . . 2013-09-25 15:15:37.619 | 2013-09-25 15:14:11,445 Starting new HTTP connection (1): 127.0.0.1 2013-09-25 15:15:37.619 | 2013-09-25 15:14:12,529 Starting new HTTP connection (1): 127.0.0.1 2013-09-25 15:15:37.620 | 2013-09-25 15:14:13,617 Starting new HTTP connection (1): 127.0.0.1 2013-09-25 15:15:37.620 | }}} 2013-09-25 15:15:37.620 | 2013-09-25 15:15:37.620 | Traceback (most recent call last): 2013-09-25 15:15:37.621 | File "tempest/scenario/test_minimum_basic.py", line 158, in test_minimum_basic_scenario 2013-09-25 15:15:37.621 | self.nova_volume_attach() 2013-09-25 15:15:37.621 | File "tempest/scenario/test_minimum_basic.py", line 120, in nova_volume_attach 2013-09-25 15:15:37.622 | self._wait_for_volume_status('in-use') 2013-09-25 15:15:37.622 | File "tempest/scenario/test_minimum_basic.py", line 48, in _wait_for_volume_status 2013-09-25 15:15:37.622 | self.volume_client.volumes, volume_id, status) 2013-09-25 15:15:37.622 | File "tempest/scenario/manager.py", line 290, in status_timeout 2013-09-25 15:15:37.622 | self._status_timeout(things, thing_id, expected_status=expected_status) 2013-09-25 15:15:37.623 | File "tempest/scenario/manager.py", line 341, in _status_timeout 2013-09-25 15:15:37.623 | raise exceptions.TimeoutException(message) 2013-09-25 15:15:37.623 | TimeoutException: Request timed out 2013-09-25 15:15:37.623 | Details: Timed out waiting for thing b5781471-5ee9-44de-944b-28c10a793b31 to become in-use To manage notifications about this bug go to: https://bugs.launchpad.net/tempest/+bug/1230354/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1239891] Re: tempest.api.object_storage.test_account_services.AccountTest fails under neutron-pg-isolated
** Project changed: tempest => neutron -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1239891 Title: tempest.api.object_storage.test_account_services.AccountTest fails under neutron-pg-isolated Status in OpenStack Neutron (virtual network service): New Bug description: http://logs.openstack.org/38/51738/1/check/check-tempest-devstack-vm- neutron-pg-isolated/73aad7a/console.html 2013-10-15 00:19:04.556 | Error in atexit._run_exitfuncs: 2013-10-15 00:19:04.556 | Traceback (most recent call last): 2013-10-15 00:19:04.556 | File "/usr/lib/python2.7/atexit.py", line 24, in _run_exitfuncs 2013-10-15 00:19:04.557 | func(*targs, **kargs) 2013-10-15 00:19:04.558 | File "tempest/test.py", line 167, in validate_tearDownClass 2013-10-15 00:19:04.558 | + str(at_exit_set)) 2013-10-15 00:19:04.558 | RuntimeError: tearDownClass does not calls the super's tearDownClass in these classes: set([]) 2013-10-15 00:19:04.559 | Error in sys.exitfunc: 2013-10-15 00:19:04.663 | 2013-10-15 00:19:04.664 | process-returncode 2013-10-15 00:19:04.664 | process-returncode ... FAIL 2013-10-15 00:19:04.980 | 2013-10-15 00:19:04.981 | == 2013-10-15 00:19:04.981 | FAIL: tearDownClass (tempest.api.object_storage.test_account_services.AccountTest) 2013-10-15 00:19:04.981 | tearDownClass (tempest.api.object_storage.test_account_services.AccountTest) 2013-10-15 00:19:04.982 | -- 2013-10-15 00:19:04.982 | _StringException: Traceback (most recent call last): 2013-10-15 00:19:04.982 | File "tempest/api/object_storage/test_account_services.py", line 41, in tearDownClass 2013-10-15 00:19:04.983 | super(AccountTest, cls).tearDownClass() 2013-10-15 00:19:04.983 | File "tempest/api/object_storage/base.py", line 77, in tearDownClass 2013-10-15 00:19:04.983 | cls.isolated_creds.clear_isolated_creds() 2013-10-15 00:19:04.984 | File "tempest/common/isolated_creds.py", line 453, in clear_isolated_creds 2013-10-15 00:19:04.984 | self._clear_isolated_net_resources() 2013-10-15 00:19:04.984 | File "tempest/common/isolated_creds.py", line 445, in _clear_isolated_net_resources 2013-10-15 00:19:04.985 | self._clear_isolated_network(network['id'], network['name']) 2013-10-15 00:19:04.985 | File "tempest/common/isolated_creds.py", line 399, in _clear_isolated_network 2013-10-15 00:19:04.985 | net_client.delete_network(network_id) 2013-10-15 00:19:04.985 | File "tempest/services/network/json/network_client.py", line 76, in delete_network 2013-10-15 00:19:04.986 | resp, body = self.delete(uri, self.headers) 2013-10-15 00:19:04.986 | File "tempest/common/rest_client.py", line 308, in delete 2013-10-15 00:19:04.986 | return self.request('DELETE', url, headers) 2013-10-15 00:19:04.987 | File "tempest/common/rest_client.py", line 436, in request 2013-10-15 00:19:04.987 | resp, resp_body) 2013-10-15 00:19:04.987 | File "tempest/common/rest_client.py", line 522, in _error_checker 2013-10-15 00:19:04.988 | raise exceptions.ComputeFault(message) 2013-10-15 00:19:04.988 | ComputeFault: Got compute fault 2013-10-15 00:19:04.988 | Details: {"NeutronError": "Request Failed: internal server error while processing your request."} 2013-10-15 00:19:04.988 | 2013-10-15 00:19:04.989 | 2013-10-15 00:19:04.989 | == 2013-10-15 00:19:04.989 | FAIL: process-returncode 2013-10-15 00:19:04.990 | process-returncode 2013-10-15 00:19:04.990 | -- 2013-10-15 00:19:04.990 | _StringException: Binary content: 2013-10-15 00:19:04.991 | traceback (test/plain; charset="utf8") 2013-10-15 00:19:04.991 | 2013-10-15 00:19:04.991 | 2013-10-15 00:19:04.991 | -- To manage notifications about this bug go to: https://bugs.launchpad.net/neutron/+bug/1239891/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260075] Re: VMware: NotAuthenticated occurred in the call to RetrievePropertiesEx
Marked High because this impacts the CI environment we use to approve other patches. If not for that context, I would mark this Medium. ** Changed in: nova Status: New => Confirmed ** Changed in: nova Importance: Undecided => High ** Changed in: nova Assignee: (unassigned) => Sabari Kumar Murugesan (smurugesan) ** Changed in: nova Milestone: None => icehouse-2 ** Also affects: openstack-vmwareapi-team Importance: Undecided Status: New ** Changed in: openstack-vmwareapi-team Status: New => Confirmed ** Changed in: openstack-vmwareapi-team Importance: Undecided => High -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1260075 Title: VMware: NotAuthenticated occurred in the call to RetrievePropertiesEx Status in OpenStack Compute (Nova): Confirmed Status in The OpenStack VMwareAPI subTeam: Confirmed Bug description: The VMware Minesweeper CI occasionally runs into this error when trying to boot an instance: 2013-12-11 04:50:15.048 20785 DEBUG nova.virt.vmwareapi.driver [-] Task [ReconfigVM_Task] (returnval){ value = "task-322" _type = "Task" } status: success _poll_task /opt/stack/nova/nova/virt/vmwareapi/driver.py:926 Reconfigured VM instance to enable vnc on port - 5986 _set_vnc_config /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1461 Instance failed to spawn Traceback (most recent call last): File "/opt/stack/nova/nova/compute/manager.py", line 1461, in _spawn block_device_info) File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 628, in spawn admin_password, network_info, block_device_info) File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 435, in spawn upload_folder, upload_name + ".vmdk")): File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1556, in _check_if_folder_file_exists "browser") File "/opt/stack/nova/nova/virt/vmwareapi/vim_util.py", line 173, in get_dynamic_property property_dict = get_dynamic_properties(vim, mobj, type, [property_name]) File "/opt/stack/nova/nova/virt/vmwareapi/vim_util.py", line 179, in get_dynamic_properties obj_content = get_object_properties(vim, None, mobj, type, property_names) File "/opt/stack/nova/nova/virt/vmwareapi/vim_util.py", line 168, in get_object_properties options=options) File "/opt/stack/nova/nova/virt/vmwareapi/vim.py", line 187, in vim_request_handler fault_checker(response) File "/opt/stack/nova/nova/virt/vmwareapi/error_util.py", line 99, in retrievepropertiesex_fault_checker exc_msg_list)) VimFaultException: Error(s) NotAuthenticated occurred in the call to RetrievePropertiesEx Full logs here for a CI build where this occurred are available here: http://162.209.83.206/logs/35303/31/ To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1260075/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1252854] Re: setUpClass ServerAddressesTest FAIL
Very unlikely tempest caused the ERROR status on the server, so adding nova. ** Also affects: nova Importance: Undecided Status: New ** Changed in: tempest Status: New => Incomplete -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1252854 Title: setUpClass ServerAddressesTest FAIL Status in OpenStack Compute (Nova): New Status in Tempest: Incomplete Bug description: I encountered the following Tempest failure during the initial Jenkins tests for the following submission to python-cinderclient: https://review.openstack.org/#/c/57245/ http://logs.openstack.org/45/57245/1/check/check-grenade-devstack- vm/1ad2c17/ 2013-11-19 16:07:09.167 | == 2013-11-19 16:07:09.167 | FAIL: setUpClass (tempest.api.compute.servers.test_server_addresses.ServerAddressesTest) 2013-11-19 16:07:09.167 | setUpClass (tempest.api.compute.servers.test_server_addresses.ServerAddressesTest) 2013-11-19 16:07:09.168 | -- 2013-11-19 16:07:09.168 | _StringException: Traceback (most recent call last): 2013-11-19 16:07:09.168 | File "tempest/api/compute/servers/test_server_addresses.py", line 31, in setUpClass 2013-11-19 16:07:09.169 | resp, cls.server = cls.create_test_server(wait_until='ACTIVE') 2013-11-19 16:07:09.169 | File "tempest/api/compute/base.py", line 118, in create_test_server 2013-11-19 16:07:09.169 | server['id'], kwargs['wait_until']) 2013-11-19 16:07:09.169 | File "tempest/services/compute/json/servers_client.py", line 160, in wait_for_server_status 2013-11-19 16:07:09.170 | extra_timeout=extra_timeout) 2013-11-19 16:07:09.170 | File "tempest/common/waiters.py", line 73, in wait_for_server_status 2013-11-19 16:07:09.170 | raise exceptions.BuildErrorException(server_id=server_id) 2013-11-19 16:07:09.171 | BuildErrorException: Server e4f08fa8-bf4c-4994-b5f9-97566a393baf failed to build and is in ERROR status To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1252854/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1239923] Re: neutron doesn't use request-ids
not a tempest bug ** No longer affects: tempest -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1239923 Title: neutron doesn't use request-ids Status in OpenStack Neutron (virtual network service): In Progress Bug description: Just like nova and cinder, neutron should use request-ids for logging and to return to users. request-ids are used to help look up the logs for a specific request. Also they make logs more usable. Here is an example of how they can be used. http://git.openstack.org/cgit/openstack/oslo- incubator/tree/openstack/common/context.py etc/cinder/cinder.conf.sample:#logging_context_format_string=%(asctime)s.%(msecs)03d %(process)d %(levelname)s %(name)s [%(request_id)s %(user)s %( To manage notifications about this bug go to: https://bugs.launchpad.net/neutron/+bug/1239923/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1242645] Re: Resource tracking does not take into account the current resources on the host
** Changed in: nova Assignee: Gary Kotton (garyk) => (unassigned) ** Changed in: nova Status: In Progress => Won't Fix ** Changed in: nova Milestone: icehouse-2 => None -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1242645 Title: Resource tracking does not take into account the current resources on the host Status in OpenStack Compute (Nova): Won't Fix Bug description: The trace below is from the log file: 2013-10-21 04:28:27.210 INFO nova.compute.resource_tracker [-] Hypervisor: free disk (GB): 11 2013-10-21 04:28:27.263 AUDIT nova.compute.resource_tracker [-] Free ram (MB): 4898 2013-10-21 04:28:27.263 AUDIT nova.compute.resource_tracker [-] Free disk (GB): 15 2013-10-21 04:28:27.263 AUDIT nova.compute.resource_tracker [-] Free VCPUS: 4 In this specific case there is 4GB used on the hypervisor - used for image cache etc. +--+-+ | Property | Value | +--+-+ | hypervisor_hostname | domain-c7(Cluster31) | | cpu_info | {"model": ["Intel(R) Xeon(R) CPU E5-2650 0 @ 2.00GHz"], "vendor": ["VMware, Inc."], "topology": {"cores": 4, "threads": 4}} | | free_disk_gb | 15 | | hypervisor_version | 51 | | disk_available_least | None | | local_gb | 15 | | free_ram_mb | 4898 | | id | 1 | | vcpus_used | 0 | | hypervisor_type | VMware vCenter Server | | local_gb_used| 0 | | memory_mb_used | 512 | | memory_mb| 5410 | | current_workload | 0 | | vcpus| 4 | | running_vms | 0 | | service_id | 5 | | service_host | os-devstack | +--+-+ To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1242645/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1226412] Re: tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML.test_rescue_paused_instance
This is a nova race bug ** Also affects: nova Importance: Undecided Status: New ** Changed in: nova Importance: Undecided => Medium ** Changed in: nova Status: New => Confirmed ** Summary changed: - tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML.test_rescue_paused_instance + guest doesn't reach PAUSED state within 200s in the gate ** Changed in: tempest Status: New => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1226412 Title: guest doesn't reach PAUSED state within 200s in the gate Status in OpenStack Compute (Nova): Confirmed Status in Tempest: Invalid Bug description: Tempest test fails :tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML.test_rescue_paused_instance Traceback (most recent call last): File "tempest/api/compute/servers/test_server_rescue.py", line 135, in test_rescue_paused_instance self.servers_client.wait_for_server_status(self.server_id, 'PAUSED') File "tempest/services/compute/xml/servers_client.py", line 340, in wait_for_server_status return waiters.wait_for_server_status(self, server_id, status) File "tempest/common/waiters.py", line 80, in wait_for_server_status raise exceptions.TimeoutException(message) TimeoutException: Request timed out Details: Server 8539b620-909c-46a6-9293-1b1add06a343 failed to reach PAUSED status within the required time (400 s). Current status: ACTIVE. see http://logs.openstack.org/55/46855/3/check/gate-tempest-devstack-vm-postgres-full/28acd2d/testr_results.html.gz To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1226412/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1240728] Re: tempest.api.compute.servers.test_server_rescue.ServerRescueTestJSON.test_rescued_vm_attach_volume is nondeterministic
** Changed in: nova Status: New => Confirmed ** Changed in: nova Importance: Undecided => High ** Changed in: cinder Importance: Undecided => High ** Changed in: cinder Status: Invalid => Confirmed -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1240728 Title: tempest.api.compute.servers.test_server_rescue.ServerRescueTestJSON.test_rescued_vm_attach_volume is nondeterministic Status in Cinder: Confirmed Status in OpenStack Compute (Nova): Confirmed Status in Tempest: New Bug description: Traceback (most recent call last): File "tempest/api/compute/servers/test_server_rescue.py", line 111, in _unrescue self.servers_client.wait_for_server_status(server_id, 'ACTIVE') File "tempest/services/compute/json/servers_client.py", line 156, in wait_for_server_status return waiters.wait_for_server_status(self, server_id, status) File "tempest/common/waiters.py", line 80, in wait_for_server_status raise exceptions.TimeoutException(message) TimeoutException: Request timed out Details: Server 802897a6-6793-4af2-9d84-8750be518380 failed to reach ACTIVE status within the required time (400 s). Current status: SHUTOFF. Sample failure: http://logs.openstack.org/51/52151/1/gate/gate- tempest-devstack-vm-full/6b393f5/ Basic query for the failure string: http://logstash.openstack.org/#eyJzZWFyY2giOiJAbWVzc2FnZTpcIkZBSUw6IHRlbXBlc3QuYXBpLmNvbXB1dGUuc2VydmVycy50ZXN0X3NlcnZlcl9yZXNjdWUuU2VydmVyUmVzY3VlVGVzdEpTT04udGVzdF9yZXNjdWVkX3ZtX2F0dGFjaF92b2x1bWVcIiIsImZpZWxkcyI6W10sIm9mZnNldCI6MCwidGltZWZyYW1lIjoiYWxsIiwiZ3JhcGhtb2RlIjoiY291bnQiLCJ0aW1lIjp7InVzZXJfaW50ZXJ2YWwiOjB9LCJzdGFtcCI6MTM4MTk2MTIyMjkwMSwibW9kZSI6IiIsImFuYWx5emVfZmllbGQiOiIifQ== To manage notifications about this bug go to: https://bugs.launchpad.net/cinder/+bug/1240728/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260406] [NEW] allow disable wsgi keepalive
Public bug reported: The wsgi server used in most if not all openstack services currently has keepalive=True by default thus keeping connections open after each request. This can cause problems when using load balancers in front of these services e.g. connections for requests that take a long time can get closed in the load balancers once a timeout has expired. This can then cause issues if a client performs a request using the same source port as a previous request that is not closed in the LB but still open in the server due to TCP packet sequecing in the LB and the new client not expecting the connection to already be open. So, it would be useful to be able to disable wsgi keepalive. ** Affects: keystone Importance: Undecided Assignee: Edward Hope-Morley (hopem) Status: In Progress ** Changed in: keystone Assignee: (unassigned) => Edward Hope-Morley (hopem) ** Changed in: keystone Status: New => In Progress -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to Keystone. https://bugs.launchpad.net/bugs/1260406 Title: allow disable wsgi keepalive Status in OpenStack Identity (Keystone): In Progress Bug description: The wsgi server used in most if not all openstack services currently has keepalive=True by default thus keeping connections open after each request. This can cause problems when using load balancers in front of these services e.g. connections for requests that take a long time can get closed in the load balancers once a timeout has expired. This can then cause issues if a client performs a request using the same source port as a previous request that is not closed in the LB but still open in the server due to TCP packet sequecing in the LB and the new client not expecting the connection to already be open. So, it would be useful to be able to disable wsgi keepalive. To manage notifications about this bug go to: https://bugs.launchpad.net/keystone/+bug/1260406/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1242916] Re: metadata server update_all expects body but doesn't get it passed to it
** No longer affects: tempest -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1242916 Title: metadata server update_all expects body but doesn't get it passed to it Status in OpenStack Compute (Nova): Confirmed Bug description: This recently started showing up in n-api. Seems like mishandling of invalid client args from a negative test in tempest. Example: http://logs.openstack.org/03/52803/3/check/check-tempest-devstack-vm-postgres-full/354d7a3/logs/screen-n-api.txt.gz 2013-10-21 20:04:51.724 20923 DEBUG routes.middleware [-] Matched PUT /6fa344aaf3034c4992bc30b3c06ad531/servers/a329912b-7874-4636-a08b-e40362e04ab2/metadata __call__ /usr/lib/python2.7/dist-packages/routes/middleware.py:100 2013-10-21 20:04:51.724 20923 DEBUG routes.middleware [-] Route path: '/{project_id}/servers/{server_id}/metadata', defaults: {'action': u'update_all', 'controller': } __call__ /usr/lib/python2.7/dist-packages/routes/middleware.py:102 2013-10-21 20:04:51.724 20923 DEBUG routes.middleware [-] Match dict: {'action': u'update_all', 'server_id': u'a329912b-7874-4636-a08b-e40362e04ab2', 'project_id': u'6fa344aaf3034c4992bc30b3c06ad531', 'controller': } __call__ /usr/lib/python2.7/dist-packages/routes/middleware.py:103 2013-10-21 20:04:51.724 DEBUG nova.api.openstack.wsgi [req-23bdfb52-dae5-42f7-a5b7-17c319ed67ff ServerMetadataTestJSON-tempest-1354548727-user ServerMetadataTestJSON-tempest-1354548727-tenant] Empty body provided in request get_body /opt/stack/new/nova/nova/api/openstack/wsgi.py:839 2013-10-21 20:04:51.724 DEBUG nova.api.openstack.wsgi [req-23bdfb52-dae5-42f7-a5b7-17c319ed67ff ServerMetadataTestJSON-tempest-1354548727-user ServerMetadataTestJSON-tempest-1354548727-tenant] Calling method > _process_stack /opt/stack/new/nova/nova/api/openstack/wsgi.py:962 2013-10-21 20:04:51.725 ERROR nova.api.openstack.wsgi [req-23bdfb52-dae5-42f7-a5b7-17c319ed67ff ServerMetadataTestJSON-tempest-1354548727-user ServerMetadataTestJSON-tempest-1354548727-tenant] Exception handling resource: update_all() takes exactly 4 arguments (3 given) 2013-10-21 20:04:51.725 20923 TRACE nova.api.openstack.wsgi Traceback (most recent call last): 2013-10-21 20:04:51.725 20923 TRACE nova.api.openstack.wsgi File "/opt/stack/new/nova/nova/api/openstack/wsgi.py", line 997, in _process_stack 2013-10-21 20:04:51.725 20923 TRACE nova.api.openstack.wsgi action_result = self.dispatch(meth, request, action_args) 2013-10-21 20:04:51.725 20923 TRACE nova.api.openstack.wsgi File "/opt/stack/new/nova/nova/api/openstack/wsgi.py", line 1078, in dispatch 2013-10-21 20:04:51.725 20923 TRACE nova.api.openstack.wsgi return method(req=request, **action_args) 2013-10-21 20:04:51.725 20923 TRACE nova.api.openstack.wsgi TypeError: update_all() takes exactly 4 arguments (3 given) 2013-10-21 20:04:51.725 20923 TRACE nova.api.openstack.wsgi 2013-10-21 20:04:51.726 DEBUG nova.api.openstack.wsgi [req-23bdfb52-dae5-42f7-a5b7-17c319ed67ff ServerMetadataTestJSON-tempest-1354548727-user ServerMetadataTestJSON-tempest-1354548727-tenant] Returning 400 to user: The server could not comply with the request since it is either malformed or otherwise incorrect. __call__ /opt/stack/new/nova/nova/api/openstack/wsgi.py:1224 To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1242916/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260423] [NEW] Email shouldn't be a mandatory attribute
Public bug reported: When using a LDAP backend, it's possible that a user won't have the "email" attribute defined, however it should still be possible to edit the other fields. Steps to reproduce (in an environment with keystone using a LDAP backend): 1. Log in as admin 2. Go to the Users dashboard 3. Select a user that doesn't have an email defined Expected result: 4. "Edit user" modal opens Actual result: 4. Error 500 Traceback: File "/usr/lib/python2.7/site-packages/django/views/generic/edit.py" in get 154. form = self.get_form(form_class) File "/opt/stack/horizon/openstack_dashboard/wsgi/../../horizon/forms/views.py" in get_form 82. return form_class(self.request, **self.get_form_kwargs()) File "/usr/lib/python2.7/site-packages/django/views/generic/edit.py" in get_form_kwargs 41. kwargs = {'initial': self.get_initial()} File "/opt/stack/horizon/openstack_dashboard/wsgi/../../openstack_dashboard/dashboards/admin/users/views.py" in get_initial 103. 'email': user.email} File "/opt/stack/python-keystoneclient/keystoneclient/base.py" in __getattr__ 425. raise AttributeError(k) Exception Type: AttributeError at /admin/users/e005aa43475b403c8babdff86ea27c37/update/ Exception Value: email ** Affects: horizon Importance: Medium Assignee: Julie Pichon (jpichon) Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1260423 Title: Email shouldn't be a mandatory attribute Status in OpenStack Dashboard (Horizon): New Bug description: When using a LDAP backend, it's possible that a user won't have the "email" attribute defined, however it should still be possible to edit the other fields. Steps to reproduce (in an environment with keystone using a LDAP backend): 1. Log in as admin 2. Go to the Users dashboard 3. Select a user that doesn't have an email defined Expected result: 4. "Edit user" modal opens Actual result: 4. Error 500 Traceback: File "/usr/lib/python2.7/site-packages/django/views/generic/edit.py" in get 154. form = self.get_form(form_class) File "/opt/stack/horizon/openstack_dashboard/wsgi/../../horizon/forms/views.py" in get_form 82. return form_class(self.request, **self.get_form_kwargs()) File "/usr/lib/python2.7/site-packages/django/views/generic/edit.py" in get_form_kwargs 41. kwargs = {'initial': self.get_initial()} File "/opt/stack/horizon/openstack_dashboard/wsgi/../../openstack_dashboard/dashboards/admin/users/views.py" in get_initial 103. 'email': user.email} File "/opt/stack/python-keystoneclient/keystoneclient/base.py" in __getattr__ 425. raise AttributeError(k) Exception Type: AttributeError at /admin/users/e005aa43475b403c8babdff86ea27c37/update/ Exception Value: email To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1260423/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1211338] Re: "Direct" vs. "direct" in impl_qpid
nova grizzly patch: https://review.openstack.org/#/c/61831/ ** Also affects: nova Importance: Undecided Status: New ** Changed in: nova Status: New => Fix Released ** Also affects: nova/grizzly Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1211338 Title: "Direct" vs. "direct" in impl_qpid Status in OpenStack Compute (Nova): Fix Released Status in OpenStack Compute (nova) grizzly series: New Status in Oslo - a Library of Common OpenStack Code: Fix Released Bug description: impl_qpid.py has {"type": "Direct"} (with a capital D) in one place. "direct" (lowercase) in others. It appears that qpid is case- sensitive about exchange types, so the version with the capital D is invalid. This ends up causing qpid to throw an error like: >> "/usr/lib/python2.6/site-packages/qpid/messaging/endpoints.py", >> line 567, in _ewait\nself.check_error()\n', ' File >> "/usr/lib/python2.6/site-packages/qpid/messaging/endpoints.py", >> line 556, in check_error\nraise self.error\n', 'NotFound: >> not-found: Exchange type not implemented: Direct >> (qpid/broker/SessionAdapter.cpp:117)(404)\n'] It should be a one-character fix. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1211338/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260423] Re: Email shouldn't be a mandatory attribute
** Also affects: horizon/havana Importance: Undecided Status: New ** Changed in: horizon/havana Importance: Undecided => Medium ** Changed in: horizon/havana Assignee: (unassigned) => Julie Pichon (jpichon) -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1260423 Title: Email shouldn't be a mandatory attribute Status in OpenStack Dashboard (Horizon): In Progress Status in OpenStack Dashboard (Horizon) havana series: New Bug description: When using a LDAP backend, it's possible that a user won't have the "email" attribute defined, however it should still be possible to edit the other fields. Steps to reproduce (in an environment with keystone using a LDAP backend): 1. Log in as admin 2. Go to the Users dashboard 3. Select a user that doesn't have an email defined Expected result: 4. "Edit user" modal opens Actual result: 4. Error 500 Traceback: File "/usr/lib/python2.7/site-packages/django/views/generic/edit.py" in get 154. form = self.get_form(form_class) File "/opt/stack/horizon/openstack_dashboard/wsgi/../../horizon/forms/views.py" in get_form 82. return form_class(self.request, **self.get_form_kwargs()) File "/usr/lib/python2.7/site-packages/django/views/generic/edit.py" in get_form_kwargs 41. kwargs = {'initial': self.get_initial()} File "/opt/stack/horizon/openstack_dashboard/wsgi/../../openstack_dashboard/dashboards/admin/users/views.py" in get_initial 103. 'email': user.email} File "/opt/stack/python-keystoneclient/keystoneclient/base.py" in __getattr__ 425. raise AttributeError(k) Exception Type: AttributeError at /admin/users/e005aa43475b403c8babdff86ea27c37/update/ Exception Value: email To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1260423/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260432] [NEW] nova-compute can't be setting up during install on trusty
Public bug reported: 1, during install: Setting up nova-compute (1:2014.1~b1-0ubuntu2) ... start: Job failed to start invoke-rc.d: initscript nova-compute, action "start" failed. dpkg: error processing nova-compute (--configure): subprocess installed post-installation script returned error exit status 1 Setting up nova-compute-kvm (1:2014.1~b1-0ubuntu2) ... Errors were encountered while processing: nova-compute E: Sub-process /usr/bin/dpkg returned an error code (1) 2, the system is latest trusty: ming@arm64:~$ sudo apt-get dist-upgrade Reading package lists... Done Building dependency tree Reading state information... Done Calculating upgrade... Done The following packages were automatically installed and are no longer required: dnsmasq-utils iputils-arping libboost-system1.53.0 libboost-thread1.53.0 libclass-isa-perl libopts25 libswitch-perl ttf-dejavu-core Use 'apt-get autoremove' to remove them. The following packages have been kept back: checkbox-cli 0 upgraded, 0 newly installed, 0 to remove and 1 not upgraded. 3, looks /usr/bin/nova-compute can't be started: ming@arm64:~$ nova-compute 2013-12-12 17:57:19.992 13823 ERROR stevedore.extension [-] Could not load 'file': (WebOb 1.3 (/usr/lib/python2.7/dist-packages), Requirement.parse('WebOb>=1.2.3,<1.3')) 2013-12-12 17:57:19.993 13823 ERROR stevedore.extension [-] (WebOb 1.3 (/usr/lib/python2.7/dist-packages), Requirement.parse('WebOb>=1.2.3,<1.3')) 2013-12-12 17:57:19.993 13823 TRACE stevedore.extension Traceback (most recent call last): 2013-12-12 17:57:19.993 13823 TRACE stevedore.extension File "/usr/lib/python2.7/dist-packages/stevedore/extension.py", line 134, in _load_plugins 2013-12-12 17:57:19.993 13823 TRACE stevedore.extension invoke_kwds, 2013-12-12 17:57:19.993 13823 TRACE stevedore.extension File "/usr/lib/python2.7/dist-packages/stevedore/extension.py", line 146, in _load_one_plugin 2013-12-12 17:57:19.993 13823 TRACE stevedore.extension plugin = ep.load() 2013-12-12 17:57:19.993 13823 TRACE stevedore.extension File "/usr/lib/python2.7/dist-packages/pkg_resources.py", line 2107, in load 2013-12-12 17:57:19.993 13823 TRACE stevedore.extension if require: self.require(env, installer) 2013-12-12 17:57:19.993 13823 TRACE stevedore.extension File "/usr/lib/python2.7/dist-packages/pkg_resources.py", line 2120, in require 2013-12-12 17:57:19.993 13823 TRACE stevedore.extension working_set.resolve(self.dist.requires(self.extras),env,installer))) 2013-12-12 17:57:19.993 13823 TRACE stevedore.extension File "/usr/lib/python2.7/dist-packages/pkg_resources.py", line 580, in resolve 2013-12-12 17:57:19.993 13823 TRACE stevedore.extension raise VersionConflict(dist,req) # XXX put more info here 2013-12-12 17:57:19.993 13823 TRACE stevedore.extension VersionConflict: (WebOb 1.3 (/usr/lib/python2.7/dist-packages), Requirement.parse('WebOb>=1.2.3,<1.3')) 2013-12-12 17:57:19.993 13823 TRACE stevedore.extension 2013-12-12 17:57:20.133 13823 ERROR nova.virt.driver [-] Compute driver option required, but not specified ** Affects: nova Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1260432 Title: nova-compute can't be setting up during install on trusty Status in OpenStack Compute (Nova): New Bug description: 1, during install: Setting up nova-compute (1:2014.1~b1-0ubuntu2) ... start: Job failed to start invoke-rc.d: initscript nova-compute, action "start" failed. dpkg: error processing nova-compute (--configure): subprocess installed post-installation script returned error exit status 1 Setting up nova-compute-kvm (1:2014.1~b1-0ubuntu2) ... Errors were encountered while processing: nova-compute E: Sub-process /usr/bin/dpkg returned an error code (1) 2, the system is latest trusty: ming@arm64:~$ sudo apt-get dist-upgrade Reading package lists... Done Building dependency tree Reading state information... Done Calculating upgrade... Done The following packages were automatically installed and are no longer required: dnsmasq-utils iputils-arping libboost-system1.53.0 libboost-thread1.53.0 libclass-isa-perl libopts25 libswitch-perl ttf-dejavu-core Use 'apt-get autoremove' to remove them. The following packages have been kept back: checkbox-cli 0 upgraded, 0 newly installed, 0 to remove and 1 not upgraded. 3, looks /usr/bin/nova-compute can't be started: ming@arm64:~$ nova-compute 2013-12-12 17:57:19.992 13823 ERROR stevedore.extension [-] Could not load 'file': (WebOb 1.3 (/usr/lib/python2.7/dist-packages), Requirement.parse('WebOb>=1.2.3,<1.3')) 2013-12-12 17:57:19.993 13823 ERROR stevedore.extension [-] (WebOb 1.3 (/usr/lib/python2.7/dist-packages), Requirement.parse('WebOb>=1.2.3,<1.3')) 2013-12-12 17:57:19.993 13823 TRACE stevedore.ext
[Yahoo-eng-team] [Bug 1260435] [NEW] Edit N1K network profile
Public bug reported: Fix issue with performing an N1K network profile update operation by ensuring only the editable fields are edited. Also with respect to new additions in the neutron N1K plugin, ensure new fields are updated in the update section. ** Affects: horizon Importance: Undecided Assignee: Abishek Subramanian (absubram) Status: In Progress ** Changed in: horizon Status: New => In Progress ** Changed in: horizon Assignee: (unassigned) => Abishek Subramanian (absubram) -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1260435 Title: Edit N1K network profile Status in OpenStack Dashboard (Horizon): In Progress Bug description: Fix issue with performing an N1K network profile update operation by ensuring only the editable fields are edited. Also with respect to new additions in the neutron N1K plugin, ensure new fields are updated in the update section. To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1260435/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260436] [NEW] Multi-nic support with N1K plugin
Public bug reported: When the cisco N1K neutron plugin is being used, an instance cannot be launched via Horizon with ability to have multiple nics. Only the first network is used for all created nics. This bug addressed that issue. ** Affects: horizon Importance: Undecided Assignee: Abishek Subramanian (absubram) Status: In Progress ** Changed in: horizon Status: New => In Progress ** Changed in: horizon Assignee: (unassigned) => Abishek Subramanian (absubram) -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1260436 Title: Multi-nic support with N1K plugin Status in OpenStack Dashboard (Horizon): In Progress Bug description: When the cisco N1K neutron plugin is being used, an instance cannot be launched via Horizon with ability to have multiple nics. Only the first network is used for all created nics. This bug addressed that issue. To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1260436/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260437] [NEW] Project name display when editing N1K profile
Public bug reported: When editing an N1K network profile, the project already associated with the network profile is not displayed. ** Affects: horizon Importance: Undecided Assignee: Abishek Subramanian (absubram) Status: In Progress ** Changed in: horizon Assignee: (unassigned) => Abishek Subramanian (absubram) ** Changed in: horizon Status: New => In Progress -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1260437 Title: Project name display when editing N1K profile Status in OpenStack Dashboard (Horizon): In Progress Bug description: When editing an N1K network profile, the project already associated with the network profile is not displayed. To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1260437/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260439] [NEW] Error message when creating a user with the default member role
Public bug reported: When you create a new user with the same role, from the roles list, as the default member role set in keystone.conf, Horizon shows the error message "Unable to add user to primary project". Even with this message, the user is created and the member role is granted to the user in the selected project. If you look the keystone.log file, you will see the following warning: " WARNING keystone.common.wsgi [-] Conflict occurred attempting to store role grant. User already has role in tenant " Example: I have the "member_role_name = _member_" entry in my keystone.conf file. When I create a User called "Test_User" in the project "Test_Project" with the role "_member_", the error message is shown. ** Affects: horizon Importance: Undecided Status: New ** Tags: error horizon message role -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1260439 Title: Error message when creating a user with the default member role Status in OpenStack Dashboard (Horizon): New Bug description: When you create a new user with the same role, from the roles list, as the default member role set in keystone.conf, Horizon shows the error message "Unable to add user to primary project". Even with this message, the user is created and the member role is granted to the user in the selected project. If you look the keystone.log file, you will see the following warning: " WARNING keystone.common.wsgi [-] Conflict occurred attempting to store role grant. User already has role in tenant " Example: I have the "member_role_name = _member_" entry in my keystone.conf file. When I create a User called "Test_User" in the project "Test_Project" with the role "_member_", the error message is shown. To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1260439/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260438] [NEW] Display N1K network profile information in network
Public bug reported: When an N1K profile is associated with a network, currently the N1K profile information is not displayed in the network detail. ** Affects: horizon Importance: Undecided Assignee: Abishek Subramanian (absubram) Status: In Progress ** Changed in: horizon Status: New => In Progress ** Changed in: horizon Assignee: (unassigned) => Abishek Subramanian (absubram) -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1260438 Title: Display N1K network profile information in network Status in OpenStack Dashboard (Horizon): In Progress Bug description: When an N1K profile is associated with a network, currently the N1K profile information is not displayed in the network detail. To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1260438/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260440] [NEW] nova-compute host is added to scheduling pool before Neutron can bind network ports on said host
Public bug reported: This is a race condition. Given a cloud with 0 compute nodes available, on a compute node: * Start up neutron-openvswitch-agent * Start up nova-compute * nova boot an instance Scenario 1: * neutron-openvswitch-agent registers with Neutron before nova tries to boot instance * port is bound to agent * instance boots with correct networking Scenario 2: * nova schedules instance to host before neutron-openvswitch-agent is registered with Neutron * nova instance fails with vif_type=binding_failed * instance is in ERROR state I would expect that Nova would not try to schedule instances on compute hosts that are not ready. Please also see this mailing list thread for more info: http://lists.openstack.org/pipermail/openstack- dev/2013-December/022084.html ** Affects: nova Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1260440 Title: nova-compute host is added to scheduling pool before Neutron can bind network ports on said host Status in OpenStack Compute (Nova): New Bug description: This is a race condition. Given a cloud with 0 compute nodes available, on a compute node: * Start up neutron-openvswitch-agent * Start up nova-compute * nova boot an instance Scenario 1: * neutron-openvswitch-agent registers with Neutron before nova tries to boot instance * port is bound to agent * instance boots with correct networking Scenario 2: * nova schedules instance to host before neutron-openvswitch-agent is registered with Neutron * nova instance fails with vif_type=binding_failed * instance is in ERROR state I would expect that Nova would not try to schedule instances on compute hosts that are not ready. Please also see this mailing list thread for more info: http://lists.openstack.org/pipermail/openstack- dev/2013-December/022084.html To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1260440/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1218279] Re: setUpClass (tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestJSON) Failed at server creation
I actually think this was the enable/disable server race. I'm going to close this for now. Reopen if an issue in the future ** Changed in: nova Status: New => Fix Released ** Changed in: tempest Status: New => Fix Released -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1218279 Title: setUpClass (tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestJSON) Failed at server creation Status in OpenStack Compute (Nova): Fix Released Status in Tempest: Fix Released Bug description: http://logs.openstack.org/78/44178/1/gate/gate-tempest-devstack-vm- postgres-full/042f00e/ 2013-08-29 08:51:47.396 | FAIL: setUpClass (tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestJSON) 2013-08-29 08:51:47.397 | setUpClass (tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestJSON) 2013-08-29 08:51:47.397 | -- 2013-08-29 08:51:47.397 | _StringException: Traceback (most recent call last): 2013-08-29 08:51:47.397 | File "tempest/api/compute/images/test_images_oneserver.py", line 50, in setUpClass 2013-08-29 08:51:47.397 | cls.tearDownClass() 2013-08-29 08:51:47.397 | File "tempest/api/compute/base.py", line 114, in tearDownClass 2013-08-29 08:51:47.398 | super(BaseComputeTest, cls).tearDownClass() 2013-08-29 08:51:47.398 | File "tempest/test.py", line 144, in tearDownClass 2013-08-29 08:51:47.398 | at_exit_set.remove(cls) 2013-08-29 08:51:47.398 | KeyError: The server creation was the actually failed. The setUpClass attempts to call the tearDownClass on error and throws a different exception, the correct exception throwing is tempest side issue, but the root cause probably not. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1218279/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260454] [NEW] Add cinder 'extend' volume functionality
Public bug reported: Cinder now has the ability to 'extend' (ie grow/expand/resize up) a volume. This functionality should be exposed through Horizon. ** Affects: horizon Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1260454 Title: Add cinder 'extend' volume functionality Status in OpenStack Dashboard (Horizon): New Bug description: Cinder now has the ability to 'extend' (ie grow/expand/resize up) a volume. This functionality should be exposed through Horizon. To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1260454/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1181567] Re: tempest: test_create_server / wait_for_server_status timeout
Not actually a tempest bug, this is a race in Nova ** Changed in: tempest Status: Incomplete => Invalid ** Summary changed: - tempest: test_create_server / wait_for_server_status timeout + Nova compute guest still stuck in BUILD state after 400s -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1181567 Title: Nova compute guest still stuck in BUILD state after 400s Status in OpenStack Compute (Nova): New Status in Tempest: Invalid Bug description: Failure occurred on https://review.openstack.org/#/c/29591/2 http://logs.openstack.org/29591/2/gate/gate-tempest-devstack-vm- quantum/23189/console.html.gz 2013-05-17 22:54:02.079 | == 2013-05-17 22:54:02.079 | ERROR: test suite for 2013-05-17 22:54:02.079 | -- 2013-05-17 22:54:02.079 | Traceback (most recent call last): 2013-05-17 22:54:02.079 | File "/usr/lib/python2.7/dist-packages/nose/suite.py", line 208, in run 2013-05-17 22:54:02.080 | self.setUp() 2013-05-17 22:54:02.080 | File "/usr/lib/python2.7/dist-packages/nose/suite.py", line 291, in setUp 2013-05-17 22:54:02.080 | self.setupContext(ancestor) 2013-05-17 22:54:02.080 | File "/usr/lib/python2.7/dist-packages/nose/suite.py", line 314, in setupContext 2013-05-17 22:54:02.080 | try_run(context, names) 2013-05-17 22:54:02.080 | File "/usr/lib/python2.7/dist-packages/nose/util.py", line 478, in try_run 2013-05-17 22:54:02.080 | return func() 2013-05-17 22:54:02.080 | File "/opt/stack/new/tempest/tempest/tests/compute/servers/test_create_server.py", line 57, in setUpClass 2013-05-17 22:54:02.080 | cls.client.wait_for_server_status(cls.server_initial['id'], 'ACTIVE') 2013-05-17 22:54:02.081 | File "/opt/stack/new/tempest/tempest/services/compute/xml/servers_client.py", line 311, in wait_for_server_status 2013-05-17 22:54:02.081 | raise exceptions.TimeoutException(message) 2013-05-17 22:54:02.081 | TimeoutException: Request timed out 2013-05-17 22:54:02.081 | Details: Request timed out 2013-05-17 22:54:02.081 | Details: Server 87c1dc14-44b1-406f-a7a0-c41876dc9111 failed to reach ACTIVE status within the required time (400 s). Current status: BUILD. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1181567/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1242898] Re: tearDownClass (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML): tearDownClass does not call the super's tearDownClass
if you look into http://logs.openstack.org/48/59948/4/check/check- tempest-dsvm-neutron-pg-isolated/dab4997/logs/screen-q-lbaas.txt.gz will see some tracebacks, it's probably an error in neutron. Note that I get it from http://logs.openstack.org/48/59948/4/check /check-tempest-dsvm-neutron-pg-isolated/dab4997/ not the original trace which is already deleted now. ** Also affects: neutron Importance: Undecided Status: New ** Changed in: tempest Status: New => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1242898 Title: tearDownClass (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML): tearDownClass does not call the super's tearDownClass Status in OpenStack Neutron (virtual network service): New Status in Tempest: Invalid Bug description: From http://logs.openstack.org/32/47432/16/check/check-tempest- devstack-vm-neutron-pg-isolated/c2a0dd3/console.html ... tearDownClass (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML) ... FAIL ... 2013-10-21 19:17:53.068 | Error in atexit._run_exitfuncs: 2013-10-21 19:17:53.068 | Traceback (most recent call last): 2013-10-21 19:17:53.068 | File "/usr/lib/python2.7/atexit.py", line 24, in _run_exitfuncs 2013-10-21 19:17:53.069 | func(*targs, **kargs) 2013-10-21 19:17:53.069 | File "tempest/test.py", line 167, in validate_tearDownClass 2013-10-21 19:17:53.069 | + str(at_exit_set)) 2013-10-21 19:17:53.069 | RuntimeError: tearDownClass does not calls the super's tearDownClass in these classes: set([]) 2013-10-21 19:17:53.070 | Error in sys.exitfunc: 2013-10-21 19:17:53.221 | 2013-10-21 19:17:53.221 | process-returncode 2013-10-21 19:17:53.221 | process-returncode ... FAIL 2013-10-21 19:17:53.614 | 2013-10-21 19:17:53.614 | == 2013-10-21 19:17:53.614 | FAIL: tearDownClass (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML) 2013-10-21 19:17:53.614 | tearDownClass (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML) 2013-10-21 19:17:53.614 | -- 2013-10-21 19:17:53.614 | _StringException: Traceback (most recent call last): 2013-10-21 19:17:53.615 | File "tempest/api/compute/servers/test_server_rescue.py", line 95, in tearDownClass 2013-10-21 19:17:53.615 | super(ServerRescueTestJSON, cls).tearDownClass() 2013-10-21 19:17:53.615 | File "tempest/api/compute/base.py", line 132, in tearDownClass 2013-10-21 19:17:53.615 | cls.isolated_creds.clear_isolated_creds() 2013-10-21 19:17:53.615 | File "tempest/common/isolated_creds.py", line 453, in clear_isolated_creds 2013-10-21 19:17:53.615 | self._clear_isolated_net_resources() 2013-10-21 19:17:53.615 | File "tempest/common/isolated_creds.py", line 445, in _clear_isolated_net_resources 2013-10-21 19:17:53.616 | self._clear_isolated_network(network['id'], network['name']) 2013-10-21 19:17:53.616 | File "tempest/common/isolated_creds.py", line 399, in _clear_isolated_network 2013-10-21 19:17:53.616 | net_client.delete_network(network_id) 2013-10-21 19:17:53.616 | File "tempest/services/network/json/network_client.py", line 76, in delete_network 2013-10-21 19:17:53.616 | resp, body = self.delete(uri, self.headers) 2013-10-21 19:17:53.616 | File "tempest/common/rest_client.py", line 308, in delete 2013-10-21 19:17:53.617 | return self.request('DELETE', url, headers) 2013-10-21 19:17:53.617 | File "tempest/common/rest_client.py", line 436, in request 2013-10-21 19:17:53.617 | resp, resp_body) 2013-10-21 19:17:53.617 | File "tempest/common/rest_client.py", line 522, in _error_checker 2013-10-21 19:17:53.617 | raise exceptions.ServerFault(message) 2013-10-21 19:17:53.617 | ServerFault: Got server fault 2013-10-21 19:17:53.617 | Details: {"NeutronError": "Request Failed: internal server error while processing your request."} 2013-10-21 19:17:53.618 | 2013-10-21 19:17:53.618 | 2013-10-21 19:17:53.618 | == 2013-10-21 19:17:53.618 | FAIL: process-returncode 2013-10-21 19:17:53.619 | process-returncode 2013-10-21 19:17:53.619 | -- 2013-10-21 19:17:53.619 | _StringException: Binary content: 2013-10-21 19:17:53.619 | traceback (test/plain; charset="utf8") 2013-10-21 19:17:53.619 | 2013-10-21 19:17:53.619 | 2013-10-21 19:17:53.620 | -- 2013-10-21 19:17:53.620 | Ran 237 tests in 914.828s 2013-10-21 19:17:53.637 | 2013-10-21 19:17:53.638 | FAILED (failures=2, skipped=8) To manage notifications
[Yahoo-eng-team] [Bug 1258379] Re: vpnservice's router must have gateway interface set
** No longer affects: tempest -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1258379 Title: vpnservice's router must have gateway interface set Status in OpenStack Neutron (virtual network service): In Progress Bug description: at line https://github.com/openstack/neutron/blob/master/neutron/services/vpn/service_drivers/ipsec.py#L172 it is obvious the router must have gateway interface set then it can be used as vpnservce router. To manage notifications about this bug go to: https://bugs.launchpad.net/neutron/+bug/1258379/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1218190] Re: Use assertEqual instead of assertEquals in unitttest
** Also affects: heat Importance: Undecided Status: New ** Changed in: heat Milestone: None => icehouse-2 ** Changed in: heat Assignee: (unassigned) => Jeff Peeler (jpeeler-z) ** Changed in: heat Status: New => In Progress -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to Keystone. https://bugs.launchpad.net/bugs/1218190 Title: Use assertEqual instead of assertEquals in unitttest Status in Orchestration API (Heat): In Progress Status in OpenStack Identity (Keystone): Fix Released Status in Python client library for Keystone: Fix Committed Status in Python client library for Neutron: Fix Committed Bug description: I noticed that [keystone, python-keystoneclient, python-neutronclient] configure tox.ini with py33 test, however, assertEquals is deprecated in py3 but ok with py2, so i think it is better to change all of assertEquals to assertEqual To manage notifications about this bug go to: https://bugs.launchpad.net/heat/+bug/1218190/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1256043] Re: Need to add Development environment files to ignore list
** No longer affects: tempest -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1256043 Title: Need to add Development environment files to ignore list Status in OpenStack Telemetry (Ceilometer): In Progress Status in OpenStack Dashboard (Horizon): Won't Fix Status in OpenStack Neutron (virtual network service): In Progress Status in Python client library for Ceilometer: In Progress Status in Python client library for Cinder: In Progress Status in Python client library for Glance: In Progress Status in Python client library for heat: Won't Fix Status in Python client library for Keystone: Fix Committed Status in Python client library for Neutron: In Progress Status in Python client library for Nova: In Progress Status in Python client library for Swift: Won't Fix Status in OpenStack Object Storage (Swift): Won't Fix Bug description: Following files generated by Eclipse development environment should be in ignore list to avoid their inclusion during a git push. .project .pydevproject To manage notifications about this bug go to: https://bugs.launchpad.net/ceilometer/+bug/1256043/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1193113] Re: DevicePathInUse exception in devstack-vm-quantum
** No longer affects: tempest -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1193113 Title: DevicePathInUse exception in devstack-vm-quantum Status in OpenStack Compute (Nova): Confirmed Bug description: I just got this during verification of one of my changes. I don't think it's related to the change (https://review.openstack.org/#/c/33478/) so I'm reporting it here before I reverify. Full log: http://logs.openstack.org/33478/1/gate/gate-tempest- devstack-vm-quantum/32609/logs/screen-n-cpu.txt.gz Also, this was for stable/grizzly. I'm not sure how to specify that in LP. 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/new/nova/nova/openstack/common/rpc/amqp.py", line 430, in _process_data 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp rval = self.proxy.dispatch(ctxt, version, method, **args) 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/new/nova/nova/openstack/common/rpc/dispatcher.py", line 133, in dispatch 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp return getattr(proxyobj, method)(ctxt, **kwargs) 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/new/nova/nova/exception.py", line 117, in wrapped 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp temp_level, payload) 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/contextlib.py", line 24, in __exit__ 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp self.gen.next() 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/new/nova/nova/exception.py", line 94, in wrapped 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp return f(self, context, *args, **kw) 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/new/nova/nova/compute/manager.py", line 209, in decorated_function 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp pass 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/contextlib.py", line 24, in __exit__ 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp self.gen.next() 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/new/nova/nova/compute/manager.py", line 195, in decorated_function 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp return function(self, context, *args, **kwargs) 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/new/nova/nova/compute/manager.py", line 237, in decorated_function 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp e, sys.exc_info()) 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/usr/lib/python2.7/contextlib.py", line 24, in __exit__ 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp self.gen.next() 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/new/nova/nova/compute/manager.py", line 224, in decorated_function 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp return function(self, context, *args, **kwargs) 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/new/nova/nova/compute/manager.py", line 2854, in reserve_block_device_name 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp return do_reserve() 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/new/nova/nova/openstack/common/lockutils.py", line 242, in inner 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp retval = f(*args, **kwargs) 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/new/nova/nova/compute/manager.py", line 2843, in do_reserve 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp context, instance, bdms, device) 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp File "/opt/stack/new/nova/nova/compute/utils.py", line 165, in get_device_name_for_instance 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp raise exception.DevicePathInUse(path=device) 2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp DevicePathInUse: The supplied device path (/dev/vdb) is in use. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1193113/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.n
[Yahoo-eng-team] [Bug 1260440] Re: nova-compute host is added to scheduling pool before Neutron can bind network ports on said host
This breaks deployment of new clouds in TripleO sometimes, and will likely break scaling too. Hence the Critical status. ** Also affects: neutron Importance: Undecided Status: New ** Also affects: tripleo Importance: Undecided Status: New ** Changed in: tripleo Status: New => Triaged ** Changed in: tripleo Importance: Undecided => Critical -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1260440 Title: nova-compute host is added to scheduling pool before Neutron can bind network ports on said host Status in OpenStack Neutron (virtual network service): New Status in OpenStack Compute (Nova): Confirmed Status in tripleo - openstack on openstack: Triaged Bug description: This is a race condition. Given a cloud with 0 compute nodes available, on a compute node: * Start up neutron-openvswitch-agent * Start up nova-compute * nova boot an instance Scenario 1: * neutron-openvswitch-agent registers with Neutron before nova tries to boot instance * port is bound to agent * instance boots with correct networking Scenario 2: * nova schedules instance to host before neutron-openvswitch-agent is registered with Neutron * nova instance fails with vif_type=binding_failed * instance is in ERROR state I would expect that Nova would not try to schedule instances on compute hosts that are not ready. Please also see this mailing list thread for more info: http://lists.openstack.org/pipermail/openstack- dev/2013-December/022084.html To manage notifications about this bug go to: https://bugs.launchpad.net/neutron/+bug/1260440/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1218391] Re: tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestXML.test_delete_image_that_is_not_yet_active spurious failure
there is only 1 hit in the last 2 weeks on this test, I actually think it's closed ** Changed in: nova Status: Confirmed => Fix Released ** No longer affects: nova ** Changed in: tempest Status: Confirmed => Fix Released ** No longer affects: tempest -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1218391 Title: tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestXML.test_delete_image_that_is_not_yet_active spurious failure Status in OpenStack Object Storage (Swift): Confirmed Bug description: Looks like this is an intermittent failure: ft45.7: tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestXML.test_delete_image_that_is_not_yet_active[gate,negative]_StringException: Empty attachments: stderr stdout Traceback (most recent call last): File "tempest/api/compute/images/test_images_oneserver.py", line 161, in test_delete_image_that_is_not_yet_active resp, body = self.client.create_image(self.server['id'], snapshot_name) File "tempest/services/compute/xml/images_client.py", line 105, in create_image str(Document(post_body)), self.headers) File "tempest/common/rest_client.py", line 260, in post return self.request('POST', url, headers, body) File "tempest/common/rest_client.py", line 388, in request resp, resp_body) File "tempest/common/rest_client.py", line 443, in _error_checker raise exceptions.Duplicate(resp_body) Duplicate: An object with that identifier already exists Details: {'message': "Cannot 'createImage' while instance is in task_state image_uploading", 'code': '409'} http://logs.openstack.org/50/41350/3/gate/gate-nova- python26/80bbbf0/testr_results.html.gz To manage notifications about this bug go to: https://bugs.launchpad.net/swift/+bug/1218391/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1248757] Re: test_snapshot_pattern fails with paramiko ssh EOFError
Removing tempest and adding nova, glance and neutron given what this test impacts: FAIL: tempest.scenario.test_snapshot_pattern.TestSnapshotPattern.test_snapshot_pattern[compute,image,network] ** Also affects: nova Importance: Undecided Status: New ** Also affects: glance Importance: Undecided Status: New ** Also affects: neutron Importance: Undecided Status: New ** No longer affects: tempest ** Tags added: gate-failure -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1248757 Title: test_snapshot_pattern fails with paramiko ssh EOFError Status in OpenStack Image Registry and Delivery Service (Glance): New Status in OpenStack Neutron (virtual network service): New Status in OpenStack Compute (Nova): New Bug description: I haven't seen this one reported yet (or seen it yet): http://logs.openstack.org/55/55455/1/check/check-tempest-devstack-vm- neutron/28d1ed7/console.html http://paste.openstack.org/show/50561/ To manage notifications about this bug go to: https://bugs.launchpad.net/glance/+bug/1248757/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1255627] Re: images.test_list_image_filters.ListImageFiltersTest fails with timeout
This non-white-listed error showed up in n-cpu: 2013-11-27 00:53:57.756 ERROR nova.virt.libvirt.driver [req- 298cf8f1-3907-4494-8b6e-61e9b88dfded ListImageFiltersTestXML- tempest-656023876-user ListImageFiltersTestXML-tempest-656023876-tenant] An error occurred while enabling hairpin mode on domain with xml: According to logstash this happened 9 times in the last two weeks. ** Changed in: nova Status: New => Confirmed ** Changed in: tempest Status: New => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1255627 Title: images.test_list_image_filters.ListImageFiltersTest fails with timeout Status in OpenStack Compute (Nova): Confirmed Status in Tempest: Invalid Bug description: Spurious failure in this test: http://logs.openstack.org/49/55749/8/check/check-tempest-devstack-vm- full/9bc94d5/console.html 2013-11-27 01:10:35.802 | == 2013-11-27 01:10:35.802 | FAIL: setUpClass (tempest.api.compute.images.test_list_image_filters.ListImageFiltersTestXML) 2013-11-27 01:10:35.803 | setUpClass (tempest.api.compute.images.test_list_image_filters.ListImageFiltersTestXML) 2013-11-27 01:10:35.803 | -- 2013-11-27 01:10:35.803 | _StringException: Traceback (most recent call last): 2013-11-27 01:10:35.804 | File "tempest/api/compute/images/test_list_image_filters.py", line 50, in setUpClass 2013-11-27 01:10:35.807 | cls.client.wait_for_image_status(cls.image1_id, 'ACTIVE') 2013-11-27 01:10:35.809 | File "tempest/services/compute/xml/images_client.py", line 153, in wait_for_image_status 2013-11-27 01:10:35.809 | raise exceptions.TimeoutException 2013-11-27 01:10:35.809 | TimeoutException: Request timed out To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1255627/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1257032] Re: nova makes calls to neutron with out considering URI size limit
*** This bug is a duplicate of bug 1228384 *** https://bugs.launchpad.net/bugs/1228384 Will close this bug as a dup of https://bugs.launchpad.net/nova/+bug/1228384 . As part of that fix, Phil has implemented chunking of server ids when querying neutron for ports of VMs. The above stack trace is for a quantum deployment, and Phil's patch will be ported over to that deployment. ** This bug has been marked a duplicate of bug 1228384 Security Group extension reads all Neutron ports for anything other that a single server -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1257032 Title: nova makes calls to neutron with out considering URI size limit Status in OpenStack Compute (Nova): New Bug description: Nova when requesting security group instance bindings for servers makes a call to Neutron. When there are many servers, the URI size grows beyond 8k with neutron throwing up 414 error message. We could easily hit this bug in our deployment as we have several VMs running. There is a similar bug while making net-list. It needed to make subnet list internally and that would result in 414 if there are several subnets that can make URI size too long. But subnet-list was internal call and fix was appropriate in the neutron client. Here the bug is while fetching sg instance bindings which is primary call. Hence I feel the bug must be fixed in neutron API consumer which is nova. Also, there must be a general framework for all APIs to not to exceed URI size limit or fix all calls with URIs which can extend beyond 8k size limit. Stacktrace for reference 2013-11-27 13:06:01.696 ERROR nova.api.openstack [req-020f17cb-ee43-4cd2-a270-767936e6546b 6abe780581924062bdb648375abcb378 b9bbb06d41a942248e8d7070e17ed89d] Caught error: 414-{'message': ''} 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack Traceback (most recent call last): 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack File "/usr/lib/python2.6/site-packages/nova/api/openstack/__init__.py", line 81, in __call__ 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack return req.get_response(self.application) 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack File "/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/request.py", line 1053, in get_response 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack application, catch_exc_info=False) 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack File "/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/request.py", line 1022, in call_application 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack app_iter = application(self.environ, start_response) 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack File "/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/dec.py", line 159, in __call__ 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack return resp(environ, start_response) 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack File "/usr/lib/python2.6/site-packages/keystoneclient/middleware/auth_token.py", line 450, in __call__ 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack return self.app(env, start_response) 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack File "/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/dec.py", line 159, in __call__ 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack return resp(environ, start_response) 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack File "/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/dec.py", line 159, in __call__ 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack return resp(environ, start_response) 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack File "/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/dec.py", line 159, in __call__ 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack return resp(environ, start_response) 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack File "/usr/lib/python2.6/site-packages/Routes-1.12.3-py2.6.egg/routes/middleware.py", line 131, in __call__ 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack response = self.app(environ, start_response) 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack File "/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/dec.py", line 159, in __call__ 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack return resp(environ, start_response) 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack File "/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/dec.py", line 147, in __call__ 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack resp = self.call_func(req, *args, **self.kwargs) 2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack
[Yahoo-eng-team] [Bug 1254772] Re: tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML setUpClass times-out on attaching volume
This shows up in n-cpu: The "model server went away" showed up 11 times in the last two weeks with the last one being on Dec. 3. This sample size is too small for me to close at this time. 2013-11-25 15:24:22.099 21076 ERROR nova.servicegroup.drivers.db [-] model server went away 2013-11-25 15:24:32.814 ERROR nova.compute.manager [req-ecacaa21-3f07-4b44-9896-8b5bd2238a19 ServersTestManualDisk-tempest-1962756300-user ServersTestManualDisk-tempest-1962756300-tenant] [instance: 1f872097-8ad8-44f8-ba03-89a14115efe0] Failed to deallocate network for instance. 2013-11-25 15:25:32.855 21076 ERROR root [-] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/new/nova/nova/compute/manager.py", line 1809, in _try_deallocate_network\nself._deallocate_network(context, instance, requested_networks)\n', ' File "/opt/stack/new/nova/nova/compute/manager.py", line 1491, in _deallocate_network\ncontext, instance, requested_networks=requested_networks)\n', ' File "/opt/stack/new/nova/nova/network/api.py", line 93, in wrapped\nreturn func(self, context, *args, **kwargs)\n', ' File "/opt/stack/new/nova/nova/network/api.py", line 318, in deallocate_for_instance\n self.network_rpcapi.deallocate_for_instance(context, **args)\n', ' File "/opt/stack/new/nova/nova/network/rpcapi.py", line 199, in deallocate_for_instance\nhost=host, requested_networks=requested_networks)\n', ' File "/opt/stack/new/nova/nova/rpcclient.py", line 85, in call\nreturn self._invoke(self.proxy.call, ctxt, method, ** kwargs)\n', ' File "/opt/stack/new/nova/nova/rpcclient.py", line 63, in _invoke\nreturn cast_or_call(ctxt, msg, **self.kwargs)\n', ' File "/opt/stack/new/nova/nova/openstack/common/rpc/proxy.py", line 130, in call\n exc.info, real_topic, msg.get(\'method\'))\n', 'Timeout: Timeout while waiting on RPC response - topic: "network", RPC method: "deallocate_for_instance" info: ""\n'] 2013-11-25 15:25:38.371 21076 ERROR nova.openstack.common.periodic_task [-] Error during ComputeManager.update_available_resource: Timeout while waiting on RPC response - topic: "conductor", RPC method: "compute_node_update" info: "" 2013-11-25 15:26:32.903 21076 ERROR root [-] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/new/nova/nova/compute/manager.py", line 1919, in _delete_instance\n self._shutdown_instance(context, db_inst, bdms)\n', ' File "/opt/stack/new/nova/nova/compute/manager.py", line 1854, in _shutdown_instance\nself._try_deallocate_network(context, instance, requested_networks)\n', ' File "/opt/stack/new/nova/nova/compute/manager.py", line 1814, in _try_deallocate_network\n self._set_instance_error_state(context, instance[\'uuid\'])\n', ' File "/opt/stack/new/nova/nova/compute/manager.py", line 484, in _set_instance_error_state\nvm_state=vm_states.ERROR)\n', ' File "/opt/stack/new/nova/nova/compute/manager.py", line 473, in _instance_update\n **kwargs)\n', ' File "/opt/stack/new/nova/nova/conductor/api.py", line 389, in instance_update\nupdates, \'conductor\')\n', ' File "/opt/stack/new/nova/nova/conductor/rpcapi.py", line 149, in instance_update\nservice=service)\n', ' File "/opt/stack/new/nova/nova/rpcclient.py", line 85, in call\nreturn self._invoke(self.proxy.call, ctxt, method, **kwargs)\n', ' File "/opt/stack/new/nova/nova/rpcclient.py", line 63, in _invoke\nreturn cast_or_call(ctxt, msg, **self.kwargs)\n', ' File "/opt/stack/new/nova/nova/openstack/common/rpc/proxy.py", line 130, in call\n exc.info, real_topic, msg.get(\'method\'))\n', 'Timeout: Timeout while waiting on RPC response - topic: "conductor", RPC method: "instance_update" info: ""\n'] 2013-11-25 15:26:32.933 21076 ERROR nova.servicegroup.drivers.db [-] Recovered model server connection! ** Also affects: nova Importance: Undecided Status: New ** Changed in: tempest Status: New => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1254772 Title: tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML setUpClass times-out on attaching volume Status in OpenStack Compute (Nova): New Status in Tempest: Invalid Bug description: 2013-11-25 15:42:45.769 | == 2013-11-25 15:42:45.770 | FAIL: setUpClass (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML) 2013-11-25 15:42:45.770 | setUpClass (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML) 2013-11-25 15:42:45.770 | -- 2013-11-25 15:42:45.770 | _StringException: Traceback (most recent call last): 2013-11-25 15:42:45.770 | File "tempest/api/compute/servers/test_server
[Yahoo-eng-team] [Bug 1260489] [NEW] --debug flag not working in neutron
Public bug reported: This is with the neutron master branch, in a single node devstack setup. The branch is at commit 3b4233873539bad62d202025529678a5b0add412. If I use the --debug flag in a neutron CLI, for example, port-list, I don't see any debug output: cloud@controllernode:/opt/stack/neutron$ neutron --debug port-list +--+--+---+-+ | id | name | mac_address | fixed_ips | +--+--+---+-+ | 6c26cdc1-acc1-439c-bb47-d343085b7b78 | | fa:16:3e:32:2c:eb | {"subnet_id": "37f15352-e816-4a03-b58c-b4d5c1fa8e2a", "ip_address": "10.0.0.2"} | | f09b14b2-3162-4212-9d91-f97b22c95f31 | | fa:16:3e:99:08:6b | {"subnet_id": "d4717b67-fd64-45ed-b22c-dedbd23afff3", "ip_address": "172.24.4.226"} | | f0ba4efd-12ca-4d56-8c7d-e879e4150a63 | | fa:16:3e:02:41:47 | {"subnet_id": "37f15352-e816-4a03-b58c-b4d5c1fa8e2a", "ip_address": "10.0.0.1"} | +--+--+---+-+ cloud@controllernode:/opt/stack/neutron$ On the other hand, if I use the --debug flag for nova, for example, nova list, I see the curl request and response showing up: cloud@controllernode:/opt/stack/neutron$ nova --debug list REQ: curl -i 'http://192.168.52.85:5000/v2.0/tokens' -X POST -H "Content-Type: application/json" -H "Accept: application/json" -H "User- Agent: python-novaclient" -d '{"auth": {"tenantName": "admin", "passwordCredentials": {"username": "admin", "password": "password"}}}' RESP: [200] CaseInsensitiveDict({'date': 'Thu, 05 Dec 2013 23:41:07 GMT', 'vary': 'X-Auth-Token', 'content-length': '8255', 'content-type': 'application/json'}) RESP BODY: {"access": {"token": {"issued_at": "2013-12-05T23:41:07.307915", "expires": "2013-12-06T23:41:07Z", "id": "MIIOkwYJKoZIhvcNAQcCoIIOhDCCDoACAQExCTAHBgUrDgMCGjCCDOkGCSqGSIb3DQEHAaCCDNoEggzWeyJhY2Nlc3MiOiB7InRva2VuIjogeyJpc3N1ZWRfYXQiOiAiMjAxMy0xMi0wNVQyMzo0MTowNy4zMDc5MTUiLCAiZXhwaXJlcyI6ICIyMDEzLTEyLTA2VDIzOjQxOjA3WiIsICJpZCI6ICJwbGFjZWhvbGRlciIsICJ0ZW5hbnQiOiB7ImRlc2NyaXB0aW9uIjogbnVsbCwgImVuYWJsZWQiOiB0cnVlLCAiaWQiOiAiYTdiMzk2MGI5NzkyNGJhYjlhNTVhOWY5ZjY4NGE4NzAiLCAibmFtZSI6ICJhZG1pbiJ9fSwgInNlcnZpY2VDYXRhbG9nIjogW3siZW5kcG9pbnRzIjogW3siYWRtaW5VUkwiOiAiaHR0cDovLzE5Mi4xNjguNTIuODU6ODc3NC92Mi9hN2IzOTYwYjk3OTI0YmFiOWE1NWE5ZjlmNjg0YTg3MCIsICJyZWdpb24iOiAiUmVnaW9uT25lIiwgImludGVybmFsVVJMIjogImh0dHA6Ly8xOTIuMTY4LjUyLjg1Ojg3NzQvdjIvYTdiMzk2MGI5NzkyNGJhYjlhNTVhOWY5ZjY4NGE4NzAiLCAiaWQiOiAiMDQyMzVjMmE1ODNlNDAwZDg1NTBkYTI0NmNiZDI1YWEiLCAicHVibGljVVJMIjogImh0dHA6Ly8xOTIuMTY4LjUyLjg1Ojg3NzQvdjIvYTdiMzk2MGI5NzkyNGJhYjlhNTVhOWY5ZjY4NGE4NzAifV0sICJlbmRwb2ludHNfbGlua3MiOiBbXSwgInR5cGUiOiAi Y29tcHV0ZSIsICJuYW1lIjogIm5vdmEifSwgeyJlbmRwb2ludHMiOiBbeyJhZG1pblVSTCI6ICJodHRwOi8vMTkyLjE2OC41Mi44NTo5Njk2LyIsICJyZWdpb24iOiAiUmVnaW9uT25lIiwgImludGVybmFsVVJMIjogImh0dHA6Ly8xOTIuMTY4LjUyLjg1Ojk2OTYvIiwgImlkIjogIjYyNWI1YzM3ZDJlYzQ4ZGRhMTRmZGZmZmMyZjBhMTY0IiwgInB1YmxpY1VSTCI6ICJodHRwOi8vMTkyLjE2OC41Mi44NTo5Njk2LyJ9XSwgImVuZHBvaW50c19saW5rcyI6IFtdLCAidHlwZSI6ICJuZXR3b3JrIiwgIm5hbWUiOiAibmV1dHJvbiJ9LCB7ImVuZHBvaW50cyI6IFt7ImFkbWluVVJMIjogImh0dHA6Ly8xOTIuMTY4LjUyLjg1Ojg3NzYvdjIvYTdiMzk2MGI5NzkyNGJhYjlhNTVhOWY5ZjY4NGE4NzAiLCAicmVnaW9uIjogIlJlZ2lvbk9uZSIsICJpbnRlcm5hbFVSTCI6ICJodHRwOi8vMTkyLjE2OC41Mi44NTo4Nzc2L3YyL2E3YjM5NjBiOTc5MjRiYWI5YTU1YTlmOWY2ODRhODcwIiwgImlkIjogIjNmODVjN2ZmZjNjMzRmNWNiMzlmMTZiMzQ2ZmY1Mjc0IiwgInB1YmxpY1VSTCI6ICJodHRwOi8vMTkyLjE2OC41Mi44NTo4Nzc2L3YyL2E3YjM5NjBiOTc5MjRiYWI5YTU1YTlmOWY2ODRhODcwIn1dLCAiZW5kcG9pbnRzX2xpbmtzIjogW10sICJ0eXBlIjogInZvbHVtZXYyIiwgIm5hbWUiOiAiY2luZGVyIn0sIHsiZW5kcG9pbnRzIjogW3siYWRtaW5VUkwiOiAiaHR0cDovLzE5Mi4xNjguNTIuODU6ODc3NC92MyIsICJyZWdpb 24iOiAiUmVnaW9uT25lIiwgImludGVybmFsVVJMIjogImh0dHA6Ly8xOTIuMTY4LjUyLjg1Ojg3NzQvdjMiLCAiaWQiOiAiYTM4NjBlZTM3MWEyNDIxNGFlYTBiODk5M2I1YTY0OTciLCAicHVibGljVVJMIjogImh0dHA6Ly8xOTIuMTY4LjUyLjg1Ojg3NzQvdjMifV0sICJlbmRwb2ludHNfbGlua3MiOiBbXSwgInR5cGUiOiAiY29tcHV0ZXYzIiwgIm5hbWUiOiAibm92YSJ9LCB7ImVuZHBvaW50cyI6IFt7ImFkbWluVVJMIjogImh0dHA6Ly8xOTIuMTY4LjUyLjg1OjMzMzMiLCAicmVnaW9uIjogIlJlZ2lvbk9uZSIsICJpbnRlcm5hbFVSTCI6ICJodHRwOi8vMTkyLjE2OC41Mi44NTozMzMzIiwgImlkIjogIjZmZTY2OTMwNjA5MTQwYWVhMTIwMTJjNWViMzViZGQ2IiwgInB1YmxpY1VSTCI6ICJodHRwOi8vMTkyLjE2OC41Mi44NTozMzMzIn1dLCAiZW5kcG9pbnRzX2xpbmtzIjogW10sICJ0eXBlIjogInMzIiwgIm5hbWUiOiAiczMifSwgeyJlbmRwb2ludHMiOiBbeyJhZG1pblVSTCI6ICJodHRwOi8vMTkyLjE2OC41Mi44NTo5MjkyIiwgInJlZ2lvbiI6ICJSZWdpb25PbmUiLCAiaW50ZXJuYWxVUkwiOiAiaHR0cDovLzE5Mi4xNjguNTIuODU6OTI5MiIsICJpZCI6ICIyMjVhMDc2ZmZiOWI0YmQxYTdmODE4N2M0NzY2M2I0NyIsICJwdWJsaWNVUkwiOiAiaHR0cDovLzE5Mi4xNjguNTIuODU6OTI5MiJ9XS
[Yahoo-eng-team] [Bug 1252947] Re: tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON fails sporadically
** Also affects: nova Importance: Undecided Status: New ** No longer affects: tempest -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1252947 Title: tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON fails sporadically Status in OpenStack Compute (Nova): New Bug description: tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON fails sporadically. See: http://logs.openstack.org/66/54966/2/check/check-tempest- devstack-vm-full/d611ed0/console.html 2013-11-19 22:24:52.379 | == 2013-11-19 22:24:52.380 | FAIL: setUpClass (tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON) 2013-11-19 22:24:52.380 | setUpClass (tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON) 2013-11-19 22:24:52.380 | -- 2013-11-19 22:24:52.380 | _StringException: Traceback (most recent call last): 2013-11-19 22:24:52.380 | File "tempest/api/compute/servers/test_servers_negative.py", line 46, in setUpClass 2013-11-19 22:24:52.380 | resp, server = cls.create_test_server(wait_until='ACTIVE') 2013-11-19 22:24:52.381 | File "tempest/api/compute/base.py", line 118, in create_test_server 2013-11-19 22:24:52.381 | server['id'], kwargs['wait_until']) 2013-11-19 22:24:52.381 | File "tempest/services/compute/json/servers_client.py", line 160, in wait_for_server_status 2013-11-19 22:24:52.381 | extra_timeout=extra_timeout) 2013-11-19 22:24:52.381 | File "tempest/common/waiters.py", line 73, in wait_for_server_status 2013-11-19 22:24:52.381 | raise exceptions.BuildErrorException(server_id=server_id) 2013-11-19 22:24:52.381 | BuildErrorException: Server 62bfeebd-8878-477f-9eac-a8b21ec5ac26 failed to build and is in ERROR status To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1252947/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1224518] Re: test_reboot_server_hard fails sporadically in swift check jobs
I don't think this is a tempest bug, this is a state transition bug in Nova ** Changed in: tempest Importance: Undecided => Low ** Changed in: tempest Status: New => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1224518 Title: test_reboot_server_hard fails sporadically in swift check jobs Status in OpenStack Compute (Nova): New Status in Tempest: Invalid Bug description: See: http://logs.openstack.org/46/46146/2/check/gate-tempest-devstack- vm-postgres-full/b2712f1/console.html 2013-09-12 04:43:17.625 | == 2013-09-12 04:43:17.649 | FAIL: tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_reboot_server_hard[gate,smoke] 2013-09-12 04:43:17.651 | tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_reboot_server_hard[gate,smoke] 2013-09-12 04:43:17.652 | -- 2013-09-12 04:43:17.652 | _StringException: Empty attachments: 2013-09-12 04:43:17.652 | stderr 2013-09-12 04:43:17.652 | stdout 2013-09-12 04:43:17.653 | 2013-09-12 04:43:17.653 | pythonlogging:'': {{{ 2013-09-12 04:43:17.653 | 2013-09-12 04:16:55,739 Request: GET http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf 2013-09-12 04:43:17.654 | 2013-09-12 04:16:55,806 Response Status: 200 2013-09-12 04:43:17.654 | 2013-09-12 04:16:55,806 Nova request id: req-cdc6b1fc-bcf2-4e9c-bea1-8bf935993cbd 2013-09-12 04:43:17.654 | 2013-09-12 04:16:55,807 Request: POST http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf/action 2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,917 Response Status: 202 2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,917 Nova request id: req-3af37dd3-0ddc-4daa-aa6f-6958a5073cc4 2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,918 Request: GET http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf 2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,986 Response Status: 200 2013-09-12 04:43:17.656 | 2013-09-12 04:16:55,986 Nova request id: req-a7298d3e-167c-4c8f-9506-6064ba811e5b . . . 2013-09-12 04:43:17.976 | 2013-09-12 04:23:35,773 Request: GET http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf 2013-09-12 04:43:17.976 | 2013-09-12 04:23:35,822 Response Status: 200 2013-09-12 04:43:17.976 | 2013-09-12 04:23:35,823 Nova request id: req-a122aded-b49b-4847-9920-b2b8b09bc0ca 2013-09-12 04:43:17.976 | }}} 2013-09-12 04:43:17.977 | 2013-09-12 04:43:17.977 | Traceback (most recent call last): 2013-09-12 04:43:17.978 | File "tempest/api/compute/servers/test_server_actions.py", line 81, in test_reboot_server_hard 2013-09-12 04:43:17.978 | self.client.wait_for_server_status(self.server_id, 'ACTIVE') 2013-09-12 04:43:17.979 | File "tempest/services/compute/json/servers_client.py", line 176, in wait_for_server_status 2013-09-12 04:43:17.979 | raise exceptions.TimeoutException(message) 2013-09-12 04:43:17.979 | TimeoutException: Request timed out 2013-09-12 04:43:17.980 | Details: Server 8ad0ad9a-3975-486f-94b4-af1c89b51aaf failed to reach ACTIVE status within the required time (400 s). Current status: HARD_REBOOT. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1224518/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1244762] Re: tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance fails sporadically
** Also affects: nova Importance: Undecided Status: New ** No longer affects: tempest -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1244762 Title: tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance fails sporadically Status in OpenStack Compute (Nova): New Bug description: See: http://logs.openstack.org/87/44787/16/check/check-tempest- devstack-vm-neutron/d2ede4d/console.html 2013-10-25 18:06:37.957 | == 2013-10-25 18:06:37.959 | FAIL: tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance[gate,smoke] 2013-10-25 18:06:37.959 | tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance[gate,smoke] 2013-10-25 18:06:37.959 | -- 2013-10-25 18:06:37.959 | _StringException: Empty attachments: 2013-10-25 18:06:37.959 | stderr 2013-10-25 18:06:37.960 | stdout 2013-10-25 18:06:37.960 | 2013-10-25 18:06:37.960 | pythonlogging:'': {{{ 2013-10-25 18:06:37.960 | 2013-10-25 17:59:08,821 state: pending 2013-10-25 18:06:37.960 | 2013-10-25 17:59:14,092 State transition "pending" ==> "error" 5 second 2013-10-25 18:06:37.961 | }}} 2013-10-25 18:06:37.961 | 2013-10-25 18:06:37.961 | Traceback (most recent call last): 2013-10-25 18:06:37.961 | File "tempest/thirdparty/boto/test_ec2_instance_run.py", line 150, in test_run_stop_terminate_instance 2013-10-25 18:06:37.961 | self.assertInstanceStateWait(instance, "running") 2013-10-25 18:06:37.961 | File "tempest/thirdparty/boto/test.py", line 356, in assertInstanceStateWait 2013-10-25 18:06:37.962 | state = self.waitInstanceState(lfunction, wait_for) 2013-10-25 18:06:37.962 | File "tempest/thirdparty/boto/test.py", line 341, in waitInstanceState 2013-10-25 18:06:37.962 | self.valid_instance_state) 2013-10-25 18:06:37.962 | File "tempest/thirdparty/boto/test.py", line 332, in state_wait_gone 2013-10-25 18:06:37.962 | self.assertIn(state, valid_set | self.gone_set) 2013-10-25 18:06:37.963 | File "/usr/local/lib/python2.7/dist-packages/testtools/testcase.py", line 328, in assertIn 2013-10-25 18:06:37.963 | self.assertThat(haystack, Contains(needle)) 2013-10-25 18:06:37.963 | File "/usr/local/lib/python2.7/dist-packages/testtools/testcase.py", line 417, in assertThat 2013-10-25 18:06:37.963 | raise MismatchError(matchee, matcher, mismatch, verbose) 2013-10-25 18:06:37.963 | MismatchError: u'error' not in set(['paused', 'terminated', 'running', 'stopped', 'pending', '_GONE', 'stopping', 'shutting-down']) To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1244762/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1258319] Re: test_reboot_server_hard fails sporadically in swift check jobs
*** This bug is a duplicate of bug 1224518 *** https://bugs.launchpad.net/bugs/1224518 ** This bug has been marked a duplicate of bug 1224518 test_reboot_server_hard fails sporadically in swift check jobs -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1258319 Title: test_reboot_server_hard fails sporadically in swift check jobs Status in OpenStack Compute (Nova): New Bug description: test_reboot_server_hard fails sporadically in swift check jobs I believe this has been reported before, but I was not able to find it. See: http://logs.openstack.org/43/60343/1/gate/gate-tempest-dsvm- full/c92d206/console.html 2013-12-05 21:29:18.174 | == 2013-12-05 21:29:18.183 | FAIL: tempest.api.compute.servers.test_server_actions.ServerActionsTestXML.test_reboot_server_hard[gate,smoke] 2013-12-05 21:29:18.186 | tempest.api.compute.servers.test_server_actions.ServerActionsTestXML.test_reboot_server_hard[gate,smoke] 2013-12-05 21:29:18.200 | -- 2013-12-05 21:29:18.206 | _StringException: Empty attachments: 2013-12-05 21:29:18.206 | stderr 2013-12-05 21:29:18.207 | stdout 2013-12-05 21:29:18.207 | 2013-12-05 21:29:18.207 | pythonlogging:'': {{{ . . . 2013-12-05 21:29:19.174 | Traceback (most recent call last): 2013-12-05 21:29:19.175 | File "tempest/api/compute/servers/test_server_actions.py", line 83, in test_reboot_server_hard 2013-12-05 21:29:19.175 | self.client.wait_for_server_status(self.server_id, 'ACTIVE') 2013-12-05 21:29:19.175 | File "tempest/services/compute/xml/servers_client.py", line 369, in wait_for_server_status 2013-12-05 21:29:19.175 | extra_timeout=extra_timeout) 2013-12-05 21:29:19.176 | File "tempest/common/waiters.py", line 82, in wait_for_server_status 2013-12-05 21:29:19.176 | raise exceptions.TimeoutException(message) 2013-12-05 21:29:19.176 | TimeoutException: Request timed out 2013-12-05 21:29:19.177 | Details: Server f313af9a-8ec1-4f77-b63f-76d9317d6423 failed to reach ACTIVE status within the required time (196 s). Current status: HARD_REBOOT. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1258319/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1250836] Re: Updating of instance metadata occasionally leads to a deadlock
not a tempest bug ** No longer affects: tempest -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1250836 Title: Updating of instance metadata occasionally leads to a deadlock Status in OpenStack Compute (Nova): New Bug description: During the tempest tests run I got the following error: 2013-11-13 10:09:18.814 ERROR nova.api.openstack [req-a3172f97-0d7c-4f8b-a7a5-bec6aad2b549 ServerMetadataTestJSON-tempest-1285971638-user ServerMetadataTestJSON-tempest-1285971638-tenant] Caught error: (OperationalError) (1213, 'Deadlock found when trying to get lock; try restarting transaction') 'INSERT INTO instance_metadata (created_at, updated_at, deleted_at, deleted, `key`, value, instance_uuid) VALUES (%s, %s, %s, %s, %s, %s, %s)' (datetime.datetime(2013, 11, 13, 10, 9, 18, 811419), None, None, 0, 'key3', 'value3', 'ba645a19-78c5-439d-9408-68f413c200f4') 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack Traceback (most recent call last): 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/opt/stack/new/nova/nova/api/openstack/__init__.py", line 119, in __call__ 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return req.get_response(self.application) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/usr/local/lib/python2.7/dist-packages/webob/request.py", line 1296, in send 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack application, catch_exc_info=False) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/usr/local/lib/python2.7/dist-packages/webob/request.py", line 1260, in call_application 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack app_iter = application(self.environ, start_response) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/usr/local/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__ 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return resp(environ, start_response) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/opt/stack/new/python-keystoneclient/keystoneclient/middleware/auth_token.py", line 571, in __call__ 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return self.app(env, start_response) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/usr/local/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__ 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return resp(environ, start_response) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/usr/local/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__ 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return resp(environ, start_response) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/usr/lib/python2.7/dist-packages/routes/middleware.py", line 131, in __call__ 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack response = self.app(environ, start_response) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/usr/local/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__ 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return resp(environ, start_response) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/usr/local/lib/python2.7/dist-packages/webob/dec.py", line 130, in __call__ 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack resp = self.call_func(req, *args, **self.kwargs) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/usr/local/lib/python2.7/dist-packages/webob/dec.py", line 195, in call_func 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return self.func(req, *args, **kwargs) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/opt/stack/new/nova/nova/api/openstack/wsgi.py", line 939, in __call__ 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack content_type, body, accept) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/opt/stack/new/nova/nova/api/openstack/wsgi.py", line 998, in _process_stack 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack action_result = self.dispatch(meth, request, action_args) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/opt/stack/new/nova/nova/api/openstack/wsgi.py", line 1079, in dispatch 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return method(req=request, **action_args) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/opt/stack/new/nova/nova/api/openstack/compute/server_metadata.py", line 67, in create 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack delete=False) 2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack File "/opt/stack/new/nova/nova/api/openstack/compute/server_metadata.py", line 120, in _update_instance_metadata 2013-11-13 10:09:18.814 22804 TRA
[Yahoo-eng-team] [Bug 1260516] [NEW] PortInUse exception leaves orphaned ports
Public bug reported: While bringing up 1000 VMs with Heat, some VMs are in error state. After deleting the VMs, not all ports are removed. The next time VMs are created, they fail due to no IP addresses left. Note, initially there are 7 ports, no VMs 804 VMs are created, 811 ports root@control01:/usr/share/pyshared/heat# nova list | grep ACT | wc -l 358 root@control01:/usr/share/pyshared/heat# neutron port-list | wc -l 811 root@control01:/usr/share/pyshared/heat# nova list | grep ERR | wc -l 270 root@control01:/usr/share/pyshared/heat# nova list | grep stack | wc -l 804 After deleting the VMs, 248 ports remain root@control01:/usr/share/pyshared/heat# nova list | grep stack | wc -l 0 root@control01:/usr/share/pyshared/heat# neutron port-list | wc -l 248 2013-12-12 20:55:35.320 20945 ERROR nova.scheduler.filter_scheduler [req-844a11 f6-66f1-4fc7-9f3b-8bacbe57a04d e95097acd0b041558f33d07f720c1bd7 354f17bf81924b278806c3e3798aa527] [instance: 2a3eadcc-5230-4ed6-bf9f-9a82d43b91c3] Error from last host: compute134 (node compute134.lab.cisco): [u'Traceback (most recent call last):\n', u' File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 1037, in _build_instance\nset_access_ip=set_access_ip)\n', u' File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 1410, in _sp awn\nLOG.exception(_(\'Instance failed to spawn\'), instance=instance)\n', u' File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 1407, in _spawn\nblock_device_info)\n', u' File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py", line 2063, in spawn\nadmin_pass=admin_password)\n', u' File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py" , line 2412, in _create_image\ncontent=files, extra_md=extra_md, network_ info=network_info)\n', u' File "/usr/lib/python2.7/dist-packages/nova/api/metadata/base.py", line 157, in __init__\ncfg = netutils.get_injected_network_t emplate(network_info)\n', u' File "/usr/lib/python2.7/dist-packages/nova/virt/ netutils.py", line 74, in get_injected_network_template\nif not (network_info and template):\n', u' File "/usr/lib/python2.7/dist-packages/nova/network/model.py", line 379, in __len__\nreturn self._sync_wrapper(fn, *args, **kwargs)\n', u' File "/usr/lib/python2.7/dist-packages/nova/network/model.py", line 366, in _sync_wrapper\nself.wait()\n', u' File "/usr/lib/python2.7/dist-p ackages/nova/network/model.py", line 398, in wait\nself[:] = self._gt.wait( )\n', u' File "/usr/lib/python2.7/dist-packages/eventlet/greenthread.py", line 168, in wait\nreturn self._exit_event.wait()\n', u' File "/usr/lib/python2.7 /dist-packages/eventlet/event.py", line 120, in wait\ncurrent.throw(*sel f._exc)\n', u' File "/usr/lib/python2.7/dist-packages/eventlet/greenthread.py", line 194, in main\nresult = function(*args, **kwargs)\n', u' File "/usr/ lib/python2.7/dist-packages/nova/compute/manager.py", line 1228, in _allocate_n etwork_async\ndhcp_options=dhcp_options)\n', u' File "/usr/lib/python2.7/dist-packages/nova/network/api.py", line 49, in wrapper\nres = f(self, context, *args, **kwargs)\n', u' File "/usr/lib/python2.7/dist-packages/nova/network/neutronv2/api.py", line 243, in allocate_for_instance\nraise exception.Po rtInUse(port_id=port_id)\n', u'PortInUse: Port 69e55016-b794-4dd9-b3f3-4e78336f bd11 is still in use.\n'] 93 2013-12-12 20:55:35.321 20945 WARNING nova.scheduler.utils [req-844a11f6-66f1-4fc7-9f3b-8bacbe57a04d e95097acd0b041558f33d07f720c1bd7 354f17bf81924b278806c3e3798aa527] Failed to scheduler_run_instance: No valid host was found. Exceeded max scheduling attempts 3 for instance 2a3eadcc-5230-4ed6-bf9f-9a82d43b91c3 94 2013-12-12 20:55:35.324 20945 WARNING nova.scheduler.utils [req-844a11f6-66f1-4fc7-9f3b-8bacbe57a04d e95097acd0b041558f33d07f720c1bd7 354f17bf81924b278806c3e3798aa527] [instance: 2a3eadcc-5230-4ed6-bf9f-9a82d43b91c3] Setting instance to ERROR state. ** Affects: neutron Importance: Undecided Status: New ** Attachment added: "server.log" https://bugs.launchpad.net/bugs/1260516/+attachment/3928318/+files/server.log -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1260516 Title: PortInUse exception leaves orphaned ports Status in OpenStack Neutron (virtual network service): New Bug description: While bringing up 1000 VMs with Heat, some VMs are in error state. After deleting the VMs, not all ports are removed. The next time VMs are created, they fail due to no IP addresses left. Note, initially there are 7 ports, no VMs 804 VMs are created, 811 ports root@control01:/usr/share/pyshared/heat# nova list | grep
[Yahoo-eng-team] [Bug 1260528] [NEW] Metering dashboard. Marker could not be found (havana)
Public bug reported: Hello, I couldn't reopen this bug https://bugs.launchpad.net/horizon/+bug/1247752 . And decided to create new one. I use latest havana release code, but also plunged into "Marker could not be found" error in horizon logs. [Thu Dec 12 22:49:15 2013] [error] Request returned failure status: 400 [Thu Dec 12 22:49:18 2013] [error] REQ: curl -i -X GET http://192.168.0.2:35357/v2.0/tenants?marker=tenant_marker&limit=21 -H "User-Agent: python-keystoneclient" -H "Forwarded: for=10.20.0.1;by=python-keystonece" [Thu Dec 12 22:49:18 2013] [error] RESP: [400] {'date': 'Thu, 12 Dec 2013 22:49:18 GMT', 'content-type': 'application/json', 'content-length': '88', 'vary': 'X-Auth-Token'} [Thu Dec 12 22:49:18 2013] [error] RESP BODY: {"error": {"message": "Marker could not be found", "code": 400, "title": "Bad Request"}} "tenant_marker" value comes from https://github.com/openstack/horizon/blob/stable/havana/openstack_dashboard/dashboards/admin/metering/views.py#L149 ** Affects: horizon Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1260528 Title: Metering dashboard. Marker could not be found (havana) Status in OpenStack Dashboard (Horizon): New Bug description: Hello, I couldn't reopen this bug https://bugs.launchpad.net/horizon/+bug/1247752 . And decided to create new one. I use latest havana release code, but also plunged into "Marker could not be found" error in horizon logs. [Thu Dec 12 22:49:15 2013] [error] Request returned failure status: 400 [Thu Dec 12 22:49:18 2013] [error] REQ: curl -i -X GET http://192.168.0.2:35357/v2.0/tenants?marker=tenant_marker&limit=21 -H "User-Agent: python-keystoneclient" -H "Forwarded: for=10.20.0.1;by=python-keystonece" [Thu Dec 12 22:49:18 2013] [error] RESP: [400] {'date': 'Thu, 12 Dec 2013 22:49:18 GMT', 'content-type': 'application/json', 'content-length': '88', 'vary': 'X-Auth-Token'} [Thu Dec 12 22:49:18 2013] [error] RESP BODY: {"error": {"message": "Marker could not be found", "code": 400, "title": "Bad Request"}} "tenant_marker" value comes from https://github.com/openstack/horizon/blob/stable/havana/openstack_dashboard/dashboards/admin/metering/views.py#L149 To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1260528/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260530] [NEW] Instances appear pingable without an ingress icmp security rule
Public bug reported: Instances appear to be pingable for a short time after a floating ip is associated even though there is no ingress icmp security group rule. tcpdump of the instance's tap device shows that the instance isn't actually responding to the ping. It appears that the router gateway interface is responding to the ping for a short time. You can reproduce this by booting an instance using a security group with only egress rules. Allocate a floating ip address. ping the ip address ( nothing will happen yet ). Associate the ip with the instance. The ping will begin responding. ** Affects: neutron Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1260530 Title: Instances appear pingable without an ingress icmp security rule Status in OpenStack Neutron (virtual network service): New Bug description: Instances appear to be pingable for a short time after a floating ip is associated even though there is no ingress icmp security group rule. tcpdump of the instance's tap device shows that the instance isn't actually responding to the ping. It appears that the router gateway interface is responding to the ping for a short time. You can reproduce this by booting an instance using a security group with only egress rules. Allocate a floating ip address. ping the ip address ( nothing will happen yet ). Associate the ip with the instance. The ping will begin responding. To manage notifications about this bug go to: https://bugs.launchpad.net/neutron/+bug/1260530/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1161988] Re: Flavor naming shouldn't include "m1"
** Changed in: devstack Status: Confirmed => Won't Fix -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1161988 Title: Flavor naming shouldn't include "m1" Status in devstack - openstack dev environments: Won't Fix Status in OpenStack Dashboard (Horizon): Won't Fix Status in OpenStack Compute (Nova): Won't Fix Status in Python client library for heat: Confirmed Status in Python client library for Nova: Won't Fix Status in Tempest: Won't Fix Bug description: Flavor naming shouldn't include "m1" ENV: devstack trunk / nova 814e109845b3b2546f60e3f537dcfe32893906a3 (grizzly) The default flavors are now: m1.nano m1.micro m1.tiny m1.small m1.medium m1.large m1.xlarge We are propagating AWS "m1" designation. This is not useful information to the OpenStack administrator or user, and it's actually possible misinformation as the "m1" on AWS suggests a specific generation of hardware. POSSIBLE SOLUTION: Drop the "m1": nano micro tiny small medium large xlarge To manage notifications about this bug go to: https://bugs.launchpad.net/devstack/+bug/1161988/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1217432] Re: timeout on AuthorizationTestJSON
This is a glance call that's failing to allocate the image ** Also affects: glance Importance: Undecided Status: New ** Also affects: nova Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to Glance. https://bugs.launchpad.net/bugs/1217432 Title: timeout on AuthorizationTestJSON Status in OpenStack Image Registry and Delivery Service (Glance): New Status in OpenStack Compute (Nova): New Status in Tempest: New Bug description: http://logs.openstack.org/59/43459/3/gate/gate-tempest-devstack-vm- full/e57504d/console.html 2013-08-27 14:39:29.384 | == 2013-08-27 14:39:29.384 | FAIL: setUpClass (tempest.api.compute.test_authorization.AuthorizationTestJSON) 2013-08-27 14:39:29.385 | setUpClass (tempest.api.compute.test_authorization.AuthorizationTestJSON) 2013-08-27 14:39:29.385 | -- 2013-08-27 14:39:29.385 | _StringException: Traceback (most recent call last): 2013-08-27 14:39:29.386 | File "tempest/api/compute/test_authorization.py", line 66, in setUpClass 2013-08-27 14:39:29.386 | cls.images_client.wait_for_image_status(image_id, 'ACTIVE') 2013-08-27 14:39:29.386 | File "tempest/services/compute/json/images_client.py", line 110, in wait_for_image_status 2013-08-27 14:39:29.386 | raise exceptions.TimeoutException 2013-08-27 14:39:29.386 | TimeoutException: Request timed out 2013-08-27 14:39:29.386 | 2013-08-27 14:39:29.387 | 2013-08-27 14:39:29.387 | == 2013-08-27 14:39:29.388 | FAIL: process-returncode 2013-08-27 14:39:29.388 | process-returncode 2013-08-27 14:39:29.416 | -- 2013-08-27 14:39:29.416 | _StringException: Binary content: 2013-08-27 14:39:29.416 | traceback (test/plain; charset="utf8") 2013-08-27 14:39:29.416 | 2013-08-27 14:39:29.417 | 2013-08-27 14:39:29.417 | -- 2013-08-27 14:39:29.418 | Ran 1152 tests in 968.915s 2013-08-27 14:39:29.418 | 2013-08-27 14:39:29.419 | FAILED (failures=2, skipped=67) To manage notifications about this bug go to: https://bugs.launchpad.net/glance/+bug/1217432/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260538] [NEW] nova-manage useage exposes action-args
You have been subscribed to a public bug: The nova-manage command exposes the action_args options during the usage output for command. E.g. $ nova-manage network modify -h usage: nova-manage network modify [-h] [--fixed_range ] [--project ] [--host ] [--disassociate-project] [--disassociate-host] [action_args [action_args ...]] positional arguments: action_args This can cause confusion as users naturally expect there to be more "actions" on commands like "modify". Even in straightforward cases, this positional argument leaks into usage. $ nova-manage db version -h usage: nova-manage db version [-h] [action_args [action_args ...]] positional arguments: action_args Please consider suppressing documentation on action_args. In addition, expose the __doc__ strings for these functions, which is done in the nova command. ** Affects: nova Importance: Undecided Status: New -- nova-manage useage exposes action-args https://bugs.launchpad.net/bugs/1260538 You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260538] [NEW] nova-manage useage exposes action-args
Public bug reported: The nova-manage command exposes the action_args options during the usage output for command. E.g. $ nova-manage network modify -h usage: nova-manage network modify [-h] [--fixed_range ] [--project ] [--host ] [--disassociate-project] [--disassociate-host] [action_args [action_args ...]] positional arguments: action_args This can cause confusion as users naturally expect there to be more "actions" on commands like "modify". Even in straightforward cases, this positional argument leaks into usage. $ nova-manage db version -h usage: nova-manage db version [-h] [action_args [action_args ...]] positional arguments: action_args Please consider suppressing documentation on action_args. In addition, expose the __doc__ strings for these functions, which is done in the nova command. ** Affects: nova Importance: Undecided Status: New ** Tags: low-hanging-fruit nova-manage user-experience ux ** Project changed: barbican => nova ** Tags added: low-hanging-fruit user-experience ** Tags added: nova-manage ux -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1260538 Title: nova-manage useage exposes action-args Status in OpenStack Compute (Nova): New Bug description: The nova-manage command exposes the action_args options during the usage output for command. E.g. $ nova-manage network modify -h usage: nova-manage network modify [-h] [--fixed_range ] [--project ] [--host ] [--disassociate-project] [--disassociate-host] [action_args [action_args ...]] positional arguments: action_args This can cause confusion as users naturally expect there to be more "actions" on commands like "modify". Even in straightforward cases, this positional argument leaks into usage. $ nova-manage db version -h usage: nova-manage db version [-h] [action_args [action_args ...]] positional arguments: action_args Please consider suppressing documentation on action_args. In addition, expose the __doc__ strings for these functions, which is done in the nova command. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1260538/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1217734] Re: FAIL: setUpClass (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML Unauthorized)
if you can't find a gate race in logstash, I'm calling it fixed ** Changed in: nova Status: Incomplete => Fix Released ** Changed in: python-keystoneclient Status: Incomplete => Fix Released ** Changed in: tempest Status: Incomplete => Fix Released ** Changed in: python-cinderclient Status: Incomplete => Fix Released -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1217734 Title: FAIL: setUpClass (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML Unauthorized) Status in OpenStack Compute (Nova): Fix Released Status in Python client library for Cinder: Fix Released Status in Python client library for Keystone: Fix Released Status in Tempest: Fix Released Bug description: http://logs.openstack.org/44/43444/4/check/gate-grenade-devstack- vm/4f78566/console.html 2013-08-28 06:32:58.510 | == 2013-08-28 06:32:58.511 | FAIL: setUpClass (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML) 2013-08-28 06:32:58.511 | setUpClass (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML) 2013-08-28 06:32:58.511 | -- 2013-08-28 06:32:58.512 | _StringException: Traceback (most recent call last): 2013-08-28 06:32:58.512 | File "tempest/api/compute/servers/test_server_rescue.py", line 52, in setUpClass 2013-08-28 06:32:58.512 | 'test_attach') 2013-08-28 06:32:58.512 | File "tempest/services/compute/xml/volumes_extensions_client.py", line 114, in create_volume 2013-08-28 06:32:58.513 | self.headers) 2013-08-28 06:32:58.513 | File "tempest/common/rest_client.py", line 260, in post 2013-08-28 06:32:58.513 | return self.request('POST', url, headers, body) 2013-08-28 06:32:58.514 | File "tempest/common/rest_client.py", line 388, in request 2013-08-28 06:32:58.514 | resp, resp_body) 2013-08-28 06:32:58.514 | File "tempest/common/rest_client.py", line 430, in _error_checker 2013-08-28 06:32:58.515 | raise exceptions.Unauthorized() 2013-08-28 06:32:58.515 | Unauthorized: Unauthorized http://logs.openstack.org/23/43723/4/gate/gate-tempest-devstack-vm- full/3fefc90/console.html The real happening time is close to: 2013-08-28 06:24:24.882 | setUpClass (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML) To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1217734/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1260440] Re: nova-compute host is added to scheduling pool before Neutron can bind network ports on said host
Reviewed: https://review.openstack.org/61608 Committed: https://git.openstack.org/cgit/openstack/tripleo-incubator/commit/?id=661884b5c7a47d01171c680c83b601d3c9a15d9f Submitter: Jenkins Branch:master commit 661884b5c7a47d01171c680c83b601d3c9a15d9f Author: Clint Byrum Date: Wed Dec 11 15:33:01 2013 -0800 Wait for Neutron L2 Agent on Compute Node The L2 Agent sometimes does not register until later on in the deployment for some reason. This is just a work-around until that bug can be properly understood. Change-Id: Idbbc977aa2e13f2026de05ae7e6571bc9dd0a498 Closes-Bug: #1260440 ** Changed in: tripleo Status: In Progress => Fix Released -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1260440 Title: nova-compute host is added to scheduling pool before Neutron can bind network ports on said host Status in OpenStack Neutron (virtual network service): New Status in OpenStack Compute (Nova): Confirmed Status in tripleo - openstack on openstack: Fix Released Bug description: This is a race condition. Given a cloud with 0 compute nodes available, on a compute node: * Start up neutron-openvswitch-agent * Start up nova-compute * nova boot an instance Scenario 1: * neutron-openvswitch-agent registers with Neutron before nova tries to boot instance * port is bound to agent * instance boots with correct networking Scenario 2: * nova schedules instance to host before neutron-openvswitch-agent is registered with Neutron * nova instance fails with vif_type=binding_failed * instance is in ERROR state I would expect that Nova would not try to schedule instances on compute hosts that are not ready. Please also see this mailing list thread for more info: http://lists.openstack.org/pipermail/openstack- dev/2013-December/022084.html To manage notifications about this bug go to: https://bugs.launchpad.net/neutron/+bug/1260440/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1249889] Re: tempest.scenario.test_volume_boot_pattern.TestVolumeBootPattern.test_volume_boot_pattern[compute, image, volume] failed
not a tempest bug, this looks to be a nova bug with some sort of race on nw attach ** Also affects: nova Importance: Undecided Status: New ** Changed in: tempest Status: New => Invalid -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1249889 Title: tempest.scenario.test_volume_boot_pattern.TestVolumeBootPattern.test_volume_boot_pattern[compute,image,volume] failed Status in OpenStack Compute (Nova): New Status in Tempest: Invalid Bug description: Traceback (most recent call last): File "tempest/scenario/test_volume_boot_pattern.py", line 144, in test_volume_boot_pattern keypair) File "tempest/scenario/test_volume_boot_pattern.py", line 93, in _ssh_to_server server.add_floating_ip(floating_ip) File "/opt/stack/new/python-novaclient/novaclient/v1_1/servers.py", line 108, in add_floating_ip self.manager.add_floating_ip(self, address, fixed_address) File "/opt/stack/new/python-novaclient/novaclient/v1_1/servers.py", line 465, in add_floating_ip self._action('addFloatingIp', server, {'address': address}) File "/opt/stack/new/python-novaclient/novaclient/v1_1/servers.py", line 993, in _action return self.api.client.post(url, body=body) File "/opt/stack/new/python-novaclient/novaclient/client.py", line 234, in post return self._cs_request(url, 'POST', **kwargs) File "/opt/stack/new/python-novaclient/novaclient/client.py", line 213, in _cs_request **kwargs) File "/opt/stack/new/python-novaclient/novaclient/client.py", line 195, in _time_request resp, body = self.request(url, method, **kwargs) File "/opt/stack/new/python-novaclient/novaclient/client.py", line 189, in request raise exceptions.from_response(resp, body, url, method) BadRequest: No nw_info cache associated with instance (HTTP 400) (Request-ID: req-4e6ed4cd-d2e8-42a2-aae6-f0a3820f71f5) To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1249889/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1225024] Re: tempest.api.compute.admin.test_hosts.HostsAdminTestXML.test_list_hosts_with_zone[gate] unexpected conductor service
Can't find it in log stash, assuming it's fixed ** Changed in: nova Status: Incomplete => Fix Released ** Changed in: tempest Status: Incomplete => Fix Released -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1225024 Title: tempest.api.compute.admin.test_hosts.HostsAdminTestXML.test_list_hosts_with_zone[gate] unexpected conductor service Status in OpenStack Compute (Nova): Fix Released Status in Tempest: Fix Released Bug description: 2013-09-13 03:44:02.095 | == 2013-09-13 03:44:02.095 | FAIL: tempest.api.compute.admin.test_hosts.HostsAdminTestXML.test_list_hosts_with_zone[gate] 2013-09-13 03:44:02.096 | tempest.api.compute.admin.test_hosts.HostsAdminTestXML.test_list_hosts_with_zone[gate] 2013-09-13 03:44:02.096 | -- 2013-09-13 03:44:02.096 | _StringException: Empty attachments: 2013-09-13 03:44:02.097 | stderr 2013-09-13 03:44:02.097 | stdout 2013-09-13 03:44:02.097 | 2013-09-13 03:44:02.098 | pythonlogging:'': {{{ 2013-09-13 03:44:02.098 | 2013-09-13 03:27:35,206 Request: GET http://127.0.0.1:8774/v2/b5dc34c995d94389b0b2a5f18851aca6/os-hosts 2013-09-13 03:44:02.098 | 2013-09-13 03:27:35,487 Response Status: 200 2013-09-13 03:44:02.099 | 2013-09-13 03:27:35,488 Nova request id: req-10082857-8c33-4472-b511-5a5945cc2da4 2013-09-13 03:44:02.099 | 2013-09-13 03:27:35,488 Request: GET http://127.0.0.1:8774/v2/b5dc34c995d94389b0b2a5f18851aca6/os-hosts?zone=internal 2013-09-13 03:44:02.099 | 2013-09-13 03:27:35,513 Response Status: 200 2013-09-13 03:44:02.099 | 2013-09-13 03:27:35,513 Nova request id: req-29968204-ebbb-468c-ab8c-d25654550a97 2013-09-13 03:44:02.100 | }}} 2013-09-13 03:44:02.100 | 2013-09-13 03:44:02.100 | Traceback (most recent call last): 2013-09-13 03:44:02.101 | File "tempest/api/compute/admin/test_hosts.py", line 51, in test_list_hosts_with_zone 2013-09-13 03:44:02.101 | self.assertIn(host, hosts) 2013-09-13 03:44:02.101 | File "/usr/local/lib/python2.7/dist-packages/testtools/testcase.py", line 328, in assertIn 2013-09-13 03:44:02.101 | self.assertThat(haystack, Contains(needle)) 2013-09-13 03:44:02.102 | File "/usr/local/lib/python2.7/dist-packages/testtools/testcase.py", line 417, in assertThat 2013-09-13 03:44:02.102 | raise MismatchError(matchee, matcher, mismatch, verbose) 2013-09-13 03:44:02.103 | MismatchError: {u'service': u'conductor', u'host_name': u'devstack-precise-hpcloud-az3-265828', u'zone': u'internal'} not in [{u'service': u'network', u'host_name': u'devstack-precise-hpcloud-az3-265828', u'zone': u'internal'}, {u'service': u'cert', u'host_name': u'devstack-precise-hpcloud-az3-265828', u'zone': u'internal'}, {u'service': u'scheduler', u'host_name': u'devstack-precise-hpcloud-az3-265828', u'zone': u'internal'}] 2013-09-13 03:44:02.103 | To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1225024/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp