[Yahoo-eng-team] [Bug 1260224] [NEW] agent model migration script does not contain ml2 plugin

2013-12-12 Thread yong sheng gong
Public bug reported:

+ mysql -uroot -proot -h127.0.0.1 -e 'DROP DATABASE IF EXISTS neutron_ml2;'
+ mysql -uroot -proot -h127.0.0.1 -e 'CREATE DATABASE neutron_ml2 CHARACTER SET 
utf8;'
+ /usr/local/bin/neutron-db-manage --config-file /etc/neutron/neutron.conf 
--config-file /etc/neutron/plugins/ml2/ml2_conf.ini upgrade head
No handlers could be found for logger "neutron.common.legacy"
INFO  [alembic.migration] Context impl MySQLImpl.
INFO  [alembic.migration] Will assume non-transactional DDL.
INFO  [alembic.migration] Running upgrade None -> folsom, folsom initial 
database
INFO  [alembic.migration] Running upgrade folsom -> 2c4af419145b, l3_support
INFO  [alembic.migration] Running upgrade 2c4af419145b -> 5a875d0e5c, ryu
INFO  [alembic.migration] Running upgrade 5a875d0e5c -> 48b6f43f7471, DB 
support for service types
INFO  [alembic.migration] Running upgrade 48b6f43f7471 -> 3cb5d900c5de, 
security_groups
INFO  [alembic.migration] Running upgrade 3cb5d900c5de -> 1d76643bcec4, 
nvp_netbinding
INFO  [alembic.migration] Running upgrade 1d76643bcec4 -> 2a6d0b51f4bb, cisco 
plugin cleanup
INFO  [alembic.migration] Running upgrade 2a6d0b51f4bb -> 1b693c095aa3, Quota 
ext support added in Grizzly
INFO  [alembic.migration] Running upgrade 1b693c095aa3 -> 1149d7de0cfa, inital 
port security
INFO  [alembic.migration] Running upgrade 1149d7de0cfa -> 49332180ca96, ryu 
plugin update
INFO  [alembic.migration] Running upgrade 49332180ca96 -> 38335592a0dc, 
nvp_portmap
INFO  [alembic.migration] Running upgrade 38335592a0dc -> 54c2c487e913, 'DB 
support for load balancing service
INFO  [alembic.migration] Running upgrade 54c2c487e913 -> 45680af419f9, nvp_qos
INFO  [alembic.migration] Running upgrade 45680af419f9 -> 1c33fa3cd1a1, Support 
routing table configuration on Router
INFO  [alembic.migration] Running upgrade 1c33fa3cd1a1 -> 363468ac592c, 
nvp_network_gw
INFO  [alembic.migration] Running upgrade 363468ac592c -> 511471cc46b, Add 
agent management extension model support
INFO  [alembic.migration] Running upgrade 511471cc46b -> 3b54bf9e29f7, NEC 
plugin sharednet
INFO  [alembic.migration] Running upgrade 3b54bf9e29f7 -> 4692d074d587, agent 
scheduler
INFO  [alembic.migration] Running upgrade 4692d074d587 -> 1341ed32cc1e, 
nvp_net_binding
INFO  [alembic.migration] Running upgrade 1341ed32cc1e -> grizzly, grizzly
INFO  [alembic.migration] Running upgrade grizzly -> f489cf14a79c, DB support 
for load balancing service (havana)
INFO  [alembic.migration] Running upgrade f489cf14a79c -> 176a85fc7d79, Add 
portbindings db
INFO  [alembic.migration] Running upgrade 176a85fc7d79 -> 32b517556ec9, remove 
TunnelIP model
INFO  [alembic.migration] Running upgrade 32b517556ec9 -> 128e042a2b68, 
ext_gw_mode
INFO  [alembic.migration] Running upgrade 128e042a2b68 -> 5ac71e65402c, 
ml2_initial
INFO  [alembic.migration] Running upgrade 5ac71e65402c -> 3cbf70257c28, 
nvp_mac_learning
INFO  [alembic.migration] Running upgrade 3cbf70257c28 -> 5918cbddab04, add 
tables for router rules support
INFO  [alembic.migration] Running upgrade 5918cbddab04 -> 3cabb850f4a5, Table 
to track port to host associations
INFO  [alembic.migration] Running upgrade 3cabb850f4a5 -> b7a8863760e, Remove 
cisco_vlan_bindings table
INFO  [alembic.migration] Running upgrade b7a8863760e -> 13de305df56e, 
nec_add_pf_name
INFO  [alembic.migration] Running upgrade 13de305df56e -> 20ae61555e95, DB 
Migration for ML2 GRE Type Driver
INFO  [alembic.migration] Running upgrade 20ae61555e95 -> 477a4488d3f4, DB 
Migration for ML2 VXLAN Type Driver
INFO  [alembic.migration] Running upgrade 477a4488d3f4 -> 2032abe8edac, LBaaS 
add status description
INFO  [alembic.migration] Running upgrade 2032abe8edac -> 52c5e4a18807, LBaaS 
Pool scheduler
INFO  [alembic.migration] Running upgrade 52c5e4a18807 -> 557edfc53098, New 
service types framework (service providers)
INFO  [alembic.migration] Running upgrade 557edfc53098 -> e6b16a30d97, Add 
cisco_provider_networks table
INFO  [alembic.migration] Running upgrade e6b16a30d97 -> 39cf3f799352, FWaaS 
Havana-2 model
INFO  [alembic.migration] Running upgrade 39cf3f799352 -> 52ff27f7567a, Support 
for VPNaaS
INFO  [alembic.migration] Running upgrade 52ff27f7567a -> 11c6e18605c8, Pool 
Monitor status field
INFO  [alembic.migration] Running upgrade 11c6e18605c8 -> 35c7c198ddea, remove 
status from HealthMonitor
INFO  [alembic.migration] Running upgrade 35c7c198ddea -> 263772d65691, Cisco 
plugin db cleanup part II
INFO  [alembic.migration] Running upgrade 263772d65691 -> c88b6b5fea3, Cisco 
N1KV tables
INFO  [alembic.migration] Running upgrade c88b6b5fea3 -> f9263d6df56, 
remove_dhcp_lease
INFO  [alembic.migration] Running upgrade f9263d6df56 -> 569e98a8132b, metering
INFO  [alembic.migration] Running upgrade 569e98a8132b -> 86cf4d88bd3, remove 
bigswitch port tracking table
INFO  [alembic.migration] Running upgrade 86cf4d88bd3 -> 3c6e57a23db4, add 
multiprovider
INFO  [alembic.migration] Running upgrade 3c6e57a23db4 -> 63afba73813, Add 
u

[Yahoo-eng-team] [Bug 1260227] [NEW] Pre-select the network when there's only one.

2013-12-12 Thread Zhenguo Niu
Public bug reported:

When launching instance with neutron, network field is required. If
there is only one, it should be pre-selected.

** Affects: horizon
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1260227

Title:
  Pre-select the network when there's only one.

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  When launching instance with neutron, network field is required. If
  there is only one, it should be pre-selected.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1260227/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260228] [NEW] Pre-select the network when there's only one.

2013-12-12 Thread Zhenguo Niu
*** This bug is a duplicate of bug 1260227 ***
https://bugs.launchpad.net/bugs/1260227

Public bug reported:

When launching instance with neutron, network field is required. If
there is only one, it should be pre-selected.

** Affects: horizon
 Importance: Undecided
 Status: New

** This bug has been marked a duplicate of bug 1260227
   Pre-select the network when there's only one.

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1260228

Title:
  Pre-select the network when there's only one.

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  When launching instance with neutron, network field is required. If
  there is only one, it should be pre-selected.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1260228/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260232] [NEW] db migration on unique constraint should not be applied to all plugins

2013-12-12 Thread Edgar Magana
Public bug reported:

There is an assumption that all "neutron plugins" creates the
plumgrid_neutron.agents table, which is not the case. I just tested big
switch and plumgrid plugins and they are failing:

INFO  [alembic.migration] Running upgrade havana -> e197124d4b9, add unique 
constraint to members
INFO  [alembic.migration] Running upgrade e197124d4b9 -> 1fcfc149aca4, Add a 
unique constraint on (agent_type, host) columns to prevent a race
condition when an agent entry is 'upserted'.
Traceback (most recent call last):
  File "/usr/local/bin/neutron-db-manage", line 10, in 
sys.exit(main())
  File "/opt/stack/neutron/neutron/db/migration/cli.py", line 143, in main
CONF.command.func(config, CONF.command.name)
  File "/opt/stack/neutron/neutron/db/migration/cli.py", line 80, in 
do_upgrade_downgrade
do_alembic_command(config, cmd, revision, sql=CONF.command.sql)
  File "/opt/stack/neutron/neutron/db/migration/cli.py", line 59, in 
do_alembic_command
getattr(alembic_command, cmd)(config, *args, **kwargs)
  File "/usr/local/lib/python2.7/dist-packages/alembic/command.py", line 124, 
in upgrade
script.run_env()
  File "/usr/local/lib/python2.7/dist-packages/alembic/script.py", line 193, in 
run_env
util.load_python_file(self.dir, 'env.py')
  File "/usr/local/lib/python2.7/dist-packages/alembic/util.py", line 177, in 
load_python_file
module = load_module(module_id, path)
  File "/usr/local/lib/python2.7/dist-packages/alembic/compat.py", line 39, in 
load_module
return imp.load_source(module_id, path, fp)
  File "/opt/stack/neutron/neutron/db/migration/alembic_migrations/env.py", 
line 105, in 
run_migrations_online()
  File "/opt/stack/neutron/neutron/db/migration/alembic_migrations/env.py", 
line 89, in run_migrations_online
options=build_options())
  File "", line 7, in run_migrations
  File "/usr/local/lib/python2.7/dist-packages/alembic/environment.py", line 
652, in run_migrations
self.get_context().run_migrations(**kw)
  File "/usr/local/lib/python2.7/dist-packages/alembic/migration.py", line 224, 
in run_migrations
change(**kw)
  File 
"/opt/stack/neutron/neutron/db/migration/alembic_migrations/versions/1fcfc149aca4_agents_unique_by_type_and_host.py",
 line 50, in upgrade
local_cols=['agent_type', 'host']
  File "", line 7, in create_unique_constraint
  File "/usr/local/lib/python2.7/dist-packages/alembic/operations.py", line 
539, in create_unique_constraint
schema=schema, **kw)
  File "/usr/local/lib/python2.7/dist-packages/alembic/ddl/impl.py", line 135, 
in add_constraint
self._exec(schema.AddConstraint(const))
  File "/usr/local/lib/python2.7/dist-packages/alembic/ddl/impl.py", line 76, 
in _exec
conn.execute(construct, *multiparams, **params)
  File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 
1449, in execute
params)
  File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 
1542, in _execute_ddl
compiled
  File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 
1698, in _execute_context
context)
  File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 
1691, in _execute_context
context)
  File "/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/default.py", 
line 331, in do_execute
cursor.execute(statement, parameters)
  File "/usr/lib/python2.7/dist-packages/MySQLdb/cursors.py", line 174, in 
execute
self.errorhandler(self, exc, value)
  File "/usr/lib/python2.7/dist-packages/MySQLdb/connections.py", line 36, in 
defaulterrorhandler
raise errorclass, errorvalue
sqlalchemy.exc.ProgrammingError: (ProgrammingError) (1146, "Table 
'plumgrid_neutron.agents' doesn't exist") 'ALTER TABLE agents ADD CONSTRAINT 
uniq_agents0agent_type0host UNIQUE (agent_type, host)' ()
++ failed
++ local r=1
+++ jobs -p
++ kill
++ set +o xtrace

** Affects: neutron
 Importance: Undecided
 Assignee: Edgar Magana (emagana)
 Status: In Progress

** Changed in: neutron
 Assignee: (unassigned) => Edgar Magana (emagana)

** Changed in: neutron
   Status: New => In Progress

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1260232

Title:
  db migration on unique constraint should not be applied to all plugins

Status in OpenStack Neutron (virtual network service):
  In Progress

Bug description:
  There is an assumption that all "neutron plugins" creates the
  plumgrid_neutron.agents table, which is not the case. I just tested
  big switch and plumgrid plugins and they are failing:

  INFO  [alembic.migration] Running upgrade havana -> e197124d4b9, add unique 
constraint to members
  INFO  [alembic.migration] Running upgrade e197124d4b9 -> 1fcfc149aca4, Add a 
unique constraint on (agent_type, host) columns to prevent a race
  condition when an agent entry is 'upserted'.
  Traceback (most recent call l

[Yahoo-eng-team] [Bug 1260233] [NEW] db migration (agents constraint) fails when using ryu plugin

2013-12-12 Thread Yoshihiro Kaneko
Public bug reported:

Ryu plugin does not support agent extension yet.
Therefore, 511471cc46b_agent_ext_model_supp.py does not contain ryu plugin, and 
agents table is not created.
However, 1fcfc149aca4_agents_unique_by_type_and_host.py does not consider this 
case.
I think that migration_for_plugins of 1fcfc149aca4 should be the same as 
511471cc46b's.

2013-12-12 01:08:44 INFO  [alembic.migration] Running upgrade e197124d4b9 -> 
1fcfc149aca4, Add a unique constraint on (agent_type, host) columns to prevent 
a race
2013-12-12 01:08:44 condition when an agent entry is 'upserted'.
2013-12-12 01:08:44 Traceback (most recent call last):
2013-12-12 01:08:44   File "/usr/local/bin/neutron-db-manage", line 10, in 

2013-12-12 01:08:44 sys.exit(main())
2013-12-12 01:08:44   File "/opt/stack/neutron/neutron/db/migration/cli.py", 
line 143, in main
2013-12-12 01:08:44 CONF.command.func(config, CONF.command.name)
2013-12-12 01:08:44   File "/opt/stack/neutron/neutron/db/migration/cli.py", 
line 80, in do_upgrade_downgrade
2013-12-12 01:08:44 do_alembic_command(config, cmd, revision, 
sql=CONF.command.sql)
2013-12-12 01:08:44   File "/opt/stack/neutron/neutron/db/migration/cli.py", 
line 59, in do_alembic_command
2013-12-12 01:08:44 getattr(alembic_command, cmd)(config, *args, **kwargs)
2013-12-12 01:08:44   File 
"/usr/local/lib/python2.7/dist-packages/alembic/command.py", line 124, in 
upgrade
2013-12-12 01:08:44 script.run_env()
2013-12-12 01:08:44   File 
"/usr/local/lib/python2.7/dist-packages/alembic/script.py", line 193, in run_env
2013-12-12 01:08:44 util.load_python_file(self.dir, 'env.py')
2013-12-12 01:08:44   File 
"/usr/local/lib/python2.7/dist-packages/alembic/util.py", line 177, in 
load_python_file
2013-12-12 01:08:44 module = load_module(module_id, path)
2013-12-12 01:08:44   File 
"/usr/local/lib/python2.7/dist-packages/alembic/compat.py", line 39, in 
load_module
2013-12-12 01:08:44 return imp.load_source(module_id, path, fp)
2013-12-12 01:08:44   File 
"/opt/stack/neutron/neutron/db/migration/alembic_migrations/env.py", line 105, 
in 
2013-12-12 01:08:44 run_migrations_online()
2013-12-12 01:08:44   File 
"/opt/stack/neutron/neutron/db/migration/alembic_migrations/env.py", line 89, 
in run_migrations_online
2013-12-12 01:08:44 options=build_options())
2013-12-12 01:08:44   File "", line 7, in run_migrations
2013-12-12 01:08:44   File 
"/usr/local/lib/python2.7/dist-packages/alembic/environment.py", line 652, in 
run_migrations
2013-12-12 01:08:45 self.get_context().run_migrations(**kw)
2013-12-12 01:08:45   File 
"/usr/local/lib/python2.7/dist-packages/alembic/migration.py", line 224, in 
run_migrations
2013-12-12 01:08:45 change(**kw)
2013-12-12 01:08:45   File 
"/opt/stack/neutron/neutron/db/migration/alembic_migrations/versions/1fcfc149aca4_agents_unique_by_type_and_host.py",
 line 50, in upgrade
2013-12-12 01:08:45 local_cols=['agent_type', 'host']
2013-12-12 01:08:45   File "", line 7, in create_unique_constraint
2013-12-12 01:08:45   File 
"/usr/local/lib/python2.7/dist-packages/alembic/operations.py", line 539, in 
create_unique_constraint
2013-12-12 01:08:45 schema=schema, **kw)
2013-12-12 01:08:45   File 
"/usr/local/lib/python2.7/dist-packages/alembic/ddl/impl.py", line 135, in 
add_constraint
2013-12-12 01:08:45 self._exec(schema.AddConstraint(const))
2013-12-12 01:08:45   File 
"/usr/local/lib/python2.7/dist-packages/alembic/ddl/impl.py", line 76, in _exec
2013-12-12 01:08:45 conn.execute(construct, *multiparams, **params)
2013-12-12 01:08:45   File 
"/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 1449, 
in execute
2013-12-12 01:08:45 params)
2013-12-12 01:08:45   File 
"/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 1542, 
in _execute_ddl
2013-12-12 01:08:45 compiled
2013-12-12 01:08:45   File 
"/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 1698, 
in _execute_context
2013-12-12 01:08:45 context)
2013-12-12 01:08:45   File 
"/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/base.py", line 1691, 
in _execute_context
2013-12-12 01:08:45 context)
2013-12-12 01:08:45   File 
"/usr/local/lib/python2.7/dist-packages/sqlalchemy/engine/default.py", line 
331, in do_execute
2013-12-12 01:08:45 cursor.execute(statement, parameters)
2013-12-12 01:08:45   File 
"/usr/lib/python2.7/dist-packages/MySQLdb/cursors.py", line 174, in execute
2013-12-12 01:08:45 self.errorhandler(self, exc, value)
2013-12-12 01:08:45   File 
"/usr/lib/python2.7/dist-packages/MySQLdb/connections.py", line 36, in 
defaulterrorhandler
2013-12-12 01:08:45 raise errorclass, errorvalue
2013-12-12 01:08:45 sqlalchemy.exc.ProgrammingError: (ProgrammingError) (1146, 
"Table 'ovs_neutron.agents' doesn't exist") 'ALTER TABLE agents ADD CONSTRAINT 
uniq_agents0agent_type0host UNIQUE (agent_type, host)' ()

** Affects: neutron
 Importance: Undecided

[Yahoo-eng-team] [Bug 1260262] [NEW] ml2 gre/vxlan type driver can lose segment_id allocation

2013-12-12 Thread Isaku Yamahata
Public bug reported:

With ML2 gre/vxlan type driver and multiple segment is specified, the
type driver can leak segment_id

** Affects: neutron
 Importance: Undecided
 Assignee: Isaku Yamahata (yamahata)
 Status: In Progress

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1260262

Title:
  ml2 gre/vxlan type driver can lose segment_id allocation

Status in OpenStack Neutron (virtual network service):
  In Progress

Bug description:
  With ML2 gre/vxlan type driver and multiple segment is specified, the
  type driver can leak segment_id

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1260262/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1252603] Re: PMTUD needs to be disabled for tunneling to work in many Grizzly environments.

2013-12-12 Thread James Page
Limiting scope to cloud-archive; raring actually has a sufficiently
recent version of openvswitch to not have this issue.


** Also affects: cloud-archive
   Importance: Undecided
   Status: New

** Changed in: quantum (Ubuntu)
   Status: New => Invalid

** Changed in: cloud-archive
   Status: New => Triaged

** Changed in: cloud-archive
   Importance: Undecided => Low

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1252603

Title:
  PMTUD needs to be disabled for tunneling to work in many Grizzly
  environments.

Status in Ubuntu Cloud Archive:
  Triaged
Status in OpenStack Neutron (virtual network service):
  Invalid
Status in “quantum” package in Ubuntu:
  Invalid

Bug description:
  In Grizzly, the version of OVS is lower than 1.9.0. As a result, tunnel path 
MTU
  discovery default value is set to 'enabled'. But internet-wide path MTU 
discovery
  rarely works, we need to add a configuration option to disable tunnel path MTU
  Discovery.

  Discussion about this issue:
Connectivity issue from within the Instances.
http://lists.openstack.org/pipermail/openstack/2013-August/000293.html

  Blog about this issue:
Path MTU discovery and GRE.
http://techbackground.blogspot.com/2013/06/path-mtu-discovery-and-gre.html

  Tunnel Path MTU Discovery default value was set to 'disabled' in OVS 1.9.0.
  Both inheritance of the Don't Fragment bit in IP tunnels (df_inherit) and path
  MTU discovery are no longer supported in OVS 1.10.0. So Linux distributions
  that ship with OVS >= 1.9.0 for Havana have no such issue.

To manage notifications about this bug go to:
https://bugs.launchpad.net/cloud-archive/+bug/1252603/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260249] Re: migration-list: 'unicode' object has no attribute 'iteritems'

2013-12-12 Thread sahid
** Also affects: nova
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1260249

Title:
  migration-list: 'unicode' object has no attribute 'iteritems'

Status in OpenStack Compute (Nova):
  New
Status in Python client library for Nova:
  New

Bug description:
  There is an AttributeError when we try to use the command "nova
  migration-list"

  Traceback (most recent call last):
File "/opt/stack/python-novaclient/novaclient/shell.py", line 721, in main
  OpenStackComputeShell().main(map(strutils.safe_decode, sys.argv[1:]))
File "/opt/stack/python-novaclient/novaclient/shell.py", line 657, in main
  args.func(self.cs, args)
File "/opt/stack/python-novaclient/novaclient/v1_1/contrib/migrations.py", 
line 71, in do_migration_list
  args.cell_name))
File "/opt/stack/python-novaclient/novaclient/v1_1/contrib/migrations.py", 
line 53, in list
  return self._list("/os-migrations%s" % query_string, "migrations")
File "/opt/stack/python-novaclient/novaclient/base.py", line 80, in _list
  for res in data if res]
File "/opt/stack/python-novaclient/novaclient/base.py", line 426, in 
__init__
  self._add_details(info)
File "/opt/stack/python-novaclient/novaclient/base.py", line 449, in 
_add_details
  for (k, v) in six.iteritems(info):
File "/usr/local/lib/python2.7/dist-packages/six.py", line 439, in iteritems
  return iter(getattr(d, _iteritems)(**kw))
  AttributeError: 'unicode' object has no attribute 'iteritems'
  ERROR: 'unicode' object has no attribute 'iteritems'

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1260249/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260265] [NEW] BaremetalHostManager cannot distinguish baremetal hosts from other hosts

2013-12-12 Thread Arata Notsu
Public bug reported:

BaremetalHostManager cloud distingush baremetal hosts by checking
"baremetal_driver" exists in capabilities or not. However, now
BaremetalHostManager cannot, because capabilities are not reported to
scheduler and BaremetalHostManager always receives empty capabilities.
As a result, BaremetalHostManager just does the same thing as the
original HostManager.

** Affects: nova
 Importance: Undecided
 Status: New


** Tags: baremetal

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1260265

Title:
  BaremetalHostManager cannot distinguish baremetal hosts from other
  hosts

Status in OpenStack Compute (Nova):
  New

Bug description:
  BaremetalHostManager cloud distingush baremetal hosts by checking
  "baremetal_driver" exists in capabilities or not. However, now
  BaremetalHostManager cannot, because capabilities are not reported to
  scheduler and BaremetalHostManager always receives empty capabilities.
  As a result, BaremetalHostManager just does the same thing as the
  original HostManager.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1260265/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260274] [NEW] NoVNC Console not showing in Internet Explorer

2013-12-12 Thread Jesse Pretorius
Public bug reported:

When accessing the NoVNC console through Internet Explorer (tested with
IE9 and IE10) the HTML5 Canvas never renders, instead showing 'Canvas
not supported'.

Environment:
 - OS: Ubuntu 12.04 LTS
 - Platform: Openstack Grizzly
 - Packages:
  nova-novncproxy 1:2013.1.3-0ubuntu1~cloud0
  novnc 2012.2~20120906+dfsg-0ubuntu4~cloud0
  python-novnc 2012.2~20120906+dfsg-0ubuntu4~cloud0

According to https://github.com/kanaka/noVNC/wiki/Browser-support NoVNC
should work with IE9 and above.

** Affects: nova
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1260274

Title:
  NoVNC Console not showing in Internet Explorer

Status in OpenStack Compute (Nova):
  New

Bug description:
  When accessing the NoVNC console through Internet Explorer (tested
  with IE9 and IE10) the HTML5 Canvas never renders, instead showing
  'Canvas not supported'.

  Environment:
   - OS: Ubuntu 12.04 LTS
   - Platform: Openstack Grizzly
   - Packages:
nova-novncproxy 1:2013.1.3-0ubuntu1~cloud0
novnc 2012.2~20120906+dfsg-0ubuntu4~cloud0
python-novnc 2012.2~20120906+dfsg-0ubuntu4~cloud0

  According to https://github.com/kanaka/noVNC/wiki/Browser-support
  NoVNC should work with IE9 and above.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1260274/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260281] [NEW] Rendering of dashboard is broken in Internet Explorer

2013-12-12 Thread Jesse Pretorius
Public bug reported:

In Internet Explorer (tested with IE9 and IE10) the rendering of various
dashboard components is broken.

 - Content section is shown below the left hand navigation menu most often, 
unless you have a super-wide screen
 - Network Topology network names do not display inside the vertical network bar
 - The rounded edges do not render
 - The buttons look funny

While I realise that some of these are due to differences in the way
that IE renders CSS we do feel that it's important to ensure that using
IE for Openstack End-Users and Administrators gives a reasonable
experience.

** Affects: horizon
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1260281

Title:
  Rendering of dashboard is broken in Internet Explorer

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  In Internet Explorer (tested with IE9 and IE10) the rendering of
  various dashboard components is broken.

   - Content section is shown below the left hand navigation menu most often, 
unless you have a super-wide screen
   - Network Topology network names do not display inside the vertical network 
bar
   - The rounded edges do not render
   - The buttons look funny

  While I realise that some of these are due to differences in the way
  that IE renders CSS we do feel that it's important to ensure that
  using IE for Openstack End-Users and Administrators gives a reasonable
  experience.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1260281/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260249] Re: migration-list: 'unicode' object has no attribute 'iteritems'

2013-12-12 Thread Joe Gordon
this looks like a novaclient bug only

** Changed in: nova
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1260249

Title:
  migration-list: 'unicode' object has no attribute 'iteritems'

Status in OpenStack Compute (Nova):
  Invalid
Status in Python client library for Nova:
  New

Bug description:
  There is an AttributeError when we try to use the command "nova
  migration-list"

  Traceback (most recent call last):
File "/opt/stack/python-novaclient/novaclient/shell.py", line 721, in main
  OpenStackComputeShell().main(map(strutils.safe_decode, sys.argv[1:]))
File "/opt/stack/python-novaclient/novaclient/shell.py", line 657, in main
  args.func(self.cs, args)
File "/opt/stack/python-novaclient/novaclient/v1_1/contrib/migrations.py", 
line 71, in do_migration_list
  args.cell_name))
File "/opt/stack/python-novaclient/novaclient/v1_1/contrib/migrations.py", 
line 53, in list
  return self._list("/os-migrations%s" % query_string, "migrations")
File "/opt/stack/python-novaclient/novaclient/base.py", line 80, in _list
  for res in data if res]
File "/opt/stack/python-novaclient/novaclient/base.py", line 426, in 
__init__
  self._add_details(info)
File "/opt/stack/python-novaclient/novaclient/base.py", line 449, in 
_add_details
  for (k, v) in six.iteritems(info):
File "/usr/local/lib/python2.7/dist-packages/six.py", line 439, in iteritems
  return iter(getattr(d, _iteritems)(**kw))
  AttributeError: 'unicode' object has no attribute 'iteritems'
  ERROR: 'unicode' object has no attribute 'iteritems'

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1260249/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260301] [NEW] Expired vnc token doesn't fail fast

2013-12-12 Thread Loganathan Parthipan
Public bug reported:

If an expired vnc token is used, I'd expect an appropriate HTTP
response. However, currently novnc just hangs at "Starting VNC
handshake".

To reproduce

1. nova get-vnc-console  novnc => returns a URI with the token
2. point browser to the returned URI after the expiry
3. page hangs at "Starting VNC handshake"

Expected response

An appropriate HTTP response

** Affects: nova
 Importance: Undecided
 Status: New

** Summary changed:

- Expired vnc token doesn't result in an HTTP 401
+ Expired vnc token doesn't fail fast

** Description changed:

- If an expired vnc token is used, I'd expect an HTTP 401. However,
- currently novnc just hangs at "Starting VNC handshake".
+ If an expired vnc token is used, I'd expect an appropriate HTTP
+ response. However, currently novnc just hangs at "Starting VNC
+ handshake".
  
  To reproduce
  
  1. nova get-vnc-console  novnc => returns a URI with the token
  2. point browser to the returned URI after the expiry
  3. page hangs at "Starting VNC handshake"
  
  Expected response
  
- HTTP 401
+ An appropriate HTTP response

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1260301

Title:
  Expired vnc token doesn't fail fast

Status in OpenStack Compute (Nova):
  New

Bug description:
  If an expired vnc token is used, I'd expect an appropriate HTTP
  response. However, currently novnc just hangs at "Starting VNC
  handshake".

  To reproduce

  1. nova get-vnc-console  novnc => returns a URI with the token
  2. point browser to the returned URI after the expiry
  3. page hangs at "Starting VNC handshake"

  Expected response

  An appropriate HTTP response

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1260301/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1253896] Re: test_minimum_basic_scenario fails with SSHException: Error reading SSH protocol banner

2013-12-12 Thread Sean Dague
This bug is not invalid for neutron, it's unfortunate that jog0 said
that earlier, as neutron is exposing this more often than nova-network.

** Changed in: neutron
   Importance: High => Critical

** Changed in: neutron
   Status: Invalid => Confirmed

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1253896

Title:
  test_minimum_basic_scenario fails with SSHException: Error reading SSH
  protocol banner

Status in OpenStack Neutron (virtual network service):
  Confirmed
Status in OpenStack Compute (Nova):
  Triaged
Status in Tempest:
  Confirmed

Bug description:
  An example of this can be found at
  http://logs.openstack.org/74/57774/2/gate/gate-tempest-devstack-vm-
  full/e592961/console.html. This test failing seems to cause the
  tearDownClass failure and the process exit code failure.

  Judging by the logs below, the VM is coming up, and the test is
  connecting to the SSH server (dropbear) running in the VM, but the
  authentication is failing. It appears that authentication is attempted
  several times before paramiko gives up causing the test to fail. I
  think this indicates there isn't a network or compute problem, instead
  is possible the client doesn't have the correct key or the authorized
  keys aren't configured properly on the server side. But these are just
  guesses, I haven't been able to get any concrete data that would
  support these theories.

  2013-11-22 05:36:33.980 | 2013-11-22 05:32:17,029 Adding  to shared resources of 
TestMinimumBasicScenario
  2013-11-22 05:36:33.980 | 2013-11-22 05:32:34,226 starting thread (client 
mode): 0x52e7e50L
  2013-11-22 05:36:33.980 | 2013-11-22 05:32:34,232 Connected (version 2.0, 
client dropbear_2012.55)
  2013-11-22 05:36:33.981 | 2013-11-22 05:32:34,237 kex 
algos:['diffie-hellman-group1-sha1', 'diffie-hellman-group14-sha1'] server 
key:['ssh-rsa', 'ssh-dss'] client encrypt:['aes128-ctr', '3des-ctr', 
'aes256-ctr', 'aes128-cbc', '3des-cbc', 'aes256-cbc', 'twofish256-cbc', 
'twofish-cbc', 'twofish128-cbc'] server encrypt:['aes128-ctr', '3des-ctr', 
'aes256-ctr', 'aes128-cbc', '3des-cbc', 'aes256-cbc', 'twofish256-cbc', 
'twofish-cbc', 'twofish128-cbc'] client mac:['hmac-sha1-96', 'hmac-sha1', 
'hmac-md5'] server mac:['hmac-sha1-96', 'hmac-sha1', 'hmac-md5'] client 
compress:['none'] server compress:['none'] client lang:[''] server lang:[''] 
kex follows?False
  2013-11-22 05:36:33.981 | 2013-11-22 05:32:34,238 Ciphers agreed: 
local=aes128-ctr, remote=aes128-ctr
  2013-11-22 05:36:33.981 | 2013-11-22 05:32:34,238 using kex 
diffie-hellman-group1-sha1; server key type ssh-rsa; cipher: local aes128-ctr, 
remote aes128-ctr; mac: local hmac-sha1, remote hmac-sha1; compression: local 
none, remote none
  2013-11-22 05:36:33.981 | 2013-11-22 05:32:34,433 Switch to new keys ...
  2013-11-22 05:36:33.982 | 2013-11-22 05:32:34,434 Adding ssh-rsa host key for 
172.24.4.227: 189c16acb93fe44ae975e1c653f1856c
  2013-11-22 05:36:33.982 | 2013-11-22 05:32:34,434 Trying SSH key 
9a9afe52a9485c15495a59b94ebca6b6
  2013-11-22 05:36:33.982 | 2013-11-22 05:32:34,437 userauth is OK
  2013-11-22 05:36:33.982 | 2013-11-22 05:32:35,104 Authentication (publickey) 
failed.
  2013-11-22 05:36:33.982 | 2013-11-22 05:32:36,693 starting thread (client 
mode): 0x52f9190L
  2013-11-22 05:36:33.982 | 2013-11-22 05:32:36,697 Connected (version 2.0, 
client dropbear_2012.55)
  2013-11-22 05:36:33.983 | 2013-11-22 05:32:36,699 kex 
algos:['diffie-hellman-group1-sha1', 'diffie-hellman-group14-sha1'] server 
key:['ssh-rsa', 'ssh-dss'] client encrypt:['aes128-ctr', '3des-ctr', 
'aes256-ctr', 'aes128-cbc', '3des-cbc', 'aes256-cbc', 'twofish256-cbc', 
'twofish-cbc', 'twofish128-cbc'] server encrypt:['aes128-ctr', '3des-ctr', 
'aes256-ctr', 'aes128-cbc', '3des-cbc', 'aes256-cbc', 'twofish256-cbc', 
'twofish-cbc', 'twofish128-cbc'] client mac:['hmac-sha1-96', 'hmac-sha1', 
'hmac-md5'] server mac:['hmac-sha1-96', 'hmac-sha1', 'hmac-md5'] client 
compress:['none'] server compress:['none'] client lang:[''] server lang:[''] 
kex follows?False
  2013-11-22 05:36:33.983 | 2013-11-22 05:32:36,699 Ciphers agreed: 
local=aes128-ctr, remote=aes128-ctr
  2013-11-22 05:36:33.983 | 2013-11-22 05:32:36,699 using kex 
diffie-hellman-group1-sha1; server key type ssh-rsa; cipher: local aes128-ctr, 
remote aes128-ctr; mac: local hmac-sha1, remote hmac-sha1; compression: local 
none, remote none
  2013-11-22 05:36:33.983 | 2013-11-22 05:32:36,903 Switch to new keys ...
  2013-11-22 05:36:33.983 | 2013-11-22 05:32:36,904 Trying SSH key 
9a9afe52a9485c15495a59b94ebca6b6
  2013-11-22 05:36:33.984 | 2013-11-22 05:32:36,906 userauth is OK
  2013-11-22 05:36:33.984 | 2013-11-22 05:32:37,438 Authentication (publickey) 
failed.
  2013-11-22 05:36:33.984 | 2013-11-22 05:32:39,035 starting thread (client 
mode): 0x24c62d0L
  2013-11-22 05:36:33.984 | 2013-11-22 05:32:39,043 Connected (version 2.0, 

[Yahoo-eng-team] [Bug 1260310] [NEW] gate-tempest-dsvm-full failure with "An error occurred while enabling hairpin mode on domain with xml"

2013-12-12 Thread Abhishek Chanda
Public bug reported:

Kibana search

http://logstash.openstack.org/#eyJzZWFyY2giOiJtZXNzYWdlOiBcIkFuIGVycm9yIG9jY3VycmVkIHdoaWxlIGVuYWJsaW5nIGhhaXJwaW4gbW9kZSBvbiBkb21haW4gd2l0aCB4bWxcIiIsImZpZWxkcyI6W10sIm9mZnNldCI6MCwidGltZWZyYW1lIjoiNjA0ODAwIiwiZ3JhcGhtb2RlIjoiY291bnQiLCJ0aW1lIjp7InVzZXJfaW50ZXJ2YWwiOjB9LCJzdGFtcCI6MTM4Njg1MjgzMTk3M30=

This is pretty infrequent

** Affects: nova
 Importance: High
 Status: Triaged

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1260310

Title:
  gate-tempest-dsvm-full failure with "An error occurred while enabling
  hairpin mode on domain with xml"

Status in OpenStack Compute (Nova):
  Triaged

Bug description:
  Kibana search

  
http://logstash.openstack.org/#eyJzZWFyY2giOiJtZXNzYWdlOiBcIkFuIGVycm9yIG9jY3VycmVkIHdoaWxlIGVuYWJsaW5nIGhhaXJwaW4gbW9kZSBvbiBkb21haW4gd2l0aCB4bWxcIiIsImZpZWxkcyI6W10sIm9mZnNldCI6MCwidGltZWZyYW1lIjoiNjA0ODAwIiwiZ3JhcGhtb2RlIjoiY291bnQiLCJ0aW1lIjp7InVzZXJfaW50ZXJ2YWwiOjB9LCJzdGFtcCI6MTM4Njg1MjgzMTk3M30=

  This is pretty infrequent

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1260310/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260314] [NEW] glance image-create with invalid store fails but still creates image

2013-12-12 Thread John Lenihan
Public bug reported:

glance checks whether or not a specified store is valid, but if it is
invalid the image has already been created.

I pulled the latest  devstack code and then ran these commands after
sourcing openrc:

ubuntu@devstack-glance:/mnt/devstack$ glance index
ID   Name   Disk Format 
 Container Format Size  
 -- 
  --
6792e9a7-f4f8-48cb-b407-80e360b8a773 cirros-0.3.1-x86_64-uecami 
 ami25165824
7808c034-3fdd-4975-af26-e7d5a15d2113 cirros-0.3.1-x86_64-uec-ramdis ari 
 ari 3714968
4efcddb2-9f20-413f-86a3-3bf69455e09b cirros-0.3.1-x86_64-uec-kernel aki 
 aki 4955792
ubuntu@devstack-glance:/mnt/devstack$ 
ubuntu@devstack-glance:/mnt/devstack$ glance -d image-create --store s3e 
--disk-format raw --container-format bare --name complete_gibberish ' -H 'Content-Type: application/octet-stream' -H 
'x-image-meta-disk_format: raw' -H 'x-image-meta-name: complete_gibberish' -d 
'', mode 'r' at 0x7f16181b6150>' 
http://10.4.36.1:9292/v1/images

HTTP/1.1 400 Bad Request
date: Thu, 12 Dec 2013 12:47:37 GMT
content-length: 52
content-type: text/plain; charset=UTF-8
x-openstack-request-id: req-c9bad6ee-d79c-41f3-bd96-d3929afd742c

400 Bad Request

Store for scheme s3e not found


Request returned failure status.
400 Bad Request
Store for scheme s3e not found
(HTTP 400)
ubuntu@devstack-glance:/mnt/devstack$ glance index
ID   Name   Disk Format 
 Container Format Size  
 -- 
  --
b26c03e4-7cdf-44fe-9187-7de315c9b38b complete_gibberish raw 
 bare221
6792e9a7-f4f8-48cb-b407-80e360b8a773 cirros-0.3.1-x86_64-uecami 
 ami25165824
7808c034-3fdd-4975-af26-e7d5a15d2113 cirros-0.3.1-x86_64-uec-ramdis ari 
 ari 3714968
4efcddb2-9f20-413f-86a3-3bf69455e09b cirros-0.3.1-x86_64-uec-kernel aki 
 aki 4955792

This problem occurs using the v1 API. If using the V2 API the '--store'
option does not seem to be present.

** Affects: glance
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Glance.
https://bugs.launchpad.net/bugs/1260314

Title:
  glance image-create with invalid store fails but still creates image

Status in OpenStack Image Registry and Delivery Service (Glance):
  New

Bug description:
  glance checks whether or not a specified store is valid, but if it is
  invalid the image has already been created.

  I pulled the latest  devstack code and then ran these commands after
  sourcing openrc:

  ubuntu@devstack-glance:/mnt/devstack$ glance index
  ID   Name   Disk 
Format  Container Format Size  
   -- 
  --
  6792e9a7-f4f8-48cb-b407-80e360b8a773 cirros-0.3.1-x86_64-uecami   
   ami25165824
  7808c034-3fdd-4975-af26-e7d5a15d2113 cirros-0.3.1-x86_64-uec-ramdis ari   
   ari 3714968
  4efcddb2-9f20-413f-86a3-3bf69455e09b cirros-0.3.1-x86_64-uec-kernel aki   
   aki 4955792
  ubuntu@devstack-glance:/mnt/devstack$ 
  ubuntu@devstack-glance:/mnt/devstack$ glance -d image-create --store s3e 
--disk-format raw --container-format bare --name complete_gibberish ' -H 'Content-Type: 
application/octet-stream' -H 'x-image-meta-disk_format: raw' -H 
'x-image-meta-name: complete_gibberish' -d '', mode 'r' at 
0x7f16181b6150>' http://10.4.36.1:9292/v1/images

  HTTP/1.1 400 Bad Request
  date: Thu, 12 Dec 2013 12:47:37 GMT
  content-length: 52
  content-type: text/plain; charset=UTF-8
  x-openstack-request-id: req-c9bad6ee-d79c-41f3-bd96-d3929afd742c

  400 Bad Request

  Store for scheme s3e not found


  Request returned failure status.
  400 Bad Request
  Store for scheme s3e not found
  (HTTP 400)
  ubuntu@devstack-glance:/mnt/devstack$ glance index
  ID   Name   Disk 
Format  Container Format Size  
   -- 
  --
  b26c03e4-7cdf-44fe-9187-7de315c9b38b complete_gibberish raw   
   bare  

[Yahoo-eng-team] [Bug 1259894] Re: tempest.api.compute.v3.servers.test_server_rescue.ServerRescueV3TestJSON.test_rescue_unrescue_instance FAILED

2013-12-12 Thread Attila Fazekas
It looks like a nova issue, adding nova to the afflicted projects.

** Also affects: nova
   Importance: Undecided
   Status: New

** Changed in: tempest
   Status: New => Incomplete

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1259894

Title:
  
tempest.api.compute.v3.servers.test_server_rescue.ServerRescueV3TestJSON.test_rescue_unrescue_instance
  FAILED

Status in OpenStack Compute (Nova):
  New
Status in Tempest:
  Incomplete

Bug description:
  
  Traceback (most recent call last):
File "tempest/api/compute/v3/servers/test_server_rescue.py", line 98, in 
test_rescue_unrescue_instance
  self.servers_client.wait_for_server_status(self.server_id, 'RESCUE')
File "tempest/services/compute/v3/json/servers_client.py", line 167, in 
wait_for_server_status
  extra_timeout=extra_timeout)
File "tempest/common/waiters.py", line 82, in wait_for_server_status
  raise exceptions.TimeoutException(message)
  TimeoutException: Request timed out
  Details: Server 9c38fcf3-50ab-4a19-86e3-5c0fab501fca failed to reach RESCUE 
status within the required time (196 s). Current status: ACTIVE.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1259894/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1257460] Re: test_rescued_vm_detach_volume Volume test_detach failed to reach in-use status within the required time (196 s).

2013-12-12 Thread Attila Fazekas
Is tempest did anything incorrectly ?

** Also affects: nova
   Importance: Undecided
   Status: New

** Also affects: cinder
   Importance: Undecided
   Status: New

** Changed in: tempest
   Status: New => Incomplete

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1257460

Title:
  test_rescued_vm_detach_volume Volume test_detach failed to reach in-
  use status within the required time (196 s).

Status in Cinder:
  New
Status in OpenStack Compute (Nova):
  New
Status in Tempest:
  Incomplete

Bug description:
  http://logs.openstack.org/76/59576/3/check/check-tempest-dsvm-
  postgres-full/af0e4e2/console.html

  
  2013-12-03 01:57:24.699 | 
==
  2013-12-03 01:57:24.699 | FAIL: 
tempest.api.compute.servers.test_server_rescue.ServerRescueTestJSON.test_rescued_vm_detach_volume[gate,negative]
  2013-12-03 01:57:24.699 | 
tempest.api.compute.servers.test_server_rescue.ServerRescueTestJSON.test_rescued_vm_detach_volume[gate,negative]
  2013-12-03 01:57:24.699 | 
--
  2013-12-03 01:57:24.700 | _StringException: Empty attachments:
  2013-12-03 01:57:24.700 |   stderr
  2013-12-03 01:57:24.700 |   stdout
  2013-12-03 01:57:24.701 | 
  2013-12-03 01:57:24.701 | pythonlogging:'': {{{
  2013-12-03 01:57:24.701 | 2013-12-03 01:34:31,546 Request: POST 
http://127.0.0.1:8774/v2/742879cfcd384dffb721c692c81376be/servers/ab993087-a194-4798-ac63-3b5495690fb6/os-volume_attachments
  2013-12-03 01:57:24.701 | 2013-12-03 01:34:31,546 Request Headers: 
{'Content-Type': 'application/json', 'Accept': 'application/json', 
'X-Auth-Token': ''}
  2013-12-03 01:57:24.701 | 2013-12-03 01:34:31,546 Request Body: 
{"volumeAttachment": {"device": "/dev/vdf", "volumeId": 
"67bc3c74-1d66-4107-9d6d-f66b6a678e52"}}
  2013-12-03 01:57:24.701 | 2013-12-03 01:34:33,730 Response Status: 200
  2013-12-03 01:57:24.702 | 2013-12-03 01:34:33,730 Nova request id: 
req-4314d02e-0522-4410-9e6c-909528fb2314
  2013-12-03 01:57:24.702 | 2013-12-03 01:34:33,731 Response Headers: 
{'content-length': '194', 'date': 'Tue, 03 Dec 2013 01:34:33 GMT', 
'content-type': 'application/json', 'connection': 'close'}
  2013-12-03 01:57:24.702 | 2013-12-03 01:34:33,731 Response Body: 
{"volumeAttachment": {"device": "/dev/vdf", "serverId": 
"ab993087-a194-4798-ac63-3b5495690fb6", "id": 
"67bc3c74-1d66-4107-9d6d-f66b6a678e52", "volumeId": 
"67bc3c74-1d66-4107-9d6d-f66b6a678e52"}}
  2013-12-03 01:57:24.702 | 2013-12-03 01:34:33,731 Request: GET 
http://127.0.0.1:8774/v2/742879cfcd384dffb721c692c81376be/os-volumes/67bc3c74-1d66-4107-9d6d-f66b6a678e52
  2013-12-03 01:57:24.702 | 2013-12-03 01:34:33,732 Request Headers: 
{'X-Auth-Token': ''}
  2013-12-03 01:57:24.702 | 2013-12-03 01:34:34,025 Response Status: 200
  2013-12-03 01:57:24.702 | 2013-12-03 01:34:34,026 Nova request id: 
req-4a9aac2a-55cb-4287-8f4c-392d3498a304

  

  2013-12-03 01:57:24.871 | Traceback (most recent call last):
  2013-12-03 01:57:24.871 |   File 
"tempest/api/compute/servers/test_server_rescue.py", line 173, in 
test_rescued_vm_detach_volume
  2013-12-03 01:57:24.871 | self.volume_to_detach['id'], 'in-use')
  2013-12-03 01:57:24.872 |   File 
"tempest/services/compute/json/volumes_extensions_client.py", line 104, in 
wait_for_volume_status
  2013-12-03 01:57:24.872 | raise exceptions.TimeoutException(message)
  2013-12-03 01:57:24.872 | TimeoutException: Request timed out
  2013-12-03 01:57:24.872 | Details: Volume test_detach failed to reach in-use 
status within the required time (196 s).

To manage notifications about this bug go to:
https://bugs.launchpad.net/cinder/+bug/1257460/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260138] Re: VMWARE: Unable to spawn instances from sparse/ide images

2013-12-12 Thread Abhishek Chanda
Duplicate of #1260139

** Changed in: nova
   Status: New => Won't Fix

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1260138

Title:
  VMWARE: Unable to spawn instances from sparse/ide images

Status in OpenStack Compute (Nova):
  Won't Fix

Bug description:
  Branch: stable/havana

  Traceback: http://paste.openstack.org/show/54855/

  Steps to reprodude:
  Upload a ide/sparse type image to glance.
  Spawn an instance from that image

  Actual Result:
  Failed to spawn an image

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1260138/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1182883] Re: List servers matching a regex fails with Quantum

2013-12-12 Thread Sean Dague
** Changed in: tempest
   Status: Confirmed => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1182883

Title:
  List servers matching a regex fails with Quantum

Status in OpenStack Neutron (virtual network service):
  Invalid
Status in OpenStack Compute (Nova):
  Invalid
Status in Tempest:
  Invalid

Bug description:
  The test
  
tempest.api.compute.servers.test_list_server_filters:ListServerFiltersTestXML.test_list_servers_filtered_by_ip_regex
  tries to search a server with only a fragment of its IP (GET
  http://XX/v2/$Tenant/servers?ip=10.0.) which calls the following
  Quantum request :
  http://XX/v2.0/ports.json?fixed_ips=ip_address%3D10.0. But it seems
  this regex search is not supporter by Quantum. Thus the tempest test
  fauls.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1182883/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1195473] Re: Need some docs about Glance notifications with examples

2013-12-12 Thread Brian Rosmaita
** No longer affects: glance

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Glance.
https://bugs.launchpad.net/bugs/1195473

Title:
  Need some docs about Glance notifications with examples

Status in OpenStack Manuals:
  Confirmed

Bug description:
  The notifications emitted by Glance were enhanced for Grizzly.  The
  developer docs have been updated, but we could probably use a listing
  of what notifications are emitted and what they look like for the
  operator manual.

To manage notifications about this bug go to:
https://bugs.launchpad.net/openstack-manuals/+bug/1195473/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1244055] Re: six has no attribute 'add_metaclass'

2013-12-12 Thread Sean Dague
** Changed in: tempest
   Status: Confirmed => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1244055

Title:
  six has no attribute 'add_metaclass'

Status in OpenStack Compute (Nova):
  Invalid
Status in Tempest:
  Fix Released

Bug description:
  I have a patch failing in gate with traces containing the following:

  2013-10-23 22:27:54.336 |   File 
"/opt/stack/new/python-novaclient/novaclient/base.py", line 166, in 
  2013-10-23 22:27:54.336 | @six.add_metaclass(abc.ABCMeta)
  2013-10-23 22:27:54.337 | AttributeError: 'module' object has no attribute 
'add_metaclass'

  For full logs, see the failing patch:
  https://review.openstack.org/#/c/52876/

  It looks like this was caused by this recent commit:
  https://review.openstack.org/#/c/52255/

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1244055/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1124674] Re: add-fixed-ip causes traceback

2013-12-12 Thread Sean Dague
** Changed in: tempest
   Status: Confirmed => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1124674

Title:
  add-fixed-ip causes traceback

Status in OpenStack Compute (Nova):
  Fix Released
Status in Tempest:
  Invalid

Bug description:
  nova add-fixed-ip leads to a traceback in nova-network
  2013-02-13 16:10:57.711 ERROR nova.openstack.common.rpc.amqp 
[req-434b8722-566c-4345-933a-5646277cd6ef demo demo] Exception during message 
handling2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp Traceback 
(most recent call last):2013-02-13 16:10:57.711 TRACE 
nova.openstack.common.rpc.amqp   File 
"/opt/stack/nova/nova/openstack/common/rpc/amqp.py", line 276, in 
_process_data2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp 
rval = self.proxy.dispatch(ctxt, version, method, **args)2013-02-13 
16:10:57.711 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/nova/nova/openstack/common/rpc/dispatcher.py", line 133, in 
dispatch2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp return 
getattr(proxyobj, method)(ctxt, **kwargs)2013-02-13 16:10:57.711 TRACE 
nova.openstack.common.rpc.amqp   File 
"/opt/stack/nova/nova/network/manager.py", line 756, in 
add_fixed_ip_to_instance2013-02-13 16:10:57.711 TRACE 
nova.openstack.common.rpc.amqp self._alloc
 ate_fixed_ips(context, instance_id, host, [network])2013-02-13 16:10:57.711 
TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/nova/nova/network/manager.py", line 212, in _allocate_fixed_ips
  2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp vpn=vpn, 
address=address)
  2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/nova/nova/network/manager.py", line 801, in allocate_fixed_ip
  2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp instance_ref 
= self.db.instance_get(context, instance_id)
  2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/nova/nova/db/api.py", line 593, in instance_get
  2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp return 
IMPL.instance_get(context, instance_id)
  2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/nova/nova/db/sqlalchemy/api.py", line 137, in wrapper
  2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp return 
f(*args, **kwargs)
  2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/nova/nova/db/sqlalchemy/api.py", line 1520, in instance_get
  2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp raise 
exception.InstanceNotFound(instance_id=instance_id)
  2013-02-13 16:10:57.711 TRACE nova.openstack.common.rpc.amqp 
InstanceNotFound: Instance f96d5c44-6c17-46cb-8b5c-057717ce076d could not be 
found.

  there seems to be quite a bit of mixing of instance['id'] with
  instance['uuid']

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1124674/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1213212] Re: test_resize_server_confirm server failed to build

2013-12-12 Thread Sean Dague
removing this as a tempest issue, as I don't think it actually is a bug
in tempest, it's a nova state bug

** Changed in: nova
   Status: New => Confirmed

** Changed in: nova
   Importance: Undecided => Medium

** No longer affects: tempest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1213212

Title:
  test_resize_server_confirm server failed to build

Status in OpenStack Compute (Nova):
  Confirmed

Bug description:
  When running tempest in parallel occasionally
  test_resize_server_confirm fails to build the server and goes into an
  error state see:

  2013-08-16 14:08:33.607 | 
==
  2013-08-16 14:08:33.607 | FAIL: 
tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_resize_server_confirm[gate,smoke]
  2013-08-16 14:08:33.607 | 
tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_resize_server_confirm[gate,smoke]
  2013-08-16 14:08:33.608 | 
--
  2013-08-16 14:08:33.608 | _StringException: Empty attachments:
  2013-08-16 14:08:33.608 |   stderr
  2013-08-16 14:08:33.609 |   stdout
  2013-08-16 14:08:33.609 | 
  2013-08-16 14:08:33.609 | Traceback (most recent call last):
  2013-08-16 14:08:33.609 |   File 
"tempest/api/compute/servers/test_server_actions.py", line 161, in 
test_resize_server_confirm
  2013-08-16 14:08:33.609 | 
self.client.wait_for_server_status(self.server_id, 'VERIFY_RESIZE')
  2013-08-16 14:08:33.609 |   File 
"tempest/services/compute/json/servers_client.py", line 165, in 
wait_for_server_status
  2013-08-16 14:08:33.609 | raise 
exceptions.BuildErrorException(server_id=server_id)
  2013-08-16 14:08:33.610 | BuildErrorException: Server 
ed3c7212-f4b6-4365-91b8-bc9e1a60 failed to build and is in ERROR status
  2013-08-16 14:08:33.610 | 
  2013-08-16 14:08:33.610 | 
  2013-08-16 14:08:33.611 | 
==

  A set of logs for this failure can be found here:
  
http://logs.openstack.org/63/42063/1/gate/gate-tempest-devstack-vm-testr-full/fa32f42/

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1213212/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1213209] Re: test_list_image_filters.py setUpClass created image never becomes active

2013-12-12 Thread Sean Dague
*** This bug is a duplicate of bug 1258635 ***
https://bugs.launchpad.net/bugs/1258635

** This bug has been marked a duplicate of bug 1258635
   Race with changing image status when snapshotting

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1213209

Title:
  test_list_image_filters.py setUpClass created image never becomes
  active

Status in OpenStack Compute (Nova):
  New
Status in Tempest:
  Confirmed

Bug description:
  When running in parallel occassionally the tests in
  test_list_image_filters fail while waiting for one of the created
  images to become active. See the logs here:

  http://logs.openstack.org/42/40342/2/gate/gate-tempest-devstack-vm-
  testr-full/dad876b/

  
  From the tempest log:

  2013-08-16 05:51:36.930 368 ERROR 
tempest.api.compute.images.test_list_image_filters [-] Request timed out
  2013-08-16 05:51:36.930 368 TRACE 
tempest.api.compute.images.test_list_image_filters Traceback (most recent call 
last):
  2013-08-16 05:51:36.930 368 TRACE 
tempest.api.compute.images.test_list_image_filters   File 
"tempest/api/compute/images/test_list_image_filters.py", line 67, in setUpClass
  2013-08-16 05:51:36.930 368 TRACE 
tempest.api.compute.images.test_list_image_filters 
cls.client.wait_for_image_status(cls.image2_id, 'ACTIVE')
  2013-08-16 05:51:36.930 368 TRACE 
tempest.api.compute.images.test_list_image_filters   File 
"tempest/services/compute/json/images_client.py", line 110, in 
wait_for_image_status
  2013-08-16 05:51:36.930 368 TRACE 
tempest.api.compute.images.test_list_image_filters raise 
exceptions.TimeoutException
  2013-08-16 05:51:36.930 368 TRACE 
tempest.api.compute.images.test_list_image_filters TimeoutException: Request 
timed out
  2013-08-16 05:51:36.930 368 TRACE 
tempest.api.compute.images.test_list_image_filters

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1213209/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1132879] Re: server reboot hard and rebuild are flaky in tempest when ssh is enabled

2013-12-12 Thread Adalberto Medeiros
** Changed in: tempest
   Status: Confirmed => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1132879

Title:
  server reboot hard and rebuild are flaky in tempest when ssh is
  enabled

Status in OpenStack Compute (Nova):
  Confirmed
Status in Tempest:
  Invalid

Bug description:
  Working on enabling back ssh access to VMs in tempest tests:

  https://review.openstack.org/#/c/22415/
  https://blueprints.launchpad.net/tempest/+spec/ssh-auth-strategy

  On the gate devstack with nova networking the hard reboot and rebuild
  test are sometimes passing and sometimes not.

  On the gate devstack with quantum networking the hard reboot and
  rebuild tests are systematically not passing, and blocking the overall
  blueprint implementation.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1132879/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260333] [NEW] Malformed property protection rules return error to end user

2013-12-12 Thread Thomas Leaman
Public bug reported:

Using a property protections file such as:

[.*]
create = @,! 
read = @
update = @
delete = @

The create operation has an invalid rule, duplicate values are not
allowed. This should probably result in the service refusing to start,
however currently the service will start and operations touching this
value will return:

500 Internal Server Error
Malformed property protection rule 'some_property': '@' and '!' are mutually 
exclusive   (HTTP 500)

to the end user. My feeling is that the end user should not receive any
information about the cause of the error, just the 500 status.

** Affects: glance
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Glance.
https://bugs.launchpad.net/bugs/1260333

Title:
  Malformed property protection rules return error to end user

Status in OpenStack Image Registry and Delivery Service (Glance):
  New

Bug description:
  Using a property protections file such as:

  [.*]
  create = @,! 
  read = @
  update = @
  delete = @

  The create operation has an invalid rule, duplicate values are not
  allowed. This should probably result in the service refusing to start,
  however currently the service will start and operations touching this
  value will return:

  500 Internal Server Error
  Malformed property protection rule 'some_property': '@' and '!' are mutually 
exclusive   (HTTP 500)

  to the end user. My feeling is that the end user should not receive
  any information about the cause of the error, just the 500 status.

To manage notifications about this bug go to:
https://bugs.launchpad.net/glance/+bug/1260333/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1254752] Re: test_volume_boot_pattern: SSH timed out

2013-12-12 Thread Attila Fazekas
Is there any sign of tempest did something incorrectly ?

** Also affects: neutron
   Importance: Undecided
   Status: New

** Changed in: tempest
   Status: New => Incomplete

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1254752

Title:
  test_volume_boot_pattern: SSH timed out

Status in OpenStack Neutron (virtual network service):
  New
Status in Tempest:
  Incomplete

Bug description:
  The test_volume_boot_pattern tests fails sporadically:

  http://logs.openstack.org/97/57797/2/check/check-tempest-devstack-vm-
  neutron/38fdc5a/console.html.gz

  2013-11-22 21:53:07.670 | Traceback (most recent call last):
  2013-11-22 21:53:07.670 |   File 
"tempest/scenario/test_volume_boot_pattern.py", line 156, in 
test_volume_boot_pattern
  2013-11-22 21:53:07.670 | ssh_client = 
self._ssh_to_server(instance_from_snapshot, keypair)
  2013-11-22 21:53:07.670 |   File 
"tempest/scenario/test_volume_boot_pattern.py", line 100, in _ssh_to_server
  2013-11-22 21:53:07.670 | private_key=keypair.private_key)
  2013-11-22 21:53:07.670 |   File "tempest/scenario/manager.py", line 475, in 
get_remote_client
  2013-11-22 21:53:07.671 | return RemoteClient(ip, username, 
pkey=private_key)
  2013-11-22 21:53:07.671 |   File 
"tempest/common/utils/linux/remote_client.py", line 47, in __init__
  2013-11-22 21:53:07.671 | if not self.ssh_client.test_connection_auth():
  2013-11-22 21:53:07.671 |   File "tempest/common/ssh.py", line 148, in 
test_connection_auth
  2013-11-22 21:53:07.671 | connection = self._get_ssh_connection()
  2013-11-22 21:53:07.672 |   File "tempest/common/ssh.py", line 76, in 
_get_ssh_connection
  2013-11-22 21:53:07.672 | password=self.password)
  2013-11-22 21:53:07.672 | SSHTimeout: Connection to the 172.24.4.230 via SSH 
timed out.
  2013-11-22 21:53:07.672 | User: cirros, Password: None

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1254752/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1251920] Re: Tempest failures due to failure to return console logs from an instance

2013-12-12 Thread Adalberto Medeiros
Since this is not hitting on tempest anymore, moving to Fix Released

** Changed in: tempest
   Status: Fix Committed => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1251920

Title:
  Tempest failures due to failure to return console logs from an
  instance

Status in OpenStack Compute (Nova):
  Invalid
Status in OpenStack Compute (nova) havana series:
  Fix Committed
Status in Tempest:
  Fix Released

Bug description:
  Logstash search:
  
http://logstash.openstack.org/#eyJzZWFyY2giOiJmaWxlbmFtZTpjb25zb2xlLmh0bWwgQU5EIG1lc3NhZ2U6XCJhc3NlcnRpb25lcnJvcjogY29uc29sZSBvdXRwdXQgd2FzIGVtcHR5XCIiLCJmaWVsZHMiOltdLCJvZmZzZXQiOjAsInRpbWVmcmFtZSI6IjYwNDgwMCIsImdyYXBobW9kZSI6ImNvdW50IiwidGltZSI6eyJ1c2VyX2ludGVydmFsIjowfSwic3RhbXAiOjEzODQ2NDEwNzIxODl9

  An example failure is http://logs.openstack.org/92/55492/8/check
  /check-tempest-devstack-vm-full/ef3a4a4/console.html

  console.html
  ===

  2013-11-16 21:54:27.998 | 2013-11-16 21:41:20,775 Request: POST 
http://127.0.0.1:8774/v2/3f6934d9aabf467aa8bc51397ccfa782/servers/10aace14-23c1-4cec-9bfd-2c873df1fbee/action
  2013-11-16 21:54:27.998 | 2013-11-16 21:41:20,776 Request Headers: 
{'Content-Type': 'application/json', 'Accept': 'application/json', 
'X-Auth-Token': ''}
  2013-11-16 21:54:27.998 | 2013-11-16 21:41:20,776 Request Body: 
{"os-getConsoleOutput": {"length": 10}}
  2013-11-16 21:54:27.998 | 2013-11-16 21:41:21,000 Response Status: 200
  2013-11-16 21:54:27.999 | 2013-11-16 21:41:21,001 Nova request id: 
req-7a2ee0ab-c977-4957-abb5-1d84191bf30c
  2013-11-16 21:54:27.999 | 2013-11-16 21:41:21,001 Response Headers: 
{'content-length': '14', 'date': 'Sat, 16 Nov 2013 21:41:20 GMT', 
'content-type': 'application/json', 'connection': 'close'}
  2013-11-16 21:54:27.999 | 2013-11-16 21:41:21,001 Response Body: {"output": 
""}
  2013-11-16 21:54:27.999 | }}}
  2013-11-16 21:54:27.999 | 
  2013-11-16 21:54:27.999 | Traceback (most recent call last):
  2013-11-16 21:54:27.999 |   File 
"tempest/api/compute/servers/test_server_actions.py", line 281, in 
test_get_console_output
  2013-11-16 21:54:28.000 | self.wait_for(get_output)
  2013-11-16 21:54:28.000 |   File "tempest/api/compute/base.py", line 133, in 
wait_for
  2013-11-16 21:54:28.000 | condition()
  2013-11-16 21:54:28.000 |   File 
"tempest/api/compute/servers/test_server_actions.py", line 278, in get_output
  2013-11-16 21:54:28.000 | self.assertTrue(output, "Console output was 
empty.")
  2013-11-16 21:54:28.000 |   File "/usr/lib/python2.7/unittest/case.py", line 
420, in assertTrue
  2013-11-16 21:54:28.000 | raise self.failureException(msg)
  2013-11-16 21:54:28.001 | AssertionError: Console output was empty.

  n-api
  

  2013-11-16 21:41:20.782 DEBUG nova.api.openstack.wsgi 
[req-7a2ee0ab-c977-4957-abb5-1d84191bf30c 
ServerActionsTestJSON-tempest-2102529866-user 
ServerActionsTestJSON-tempest-2102529866-tenant] Action: 'action', body: 
{"os-getConsoleOutput": {"length": 10}} _process_stack 
/opt/stack/new/nova/nova/api/openstack/wsgi.py:963
  2013-11-16 21:41:20.782 DEBUG nova.api.openstack.wsgi 
[req-7a2ee0ab-c977-4957-abb5-1d84191bf30c 
ServerActionsTestJSON-tempest-2102529866-user 
ServerActionsTestJSON-tempest-2102529866-tenant] Calling method > _process_stack 
/opt/stack/new/nova/nova/api/openstack/wsgi.py:964
  2013-11-16 21:41:20.865 DEBUG nova.openstack.common.rpc.amqp 
[req-7a2ee0ab-c977-4957-abb5-1d84191bf30c 
ServerActionsTestJSON-tempest-2102529866-user 
ServerActionsTestJSON-tempest-2102529866-tenant] Making synchronous call on 
compute.devstack-precise-hpcloud-az2-663635 ... multicall 
/opt/stack/new/nova/nova/openstack/common/rpc/amqp.py:553
  2013-11-16 21:41:20.866 DEBUG nova.openstack.common.rpc.amqp 
[req-7a2ee0ab-c977-4957-abb5-1d84191bf30c 
ServerActionsTestJSON-tempest-2102529866-user 
ServerActionsTestJSON-tempest-2102529866-tenant] MSG_ID is 
a93dceabf6a441eb850b5fbb012d661f multicall 
/opt/stack/new/nova/nova/openstack/common/rpc/amqp.py:556
  2013-11-16 21:41:20.866 DEBUG nova.openstack.common.rpc.amqp 
[req-7a2ee0ab-c977-4957-abb5-1d84191bf30c 
ServerActionsTestJSON-tempest-2102529866-user 
ServerActionsTestJSON-tempest-2102529866-tenant] UNIQUE_ID is 
706ab69dc066440fbe1bd7766b73d953. _add_unique_id 
/opt/stack/new/nova/nova/openstack/common/rpc/amqp.py:341
  2013-11-16 21:41:20.869 22679 DEBUG amqp [-] Closed channel #1 _do_close 
/usr/local/lib/python2.7/dist-packages/amqp/channel.py:95
  2013-11-16 21:41:20.869 22679 DEBUG amqp [-] using channel_id: 1 __init__ 
/usr/local/lib/python2.7/dist-packages/amqp/channel.py:71
  2013-11-16 21:41:20.870 22679 DEBUG amqp [-] Channel open _open_ok 
/usr/local/lib/python2.7/dist-packages/amqp/channel.py:429
  2013-11-16 21:41:20.999 INFO nova.osapi_compute.wsgi.server 
[req-7a2ee0ab-c977-4957-abb5-1d84191bf30c 
ServerActionsTestJSON-tempest-2102529866-user 
Se

[Yahoo-eng-team] [Bug 1213215] Re: ServerRescueTest tearDownClass fails with volume status being in-use

2013-12-12 Thread Sean Dague
** Also affects: cinder
   Importance: Undecided
   Status: New

** Changed in: cinder
   Importance: Undecided => High

** Changed in: nova
   Importance: Undecided => High

** Changed in: cinder
   Status: New => Confirmed

** Changed in: nova
   Status: New => Confirmed

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1213215

Title:
  ServerRescueTest tearDownClass fails with volume status being in-use

Status in Cinder:
  Confirmed
Status in OpenStack Compute (Nova):
  Confirmed
Status in Tempest:
  Confirmed

Bug description:
  Occasionally running tempest in parallel will fail several tests with
  timeout errors. The only nontimeout failure message is that the
  ServerRescueTest failed to delete a volume because it was still marked
  as in use. My guess is that the leftover volume is somehow interfering
  with the other tests causing them to timeout. But, I haven't looked at
  the logs in detail so it's just a wild guess.

  
  2013-08-16 14:11:42.074 | 
==
  2013-08-16 14:11:42.075 | FAIL: 
tempest.api.compute.servers.test_disk_config.ServerDiskConfigTestJSON.test_rebuild_server_with_auto_disk_config[gate]
  2013-08-16 14:11:42.075 | 
tempest.api.compute.servers.test_disk_config.ServerDiskConfigTestJSON.test_rebuild_server_with_auto_disk_config[gate]
  2013-08-16 14:11:42.075 | 
--
  2013-08-16 14:11:42.075 | _StringException: Empty attachments:
  2013-08-16 14:11:42.075 |   stderr
  2013-08-16 14:11:42.076 |   stdout
  2013-08-16 14:11:42.076 | 
  2013-08-16 14:11:42.076 | Traceback (most recent call last):
  2013-08-16 14:11:42.076 |   File 
"tempest/api/compute/servers/test_disk_config.py", line 64, in 
test_rebuild_server_with_auto_disk_config
  2013-08-16 14:11:42.076 | wait_until='ACTIVE')
  2013-08-16 14:11:42.076 |   File "tempest/api/compute/base.py", line 140, in 
create_server
  2013-08-16 14:11:42.076 | server['id'], kwargs['wait_until'])
  2013-08-16 14:11:42.077 |   File 
"tempest/services/compute/json/servers_client.py", line 160, in 
wait_for_server_status
  2013-08-16 14:11:42.077 | time.sleep(self.build_interval)
  2013-08-16 14:11:42.077 |   File 
"/usr/local/lib/python2.7/dist-packages/fixtures/_fixtures/timeout.py", line 
52, in signal_handler
  2013-08-16 14:11:42.077 | raise TimeoutException()
  2013-08-16 14:11:42.077 | TimeoutException
  2013-08-16 14:11:42.077 | 
  2013-08-16 14:11:42.077 | 
  2013-08-16 14:11:42.078 | 
==
  2013-08-16 14:11:42.078 | FAIL: setUpClass 
(tempest.api.compute.images.test_image_metadata.ImagesMetadataTestXML)
  2013-08-16 14:11:42.078 | setUpClass 
(tempest.api.compute.images.test_image_metadata.ImagesMetadataTestXML)
  2013-08-16 14:11:42.078 | 
--
  2013-08-16 14:11:42.078 | _StringException: Traceback (most recent call last):
  2013-08-16 14:11:42.078 |   File 
"tempest/api/compute/images/test_image_metadata.py", line 46, in setUpClass
  2013-08-16 14:11:42.078 | cls.client.wait_for_image_status(cls.image_id, 
'ACTIVE')
  2013-08-16 14:11:42.079 |   File 
"tempest/services/compute/xml/images_client.py", line 167, in 
wait_for_image_status
  2013-08-16 14:11:42.079 | raise exceptions.TimeoutException
  2013-08-16 14:11:42.079 | TimeoutException: Request timed out
  2013-08-16 14:11:42.079 | 
  2013-08-16 14:11:42.079 | 
  2013-08-16 14:11:42.079 | 
==
  2013-08-16 14:11:42.079 | FAIL: 
tempest.api.compute.servers.test_server_rescue.ServerRescueTestJSON.test_rescued_vm_detach_volume[gate,negative]
  2013-08-16 14:11:42.080 | 
tempest.api.compute.servers.test_server_rescue.ServerRescueTestJSON.test_rescued_vm_detach_volume[gate,negative]
  2013-08-16 14:11:42.080 | 
--
  2013-08-16 14:11:42.080 | _StringException: Empty attachments:
  2013-08-16 14:11:42.080 |   stderr
  2013-08-16 14:11:42.080 |   stdout
  2013-08-16 14:11:42.080 | 
  2013-08-16 14:11:42.081 | Traceback (most recent call last):
  2013-08-16 14:11:42.081 |   File 
"tempest/api/compute/servers/test_server_rescue.py", line 184, in 
test_rescued_vm_detach_volume
  2013-08-16 14:11:42.081 | 
self.servers_client.wait_for_server_status(self.server_id, 'RESCUE')
  2013-08-16 14:11:42.081 |   File 
"tempest/services/compute/json/servers_client.py", line 160, in 
wait_for_server_status
  2013-08-16 14:11:42.081 | time.sleep(self.build_interval)
  2013-08-16 14:11:42.081 |   File 
"/usr/local/lib/python2.7/dist-packages/fixtures/_fixtures/timeout.py", line 
52, in signal_handler
  2013-08-16 14:11:42.081 | raise Timeo

[Yahoo-eng-team] [Bug 1257070] Re: test_glance_timeout flakey fail

2013-12-12 Thread Attila Fazekas
Glance configured with one worker, is it possible it is too busy because
of another orations?

Do we need to increase the timeout  and/or the workers ?

** Changed in: tempest
   Status: New => Incomplete

** Also affects: glance
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Glance.
https://bugs.launchpad.net/bugs/1257070

Title:
  test_glance_timeout flakey fail

Status in OpenStack Image Registry and Delivery Service (Glance):
  New
Status in Tempest:
  Incomplete

Bug description:
  Transient fail for
  tempest.cli.simple_read_only.test_glance.SimpleReadOnlyGlanceClientTest
  test_glance_timeout

  http://logs.openstack.org/66/55766/3/gate/gate-tempest-devstack-vm-
  postgres-full/a807434/testr_results.html.gz

  ft254.6: 
tempest.cli.simple_read_only.test_glance.SimpleReadOnlyGlanceClientTest.test_glance_timeout_StringException:
 Empty attachments:
stderr
stdout

  pythonlogging:'': {{{
  2013-11-29 07:57:20,345 running: '/usr/local/bin/glance --os-username admin 
--os-tenant-name admin --os-password secret --os-auth-url 
http://127.0.0.1:5000/v2.0/  --timeout 15 image-list '
  2013-11-29 07:57:37,633 output of /usr/local/bin/glance --os-username admin 
--os-tenant-name admin --os-password secret --os-auth-url 
http://127.0.0.1:5000/v2.0/  --timeout 15 image-list :

  2013-11-29 07:57:37,635 error output of /usr/local/bin/glance --os-username 
admin --os-tenant-name admin --os-password secret --os-auth-url 
http://127.0.0.1:5000/v2.0/  --timeout 15 image-list :
  Error communicating with http://127.0.0.1:9292 timed out
  }}}

  Traceback (most recent call last):
File "tempest/cli/simple_read_only/test_glance.py", line 89, in 
test_glance_timeout
  self.glance('image-list', flags='--timeout %d' % CONF.cli.timeout)
File "tempest/cli/__init__.py", line 81, in glance
  'glance', action, flags, params, admin, fail_ok)
File "tempest/cli/__init__.py", line 110, in cmd_with_auth
  return self.cmd(cmd, action, flags, params, fail_ok)
File "tempest/cli/__init__.py", line 132, in cmd
  stderr=result_err)
  CommandFailed: Command '['/usr/local/bin/glance', '--os-username', 'admin', 
'--os-tenant-name', 'admin', '--os-password', 'secret', '--os-auth-url', 
'http://127.0.0.1:5000/v2.0/', '--timeout', '15', 'image-list']' returned 
non-zero exit status 1

To manage notifications about this bug go to:
https://bugs.launchpad.net/glance/+bug/1257070/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1221899] Re: test_resize_server_from_auto_to_manual: server failed to reach VERIFY_RESIZE status within the required time

2013-12-12 Thread Sean Dague
** Changed in: tempest
   Status: New => Incomplete

** Also affects: nova
   Importance: Undecided
   Status: New

** Changed in: tempest
   Status: Incomplete => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1221899

Title:
  test_resize_server_from_auto_to_manual: server failed to reach
  VERIFY_RESIZE status within the required time

Status in OpenStack Compute (Nova):
  New
Status in Tempest:
  Invalid

Bug description:
  2013-09-06 19:24:49.608 | Traceback (most recent call last):
  2013-09-06 19:24:49.608 |   File 
"tempest/api/compute/servers/test_disk_config.py", line 114, in 
test_resize_server_from_auto_to_manual
  2013-09-06 19:24:49.609 | 
self.client.wait_for_server_status(server['id'], 'VERIFY_RESIZE')
  2013-09-06 19:24:49.609 |   File 
"tempest/services/compute/xml/servers_client.py", line 331, in 
wait_for_server_status
  2013-09-06 19:24:49.609 | raise exceptions.TimeoutException(message)
  2013-09-06 19:24:49.609 | TimeoutException: Request timed out
  2013-09-06 19:24:49.609 | Details: Server 
dabbdc8d-3194-4e88-bc9c-c897a1fe5f78 failed to reach VERIFY_RESIZE status 
within the required time (400 s). Current status: RESIZE.

  
  
http://logs.openstack.org/48/45248/2/check/gate-tempest-devstack-vm-full/66d555c/

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1221899/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1218582] Re: ServerActionsTestJSON.test_pause_unpause_server fails with a timeout, other failures as a side effect

2013-12-12 Thread Sean Dague
This is actually also a nova bug. Nova didn't move this into a paused
state correctly.

** Also affects: nova
   Importance: Undecided
   Status: New

** Changed in: tempest
   Importance: Undecided => Medium

** Changed in: tempest
   Status: New => Confirmed

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1218582

Title:
  ServerActionsTestJSON.test_pause_unpause_server fails with a timeout,
  other failures as a side effect

Status in OpenStack Compute (Nova):
  New
Status in Tempest:
  Confirmed

Bug description:
  In this case, it looks like a test for pausing an instance got stuck
  (reported as running the longest at 250 seconds).  A number of other
  tests failed as a side effect it seems.  They all report that they
  can't do what they needed to do because the instance is still in a
  pausing task state.

  
http://logs.openstack.org/69/42769/7/gate/gate-tempest-devstack-vm-full/b2879b7
  
http://logs.openstack.org/69/42769/7/gate/gate-tempest-devstack-vm-full/b2879b7/console.html
  https://review.openstack.org/#/c/42769/

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1218582/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1259553] Re: ListServersNegativeTestXML flakey ERROR generation

2013-12-12 Thread Sean Dague
This isn't a tempest bug, because there wouldnt' be a code fix in
tempest to fix it

** Changed in: tempest
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1259553

Title:
  ListServersNegativeTestXML flakey ERROR generation

Status in OpenStack Compute (Nova):
  New
Status in Tempest:
  Invalid

Bug description:
  Looks like ListServersNegativeTestXML can trigger a race in Nova which
  causes an ERROR to be logged:

  
  2013-12-10 12:41:34.628 ERROR nova.network.manager 
[req-7430295d-9c23-4a46-a755-f1e93aa53c6f 
ListServersNegativeTestXML-tempest-1313593245-user 
ListServersNegativeTestXML-tempest-1313593245-tenant] Unable to release 
10.1.0.10 because vif doesn't exist.

  http://logs.openstack.org/12/61012/2/check/check-tempest-dsvm-
  full/c92bea6/logs/screen-n-net.txt.gz?level=ERROR

  Probably related to https://code.launchpad.net/bugs/968457 which is meant to 
be fixed, but the comment in the nova
  code suggests it might not be.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1259553/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260249] Re: migration-list: 'unicode' object has no attribute 'iteritems'

2013-12-12 Thread Joe Gordon
** Changed in: nova
   Status: Invalid => New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1260249

Title:
  migration-list: 'unicode' object has no attribute 'iteritems'

Status in OpenStack Compute (Nova):
  New
Status in Python client library for Nova:
  In Progress

Bug description:
  There is an AttributeError when we try to use the command "nova
  migration-list"

  Traceback (most recent call last):
File "/opt/stack/python-novaclient/novaclient/shell.py", line 721, in main
  OpenStackComputeShell().main(map(strutils.safe_decode, sys.argv[1:]))
File "/opt/stack/python-novaclient/novaclient/shell.py", line 657, in main
  args.func(self.cs, args)
File "/opt/stack/python-novaclient/novaclient/v1_1/contrib/migrations.py", 
line 71, in do_migration_list
  args.cell_name))
File "/opt/stack/python-novaclient/novaclient/v1_1/contrib/migrations.py", 
line 53, in list
  return self._list("/os-migrations%s" % query_string, "migrations")
File "/opt/stack/python-novaclient/novaclient/base.py", line 80, in _list
  for res in data if res]
File "/opt/stack/python-novaclient/novaclient/base.py", line 426, in 
__init__
  self._add_details(info)
File "/opt/stack/python-novaclient/novaclient/base.py", line 449, in 
_add_details
  for (k, v) in six.iteritems(info):
File "/usr/local/lib/python2.7/dist-packages/six.py", line 439, in iteritems
  return iter(getattr(d, _iteritems)(**kw))
  AttributeError: 'unicode' object has no attribute 'iteritems'
  ERROR: 'unicode' object has no attribute 'iteritems'

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1260249/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1258682] Re: timeout causing gate-tempest-dsvm-full to fail

2013-12-12 Thread Sean Dague
If the fix is increasing the timeout in the gate, it's not a tempest
bug. It looks like in this case libvirt went off the rails, so nova is
probably a good bug choice

** Also affects: openstack-ci
   Importance: Undecided
   Status: New

** Changed in: tempest
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1258682

Title:
  timeout causing gate-tempest-dsvm-full to fail

Status in OpenStack Compute (Nova):
  New
Status in OpenStack Core Infrastructure:
  New
Status in Tempest:
  Invalid

Bug description:
  This has happened several times. A recent example is in
  https://jenkins02.openstack.org/job/gate-tempest-dsvm-full/775/console

  There are several mentions of FAIL in the logs, but since the job
  timed out, no console logs were saved.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1258682/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260359] [NEW] Adding a member to a LBaaS pool when there are no servers available it shows a "Success: Added member(s)." message.

2013-12-12 Thread alejandro emanuel paredes
Public bug reported:

Steps to reproduce:
1) Make sure there are no running instances
2) Go to "Load Balancers" tab / "Add Member"
3) In the "Add Member" window click on "Add"

Issue: No members are added and a "Success: Added member(s)." is shown.

** Affects: horizon
 Importance: Undecided
 Status: New


** Tags: lbaas

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1260359

Title:
  Adding a member to a LBaaS pool when there are no servers available it
  shows a "Success: Added member(s)." message.

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  Steps to reproduce:
  1) Make sure there are no running instances
  2) Go to "Load Balancers" tab / "Add Member"
  3) In the "Add Member" window click on "Add"

  Issue: No members are added and a "Success: Added member(s)." is
  shown.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1260359/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1258601] Re: nova.network.manager: Unable to release because vif doesn't exist.

2013-12-12 Thread David Kranz
*** This bug is a duplicate of bug 1258848 ***
https://bugs.launchpad.net/bugs/1258848

Please include a pointer to the log file for such reports. According to
logstash this has hit 48 times in the last two weeks which is a very low
failure rate. Ideally flaky bugs like this would be fixed. If the nova
team wants to silence this a patch can be submitted to the whitelist in
tempest.

** Changed in: tempest
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1258601

Title:
  nova.network.manager: Unable to release  because vif doesn't
  exist.

Status in OpenStack Compute (Nova):
  New
Status in Tempest:
  Invalid

Bug description:
  This error shows un in nova-network log.

  Not sure if it needs to be whitelisted.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1258601/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1153926] Re: flavor show shouldn't read deleted flavors.

2013-12-12 Thread Sean Dague
** No longer affects: tempest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1153926

Title:
  flavor show shouldn't read deleted flavors.

Status in OpenStack Dashboard (Horizon):
  In Progress
Status in OpenStack Compute (Nova):
  In Progress
Status in Python client library for Nova:
  In Progress

Bug description:
  An instance type is created by:

  return db.instance_type_create(context.get_admin_context(), kwargs)

  which uses the read_deleted="no" from the admin context.

  This means, as seen in nova/tests/test_instance_types.py:

  def test_read_deleted_false_converting_flavorid(self):
  """
  Ensure deleted instance types are not returned when not needed (for
  example when creating a server and attempting to translate from
  flavorid to instance_type_id.
  """
  instance_types.create("instance_type1", 256, 1, 120, 100, "test1")
  instance_types.destroy("instance_type1")
  instance_types.create("instance_type1_redo", 256, 1, 120, 100, "test1")

  instance_type = instance_types.get_instance_type_by_flavor_id(
  "test1", read_deleted="no")
  self.assertEqual("instance_type1_redo", instance_type["name"])

  flavors with colliding ids can exist in the database.

  From the test we see this looks intended, however it results in
  undesirable results if we consider the following scenario.

  For 'show' in the flavors api, it uses read_deleted="yes". The reason
  for this is if a vm was created in the past with a now-deleted flavor,
  'nova show' can still show the flavor name that was specified for that
  vm creation. The flavor name is retrieved using the flavor id stored
  with the instance.

  Well, if there are colliding flavor ids in the database, the first of
  the duplicates will be picked, and it may not be the correct flavor
  for the vm.

  This leads me to believe that maybe at flavor create time, colliding
  ids should not be allowed, i.e. use

  return db.instance_type_create(context.get_admin_context(read_deleted="yes"),
 kwargs)

  to prevent the possibility of colliding flavor ids.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1153926/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1224518] Re: test_reboot_server_hard fails sporadically in swift check jobs

2013-12-12 Thread Attila Fazekas
10 results  in logstash with  "Current status: HARD_REBOOT" .

** Also affects: nova
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1224518

Title:
  test_reboot_server_hard fails sporadically in swift check jobs

Status in OpenStack Compute (Nova):
  New
Status in Tempest:
  New

Bug description:
  See: http://logs.openstack.org/46/46146/2/check/gate-tempest-devstack-
  vm-postgres-full/b2712f1/console.html

  2013-09-12 04:43:17.625 | 
==
  2013-09-12 04:43:17.649 | FAIL: 
tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_reboot_server_hard[gate,smoke]
  2013-09-12 04:43:17.651 | 
tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_reboot_server_hard[gate,smoke]
  2013-09-12 04:43:17.652 | 
--
  2013-09-12 04:43:17.652 | _StringException: Empty attachments:
  2013-09-12 04:43:17.652 |   stderr
  2013-09-12 04:43:17.652 |   stdout
  2013-09-12 04:43:17.653 | 
  2013-09-12 04:43:17.653 | pythonlogging:'': {{{
  2013-09-12 04:43:17.653 | 2013-09-12 04:16:55,739 Request: GET 
http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf
  2013-09-12 04:43:17.654 | 2013-09-12 04:16:55,806 Response Status: 200
  2013-09-12 04:43:17.654 | 2013-09-12 04:16:55,806 Nova request id: 
req-cdc6b1fc-bcf2-4e9c-bea1-8bf935993cbd
  2013-09-12 04:43:17.654 | 2013-09-12 04:16:55,807 Request: POST 
http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf/action
  2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,917 Response Status: 202
  2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,917 Nova request id: 
req-3af37dd3-0ddc-4daa-aa6f-6958a5073cc4
  2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,918 Request: GET 
http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf
  2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,986 Response Status: 200
  2013-09-12 04:43:17.656 | 2013-09-12 04:16:55,986 Nova request id: 
req-a7298d3e-167c-4c8f-9506-6064ba811e5b

  .
  .
  .

  2013-09-12 04:43:17.976 | 2013-09-12 04:23:35,773 Request: GET 
http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf
  2013-09-12 04:43:17.976 | 2013-09-12 04:23:35,822 Response Status: 200
  2013-09-12 04:43:17.976 | 2013-09-12 04:23:35,823 Nova request id: 
req-a122aded-b49b-4847-9920-b2b8b09bc0ca
  2013-09-12 04:43:17.976 | }}}
  2013-09-12 04:43:17.977 | 
  2013-09-12 04:43:17.977 | Traceback (most recent call last):
  2013-09-12 04:43:17.978 |   File 
"tempest/api/compute/servers/test_server_actions.py", line 81, in 
test_reboot_server_hard
  2013-09-12 04:43:17.978 | 
self.client.wait_for_server_status(self.server_id, 'ACTIVE')
  2013-09-12 04:43:17.979 |   File 
"tempest/services/compute/json/servers_client.py", line 176, in 
wait_for_server_status
  2013-09-12 04:43:17.979 | raise exceptions.TimeoutException(message)
  2013-09-12 04:43:17.979 | TimeoutException: Request timed out
  2013-09-12 04:43:17.980 | Details: Server 
8ad0ad9a-3975-486f-94b4-af1c89b51aaf failed to reach ACTIVE status within the 
required time (400 s). Current status: HARD_REBOOT.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1224518/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1259542] Re: Send OS distribution on API headers

2013-12-12 Thread Dolph Mathews
Moved to oslo as there's nothing keystone-specific about this.

** Changed in: keystone
   Importance: Undecided => Wishlist

** Project changed: keystone => oslo

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Keystone.
https://bugs.launchpad.net/bugs/1259542

Title:
  Send OS distribution on API headers

Status in Oslo - a Library of Common OpenStack Code:
  In Progress

Bug description:
  It should be interesting to send OS distribution in API headers, to be
  able to detect the OS that is serving API calls and collect that on
  stats.

To manage notifications about this bug go to:
https://bugs.launchpad.net/oslo/+bug/1259542/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1152623] Re: RFC2616 section 9.7 status code vs. nova server delete

2013-12-12 Thread Sean Dague
Until this is changed in nova, it's not actually appropriate to have a
tempest issue

** No longer affects: tempest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1152623

Title:
  RFC2616 section 9.7 status code vs. nova server delete

Status in OpenStack Compute (Nova):
  In Progress

Bug description:
  In REST client implementation is common good practice, when:
  - request causes an synchronous and asynchronous effect , and
  - the synchronous operation has any immediately visible effect ie. immediate 
subsequent request showing any change,
  we should emphasize the synchronous behavior in the responses (Status code)  
(Or responding in way which does not distinguish the two cases).

  However if the HTTP method is DELETE, the rule is the opposite! 
  If the resource on the request URL does not deleted the service MUST NOT 
response with 204.

  "
 A successful response SHOULD be 200 (OK) if the response includes an
 entity describing the status, 202 (Accepted) if the action has not
 yet been enacted, or 204 (No Content) if the action has been enacted
 but the response does not include an entity.
  " by RFC2616 section 9.7

  It means if a DELETE request responded with 204 status code, I MUST
  get 404 in an immediate subsequent request, unless concurrent
  operation recreated the resource.

  
  $ nova --debug delete ab0ebda6-2c21-4258-8934-1005b970fee5 ; nova --debug 
show ab0ebda6-2c21-4258-8934-1005b970fee5

  Part of the output in the received order:
  -
  REQ: curl -i 
http://10.34.69.149:8774/v2/89a38fe6d3194864995ab0872905a65e/servers/ab0ebda6-2c21-4258-8934-1005b970fee5
 -X DELETE -H "X-Auth-Project-Id: admin" -H "User-Agent: python-novaclient" -H 
"Accept: application/json" -H "X-Auth-Token: c35f5783528d4131bf100604b2fabd6c"

  send: u'DELETE 
/v2/89a38fe6d3194864995ab0872905a65e/servers/ab0ebda6-2c21-4258-8934-1005b970fee5
 HTTP/1.1\r\nHost: 10.34.69.149:8774\r\nx-auth-project-id: 
admin\r\nx-auth-token: c35f5783528d4131bf100604b2fabd6c\r\naccept-encoding: 
gzip, deflate\r\naccept: application/json\r\nuser-agent: 
python-novaclient\r\n\r\n'
  reply: 'HTTP/1.1 204 No Content\r\n'
  header: Content-Length: 0
  header: X-Compute-Request-Id: req-53e3503a-8d73-4ffc-ba43-4bd5659a9e22
  header: Content-Type: application/json
  header: Date: Sat, 02 Mar 2013 18:26:21 GMT
  RESP:{'date': 'Sat, 02 Mar 2013 18:26:21 GMT', 'status': '204', 
'content-length': '0', 'content-type': 'application/json', 
'x-compute-request-id': 'req-53e3503a-8d73-4ffc-ba43-4bd5659a9e22'} 
  -
  REQ: curl -i 
http://10.34.69.149:8774/v2/89a38fe6d3194864995ab0872905a65e/servers/ab0ebda6-2c21-4258-8934-1005b970fee5
 -X GET -H "X-Auth-Project-Id: admin" -H "User-Agent: python-novaclient" -H 
"Accept: application/json" -H "X-Auth-Token: f74d6c7226c14915a26a81b540d43f3b"

  connect: (10.34.69.149, 8774)
  send: u'GET 
/v2/89a38fe6d3194864995ab0872905a65e/servers/ab0ebda6-2c21-4258-8934-1005b970fee5
 HTTP/1.1\r\nHost: 10.34.69.149:8774\r\nx-auth-project-id: 
admin\r\nx-auth-token: f74d6c7226c14915a26a81b540d43f3b\r\naccept-encoding: 
gzip, deflate\r\naccept: application/json\r\nuser-agent: 
python-novaclient\r\n\r\n'
  reply: 'HTTP/1.1 200 OK\r\n'
  header: X-Compute-Request-Id: req-80c97c68-0b44-4650-b027-84a85ee04b86
  header: Content-Type: application/json
  header: Content-Length: 1502
  header: Date: Sat, 02 Mar 2013 18:26:21 GMT
  RESP:{'status': '200', 'content-length': '1502', 'content-location': 
u'http://10.34.69.149:8774/v2/89a38fe6d3194864995ab0872905a65e/servers/ab0ebda6-2c21-4258-8934-1005b970fee5',
 'x-compute-request-id': 'req-80c97c68-0b44-4650-b027-84a85ee04b86', 'date': 
'Sat, 02 Mar 2013 18:26:21 GMT', 'content-type': 'application/json'} {"server": 
{"status": "ACTIVE", "updated": "2013-03-02T18:26:21Z", "hostId": 
"31bdffcdffd5b869b87c9be3cdd700e29c4a08286d6d306622b4815a", 
"OS-EXT-SRV-ATTR:host": "new32.lithium.rhev.lab.eng.brq.redhat.com", 
"addresses": {"novanetwork": [{"version": 4, "addr": "192.168.32.2"}]}, 
"links": [{"href": 
"http://10.34.69.149:8774/v2/89a38fe6d3194864995ab0872905a65e/servers/ab0ebda6-2c21-4258-8934-1005b970fee5";,
 "rel": "self"}, {"href": 
"http://10.34.69.149:8774/89a38fe6d3194864995ab0872905a65e/servers/ab0ebda6-2c21-4258-8934-1005b970fee5";,
 "rel": "bookmark"}], "key_name": null, "image": {"id": 
"12e9c131-aaf4-4f73-9659-ed2da9759cd2", "links": [{"href": "http://10.34.69.149:
 
8774/89a38fe6d3194864995ab0872905a65e/images/12e9c131-aaf4-4f73-9659-ed2da9759cd2",
 "rel": "bookmark"}]}, "OS-EXT-STS:task_state": "deleting", 
"OS-EXT-STS:vm_state": "active", "OS-EXT-SRV-ATTR:instance_name": 
"instance-0003", "OS-EXT-SRV-ATTR:hypervisor_hostname": 
"new32.lithium.rhev.lab.eng.brq.redhat.com", "flavor": {"id": "1", "links": 
[{"href": 
"http://10.34.69

[Yahoo-eng-team] [Bug 1006725] Re: Incorrect error returned during Create Image and multi byte characters used for Image name

2013-12-12 Thread Sean Dague
Definitely not fixed on the nova side. Attempts to unskip the bug
generated issues.

** Changed in: nova
   Status: Invalid => Confirmed

** Changed in: nova
   Importance: Low => High

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1006725

Title:
  Incorrect error returned during Create Image and multi byte characters
  used for Image name

Status in OpenStack Compute (Nova):
  Confirmed
Status in Tempest:
  In Progress

Bug description:
  Our tempest tests that checks for 400 Bad Request return code fails
  with a ComputeFault instead.

  Pass multi-byte character image name during Create Image
  Actual Response Code: ComputeFault, 500 
  Expected Response Code: 400 Bad Request

  
  Return an error if the server name has a multi-byte character ... FAIL

  ==
  FAIL: Return an error if the server name has a multi-byte character
  --
  Traceback (most recent call last):
File "/opt/stack/tempest/tests/test_images.py", line 251, in 
test_create_image_specify_multibyte_character_server_name
  self.fail("Should return 400 Bad Request if multi byte characters"
  AssertionError: Should return 400 Bad Request if multi byte characters are 
used for image name
   >> begin captured logging << 
  tempest.config: INFO: Using tempest config file 
/opt/stack/tempest/etc/tempest.conf
  tempest.common.rest_client: ERROR: Request URL: 
http://10.2.3.164:8774/v2/1aeac1cfbfdd43c2845b2cb3a4f15790/images/24ceff93-1af3-41ab-802f-9fc4d8b90b69
  tempest.common.rest_client: ERROR: Request Body: None
  tempest.common.rest_client: ERROR: Response Headers: {'date': 'Thu, 31 May 
2012 06:02:33 GMT', 'status': '404', 'content-length': '62', 'content-type': 
'application/json; charset=UTF-8', 'x-compute-request-id': 
'req-7a15d284-e934-47a1-87f4-7746e949c7a2'}
  tempest.common.rest_client: ERROR: Response Body: {"itemNotFound": 
{"message": "Image not found.", "code": 404}}
  tempest.common.rest_client: ERROR: Request URL: 
http://10.2.3.164:8774/v2/1aeac1cfbfdd43c2845b2cb3a4f15790/servers/ecb51dfb-493d-4ef8-9178-1adc3d96a04d/action
  tempest.common.rest_client: ERROR: Request Body: {"createImage": {"name": 
"\ufeff43802479847"}}
  tempest.common.rest_client: ERROR: Response Headers: {'date': 'Thu, 31 May 
2012 06:02:44 GMT', 'status': '500', 'content-length': '128', 'content-type': 
'application/json; charset=UTF-8', 'x-compute-request-id': 
'req-1a9505f5-4dfc-44e7-b04a-f8daec0f956e'}
  tempest.common.rest_client: ERROR: Response Body: {u'computeFault': 
{u'message': u'The server has either erred or is incapable of performing the 
requested operation.', u'code': 500}}
  - >> end captured logging << -

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1006725/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1226943] Re: Need Use built-in print() function instead of print statement

2013-12-12 Thread Sean Dague
** No longer affects: tempest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1226943

Title:
  Need Use built-in  print() function  instead of print  statement

Status in OpenStack Telemetry (Ceilometer):
  Fix Released
Status in Cinder:
  Fix Released
Status in OpenStack Image Registry and Delivery Service (Glance):
  Fix Released
Status in Orchestration API (Heat):
  Fix Released
Status in OpenStack Identity (Keystone):
  In Progress
Status in OpenStack Neutron (virtual network service):
  Fix Released
Status in OpenStack Compute (Nova):
  Fix Released

Bug description:
  In python 3 print statement is not supported, so we should  use only print() 
functions.
  built-in function was introduce in  python 2.6
  http://www.python.org/dev/peps/pep-3105/

  Note :This function is not normally available as a built-in since the name 
print is recognized as the print statement. 
  To disable the statement and use the print() function, use this future 
statement at the top of your module:
  from __future__ import print_function

  http://docs.python.org/2/library/functions.html#print

To manage notifications about this bug go to:
https://bugs.launchpad.net/ceilometer/+bug/1226943/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1216903] Re: logical_resource_id disappeared in favor of resource_name

2013-12-12 Thread Sean Dague
** Changed in: tempest
   Status: Fix Committed => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1216903

Title:
  logical_resource_id  disappeared in favor of resource_name

Status in OpenStack Dashboard (Horizon):
  Fix Released
Status in Python client library for heat:
  Fix Released
Status in Tempest:
  Fix Released

Bug description:
  impact from Heat side :

  https://review.openstack.org/#/c/43391/

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1216903/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1239891] [NEW] tempest.api.object_storage.test_account_services.AccountTest fails under neutron-pg-isolated

2013-12-12 Thread Launchpad Bug Tracker
You have been subscribed to a public bug:

http://logs.openstack.org/38/51738/1/check/check-tempest-devstack-vm-
neutron-pg-isolated/73aad7a/console.html

2013-10-15 00:19:04.556 | Error in atexit._run_exitfuncs:
2013-10-15 00:19:04.556 | Traceback (most recent call last):
2013-10-15 00:19:04.556 |   File "/usr/lib/python2.7/atexit.py", line 24, in 
_run_exitfuncs
2013-10-15 00:19:04.557 | func(*targs, **kargs)
2013-10-15 00:19:04.558 |   File "tempest/test.py", line 167, in 
validate_tearDownClass
2013-10-15 00:19:04.558 | + str(at_exit_set))
2013-10-15 00:19:04.558 | RuntimeError: tearDownClass does not calls the 
super's tearDownClass in these classes: set([])
2013-10-15 00:19:04.559 | Error in sys.exitfunc:
2013-10-15 00:19:04.663 | 
2013-10-15 00:19:04.664 | process-returncode
2013-10-15 00:19:04.664 | process-returncode ... FAIL
2013-10-15 00:19:04.980 | 
2013-10-15 00:19:04.981 | 
==
2013-10-15 00:19:04.981 | FAIL: tearDownClass 
(tempest.api.object_storage.test_account_services.AccountTest)
2013-10-15 00:19:04.981 | tearDownClass 
(tempest.api.object_storage.test_account_services.AccountTest)
2013-10-15 00:19:04.982 | 
--
2013-10-15 00:19:04.982 | _StringException: Traceback (most recent call last):
2013-10-15 00:19:04.982 |   File 
"tempest/api/object_storage/test_account_services.py", line 41, in tearDownClass
2013-10-15 00:19:04.983 | super(AccountTest, cls).tearDownClass()
2013-10-15 00:19:04.983 |   File "tempest/api/object_storage/base.py", line 77, 
in tearDownClass
2013-10-15 00:19:04.983 | cls.isolated_creds.clear_isolated_creds()
2013-10-15 00:19:04.984 |   File "tempest/common/isolated_creds.py", line 453, 
in clear_isolated_creds
2013-10-15 00:19:04.984 | self._clear_isolated_net_resources()
2013-10-15 00:19:04.984 |   File "tempest/common/isolated_creds.py", line 445, 
in _clear_isolated_net_resources
2013-10-15 00:19:04.985 | self._clear_isolated_network(network['id'], 
network['name'])
2013-10-15 00:19:04.985 |   File "tempest/common/isolated_creds.py", line 399, 
in _clear_isolated_network
2013-10-15 00:19:04.985 | net_client.delete_network(network_id)
2013-10-15 00:19:04.985 |   File 
"tempest/services/network/json/network_client.py", line 76, in delete_network
2013-10-15 00:19:04.986 | resp, body = self.delete(uri, self.headers)
2013-10-15 00:19:04.986 |   File "tempest/common/rest_client.py", line 308, in 
delete
2013-10-15 00:19:04.986 | return self.request('DELETE', url, headers)
2013-10-15 00:19:04.987 |   File "tempest/common/rest_client.py", line 436, in 
request
2013-10-15 00:19:04.987 | resp, resp_body)
2013-10-15 00:19:04.987 |   File "tempest/common/rest_client.py", line 522, in 
_error_checker
2013-10-15 00:19:04.988 | raise exceptions.ComputeFault(message)
2013-10-15 00:19:04.988 | ComputeFault: Got compute fault
2013-10-15 00:19:04.988 | Details: {"NeutronError": "Request Failed: internal 
server error while processing your request."}
2013-10-15 00:19:04.988 | 
2013-10-15 00:19:04.989 | 
2013-10-15 00:19:04.989 | 
==
2013-10-15 00:19:04.989 | FAIL: process-returncode
2013-10-15 00:19:04.990 | process-returncode
2013-10-15 00:19:04.990 | 
--
2013-10-15 00:19:04.990 | _StringException: Binary content:
2013-10-15 00:19:04.991 |   traceback (test/plain; charset="utf8")
2013-10-15 00:19:04.991 | 
2013-10-15 00:19:04.991 | 
2013-10-15 00:19:04.991 | 
--

** Affects: neutron
 Importance: Undecided
 Status: New

-- 
tempest.api.object_storage.test_account_services.AccountTest fails under 
neutron-pg-isolated
https://bugs.launchpad.net/bugs/1239891
You received this bug notification because you are a member of Yahoo! 
Engineering Team, which is subscribed to neutron.

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1232971] Re: tempest gating error: test_run_stop_terminate_instance_with_tags

2013-12-12 Thread Sean Dague
** Changed in: tempest
   Status: Confirmed => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1232971

Title:
  tempest gating error: test_run_stop_terminate_instance_with_tags

Status in OpenStack Neutron (virtual network service):
  Confirmed
Status in Tempest:
  Fix Released

Bug description:
  2013-09-29 21:56:27.035 | 
==
  2013-09-29 21:56:27.035 | FAIL: 
tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance_with_tags[gate,smoke]
  2013-09-29 21:56:27.035 | 
tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance_with_tags[gate,smoke]
  2013-09-29 21:56:27.035 | 
--
  2013-09-29 21:56:27.036 | _StringException: Empty attachments:
  2013-09-29 21:56:27.036 |   stderr
  2013-09-29 21:56:27.036 |   stdout
  2013-09-29 21:56:27.036 | 
  2013-09-29 21:56:27.036 | pythonlogging:'': {{{2013-09-29 21:42:26,545 state: 
pending}}}
  2013-09-29 21:56:27.036 | 
  2013-09-29 21:56:27.037 | Traceback (most recent call last):
  2013-09-29 21:56:27.037 |   File 
"tempest/thirdparty/boto/test_ec2_instance_run.py", line 175, in 
test_run_stop_terminate_instance_with_tags
  2013-09-29 21:56:27.037 | self.assertInstanceStateWait(instance, 
"running")
  2013-09-29 21:56:27.037 |   File "tempest/thirdparty/boto/test.py", line 356, 
in assertInstanceStateWait
  2013-09-29 21:56:27.037 | state = self.waitInstanceState(lfunction, 
wait_for)
  2013-09-29 21:56:27.037 |   File "tempest/thirdparty/boto/test.py", line 341, 
in waitInstanceState
  2013-09-29 21:56:27.037 | self.valid_instance_state)
  2013-09-29 21:56:27.038 |   File "tempest/thirdparty/boto/test.py", line 331, 
in state_wait_gone
  2013-09-29 21:56:27.038 | state = state_wait(lfunction, final_set, 
valid_set)
  2013-09-29 21:56:27.038 |   File "tempest/thirdparty/boto/utils/wait.py", 
line 57, in state_wait
  2013-09-29 21:56:27.038 | (dtime, final_set, status))
  2013-09-29 21:56:27.038 | AssertionError: State change timeout 
exceeded!(400s) While waitingfor set(['running', '_GONE']) at "pending"

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1232971/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1232303] Re: FAIL: tempest test_large_ops_scenario - failed to get to expected status. In ERROR state.

2013-12-12 Thread Sean Dague
** No longer affects: tempest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1232303

Title:
  FAIL: tempest test_large_ops_scenario - failed to get to expected
  status. In ERROR state.

Status in OpenStack Neutron (virtual network service):
  New
Status in OpenStack Compute (Nova):
  New

Bug description:
  http://logs.openstack.org/13/48513/2/check/check-tempest-devstack-vm-
  large-ops/8ce3344/console.html

  2013-09-27 21:08:58.498 | 
==
  2013-09-27 21:08:58.498 | FAIL: 
tempest.scenario.test_large_ops.TestLargeOpsScenario.test_large_ops_scenario[compute,image]
  2013-09-27 21:08:58.498 | tags: worker-0
  2013-09-27 21:08:58.498 | 
--
  2013-09-27 21:08:58.498 | Empty attachments:
  2013-09-27 21:08:58.498 |   stderr
  2013-09-27 21:08:58.499 |   stdout
  2013-09-27 21:08:58.499 | 
  2013-09-27 21:08:58.499 | pythonlogging:'': {{{
  2013-09-27 21:08:58.499 | 2013-09-27 21:04:54,743 Starting new HTTP 
connection (1): 127.0.0.1
  2013-09-27 21:08:58.499 | 2013-09-27 21:04:54,864 Starting new HTTP 
connection (1): 127.0.0.1
  2013-09-27 21:08:58.499 | }}}
  2013-09-27 21:08:58.500 | 
  2013-09-27 21:08:58.500 | Traceback (most recent call last):
  2013-09-27 21:08:58.500 |   File "tempest/scenario/test_large_ops.py", line 
105, in test_large_ops_scenario
  2013-09-27 21:08:58.500 | self.nova_boot()
  2013-09-27 21:08:58.500 |   File "tempest/scenario/test_large_ops.py", line 
98, in nova_boot
  2013-09-27 21:08:58.500 | self._wait_for_server_status('ACTIVE')
  2013-09-27 21:08:58.501 |   File "tempest/scenario/test_large_ops.py", line 
42, in _wait_for_server_status
  2013-09-27 21:08:58.501 | self.compute_client.servers, server.id, status)
  2013-09-27 21:08:58.501 |   File "tempest/scenario/manager.py", line 290, in 
status_timeout
  2013-09-27 21:08:58.501 | self._status_timeout(things, thing_id, 
expected_status=expected_status)
  2013-09-27 21:08:58.501 |   File "tempest/scenario/manager.py", line 338, in 
_status_timeout
  2013-09-27 21:08:58.501 | self.config.compute.build_interval):
  2013-09-27 21:08:58.502 |   File "tempest/test.py", line 237, in 
call_until_true
  2013-09-27 21:08:58.502 | if func():
  2013-09-27 21:08:58.502 |   File "tempest/scenario/manager.py", line 329, in 
check_status
  2013-09-27 21:08:58.502 | raise exceptions.BuildErrorException(message)
  2013-09-27 21:08:58.502 | BuildErrorException: Server %(server_id)s failed to 
build and is in ERROR status
  2013-09-27 21:08:58.503 | Details:  failed to get 
to expected status.   In ERROR state.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1232303/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1230407] Re: VMs can't progress through state changes because Neutron is deadlocking on it's database queries, and thus leaving networks in inconsistent states

2013-12-12 Thread Sean Dague
** No longer affects: tempest

** No longer affects: nova

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1230407

Title:
  VMs can't progress through state changes because Neutron is
  deadlocking on it's database queries, and thus leaving networks in
  inconsistent states

Status in OpenStack Neutron (virtual network service):
  Confirmed

Bug description:
  This is most often seen with the "State change timeout exceeded" in
  the tempest logs.

  2013-09-25 16:03:28.319 | FAIL: 
tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance_with_tags[gate,smoke]
  2013-09-25 16:03:28.319 | 
tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance_with_tags[gate,smoke]
  2013-09-25 16:03:28.319 | 
--
  2013-09-25 16:03:28.319 | _StringException: Empty attachments:
  2013-09-25 16:03:28.319 |   stderr
  2013-09-25 16:03:28.320 |   stdout
  2013-09-25 16:03:28.320 |
  2013-09-25 16:03:28.320 | pythonlogging:'': {{{2013-09-25 15:49:34,792 state: 
pending}}}
  2013-09-25 16:03:28.320 |
  2013-09-25 16:03:28.320 | Traceback (most recent call last):
  2013-09-25 16:03:28.320 |   File 
"tempest/thirdparty/boto/test_ec2_instance_run.py", line 175, in 
test_run_stop_terminate_instance_with_tags
  2013-09-25 16:03:28.320 | self.assertInstanceStateWait(instance, 
"running")
  2013-09-25 16:03:28.321 |   File "tempest/thirdparty/boto/test.py", line 356, 
in assertInstanceStateWait
  2013-09-25 16:03:28.321 | state = self.waitInstanceState(lfunction, 
wait_for)
  2013-09-25 16:03:28.321 |   File "tempest/thirdparty/boto/test.py", line 341, 
in waitInstanceState
  2013-09-25 16:03:28.321 | self.valid_instance_state)
  2013-09-25 16:03:28.321 |   File "tempest/thirdparty/boto/test.py", line 331, 
in state_wait_gone
  2013-09-25 16:03:28.321 | state = state_wait(lfunction, final_set, 
valid_set)
  2013-09-25 16:03:28.322 |   File "tempest/thirdparty/boto/utils/wait.py", 
line 57, in state_wait
  2013-09-25 16:03:28.322 | (dtime, final_set, status))
  2013-09-25 16:03:28.322 | AssertionError: State change timeout 
exceeded!(400s) While waitingfor set(['running', '_GONE']) at "pending"

  full log: http://logs.openstack.org/38/47438/1/gate/gate-tempest-
  devstack-vm-neutron/93db162/

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1230407/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1230354] Re: tempest.scenario.test_minimum_basic.TestMinimumBasicScenario.test_minimum_basic_scenario fails sporadically

2013-12-12 Thread Sean Dague
This bug is basically useless. It doesn't get far enough to get to any
root causes. I'm marking this as invalid and the narrower bugs can be
refiled.

** No longer affects: cinder

** No longer affects: nova

** Changed in: tempest
   Status: Confirmed => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1230354

Title:
  
tempest.scenario.test_minimum_basic.TestMinimumBasicScenario.test_minimum_basic_scenario
  fails sporadically

Status in Tempest:
  Invalid

Bug description:
  See: http://logs.openstack.org/97/44097/18/check/gate-tempest-
  devstack-vm-postgres-full/270e611/console.html

  2013-09-25 15:15:37.447 | 
==
  2013-09-25 15:15:37.463 | FAIL: 
tempest.scenario.test_minimum_basic.TestMinimumBasicScenario.test_minimum_basic_scenario[compute,image,network,volume]
  2013-09-25 15:15:37.464 | 
tempest.scenario.test_minimum_basic.TestMinimumBasicScenario.test_minimum_basic_scenario[compute,image,network,volume]
  2013-09-25 15:15:37.464 | 
--
  2013-09-25 15:15:37.464 | _StringException: Empty attachments:
  2013-09-25 15:15:37.464 |   stderr
  2013-09-25 15:15:37.464 |   stdout
  2013-09-25 15:15:37.464 | 
  2013-09-25 15:15:37.465 | pythonlogging:'': {{{
  2013-09-25 15:15:37.465 | 2013-09-25 15:07:17,364 Starting new HTTP 
connection (1): 127.0.0.1
  2013-09-25 15:15:37.465 | 2013-09-25 15:07:17,628 Starting new HTTP 
connection (1): 127.0.0.1
  2013-09-25 15:15:37.465 | 2013-09-25 15:07:29,212 Starting new HTTP 
connection (1): 127.0.0.1
  .
  .
  .
  2013-09-25 15:15:37.619 | 2013-09-25 15:14:11,445 Starting new HTTP 
connection (1): 127.0.0.1
  2013-09-25 15:15:37.619 | 2013-09-25 15:14:12,529 Starting new HTTP 
connection (1): 127.0.0.1
  2013-09-25 15:15:37.620 | 2013-09-25 15:14:13,617 Starting new HTTP 
connection (1): 127.0.0.1
  2013-09-25 15:15:37.620 | }}}
  2013-09-25 15:15:37.620 | 
  2013-09-25 15:15:37.620 | Traceback (most recent call last):
  2013-09-25 15:15:37.621 |   File "tempest/scenario/test_minimum_basic.py", 
line 158, in test_minimum_basic_scenario
  2013-09-25 15:15:37.621 | self.nova_volume_attach()
  2013-09-25 15:15:37.621 |   File "tempest/scenario/test_minimum_basic.py", 
line 120, in nova_volume_attach
  2013-09-25 15:15:37.622 | self._wait_for_volume_status('in-use')
  2013-09-25 15:15:37.622 |   File "tempest/scenario/test_minimum_basic.py", 
line 48, in _wait_for_volume_status
  2013-09-25 15:15:37.622 | self.volume_client.volumes, volume_id, status)
  2013-09-25 15:15:37.622 |   File "tempest/scenario/manager.py", line 290, in 
status_timeout
  2013-09-25 15:15:37.622 | self._status_timeout(things, thing_id, 
expected_status=expected_status)
  2013-09-25 15:15:37.623 |   File "tempest/scenario/manager.py", line 341, in 
_status_timeout
  2013-09-25 15:15:37.623 | raise exceptions.TimeoutException(message)
  2013-09-25 15:15:37.623 | TimeoutException: Request timed out
  2013-09-25 15:15:37.623 | Details: Timed out waiting for thing 
b5781471-5ee9-44de-944b-28c10a793b31   to become in-use

To manage notifications about this bug go to:
https://bugs.launchpad.net/tempest/+bug/1230354/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1239891] Re: tempest.api.object_storage.test_account_services.AccountTest fails under neutron-pg-isolated

2013-12-12 Thread Attila Fazekas
** Project changed: tempest => neutron

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1239891

Title:
  tempest.api.object_storage.test_account_services.AccountTest fails
  under neutron-pg-isolated

Status in OpenStack Neutron (virtual network service):
  New

Bug description:
  http://logs.openstack.org/38/51738/1/check/check-tempest-devstack-vm-
  neutron-pg-isolated/73aad7a/console.html

  2013-10-15 00:19:04.556 | Error in atexit._run_exitfuncs:
  2013-10-15 00:19:04.556 | Traceback (most recent call last):
  2013-10-15 00:19:04.556 |   File "/usr/lib/python2.7/atexit.py", line 24, in 
_run_exitfuncs
  2013-10-15 00:19:04.557 | func(*targs, **kargs)
  2013-10-15 00:19:04.558 |   File "tempest/test.py", line 167, in 
validate_tearDownClass
  2013-10-15 00:19:04.558 | + str(at_exit_set))
  2013-10-15 00:19:04.558 | RuntimeError: tearDownClass does not calls the 
super's tearDownClass in these classes: set([])
  2013-10-15 00:19:04.559 | Error in sys.exitfunc:
  2013-10-15 00:19:04.663 | 
  2013-10-15 00:19:04.664 | process-returncode
  2013-10-15 00:19:04.664 | process-returncode ... FAIL
  2013-10-15 00:19:04.980 | 
  2013-10-15 00:19:04.981 | 
==
  2013-10-15 00:19:04.981 | FAIL: tearDownClass 
(tempest.api.object_storage.test_account_services.AccountTest)
  2013-10-15 00:19:04.981 | tearDownClass 
(tempest.api.object_storage.test_account_services.AccountTest)
  2013-10-15 00:19:04.982 | 
--
  2013-10-15 00:19:04.982 | _StringException: Traceback (most recent call last):
  2013-10-15 00:19:04.982 |   File 
"tempest/api/object_storage/test_account_services.py", line 41, in tearDownClass
  2013-10-15 00:19:04.983 | super(AccountTest, cls).tearDownClass()
  2013-10-15 00:19:04.983 |   File "tempest/api/object_storage/base.py", line 
77, in tearDownClass
  2013-10-15 00:19:04.983 | cls.isolated_creds.clear_isolated_creds()
  2013-10-15 00:19:04.984 |   File "tempest/common/isolated_creds.py", line 
453, in clear_isolated_creds
  2013-10-15 00:19:04.984 | self._clear_isolated_net_resources()
  2013-10-15 00:19:04.984 |   File "tempest/common/isolated_creds.py", line 
445, in _clear_isolated_net_resources
  2013-10-15 00:19:04.985 | self._clear_isolated_network(network['id'], 
network['name'])
  2013-10-15 00:19:04.985 |   File "tempest/common/isolated_creds.py", line 
399, in _clear_isolated_network
  2013-10-15 00:19:04.985 | net_client.delete_network(network_id)
  2013-10-15 00:19:04.985 |   File 
"tempest/services/network/json/network_client.py", line 76, in delete_network
  2013-10-15 00:19:04.986 | resp, body = self.delete(uri, self.headers)
  2013-10-15 00:19:04.986 |   File "tempest/common/rest_client.py", line 308, 
in delete
  2013-10-15 00:19:04.986 | return self.request('DELETE', url, headers)
  2013-10-15 00:19:04.987 |   File "tempest/common/rest_client.py", line 436, 
in request
  2013-10-15 00:19:04.987 | resp, resp_body)
  2013-10-15 00:19:04.987 |   File "tempest/common/rest_client.py", line 522, 
in _error_checker
  2013-10-15 00:19:04.988 | raise exceptions.ComputeFault(message)
  2013-10-15 00:19:04.988 | ComputeFault: Got compute fault
  2013-10-15 00:19:04.988 | Details: {"NeutronError": "Request Failed: internal 
server error while processing your request."}
  2013-10-15 00:19:04.988 | 
  2013-10-15 00:19:04.989 | 
  2013-10-15 00:19:04.989 | 
==
  2013-10-15 00:19:04.989 | FAIL: process-returncode
  2013-10-15 00:19:04.990 | process-returncode
  2013-10-15 00:19:04.990 | 
--
  2013-10-15 00:19:04.990 | _StringException: Binary content:
  2013-10-15 00:19:04.991 |   traceback (test/plain; charset="utf8")
  2013-10-15 00:19:04.991 | 
  2013-10-15 00:19:04.991 | 
  2013-10-15 00:19:04.991 | 
--

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1239891/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260075] Re: VMware: NotAuthenticated occurred in the call to RetrievePropertiesEx

2013-12-12 Thread Shawn Hartsock
Marked High because this impacts the CI environment we use to approve
other patches. If not for that context, I would mark this Medium.

** Changed in: nova
   Status: New => Confirmed

** Changed in: nova
   Importance: Undecided => High

** Changed in: nova
 Assignee: (unassigned) => Sabari Kumar Murugesan (smurugesan)

** Changed in: nova
Milestone: None => icehouse-2

** Also affects: openstack-vmwareapi-team
   Importance: Undecided
   Status: New

** Changed in: openstack-vmwareapi-team
   Status: New => Confirmed

** Changed in: openstack-vmwareapi-team
   Importance: Undecided => High

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1260075

Title:
  VMware: NotAuthenticated occurred in the call to RetrievePropertiesEx

Status in OpenStack Compute (Nova):
  Confirmed
Status in The OpenStack VMwareAPI subTeam:
  Confirmed

Bug description:
  The VMware Minesweeper CI occasionally runs into this error when
  trying to boot an instance:

  2013-12-11 04:50:15.048 20785 DEBUG nova.virt.vmwareapi.driver [-] Task 
[ReconfigVM_Task] (returnval){
 value = "task-322"
 _type = "Task"
   } status: success _poll_task 
/opt/stack/nova/nova/virt/vmwareapi/driver.py:926
  Reconfigured VM instance to enable vnc on port - 5986 _set_vnc_config 
/opt/stack/nova/nova/virt/vmwareapi/vmops.py:1461
  Instance failed to spawn
  Traceback (most recent call last):
File "/opt/stack/nova/nova/compute/manager.py", line 1461, in _spawn
  block_device_info)
File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 628, in spawn
  admin_password, network_info, block_device_info)
File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 435, in spawn
  upload_folder, upload_name + ".vmdk")):
File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1556, in 
_check_if_folder_file_exists
  "browser")
File "/opt/stack/nova/nova/virt/vmwareapi/vim_util.py", line 173, in 
get_dynamic_property
  property_dict = get_dynamic_properties(vim, mobj, type, [property_name])
File "/opt/stack/nova/nova/virt/vmwareapi/vim_util.py", line 179, in 
get_dynamic_properties
  obj_content = get_object_properties(vim, None, mobj, type, property_names)
File "/opt/stack/nova/nova/virt/vmwareapi/vim_util.py", line 168, in 
get_object_properties
  options=options)
File "/opt/stack/nova/nova/virt/vmwareapi/vim.py", line 187, in 
vim_request_handler
  fault_checker(response)
File "/opt/stack/nova/nova/virt/vmwareapi/error_util.py", line 99, in 
retrievepropertiesex_fault_checker
  exc_msg_list))
  VimFaultException: Error(s) NotAuthenticated occurred in the call to 
RetrievePropertiesEx

  Full logs here for a CI build where this occurred are available here:
  http://162.209.83.206/logs/35303/31/

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1260075/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1252854] Re: setUpClass ServerAddressesTest FAIL

2013-12-12 Thread Attila Fazekas
Very unlikely tempest caused the ERROR status on the server, so adding
nova.

** Also affects: nova
   Importance: Undecided
   Status: New

** Changed in: tempest
   Status: New => Incomplete

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1252854

Title:
  setUpClass ServerAddressesTest FAIL

Status in OpenStack Compute (Nova):
  New
Status in Tempest:
  Incomplete

Bug description:
  I encountered the following Tempest failure during the initial Jenkins
  tests for the following submission to python-cinderclient:

  https://review.openstack.org/#/c/57245/

  http://logs.openstack.org/45/57245/1/check/check-grenade-devstack-
  vm/1ad2c17/

  2013-11-19 16:07:09.167 | 
==
  2013-11-19 16:07:09.167 | FAIL: setUpClass 
(tempest.api.compute.servers.test_server_addresses.ServerAddressesTest)
  2013-11-19 16:07:09.167 | setUpClass 
(tempest.api.compute.servers.test_server_addresses.ServerAddressesTest)
  2013-11-19 16:07:09.168 | 
--
  2013-11-19 16:07:09.168 | _StringException: Traceback (most recent call last):
  2013-11-19 16:07:09.168 |   File 
"tempest/api/compute/servers/test_server_addresses.py", line 31, in setUpClass
  2013-11-19 16:07:09.169 | resp, cls.server = 
cls.create_test_server(wait_until='ACTIVE')
  2013-11-19 16:07:09.169 |   File "tempest/api/compute/base.py", line 118, in 
create_test_server
  2013-11-19 16:07:09.169 | server['id'], kwargs['wait_until'])
  2013-11-19 16:07:09.169 |   File 
"tempest/services/compute/json/servers_client.py", line 160, in 
wait_for_server_status
  2013-11-19 16:07:09.170 | extra_timeout=extra_timeout)
  2013-11-19 16:07:09.170 |   File "tempest/common/waiters.py", line 73, in 
wait_for_server_status
  2013-11-19 16:07:09.170 | raise 
exceptions.BuildErrorException(server_id=server_id)
  2013-11-19 16:07:09.171 | BuildErrorException: Server 
e4f08fa8-bf4c-4994-b5f9-97566a393baf failed to build and is in ERROR status

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1252854/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1239923] Re: neutron doesn't use request-ids

2013-12-12 Thread Sean Dague
not a tempest bug

** No longer affects: tempest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1239923

Title:
  neutron doesn't use request-ids

Status in OpenStack Neutron (virtual network service):
  In Progress

Bug description:
  Just like nova and cinder, neutron should use request-ids for logging
  and to return to users.

  request-ids are used to help look up the logs for a specific request.

  Also they make logs more usable.  Here is an example of how they can
  be used. http://git.openstack.org/cgit/openstack/oslo-
  incubator/tree/openstack/common/context.py

  
etc/cinder/cinder.conf.sample:#logging_context_format_string=%(asctime)s.%(msecs)03d
  %(process)d %(levelname)s %(name)s [%(request_id)s %(user)s %(

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1239923/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1242645] Re: Resource tracking does not take into account the current resources on the host

2013-12-12 Thread Gary Kotton
** Changed in: nova
 Assignee: Gary Kotton (garyk) => (unassigned)

** Changed in: nova
   Status: In Progress => Won't Fix

** Changed in: nova
Milestone: icehouse-2 => None

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1242645

Title:
  Resource tracking does not take into account the current resources on
  the host

Status in OpenStack Compute (Nova):
  Won't Fix

Bug description:
  The trace below is from the log file:

  2013-10-21 04:28:27.210 INFO nova.compute.resource_tracker [-] Hypervisor: 
free disk (GB): 11
  2013-10-21 04:28:27.263 AUDIT nova.compute.resource_tracker [-] Free ram 
(MB): 4898
  2013-10-21 04:28:27.263 AUDIT nova.compute.resource_tracker [-] Free disk 
(GB): 15
  2013-10-21 04:28:27.263 AUDIT nova.compute.resource_tracker [-] Free VCPUS: 4

  In this specific case there is 4GB used on the hypervisor - used for
  image cache etc.

  
+--+-+
  | Property | Value
   |
  
+--+-+
  | hypervisor_hostname  | domain-c7(Cluster31) 
   |
  | cpu_info | {"model": ["Intel(R) Xeon(R) CPU E5-2650 0 @ 
2.00GHz"], "vendor": ["VMware, Inc."], "topology": {"cores": 4, "threads": 4}} |
  | free_disk_gb | 15   
   |
  | hypervisor_version   | 51   
   |
  | disk_available_least | None 
   |
  | local_gb | 15   
   |
  | free_ram_mb  | 4898 
   |
  | id   | 1
   |
  | vcpus_used   | 0
   |
  | hypervisor_type  | VMware vCenter Server
   |
  | local_gb_used| 0
   |
  | memory_mb_used   | 512  
   |
  | memory_mb| 5410 
   |
  | current_workload | 0
   |
  | vcpus| 4
   |
  | running_vms  | 0
   |
  | service_id   | 5
   |
  | service_host | os-devstack  
   |
  
+--+-+

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1242645/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1226412] Re: tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML.test_rescue_paused_instance

2013-12-12 Thread Sean Dague
This is a nova race bug

** Also affects: nova
   Importance: Undecided
   Status: New

** Changed in: nova
   Importance: Undecided => Medium

** Changed in: nova
   Status: New => Confirmed

** Summary changed:

- 
tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML.test_rescue_paused_instance
+ guest doesn't reach PAUSED state within 200s in the gate

** Changed in: tempest
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1226412

Title:
  guest doesn't reach PAUSED state within 200s in the gate

Status in OpenStack Compute (Nova):
  Confirmed
Status in Tempest:
  Invalid

Bug description:
  Tempest test fails
  
:tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML.test_rescue_paused_instance

  
  Traceback (most recent call last):
File "tempest/api/compute/servers/test_server_rescue.py", line 135, in 
test_rescue_paused_instance
  self.servers_client.wait_for_server_status(self.server_id, 'PAUSED')
File "tempest/services/compute/xml/servers_client.py", line 340, in 
wait_for_server_status
  return waiters.wait_for_server_status(self, server_id, status)
File "tempest/common/waiters.py", line 80, in wait_for_server_status
  raise exceptions.TimeoutException(message)
  TimeoutException: Request timed out
  Details: Server 8539b620-909c-46a6-9293-1b1add06a343 failed to reach PAUSED 
status within the required time (400 s). Current status: ACTIVE.

  
  see 
http://logs.openstack.org/55/46855/3/check/gate-tempest-devstack-vm-postgres-full/28acd2d/testr_results.html.gz

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1226412/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1240728] Re: tempest.api.compute.servers.test_server_rescue.ServerRescueTestJSON.test_rescued_vm_attach_volume is nondeterministic

2013-12-12 Thread Sean Dague
** Changed in: nova
   Status: New => Confirmed

** Changed in: nova
   Importance: Undecided => High

** Changed in: cinder
   Importance: Undecided => High

** Changed in: cinder
   Status: Invalid => Confirmed

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1240728

Title:
  
tempest.api.compute.servers.test_server_rescue.ServerRescueTestJSON.test_rescued_vm_attach_volume
  is nondeterministic

Status in Cinder:
  Confirmed
Status in OpenStack Compute (Nova):
  Confirmed
Status in Tempest:
  New

Bug description:
  Traceback (most recent call last):
    File "tempest/api/compute/servers/test_server_rescue.py", line 111, in 
_unrescue
  self.servers_client.wait_for_server_status(server_id, 'ACTIVE')
    File "tempest/services/compute/json/servers_client.py", line 156, in 
wait_for_server_status
  return waiters.wait_for_server_status(self, server_id, status)
    File "tempest/common/waiters.py", line 80, in wait_for_server_status
  raise exceptions.TimeoutException(message)
  TimeoutException: Request timed out
  Details: Server 802897a6-6793-4af2-9d84-8750be518380 failed to reach ACTIVE 
status within the required time (400 s). Current status: SHUTOFF.

  Sample failure: http://logs.openstack.org/51/52151/1/gate/gate-
  tempest-devstack-vm-full/6b393f5/

  Basic query for the failure string:

  
http://logstash.openstack.org/#eyJzZWFyY2giOiJAbWVzc2FnZTpcIkZBSUw6IHRlbXBlc3QuYXBpLmNvbXB1dGUuc2VydmVycy50ZXN0X3NlcnZlcl9yZXNjdWUuU2VydmVyUmVzY3VlVGVzdEpTT04udGVzdF9yZXNjdWVkX3ZtX2F0dGFjaF92b2x1bWVcIiIsImZpZWxkcyI6W10sIm9mZnNldCI6MCwidGltZWZyYW1lIjoiYWxsIiwiZ3JhcGhtb2RlIjoiY291bnQiLCJ0aW1lIjp7InVzZXJfaW50ZXJ2YWwiOjB9LCJzdGFtcCI6MTM4MTk2MTIyMjkwMSwibW9kZSI6IiIsImFuYWx5emVfZmllbGQiOiIifQ==

To manage notifications about this bug go to:
https://bugs.launchpad.net/cinder/+bug/1240728/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260406] [NEW] allow disable wsgi keepalive

2013-12-12 Thread Edward Hope-Morley
Public bug reported:

The wsgi server used in most if not  all openstack services currently
has keepalive=True by default thus keeping connections open after each
request. This can cause problems when using load balancers in front of
these services e.g. connections for requests that take a long time can
get closed in the load balancers once a timeout has expired. This can
then cause issues if a client performs a request using the same source
port as a previous request that is not closed in the LB but still open
in the server due to TCP packet sequecing in the LB and the new client
not expecting the connection to already be open. So, it would be useful
to be able to disable wsgi keepalive.

** Affects: keystone
 Importance: Undecided
 Assignee: Edward Hope-Morley (hopem)
 Status: In Progress

** Changed in: keystone
 Assignee: (unassigned) => Edward Hope-Morley (hopem)

** Changed in: keystone
   Status: New => In Progress

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Keystone.
https://bugs.launchpad.net/bugs/1260406

Title:
  allow disable wsgi keepalive

Status in OpenStack Identity (Keystone):
  In Progress

Bug description:
  The wsgi server used in most if not  all openstack services currently
  has keepalive=True by default thus keeping connections open after each
  request. This can cause problems when using load balancers in front of
  these services e.g. connections for requests that take a long time can
  get closed in the load balancers once a timeout has expired. This can
  then cause issues if a client performs a request using the same source
  port as a previous request that is not closed in the LB but still open
  in the server due to TCP packet sequecing in the LB and the new client
  not expecting the connection to already be open. So, it would be
  useful to be able to disable wsgi keepalive.

To manage notifications about this bug go to:
https://bugs.launchpad.net/keystone/+bug/1260406/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1242916] Re: metadata server update_all expects body but doesn't get it passed to it

2013-12-12 Thread Sean Dague
** No longer affects: tempest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1242916

Title:
  metadata server update_all expects body but doesn't get it passed to
  it

Status in OpenStack Compute (Nova):
  Confirmed

Bug description:
  This recently started showing up in n-api. Seems like mishandling of invalid 
client args from a negative test in tempest.
  Example: 
http://logs.openstack.org/03/52803/3/check/check-tempest-devstack-vm-postgres-full/354d7a3/logs/screen-n-api.txt.gz

  2013-10-21 20:04:51.724 20923 DEBUG routes.middleware [-] Matched PUT 
/6fa344aaf3034c4992bc30b3c06ad531/servers/a329912b-7874-4636-a08b-e40362e04ab2/metadata
 __call__ /usr/lib/python2.7/dist-packages/routes/middleware.py:100
  2013-10-21 20:04:51.724 20923 DEBUG routes.middleware [-] Route path: 
'/{project_id}/servers/{server_id}/metadata', defaults: {'action': 
u'update_all', 'controller': } __call__ /usr/lib/python2.7/dist-packages/routes/middleware.py:102
  2013-10-21 20:04:51.724 20923 DEBUG routes.middleware [-] Match dict: 
{'action': u'update_all', 'server_id': u'a329912b-7874-4636-a08b-e40362e04ab2', 
'project_id': u'6fa344aaf3034c4992bc30b3c06ad531', 'controller': 
} __call__ 
/usr/lib/python2.7/dist-packages/routes/middleware.py:103
  2013-10-21 20:04:51.724 DEBUG nova.api.openstack.wsgi 
[req-23bdfb52-dae5-42f7-a5b7-17c319ed67ff 
ServerMetadataTestJSON-tempest-1354548727-user 
ServerMetadataTestJSON-tempest-1354548727-tenant] Empty body provided in 
request get_body /opt/stack/new/nova/nova/api/openstack/wsgi.py:839
  2013-10-21 20:04:51.724 DEBUG nova.api.openstack.wsgi 
[req-23bdfb52-dae5-42f7-a5b7-17c319ed67ff 
ServerMetadataTestJSON-tempest-1354548727-user 
ServerMetadataTestJSON-tempest-1354548727-tenant] Calling method > _process_stack 
/opt/stack/new/nova/nova/api/openstack/wsgi.py:962
  2013-10-21 20:04:51.725 ERROR nova.api.openstack.wsgi 
[req-23bdfb52-dae5-42f7-a5b7-17c319ed67ff 
ServerMetadataTestJSON-tempest-1354548727-user 
ServerMetadataTestJSON-tempest-1354548727-tenant] Exception handling resource: 
update_all() takes exactly 4 arguments (3 given)
  2013-10-21 20:04:51.725 20923 TRACE nova.api.openstack.wsgi Traceback (most 
recent call last):
  2013-10-21 20:04:51.725 20923 TRACE nova.api.openstack.wsgi   File 
"/opt/stack/new/nova/nova/api/openstack/wsgi.py", line 997, in _process_stack
  2013-10-21 20:04:51.725 20923 TRACE nova.api.openstack.wsgi action_result 
= self.dispatch(meth, request, action_args)
  2013-10-21 20:04:51.725 20923 TRACE nova.api.openstack.wsgi   File 
"/opt/stack/new/nova/nova/api/openstack/wsgi.py", line 1078, in dispatch
  2013-10-21 20:04:51.725 20923 TRACE nova.api.openstack.wsgi return 
method(req=request, **action_args)
  2013-10-21 20:04:51.725 20923 TRACE nova.api.openstack.wsgi TypeError: 
update_all() takes exactly 4 arguments (3 given)
  2013-10-21 20:04:51.725 20923 TRACE nova.api.openstack.wsgi 
  2013-10-21 20:04:51.726 DEBUG nova.api.openstack.wsgi 
[req-23bdfb52-dae5-42f7-a5b7-17c319ed67ff 
ServerMetadataTestJSON-tempest-1354548727-user 
ServerMetadataTestJSON-tempest-1354548727-tenant] Returning 400 to user: The 
server could not comply with the request since it is either malformed or 
otherwise incorrect. __call__ 
/opt/stack/new/nova/nova/api/openstack/wsgi.py:1224

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1242916/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260423] [NEW] Email shouldn't be a mandatory attribute

2013-12-12 Thread Julie Pichon
Public bug reported:

When using a LDAP backend, it's possible that a user won't have the
"email" attribute defined, however it should still be possible to edit
the other fields.

Steps to reproduce (in an environment with keystone using a LDAP backend):
1. Log in as admin
2. Go to the Users dashboard
3. Select a user that doesn't have an email defined

Expected result:
4. "Edit user" modal opens

Actual result:
4. Error 500

Traceback:
File "/usr/lib/python2.7/site-packages/django/views/generic/edit.py" in get
  154. form = self.get_form(form_class)
File "/opt/stack/horizon/openstack_dashboard/wsgi/../../horizon/forms/views.py" 
in get_form
  82. return form_class(self.request, **self.get_form_kwargs())
File "/usr/lib/python2.7/site-packages/django/views/generic/edit.py" in 
get_form_kwargs
  41. kwargs = {'initial': self.get_initial()}
File 
"/opt/stack/horizon/openstack_dashboard/wsgi/../../openstack_dashboard/dashboards/admin/users/views.py"
 in get_initial
  103. 'email': user.email}
File "/opt/stack/python-keystoneclient/keystoneclient/base.py" in __getattr__
  425. raise AttributeError(k)

Exception Type: AttributeError at 
/admin/users/e005aa43475b403c8babdff86ea27c37/update/
Exception Value: email

** Affects: horizon
 Importance: Medium
 Assignee: Julie Pichon (jpichon)
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1260423

Title:
  Email shouldn't be a mandatory attribute

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  When using a LDAP backend, it's possible that a user won't have the
  "email" attribute defined, however it should still be possible to edit
  the other fields.

  Steps to reproduce (in an environment with keystone using a LDAP backend):
  1. Log in as admin
  2. Go to the Users dashboard
  3. Select a user that doesn't have an email defined

  Expected result:
  4. "Edit user" modal opens

  Actual result:
  4. Error 500

  Traceback:
  File "/usr/lib/python2.7/site-packages/django/views/generic/edit.py" in get
154. form = self.get_form(form_class)
  File 
"/opt/stack/horizon/openstack_dashboard/wsgi/../../horizon/forms/views.py" in 
get_form
82. return form_class(self.request, **self.get_form_kwargs())
  File "/usr/lib/python2.7/site-packages/django/views/generic/edit.py" in 
get_form_kwargs
41. kwargs = {'initial': self.get_initial()}
  File 
"/opt/stack/horizon/openstack_dashboard/wsgi/../../openstack_dashboard/dashboards/admin/users/views.py"
 in get_initial
103. 'email': user.email}
  File "/opt/stack/python-keystoneclient/keystoneclient/base.py" in __getattr__
425. raise AttributeError(k)

  Exception Type: AttributeError at 
/admin/users/e005aa43475b403c8babdff86ea27c37/update/
  Exception Value: email

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1260423/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1211338] Re: "Direct" vs. "direct" in impl_qpid

2013-12-12 Thread Russell Bryant
nova grizzly patch: https://review.openstack.org/#/c/61831/

** Also affects: nova
   Importance: Undecided
   Status: New

** Changed in: nova
   Status: New => Fix Released

** Also affects: nova/grizzly
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1211338

Title:
  "Direct" vs. "direct" in impl_qpid

Status in OpenStack Compute (Nova):
  Fix Released
Status in OpenStack Compute (nova) grizzly series:
  New
Status in Oslo - a Library of Common OpenStack Code:
  Fix Released

Bug description:
  impl_qpid.py has {"type": "Direct"} (with a capital D) in one place.
  "direct" (lowercase) in others.  It appears that qpid is case-
  sensitive about exchange types, so the version with the capital D is
  invalid.  This ends up causing qpid to throw an error like:

  >> "/usr/lib/python2.6/site-packages/qpid/messaging/endpoints.py",
  >> line 567, in _ewait\nself.check_error()\n', '  File
  >> "/usr/lib/python2.6/site-packages/qpid/messaging/endpoints.py",
  >> line 556, in check_error\nraise self.error\n', 'NotFound:
  >> not-found: Exchange type not implemented: Direct
  >> (qpid/broker/SessionAdapter.cpp:117)(404)\n']

  It should be a one-character fix.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1211338/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260423] Re: Email shouldn't be a mandatory attribute

2013-12-12 Thread Julie Pichon
** Also affects: horizon/havana
   Importance: Undecided
   Status: New

** Changed in: horizon/havana
   Importance: Undecided => Medium

** Changed in: horizon/havana
 Assignee: (unassigned) => Julie Pichon (jpichon)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1260423

Title:
  Email shouldn't be a mandatory attribute

Status in OpenStack Dashboard (Horizon):
  In Progress
Status in OpenStack Dashboard (Horizon) havana series:
  New

Bug description:
  When using a LDAP backend, it's possible that a user won't have the
  "email" attribute defined, however it should still be possible to edit
  the other fields.

  Steps to reproduce (in an environment with keystone using a LDAP backend):
  1. Log in as admin
  2. Go to the Users dashboard
  3. Select a user that doesn't have an email defined

  Expected result:
  4. "Edit user" modal opens

  Actual result:
  4. Error 500

  Traceback:
  File "/usr/lib/python2.7/site-packages/django/views/generic/edit.py" in get
154. form = self.get_form(form_class)
  File 
"/opt/stack/horizon/openstack_dashboard/wsgi/../../horizon/forms/views.py" in 
get_form
82. return form_class(self.request, **self.get_form_kwargs())
  File "/usr/lib/python2.7/site-packages/django/views/generic/edit.py" in 
get_form_kwargs
41. kwargs = {'initial': self.get_initial()}
  File 
"/opt/stack/horizon/openstack_dashboard/wsgi/../../openstack_dashboard/dashboards/admin/users/views.py"
 in get_initial
103. 'email': user.email}
  File "/opt/stack/python-keystoneclient/keystoneclient/base.py" in __getattr__
425. raise AttributeError(k)

  Exception Type: AttributeError at 
/admin/users/e005aa43475b403c8babdff86ea27c37/update/
  Exception Value: email

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1260423/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260432] [NEW] nova-compute can't be setting up during install on trusty

2013-12-12 Thread Ming Lei
Public bug reported:


1, during install:
Setting up nova-compute (1:2014.1~b1-0ubuntu2) ...
start: Job failed to start
invoke-rc.d: initscript nova-compute, action "start" failed.
dpkg: error processing nova-compute (--configure):
 subprocess installed post-installation script returned error exit status 1
Setting up nova-compute-kvm (1:2014.1~b1-0ubuntu2) ...
Errors were encountered while processing:
 nova-compute
E: Sub-process /usr/bin/dpkg returned an error code (1)

2, the system is latest trusty:
ming@arm64:~$ sudo apt-get dist-upgrade
Reading package lists... Done
Building dependency tree   
Reading state information... Done
Calculating upgrade... Done
The following packages were automatically installed and are no longer required:
  dnsmasq-utils iputils-arping libboost-system1.53.0 libboost-thread1.53.0
  libclass-isa-perl libopts25 libswitch-perl ttf-dejavu-core
Use 'apt-get autoremove' to remove them.
The following packages have been kept back:
  checkbox-cli
0 upgraded, 0 newly installed, 0 to remove and 1 not upgraded.

3, looks /usr/bin/nova-compute can't be started:
ming@arm64:~$ nova-compute 
2013-12-12 17:57:19.992 13823 ERROR stevedore.extension [-] Could not load 
'file': (WebOb 1.3 (/usr/lib/python2.7/dist-packages), 
Requirement.parse('WebOb>=1.2.3,<1.3'))
2013-12-12 17:57:19.993 13823 ERROR stevedore.extension [-] (WebOb 1.3 
(/usr/lib/python2.7/dist-packages), Requirement.parse('WebOb>=1.2.3,<1.3'))
2013-12-12 17:57:19.993 13823 TRACE stevedore.extension Traceback (most recent 
call last):
2013-12-12 17:57:19.993 13823 TRACE stevedore.extension   File 
"/usr/lib/python2.7/dist-packages/stevedore/extension.py", line 134, in 
_load_plugins
2013-12-12 17:57:19.993 13823 TRACE stevedore.extension invoke_kwds,
2013-12-12 17:57:19.993 13823 TRACE stevedore.extension   File 
"/usr/lib/python2.7/dist-packages/stevedore/extension.py", line 146, in 
_load_one_plugin
2013-12-12 17:57:19.993 13823 TRACE stevedore.extension plugin = ep.load()
2013-12-12 17:57:19.993 13823 TRACE stevedore.extension   File 
"/usr/lib/python2.7/dist-packages/pkg_resources.py", line 2107, in load
2013-12-12 17:57:19.993 13823 TRACE stevedore.extension if require: 
self.require(env, installer)
2013-12-12 17:57:19.993 13823 TRACE stevedore.extension   File 
"/usr/lib/python2.7/dist-packages/pkg_resources.py", line 2120, in require
2013-12-12 17:57:19.993 13823 TRACE stevedore.extension 
working_set.resolve(self.dist.requires(self.extras),env,installer)))
2013-12-12 17:57:19.993 13823 TRACE stevedore.extension   File 
"/usr/lib/python2.7/dist-packages/pkg_resources.py", line 580, in resolve
2013-12-12 17:57:19.993 13823 TRACE stevedore.extension raise 
VersionConflict(dist,req) # XXX put more info here
2013-12-12 17:57:19.993 13823 TRACE stevedore.extension VersionConflict: (WebOb 
1.3 (/usr/lib/python2.7/dist-packages), Requirement.parse('WebOb>=1.2.3,<1.3'))
2013-12-12 17:57:19.993 13823 TRACE stevedore.extension 
2013-12-12 17:57:20.133 13823 ERROR nova.virt.driver [-] Compute driver option 
required, but not specified

** Affects: nova
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1260432

Title:
  nova-compute can't be setting up during install on trusty

Status in OpenStack Compute (Nova):
  New

Bug description:
  
  1, during install:
  Setting up nova-compute (1:2014.1~b1-0ubuntu2) ...
  start: Job failed to start
  invoke-rc.d: initscript nova-compute, action "start" failed.
  dpkg: error processing nova-compute (--configure):
   subprocess installed post-installation script returned error exit status 1
  Setting up nova-compute-kvm (1:2014.1~b1-0ubuntu2) ...
  Errors were encountered while processing:
   nova-compute
  E: Sub-process /usr/bin/dpkg returned an error code (1)

  2, the system is latest trusty:
  ming@arm64:~$ sudo apt-get dist-upgrade
  Reading package lists... Done
  Building dependency tree   
  Reading state information... Done
  Calculating upgrade... Done
  The following packages were automatically installed and are no longer 
required:
dnsmasq-utils iputils-arping libboost-system1.53.0 libboost-thread1.53.0
libclass-isa-perl libopts25 libswitch-perl ttf-dejavu-core
  Use 'apt-get autoremove' to remove them.
  The following packages have been kept back:
checkbox-cli
  0 upgraded, 0 newly installed, 0 to remove and 1 not upgraded.

  3, looks /usr/bin/nova-compute can't be started:
  ming@arm64:~$ nova-compute 
  2013-12-12 17:57:19.992 13823 ERROR stevedore.extension [-] Could not load 
'file': (WebOb 1.3 (/usr/lib/python2.7/dist-packages), 
Requirement.parse('WebOb>=1.2.3,<1.3'))
  2013-12-12 17:57:19.993 13823 ERROR stevedore.extension [-] (WebOb 1.3 
(/usr/lib/python2.7/dist-packages), Requirement.parse('WebOb>=1.2.3,<1.3'))
  2013-12-12 17:57:19.993 13823 TRACE stevedore.ext

[Yahoo-eng-team] [Bug 1260435] [NEW] Edit N1K network profile

2013-12-12 Thread Abishek Subramanian
Public bug reported:

Fix issue with performing an N1K network profile update operation by ensuring 
only the editable fields are edited.
Also with respect to new additions in the neutron N1K plugin, ensure new fields 
are updated in the update section.

** Affects: horizon
 Importance: Undecided
 Assignee: Abishek Subramanian (absubram)
 Status: In Progress

** Changed in: horizon
   Status: New => In Progress

** Changed in: horizon
 Assignee: (unassigned) => Abishek Subramanian (absubram)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1260435

Title:
  Edit N1K network profile

Status in OpenStack Dashboard (Horizon):
  In Progress

Bug description:
  Fix issue with performing an N1K network profile update operation by ensuring 
only the editable fields are edited.
  Also with respect to new additions in the neutron N1K plugin, ensure new 
fields are updated in the update section.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1260435/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260436] [NEW] Multi-nic support with N1K plugin

2013-12-12 Thread Abishek Subramanian
Public bug reported:

When the cisco N1K neutron plugin is being used, an instance cannot be launched 
via Horizon with ability to have multiple nics.
Only the first network is used for all created nics.

This bug addressed that issue.

** Affects: horizon
 Importance: Undecided
 Assignee: Abishek Subramanian (absubram)
 Status: In Progress

** Changed in: horizon
   Status: New => In Progress

** Changed in: horizon
 Assignee: (unassigned) => Abishek Subramanian (absubram)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1260436

Title:
  Multi-nic support with N1K plugin

Status in OpenStack Dashboard (Horizon):
  In Progress

Bug description:
  When the cisco N1K neutron plugin is being used, an instance cannot be 
launched via Horizon with ability to have multiple nics.
  Only the first network is used for all created nics.

  This bug addressed that issue.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1260436/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260437] [NEW] Project name display when editing N1K profile

2013-12-12 Thread Abishek Subramanian
Public bug reported:

When editing an N1K network profile, the project already associated with
the network profile is not displayed.

** Affects: horizon
 Importance: Undecided
 Assignee: Abishek Subramanian (absubram)
 Status: In Progress

** Changed in: horizon
 Assignee: (unassigned) => Abishek Subramanian (absubram)

** Changed in: horizon
   Status: New => In Progress

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1260437

Title:
  Project name display when editing N1K profile

Status in OpenStack Dashboard (Horizon):
  In Progress

Bug description:
  When editing an N1K network profile, the project already associated
  with the network profile is not displayed.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1260437/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260439] [NEW] Error message when creating a user with the default member role

2013-12-12 Thread Guilherme Birk
Public bug reported:

When you create a new user with the same role, from the roles list, as
the default member role set in keystone.conf, Horizon shows the error
message "Unable to add user to primary project". Even with this message,
the user is created and the member role is granted to the user in the
selected project.

If you look the keystone.log file, you will see the following warning:

" WARNING keystone.common.wsgi [-] Conflict occurred attempting to store
role grant. User  already has role  in tenant
"

Example:
I have the "member_role_name = _member_" entry in my keystone.conf file.
When I create a User called "Test_User" in the project "Test_Project" with the 
role "_member_", the error message is shown.

** Affects: horizon
 Importance: Undecided
 Status: New


** Tags: error horizon message role

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1260439

Title:
  Error message when creating a user with the default member role

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  When you create a new user with the same role, from the roles list, as
  the default member role set in keystone.conf, Horizon shows the error
  message "Unable to add user to primary project". Even with this
  message, the user is created and the member role is granted to the
  user in the selected project.

  If you look the keystone.log file, you will see the following warning:

  " WARNING keystone.common.wsgi [-] Conflict occurred attempting to
  store role grant. User  already has role  in tenant
  "

  Example:
  I have the "member_role_name = _member_" entry in my keystone.conf file.
  When I create a User called "Test_User" in the project "Test_Project" with 
the role "_member_", the error message is shown.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1260439/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260438] [NEW] Display N1K network profile information in network

2013-12-12 Thread Abishek Subramanian
Public bug reported:

When an N1K profile is associated with a network, currently the N1K
profile information is not displayed in the network detail.

** Affects: horizon
 Importance: Undecided
 Assignee: Abishek Subramanian (absubram)
 Status: In Progress

** Changed in: horizon
   Status: New => In Progress

** Changed in: horizon
 Assignee: (unassigned) => Abishek Subramanian (absubram)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1260438

Title:
  Display N1K network profile information in network

Status in OpenStack Dashboard (Horizon):
  In Progress

Bug description:
  When an N1K profile is associated with a network, currently the N1K
  profile information is not displayed in the network detail.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1260438/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260440] [NEW] nova-compute host is added to scheduling pool before Neutron can bind network ports on said host

2013-12-12 Thread Clint Byrum
Public bug reported:

This is a race condition.

Given a cloud with 0 compute nodes available, on a compute node:
* Start up neutron-openvswitch-agent
* Start up nova-compute
* nova boot an instance

Scenario 1:
* neutron-openvswitch-agent registers with Neutron before nova tries to boot 
instance
* port is bound to agent
* instance boots with correct networking

Scenario 2:
* nova schedules instance to host before neutron-openvswitch-agent is 
registered with Neutron
* nova instance fails with vif_type=binding_failed
* instance is in ERROR state

I would expect that Nova would not try to schedule instances on compute
hosts that are not ready.

Please also see this mailing list thread for more info:

http://lists.openstack.org/pipermail/openstack-
dev/2013-December/022084.html

** Affects: nova
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1260440

Title:
  nova-compute host is added to scheduling pool before Neutron can bind
  network ports on said host

Status in OpenStack Compute (Nova):
  New

Bug description:
  This is a race condition.

  Given a cloud with 0 compute nodes available, on a compute node:
  * Start up neutron-openvswitch-agent
  * Start up nova-compute
  * nova boot an instance

  Scenario 1:
  * neutron-openvswitch-agent registers with Neutron before nova tries to boot 
instance
  * port is bound to agent
  * instance boots with correct networking

  Scenario 2:
  * nova schedules instance to host before neutron-openvswitch-agent is 
registered with Neutron
  * nova instance fails with vif_type=binding_failed
  * instance is in ERROR state

  I would expect that Nova would not try to schedule instances on
  compute hosts that are not ready.

  Please also see this mailing list thread for more info:

  http://lists.openstack.org/pipermail/openstack-
  dev/2013-December/022084.html

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1260440/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1218279] Re: setUpClass (tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestJSON) Failed at server creation

2013-12-12 Thread Sean Dague
I actually think this was the enable/disable server race. I'm going to
close this for now. Reopen if an issue in the future

** Changed in: nova
   Status: New => Fix Released

** Changed in: tempest
   Status: New => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1218279

Title:
  setUpClass
  (tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestJSON)
  Failed at server creation

Status in OpenStack Compute (Nova):
  Fix Released
Status in Tempest:
  Fix Released

Bug description:
  http://logs.openstack.org/78/44178/1/gate/gate-tempest-devstack-vm-
  postgres-full/042f00e/

  2013-08-29 08:51:47.396 | FAIL: setUpClass 
(tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestJSON)
  2013-08-29 08:51:47.397 | setUpClass 
(tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestJSON)
  2013-08-29 08:51:47.397 | 
--
  2013-08-29 08:51:47.397 | _StringException: Traceback (most recent call last):
  2013-08-29 08:51:47.397 |   File 
"tempest/api/compute/images/test_images_oneserver.py", line 50, in setUpClass
  2013-08-29 08:51:47.397 | cls.tearDownClass()
  2013-08-29 08:51:47.397 |   File "tempest/api/compute/base.py", line 114, in 
tearDownClass
  2013-08-29 08:51:47.398 | super(BaseComputeTest, cls).tearDownClass()
  2013-08-29 08:51:47.398 |   File "tempest/test.py", line 144, in tearDownClass
  2013-08-29 08:51:47.398 | at_exit_set.remove(cls)
  2013-08-29 08:51:47.398 | KeyError: 

  The server creation was the actually failed.
  The setUpClass attempts to call the tearDownClass on error and throws a 
different exception,
  the correct exception throwing is tempest side issue, but the root cause 
probably not.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1218279/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260454] [NEW] Add cinder 'extend' volume functionality

2013-12-12 Thread Jesse Pretorius
Public bug reported:

Cinder now has the ability to 'extend' (ie grow/expand/resize up) a
volume. This functionality should be exposed through Horizon.

** Affects: horizon
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1260454

Title:
  Add cinder 'extend' volume functionality

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  Cinder now has the ability to 'extend' (ie grow/expand/resize up) a
  volume. This functionality should be exposed through Horizon.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1260454/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1181567] Re: tempest: test_create_server / wait_for_server_status timeout

2013-12-12 Thread Sean Dague
Not actually a tempest bug, this is a race in Nova

** Changed in: tempest
   Status: Incomplete => Invalid

** Summary changed:

- tempest: test_create_server / wait_for_server_status timeout
+ Nova compute guest still stuck in BUILD state after 400s

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1181567

Title:
  Nova compute guest still stuck in BUILD state after 400s

Status in OpenStack Compute (Nova):
  New
Status in Tempest:
  Invalid

Bug description:
  Failure occurred on https://review.openstack.org/#/c/29591/2

  http://logs.openstack.org/29591/2/gate/gate-tempest-devstack-vm-
  quantum/23189/console.html.gz

  2013-05-17 22:54:02.079 | 
==
  2013-05-17 22:54:02.079 | ERROR: test suite for 
  2013-05-17 22:54:02.079 | 
--
  2013-05-17 22:54:02.079 | Traceback (most recent call last):
  2013-05-17 22:54:02.079 |   File 
"/usr/lib/python2.7/dist-packages/nose/suite.py", line 208, in run
  2013-05-17 22:54:02.080 | self.setUp()
  2013-05-17 22:54:02.080 |   File 
"/usr/lib/python2.7/dist-packages/nose/suite.py", line 291, in setUp
  2013-05-17 22:54:02.080 | self.setupContext(ancestor)
  2013-05-17 22:54:02.080 |   File 
"/usr/lib/python2.7/dist-packages/nose/suite.py", line 314, in setupContext
  2013-05-17 22:54:02.080 | try_run(context, names)
  2013-05-17 22:54:02.080 |   File 
"/usr/lib/python2.7/dist-packages/nose/util.py", line 478, in try_run
  2013-05-17 22:54:02.080 | return func()
  2013-05-17 22:54:02.080 |   File 
"/opt/stack/new/tempest/tempest/tests/compute/servers/test_create_server.py", 
line 57, in setUpClass
  2013-05-17 22:54:02.080 | 
cls.client.wait_for_server_status(cls.server_initial['id'], 'ACTIVE')
  2013-05-17 22:54:02.081 |   File 
"/opt/stack/new/tempest/tempest/services/compute/xml/servers_client.py", line 
311, in wait_for_server_status
  2013-05-17 22:54:02.081 | raise exceptions.TimeoutException(message)
  2013-05-17 22:54:02.081 | TimeoutException: Request timed out
  2013-05-17 22:54:02.081 | Details: Request timed out
  2013-05-17 22:54:02.081 | Details: Server 
87c1dc14-44b1-406f-a7a0-c41876dc9111 failed to reach ACTIVE status within the 
required time (400 s). Current status: BUILD.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1181567/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1242898] Re: tearDownClass (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML): tearDownClass does not call the super's tearDownClass

2013-12-12 Thread Mauro Sergio Martins Rodrigues
if you look into http://logs.openstack.org/48/59948/4/check/check-
tempest-dsvm-neutron-pg-isolated/dab4997/logs/screen-q-lbaas.txt.gz will
see some tracebacks, it's probably an error in neutron.

Note that I get it from http://logs.openstack.org/48/59948/4/check
/check-tempest-dsvm-neutron-pg-isolated/dab4997/ not the original trace
which is already deleted now.

** Also affects: neutron
   Importance: Undecided
   Status: New

** Changed in: tempest
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1242898

Title:
  tearDownClass
  (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML):
  tearDownClass does not call the super's tearDownClass

Status in OpenStack Neutron (virtual network service):
  New
Status in Tempest:
  Invalid

Bug description:
  From http://logs.openstack.org/32/47432/16/check/check-tempest-
  devstack-vm-neutron-pg-isolated/c2a0dd3/console.html

  ...

  tearDownClass
  (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML)
  ... FAIL

  ...

  2013-10-21 19:17:53.068 | Error in atexit._run_exitfuncs:
  2013-10-21 19:17:53.068 | Traceback (most recent call last):
  2013-10-21 19:17:53.068 |   File "/usr/lib/python2.7/atexit.py", line 24, in 
_run_exitfuncs
  2013-10-21 19:17:53.069 | func(*targs, **kargs)
  2013-10-21 19:17:53.069 |   File "tempest/test.py", line 167, in 
validate_tearDownClass
  2013-10-21 19:17:53.069 | + str(at_exit_set))
  2013-10-21 19:17:53.069 | RuntimeError: tearDownClass does not calls the 
super's tearDownClass in these classes: set([])
  2013-10-21 19:17:53.070 | Error in sys.exitfunc:
  2013-10-21 19:17:53.221 | 
  2013-10-21 19:17:53.221 | process-returncode
  2013-10-21 19:17:53.221 | process-returncode ... FAIL
  2013-10-21 19:17:53.614 | 
  2013-10-21 19:17:53.614 | 
==
  2013-10-21 19:17:53.614 | FAIL: tearDownClass 
(tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML)
  2013-10-21 19:17:53.614 | tearDownClass 
(tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML)
  2013-10-21 19:17:53.614 | 
--
  2013-10-21 19:17:53.614 | _StringException: Traceback (most recent call last):
  2013-10-21 19:17:53.615 |   File 
"tempest/api/compute/servers/test_server_rescue.py", line 95, in tearDownClass
  2013-10-21 19:17:53.615 | super(ServerRescueTestJSON, cls).tearDownClass()
  2013-10-21 19:17:53.615 |   File "tempest/api/compute/base.py", line 132, in 
tearDownClass
  2013-10-21 19:17:53.615 | cls.isolated_creds.clear_isolated_creds()
  2013-10-21 19:17:53.615 |   File "tempest/common/isolated_creds.py", line 
453, in clear_isolated_creds
  2013-10-21 19:17:53.615 | self._clear_isolated_net_resources()
  2013-10-21 19:17:53.615 |   File "tempest/common/isolated_creds.py", line 
445, in _clear_isolated_net_resources
  2013-10-21 19:17:53.616 | self._clear_isolated_network(network['id'], 
network['name'])
  2013-10-21 19:17:53.616 |   File "tempest/common/isolated_creds.py", line 
399, in _clear_isolated_network
  2013-10-21 19:17:53.616 | net_client.delete_network(network_id)
  2013-10-21 19:17:53.616 |   File 
"tempest/services/network/json/network_client.py", line 76, in delete_network
  2013-10-21 19:17:53.616 | resp, body = self.delete(uri, self.headers)
  2013-10-21 19:17:53.616 |   File "tempest/common/rest_client.py", line 308, 
in delete
  2013-10-21 19:17:53.617 | return self.request('DELETE', url, headers)
  2013-10-21 19:17:53.617 |   File "tempest/common/rest_client.py", line 436, 
in request
  2013-10-21 19:17:53.617 | resp, resp_body)
  2013-10-21 19:17:53.617 |   File "tempest/common/rest_client.py", line 522, 
in _error_checker
  2013-10-21 19:17:53.617 | raise exceptions.ServerFault(message)
  2013-10-21 19:17:53.617 | ServerFault: Got server fault
  2013-10-21 19:17:53.617 | Details: {"NeutronError": "Request Failed: internal 
server error while processing your request."}
  2013-10-21 19:17:53.618 | 
  2013-10-21 19:17:53.618 | 
  2013-10-21 19:17:53.618 | 
==
  2013-10-21 19:17:53.618 | FAIL: process-returncode
  2013-10-21 19:17:53.619 | process-returncode
  2013-10-21 19:17:53.619 | 
--
  2013-10-21 19:17:53.619 | _StringException: Binary content:
  2013-10-21 19:17:53.619 |   traceback (test/plain; charset="utf8")
  2013-10-21 19:17:53.619 | 
  2013-10-21 19:17:53.619 | 
  2013-10-21 19:17:53.620 | 
--
  2013-10-21 19:17:53.620 | Ran 237 tests in 914.828s
  2013-10-21 19:17:53.637 | 
  2013-10-21 19:17:53.638 | FAILED (failures=2, skipped=8)

To manage notifications

[Yahoo-eng-team] [Bug 1258379] Re: vpnservice's router must have gateway interface set

2013-12-12 Thread Sean Dague
** No longer affects: tempest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1258379

Title:
  vpnservice's router must have gateway interface set

Status in OpenStack Neutron (virtual network service):
  In Progress

Bug description:
  at line
  
https://github.com/openstack/neutron/blob/master/neutron/services/vpn/service_drivers/ipsec.py#L172

  it is obvious the router must have gateway interface set  then it can
  be used as vpnservce router.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1258379/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1218190] Re: Use assertEqual instead of assertEquals in unitttest

2013-12-12 Thread Jeff Peeler
** Also affects: heat
   Importance: Undecided
   Status: New

** Changed in: heat
Milestone: None => icehouse-2

** Changed in: heat
 Assignee: (unassigned) => Jeff Peeler (jpeeler-z)

** Changed in: heat
   Status: New => In Progress

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Keystone.
https://bugs.launchpad.net/bugs/1218190

Title:
  Use assertEqual instead of assertEquals in unitttest

Status in Orchestration API (Heat):
  In Progress
Status in OpenStack Identity (Keystone):
  Fix Released
Status in Python client library for Keystone:
  Fix Committed
Status in Python client library for Neutron:
  Fix Committed

Bug description:
  I noticed that [keystone, python-keystoneclient, python-neutronclient]
  configure tox.ini with py33 test, however, assertEquals is deprecated
  in py3 but ok with py2, so i think it is better to change all of
  assertEquals to assertEqual

To manage notifications about this bug go to:
https://bugs.launchpad.net/heat/+bug/1218190/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1256043] Re: Need to add Development environment files to ignore list

2013-12-12 Thread Sean Dague
** No longer affects: tempest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1256043

Title:
  Need to add Development environment files to ignore list

Status in OpenStack Telemetry (Ceilometer):
  In Progress
Status in OpenStack Dashboard (Horizon):
  Won't Fix
Status in OpenStack Neutron (virtual network service):
  In Progress
Status in Python client library for Ceilometer:
  In Progress
Status in Python client library for Cinder:
  In Progress
Status in Python client library for Glance:
  In Progress
Status in Python client library for heat:
  Won't Fix
Status in Python client library for Keystone:
  Fix Committed
Status in Python client library for Neutron:
  In Progress
Status in Python client library for Nova:
  In Progress
Status in Python client library for Swift:
  Won't Fix
Status in OpenStack Object Storage (Swift):
  Won't Fix

Bug description:
  Following files generated by Eclipse development environment should be
  in ignore list to avoid their inclusion during a git push.

  .project
  .pydevproject

To manage notifications about this bug go to:
https://bugs.launchpad.net/ceilometer/+bug/1256043/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1193113] Re: DevicePathInUse exception in devstack-vm-quantum

2013-12-12 Thread Sean Dague
** No longer affects: tempest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1193113

Title:
  DevicePathInUse exception in devstack-vm-quantum

Status in OpenStack Compute (Nova):
  Confirmed

Bug description:
  I just got this during verification of one of my changes.  I don't
  think it's related to the change
  (https://review.openstack.org/#/c/33478/) so I'm reporting it here
  before I reverify.

  Full log: http://logs.openstack.org/33478/1/gate/gate-tempest-
  devstack-vm-quantum/32609/logs/screen-n-cpu.txt.gz

  Also, this was for stable/grizzly.  I'm not sure how to specify that
  in LP.

  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/new/nova/nova/openstack/common/rpc/amqp.py", line 430, in 
_process_data
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp rval = 
self.proxy.dispatch(ctxt, version, method, **args)
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/new/nova/nova/openstack/common/rpc/dispatcher.py", line 133, in 
dispatch
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp return 
getattr(proxyobj, method)(ctxt, **kwargs)
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/new/nova/nova/exception.py", line 117, in wrapped
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp 
temp_level, payload)
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/usr/lib/python2.7/contextlib.py", line 24, in __exit__
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp 
self.gen.next()
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/new/nova/nova/exception.py", line 94, in wrapped
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp return 
f(self, context, *args, **kw)
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/new/nova/nova/compute/manager.py", line 209, in decorated_function
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp pass
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/usr/lib/python2.7/contextlib.py", line 24, in __exit__
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp 
self.gen.next()
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/new/nova/nova/compute/manager.py", line 195, in decorated_function
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp return 
function(self, context, *args, **kwargs)
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/new/nova/nova/compute/manager.py", line 237, in decorated_function
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp e, 
sys.exc_info())
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/usr/lib/python2.7/contextlib.py", line 24, in __exit__
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp 
self.gen.next()
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/new/nova/nova/compute/manager.py", line 224, in decorated_function
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp return 
function(self, context, *args, **kwargs)
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/new/nova/nova/compute/manager.py", line 2854, in 
reserve_block_device_name
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp return 
do_reserve()
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/new/nova/nova/openstack/common/lockutils.py", line 242, in inner
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp retval 
= f(*args, **kwargs)
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/new/nova/nova/compute/manager.py", line 2843, in do_reserve
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp 
context, instance, bdms, device)
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp   File 
"/opt/stack/new/nova/nova/compute/utils.py", line 165, in 
get_device_name_for_instance
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp raise 
exception.DevicePathInUse(path=device)
  2013-06-20 19:26:25.981 23879 TRACE nova.openstack.common.rpc.amqp 
DevicePathInUse: The supplied device path (/dev/vdb) is in use.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1193113/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.n

[Yahoo-eng-team] [Bug 1260440] Re: nova-compute host is added to scheduling pool before Neutron can bind network ports on said host

2013-12-12 Thread Clint Byrum
This breaks deployment of new clouds in TripleO sometimes, and will
likely break scaling too. Hence the Critical status.

** Also affects: neutron
   Importance: Undecided
   Status: New

** Also affects: tripleo
   Importance: Undecided
   Status: New

** Changed in: tripleo
   Status: New => Triaged

** Changed in: tripleo
   Importance: Undecided => Critical

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1260440

Title:
  nova-compute host is added to scheduling pool before Neutron can bind
  network ports on said host

Status in OpenStack Neutron (virtual network service):
  New
Status in OpenStack Compute (Nova):
  Confirmed
Status in tripleo - openstack on openstack:
  Triaged

Bug description:
  This is a race condition.

  Given a cloud with 0 compute nodes available, on a compute node:
  * Start up neutron-openvswitch-agent
  * Start up nova-compute
  * nova boot an instance

  Scenario 1:
  * neutron-openvswitch-agent registers with Neutron before nova tries to boot 
instance
  * port is bound to agent
  * instance boots with correct networking

  Scenario 2:
  * nova schedules instance to host before neutron-openvswitch-agent is 
registered with Neutron
  * nova instance fails with vif_type=binding_failed
  * instance is in ERROR state

  I would expect that Nova would not try to schedule instances on
  compute hosts that are not ready.

  Please also see this mailing list thread for more info:

  http://lists.openstack.org/pipermail/openstack-
  dev/2013-December/022084.html

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1260440/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1218391] Re: tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestXML.test_delete_image_that_is_not_yet_active spurious failure

2013-12-12 Thread Sean Dague
there is only 1 hit in the last 2 weeks on this test, I actually think
it's closed

** Changed in: nova
   Status: Confirmed => Fix Released

** No longer affects: nova

** Changed in: tempest
   Status: Confirmed => Fix Released

** No longer affects: tempest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1218391

Title:
  
tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestXML.test_delete_image_that_is_not_yet_active
  spurious failure

Status in OpenStack Object Storage (Swift):
  Confirmed

Bug description:
  Looks like this is an intermittent failure:

  ft45.7: 
tempest.api.compute.images.test_images_oneserver.ImagesOneServerTestXML.test_delete_image_that_is_not_yet_active[gate,negative]_StringException:
 Empty attachments:
stderr
stdout

  Traceback (most recent call last):
File "tempest/api/compute/images/test_images_oneserver.py", line 161, in 
test_delete_image_that_is_not_yet_active
  resp, body = self.client.create_image(self.server['id'], snapshot_name)
File "tempest/services/compute/xml/images_client.py", line 105, in 
create_image
  str(Document(post_body)), self.headers)
File "tempest/common/rest_client.py", line 260, in post
  return self.request('POST', url, headers, body)
File "tempest/common/rest_client.py", line 388, in request
  resp, resp_body)
File "tempest/common/rest_client.py", line 443, in _error_checker
  raise exceptions.Duplicate(resp_body)
  Duplicate: An object with that identifier already exists

  Details: {'message': "Cannot 'createImage' while instance is in
  task_state image_uploading", 'code': '409'}

  http://logs.openstack.org/50/41350/3/gate/gate-nova-
  python26/80bbbf0/testr_results.html.gz

To manage notifications about this bug go to:
https://bugs.launchpad.net/swift/+bug/1218391/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1248757] Re: test_snapshot_pattern fails with paramiko ssh EOFError

2013-12-12 Thread Matt Riedemann
Removing tempest and adding nova, glance and neutron given what this
test impacts:

FAIL:
tempest.scenario.test_snapshot_pattern.TestSnapshotPattern.test_snapshot_pattern[compute,image,network]

** Also affects: nova
   Importance: Undecided
   Status: New

** Also affects: glance
   Importance: Undecided
   Status: New

** Also affects: neutron
   Importance: Undecided
   Status: New

** No longer affects: tempest

** Tags added: gate-failure

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1248757

Title:
  test_snapshot_pattern fails with paramiko ssh EOFError

Status in OpenStack Image Registry and Delivery Service (Glance):
  New
Status in OpenStack Neutron (virtual network service):
  New
Status in OpenStack Compute (Nova):
  New

Bug description:
  I haven't seen this one reported yet (or seen it yet):

  http://logs.openstack.org/55/55455/1/check/check-tempest-devstack-vm-
  neutron/28d1ed7/console.html

  http://paste.openstack.org/show/50561/

To manage notifications about this bug go to:
https://bugs.launchpad.net/glance/+bug/1248757/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1255627] Re: images.test_list_image_filters.ListImageFiltersTest fails with timeout

2013-12-12 Thread David Kranz
This non-white-listed error showed up in n-cpu:

2013-11-27 00:53:57.756 ERROR nova.virt.libvirt.driver [req-
298cf8f1-3907-4494-8b6e-61e9b88dfded ListImageFiltersTestXML-
tempest-656023876-user ListImageFiltersTestXML-tempest-656023876-tenant]
An error occurred while enabling hairpin mode on domain with xml:


According to logstash this happened 9 times in the last two weeks.

** Changed in: nova
   Status: New => Confirmed

** Changed in: tempest
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1255627

Title:
  images.test_list_image_filters.ListImageFiltersTest fails with timeout

Status in OpenStack Compute (Nova):
  Confirmed
Status in Tempest:
  Invalid

Bug description:
  Spurious failure in this test:

  http://logs.openstack.org/49/55749/8/check/check-tempest-devstack-vm-
  full/9bc94d5/console.html

  2013-11-27 01:10:35.802 | 
==
  2013-11-27 01:10:35.802 | FAIL: setUpClass 
(tempest.api.compute.images.test_list_image_filters.ListImageFiltersTestXML)
  2013-11-27 01:10:35.803 | setUpClass 
(tempest.api.compute.images.test_list_image_filters.ListImageFiltersTestXML)
  2013-11-27 01:10:35.803 | 
--
  2013-11-27 01:10:35.803 | _StringException: Traceback (most recent call last):
  2013-11-27 01:10:35.804 |   File 
"tempest/api/compute/images/test_list_image_filters.py", line 50, in setUpClass
  2013-11-27 01:10:35.807 | cls.client.wait_for_image_status(cls.image1_id, 
'ACTIVE')
  2013-11-27 01:10:35.809 |   File 
"tempest/services/compute/xml/images_client.py", line 153, in 
wait_for_image_status
  2013-11-27 01:10:35.809 | raise exceptions.TimeoutException
  2013-11-27 01:10:35.809 | TimeoutException: Request timed out

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1255627/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1257032] Re: nova makes calls to neutron with out considering URI size limit

2013-12-12 Thread Venkata Siva Vijayendra Bhamidipati
*** This bug is a duplicate of bug 1228384 ***
https://bugs.launchpad.net/bugs/1228384

Will close this bug as a dup of
https://bugs.launchpad.net/nova/+bug/1228384 . As part of that fix, Phil
has implemented chunking of server ids when querying neutron for ports
of VMs. The above stack trace is for a quantum deployment, and Phil's
patch will be ported over to that deployment.

** This bug has been marked a duplicate of bug 1228384
   Security Group extension reads all Neutron ports for anything other that a 
single server

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1257032

Title:
  nova makes calls to neutron with out considering URI size limit

Status in OpenStack Compute (Nova):
  New

Bug description:
  Nova when requesting security group instance bindings for servers
  makes a call to Neutron. When there are many servers, the URI size
  grows beyond 8k with neutron throwing up 414 error message.

  
  We could easily hit this bug in our deployment as we have several VMs running.

  There is a similar bug while making net-list. It needed to make subnet
  list internally and that would result in 414 if there are several
  subnets that can make URI size too long. But subnet-list was internal
  call and fix was appropriate in the neutron client.

  Here the bug is while fetching sg instance bindings which is primary
  call. Hence I feel the bug must be fixed in neutron API consumer which
  is nova.

  Also, there must be a general framework for all APIs to not to exceed
  URI size limit or fix all calls  with URIs which can extend beyond 8k
  size limit.

  Stacktrace for reference
  2013-11-27 13:06:01.696 ERROR nova.api.openstack 
[req-020f17cb-ee43-4cd2-a270-767936e6546b 6abe780581924062bdb648375abcb378 
b9bbb06d41a942248e8d7070e17ed89d] Caught error: 414-{'message': ''}
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack Traceback (most recent 
call last):
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack   File 
"/usr/lib/python2.6/site-packages/nova/api/openstack/__init__.py", line 81, in 
__call__
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack return 
req.get_response(self.application)
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack   File 
"/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/request.py", line 
1053, in get_response
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack application, 
catch_exc_info=False)
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack   File 
"/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/request.py", line 
1022, in call_application
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack app_iter = 
application(self.environ, start_response)
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack   File 
"/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/dec.py", line 
159, in __call__
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack return 
resp(environ, start_response)
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack   File 
"/usr/lib/python2.6/site-packages/keystoneclient/middleware/auth_token.py", 
line 450, in __call__
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack return 
self.app(env, start_response)
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack   File 
"/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/dec.py", line 
159, in __call__
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack return 
resp(environ, start_response)
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack   File 
"/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/dec.py", line 
159, in __call__
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack return 
resp(environ, start_response)
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack   File 
"/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/dec.py", line 
159, in __call__
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack return 
resp(environ, start_response)
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack   File 
"/usr/lib/python2.6/site-packages/Routes-1.12.3-py2.6.egg/routes/middleware.py",
 line 131, in __call__
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack response = 
self.app(environ, start_response)
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack   File 
"/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/dec.py", line 
159, in __call__
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack return 
resp(environ, start_response)
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack   File 
"/usr/lib/python2.6/site-packages/WebOb-1.0.8-py2.6.egg/webob/dec.py", line 
147, in __call__
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack resp = 
self.call_func(req, *args, **self.kwargs)
  2013-11-27 13:06:01.696 30107 TRACE nova.api.openstack

[Yahoo-eng-team] [Bug 1254772] Re: tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML setUpClass times-out on attaching volume

2013-12-12 Thread David Kranz
This shows up in n-cpu:

The "model server went away" showed up 11 times in the last two weeks
with the last one being on Dec. 3. This sample size is too small for me
to close at this time.

2013-11-25 15:24:22.099 21076 ERROR nova.servicegroup.drivers.db [-] model 
server went away
2013-11-25 15:24:32.814 ERROR nova.compute.manager 
[req-ecacaa21-3f07-4b44-9896-8b5bd2238a19 
ServersTestManualDisk-tempest-1962756300-user 
ServersTestManualDisk-tempest-1962756300-tenant] [instance: 
1f872097-8ad8-44f8-ba03-89a14115efe0] Failed to deallocate network for instance.
2013-11-25 15:25:32.855 21076 ERROR root [-] Original exception being dropped: 
['Traceback (most recent call last):\n', '  File 
"/opt/stack/new/nova/nova/compute/manager.py", line 1809, in 
_try_deallocate_network\nself._deallocate_network(context, instance, 
requested_networks)\n', '  File "/opt/stack/new/nova/nova/compute/manager.py", 
line 1491, in _deallocate_network\ncontext, instance, 
requested_networks=requested_networks)\n', '  File 
"/opt/stack/new/nova/nova/network/api.py", line 93, in wrapped\nreturn 
func(self, context, *args, **kwargs)\n', '  File 
"/opt/stack/new/nova/nova/network/api.py", line 318, in 
deallocate_for_instance\n
self.network_rpcapi.deallocate_for_instance(context, **args)\n', '  File 
"/opt/stack/new/nova/nova/network/rpcapi.py", line 199, in 
deallocate_for_instance\nhost=host, 
requested_networks=requested_networks)\n', '  File 
"/opt/stack/new/nova/nova/rpcclient.py", line 85, in call\nreturn 
self._invoke(self.proxy.call, ctxt, method, **
 kwargs)\n', '  File "/opt/stack/new/nova/nova/rpcclient.py", line 63, in 
_invoke\nreturn cast_or_call(ctxt, msg, **self.kwargs)\n', '  File 
"/opt/stack/new/nova/nova/openstack/common/rpc/proxy.py", line 130, in call\n   
 exc.info, real_topic, msg.get(\'method\'))\n', 'Timeout: Timeout while waiting 
on RPC response - topic: "network", RPC method: "deallocate_for_instance" info: 
""\n']
2013-11-25 15:25:38.371 21076 ERROR nova.openstack.common.periodic_task [-] 
Error during ComputeManager.update_available_resource: Timeout while waiting on 
RPC response - topic: "conductor", RPC method: "compute_node_update" info: 
""
2013-11-25 15:26:32.903 21076 ERROR root [-] Original exception being dropped: 
['Traceback (most recent call last):\n', '  File 
"/opt/stack/new/nova/nova/compute/manager.py", line 1919, in _delete_instance\n 
   self._shutdown_instance(context, db_inst, bdms)\n', '  File 
"/opt/stack/new/nova/nova/compute/manager.py", line 1854, in 
_shutdown_instance\nself._try_deallocate_network(context, instance, 
requested_networks)\n', '  File "/opt/stack/new/nova/nova/compute/manager.py", 
line 1814, in _try_deallocate_network\n
self._set_instance_error_state(context, instance[\'uuid\'])\n', '  File 
"/opt/stack/new/nova/nova/compute/manager.py", line 484, in 
_set_instance_error_state\nvm_state=vm_states.ERROR)\n', '  File 
"/opt/stack/new/nova/nova/compute/manager.py", line 473, in _instance_update\n  
  **kwargs)\n', '  File "/opt/stack/new/nova/nova/conductor/api.py", line 389, 
in instance_update\nupdates, \'conductor\')\n', '  File 
"/opt/stack/new/nova/nova/conductor/rpcapi.py", line
  149, in instance_update\nservice=service)\n', '  File 
"/opt/stack/new/nova/nova/rpcclient.py", line 85, in call\nreturn 
self._invoke(self.proxy.call, ctxt, method, **kwargs)\n', '  File 
"/opt/stack/new/nova/nova/rpcclient.py", line 63, in _invoke\nreturn 
cast_or_call(ctxt, msg, **self.kwargs)\n', '  File 
"/opt/stack/new/nova/nova/openstack/common/rpc/proxy.py", line 130, in call\n   
 exc.info, real_topic, msg.get(\'method\'))\n', 'Timeout: Timeout while waiting 
on RPC response - topic: "conductor", RPC method: "instance_update" info: 
""\n']
2013-11-25 15:26:32.933 21076 ERROR nova.servicegroup.drivers.db [-] Recovered 
model server connection!


** Also affects: nova
   Importance: Undecided
   Status: New

** Changed in: tempest
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1254772

Title:
  tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML
  setUpClass times-out on attaching volume

Status in OpenStack Compute (Nova):
  New
Status in Tempest:
  Invalid

Bug description:
  2013-11-25 15:42:45.769 | 
==
  2013-11-25 15:42:45.770 | FAIL: setUpClass 
(tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML)
  2013-11-25 15:42:45.770 | setUpClass 
(tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML)
  2013-11-25 15:42:45.770 | 
--
  2013-11-25 15:42:45.770 | _StringException: Traceback (most recent call last):
  2013-11-25 15:42:45.770 |   File 
"tempest/api/compute/servers/test_server

[Yahoo-eng-team] [Bug 1260489] [NEW] --debug flag not working in neutron

2013-12-12 Thread Venkata Siva Vijayendra Bhamidipati
Public bug reported:

This is with the neutron master branch, in a single node devstack setup.
The branch is at commit 3b4233873539bad62d202025529678a5b0add412.

If I use the --debug flag in a neutron CLI, for example, port-list, I
don't see any debug output:

cloud@controllernode:/opt/stack/neutron$ neutron --debug port-list
+--+--+---+-+
| id   | name | mac_address   | fixed_ips   
|
+--+--+---+-+
| 6c26cdc1-acc1-439c-bb47-d343085b7b78 |  | fa:16:3e:32:2c:eb | 
{"subnet_id": "37f15352-e816-4a03-b58c-b4d5c1fa8e2a", "ip_address": "10.0.0.2"} 
|
| f09b14b2-3162-4212-9d91-f97b22c95f31 |  | fa:16:3e:99:08:6b | 
{"subnet_id": "d4717b67-fd64-45ed-b22c-dedbd23afff3", "ip_address": 
"172.24.4.226"} |
| f0ba4efd-12ca-4d56-8c7d-e879e4150a63 |  | fa:16:3e:02:41:47 | 
{"subnet_id": "37f15352-e816-4a03-b58c-b4d5c1fa8e2a", "ip_address": "10.0.0.1"} 
|
+--+--+---+-+
cloud@controllernode:/opt/stack/neutron$ 


On the other hand, if I use the --debug flag for nova, for example, nova list, 
I see the curl request and response showing up:


cloud@controllernode:/opt/stack/neutron$ nova --debug list

REQ: curl -i 'http://192.168.52.85:5000/v2.0/tokens' -X POST -H
"Content-Type: application/json" -H "Accept: application/json" -H "User-
Agent: python-novaclient" -d '{"auth": {"tenantName": "admin",
"passwordCredentials": {"username": "admin", "password": "password"}}}'

RESP: [200] CaseInsensitiveDict({'date': 'Thu, 05 Dec 2013 23:41:07 GMT', 
'vary': 'X-Auth-Token', 'content-length': '8255', 'content-type': 
'application/json'})
RESP BODY: {"access": {"token": {"issued_at": "2013-12-05T23:41:07.307915", 
"expires": "2013-12-06T23:41:07Z", "id": 
"MIIOkwYJKoZIhvcNAQcCoIIOhDCCDoACAQExCTAHBgUrDgMCGjCCDOkGCSqGSIb3DQEHAaCCDNoEggzWeyJhY2Nlc3MiOiB7InRva2VuIjogeyJpc3N1ZWRfYXQiOiAiMjAxMy0xMi0wNVQyMzo0MTowNy4zMDc5MTUiLCAiZXhwaXJlcyI6ICIyMDEzLTEyLTA2VDIzOjQxOjA3WiIsICJpZCI6ICJwbGFjZWhvbGRlciIsICJ0ZW5hbnQiOiB7ImRlc2NyaXB0aW9uIjogbnVsbCwgImVuYWJsZWQiOiB0cnVlLCAiaWQiOiAiYTdiMzk2MGI5NzkyNGJhYjlhNTVhOWY5ZjY4NGE4NzAiLCAibmFtZSI6ICJhZG1pbiJ9fSwgInNlcnZpY2VDYXRhbG9nIjogW3siZW5kcG9pbnRzIjogW3siYWRtaW5VUkwiOiAiaHR0cDovLzE5Mi4xNjguNTIuODU6ODc3NC92Mi9hN2IzOTYwYjk3OTI0YmFiOWE1NWE5ZjlmNjg0YTg3MCIsICJyZWdpb24iOiAiUmVnaW9uT25lIiwgImludGVybmFsVVJMIjogImh0dHA6Ly8xOTIuMTY4LjUyLjg1Ojg3NzQvdjIvYTdiMzk2MGI5NzkyNGJhYjlhNTVhOWY5ZjY4NGE4NzAiLCAiaWQiOiAiMDQyMzVjMmE1ODNlNDAwZDg1NTBkYTI0NmNiZDI1YWEiLCAicHVibGljVVJMIjogImh0dHA6Ly8xOTIuMTY4LjUyLjg1Ojg3NzQvdjIvYTdiMzk2MGI5NzkyNGJhYjlhNTVhOWY5ZjY4NGE4NzAifV0sICJlbmRwb2ludHNfbGlua3MiOiBbXSwgInR5cGUiOiAi
 
Y29tcHV0ZSIsICJuYW1lIjogIm5vdmEifSwgeyJlbmRwb2ludHMiOiBbeyJhZG1pblVSTCI6ICJodHRwOi8vMTkyLjE2OC41Mi44NTo5Njk2LyIsICJyZWdpb24iOiAiUmVnaW9uT25lIiwgImludGVybmFsVVJMIjogImh0dHA6Ly8xOTIuMTY4LjUyLjg1Ojk2OTYvIiwgImlkIjogIjYyNWI1YzM3ZDJlYzQ4ZGRhMTRmZGZmZmMyZjBhMTY0IiwgInB1YmxpY1VSTCI6ICJodHRwOi8vMTkyLjE2OC41Mi44NTo5Njk2LyJ9XSwgImVuZHBvaW50c19saW5rcyI6IFtdLCAidHlwZSI6ICJuZXR3b3JrIiwgIm5hbWUiOiAibmV1dHJvbiJ9LCB7ImVuZHBvaW50cyI6IFt7ImFkbWluVVJMIjogImh0dHA6Ly8xOTIuMTY4LjUyLjg1Ojg3NzYvdjIvYTdiMzk2MGI5NzkyNGJhYjlhNTVhOWY5ZjY4NGE4NzAiLCAicmVnaW9uIjogIlJlZ2lvbk9uZSIsICJpbnRlcm5hbFVSTCI6ICJodHRwOi8vMTkyLjE2OC41Mi44NTo4Nzc2L3YyL2E3YjM5NjBiOTc5MjRiYWI5YTU1YTlmOWY2ODRhODcwIiwgImlkIjogIjNmODVjN2ZmZjNjMzRmNWNiMzlmMTZiMzQ2ZmY1Mjc0IiwgInB1YmxpY1VSTCI6ICJodHRwOi8vMTkyLjE2OC41Mi44NTo4Nzc2L3YyL2E3YjM5NjBiOTc5MjRiYWI5YTU1YTlmOWY2ODRhODcwIn1dLCAiZW5kcG9pbnRzX2xpbmtzIjogW10sICJ0eXBlIjogInZvbHVtZXYyIiwgIm5hbWUiOiAiY2luZGVyIn0sIHsiZW5kcG9pbnRzIjogW3siYWRtaW5VUkwiOiAiaHR0cDovLzE5Mi4xNjguNTIuODU6ODc3NC92MyIsICJyZWdpb
 
24iOiAiUmVnaW9uT25lIiwgImludGVybmFsVVJMIjogImh0dHA6Ly8xOTIuMTY4LjUyLjg1Ojg3NzQvdjMiLCAiaWQiOiAiYTM4NjBlZTM3MWEyNDIxNGFlYTBiODk5M2I1YTY0OTciLCAicHVibGljVVJMIjogImh0dHA6Ly8xOTIuMTY4LjUyLjg1Ojg3NzQvdjMifV0sICJlbmRwb2ludHNfbGlua3MiOiBbXSwgInR5cGUiOiAiY29tcHV0ZXYzIiwgIm5hbWUiOiAibm92YSJ9LCB7ImVuZHBvaW50cyI6IFt7ImFkbWluVVJMIjogImh0dHA6Ly8xOTIuMTY4LjUyLjg1OjMzMzMiLCAicmVnaW9uIjogIlJlZ2lvbk9uZSIsICJpbnRlcm5hbFVSTCI6ICJodHRwOi8vMTkyLjE2OC41Mi44NTozMzMzIiwgImlkIjogIjZmZTY2OTMwNjA5MTQwYWVhMTIwMTJjNWViMzViZGQ2IiwgInB1YmxpY1VSTCI6ICJodHRwOi8vMTkyLjE2OC41Mi44NTozMzMzIn1dLCAiZW5kcG9pbnRzX2xpbmtzIjogW10sICJ0eXBlIjogInMzIiwgIm5hbWUiOiAiczMifSwgeyJlbmRwb2ludHMiOiBbeyJhZG1pblVSTCI6ICJodHRwOi8vMTkyLjE2OC41Mi44NTo5MjkyIiwgInJlZ2lvbiI6ICJSZWdpb25PbmUiLCAiaW50ZXJuYWxVUkwiOiAiaHR0cDovLzE5Mi4xNjguNTIuODU6OTI5MiIsICJpZCI6ICIyMjVhMDc2ZmZiOWI0YmQxYTdmODE4N2M0NzY2M2I0NyIsICJwdWJsaWNVUkwiOiAiaHR0cDovLzE5Mi4xNjguNTIuODU6OTI5MiJ9XS

[Yahoo-eng-team] [Bug 1252947] Re: tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON fails sporadically

2013-12-12 Thread Sean Dague
** Also affects: nova
   Importance: Undecided
   Status: New

** No longer affects: tempest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1252947

Title:
  tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON
  fails sporadically

Status in OpenStack Compute (Nova):
  New

Bug description:
  tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON
  fails sporadically.

  See:  http://logs.openstack.org/66/54966/2/check/check-tempest-
  devstack-vm-full/d611ed0/console.html

  2013-11-19 22:24:52.379 | 
==
  2013-11-19 22:24:52.380 | FAIL: setUpClass 
(tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON)
  2013-11-19 22:24:52.380 | setUpClass 
(tempest.api.compute.servers.test_servers_negative.ServersNegativeTestJSON)
  2013-11-19 22:24:52.380 | 
--
  2013-11-19 22:24:52.380 | _StringException: Traceback (most recent call last):
  2013-11-19 22:24:52.380 |   File 
"tempest/api/compute/servers/test_servers_negative.py", line 46, in setUpClass
  2013-11-19 22:24:52.380 | resp, server = 
cls.create_test_server(wait_until='ACTIVE')
  2013-11-19 22:24:52.381 |   File "tempest/api/compute/base.py", line 118, in 
create_test_server
  2013-11-19 22:24:52.381 | server['id'], kwargs['wait_until'])
  2013-11-19 22:24:52.381 |   File 
"tempest/services/compute/json/servers_client.py", line 160, in 
wait_for_server_status
  2013-11-19 22:24:52.381 | extra_timeout=extra_timeout)
  2013-11-19 22:24:52.381 |   File "tempest/common/waiters.py", line 73, in 
wait_for_server_status
  2013-11-19 22:24:52.381 | raise 
exceptions.BuildErrorException(server_id=server_id)
  2013-11-19 22:24:52.381 | BuildErrorException: Server 
62bfeebd-8878-477f-9eac-a8b21ec5ac26 failed to build and is in ERROR status

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1252947/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1224518] Re: test_reboot_server_hard fails sporadically in swift check jobs

2013-12-12 Thread Sean Dague
I don't think this is a tempest bug, this is a state transition bug in
Nova

** Changed in: tempest
   Importance: Undecided => Low

** Changed in: tempest
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1224518

Title:
  test_reboot_server_hard fails sporadically in swift check jobs

Status in OpenStack Compute (Nova):
  New
Status in Tempest:
  Invalid

Bug description:
  See: http://logs.openstack.org/46/46146/2/check/gate-tempest-devstack-
  vm-postgres-full/b2712f1/console.html

  2013-09-12 04:43:17.625 | 
==
  2013-09-12 04:43:17.649 | FAIL: 
tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_reboot_server_hard[gate,smoke]
  2013-09-12 04:43:17.651 | 
tempest.api.compute.servers.test_server_actions.ServerActionsTestJSON.test_reboot_server_hard[gate,smoke]
  2013-09-12 04:43:17.652 | 
--
  2013-09-12 04:43:17.652 | _StringException: Empty attachments:
  2013-09-12 04:43:17.652 |   stderr
  2013-09-12 04:43:17.652 |   stdout
  2013-09-12 04:43:17.653 | 
  2013-09-12 04:43:17.653 | pythonlogging:'': {{{
  2013-09-12 04:43:17.653 | 2013-09-12 04:16:55,739 Request: GET 
http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf
  2013-09-12 04:43:17.654 | 2013-09-12 04:16:55,806 Response Status: 200
  2013-09-12 04:43:17.654 | 2013-09-12 04:16:55,806 Nova request id: 
req-cdc6b1fc-bcf2-4e9c-bea1-8bf935993cbd
  2013-09-12 04:43:17.654 | 2013-09-12 04:16:55,807 Request: POST 
http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf/action
  2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,917 Response Status: 202
  2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,917 Nova request id: 
req-3af37dd3-0ddc-4daa-aa6f-6958a5073cc4
  2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,918 Request: GET 
http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf
  2013-09-12 04:43:17.655 | 2013-09-12 04:16:55,986 Response Status: 200
  2013-09-12 04:43:17.656 | 2013-09-12 04:16:55,986 Nova request id: 
req-a7298d3e-167c-4c8f-9506-6064ba811e5b

  .
  .
  .

  2013-09-12 04:43:17.976 | 2013-09-12 04:23:35,773 Request: GET 
http://127.0.0.1:8774/v2/83ed6f49279b4292a00b32397d2f52fb/servers/8ad0ad9a-3975-486f-94b4-af1c89b51aaf
  2013-09-12 04:43:17.976 | 2013-09-12 04:23:35,822 Response Status: 200
  2013-09-12 04:43:17.976 | 2013-09-12 04:23:35,823 Nova request id: 
req-a122aded-b49b-4847-9920-b2b8b09bc0ca
  2013-09-12 04:43:17.976 | }}}
  2013-09-12 04:43:17.977 | 
  2013-09-12 04:43:17.977 | Traceback (most recent call last):
  2013-09-12 04:43:17.978 |   File 
"tempest/api/compute/servers/test_server_actions.py", line 81, in 
test_reboot_server_hard
  2013-09-12 04:43:17.978 | 
self.client.wait_for_server_status(self.server_id, 'ACTIVE')
  2013-09-12 04:43:17.979 |   File 
"tempest/services/compute/json/servers_client.py", line 176, in 
wait_for_server_status
  2013-09-12 04:43:17.979 | raise exceptions.TimeoutException(message)
  2013-09-12 04:43:17.979 | TimeoutException: Request timed out
  2013-09-12 04:43:17.980 | Details: Server 
8ad0ad9a-3975-486f-94b4-af1c89b51aaf failed to reach ACTIVE status within the 
required time (400 s). Current status: HARD_REBOOT.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1224518/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1244762] Re: tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance fails sporadically

2013-12-12 Thread Sean Dague
** Also affects: nova
   Importance: Undecided
   Status: New

** No longer affects: tempest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1244762

Title:
  
tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance
  fails sporadically

Status in OpenStack Compute (Nova):
  New

Bug description:
  See: http://logs.openstack.org/87/44787/16/check/check-tempest-
  devstack-vm-neutron/d2ede4d/console.html

  2013-10-25 18:06:37.957 | 
==
  2013-10-25 18:06:37.959 | FAIL: 
tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance[gate,smoke]
  2013-10-25 18:06:37.959 | 
tempest.thirdparty.boto.test_ec2_instance_run.InstanceRunTest.test_run_stop_terminate_instance[gate,smoke]
  2013-10-25 18:06:37.959 | 
--
  2013-10-25 18:06:37.959 | _StringException: Empty attachments:
  2013-10-25 18:06:37.959 |   stderr
  2013-10-25 18:06:37.960 |   stdout
  2013-10-25 18:06:37.960 | 
  2013-10-25 18:06:37.960 | pythonlogging:'': {{{
  2013-10-25 18:06:37.960 | 2013-10-25 17:59:08,821 state: pending
  2013-10-25 18:06:37.960 | 2013-10-25 17:59:14,092 State transition "pending" 
==> "error" 5 second
  2013-10-25 18:06:37.961 | }}}
  2013-10-25 18:06:37.961 | 
  2013-10-25 18:06:37.961 | Traceback (most recent call last):
  2013-10-25 18:06:37.961 |   File 
"tempest/thirdparty/boto/test_ec2_instance_run.py", line 150, in 
test_run_stop_terminate_instance
  2013-10-25 18:06:37.961 | self.assertInstanceStateWait(instance, 
"running")
  2013-10-25 18:06:37.961 |   File "tempest/thirdparty/boto/test.py", line 356, 
in assertInstanceStateWait
  2013-10-25 18:06:37.962 | state = self.waitInstanceState(lfunction, 
wait_for)
  2013-10-25 18:06:37.962 |   File "tempest/thirdparty/boto/test.py", line 341, 
in waitInstanceState
  2013-10-25 18:06:37.962 | self.valid_instance_state)
  2013-10-25 18:06:37.962 |   File "tempest/thirdparty/boto/test.py", line 332, 
in state_wait_gone
  2013-10-25 18:06:37.962 | self.assertIn(state, valid_set | self.gone_set)
  2013-10-25 18:06:37.963 |   File 
"/usr/local/lib/python2.7/dist-packages/testtools/testcase.py", line 328, in 
assertIn
  2013-10-25 18:06:37.963 | self.assertThat(haystack, Contains(needle))
  2013-10-25 18:06:37.963 |   File 
"/usr/local/lib/python2.7/dist-packages/testtools/testcase.py", line 417, in 
assertThat
  2013-10-25 18:06:37.963 | raise MismatchError(matchee, matcher, mismatch, 
verbose)
  2013-10-25 18:06:37.963 | MismatchError: u'error' not in set(['paused', 
'terminated', 'running', 'stopped', 'pending', '_GONE', 'stopping', 
'shutting-down'])

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1244762/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1258319] Re: test_reboot_server_hard fails sporadically in swift check jobs

2013-12-12 Thread Sean Dague
*** This bug is a duplicate of bug 1224518 ***
https://bugs.launchpad.net/bugs/1224518

** This bug has been marked a duplicate of bug 1224518
   test_reboot_server_hard fails sporadically in swift check jobs

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1258319

Title:
  test_reboot_server_hard fails sporadically in swift check jobs

Status in OpenStack Compute (Nova):
  New

Bug description:
  test_reboot_server_hard fails sporadically in swift check jobs

  I believe this has been reported before, but I was not able to find
  it.

  See: http://logs.openstack.org/43/60343/1/gate/gate-tempest-dsvm-
  full/c92d206/console.html

  2013-12-05 21:29:18.174 | 
==
  2013-12-05 21:29:18.183 | FAIL: 
tempest.api.compute.servers.test_server_actions.ServerActionsTestXML.test_reboot_server_hard[gate,smoke]
  2013-12-05 21:29:18.186 | 
tempest.api.compute.servers.test_server_actions.ServerActionsTestXML.test_reboot_server_hard[gate,smoke]
  2013-12-05 21:29:18.200 | 
--
  2013-12-05 21:29:18.206 | _StringException: Empty attachments:
  2013-12-05 21:29:18.206 |   stderr
  2013-12-05 21:29:18.207 |   stdout
  2013-12-05 21:29:18.207 | 
  2013-12-05 21:29:18.207 | pythonlogging:'': {{{

  .
  .
  .

  2013-12-05 21:29:19.174 | Traceback (most recent call last):
  2013-12-05 21:29:19.175 |   File 
"tempest/api/compute/servers/test_server_actions.py", line 83, in 
test_reboot_server_hard
  2013-12-05 21:29:19.175 | 
self.client.wait_for_server_status(self.server_id, 'ACTIVE')
  2013-12-05 21:29:19.175 |   File 
"tempest/services/compute/xml/servers_client.py", line 369, in 
wait_for_server_status
  2013-12-05 21:29:19.175 | extra_timeout=extra_timeout)
  2013-12-05 21:29:19.176 |   File "tempest/common/waiters.py", line 82, in 
wait_for_server_status
  2013-12-05 21:29:19.176 | raise exceptions.TimeoutException(message)
  2013-12-05 21:29:19.176 | TimeoutException: Request timed out
  2013-12-05 21:29:19.177 | Details: Server 
f313af9a-8ec1-4f77-b63f-76d9317d6423 failed to reach ACTIVE status within the 
required time (196 s). Current status: HARD_REBOOT.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1258319/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1250836] Re: Updating of instance metadata occasionally leads to a deadlock

2013-12-12 Thread Sean Dague
not a tempest bug

** No longer affects: tempest

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1250836

Title:
  Updating of instance metadata occasionally leads to a deadlock

Status in OpenStack Compute (Nova):
  New

Bug description:
  During the tempest tests run I got the following error:

  2013-11-13 10:09:18.814 ERROR nova.api.openstack 
[req-a3172f97-0d7c-4f8b-a7a5-bec6aad2b549 
ServerMetadataTestJSON-tempest-1285971638-user 
ServerMetadataTestJSON-tempest-1285971638-tenant] Caught error: 
(OperationalError) (1213, 'Deadlock found when trying to get lock; try 
restarting transaction') 'INSERT INTO instance_metadata (created_at, 
updated_at, deleted_at, deleted, `key`, value, instance_uuid) VALUES (%s, %s, 
%s, %s, %s, %s, %s)' (datetime.datetime(2013, 11, 13, 10, 9, 18, 811419), None, 
None, 0, 'key3', 'value3', 'ba645a19-78c5-439d-9408-68f413c200f4')
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack Traceback (most recent 
call last):
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/opt/stack/new/nova/nova/api/openstack/__init__.py", line 119, in __call__
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return 
req.get_response(self.application)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/usr/local/lib/python2.7/dist-packages/webob/request.py", line 1296, in send
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack application, 
catch_exc_info=False)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/usr/local/lib/python2.7/dist-packages/webob/request.py", line 1260, in 
call_application
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack app_iter = 
application(self.environ, start_response)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/usr/local/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return 
resp(environ, start_response)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/opt/stack/new/python-keystoneclient/keystoneclient/middleware/auth_token.py", 
line 571, in __call__
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return 
self.app(env, start_response)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/usr/local/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return 
resp(environ, start_response)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/usr/local/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return 
resp(environ, start_response)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/usr/lib/python2.7/dist-packages/routes/middleware.py", line 131, in __call__
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack response = 
self.app(environ, start_response)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/usr/local/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return 
resp(environ, start_response)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/usr/local/lib/python2.7/dist-packages/webob/dec.py", line 130, in __call__
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack resp = 
self.call_func(req, *args, **self.kwargs)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/usr/local/lib/python2.7/dist-packages/webob/dec.py", line 195, in call_func
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return 
self.func(req, *args, **kwargs)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/opt/stack/new/nova/nova/api/openstack/wsgi.py", line 939, in __call__
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack content_type, 
body, accept)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/opt/stack/new/nova/nova/api/openstack/wsgi.py", line 998, in _process_stack
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack action_result = 
self.dispatch(meth, request, action_args)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/opt/stack/new/nova/nova/api/openstack/wsgi.py", line 1079, in dispatch
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack return 
method(req=request, **action_args)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/opt/stack/new/nova/nova/api/openstack/compute/server_metadata.py", line 67, 
in create
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack delete=False)
  2013-11-13 10:09:18.814 22804 TRACE nova.api.openstack   File 
"/opt/stack/new/nova/nova/api/openstack/compute/server_metadata.py", line 120, 
in _update_instance_metadata
  2013-11-13 10:09:18.814 22804 TRA

[Yahoo-eng-team] [Bug 1260516] [NEW] PortInUse exception leaves orphaned ports

2013-12-12 Thread Robert Pothier
Public bug reported:

While bringing up 1000 VMs with Heat, some VMs are in error state.
After deleting the VMs, not all ports are removed.
The next time VMs are created, they fail due to no IP addresses left.


Note, initially there are 7 ports, no VMs
804 VMs are created, 811 ports

root@control01:/usr/share/pyshared/heat# nova list | grep ACT | wc -l
358
root@control01:/usr/share/pyshared/heat# neutron port-list | wc -l
811
root@control01:/usr/share/pyshared/heat# nova list | grep ERR | wc -l
270
root@control01:/usr/share/pyshared/heat# nova list | grep stack | wc -l
804

After deleting the VMs, 248 ports remain


root@control01:/usr/share/pyshared/heat# nova list | grep stack | wc -l
0
root@control01:/usr/share/pyshared/heat# neutron port-list | wc -l
248



2013-12-12 20:55:35.320 20945 ERROR nova.scheduler.filter_scheduler [req-844a11 
   f6-66f1-4fc7-9f3b-8bacbe57a04d e95097acd0b041558f33d07f720c1bd7 
354f17bf81924b278806c3e3798aa527] [instance: 
2a3eadcc-5230-4ed6-bf9f-9a82d43b91c3] Error from last host: compute134 
(node compute134.lab.cisco): [u'Traceback (most recent call last):\n', u'  
File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 1037, 
in _build_instance\nset_access_ip=set_access_ip)\n', u'  File 
"/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 1410, in _sp   
 awn\nLOG.exception(_(\'Instance failed to spawn\'), instance=instance)\n', 
u'  File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 
1407, in _spawn\nblock_device_info)\n', u'  File 
"/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py", line 2063, 
in spawn\nadmin_pass=admin_password)\n', u'  File 
"/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py"
 , line 2412, in _create_image\ncontent=files, extra_md=extra_md, network_  
  info=network_info)\n', u'  File 
"/usr/lib/python2.7/dist-packages/nova/api/metadata/base.py", line 157, in 
__init__\ncfg = netutils.get_injected_network_t
emplate(network_info)\n', u'  File "/usr/lib/python2.7/dist-packages/nova/virt/ 
   netutils.py", line 74, in get_injected_network_template\nif not 
(network_info and template):\n', u'  File 
"/usr/lib/python2.7/dist-packages/nova/network/model.py", line 379, in 
__len__\nreturn self._sync_wrapper(fn, *args, **kwargs)\n', u'  File 
"/usr/lib/python2.7/dist-packages/nova/network/model.py", line 366, in 
_sync_wrapper\nself.wait()\n', u'  File "/usr/lib/python2.7/dist-p
ackages/nova/network/model.py", line 398, in wait\nself[:] = self._gt.wait( 
   )\n', u'  File "/usr/lib/python2.7/dist-packages/eventlet/greenthread.py", 
line 168, in wait\nreturn self._exit_event.wait()\n', u'  File 
"/usr/lib/python2.7
 /dist-packages/eventlet/event.py", line 120, in wait\ncurrent.throw(*sel   
 f._exc)\n', u'  File 
"/usr/lib/python2.7/dist-packages/eventlet/greenthread.py", line 194, in 
main\nresult = function(*args, **kwargs)\n', u'  File "/usr/
lib/python2.7/dist-packages/nova/compute/manager.py", line 1228, in _allocate_n 
   etwork_async\ndhcp_options=dhcp_options)\n', u'  File 
"/usr/lib/python2.7/dist-packages/nova/network/api.py", line 49, in 
wrapper\nres = f(self, context, *args, **kwargs)\n', u'  File 
"/usr/lib/python2.7/dist-packages/nova/network/neutronv2/api.py", line 243, 
in allocate_for_instance\nraise exception.Po
rtInUse(port_id=port_id)\n', u'PortInUse: Port 69e55016-b794-4dd9-b3f3-4e78336f 
   bd11 is still in use.\n']
 93 2013-12-12 20:55:35.321 20945 WARNING nova.scheduler.utils 
[req-844a11f6-66f1-4fc7-9f3b-8bacbe57a04d e95097acd0b041558f33d07f720c1bd7 
354f17bf81924b278806c3e3798aa527] Failed to scheduler_run_instance: No 
valid host was found. Exceeded max scheduling attempts 3 for instance 
2a3eadcc-5230-4ed6-bf9f-9a82d43b91c3
 94 2013-12-12 20:55:35.324 20945 WARNING nova.scheduler.utils 
[req-844a11f6-66f1-4fc7-9f3b-8bacbe57a04d e95097acd0b041558f33d07f720c1bd7 
354f17bf81924b278806c3e3798aa527] [instance: 
2a3eadcc-5230-4ed6-bf9f-9a82d43b91c3] Setting instance to ERROR state.

** Affects: neutron
 Importance: Undecided
 Status: New

** Attachment added: "server.log"
   https://bugs.launchpad.net/bugs/1260516/+attachment/3928318/+files/server.log

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1260516

Title:
  PortInUse exception leaves orphaned ports

Status in OpenStack Neutron (virtual network service):
  New

Bug description:
  While bringing up 1000 VMs with Heat, some VMs are in error state.
  After deleting the VMs, not all ports are removed.
  The next time VMs are created, they fail due to no IP addresses left.

  
  Note, initially there are 7 ports, no VMs
  804 VMs are created, 811 ports

  root@control01:/usr/share/pyshared/heat# nova list | grep 

[Yahoo-eng-team] [Bug 1260528] [NEW] Metering dashboard. Marker could not be found (havana)

2013-12-12 Thread Roman Sokolkov
Public bug reported:

Hello,

I couldn't reopen this bug
https://bugs.launchpad.net/horizon/+bug/1247752 . And decided to create
new one.

I use latest havana release code, but also plunged into "Marker could
not be found" error in horizon logs.

[Thu Dec 12 22:49:15 2013] [error] Request returned failure status: 400
[Thu Dec 12 22:49:18 2013] [error] REQ: curl -i -X GET 
http://192.168.0.2:35357/v2.0/tenants?marker=tenant_marker&limit=21 -H 
"User-Agent: python-keystoneclient" -H "Forwarded: 
for=10.20.0.1;by=python-keystonece"
[Thu Dec 12 22:49:18 2013] [error] RESP: [400] {'date': 'Thu, 12 Dec 2013 
22:49:18 GMT', 'content-type': 'application/json', 'content-length': '88', 
'vary': 'X-Auth-Token'}
[Thu Dec 12 22:49:18 2013] [error] RESP BODY: {"error": {"message": "Marker 
could not be found", "code": 400, "title": "Bad Request"}}

"tenant_marker" value comes from
https://github.com/openstack/horizon/blob/stable/havana/openstack_dashboard/dashboards/admin/metering/views.py#L149

** Affects: horizon
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1260528

Title:
  Metering dashboard. Marker could not be found (havana)

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  Hello,

  I couldn't reopen this bug
  https://bugs.launchpad.net/horizon/+bug/1247752 . And decided to
  create new one.

  I use latest havana release code, but also plunged into "Marker could
  not be found" error in horizon logs.

  [Thu Dec 12 22:49:15 2013] [error] Request returned failure status: 400
  [Thu Dec 12 22:49:18 2013] [error] REQ: curl -i -X GET 
http://192.168.0.2:35357/v2.0/tenants?marker=tenant_marker&limit=21 -H 
"User-Agent: python-keystoneclient" -H "Forwarded: 
for=10.20.0.1;by=python-keystonece"
  [Thu Dec 12 22:49:18 2013] [error] RESP: [400] {'date': 'Thu, 12 Dec 2013 
22:49:18 GMT', 'content-type': 'application/json', 'content-length': '88', 
'vary': 'X-Auth-Token'}
  [Thu Dec 12 22:49:18 2013] [error] RESP BODY: {"error": {"message": "Marker 
could not be found", "code": 400, "title": "Bad Request"}}

  "tenant_marker" value comes from
  
https://github.com/openstack/horizon/blob/stable/havana/openstack_dashboard/dashboards/admin/metering/views.py#L149

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1260528/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260530] [NEW] Instances appear pingable without an ingress icmp security rule

2013-12-12 Thread Ed Bak
Public bug reported:

Instances appear to be pingable for a short time after a floating ip is
associated even though there is no ingress icmp security group rule.
tcpdump of the instance's tap device shows that the instance isn't
actually responding to the ping.  It appears that the router gateway
interface is responding to the ping for a short time.  You can reproduce
this by booting an instance using a security group with only egress
rules.  Allocate a floating ip address. ping the ip address ( nothing
will happen yet ).  Associate the ip with the instance.  The ping will
begin responding.

** Affects: neutron
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1260530

Title:
  Instances appear pingable without an ingress icmp security rule

Status in OpenStack Neutron (virtual network service):
  New

Bug description:
  Instances appear to be pingable for a short time after a floating ip
  is associated even though there is no ingress icmp security group
  rule.  tcpdump of the instance's tap device shows that the instance
  isn't actually responding to the ping.  It appears that the router
  gateway interface is responding to the ping for a short time.  You can
  reproduce this by booting an instance using a security group with only
  egress rules.  Allocate a floating ip address. ping the ip address (
  nothing will happen yet ).  Associate the ip with the instance.  The
  ping will begin responding.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1260530/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1161988] Re: Flavor naming shouldn't include "m1"

2013-12-12 Thread Dean Troyer
** Changed in: devstack
   Status: Confirmed => Won't Fix

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1161988

Title:
  Flavor naming shouldn't include "m1"

Status in devstack - openstack dev environments:
  Won't Fix
Status in OpenStack Dashboard (Horizon):
  Won't Fix
Status in OpenStack Compute (Nova):
  Won't Fix
Status in Python client library for heat:
  Confirmed
Status in Python client library for Nova:
  Won't Fix
Status in Tempest:
  Won't Fix

Bug description:
  Flavor naming shouldn't include "m1"

  ENV: devstack trunk / nova 814e109845b3b2546f60e3f537dcfe32893906a3
  (grizzly)

  The default flavors are now:
  m1.nano 
  m1.micro
  m1.tiny 
  m1.small 
  m1.medium 
  m1.large 
  m1.xlarge

  We are propagating AWS "m1" designation. This is not useful
  information to the OpenStack administrator or user, and it's actually
  possible misinformation as the "m1" on AWS suggests a specific
  generation of hardware.

  POSSIBLE SOLUTION:

  Drop the "m1":
  nano 
  micro
  tiny 
  small 
  medium 
  large 
  xlarge

To manage notifications about this bug go to:
https://bugs.launchpad.net/devstack/+bug/1161988/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1217432] Re: timeout on AuthorizationTestJSON

2013-12-12 Thread Sean Dague
This is a glance call that's failing to allocate the image

** Also affects: glance
   Importance: Undecided
   Status: New

** Also affects: nova
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Glance.
https://bugs.launchpad.net/bugs/1217432

Title:
  timeout on AuthorizationTestJSON

Status in OpenStack Image Registry and Delivery Service (Glance):
  New
Status in OpenStack Compute (Nova):
  New
Status in Tempest:
  New

Bug description:
  http://logs.openstack.org/59/43459/3/gate/gate-tempest-devstack-vm-
  full/e57504d/console.html

  2013-08-27 14:39:29.384 | 
==
  2013-08-27 14:39:29.384 | FAIL: setUpClass 
(tempest.api.compute.test_authorization.AuthorizationTestJSON)
  2013-08-27 14:39:29.385 | setUpClass 
(tempest.api.compute.test_authorization.AuthorizationTestJSON)
  2013-08-27 14:39:29.385 | 
--
  2013-08-27 14:39:29.385 | _StringException: Traceback (most recent call last):
  2013-08-27 14:39:29.386 |   File "tempest/api/compute/test_authorization.py", 
line 66, in setUpClass
  2013-08-27 14:39:29.386 | 
cls.images_client.wait_for_image_status(image_id, 'ACTIVE')
  2013-08-27 14:39:29.386 |   File 
"tempest/services/compute/json/images_client.py", line 110, in 
wait_for_image_status
  2013-08-27 14:39:29.386 | raise exceptions.TimeoutException
  2013-08-27 14:39:29.386 | TimeoutException: Request timed out
  2013-08-27 14:39:29.386 | 
  2013-08-27 14:39:29.387 | 
  2013-08-27 14:39:29.387 | 
==
  2013-08-27 14:39:29.388 | FAIL: process-returncode
  2013-08-27 14:39:29.388 | process-returncode
  2013-08-27 14:39:29.416 | 
--
  2013-08-27 14:39:29.416 | _StringException: Binary content:
  2013-08-27 14:39:29.416 |   traceback (test/plain; charset="utf8")
  2013-08-27 14:39:29.416 | 
  2013-08-27 14:39:29.417 | 
  2013-08-27 14:39:29.417 | 
--
  2013-08-27 14:39:29.418 | Ran 1152 tests in 968.915s
  2013-08-27 14:39:29.418 | 
  2013-08-27 14:39:29.419 | FAILED (failures=2, skipped=67)

To manage notifications about this bug go to:
https://bugs.launchpad.net/glance/+bug/1217432/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260538] [NEW] nova-manage useage exposes action-args

2013-12-12 Thread Launchpad Bug Tracker
You have been subscribed to a public bug:

The nova-manage command exposes the action_args options during the usage
output for command.

E.g.
$ nova-manage network modify -h
usage: nova-manage network modify [-h] [--fixed_range ]
  [--project ] [--host ]
  [--disassociate-project]
  [--disassociate-host]
  [action_args [action_args ...]]

positional arguments:
  action_args



This can cause confusion as users naturally expect there to be more
"actions" on commands like "modify". Even in straightforward cases, this
positional argument leaks into usage.

$ nova-manage db version -h
usage: nova-manage db version [-h] [action_args [action_args ...]]

positional arguments:
  action_args

Please consider suppressing documentation on action_args. In addition,
expose the __doc__ strings for these functions, which is done in the
nova command.

** Affects: nova
 Importance: Undecided
 Status: New

-- 
nova-manage useage exposes action-args
https://bugs.launchpad.net/bugs/1260538
You received this bug notification because you are a member of Yahoo! 
Engineering Team, which is subscribed to OpenStack Compute (nova).

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260538] [NEW] nova-manage useage exposes action-args

2013-12-12 Thread Scott Devoid
Public bug reported:

The nova-manage command exposes the action_args options during the usage
output for command.

E.g.
$ nova-manage network modify -h
usage: nova-manage network modify [-h] [--fixed_range ]
  [--project ] [--host ]
  [--disassociate-project]
  [--disassociate-host]
  [action_args [action_args ...]]

positional arguments:
  action_args



This can cause confusion as users naturally expect there to be more
"actions" on commands like "modify". Even in straightforward cases, this
positional argument leaks into usage.

$ nova-manage db version -h
usage: nova-manage db version [-h] [action_args [action_args ...]]

positional arguments:
  action_args

Please consider suppressing documentation on action_args. In addition,
expose the __doc__ strings for these functions, which is done in the
nova command.

** Affects: nova
 Importance: Undecided
 Status: New


** Tags: low-hanging-fruit nova-manage user-experience ux

** Project changed: barbican => nova

** Tags added: low-hanging-fruit user-experience

** Tags added: nova-manage ux

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1260538

Title:
  nova-manage useage exposes action-args

Status in OpenStack Compute (Nova):
  New

Bug description:
  The nova-manage command exposes the action_args options during the
  usage output for command.

  E.g.
  $ nova-manage network modify -h
  usage: nova-manage network modify [-h] [--fixed_range ]
[--project ] [--host ]
[--disassociate-project]
[--disassociate-host]
[action_args [action_args ...]]

  positional arguments:
action_args

  

  This can cause confusion as users naturally expect there to be more
  "actions" on commands like "modify". Even in straightforward cases,
  this positional argument leaks into usage.

  $ nova-manage db version -h
  usage: nova-manage db version [-h] [action_args [action_args ...]]

  positional arguments:
action_args

  Please consider suppressing documentation on action_args. In addition,
  expose the __doc__ strings for these functions, which is done in the
  nova command.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1260538/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1217734] Re: FAIL: setUpClass (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML Unauthorized)

2013-12-12 Thread Sean Dague
if you can't find a gate race in logstash, I'm calling it fixed

** Changed in: nova
   Status: Incomplete => Fix Released

** Changed in: python-keystoneclient
   Status: Incomplete => Fix Released

** Changed in: tempest
   Status: Incomplete => Fix Released

** Changed in: python-cinderclient
   Status: Incomplete => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1217734

Title:
  FAIL: setUpClass
  (tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML
  Unauthorized)

Status in OpenStack Compute (Nova):
  Fix Released
Status in Python client library for Cinder:
  Fix Released
Status in Python client library for Keystone:
  Fix Released
Status in Tempest:
  Fix Released

Bug description:
  http://logs.openstack.org/44/43444/4/check/gate-grenade-devstack-
  vm/4f78566/console.html

  2013-08-28 06:32:58.510 | 
==
  2013-08-28 06:32:58.511 | FAIL: setUpClass 
(tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML)
  2013-08-28 06:32:58.511 | setUpClass 
(tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML)
  2013-08-28 06:32:58.511 | 
--
  2013-08-28 06:32:58.512 | _StringException: Traceback (most recent call last):
  2013-08-28 06:32:58.512 |   File 
"tempest/api/compute/servers/test_server_rescue.py", line 52, in setUpClass
  2013-08-28 06:32:58.512 | 'test_attach')
  2013-08-28 06:32:58.512 |   File 
"tempest/services/compute/xml/volumes_extensions_client.py", line 114, in 
create_volume
  2013-08-28 06:32:58.513 | self.headers)
  2013-08-28 06:32:58.513 |   File "tempest/common/rest_client.py", line 260, 
in post
  2013-08-28 06:32:58.513 | return self.request('POST', url, headers, body)
  2013-08-28 06:32:58.514 |   File "tempest/common/rest_client.py", line 388, 
in request
  2013-08-28 06:32:58.514 | resp, resp_body)
  2013-08-28 06:32:58.514 |   File "tempest/common/rest_client.py", line 430, 
in _error_checker
  2013-08-28 06:32:58.515 | raise exceptions.Unauthorized()
  2013-08-28 06:32:58.515 | Unauthorized: Unauthorized

  http://logs.openstack.org/23/43723/4/gate/gate-tempest-devstack-vm-
  full/3fefc90/console.html

  The real happening time is close to:
  2013-08-28 06:24:24.882 | setUpClass 
(tempest.api.compute.servers.test_server_rescue.ServerRescueTestXML)

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1217734/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260440] Re: nova-compute host is added to scheduling pool before Neutron can bind network ports on said host

2013-12-12 Thread OpenStack Infra
Reviewed:  https://review.openstack.org/61608
Committed: 
https://git.openstack.org/cgit/openstack/tripleo-incubator/commit/?id=661884b5c7a47d01171c680c83b601d3c9a15d9f
Submitter: Jenkins
Branch:master

commit 661884b5c7a47d01171c680c83b601d3c9a15d9f
Author: Clint Byrum 
Date:   Wed Dec 11 15:33:01 2013 -0800

Wait for Neutron L2 Agent on Compute Node

The L2 Agent sometimes does not register until later on in the
deployment for some reason. This is just a work-around until that bug
can be properly understood.

Change-Id: Idbbc977aa2e13f2026de05ae7e6571bc9dd0a498
Closes-Bug: #1260440


** Changed in: tripleo
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1260440

Title:
  nova-compute host is added to scheduling pool before Neutron can bind
  network ports on said host

Status in OpenStack Neutron (virtual network service):
  New
Status in OpenStack Compute (Nova):
  Confirmed
Status in tripleo - openstack on openstack:
  Fix Released

Bug description:
  This is a race condition.

  Given a cloud with 0 compute nodes available, on a compute node:
  * Start up neutron-openvswitch-agent
  * Start up nova-compute
  * nova boot an instance

  Scenario 1:
  * neutron-openvswitch-agent registers with Neutron before nova tries to boot 
instance
  * port is bound to agent
  * instance boots with correct networking

  Scenario 2:
  * nova schedules instance to host before neutron-openvswitch-agent is 
registered with Neutron
  * nova instance fails with vif_type=binding_failed
  * instance is in ERROR state

  I would expect that Nova would not try to schedule instances on
  compute hosts that are not ready.

  Please also see this mailing list thread for more info:

  http://lists.openstack.org/pipermail/openstack-
  dev/2013-December/022084.html

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1260440/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1249889] Re: tempest.scenario.test_volume_boot_pattern.TestVolumeBootPattern.test_volume_boot_pattern[compute, image, volume] failed

2013-12-12 Thread Sean Dague
not a tempest bug, this looks to be a nova bug with some sort of race on
nw attach

** Also affects: nova
   Importance: Undecided
   Status: New

** Changed in: tempest
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1249889

Title:
  
tempest.scenario.test_volume_boot_pattern.TestVolumeBootPattern.test_volume_boot_pattern[compute,image,volume]
  failed

Status in OpenStack Compute (Nova):
  New
Status in Tempest:
  Invalid

Bug description:
  Traceback (most recent call last):
File "tempest/scenario/test_volume_boot_pattern.py", line 144, in 
test_volume_boot_pattern
  keypair)
File "tempest/scenario/test_volume_boot_pattern.py", line 93, in 
_ssh_to_server
  server.add_floating_ip(floating_ip)
File "/opt/stack/new/python-novaclient/novaclient/v1_1/servers.py", line 
108, in add_floating_ip
  self.manager.add_floating_ip(self, address, fixed_address)
File "/opt/stack/new/python-novaclient/novaclient/v1_1/servers.py", line 
465, in add_floating_ip
  self._action('addFloatingIp', server, {'address': address})
File "/opt/stack/new/python-novaclient/novaclient/v1_1/servers.py", line 
993, in _action
  return self.api.client.post(url, body=body)
File "/opt/stack/new/python-novaclient/novaclient/client.py", line 234, in 
post
  return self._cs_request(url, 'POST', **kwargs)
File "/opt/stack/new/python-novaclient/novaclient/client.py", line 213, in 
_cs_request
  **kwargs)
File "/opt/stack/new/python-novaclient/novaclient/client.py", line 195, in 
_time_request
  resp, body = self.request(url, method, **kwargs)
File "/opt/stack/new/python-novaclient/novaclient/client.py", line 189, in 
request
  raise exceptions.from_response(resp, body, url, method)
  BadRequest: No nw_info cache associated with instance (HTTP 400) (Request-ID: 
req-4e6ed4cd-d2e8-42a2-aae6-f0a3820f71f5)

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1249889/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1225024] Re: tempest.api.compute.admin.test_hosts.HostsAdminTestXML.test_list_hosts_with_zone[gate] unexpected conductor service

2013-12-12 Thread Sean Dague
Can't find it in log stash, assuming it's fixed

** Changed in: nova
   Status: Incomplete => Fix Released

** Changed in: tempest
   Status: Incomplete => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1225024

Title:
  
tempest.api.compute.admin.test_hosts.HostsAdminTestXML.test_list_hosts_with_zone[gate]
  unexpected conductor service

Status in OpenStack Compute (Nova):
  Fix Released
Status in Tempest:
  Fix Released

Bug description:
  2013-09-13 03:44:02.095 | 
==
  2013-09-13 03:44:02.095 | FAIL: 
tempest.api.compute.admin.test_hosts.HostsAdminTestXML.test_list_hosts_with_zone[gate]
  2013-09-13 03:44:02.096 | 
tempest.api.compute.admin.test_hosts.HostsAdminTestXML.test_list_hosts_with_zone[gate]
  2013-09-13 03:44:02.096 | 
--
  2013-09-13 03:44:02.096 | _StringException: Empty attachments:
  2013-09-13 03:44:02.097 |   stderr
  2013-09-13 03:44:02.097 |   stdout
  2013-09-13 03:44:02.097 | 
  2013-09-13 03:44:02.098 | pythonlogging:'': {{{
  2013-09-13 03:44:02.098 | 2013-09-13 03:27:35,206 Request: GET 
http://127.0.0.1:8774/v2/b5dc34c995d94389b0b2a5f18851aca6/os-hosts
  2013-09-13 03:44:02.098 | 2013-09-13 03:27:35,487 Response Status: 200
  2013-09-13 03:44:02.099 | 2013-09-13 03:27:35,488 Nova request id: 
req-10082857-8c33-4472-b511-5a5945cc2da4
  2013-09-13 03:44:02.099 | 2013-09-13 03:27:35,488 Request: GET 
http://127.0.0.1:8774/v2/b5dc34c995d94389b0b2a5f18851aca6/os-hosts?zone=internal
  2013-09-13 03:44:02.099 | 2013-09-13 03:27:35,513 Response Status: 200
  2013-09-13 03:44:02.099 | 2013-09-13 03:27:35,513 Nova request id: 
req-29968204-ebbb-468c-ab8c-d25654550a97
  2013-09-13 03:44:02.100 | }}}
  2013-09-13 03:44:02.100 | 
  2013-09-13 03:44:02.100 | Traceback (most recent call last):
  2013-09-13 03:44:02.101 |   File "tempest/api/compute/admin/test_hosts.py", 
line 51, in test_list_hosts_with_zone
  2013-09-13 03:44:02.101 | self.assertIn(host, hosts)
  2013-09-13 03:44:02.101 |   File 
"/usr/local/lib/python2.7/dist-packages/testtools/testcase.py", line 328, in 
assertIn
  2013-09-13 03:44:02.101 | self.assertThat(haystack, Contains(needle))
  2013-09-13 03:44:02.102 |   File 
"/usr/local/lib/python2.7/dist-packages/testtools/testcase.py", line 417, in 
assertThat
  2013-09-13 03:44:02.102 | raise MismatchError(matchee, matcher, mismatch, 
verbose)
  2013-09-13 03:44:02.103 | MismatchError: {u'service': u'conductor', 
u'host_name': u'devstack-precise-hpcloud-az3-265828', u'zone': u'internal'} not 
in [{u'service': u'network', u'host_name': 
u'devstack-precise-hpcloud-az3-265828', u'zone': u'internal'}, {u'service': 
u'cert', u'host_name': u'devstack-precise-hpcloud-az3-265828', u'zone': 
u'internal'}, {u'service': u'scheduler', u'host_name': 
u'devstack-precise-hpcloud-az3-265828', u'zone': u'internal'}]
  2013-09-13 03:44:02.103 |

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1225024/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


  1   2   >