On 1/16/2014 10:26 AM, Mark Vlcek (mavlcek) wrote:
I see the following related packages on my VM:

libguestfs.x86_64                      1:1.16.34-2.el6
python-libguestfs.x86_64               1:1.16.34-2.el6

But neither were able to be updated. Any other thoughts?

Thanks,
Mark

From: Yaguang Tang <heut2...@gmail.com <mailto:heut2...@gmail.com>>
Reply-To: "heut2...@gmail.com <mailto:heut2...@gmail.com>"
<heut2...@gmail.com <mailto:heut2...@gmail.com>>
Date: Thursday, January 16, 2014 8:44 AM
To: Mark Vlcek <mavl...@cisco.com <mailto:mavl...@cisco.com>>
Cc: "openstack@lists.openstack.org
<mailto:openstack@lists.openstack.org>" <openstack@lists.openstack.org
<mailto:openstack@lists.openstack.org>>
Subject: Re: [Openstack] [nova] No valid host found

from the exception we know an argument is not supported by python
guestfs ,, I tested this in ubuntu 13.10, it works well,

ii  libguestfs0:amd64                         1:1.22.4-1ubuntu1
                         amd64        guest disk image management system
- shared library
ii  python-guestfs                            1:1.22.4-1ubuntu1
                         amd64        guest disk image management system
- Python bindings

you can try to upgrade the related packages.


2014/1/14 Mark Vlcek (mavlcek) <mavl...@cisco.com
<mailto:mavl...@cisco.com>>

    Following up, when I attempt to launch a medium flavor image, here's
    the output in the nova scheduler log file:

    Error from last host: [u'Traceback (most recent call last):\n', u'
      File "/opt/stack/nova/nova/compute/manager.py", line 1054, in
    _build_instance\n    set_access_ip=set_access_ip)\n', u'  File
    "/opt/stack/nova/nova/compute/manager.py", line 357, in
    decorated_function\n    return function(self, context, *args,
    **kwargs)\n', u'  File "/opt/stack/nova/nova/compute/manager.py",
    line 1463, in _spawn\n    LOG.exception(_(\'Instance failed to
    spawn\'), instance=instance)\n', u'  File
    "/opt/stack/nova/nova/openstack/common/excutils.py", line 68, in
    __exit__\n    six.reraise(self.type_, self.value, self.tb)\n', u'
      File "/opt/stack/nova/nova/compute/manager.py", line 1460, in
    _spawn\n    block_device_info)\n', u'  File
    "/opt/stack/nova/nova/virt/libvirt/driver.py", line 2182, in spawn\n
      admin_pass=admin_password)\n', u'  File
    "/opt/stack/nova/nova/virt/libvirt/driver.py", line 2485, in
    _create_image\n    project_id=instance[\'project_id\'])\n', u'  File
    "/opt/stack/nova/nova/virt/libvirt/imagebackend.py", line 187, in
    cache\n    *args, **kwargs)\n', u'  File
    "/opt/stack/nova/nova/virt/libvirt/imagebackend.py", line 337, in
    create_image\n    copy_qcow2_image(base, self.path, size)\n', u'
      File "/opt/stack/nova/nova/openstack/common/lockutils.py", line
    249, in inner\n    return f(*args, **kwargs)\n', u'  File
    "/opt/stack/nova/nova/virt/libvirt/imagebackend.py", line 303, in
    copy_qcow2_image\n    disk.extend(target, size, use_cow=True)\n', u'
      File "/opt/stack/nova/nova/virt/disk/api.py", line 149, in
    extend\n    if not is_image_partitionless(image, use_cow):\n', u'
      File "/opt/stack/nova/nova/virt/disk/api.py", line 189, in
    is_image_partitionless\n    fs.setup()\n', u'  File
    "/opt/stack/nova/nova/virt/disk/vfs/guestfs.py", line 111, in
    setup\n    self.handle =
    tpool.Proxy(guestfs.GuestFS(close_on_exit=False))\n', u"TypeError:
    __init__() got an unexpected keyword argument 'close_on_exit'\n"]

    And here is the contents of my nova.conf file:

    [DEFAULT]
    flat_interface = eth1
    flat_network_bridge = br100
    vlan_interface = eth1
    public_interface = eth1
    network_manager = nova.network.manager.FlatDHCPManager
    firewall_driver = nova.virt.libvirt.firewall.IptablesFirewallDriver
    compute_driver = libvirt.LibvirtDriver
    use_usb_tablet = False
    libvirt_cpu_mode = none
    libvirt_type = qemu
    glance_api_servers = [my-ip-address]:9292
    rabbit_password = [my-rabbit-pw]
    rabbit_host = localhost
    rpc_backend = nova.openstack.common.rpc.impl_kombu
    ec2_dmz_host = [my-ip-address]
    vncserver_proxyclient_address = 127.0.0.1
    vncserver_listen = 127.0.0.1
    vnc_enabled = true
    xvpvncproxy_base_url = http://[my-ip-address]:6081/console
    novncproxy_base_url = http://[my-ip-address]:6080/vnc_auto.html
    notification_driver = nova.openstack.common.notifier.rpc_notifier
    notify_on_state_change = vm_and_task_state
    instance_usage_audit_period = hour
    instance_usage_audit = True
    logging_exception_prefix = %(color)s%(asctime)s.%(msecs)03d TRACE
    %(name)s %(instance)s
    logging_debug_format_suffix = from (pid=%(process)d) %(funcName)s
    %(pathname)s:%(lineno)d
    logging_default_format_string = %(asctime)s.%(msecs)03d
    %(color)s%(levelname)s %(name)s [-%(color)s]
    %(instance)s%(color)s%(message)s
    logging_context_format_string = %(asctime)s.%(msecs)03d
    %(color)s%(levelname)s %(name)s [%(request_id)s %(user_name)s
    %(project_name)s%(color)s] %(instance)s%(color)s%(message)s
    force_config_drive = always
    instances_path = /opt/stack/data/nova/instances
    lock_path = /opt/stack/data/nova
    state_path = /opt/stack/data/nova
    volume_api_class = nova.volume.cinder.API
    enabled_apis = ec2,osapi_compute,metadata
    bindir = /usr/bin
    instance_name_template = instance-%08x
    fatal_deprecations = True
    sql_connection = mysql://root:[mysql-pw]@127.0.0.1/nova?charset=utf8
    <http://127.0.0.1/nova?charset=utf8>
    metadata_workers = 4
    ec2_workers = 4
    osapi_compute_workers = 4
    my_ip = [my-ip-address]
    osapi_compute_extension =
    nova.api.openstack.compute.contrib.standard_extensions
    s3_port = 3333
    s3_host = [my-ip-address]
    default_floating_pool = public
    fixed_range =
    force_dhcp_release = True
    dhcpbridge_flagfile = /etc/nova/nova.conf
    scheduler_driver = nova.scheduler.filter_scheduler.FilterScheduler
    rootwrap_config = /etc/nova/rootwrap.conf
    api_paste_config = /etc/nova/api-paste.ini
    allow_resize_to_same_host = True
    auth_strategy = keystone
    debug = True
    verbose = True

    [conductor]
    workers = 4

    [osapi_v3]
    enabled = True

    [keystone_authtoken]
    signing_dir = /var/cache/nova
    admin_password = [my-admin-pw]
    admin_user = nova
    cafile =
    admin_tenant_name = service
    auth_protocol = http
    auth_port = 35357
    auth_host = [my-ip-address]

    [spice]
    enabled = false
    html5proxy_base_url = http://[my-ip-address]:6082/spice_auto.html

    Thanks,
    Mark

    From: Mark Vlcek <mavl...@cisco.com <mailto:mavl...@cisco.com>>
    Date: Monday, January 13, 2014 3:41 PM
    To: "openstack@lists.openstack.org
    <mailto:openstack@lists.openstack.org>"
    <openstack@lists.openstack.org <mailto:openstack@lists.openstack.org>>
    Subject: [nova] No valid host found

    Hi everyone,

    Whenever I try to launch an image instance beyond 'micro' (anything
    that has more than 0 GB disk space) I get the "no valid host found"
    error, even though my disk has 100+ GB of space, even displayed in
    the dashboard.

    The error in the nova scheduler says "Flavor's disk is too small for
    requested image." Any suggestions as to what I should try or what
    may be happening?

    FYI I'm running a DevStack based install of OpenStack.

    Thanks,
    Mark


    _______________________________________________
    Mailing list:
    http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
    Post to     : openstack@lists.openstack.org
    <mailto:openstack@lists.openstack.org>
    Unsubscribe :
    http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack




--
Tang Yaguang

Canonical Ltd. | www.ubuntu.com <http://www.ubuntu.com> |
www.canonical.com <http://www.canonical.com>
Mobile:  +86 152 1094 6968
gpg key: 0x187F664F


_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : openstack@lists.openstack.org
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack


A late response now but this has been fixed:

https://review.openstack.org/70354

--

Thanks,

Matt Riedemann


_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : openstack@lists.openstack.org
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack

Reply via email to