Hi all, I'm trying to enable numa for my instance based on https://blueprints.launchpad.net/nova/+spec/virt-driver-numa-placement.
I'm working under Ubuntu 14.10, with libvirt 1.2.8: virsh -v 1.2.8 My compute node has only 1 numa node: dmesg |grep numa [ 0.000000] mempolicy: Enabling automatic NUMA balancing. Configure with numa_balancing= or the kernel.numa_balancing sysctl numactl --hardware available: 1 nodes (0) node 0 cpus: 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 node 0 size: 15983 MB node 0 free: 8085 MB node distances: node 0 0: 10 I installed a fresh devstack. And updated flavor m1.medium with: nova flavor-key m1.medium set hw:numa_nodes=1 nova flavor-key m1.medium set hw:mem_page_size=large Then I try to start instance with command: nova boot --flavor m1.medium --image cirros-0.3.2-x86_64-uec --nic net-id=9b2afc82-b9d0-49ce-be21-732f3af506eb test1 The instance start failed due to reason: | fault | {"message": "No valid host was found. There are not enough hosts available.", "code": 500, "details": " File \"/opt/stack/nova/nova/conductor/manager.py\", line 651, in build_instances | | | request_spec, filter_properties) | | | File \"/opt/stack/nova/nova/scheduler/utils.py\", line 333, in wrapped | | | return func(*args, **kwargs) | | | File \"/opt/stack/nova/nova/scheduler/client/__init__.py\", line 52, in select_destinations | | | context, request_spec, filter_properties) | | | File \"/opt/stack/nova/nova/scheduler/client/__init__.py\", line 37, in __run_method | | | return getattr(self.instance, __name)(*args, **kwargs) | | | File \"/opt/stack/nova/nova/scheduler/client/query.py\", line 34, in select_destinations | | | context, request_spec, filter_properties) | | | File \"/opt/stack/nova/nova/scheduler/rpcapi.py\", line 114, in select_destinations | | | request_spec=request_spec, filter_properties=filter_properties) | | | File \"/usr/local/lib/python2.7/dist-packages/oslo_messaging/rpc/client.py\", line 156, in call | | | retry=self.retry) | | | File \"/usr/local/lib/python2.7/dist-packages/oslo_messaging/transport.py\", line 90, in _send | | | timeout=timeout, retry=retry) | | | File \"/usr/local/lib/python2.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py\", line 417, in send | | | retry=retry) | | | File \"/usr/local/lib/python2.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py\", line 408, in _send | | | raise result | | | ", "created": "2015-02-02T06:23:41Z"} | The log in compute.log: 2015-02-02 14:59:47.736 AUDIT nova.compute.manager [req-7e1b0ad2-0349-4afc-961b-322b3334f67a admin demo] [instance: 1a555c06-b1fd-405d-bc1a-d20a8ac1b7fb] Starting instance... 2015-02-02 14:59:47.738 DEBUG oslo_messaging._drivers.amqpdriver [-] MSG_ID is c323d0b819414dc38fe0188baef2cc98 from (pid=22569) _send /usr/local/lib/python2.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py:378 2015-02-02 14:59:47.738 DEBUG oslo_messaging._drivers.amqp [-] UNIQUE_ID is 2ce2555b27e6438db8c8d9043ab2a2f7. from (pid=22569) _add_unique_id /usr/local/lib/python2.7/dist-packages/oslo_messaging/_drivers/amqp.py:224 2015-02-02 14:59:47.900 DEBUG oslo_concurrency.lockutils [-] Lock "compute_resources" acquired by "instance_claim" :: waited 0.000s from (pid=22569) inner /usr/local/lib/python2.7/dist-packages/oslo_concurrency/lockutils.py:430 2015-02-02 14:59:47.900 DEBUG nova.compute.resource_tracker [-] Memory overhead for 4096 MB instance; 0 MB from (pid=22569) instance_claim /opt/stack/nova/nova/compute/resource_tracker.py:130 2015-02-02 14:59:47.902 AUDIT nova.compute.claims [-] [instance: 1a555c06-b1fd-405d-bc1a-d20a8ac1b7fb] Attempting claim: memory 4096 MB, disk 40 GB 2015-02-02 14:59:47.902 AUDIT nova.compute.claims [-] [instance: 1a555c06-b1fd-405d-bc1a-d20a8ac1b7fb] Total memory: 15983 MB, used: 512.00 MB 2015-02-02 14:59:47.902 AUDIT nova.compute.claims [-] [instance: 1a555c06-b1fd-405d-bc1a-d20a8ac1b7fb] memory limit: 23974.50 MB, free: 23462.50 MB 2015-02-02 14:59:47.902 AUDIT nova.compute.claims [-] [instance: 1a555c06-b1fd-405d-bc1a-d20a8ac1b7fb] Total disk: 915 GB, used: 0.00 GB 2015-02-02 14:59:47.902 AUDIT nova.compute.claims [-] [instance: 1a555c06-b1fd-405d-bc1a-d20a8ac1b7fb] disk limit not specified, defaulting to unlimited 2015-02-02 14:59:47.903 DEBUG oslo_messaging._drivers.amqpdriver [-] MSG_ID is 8010409e00644a43a4672f705e83a6cd from (pid=22569) _send /usr/local/lib/python2.7/dist-packages/oslo_messaging/_drivers/amqpdriver.py:378 2015-02-02 14:59:47.903 DEBUG oslo_messaging._drivers.amqp [-] UNIQUE_ID is 2c8e3b9dc627498d878fc7a8db7e92eb. from (pid=22569) _add_unique_id /usr/local/lib/python2.7/dist-packages/oslo_messaging/_drivers/amqp.py:224 2015-02-02 14:59:47.913 DEBUG nova.compute.resources.vcpu [-] Total CPUs: 24 VCPUs, used: 0.00 VCPUs from (pid=22569) test /opt/stack/nova/nova/compute/resources/vcpu.py:51 2015-02-02 14:59:47.913 DEBUG nova.compute.resources.vcpu [-] CPUs limit not specified, defaulting to unlimited from (pid=22569) test /opt/stack/nova/nova/compute/resources/vcpu.py:55 2015-02-02 14:59:47.913 DEBUG oslo_concurrency.lockutils [-] Lock "compute_resources" released by "instance_claim" :: held 0.013s from (pid=22569) inner /usr/local/lib/python2.7/dist-packages/oslo_concurrency/lockutils.py:442 2015-02-02 14:59:47.913 DEBUG nova.compute.manager [-] [instance: 1a555c06-b1fd-405d-bc1a-d20a8ac1b7fb] Insufficient compute resources: Requested instance NUMA topology cannot fit the given host NUMA topology. from (pid=22569) _build_and _run_instance /opt/stack/nova/nova/compute/manager.py:2183 2015-02-02 14:59:47.914 DEBUG nova.compute.utils [-] [instance: 1a555c06-b1fd-405d-bc1a-d20a8ac1b7fb] Insufficient compute resources: Requested instance NUMA topology cannot fit the given host NUMA topology. from (pid=22569) notify_about _instance_usage /opt/stack/nova/nova/compute/utils.py:324 2015-02-02 14:59:47.914 DEBUG nova.compute.manager [-] [instance: 1a555c06-b1fd-405d-bc1a-d20a8ac1b7fb] Build of instance 1a555c06-b1fd-405d-bc1a-d20a8ac1b7fb was re-scheduled: Insufficient compute resources: Requested instance NUMA topo logy cannot fit the given host NUMA topology. from (pid=22569) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2080 Anyone know why this happened ?? Also, if the compute node can't support instance NUMA topology , isn't it should failed in scheduler ??? Thanks. -chen
_______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack