Mike, Here is the bug which I reported https://bugs.launchpad.net/bugs/1795920
Cc'ing: Sean Sent from my iPhone > On Nov 12, 2018, at 8:27 AM, Satish Patel <satish....@gmail.com> wrote: > > Mike, > > I had same issue month ago when I roll out sriov in my cloud and this is what > I did to solve this issue. Set following in flavor > > hw:numa_nodes=2 > > It will spread out instance vcpu across numa, yes there will be little > penalty but if you tune your application according they you are good > > Yes this is bug I have already open ticket and I believe folks are working on > it but its not simple fix. They may release new feature in coming oprnstack > release. > > Sent from my iPhone > >> On Nov 11, 2018, at 9:25 PM, Mike Joseph <m...@mode.net> wrote: >> >> Hi folks, >> >> It appears that the numa_policy attribute of a PCI alias is ignored for >> flavors referencing that alias if the flavor also has >> hw:cpu_policy=dedicated set. The alias config is: >> >> alias = { "name": "mlx", "device_type": "type-VF", "vendor_id": "15b3", >> "product_id": "1004", "numa_policy": "preferred" } >> >> And the flavor config is: >> >> { >> "OS-FLV-DISABLED:disabled": false, >> "OS-FLV-EXT-DATA:ephemeral": 0, >> "access_project_ids": null, >> "disk": 10, >> "id": "221e1bcd-2dde-48e6-bd09-820012198908", >> "name": "vm-2", >> "os-flavor-access:is_public": true, >> "properties": "hw:cpu_policy='dedicated', pci_passthrough:alias='mlx:1'", >> "ram": 8192, >> "rxtx_factor": 1.0, >> "swap": "", >> "vcpus": 2 >> } >> >> In short, our compute nodes have an SR-IOV Mellanox NIC (ConnectX-3) with 16 >> VFs configured. We wish to expose these VFs to VMs that schedule on the >> host. However, the NIC is in NUMA region 0 which means that only half of >> the compute node's CPU cores would be usable if we required VM affinity to >> the NIC's NUMA region. But we don't need that, since we are okay with >> cross-region access to the PCI device. >> >> However, we do need CPU pinning to work, in order to have efficient cache >> hits on our VM processes. Therefore, we still want to pin our vCPUs to >> pCPUs, even if the pins end up on on a NUMA region opposite of the NIC. The >> spec for numa_policy seem to indicate that this is exactly the intent of the >> option: >> >> https://specs.openstack.org/openstack/nova-specs/specs/queens/implemented/share-pci-between-numa-nodes.html >> >> But, with the above config, we still get PCI affinity scheduling errors: >> >> 'Insufficient compute resources: Requested instance NUMA topology together >> with requested PCI devices cannot fit the given host NUMA topology.' >> >> This strikes me as a bug, but perhaps I am missing something here? >> >> Thanks, >> MJ >> _______________________________________________ >> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >> Post to : openstack@lists.openstack.org >> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
_______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack