Hi Jeff, Glad to hear that this is working for you ? What bonding mode do you use?
There is a limitation, as you saw, we cannot add additional MAC to the VF from DPDK. You can set MAC of the VF on VM by the same way you did on the pf. Let me know if you have any additional questions Best Regards, Olga From: Jesper Wramberg [mailto:jesper.wramb...@gmail.com] Sent: Monday, October 12, 2015 7:03 PM To: Olga Shern <olgas at mellanox.com> Cc: dev at dpdk.org Subject: Re: [dpdk-dev] Segmentation fault when bonding ports on Mellanox ConnectX-3 Hi again, The patches worked great and the DPDK bonding API functions with the ConnectX-3 now.. yay :-) I have run into some new trouble however and since its related I figured I would ask here if anyone could help. I am using SR-IOV to create a dual-port VF. When trying to bond the ports on this VF it seems that DPDK cannot set the mac address on the slaves. As a result, I am unable to receive data on the bonded port. As a workaround, I can set the mac address on the two VF ports manually in Linux using "ip link set <pf interface> vf 0 mac ..." after which I can start DPDK and receive data on the bonded port as wanted. Is there a way to allow DPDK to change the mac address so I avoid the workaround ? I have done some searching but couldn't find anything. Regards, Jesper Wramberg 2015-10-09 0:29 GMT+02:00 Olga Shern <olgas at mellanox.com<mailto:olgas at mellanox.com>>: For the sake of clarity, I assume you mean the patches about: "eal: new interrupt handler type" "mlx4: handle interrupts" [Olga ] Yes, you are right Best Regards, Olga 2015-10-08 17:36 GMT+02:00 Olga Shern <olgas at mellanox.com<mailto:olgas at mellanox.com>>: Hi Jesper, Bonding pmd is not supported with dpdk 2.1 on Mellanox nic We just sent patches to support async link events. Without these patches it will not work. Best Regards Olga Sent from Samsung Mobile. -------- Original message -------- From: Jesper Wramberg Date:08/10/2015 4:25 PM (GMT+00:00) To: dev at dpdk.org<mailto:dev at dpdk.org> Subject: [dpdk-dev] Segmentation fault when bonding ports on Mellanox ConnectX-3 Hi all, I was wondering if anyone has any experience with the ConnectX-3 card from Mellanox ? I have a server with such a card I can't seem to get link bonding to work. I've installed the necessary kernel modules, etc. and the card works as expected when testing it using e.g. the layer 2 forwarding example. If i try to run the bond example, however, I get a segmentation fault when the "rte_eth_bond_slave_add" function is called. Originally I wanted to bond the ports using the EAL cmd line option but the card only has one pci address :( Does anyone have any idea what could be causing this behavior ? I am running fw version 2.30.x which I have considered upgrading. Besides this I have been wondering if I have to do anything in Linux since the PMD uses the Linux drivers for control. I haven't been able to find any information on it though. Regards, Jesper Wramberg