I built testpmd binary on both r740 and r230, and ran the test. I did see testpmd reports some link status change on r230 server. testpmd report on r740 is stabler. no status change reported.
r230 log ================================ Press enter to exit Port 0: link state change event Port 1: link state change event Port 1: link state change event r740 log ================================ Press enter to exitx0 - TX RS bit threshold=32 If it is a dpdk bug, what shall I do? Report to dpdk mailing list? On Fri, Oct 18, 2019 at 11:55 AM Chuan Han via Lists.Fd.Io <chuanhan= google....@lists.fd.io> wrote: > So, it is a dpdk bug? > > I am new to dpdk/vpp. > > How do I run dpdk testpmd? Shall I install dpdk separately on the r230 > server? Are there any steps to follow? > > On Fri, Oct 18, 2019 at 10:30 AM Damjan Marion <dmar...@me.com> wrote: > >> In this case we are purely relying on link state provided by DPDK. >> Have you tried to check if same problem exists with DPDK testpmd app? >> >> >> On 18 Oct 2019, at 10:26, Chuan Han via Lists.Fd.Io < >> chuanhan=google....@lists.fd.io> wrote: >> >> I cleaned up startup config a bit. I can still see eth0 down. >> >> See attachment for config file and log. There are some errors in log, but >> I am not sure they are worrisome or not. >> >> On Thu, Oct 17, 2019 at 5:22 PM Florin Coras <fcoras.li...@gmail.com> >> wrote: >> >>> This looks like a DPDK issue, but I’ll let Damjan be the judge of that. >>> >>> To see if this is a config issues, could you simplify your startup >>> config by >>> - removing “workers 0” from the two nics and adding “num-rx-queues 2” to >>> the nics or to the default stanza, if you’re running with 2 workers >>> - comment out the cryptodev config >>> >>> If the two nics don’t come up, check if there’s any obvious dpdk error >>> in “show log”. >>> >>> Florin >>> >>> On Oct 17, 2019, at 4:56 PM, Chuan Han via Lists.Fd.Io >>> <http://lists.fd.io/> <chuanhan=google....@lists.fd.io> wrote: >>> >>> I tried disabling autoneg on R740 side. It is not allowed too. If vpp >>> cannot allow two nics to be successfully added to the same vpp instance, it >>> seems to be a bug. Is it something which can be easily spotted in the code >>> base? >>> >>> It is also not possible to enforce symmetricity on internet. The other >>> party can do anything as long as basic ping works. >>> >>> On Thu, Oct 17, 2019 at 3:55 PM Chuan Han <chuan...@google.com> wrote: >>> >>>> If I only put one phy nic, i.e., eth0, to vpp, 'sh hardware' shows it >>>> is up. If I put both eth0 and eth1 in vpp, eth0 is always down. It seems >>>> something is wrong with the nic or vpp does not support this type of >>>> hardware? >>>> >>>> We tried enabling autoneg on R230. It is not allowed. To avoid >>>> asymmetric settings, disabling autoneg on R740 will help? >>>> >>>> On Thu, Oct 17, 2019 at 3:46 PM Balaji Venkatraman (balajiv) < >>>> bala...@cisco.com> wrote: >>>> >>>>> It plays a role if it is asymmetric at both ends. You could enable it >>>>> at both ends and check. >>>>> >>>>> On Oct 17, 2019, at 3:15 PM, Chuan Han <chuan...@google.com> wrote: >>>>> >>>>> >>>>> I rebooted the r230 machine and found the phy nic corresponding to eth >>>>> has autoneg off. >>>>> >>>>> root@esdn-relay:~/gnxi/perf_testing/r230# ethtool enp6s0f1 >>>>> Settings for enp6s0f1: >>>>> Supported ports: [ FIBRE ] >>>>> Supported link modes: 10000baseT/Full >>>>> Supported pause frame use: Symmetric >>>>> Supports auto-negotiation: No >>>>> Supported FEC modes: Not reported >>>>> Advertised link modes: 10000baseT/Full >>>>> Advertised pause frame use: Symmetric >>>>> Advertised auto-negotiation: No >>>>> Advertised FEC modes: Not reported >>>>> Speed: 10000Mb/s >>>>> Duplex: Full >>>>> * Port: Direct Attach Copper* >>>>> PHYAD: 0 >>>>> Transceiver: internal >>>>> * Auto-negotiation: off* >>>>> Supports Wake-on: d >>>>> Wake-on: d >>>>> Current message level: 0x00000007 (7) >>>>> drv probe link >>>>> Link detected: yes >>>>> root@esdn-relay:~/gnxi/perf_testing/r230# >>>>> >>>>> On r740, autoneg is on. It is copper. >>>>> >>>>> root@esdn-lab:~/gnxi/perf_testing/r740/vpp# ethtool eno3 >>>>> Settings for eno3: >>>>> Supported ports: [ TP ] >>>>> Supported link modes: 100baseT/Full >>>>> 1000baseT/Full >>>>> 10000baseT/Full >>>>> Supported pause frame use: Symmetric >>>>> Supports auto-negotiation: Yes >>>>> Supported FEC modes: Not reported >>>>> Advertised link modes: 100baseT/Full >>>>> 1000baseT/Full >>>>> 10000baseT/Full >>>>> Advertised pause frame use: Symmetric >>>>> Advertised auto-negotiation: Yes >>>>> Advertised FEC modes: Not reported >>>>> Speed: 10000Mb/s >>>>> Duplex: Full >>>>> * Port: Twisted Pair* >>>>> PHYAD: 0 >>>>> Transceiver: internal >>>>> * Auto-negotiation: on* >>>>> MDI-X: Unknown >>>>> Supports Wake-on: umbg >>>>> Wake-on: g >>>>> Current message level: 0x00000007 (7) >>>>> drv probe link >>>>> Link detected: yes >>>>> root@esdn-lab:~/gnxi/perf_testing/r740/vpp# >>>>> >>>>> not clear if this plays a role or not. >>>>> >>>>> On Thu, Oct 17, 2019 at 2:41 PM Chuan Han via Lists.Fd.Io >>>>> <http://lists.fd.io/> <chuanhan=google....@lists.fd.io> wrote: >>>>> >>>>>> Restarting ixia controller does not help. We ended up with both ixia >>>>>> ports having '!'. >>>>>> >>>>>> We are not sure how ixia port plays a role here. eth0 interfaces are >>>>>> the interfaces connecting two servers, not to ixia. >>>>>> >>>>>> On Thu, Oct 17, 2019 at 11:26 AM Balaji Venkatraman (balajiv) < >>>>>> bala...@cisco.com> wrote: >>>>>> >>>>>>> Hi Chuan, >>>>>>> >>>>>>> >>>>>>> >>>>>>> Could you please try to reset the ixia controller connected to port >>>>>>> 4? >>>>>>> >>>>>>> I have seen issues with ‘!’ on ixia. Given the carrier on eth0 is >>>>>>> down, I suspect the ixia port. >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> Regards, >>>>>>> >>>>>>> Balaji. >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> *From: *Chuan Han <chuan...@google.com> >>>>>>> *Date: *Thursday, October 17, 2019 at 11:09 AM >>>>>>> *To: *"Balaji Venkatraman (balajiv)" <bala...@cisco.com> >>>>>>> *Cc: *"vpp-dev@lists.fd.io" <vpp-dev@lists.fd.io>, Arivudainambi >>>>>>> Appachi gounder <aappa...@google.com>, Jerry Cen <zhiw...@google.com >>>>>>> > >>>>>>> *Subject: *Re: [vpp-dev] Basic l2 bridging does not work >>>>>>> >>>>>>> >>>>>>> >>>>>>> Yes. It is unidirectional stream from port 1 to port 4. >>>>>>> >>>>>>> >>>>>>> >>>>>>> Another engineer, Nambi, configured ixia. What he showed me >>>>>>> yesterday is that xia port connected to port 1 is green and good. ixia >>>>>>> port >>>>>>> connected to port 4 is green but has a red exclamation mark, which means >>>>>>> ping does not work. >>>>>>> >>>>>>> >>>>>>> >>>>>>> We also found eth0 on R230 is down shown by "show hardware eth0" >>>>>>> command. However "show int" shows it is up. >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> vpp# sh hardware-interfaces eth0 >>>>>>> Name Idx Link Hardware >>>>>>> eth0 2 down eth0 >>>>>>> Link speed: unknown >>>>>>> Ethernet address b4:96:91:23:1e:d6 >>>>>>> Intel 82599 >>>>>>> carrier down >>>>>>> flags: admin-up promisc pmd rx-ip4-cksum >>>>>>> rx: queues 1 (max 128), desc 512 (min 32 max 4096 align 8) >>>>>>> tx: queues 3 (max 64), desc 512 (min 32 max 4096 align 8) >>>>>>> pci: device 8086:154d subsystem 8086:7b11 address 0000:06:00.01 >>>>>>> numa 0 >>>>>>> max rx packet len: 15872 >>>>>>> promiscuous: unicast on all-multicast on >>>>>>> vlan offload: strip off filter off qinq off >>>>>>> rx offload avail: vlan-strip ipv4-cksum udp-cksum tcp-cksum >>>>>>> tcp-lro >>>>>>> macsec-strip vlan-filter vlan-extend >>>>>>> jumbo-frame scatter >>>>>>> security keep-crc >>>>>>> rx offload active: ipv4-cksum >>>>>>> tx offload avail: vlan-insert ipv4-cksum udp-cksum tcp-cksum >>>>>>> sctp-cksum >>>>>>> tcp-tso macsec-insert multi-segs security >>>>>>> tx offload active: none >>>>>>> rss avail: ipv4-tcp ipv4-udp ipv4 ipv6-tcp-ex >>>>>>> ipv6-udp-ex ipv6-tcp >>>>>>> ipv6-udp ipv6-ex ipv6 >>>>>>> rss active: none >>>>>>> tx burst function: (nil) >>>>>>> rx burst function: ixgbe_recv_pkts_vec >>>>>>> >>>>>>> rx frames ok 33278 >>>>>>> rx bytes ok 3960082 >>>>>>> extended stats: >>>>>>> rx good packets 33278 >>>>>>> rx good bytes 3960082 >>>>>>> rx q0packets 33278 >>>>>>> rx q0bytes 3960082 >>>>>>> rx size 65 to 127 packets 33278 >>>>>>> rx multicast packets 33278 >>>>>>> rx total packets 33278 >>>>>>> rx total bytes 3960082 >>>>>>> vpp# sh int >>>>>>> Name Idx State MTU >>>>>>> (L3/IP4/IP6/MPLS) Counter Count >>>>>>> eth0 2 up 9000/0/0/0 >>>>>>> rx packets 33279 >>>>>>> rx >>>>>>> bytes 3960201 >>>>>>> >>>>>>> drops >>>>>>> 5 >>>>>>> punt >>>>>>> 1 >>>>>>> >>>>>>> tx-error >>>>>>> 33274 >>>>>>> eth1 1 up 9000/0/0/0 >>>>>>> rx packets 33274 >>>>>>> rx >>>>>>> bytes 3959606 >>>>>>> tx >>>>>>> packets 33273 >>>>>>> tx >>>>>>> bytes 3959487 >>>>>>> >>>>>>> drops >>>>>>> 33274 >>>>>>> >>>>>>> tx-error >>>>>>> 3 >>>>>>> local0 0 down 0/0/0/0 >>>>>>> vpp# >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Thu, Oct 17, 2019 at 10:54 AM Balaji Venkatraman (balajiv) < >>>>>>> bala...@cisco.com> wrote: >>>>>>> >>>>>>> Hi Chuan, >>>>>>> >>>>>>> >>>>>>> >>>>>>> I assume u have unidirectional stream ? ixia->1->2->3->4->ixia? >>>>>>> >>>>>>> >>>>>>> >>>>>>> vpp# sh int >>>>>>> Name Idx State MTU >>>>>>> (L3/IP4/IP6/MPLS) Counter Count >>>>>>> eth0 2 up 9000/0/0/0 >>>>>>> rx packets 30925 >>>>>>> rx >>>>>>> bytes 3680075 >>>>>>> >>>>>>> drops >>>>>>> 5 >>>>>>> punt >>>>>>> 1 >>>>>>> >>>>>>> tx-error >>>>>>> 30920 >>>>>>> eth1 1 up 9000/0/0/0 >>>>>>> rx packets 30920 <<< packets are received on port 3 >>>>>>> rx >>>>>>> bytes 3679480 >>>>>>> tx >>>>>>> packets 30919 >>>>>>> tx >>>>>>> bytes 3679361 >>>>>>> >>>>>>> drops >>>>>>> 30920 <<< all dropped at port 3 >>>>>>> >>>>>>> tx-error >>>>>>> 3 >>>>>>> local0 0 down 0/0/0/0 >>>>>>> >>>>>>> >>>>>>> >>>>>>> On sh error logs on R 230 we see >>>>>>> >>>>>>> 1 ethernet-input l3 mac mismatch >>>>>>> <<<< >>>>>>> 3 eth1-output interface is down >>>>>>> 30922 eth0-output interface is down >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> Do u see the arp getting resolved on ixia? The mac on ixia at port >>>>>>> with 172.16.1.2/24 should be seen on its other port. Are the ixia >>>>>>> ports up at both ends? >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> >>>>>>> Regards, >>>>>>> >>>>>>> Balaji. >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> *From: *<vpp-dev@lists.fd.io> on behalf of "Chuan Han via >>>>>>> Lists.Fd.Io <http://lists.fd.io/>" <chuanhan=google....@lists.fd.io> >>>>>>> *Reply-To: *"chuan...@google.com" <chuan...@google.com> >>>>>>> *Date: *Thursday, October 17, 2019 at 9:59 AM >>>>>>> *To: *"Balaji Venkatraman (balajiv)" <bala...@cisco.com> >>>>>>> *Cc: *"vpp-dev@lists.fd.io" <vpp-dev@lists.fd.io> >>>>>>> *Subject: *Re: [vpp-dev] Basic l2 bridging does not work >>>>>>> >>>>>>> >>>>>>> >>>>>>> It seems R740 vpp works fine. All packets coming from port 1 go to >>>>>>> port 2. >>>>>>> >>>>>>> >>>>>>> >>>>>>> vpp# sh int >>>>>>> Name Idx State MTU >>>>>>> (L3/IP4/IP6/MPLS) Counter Count >>>>>>> eth0 2 up 9000/0/0/0 >>>>>>> tx packets 30895 >>>>>>> tx >>>>>>> bytes 3676505 >>>>>>> eth1 1 up 9000/0/0/0 >>>>>>> rx packets 30895 >>>>>>> rx >>>>>>> bytes 3676505 >>>>>>> local0 0 down 0/0/0/0 >>>>>>> vpp# sh int >>>>>>> Name Idx State MTU >>>>>>> (L3/IP4/IP6/MPLS) Counter Count >>>>>>> eth0 2 up 9000/0/0/0 >>>>>>> tx packets 30897 >>>>>>> tx >>>>>>> bytes 3676743 >>>>>>> eth1 1 up 9000/0/0/0 >>>>>>> rx packets 30897 >>>>>>> rx >>>>>>> bytes 3676743 >>>>>>> local0 0 down 0/0/0/0 >>>>>>> vpp# sh error >>>>>>> Count Node Reason >>>>>>> 30899 l2-output L2 output packets >>>>>>> 30899 l2-learn L2 learn packets >>>>>>> 1 l2-learn L2 learn misses >>>>>>> 30899 l2-input L2 input packets >>>>>>> 30899 l2-flood L2 flood packets >>>>>>> vpp# >>>>>>> >>>>>>> >>>>>>> >>>>>>> The drop happened on R230 vpp. Port 3 dropped all pkts complaining >>>>>>> about down interface. However, show command shows interfaces are up. >>>>>>> >>>>>>> >>>>>>> >>>>>>> vpp# sh int >>>>>>> Name Idx State MTU >>>>>>> (L3/IP4/IP6/MPLS) Counter Count >>>>>>> eth0 2 up 9000/0/0/0 >>>>>>> rx packets 30925 >>>>>>> rx >>>>>>> bytes 3680075 >>>>>>> >>>>>>> drops >>>>>>> 5 >>>>>>> punt >>>>>>> 1 >>>>>>> >>>>>>> tx-error >>>>>>> 30920 >>>>>>> eth1 1 up 9000/0/0/0 >>>>>>> rx packets 30920 >>>>>>> rx >>>>>>> bytes 3679480 >>>>>>> tx >>>>>>> packets 30919 >>>>>>> tx >>>>>>> bytes 3679361 >>>>>>> >>>>>>> drops >>>>>>> 30920 >>>>>>> >>>>>>> tx-error >>>>>>> 3 >>>>>>> local0 0 down 0/0/0/0 >>>>>>> vpp# sh error >>>>>>> Count Node Reason >>>>>>> 2 llc-input unknown llc >>>>>>> ssap/dsap >>>>>>> 61846 l2-output L2 output packets >>>>>>> 61846 l2-learn L2 learn packets >>>>>>> 2 l2-learn L2 learn misses >>>>>>> 61846 l2-input L2 input packets >>>>>>> 61846 l2-flood L2 flood packets >>>>>>> 1 ethernet-input l3 mac mismatch >>>>>>> 3 eth1-output interface is down >>>>>>> 30922 eth0-output interface is down >>>>>>> vpp# >>>>>>> >>>>>>> >>>>>>> >>>>>>> Not sure how to check mac issues. Can you explain a bit more? Here >>>>>>> is what I can see on R230 vpp. >>>>>>> >>>>>>> >>>>>>> >>>>>>> vpp# show bridge-domain 1 detail >>>>>>> BD-ID Index BSN Age(min) Learning U-Forwrd UU-Flood >>>>>>> Flooding ARP-Term arp-ufwd BVI-Intf >>>>>>> 1 1 0 off on on flood >>>>>>> on off off N/A >>>>>>> >>>>>>> Interface If-idx ISN SHG BVI TxFlood >>>>>>> VLAN-Tag-Rewrite >>>>>>> eth0 2 1 0 - * >>>>>>> none >>>>>>> eth1 1 1 0 - * >>>>>>> none >>>>>>> vpp# sh l2fib verbose >>>>>>> Mac-Address BD-Idx If-Idx BSN-ISN Age(min) static filter >>>>>>> bvi Interface-Name >>>>>>> 28:99:3a:f4:3a:a6 1 2 0/1 - - - - >>>>>>> eth0 >>>>>>> 28:99:3a:f4:3a:9c 1 1 0/1 - - - - >>>>>>> eth1 >>>>>>> L2FIB total/learned entries: 2/2 Last scan time: 0.0000e0sec Learn >>>>>>> limit: 4194304 >>>>>>> vpp# >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Wed, Oct 16, 2019 at 6:01 PM Balaji Venkatraman (balajiv) < >>>>>>> bala...@cisco.com> wrote: >>>>>>> >>>>>>> >>>>>>> +-------------------------------------------------------------------------+ >>>>>>> >>>>>>> | >>>>>>> | >>>>>>> >>>>>>> | >>>>>>> | >>>>>>> >>>>>>> | >>>>>>> IXIA | >>>>>>> >>>>>>> | >>>>>>> | >>>>>>> >>>>>>> | >>>>>>> | >>>>>>> >>>>>>> >>>>>>> +-----------------------------------------------------------^-------------+ >>>>>>> >>>>>>> |172.16.1.1/24 | >>>>>>> 172.16.1.2/24 >>>>>>> >>>>>>> | | >>>>>>> >>>>>>> | | >>>>>>> >>>>>>> |eth0 | eth0 >>>>>>> >>>>>>> +-----------v-------------+ >>>>>>> +------------+-----------+ >>>>>>> >>>>>>> | 1 | | >>>>>>> 4 | >>>>>>> >>>>>>> | | >>>>>>> | | >>>>>>> >>>>>>> | | >>>>>>> | | >>>>>>> >>>>>>> | | >>>>>>> | | >>>>>>> >>>>>>> | |eth1 eth1 >>>>>>> | | >>>>>>> >>>>>>> | VPP1 2 +--------------------> 3 VPP >>>>>>> 2 | >>>>>>> >>>>>>> | | | >>>>>>> | >>>>>>> >>>>>>> | | >>>>>>> | | >>>>>>> >>>>>>> | | >>>>>>> | | >>>>>>> >>>>>>> | | >>>>>>> | | >>>>>>> >>>>>>> +-------------------------+ >>>>>>> +------------------------+ >>>>>>> >>>>>>> R 740 R 230 >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> Might help if you could check if the packet counts at ingress (port >>>>>>> 1) & egress (port 2) match. Similarly 3 & 4. And the mac entries seen on >>>>>>> both vpp(s). ARP req/rep tracing might also help. >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> /- >>>>>>> >>>>>>> Balaji >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> *From: *<vpp-dev@lists.fd.io> on behalf of "Damjan Marion via >>>>>>> Lists.Fd.Io <http://lists.fd.io/>" <dmarion=me....@lists.fd.io> >>>>>>> *Reply-To: *"dmar...@me.com" <dmar...@me.com> >>>>>>> *Date: *Wednesday, October 16, 2019 at 5:12 PM >>>>>>> *To: *"chuan...@google.com" <chuan...@google.com> >>>>>>> *Cc: *"vpp-dev@lists.fd.io" <vpp-dev@lists.fd.io> >>>>>>> *Subject: *Re: [vpp-dev] Basic l2 bridging does not work >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> On 16 Oct 2019, at 16:14, Chuan Han via Lists.Fd.Io >>>>>>> <http://lists.fd.io/> <chuanhan=google....@lists.fd.io> wrote: >>>>>>> >>>>>>> >>>>>>> >>>>>>> Hi, vpp experts, >>>>>>> >>>>>>> >>>>>>> >>>>>>> We are trying to make basic l2 bridge works within vpp. >>>>>>> >>>>>>> >>>>>>> >>>>>>> We have two servers: r230 and r740, each of which has two phy nics. >>>>>>> Two servers are connected via cable. On each server, we bring these two >>>>>>> nics into the same vpp instance and put them into the same l2 bridge. We >>>>>>> tried sending traffic using ixia. However, ixia shows ping does not >>>>>>> work. >>>>>>> >>>>>>> >>>>>>> >>>>>>> I attached the topology, vpp conf files, startup conf file, and logs. >>>>>>> >>>>>>> >>>>>>> >>>>>>> Please advise where we could make it wrong. >>>>>>> >>>>>>> >>>>>>> >>>>>>> Thanks. >>>>>>> >>>>>>> Chuan >>>>>>> >>>>>>> <r230 vpp.conf><r740 vpp.conf><r230 vpp startup.cfg><r740 vpp >>>>>>> startup.cfg><r740.log><r230.log><vpp testbed - >>>>>>> bridge.pdf>-=-=-=-=-=-=-=-=-=-=-=- >>>>>>> Links: You receive all messages sent to this group. >>>>>>> >>>>>>> View/Reply Online (#14189): >>>>>>> https://lists.fd.io/g/vpp-dev/message/14189 >>>>>>> Mute This Topic: https://lists.fd.io/mt/34655826/675642 >>>>>>> Group Owner: vpp-dev+ow...@lists.fd.io >>>>>>> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [dmar...@me.com] >>>>>>> -=-=-=-=-=-=-=-=-=-=-=- >>>>>>> >>>>>>> >>>>>>> >>>>>>> On the 1st look everything look ok including the packet trace. >>>>>>> >>>>>>> Can you try to clear counters* and enable packet trace on both >>>>>>> instances. >>>>>>> >>>>>>> Then send known number of packets and find put where drop happens by >>>>>>> looking into same outputs you already shared..... >>>>>>> >>>>>>> >>>>>>> >>>>>>> * "clear int", "clear run", "clear trace" >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Damjan >>>>>>> >>>>>>> >>>>>>> >>>>>>> -=-=-=-=-=-=-=-=-=-=-=- >>>>>> Links: You receive all messages sent to this group. >>>>>> >>>>>> View/Reply Online (#14211): >>>>>> https://lists.fd.io/g/vpp-dev/message/14211 >>>>>> Mute This Topic: https://lists.fd.io/mt/34655826/1991531 >>>>>> Group Owner: vpp-dev+ow...@lists.fd.io >>>>>> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [ >>>>>> chuan...@google.com] >>>>>> -=-=-=-=-=-=-=-=-=-=-=- >>>>>> >>>>> -=-=-=-=-=-=-=-=-=-=-=- >>> Links: You receive all messages sent to this group. >>> >>> View/Reply Online (#14219): https://lists.fd.io/g/vpp-dev/message/14219 >>> Mute This Topic: https://lists.fd.io/mt/34655826/675152 >>> Group Owner: vpp-dev+ow...@lists.fd.io >>> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [ >>> fcoras.li...@gmail.com] >>> -=-=-=-=-=-=-=-=-=-=-=- >>> >>> >>> <log><vpp startup.conf>-=-=-=-=-=-=-=-=-=-=-=- >> Links: You receive all messages sent to this group. >> >> View/Reply Online (#14240): https://lists.fd.io/g/vpp-dev/message/14240 >> Mute This Topic: https://lists.fd.io/mt/34655826/675642 >> Group Owner: vpp-dev+ow...@lists.fd.io >> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [dmar...@me.com] >> -=-=-=-=-=-=-=-=-=-=-=- >> >> >> -- >> Damjan >> >> -=-=-=-=-=-=-=-=-=-=-=- > Links: You receive all messages sent to this group. > > View/Reply Online (#14243): https://lists.fd.io/g/vpp-dev/message/14243 > Mute This Topic: https://lists.fd.io/mt/34655826/1991531 > Group Owner: vpp-dev+ow...@lists.fd.io > Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [chuan...@google.com] > -=-=-=-=-=-=-=-=-=-=-=- >
root@esdn-relay:~/dpdk/x86_64-native-linuxapp-gcc/app# ./testpmd EAL: Detected 8 lcore(s) EAL: Detected 1 NUMA nodes EAL: Multi-process socket /var/run/dpdk/rte/mp_socket EAL: No available hugepages reported in hugepages-1048576kB EAL: Probing VFIO support... EAL: PCI device 0000:06:00.0 on NUMA socket -1 EAL: Invalid NUMA socket, default to 0 EAL: probe driver: 8086:154d net_ixgbe EAL: PCI device 0000:06:00.1 on NUMA socket -1 EAL: Invalid NUMA socket, default to 0 EAL: probe driver: 8086:154d net_ixgbe testpmd: create a new mbuf pool <mbuf_pool_socket_0>: n=203456, size=2176, socket=0 testpmd: preferred mempool ops selected: ring_mp_mc Configuring Port 0 (socket 0) Port 0: B4:96:91:23:1E:D4 Configuring Port 1 (socket 0) Port 1: B4:96:91:23:1E:D6 Checking link statuses... Done No commandline core given, start packet forwarding io packet forwarding - ports=2 - cores=1 - streams=2 - NUMA support enabled, MP allocation mode: native Logical Core 1 (socket 0) forwards packets on 2 streams: RX P=0/Q=0 (socket 0) -> TX P=1/Q=0 (socket 0) peer=02:00:00:00:00:01 RX P=1/Q=0 (socket 0) -> TX P=0/Q=0 (socket 0) peer=02:00:00:00:00:00 io packet forwarding packets/burst=32 nb forwarding cores=1 - nb forwarding ports=2 port 0: RX queue number: 1 Tx queue number: 1 Rx offloads=0x0 Tx offloads=0x0 RX queue: 0 RX desc=256 - RX free threshold=32 RX threshold registers: pthresh=8 hthresh=8 wthresh=0 RX Offloads=0x0 TX queue: 0 TX desc=256 - TX free threshold=32 TX threshold registers: pthresh=32 hthresh=0 wthresh=0 TX offloads=0x0 - TX RS bit threshold=32 port 1: RX queue number: 1 Tx queue number: 1 Rx offloads=0x0 Tx offloads=0x0 RX queue: 0 RX desc=256 - RX free threshold=32 RX threshold registers: pthresh=8 hthresh=8 wthresh=0 RX Offloads=0x0 TX queue: 0 TX desc=256 - TX free threshold=32 TX threshold registers: pthresh=32 hthresh=0 wthresh=0 TX offloads=0x0 - TX RS bit threshold=32 Press enter to exit Port 0: link state change event Port 1: link state change event Port 1: link state change event
root@esdn-lab:~/dpdk/x86_64-native-linuxapp-gcc/app# ./testpmd EAL: Detected 96 lcore(s) EAL: Detected 2 NUMA nodes EAL: Multi-process socket /var/run/dpdk/rte/mp_socket EAL: Probing VFIO support... EAL: PCI device 0000:19:00.0 on NUMA socket 0 EAL: probe driver: 8086:1563 net_ixgbe EAL: PCI device 0000:19:00.1 on NUMA socket 0 EAL: probe driver: 8086:1563 net_ixgbe EAL: PCI device 0000:1a:00.0 on NUMA socket 0 EAL: probe driver: 8086:1563 net_ixgbe EAL: PCI device 0000:1a:00.1 on NUMA socket 0 EAL: probe driver: 8086:1563 net_ixgbe testpmd: create a new mbuf pool <mbuf_pool_socket_0>: n=907456, size=2176, socket=0 testpmd: preferred mempool ops selected: ring_mp_mc testpmd: create a new mbuf pool <mbuf_pool_socket_1>: n=907456, size=2176, socket=1 testpmd: preferred mempool ops selected: ring_mp_mc Configuring Port 0 (socket 0) Port 0: 24:6E:96:B4:B2:05 Configuring Port 1 (socket 0) Port 1: 24:6E:96:B4:B2:06 Checking link statuses... Done No commandline core given, start packet forwarding io packet forwarding - ports=2 - cores=1 - streams=2 - NUMA support enabled, MP allocation mode: native Logical Core 1 (socket 1) forwards packets on 2 streams: RX P=0/Q=0 (socket 0) -> TX P=1/Q=0 (socket 0) peer=02:00:00:00:00:01 RX P=1/Q=0 (socket 0) -> TX P=0/Q=0 (socket 0) peer=02:00:00:00:00:00 io packet forwarding packets/burst=32 nb forwarding cores=1 - nb forwarding ports=2 port 0: RX queue number: 1 Tx queue number: 1 Rx offloads=0x0 Tx of[81459.659018] irq 209: nobody cared (try booting with the "irqpoll" option) floads=0x0 [81459.666782] handlers: [81459.670411] [<000000006be52f14>] uio_interrupt [uio] [81459.675385] Disabling IRQ #209 RX queue: 0 RX desc=256 - RX free threshold=32 RX threshold registers: pthresh=8 hthresh=8 wthresh=0 RX Offloads=0x0 TX queue: 0 TX desc=256 - TX free threshold=32 TX threshold registers: pthresh=32 hthresh=0 wthresh=0 TX offloads=0x0 - TX RS bit threshold=32 port 1: RX queue number: 1 Tx queue number: 1 Rx offloads=0x0 Tx offloads=0x0 RX queue: 0 RX desc=256 - RX free threshold=32 RX threshold registers: pthresh=8 hthresh=8 wthresh=0 RX Offloads=0x0 TX queue: 0 TX desc=256 - TX free threshold=32 TX threshold registers: pthresh=32 hthresh=0 wthresh=0 Press enter to exitx0 - TX RS bit threshold=32
-=-=-=-=-=-=-=-=-=-=-=- Links: You receive all messages sent to this group. View/Reply Online (#14244): https://lists.fd.io/g/vpp-dev/message/14244 Mute This Topic: https://lists.fd.io/mt/34655826/21656 Group Owner: vpp-dev+ow...@lists.fd.io Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [arch...@mail-archive.com] -=-=-=-=-=-=-=-=-=-=-=-