What about trying to swap NIC and/or cable? 

> 
> On 18 Oct 2019, at 16:05, Chuan Han <chuan...@google.com> wrote:
> 
> 
> Hi, Damjan,
> 
> It seems the bug is in vpp.
> 
> On R230, vpp shows eth0 is down.
> 
> vpp# sh hardware-interfaces eth0
>               Name                Idx   Link  Hardware
> eth0                               2    down  eth0
>   Link speed: unknown
>   Ethernet address b4:96:91:23:1e:d6
>   Intel 82599
>     carrier down
>     flags: admin-up promisc pmd rx-ip4-cksum
>     rx: queues 2 (max 128), desc 512 (min 32 max 4096 align 8)
>     tx: queues 3 (max 64), desc 512 (min 32 max 4096 align 8)
>     pci: device 8086:154d subsystem 8086:7b11 address 0000:06:00.01 numa 0
>     max rx packet len: 15872
>     promiscuous: unicast on all-multicast on
>     vlan offload: strip off filter off qinq off
>     rx offload avail:  vlan-strip ipv4-cksum udp-cksum tcp-cksum tcp-lro
>                        macsec-strip vlan-filter vlan-extend jumbo-frame 
> scatter
>                        security keep-crc
>     rx offload active: ipv4-cksum
>     tx offload avail:  vlan-insert ipv4-cksum udp-cksum tcp-cksum sctp-cksum
>                        tcp-tso macsec-insert multi-segs security
>     tx offload active: none
>     rss avail:         ipv4-tcp ipv4-udp ipv4 ipv6-tcp-ex ipv6-udp-ex ipv6-tcp
>                        ipv6-udp ipv6-ex ipv6
>     rss active:        none
>     tx burst function: (nil)
>     rx burst function: ixgbe_recv_pkts_vec
> 
>     extended stats:
>       mac local errors                                   318
> vpp#
> 
> However, the testpmd tool shows it is up.
> 
> testpmd> show port summary 1
> Number of available ports: 2
> Port MAC Address       Name         Driver         Status   Link
> 1    B4:96:91:23:1E:D6 0000:06:00.1 net_ixgbe      up       10000Mbps
> testpmd> 
> 
> Does this prove something wrong on vpp side?
> 
> Thanks.
> Chuan
> 
> On Fri, Oct 18, 2019 at 3:06 PM Chuan Han <chuan...@google.com> wrote:
>> I built testpmd binary on both r740 and r230, and ran the test. I did see 
>> testpmd reports some link status change on r230 server. testpmd report on 
>> r740 is stabler. no status change reported.  
>> 
>> r230 log
>> ================================
>> Press enter to exit
>> 
>> Port 0: link state change event
>> 
>> Port 1: link state change event
>> 
>> Port 1: link state change event
>> 
>> r740 log
>> ================================
>> Press enter to exitx0 - TX RS bit threshold=32
>> 
>> If it is a dpdk bug, what shall I do? Report to dpdk mailing list?
>> 
>> On Fri, Oct 18, 2019 at 11:55 AM Chuan Han via Lists.Fd.Io 
>> <chuanhan=google....@lists.fd.io> wrote:
>>> So, it is a dpdk bug?
>>> 
>>> I am new to dpdk/vpp. 
>>> 
>>> How do I run dpdk testpmd? Shall I install dpdk separately on the r230 
>>> server? Are there any steps to follow? 
>>> 
>>> On Fri, Oct 18, 2019 at 10:30 AM Damjan Marion <dmar...@me.com> wrote:
>>>> In this case we are purely relying on link state provided by DPDK.
>>>> Have you tried to check if same problem exists with DPDK testpmd app?
>>>> 
>>>> 
>>>>> On 18 Oct 2019, at 10:26, Chuan Han via Lists.Fd.Io 
>>>>> <chuanhan=google....@lists.fd.io> wrote:
>>>>> 
>>>>> I cleaned up startup config a bit. I can still see eth0 down. 
>>>>> 
>>>>> See attachment for config file and log. There are some errors in log, but 
>>>>> I am not sure they are worrisome or not. 
>>>>> 
>>>>> On Thu, Oct 17, 2019 at 5:22 PM Florin Coras <fcoras.li...@gmail.com> 
>>>>> wrote:
>>>>>> This looks like a DPDK issue, but I’ll let Damjan be the judge of that. 
>>>>>> 
>>>>>> To see if this is a config issues, could you simplify your startup 
>>>>>> config by
>>>>>> - removing “workers 0” from the two nics and adding “num-rx-queues 2” to 
>>>>>> the nics or to the default stanza, if you’re running with 2 workers
>>>>>> - comment out the cryptodev config 
>>>>>> 
>>>>>> If the two nics don’t come up, check if there’s any obvious dpdk error 
>>>>>> in “show log”.
>>>>>> 
>>>>>> Florin 
>>>>>> 
>>>>>>> On Oct 17, 2019, at 4:56 PM, Chuan Han via Lists.Fd.Io 
>>>>>>> <chuanhan=google....@lists.fd.io> wrote:
>>>>>>> 
>>>>>>> I tried disabling autoneg on R740 side. It is not allowed too. If vpp 
>>>>>>> cannot allow two nics to be successfully added to the same vpp 
>>>>>>> instance, it seems to be a bug. Is it something which can be easily 
>>>>>>> spotted in the code base? 
>>>>>>> 
>>>>>>> It is also not possible to enforce symmetricity on internet. The other 
>>>>>>> party can do anything as long as basic ping works. 
>>>>>>> 
>>>>>>> On Thu, Oct 17, 2019 at 3:55 PM Chuan Han <chuan...@google.com> wrote:
>>>>>>>> If I only put one phy nic, i.e., eth0, to vpp, 'sh hardware' shows it 
>>>>>>>> is up. If I put both eth0 and eth1 in vpp, eth0 is always down. It 
>>>>>>>> seems something is wrong with the nic or vpp does not support this 
>>>>>>>> type of hardware? 
>>>>>>>> 
>>>>>>>> We tried enabling autoneg on R230. It is not allowed. To avoid 
>>>>>>>> asymmetric settings, disabling autoneg on R740 will help? 
>>>>>>>> 
>>>>>>>> On Thu, Oct 17, 2019 at 3:46 PM Balaji Venkatraman (balajiv) 
>>>>>>>> <bala...@cisco.com> wrote:
>>>>>>>>> It plays a role if it is asymmetric at both ends. You could enable it 
>>>>>>>>> at both ends and check. 
>>>>>>>>> 
>>>>>>>>>> On Oct 17, 2019, at 3:15 PM, Chuan Han <chuan...@google.com> wrote:
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> I rebooted the r230 machine and found the phy nic corresponding to 
>>>>>>>>>> eth has autoneg off. 
>>>>>>>>>> 
>>>>>>>>>> root@esdn-relay:~/gnxi/perf_testing/r230# ethtool enp6s0f1
>>>>>>>>>> Settings for enp6s0f1:
>>>>>>>>>>         Supported ports: [ FIBRE ]
>>>>>>>>>>         Supported link modes:   10000baseT/Full 
>>>>>>>>>>         Supported pause frame use: Symmetric
>>>>>>>>>>         Supports auto-negotiation: No
>>>>>>>>>>         Supported FEC modes: Not reported
>>>>>>>>>>         Advertised link modes:  10000baseT/Full 
>>>>>>>>>>         Advertised pause frame use: Symmetric
>>>>>>>>>>         Advertised auto-negotiation: No
>>>>>>>>>>         Advertised FEC modes: Not reported
>>>>>>>>>>         Speed: 10000Mb/s
>>>>>>>>>>         Duplex: Full
>>>>>>>>>>         Port: Direct Attach Copper
>>>>>>>>>>         PHYAD: 0
>>>>>>>>>>         Transceiver: internal
>>>>>>>>>>         Auto-negotiation: off
>>>>>>>>>>         Supports Wake-on: d
>>>>>>>>>>         Wake-on: d
>>>>>>>>>>         Current message level: 0x00000007 (7)
>>>>>>>>>>                                drv probe link
>>>>>>>>>>         Link detected: yes
>>>>>>>>>> root@esdn-relay:~/gnxi/perf_testing/r230# 
>>>>>>>>>> 
>>>>>>>>>> On r740, autoneg is on. It is copper. 
>>>>>>>>>> 
>>>>>>>>>> root@esdn-lab:~/gnxi/perf_testing/r740/vpp# ethtool eno3
>>>>>>>>>> Settings for eno3:
>>>>>>>>>>         Supported ports: [ TP ]
>>>>>>>>>>         Supported link modes:   100baseT/Full 
>>>>>>>>>>                                 1000baseT/Full 
>>>>>>>>>>                                 10000baseT/Full 
>>>>>>>>>>         Supported pause frame use: Symmetric
>>>>>>>>>>         Supports auto-negotiation: Yes
>>>>>>>>>>         Supported FEC modes: Not reported
>>>>>>>>>>         Advertised link modes:  100baseT/Full 
>>>>>>>>>>                                 1000baseT/Full 
>>>>>>>>>>                                 10000baseT/Full 
>>>>>>>>>>         Advertised pause frame use: Symmetric
>>>>>>>>>>         Advertised auto-negotiation: Yes
>>>>>>>>>>         Advertised FEC modes: Not reported
>>>>>>>>>>         Speed: 10000Mb/s
>>>>>>>>>>         Duplex: Full
>>>>>>>>>>         Port: Twisted Pair
>>>>>>>>>>         PHYAD: 0
>>>>>>>>>>         Transceiver: internal
>>>>>>>>>>         Auto-negotiation: on
>>>>>>>>>>         MDI-X: Unknown
>>>>>>>>>>         Supports Wake-on: umbg
>>>>>>>>>>         Wake-on: g
>>>>>>>>>>         Current message level: 0x00000007 (7)
>>>>>>>>>>                                drv probe link
>>>>>>>>>>         Link detected: yes
>>>>>>>>>> root@esdn-lab:~/gnxi/perf_testing/r740/vpp# 
>>>>>>>>>> 
>>>>>>>>>> not clear if this plays a role or not. 
>>>>>>>>>> 
>>>>>>>>>> On Thu, Oct 17, 2019 at 2:41 PM Chuan Han via Lists.Fd.Io 
>>>>>>>>>> <chuanhan=google....@lists.fd.io> wrote:
>>>>>>>>>>> Restarting ixia controller does not help. We ended up with both 
>>>>>>>>>>> ixia ports having '!'. 
>>>>>>>>>>> 
>>>>>>>>>>> We are not sure how ixia port plays a role here. eth0 interfaces 
>>>>>>>>>>> are the interfaces connecting two servers, not to ixia. 
>>>>>>>>>>> 
>>>>>>>>>>> On Thu, Oct 17, 2019 at 11:26 AM Balaji Venkatraman (balajiv) 
>>>>>>>>>>> <bala...@cisco.com> wrote:
>>>>>>>>>>>> Hi Chuan,
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> Could you please try to reset the ixia controller connected to 
>>>>>>>>>>>> port 4?
>>>>>>>>>>>> 
>>>>>>>>>>>> I have seen issues with ‘!’ on ixia. Given the carrier on eth0 is 
>>>>>>>>>>>> down, I suspect the ixia port.
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> --
>>>>>>>>>>>> 
>>>>>>>>>>>> Regards,
>>>>>>>>>>>> 
>>>>>>>>>>>> Balaji. 
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> From: Chuan Han <chuan...@google.com>
>>>>>>>>>>>> Date: Thursday, October 17, 2019 at 11:09 AM
>>>>>>>>>>>> To: "Balaji Venkatraman (balajiv)" <bala...@cisco.com>
>>>>>>>>>>>> Cc: "vpp-dev@lists.fd.io" <vpp-dev@lists.fd.io>, Arivudainambi 
>>>>>>>>>>>> Appachi gounder <aappa...@google.com>, Jerry Cen 
>>>>>>>>>>>> <zhiw...@google.com>
>>>>>>>>>>>> Subject: Re: [vpp-dev] Basic l2 bridging does not work
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> Yes. It is unidirectional stream from port 1 to port 4.  
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> Another engineer, Nambi, configured ixia. What he showed me 
>>>>>>>>>>>> yesterday is that xia port connected to port 1 is green and good. 
>>>>>>>>>>>> ixia port connected to port 4 is green but has a red exclamation 
>>>>>>>>>>>> mark, which means ping does not work. 
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> We also found eth0 on R230 is down shown by "show hardware eth0" 
>>>>>>>>>>>> command. However "show int" shows it is up.
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> vpp# sh hardware-interfaces eth0
>>>>>>>>>>>>               Name                Idx   Link  Hardware
>>>>>>>>>>>> eth0                               2    down  eth0
>>>>>>>>>>>>   Link speed: unknown
>>>>>>>>>>>>   Ethernet address b4:96:91:23:1e:d6
>>>>>>>>>>>>   Intel 82599
>>>>>>>>>>>>     carrier down 
>>>>>>>>>>>>     flags: admin-up promisc pmd rx-ip4-cksum
>>>>>>>>>>>>     rx: queues 1 (max 128), desc 512 (min 32 max 4096 align 8)
>>>>>>>>>>>>     tx: queues 3 (max 64), desc 512 (min 32 max 4096 align 8)
>>>>>>>>>>>>     pci: device 8086:154d subsystem 8086:7b11 address 
>>>>>>>>>>>> 0000:06:00.01 numa 0
>>>>>>>>>>>>     max rx packet len: 15872
>>>>>>>>>>>>     promiscuous: unicast on all-multicast on
>>>>>>>>>>>>     vlan offload: strip off filter off qinq off
>>>>>>>>>>>>     rx offload avail:  vlan-strip ipv4-cksum udp-cksum tcp-cksum 
>>>>>>>>>>>> tcp-lro 
>>>>>>>>>>>>                        macsec-strip vlan-filter vlan-extend 
>>>>>>>>>>>> jumbo-frame scatter 
>>>>>>>>>>>>                        security keep-crc 
>>>>>>>>>>>>     rx offload active: ipv4-cksum 
>>>>>>>>>>>>     tx offload avail:  vlan-insert ipv4-cksum udp-cksum tcp-cksum 
>>>>>>>>>>>> sctp-cksum 
>>>>>>>>>>>>                        tcp-tso macsec-insert multi-segs security 
>>>>>>>>>>>>     tx offload active: none
>>>>>>>>>>>>     rss avail:         ipv4-tcp ipv4-udp ipv4 ipv6-tcp-ex 
>>>>>>>>>>>> ipv6-udp-ex ipv6-tcp 
>>>>>>>>>>>>                        ipv6-udp ipv6-ex ipv6 
>>>>>>>>>>>>     rss active:        none
>>>>>>>>>>>>     tx burst function: (nil)
>>>>>>>>>>>>     rx burst function: ixgbe_recv_pkts_vec
>>>>>>>>>>>> 
>>>>>>>>>>>>     rx frames ok                                       33278
>>>>>>>>>>>>     rx bytes ok                                      3960082
>>>>>>>>>>>>     extended stats:
>>>>>>>>>>>>       rx good packets                                  33278
>>>>>>>>>>>>       rx good bytes                                  3960082
>>>>>>>>>>>>       rx q0packets                                     33278
>>>>>>>>>>>>       rx q0bytes                                     3960082
>>>>>>>>>>>>       rx size 65 to 127 packets                        33278
>>>>>>>>>>>>       rx multicast packets                             33278
>>>>>>>>>>>>       rx total packets                                 33278
>>>>>>>>>>>>       rx total bytes                                 3960082
>>>>>>>>>>>> vpp# sh int
>>>>>>>>>>>>               Name               Idx    State  MTU 
>>>>>>>>>>>> (L3/IP4/IP6/MPLS)     Counter          Count     
>>>>>>>>>>>> eth0                              2      up          9000/0/0/0    
>>>>>>>>>>>>  rx packets                 33279
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  rx bytes                 3960201
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  drops                          5
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  punt                           1
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  tx-error                   33274
>>>>>>>>>>>> eth1                              1      up          9000/0/0/0    
>>>>>>>>>>>>  rx packets                 33274
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  rx bytes                 3959606
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  tx packets                 33273
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  tx bytes                 3959487
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  drops                      33274
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  tx-error                       3
>>>>>>>>>>>> local0                            0     down          0/0/0/0      
>>>>>>>>>>>>  
>>>>>>>>>>>> vpp# 
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> On Thu, Oct 17, 2019 at 10:54 AM Balaji Venkatraman (balajiv) 
>>>>>>>>>>>> <bala...@cisco.com> wrote:
>>>>>>>>>>>> 
>>>>>>>>>>>> Hi Chuan,
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> I assume u have unidirectional stream ? ixia->1->2->3->4->ixia?
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> vpp# sh int
>>>>>>>>>>>>               Name               Idx    State  MTU 
>>>>>>>>>>>> (L3/IP4/IP6/MPLS)     Counter          Count     
>>>>>>>>>>>> eth0                              2      up          9000/0/0/0    
>>>>>>>>>>>>  rx packets                 30925
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  rx bytes                 3680075
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  drops                          5
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  punt                           1
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  tx-error                   30920 
>>>>>>>>>>>> eth1                              1      up          9000/0/0/0    
>>>>>>>>>>>>  rx packets                 30920 <<< packets are received on port 
>>>>>>>>>>>> 3
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  rx bytes                 3679480
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  tx packets                 30919
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  tx bytes                 3679361
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  drops                      30920 <<< all dropped at port 3
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  tx-error                       3
>>>>>>>>>>>> local0                            0     down          0/0/0/0      
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> On sh error logs on R 230 we see
>>>>>>>>>>>> 
>>>>>>>>>>>>          1             ethernet-input             l3 mac mismatch 
>>>>>>>>>>>> <<<<
>>>>>>>>>>>>          3               eth1-output              interface is down
>>>>>>>>>>>>      30922               eth0-output              interface is down
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> Do u see the arp getting resolved on ixia? The mac on ixia at port 
>>>>>>>>>>>> with 172.16.1.2/24 should be seen on its other port. Are the ixia 
>>>>>>>>>>>> ports up at both ends?
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> --
>>>>>>>>>>>> 
>>>>>>>>>>>> Regards,
>>>>>>>>>>>> 
>>>>>>>>>>>> Balaji. 
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> From: <vpp-dev@lists.fd.io> on behalf of "Chuan Han via 
>>>>>>>>>>>> Lists.Fd.Io" <chuanhan=google....@lists.fd.io>
>>>>>>>>>>>> Reply-To: "chuan...@google.com" <chuan...@google.com>
>>>>>>>>>>>> Date: Thursday, October 17, 2019 at 9:59 AM
>>>>>>>>>>>> To: "Balaji Venkatraman (balajiv)" <bala...@cisco.com>
>>>>>>>>>>>> Cc: "vpp-dev@lists.fd.io" <vpp-dev@lists.fd.io>
>>>>>>>>>>>> Subject: Re: [vpp-dev] Basic l2 bridging does not work
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> It seems R740 vpp works fine. All packets coming from port 1 go to 
>>>>>>>>>>>> port 2. 
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> vpp# sh int
>>>>>>>>>>>>               Name               Idx    State  MTU 
>>>>>>>>>>>> (L3/IP4/IP6/MPLS)     Counter          Count     
>>>>>>>>>>>> eth0                              2      up          9000/0/0/0    
>>>>>>>>>>>>  tx packets                 30895
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  tx bytes                 3676505
>>>>>>>>>>>> eth1                              1      up          9000/0/0/0    
>>>>>>>>>>>>  rx packets                 30895
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  rx bytes                 3676505
>>>>>>>>>>>> local0                            0     down          0/0/0/0      
>>>>>>>>>>>>  
>>>>>>>>>>>> vpp# sh int
>>>>>>>>>>>>               Name               Idx    State  MTU 
>>>>>>>>>>>> (L3/IP4/IP6/MPLS)     Counter          Count     
>>>>>>>>>>>> eth0                              2      up          9000/0/0/0    
>>>>>>>>>>>>  tx packets                 30897
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  tx bytes                 3676743
>>>>>>>>>>>> eth1                              1      up          9000/0/0/0    
>>>>>>>>>>>>  rx packets                 30897
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  rx bytes                 3676743
>>>>>>>>>>>> local0                            0     down          0/0/0/0      
>>>>>>>>>>>>  
>>>>>>>>>>>> vpp# sh error
>>>>>>>>>>>>    Count                    Node                  Reason
>>>>>>>>>>>>      30899                l2-output               L2 output packets
>>>>>>>>>>>>      30899                l2-learn                L2 learn packets
>>>>>>>>>>>>          1                l2-learn                L2 learn misses
>>>>>>>>>>>>      30899                l2-input                L2 input packets
>>>>>>>>>>>>      30899                l2-flood                L2 flood packets
>>>>>>>>>>>> vpp# 
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> The drop happened on R230 vpp. Port 3 dropped all pkts complaining 
>>>>>>>>>>>> about down interface. However, show command shows interfaces are 
>>>>>>>>>>>> up. 
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> vpp# sh int
>>>>>>>>>>>>               Name               Idx    State  MTU 
>>>>>>>>>>>> (L3/IP4/IP6/MPLS)     Counter          Count     
>>>>>>>>>>>> eth0                              2      up          9000/0/0/0    
>>>>>>>>>>>>  rx packets                 30925
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  rx bytes                 3680075
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  drops                          5
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  punt                           1
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  tx-error                   30920
>>>>>>>>>>>> eth1                              1      up          9000/0/0/0    
>>>>>>>>>>>>  rx packets                 30920
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  rx bytes                 3679480
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  tx packets                 30919
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  tx bytes                 3679361
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  drops                      30920
>>>>>>>>>>>>                                                                    
>>>>>>>>>>>>  tx-error                       3
>>>>>>>>>>>> local0                            0     down          0/0/0/0      
>>>>>>>>>>>>  
>>>>>>>>>>>> vpp# sh error
>>>>>>>>>>>>    Count                    Node                  Reason
>>>>>>>>>>>>          2                llc-input               unknown llc 
>>>>>>>>>>>> ssap/dsap
>>>>>>>>>>>>      61846                l2-output               L2 output packets
>>>>>>>>>>>>      61846                l2-learn                L2 learn packets
>>>>>>>>>>>>          2                l2-learn                L2 learn misses
>>>>>>>>>>>>      61846                l2-input                L2 input packets
>>>>>>>>>>>>      61846                l2-flood                L2 flood packets
>>>>>>>>>>>>          1             ethernet-input             l3 mac mismatch
>>>>>>>>>>>>          3               eth1-output              interface is down
>>>>>>>>>>>>      30922               eth0-output              interface is down
>>>>>>>>>>>> vpp# 
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> Not sure how to check mac issues. Can you explain a bit more? Here 
>>>>>>>>>>>> is what I can see on R230 vpp. 
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> vpp# show bridge-domain 1 detail
>>>>>>>>>>>>   BD-ID   Index   BSN  Age(min)  Learning  U-Forwrd   UU-Flood   
>>>>>>>>>>>> Flooding  ARP-Term  arp-ufwd   BVI-Intf 
>>>>>>>>>>>>     1       1      0     off        on        on       flood       
>>>>>>>>>>>>  on       off       off        N/A    
>>>>>>>>>>>> 
>>>>>>>>>>>>            Interface           If-idx ISN  SHG  BVI  TxFlood       
>>>>>>>>>>>>  VLAN-Tag-Rewrite       
>>>>>>>>>>>>              eth0                2     1    0    -      *          
>>>>>>>>>>>>        none             
>>>>>>>>>>>>              eth1                1     1    0    -      *          
>>>>>>>>>>>>        none             
>>>>>>>>>>>> vpp# sh l2fib verbose           
>>>>>>>>>>>>     Mac-Address     BD-Idx If-Idx BSN-ISN Age(min) static filter 
>>>>>>>>>>>> bvi         Interface-Name        
>>>>>>>>>>>>  28:99:3a:f4:3a:a6    1      2      0/1      -       -      -     
>>>>>>>>>>>> -               eth0             
>>>>>>>>>>>>  28:99:3a:f4:3a:9c    1      1      0/1      -       -      -     
>>>>>>>>>>>> -               eth1             
>>>>>>>>>>>> L2FIB total/learned entries: 2/2  Last scan time: 0.0000e0sec  
>>>>>>>>>>>> Learn limit: 4194304 
>>>>>>>>>>>> vpp# 
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> On Wed, Oct 16, 2019 at 6:01 PM Balaji Venkatraman (balajiv) 
>>>>>>>>>>>> <bala...@cisco.com> wrote:
>>>>>>>>>>>> 
>>>>>>>>>>>> +-------------------------------------------------------------------------+
>>>>>>>>>>>> 
>>>>>>>>>>>> |                                                                  
>>>>>>>>>>>>        |
>>>>>>>>>>>> 
>>>>>>>>>>>> |                                                                  
>>>>>>>>>>>>        |
>>>>>>>>>>>> 
>>>>>>>>>>>> |                                   IXIA                           
>>>>>>>>>>>>        |
>>>>>>>>>>>> 
>>>>>>>>>>>> |                                                                  
>>>>>>>>>>>>        |
>>>>>>>>>>>> 
>>>>>>>>>>>> |                                                                  
>>>>>>>>>>>>        |
>>>>>>>>>>>> 
>>>>>>>>>>>> +-----------------------------------------------------------^-------------+
>>>>>>>>>>>> 
>>>>>>>>>>>>             |172.16.1.1/24                                  
>>>>>>>>>>>> |172.16.1.2/24
>>>>>>>>>>>> 
>>>>>>>>>>>>             |                                               |
>>>>>>>>>>>> 
>>>>>>>>>>>>             |                                               |
>>>>>>>>>>>> 
>>>>>>>>>>>>             |eth0                                           | eth0
>>>>>>>>>>>> 
>>>>>>>>>>>> +-----------v-------------+                    
>>>>>>>>>>>> +------------+-----------+
>>>>>>>>>>>> 
>>>>>>>>>>>> |           1             |                    |            4      
>>>>>>>>>>>>      |
>>>>>>>>>>>> 
>>>>>>>>>>>> |                         |                    |                   
>>>>>>>>>>>>      |
>>>>>>>>>>>> 
>>>>>>>>>>>> |                         |                    |                   
>>>>>>>>>>>>      |
>>>>>>>>>>>> 
>>>>>>>>>>>> |                         |                    |                   
>>>>>>>>>>>>      |
>>>>>>>>>>>> 
>>>>>>>>>>>> |                         |eth1          eth1  |                   
>>>>>>>>>>>>      |
>>>>>>>>>>>> 
>>>>>>>>>>>> |        VPP1           2 +--------------------> 3        VPP 2    
>>>>>>>>>>>>      |
>>>>>>>>>>>> 
>>>>>>>>>>>> |                         |                    |                   
>>>>>>>>>>>>      |
>>>>>>>>>>>> 
>>>>>>>>>>>> |                         |                    |                   
>>>>>>>>>>>>      |
>>>>>>>>>>>> 
>>>>>>>>>>>> |                         |                    |                   
>>>>>>>>>>>>      |
>>>>>>>>>>>> 
>>>>>>>>>>>> |                         |                    |                   
>>>>>>>>>>>>      |
>>>>>>>>>>>> 
>>>>>>>>>>>> +-------------------------+                    
>>>>>>>>>>>> +------------------------+
>>>>>>>>>>>> 
>>>>>>>>>>>>          R 740                                           R 230
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> Might help if you could check if the packet counts at ingress 
>>>>>>>>>>>> (port 1) & egress (port 2) match. Similarly 3 & 4. And the mac 
>>>>>>>>>>>> entries seen on both vpp(s). ARP req/rep tracing might also help.
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> /-
>>>>>>>>>>>> 
>>>>>>>>>>>> Balaji
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> From: <vpp-dev@lists.fd.io> on behalf of "Damjan Marion via 
>>>>>>>>>>>> Lists.Fd.Io" <dmarion=me....@lists.fd.io>
>>>>>>>>>>>> Reply-To: "dmar...@me.com" <dmar...@me.com>
>>>>>>>>>>>> Date: Wednesday, October 16, 2019 at 5:12 PM
>>>>>>>>>>>> To: "chuan...@google.com" <chuan...@google.com>
>>>>>>>>>>>> Cc: "vpp-dev@lists.fd.io" <vpp-dev@lists.fd.io>
>>>>>>>>>>>> Subject: Re: [vpp-dev] Basic l2 bridging does not work
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> On 16 Oct 2019, at 16:14, Chuan Han via Lists.Fd.Io 
>>>>>>>>>>>> <chuanhan=google....@lists.fd.io> wrote:
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> Hi, vpp experts, 
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> We are trying to make basic l2 bridge works within vpp. 
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> We have two servers: r230 and r740, each of which has two phy 
>>>>>>>>>>>> nics. Two servers are connected via cable. On each server, we 
>>>>>>>>>>>> bring these two nics into the same vpp instance and put them into 
>>>>>>>>>>>> the same l2 bridge. We tried sending traffic using ixia. However, 
>>>>>>>>>>>> ixia shows ping does not work. 
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> I attached the topology, vpp conf files, startup conf file, and 
>>>>>>>>>>>> logs.
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> Please advise where we could make it wrong.
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> Thanks.
>>>>>>>>>>>> 
>>>>>>>>>>>> Chuan
>>>>>>>>>>>> 
>>>>>>>>>>>> <r230 vpp.conf><r740 vpp.conf><r230 vpp startup.cfg><r740 vpp 
>>>>>>>>>>>> startup.cfg><r740.log><r230.log><vpp testbed - 
>>>>>>>>>>>> bridge.pdf>-=-=-=-=-=-=-=-=-=-=-=-
>>>>>>>>>>>> Links: You receive all messages sent to this group.
>>>>>>>>>>>> 
>>>>>>>>>>>> View/Reply Online (#14189): 
>>>>>>>>>>>> https://lists.fd.io/g/vpp-dev/message/14189
>>>>>>>>>>>> Mute This Topic: https://lists.fd.io/mt/34655826/675642
>>>>>>>>>>>> Group Owner: vpp-dev+ow...@lists.fd.io
>>>>>>>>>>>> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub  [dmar...@me.com]
>>>>>>>>>>>> -=-=-=-=-=-=-=-=-=-=-=-
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> On the 1st look everything look ok including the packet trace.
>>>>>>>>>>>> 
>>>>>>>>>>>> Can you try to clear counters* and enable packet trace on both 
>>>>>>>>>>>> instances.
>>>>>>>>>>>> 
>>>>>>>>>>>> Then send known number of packets and find put where drop happens 
>>>>>>>>>>>> by looking into same outputs you already shared.....
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> * "clear int", "clear run", "clear trace"
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>>> -- 
>>>>>>>>>>>> Damjan
>>>>>>>>>>>> 
>>>>>>>>>>>>  
>>>>>>>>>>>> 
>>>>>>>>>>> -=-=-=-=-=-=-=-=-=-=-=-
>>>>>>>>>>> Links: You receive all messages sent to this group.
>>>>>>>>>>> 
>>>>>>>>>>> View/Reply Online (#14211): 
>>>>>>>>>>> https://lists.fd.io/g/vpp-dev/message/14211
>>>>>>>>>>> Mute This Topic: https://lists.fd.io/mt/34655826/1991531
>>>>>>>>>>> Group Owner: vpp-dev+ow...@lists.fd.io
>>>>>>>>>>> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub  
>>>>>>>>>>> [chuan...@google.com]
>>>>>>>>>>> -=-=-=-=-=-=-=-=-=-=-=-
>>>>>>> -=-=-=-=-=-=-=-=-=-=-=-
>>>>>>> Links: You receive all messages sent to this group.
>>>>>>> 
>>>>>>> View/Reply Online (#14219): https://lists.fd.io/g/vpp-dev/message/14219
>>>>>>> Mute This Topic: https://lists.fd.io/mt/34655826/675152
>>>>>>> Group Owner: vpp-dev+ow...@lists.fd.io
>>>>>>> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub  
>>>>>>> [fcoras.li...@gmail.com]
>>>>>>> -=-=-=-=-=-=-=-=-=-=-=-
>>>>>> 
>>>>> <log><vpp startup.conf>-=-=-=-=-=-=-=-=-=-=-=-
>>>>> Links: You receive all messages sent to this group.
>>>>> 
>>>>> View/Reply Online (#14240): https://lists.fd.io/g/vpp-dev/message/14240
>>>>> Mute This Topic: https://lists.fd.io/mt/34655826/675642
>>>>> Group Owner: vpp-dev+ow...@lists.fd.io
>>>>> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub  [dmar...@me.com]
>>>>> -=-=-=-=-=-=-=-=-=-=-=-
>>>> 
>>>> -- 
>>>> Damjan
>>>> 
>>> -=-=-=-=-=-=-=-=-=-=-=-
>>> Links: You receive all messages sent to this group.
>>> 
>>> View/Reply Online (#14243): https://lists.fd.io/g/vpp-dev/message/14243
>>> Mute This Topic: https://lists.fd.io/mt/34655826/1991531
>>> Group Owner: vpp-dev+ow...@lists.fd.io
>>> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub  [chuan...@google.com]
>>> -=-=-=-=-=-=-=-=-=-=-=-
-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.

View/Reply Online (#14246): https://lists.fd.io/g/vpp-dev/message/14246
Mute This Topic: https://lists.fd.io/mt/34655826/21656
Group Owner: vpp-dev+ow...@lists.fd.io
Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub  [arch...@mail-archive.com]
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to