Perfect! Thanks for checking!

Florin

> On Aug 23, 2018, at 7:35 PM, Yu, Ping <ping...@intel.com> wrote:
> 
> Thanks Florin, and this fix works pretty well.  
> Before the fix, it has problem to handle >500 sessions, and in this patch, it 
> works fine even >10000 sessions. J
>  
>   <>
>  <>From: vpp-dev@lists.fd.io [mailto:vpp-dev@lists.fd.io] On Behalf Of Florin 
> Coras
> Sent: Friday, August 24, 2018 9:37 AM
> To: Yu, Ping <ping...@intel.com>
> Cc: vpp-dev@lists.fd.io; Florin Coras (fcoras) <fco...@cisco.com>
> Subject: Re: [vpp-dev] TCP retransmission
>  
> Ping, 
>  
> Does this [1] fix the issues you’ve encountered?
>  
> Thanks, 
> Florin
>  
> [1] https://gerrit.fd.io/r/#/c/14453/ <https://gerrit.fd.io/r/#/c/14453/>
> 
> 
> On Aug 23, 2018, at 6:54 AM, Yu, Ping <ping...@intel.com 
> <mailto:ping...@intel.com>> wrote:
>  
> By checking these active session, we can see that in client side, these 
> sessions has sent out 8192 bytes, and the received < 8192 bytes, so echo 
> client is still waiting for packets.
>  
> Rx fifo: cursize 0 nitems 65536 has_event 0
> head 4284 tail 4284 segment manager 4
> vpp session 62 thread 0 app session 237 thread 0
> ooo pool 1 active elts newest 4294967295
>    [2856, 3908], len 1052, next -1, prev -1
> Tx fifo: cursize 0 nitems 65536 has_event 0
>  
>  
> While at the server side, we can see even if server has received 8192 bytes 
> from clients, and even if the tx tail is set to be 8192, but the head is 
> <8192.
>  
>  
> Rx fifo: cursize 0 nitems 65536 has_event 0
> head 8192 tail 8192 segment manager 2
> vpp session 4 thread 0 app session 0 thread 0
> ooo pool 0 active elts newest 4294967295
> Tx fifo: cursize 3908 nitems 65536 has_event 0
> head 4284 tail 8192 segment manager 2
>  
> thanks
> Ping
>  
>  
> From: Yu, Ping 
> Sent: Thursday, August 23, 2018 4:49 PM
> To: Florin Coras <fcoras.li...@gmail.com <mailto:fcoras.li...@gmail.com>>
> Cc: vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>; Florin Coras (fcoras) 
> <fco...@cisco.com <mailto:fco...@cisco.com>>; Yu, Ping <ping...@intel.com 
> <mailto:ping...@intel.com>>
> Subject: RE: [vpp-dev] TCP retransmission
>  
> Verbose 1 also shows some info below. Interesting to see the session index 
> are continuous.
>  
> [#0][T] 10.10.1.1:1111->10.10.1.2:23630           ESTABLISHED    0         
> 63056     71
> [#0][T] 10.10.1.1:1111->10.10.1.2:59371           ESTABLISHED    0         
> 63056     72
> [#0][T] 10.10.1.1:1111->10.10.1.2:20412           ESTABLISHED    0         
> 63056     73
> [#0][T] 10.10.1.1:1111->10.10.1.2:2961            ESTABLISHED    0         
> 63056     74
> [#0][T] 10.10.1.1:1111->10.10.1.2:52858           ESTABLISHED    0         
> 63056     75
> [#0][T] 10.10.1.1:1111->10.10.1.2:58311           ESTABLISHED    0         
> 63056     76
> [#0][T] 10.10.1.1:1111->10.10.1.2:49160           ESTABLISHED    0         
> 63056     77
> [#0][T] 10.10.1.1:1111->10.10.1.2:38413           ESTABLISHED    0         
> 63056     78
> [#0][T] 10.10.1.1:1111->10.10.1.2:65510           ESTABLISHED    0         
> 63056     79
> [#0][T] 10.10.1.1:1111->10.10.1.2:3043            ESTABLISHED    0         
> 63056     80
> [#0][T] 10.10.1.1:1111->10.10.1.2:15764           ESTABLISHED    0         
> 63056     81
> [#0][T] 10.10.1.1:1111->10.10.1.2:21193           ESTABLISHED    0         
> 63056     82
> [#0][T] 10.10.1.1:1111->10.10.1.2:56466           ESTABLISHED    0         
> 63056     83
> [#0][T] 10.10.1.1:1111->10.10.1.2:64575           ESTABLISHED    0         
> 63056     84
> [#0][T] 10.10.1.1:1111->10.10.1.2:54368           ESTABLISHED    0         
> 63056     85
> [#0][T] 10.10.1.1:1111->10.10.1.2:32197           ESTABLISHED    0         
> 63056     86
> [#0][T] 10.10.1.1:1111->10.10.1.2:36990           ESTABLISHED    0         
> 63056     87
> [#0][T] 10.10.1.1:1111->10.10.1.2:37083           ESTABLISHED    0         
> 63056     88
> [#0][T] 10.10.1.1:1111->10.10.1.2:11431           ESTABLISHED    0         
> 61628     411
> [#0][T] 10.10.1.1:1111->10.10.1.2:49626           ESTABLISHED    0         
> 61628     412
> [#0][T] 10.10.1.1:1111->10.10.1.2:42865           ESTABLISHED    0         
> 61628     413
> [#0][T] 10.10.1.1:1111->10.10.1.2:31260           ESTABLISHED    0         
> 61628     414
> [#0][T] 10.10.1.1:1111->10.10.1.2:20171           ESTABLISHED    0         
> 61628     415
> [#0][T] 10.10.1.1:1111->10.10.1.2:54702           ESTABLISHED    0         
> 61628     416
> [#0][T] 10.10.1.1:1111->10.10.1.2:8501            ESTABLISHED    0         
> 61628     417
> [#0][T] 10.10.1.1:1111->10.10.1.2:49424           ESTABLISHED    0         
> 61628     418
>  
>  
>  
> From: Yu, Ping 
> Sent: Thursday, August 23, 2018 4:40 PM
> To: Florin Coras <fcoras.li...@gmail.com <mailto:fcoras.li...@gmail.com>>
> Cc: vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>; Florin Coras (fcoras) 
> <fco...@cisco.com <mailto:fco...@cisco.com>>; Yu, Ping <ping...@intel.com 
> <mailto:ping...@intel.com>>
> Subject: RE: [vpp-dev] TCP retransmission
>  
> Hi, Florin,
>  
> Attached is the some active session live after echo client timeout.
>  
>  
> [#0][T] 10.10.1.1:1111->10.10.1.2:61337           ESTABLISHED
> flags: Recovery timers: []
> snd_una 4285 snd_nxt 4285 snd_una_max 4285 rcv_nxt 8193 rcv_las 8193
> snd_wnd 65536 rcv_wnd 65536 snd_wl1 8193 snd_wl2 4285
> flight size 0 out space 2856 cc space 2856 rcv_wnd_av 65536
> cong recovery cwnd 2856 ssthresh 2856 rtx_bytes 0 bytes_acked 0
> prev_ssthresh 524288 snd_congestion 8193 dupack 0 limited_transmit 1746905375
> tsecr 1714832397 tsecr_last_ack 1714832397
> rto 200 rto_boff 0 srtt 6 rttvar 2 rtt_ts 0 rtt_seq 2548066206
> tsval_recent 81891189 tsval_recent_age 13028
> scoreboard: sacked_bytes 0 last_sacked_bytes 0 lost_bytes 0
> last_bytes_delivered 0 high_sacked 2548066206 snd_una_adv 0
> cur_rxt_hole 4294967295 high_rxt 0 rescue_rxt 0
> Rx fifo: cursize 0 nitems 65536 has_event 0
> head 8192 tail 8192 segment manager 2
> vpp session 4 thread 0 app session 0 thread 0
> ooo pool 0 active elts newest 4294967295
> Tx fifo: cursize 3908 nitems 65536 has_event 0
> head 4284 tail 8192 segment manager 2
> vpp session 4 thread 0 app session 0 thread 0
> ooo pool 0 active elts newest 4294967295
> [#0][T] 10.10.1.1:1111->10.10.1.2:2210            ESTABLISHED
> flags: Recovery timers: []
> snd_una 4285 snd_nxt 4285 snd_una_max 4285 rcv_nxt 8193 rcv_las 8193
> snd_wnd 65536 rcv_wnd 65536 snd_wl1 8193 snd_wl2 4285
> flight size 0 out space 2856 cc space 2856 rcv_wnd_av 65536
> cong recovery cwnd 2856 ssthresh 2856 rtx_bytes 0 bytes_acked 0
> prev_ssthresh 524288 snd_congestion 8193 dupack 0 limited_transmit 1746905375
> tsecr 1714832397 tsecr_last_ack 1714832397
> rto 200 rto_boff 0 srtt 6 rttvar 2 rtt_ts 0 rtt_seq 2548066206
> tsval_recent 81891189 tsval_recent_age 13028
> scoreboard: sacked_bytes 0 last_sacked_bytes 0 lost_bytes 0
> last_bytes_delivered 0 high_sacked 2548066206 snd_una_adv 0
> cur_rxt_hole 4294967295 high_rxt 0 rescue_rxt 0
> Rx fifo: cursize 0 nitems 65536 has_event 0
> head 8192 tail 8192 segment manager 2
> vpp session 5 thread 0 app session 0 thread 0
> ooo pool 0 active elts newest 4294967295
> Tx fifo: cursize 3908 nitems 65536 has_event 0
> head 4284 tail 8192 segment manager 2
> vpp session 5 thread 0 app session 0 thread 0
> ooo pool 0 active elts newest 4294967295
> [#0][T] 10.10.1.1:1111->10.10.1.2:30199           ESTABLISHED
> flags: Recovery timers: []
> snd_una 5713 snd_nxt 5713 snd_una_max 5713 rcv_nxt 8193 rcv_las 8193
> snd_wnd 65536 rcv_wnd 65536 snd_wl1 8193 snd_wl2 5713
> flight size 0 out space 2856 cc space 2856 rcv_wnd_av 65536
> cong recovery cwnd 2856 ssthresh 2856 rtx_bytes 0 bytes_acked 0
> prev_ssthresh 524288 snd_congestion 8193 dupack 0 limited_transmit 1748569900
> tsecr 1714832397 tsecr_last_ack 1714832397
> rto 200 rto_boff 0 srtt 6 rttvar 1 rtt_ts 0 rtt_seq 2546405589
> tsval_recent 81891189 tsval_recent_age 13028
> scoreboard: sacked_bytes 0 last_sacked_bytes 0 lost_bytes 0
> last_bytes_delivered 0 high_sacked 2546403109 snd_una_adv 0
> cur_rxt_hole 4294967295 high_rxt 0 rescue_rxt 0
> Rx fifo: cursize 0 nitems 65536 has_event 0
> head 8192 tail 8192 segment manager 2
> vpp session 30 thread 0 app session 0 thread 0
> ooo pool 0 active elts newest 4294967295
> Tx fifo: cursize 2480 nitems 65536 has_event 0
> head 5712 tail 8192 segment manager 2
> vpp session 30 thread 0 app session 0 thread 0
> ooo pool 0 active elts newest 4294967295
> [#0][T] 10.10.1.1:1111->10.10.1.2:48888           ESTABLISHED
> flags: Recovery timers: []
> snd_una 5713 snd_nxt 5713 snd_una_max 5713 rcv_nxt 8193 rcv_las 8193
> snd_wnd 65536 rcv_wnd 65536 snd_wl1 8193 snd_wl2 5713
> flight size 0 out space 2856 cc space 2856 rcv_wnd_av 65536
> cong recovery cwnd 2856 ssthresh 2856 rtx_bytes 0 bytes_acked 0
> prev_ssthresh 524288 snd_congestion 8193 dupack 0 limited_transmit 1748569900
> tsecr 1714832397 tsecr_last_ack 1714832397
> rto 200 rto_boff 0 srtt 6 rttvar 1 rtt_ts 0 rtt_seq 2546405589
> tsval_recent 81891189 tsval_recent_age 13028
> scoreboard: sacked_bytes 0 last_sacked_bytes 0 lost_bytes 0
> last_bytes_delivered 0 high_sacked 2546403109 snd_una_adv 0
> cur_rxt_hole 4294967295 high_rxt 0 rescue_rxt 0
> Rx fifo: cursize 0 nitems 65536 has_event 0
> head 8192 tail 8192 segment manager 2
> vpp session 31 thread 0 app session 0 thread 0
> ooo pool 0 active elts newest 4294967295
> Tx fifo: cursize 2480 nitems 65536 has_event 0
> head 5712 tail 8192 segment manager 2
> vpp session 31 thread 0 app session 0 thread 0
> ooo pool 0 active elts newest 4294967295
>  
> From: Florin Coras [mailto:fcoras.li...@gmail.com 
> <mailto:fcoras.li...@gmail.com>] 
> Sent: Tuesday, August 21, 2018 11:18 PM
> To: Yu, Ping <ping...@intel.com <mailto:ping...@intel.com>>
> Cc: vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>; Florin Coras (fcoras) 
> <fco...@cisco.com <mailto:fco...@cisco.com>>
> Subject: Re: [vpp-dev] TCP retransmission
>  
> Hi Ping, 
>  
> Just from this, I can’t tell precisely what the issue is. Once the test 
> hangs, do “show session verbose 2” and you should get the detailed status of 
> the sessions that froze. I’ll take a look at those once I’m back in the 
> office. Retransmissions should work fine but one never knows. On the other 
> hand, the echo apps may be missing some io events. Did vpp report any other 
> error message? 
>  
> Florin
>  
> 
> On Aug 21, 2018, at 7:14 AM, Yu, Ping <ping...@intel.com 
> <mailto:ping...@intel.com>> wrote:
>  
> Hello, all
>  
> I am now debugging an issue observed in echo server/client. If nclients == 
> small number, then everything goes smoothly, but if nclients is big, such as 
> 500, 1000, echo client will hang without printing final report, and debug 
> shows that echo client is waiting for response from server as it is expecting 
> server to send bytes_to_receive to client.  The problems occurs in both TCP 
> and TLS protocol.
>  
> if (sp->bytes_to_receive > 0)
>         {
>           delete_session = 0;
>         }
>  
> My assumption is that due to the big traffic, some packet is missing and 
> re-transmission does not work well. Below the info for normal case, where 
> nclients == 100.
>  
> DBGvpp# show error
>    Count                    Node                  Reason
>        600              session-queue             Packets transmitted
>        100            tcp4-rcv-process            Pure ACKs received
>        100            tcp4-rcv-process            FINs received
>        600            tcp4-established            Packets pushed into rx fifo
>        600            tcp4-established            Pure ACKs received
>         32                ip4-glean               address overflow drops
>          1                ip4-glean               ARP requests sent
>          1                arp-input               ARP request IP4 source 
> address learned
>  
>  
> Server:
> DBGvpp# show error
>    Count                    Node                  Reason
>        600              session-queue             Packets transmitted
>        100               tcp4-listen              SYNs received
>        200            tcp4-rcv-process            Pure ACKs received
>        600            tcp4-established            Packets pushed into rx fifo
>        600            tcp4-established            Pure ACKs received
>        100            tcp4-established            FINs received
>          1                arp-input               ARP replies sent
>  
>  
> Below is the info when nclients == 500.  
>  
> Client:
>  
> DBGvpp# show error
>    Count                    Node                  Reason
>       3104              session-queue             Packets transmitted
>        498            tcp4-rcv-process            Pure ACKs received
>        498            tcp4-rcv-process            FINs received
>       2998            tcp4-established            Packets pushed into rx fifo
>       2892            tcp4-established            Pure ACKs received
>        106            tcp4-established            Duplicate ACK
>  
> Server:
> DBGvpp# show error
>    Count                    Node                  Reason
>       2998              session-queue             Packets transmitted
>        500               tcp4-listen              SYNs received
>        998            tcp4-rcv-process            Pure ACKs received
>       2996            tcp4-established            Packets pushed into rx fifo/
>          2            tcp4-established            OOO packets pushed into rx 
> fifo
>       2476            tcp4-established            Pure ACKs received
>        498            tcp4-established            FINs received
>  
> We can see that in clients side, packets transmission is 3104, which is 
> bigger than normal case 600*5, but in the server side, the recived packet is 
> only 2996, which is below 3000, which causes only 498 FIN is sent from 
> client, so 2 live session makes “test echo” hang until timeout. 
>  
>  
> Thanks
> Ping
>  
> -=-=-=-=-=-=-=-=-=-=-=-
> Links: You receive all messages sent to this group.
> 
> View/Reply Online (#10233): https://lists.fd.io/g/vpp-dev/message/10233 
> <https://lists.fd.io/g/vpp-dev/message/10233>
> Mute This Topic: https://lists.fd.io/mt/24876529/675152 
> <https://lists.fd.io/mt/24876529/675152>
> Group Owner: vpp-dev+ow...@lists.fd.io <mailto:vpp-dev+ow...@lists.fd.io>
> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub 
> <https://lists.fd.io/g/vpp-dev/unsub>  [fcoras.li...@gmail.com 
> <mailto:fcoras.li...@gmail.com>]
> -=-=-=-=-=-=-=-=-=-=-=-
>  
> -=-=-=-=-=-=-=-=-=-=-=-
> Links: You receive all messages sent to this group.
> 
> View/Reply Online (#10276): https://lists.fd.io/g/vpp-dev/message/10276 
> <https://lists.fd.io/g/vpp-dev/message/10276>
> Mute This Topic: https://lists.fd.io/mt/24876529/675152 
> <https://lists.fd.io/mt/24876529/675152>
> Group Owner: vpp-dev+ow...@lists.fd.io <mailto:vpp-dev+ow...@lists.fd.io>
> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub 
> <https://lists.fd.io/g/vpp-dev/unsub>  [fcoras.li...@gmail.com 
> <mailto:fcoras.li...@gmail.com>]
> -=-=-=-=-=-=-=-=-=-=-=-

-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.

View/Reply Online (#10277): https://lists.fd.io/g/vpp-dev/message/10277
Mute This Topic: https://lists.fd.io/mt/24876529/21656
Group Owner: vpp-dev+ow...@lists.fd.io
Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub  [arch...@mail-archive.com]
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to