hum, do i get a speed ticket?
i did some tests before applying your patches:

lizard> ./tbench 1 dev
.1 clients started
..............+*
Throughput 6.10567 MB/sec (NB=7.63209 MB/sec  61.0567 MBit/sec)
lizard> ./tbench 2 dev
..2 clients started
........................+....+**
Throughput 7.71796 MB/sec (NB=9.64745 MB/sec  77.1796 MBit/sec)
lizard> ./tbench 3 dev
...3 clients started
.....................................+....+.+***
Throughput 9.17012 MB/sec (NB=11.4627 MB/sec  91.7012 MBit/sec)
lizard> ./tbench 4 dev
....4 clients started
........................................................++++****
Throughput 10.3365 MB/sec (NB=12.9207 MB/sec  103.365 MBit/sec)
lizard> ./tbench 5 dev
.....5 clients started
.............................................................+.......++.+.+****
*
Throughput 10.5219 MB/sec (NB=13.1523 MB/sec  105.219 MBit/sec)

>     Richard (and others), please try this patch.  With this patch I
>     get the following between two machines connected via a 100BaseTX
>     switch (full duplex):
> 
>       ----------------
> 
>     test1:/home/dillon/dbench> ./tbench 1 test2
>     .1 clients started
>     ..............+*
>     Throughput 6.13925 MB/sec (NB=7.67406 MB/sec  61.3925 MBit/sec)  1 procs
>     test1:/home/dillon/dbench> ./tbench 2 test2
>     ..2 clients started
>     ............................++**
>     Throughput 8.37795 MB/sec (NB=10.4724 MB/sec  83.7795 MBit/sec)  2 procs
> 
>       ----------------
> 
>      On localhost I get:
> 
>       ----------------
> 
>     test1:/home/dillon/dbench> ./tbench 1 localhost
>     .1 clients started
>     ..............+*
>     Throughput 25.7156 MB/sec (NB=32.1445 MB/sec  257.156 MBit/sec)  1 procs
>     test1:/home/dillon/dbench> ./tbench 2 localhost
>     ..2 clients started
>     ............................++**
>     Throughput 36.5428 MB/sec (NB=45.6785 MB/sec  365.428 MBit/sec)  2 procs
>     test1:/home/dillon/dbench> 
> 
>       ----------------
> 
>     This is WITHOUT changing the default send and receive tcp buffers..
>     they're both 16384.
> 
>     The bug I found is that when recv() is used with MSG_WAITALL, 
>     which is what tbench does, soreceive() will block waiting for all
>     available input WITHOUT ever calling pr->pr_usrreqs->pru_rcvd(),
>     which means that if the sender filled up the receive buffer (16K default)
>     the receiver will never ack the 0 window... that is until the idle code
>     takes over after 5 seconds.
> 
>                                       -Matt
> 
> Index: uipc_socket.c
> ===================================================================
> RCS file: /home/ncvs/src/sys/kern/uipc_socket.c,v
> retrieving revision 1.68.2.16
> diff -u -r1.68.2.16 uipc_socket.c
> --- uipc_socket.c     2001/06/14 20:46:06     1.68.2.16
> +++ uipc_socket.c     2001/12/01 21:09:13
> @@ -910,6 +910,14 @@
>                   !sosendallatonce(so) && !nextrecord) {
>                       if (so->so_error || so->so_state & SS_CANTRCVMORE)
>                               break;
> +                     /*
> +                      * The window might have closed to zero, make
> +                      * sure we send an ack now that we've drained
> +                      * the buffer or we might end up blocking until
> +                      * the idle takes over (5 seconds).
> +                      */
> +                     if (pr->pr_flags & PR_WANTRCVD && so->so_pcb)
> +                             (*pr->pr_usrreqs->pru_rcvd)(so, flags);
>                       error = sbwait(&so->so_rcv);
>                       if (error) {
>                               sbunlock(&so->so_rcv);
> 
> To Unsubscribe: send mail to [EMAIL PROTECTED]
> with "unsubscribe freebsd-hackers" in the body of the message
> 



To Unsubscribe: send mail to [EMAIL PROTECTED]
with "unsubscribe freebsd-hackers" in the body of the message

Reply via email to