> On Oct 24, 2014, at 12:52 AM, Ergin Ozekes <ergin.oze...@gmail.com> wrote: > > Hi All; > > My problem's solution is as below: > > ############################################################################## > # Buffer size > ############################################################################## > CONFIG proxy.config.net.sock_recv_buffer_size_in INT 10485760 > CONFIG proxy.config.net.sock_recv_buffer_size_out INT 10485760 > CONFIG proxy.config.net.sock_send_buffer_size_in INT 10485760 > CONFIG proxy.config.net.sock_send_buffer_size_out INT 10485760
Great! You might want to check the buffers on your network adaptors as well (ethtool -g IIRC) > > Best Regards, > Ergin > > On Thu, Oct 23, 2014 at 6:33 PM, Ergin Ozekes <ergin.oze...@gmail.com> > wrote: > >> Hi all; >> >> In according to my investigation on the ats read >> performance,NetHandler::mainNetEvent is the top function. On my system >> lots of buffer over run is exist. In my system net input interface has 250 >> Mbps traffic but on the client output only 125 Mbps is exist. Half of the >> downloaded content is returning to client. >> >> Does anyone knows the reason of this case? >> And how can I fixe it? >> >> This is the "perf top" output: >> 4.82% libpcre.so.3.13.1 [.] match >> >> >> 3.27% [kernel] [k] native_write_msr_safe >> >> >> 2.09% [kernel] [k] ipt_do_table >> >> >> 1.60% [kernel] [k] __schedule >> >> >> 1.57% [kernel] [k] menu_select >> >> >> 1.43% libc-2.13.so [.] 0x00000000000e92b0 >> >> >> 1.35% [kernel] [k] find_busiest_group >> >> >> 1.16% traffic_server [.] EThread::execute() >> >> >> 1.12% [kernel] [k] copy_user_generic_string >> >> >> 1.10% [kernel] [k] nf_iterate >> >> >> 1.04% [kernel] [k] _raw_spin_lock_irqsave >> >> >> 1.03% [kernel] [k] int_sqrt >> >> >> *1.02% traffic_server [.] NetHandler::mainNetEvent(int, >> Event*)* >> >> 0.96% [bnx2x] [k] bnx2x_rx_int >> >> >> 0.93% [kernel] [k] _raw_spin_lock >> >> >> 0.90% [kernel] [k] htable_selective_cleanup >> >> >> 0.86% [kernel] [k] cpuidle_enter_state >> >> >> 0.83% [kernel] [k] enqueue_task_fair >> >> >> 0.83% [kernel] [k] tcp_packet >> >> >> 0.76% [kernel] [k] apic_timer_interrupt >> >> >> 0.76% [kernel] [k] timerqueue_add >> >> >> 0.76% [kernel] [k] idle_cpu >> >> >> 0.71% [bnx2x] [k] bnx2x_start_xmit >> >> >> 0.67% [kernel] [k] rb_erase >> >> >> *0.64% traffic_server [.] read_from_net(NetHandler*, >> UnixNetVConnection*, EThread*) * >> >> >> >> On Thu, Oct 23, 2014 at 12:33 PM, Ergin Ozekes <ergin.oze...@gmail.com> >> wrote: >> >>> Hi all; >>> >>> My current working network bandwidth is 350 Mbps. >>> My netstat -st output prints out below value increasingly. >>> >>> 1555903887 packets collapsed in receive queue due to low socket buffer >>> 1555908175 packets collapsed in receive queue due to low socket buffer >>> 1555912925 packets collapsed in receive queue due to low socket buffer >>> 1555920054 packets collapsed in receive queue due to low socket buffer >>> 1555929162 packets collapsed in receive queue due to low socket buffer >>> 1555938162 packets collapsed in receive queue due to low socket buffer >>> 1555945682 packets collapsed in receive queue due to low socket buffer >>> 1555951783 packets collapsed in receive queue due to low socket buffer >>> 1555959318 packets collapsed in receive queue due to low socket buffer >>> 1555962474 packets collapsed in receive queue due to low socket buffer >>> 1555969574 packets collapsed in receive queue due to low socket buffer >>> >>> I've increased to socket buffer size, backlog queue size and memlock >>> value was set to unlimited for ats user. How can I fix it completely? >>> BTW, How can I increase read from socket(connection) performance of the >>> ATS? >>> >>> >>> Best Regards, >>> Ergin Ozekes >>> >>> On Thu, Oct 23, 2014 at 12:16 PM, Ergin Ozekes <ergin.oze...@gmail.com> >>> wrote: >>> >>>> Hi all; >>>> >>>> My current working network bandwidth is 350 Mbps. >>>> My netstat -st output prints out below value increasingly. >>>> >>>> 1555903887 packets collapsed in receive queue due to low socket >>>> buffer >>>> 1555908175 packets collapsed in receive queue due to low socket >>>> buffer >>>> 1555912925 packets collapsed in receive queue due to low socket >>>> buffer >>>> 1555920054 packets collapsed in receive queue due to low socket >>>> buffer >>>> 1555929162 packets collapsed in receive queue due to low socket >>>> buffer >>>> 1555938162 packets collapsed in receive queue due to low socket >>>> buffer >>>> 1555945682 packets collapsed in receive queue due to low socket >>>> buffer >>>> 1555951783 packets collapsed in receive queue due to low socket >>>> buffer >>>> 1555959318 packets collapsed in receive queue due to low socket >>>> buffer >>>> 1555962474 packets collapsed in receive queue due to low socket >>>> buffer >>>> 1555969574 packets collapsed in receive queue due to low socket >>>> buffer >>>> >>>> I've increased to socket buffer size, backlog queue size and memlock >>>> value was set to unlimited for ats user. How can I fix it completely? >>>> >>>> Best Regards, >>>> Ergin Ozekes >>>> >>> >>> >>