Sounds like at this point this discussion needs to be moved off the nginx mailing list.
> On Jan 31, 2016, at 10:48, shahzaib shahzaib <shahzaib...@gmail.com> wrote: > > The server is using ports 18 and 19 and those port are configured with speed > 1000 > > > > LH26876_SW2#sh run int g 0/18 > > ! > > interface GigabitEthernet 0/18 > > description LH28765_3 > > no ip address > > speed 1000 > > ! > > port-channel-protocol LACP > > port-channel 3 mode active > > no shutdown > > LH26876_SW2#sh run int g 0/19 > > ! > > interface GigabitEthernet 0/19 > > description LH28765_3 > > no ip address > > speed 1000 > > ! > > port-channel-protocol LACP > > port-channel 3 mode active > > no shutdown > > LH26876_SW2# > > > > ------------------------------ > > Is it alright ? > > Regards. > > Shahzaib > > >> On Sun, Jan 31, 2016 at 11:18 PM, shahzaib shahzaib <shahzaib...@gmail.com> >> wrote: >> Hi, >> >> Thanks a lot for response. Now i am doubting that issue is on network layer >> as i can examine lots of retransmitted packets in netstat -s output. Here is >> the server's status : >> >> http://prntscr.com/9xa6z2 >> >> Following is the thread with same mentioned issue : >> >> http://serverfault.com/questions/218101/freebsd-8-1-unstable-network-connection >> >> This is what he said in thread : >> >> "I ran into a problem with Cisco Switchs forcing Negotiation of network >> speeds. This caused intermittent errors and retransmissions. The result was >> file transfers being really slow. May not be the cases, but you can turn of >> speed negotiation with miitools (if I recall correctly, been a long time)." >> >> >>Can you replicate using ftp, scp? >> Yes, we recently tried downloading file over FTP and encountered the same >> slow transfer rate. >> >> >>What's the output of zpool iostat (and the overal zpool/zfs >> >>configuration)?Also do you have ZFS on top of hardware raid ? In general >> >>just 12 SATA disks won't have a lot of IOps (especially random read) >> >>unless it all hits ZFS Arc (can/should be monitored), even more if there >> >>is a hardware raid underneath (in your place would flash the HBA with IT >> >>firmware so you get plain jbods managed by ZFS). >> >> zpool iostat is quite stable yet. We're using HBA LSI-9211 , so its not >> hardware controller as FreeBSD recommends to use HBA in order to directly >> access all drives for scrubbing and data-integrity purposes. Do you >> recommend Hardware-Raid ? Following is the scrnshot of ARC status : >> >> http://prntscr.com/9xaf9p >> >> >>How is your switch configured? How are the links negotiated, make sure >> >>both sides of both links are full duplex 1gig. Look for crc or input >> >>errors on the interface side. >> On My side, i can see that both interfaces have Fully-Duplex port. Regarding >> crc / input errors, is there any command i can use to check that on FreeBSD ? >> >> Regards. >> Shahzaib >> >> >>> On Sun, Jan 31, 2016 at 11:04 PM, Payam Chychi <pchy...@gmail.com> wrote: >>> Hi, >>> >>> Forget the application layer being the problem until you have successfully >>> replicated the problem in several different setups. >>> >>> Are you monitoring both links utilization levels? Really sounds like a >>> network layer problem or something with your ip stack. >>> >>> Can you replicate using ftp, scp? >>> >>> How is your switch configured? How are the links negotiated, make sure both >>> sides of both links are full duplex 1gig. Look for crc or input errors on >>> the interface side. >>> >>> How many packets are you pushing? Make sure the switch isnt activating >>> unicast limiting. >>> >>> Lots of things to check... Would help if you can help us understand what >>> tests youve done to determine its nginx. >>> >>> Thanks >>> >>> -- >>> Payam Chychi >>> Network Engineer / Security Specialist >>> >>>> On Sunday, January 31, 2016 at 9:33 AM, Reinis Rozitis wrote: >>>> >>>> This is a bit out of scope of nginx but .. >>>> >>>>> could be network issue or LACP issue but doesn't looks like it is >>>> >>>> How did you determine this? >>>> Can you generate more than 1 Gbps (without nginx)? >>>> >>>> >>>>> 12 x 3TB SATA Raid-10 (HBA LSI-9211) >>>>> ZFS FileSystem with 18TB usable space >>>> >>>>> Please i need guidance to handle with this problem, i am sure that some >>>>> value needs to tweak. >>>> >>>> What's the output of zpool iostat (and the overal zpool/zfs configuration)? >>>> >>>> Also do you have ZFS on top of hardware raid ? >>>> >>>> In general just 12 SATA disks won't have a lot of IOps (especially random >>>> read) unless it all hits ZFS Arc (can/should be monitored), even more if >>>> there is a hardware raid underneath (in your place would flash the HBA with >>>> IT firmware so you get plain jbods managed by ZFS). >>>> >>>> rr >>>> >>>> _______________________________________________ >>>> nginx mailing list >>>> nginx@nginx.org >>>> http://mailman.nginx.org/mailman/listinfo/nginx >>> >>> >>> _______________________________________________ >>> nginx mailing list >>> nginx@nginx.org >>> http://mailman.nginx.org/mailman/listinfo/nginx > > _______________________________________________ > nginx mailing list > nginx@nginx.org > http://mailman.nginx.org/mailman/listinfo/nginx
_______________________________________________ nginx mailing list nginx@nginx.org http://mailman.nginx.org/mailman/listinfo/nginx