sigh. remove this bullshit and start over.
* Steve Johnson <maill...@sjohnson.info> [2011-02-01 22:38]: > Ok, thanks for the tips. I did not have any ifq drops, but have still just > increased the net.inet.icmp.errppslimit to 10000 (from the 1000 that was > before and shown below) and will see if that helps anything. Thanks also for > the clarification on the match counter. > > I had forgotten to also include the sysctl changes that I had made as well, > mostly based from calomel.org, which were the following: > > kern.maxclusters=128000 > net.inet.icmp.errppslimit=1000 > net.inet.ip.ifq.maxlen=1536 > net.inet.ip.mtudisc=0 > net.inet.ip.ttl=254 > net.inet.ipcomp.enable=1 > net.inet.tcp.ackonpush=1 > net.inet.tcp.ecn=1 > net.inet.tcp.mssdflt=1472 > net.inet.tcp.recvspace=262144 > net.inet.tcp.rfc1323=1 > net.inet.tcp.rfc3390=1 > net.inet.tcp.sack=1 > net.inet.tcp.sendspace=262144 > net.inet.udp.recvspace=262144 > net.inet.udp.sendspace=262144 > vm.swapencrypt.enable=1 > > On Tue, Feb 1, 2011 at 3:15 PM, Henning Brauer <lists-open...@bsws.de>wrote: > > > * Steve Johnson <maill...@sjohnson.info> [2011-02-01 20:35]: > > > I currently have a system that has no match rule in the ruleset, but that > > > uses tables for a big chunk of the traffic, including our monitoring > > station > > > that has a pretty high SNMP request rate. That system has a state table > > that > > > usually stabilizes between 15-20K sessions, with a session search rate of > > > around 10K. The states limit has been raised to 100000 and the frags to > > > 10000, but all other limits are set to default values. > > > > you can increase that much more. the times where kmem was a very > > scarce ressource are long over. > > > > > However, the "match" > > > counter always states a rate between 199/200 per second. > > > > the counter has nothing to do with match rules. it is increased any > > time a rule matches, regardless of the type. > > > > > During some heavy > > > traffic period, we are getting some failures from the monitoring system > > and > > > the only thing that seems possibly out of health for the system is the > > match > > > counter rate. System processor and memory are fine and there is no other > > > noticeable impact, but clearly the monitoring tool is seeing an impact, > > as > > > it didn't reflect something this behavior before we implemented the PF > > > systems. > > > > you might hit some other limit, not necessarily pf. start with > > checking sysctl net.inet.ifq - in particular drops, and increase > > maxlen if you see it increasing. > > depending on how you monitor you might also run into the icmp err rate > > limit, play with the net.inet.icmp.errppslimit sysctl. > > > > -- > > Henning Brauer, h...@bsws.de, henn...@openbsd.org > > BS Web Services, http://bsws.de > > Full-Service ISP - Secure Hosting, Mail and DNS Services > > Dedicated Servers, Rootservers, Application Hosting > -- Henning Brauer, h...@bsws.de, henn...@openbsd.org BS Web Services, http://bsws.de Full-Service ISP - Secure Hosting, Mail and DNS Services Dedicated Servers, Rootservers, Application Hosting