Kris Deugau put forth on 9/29/2010 2:33 PM: > Hmm, no, less than 100M: > > PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND > 28776 rbldns 20 0 81740 65m 700 S 0 3.3 118:49.42 rbldnsd
I was going by information I received from another list. I don't use the data feed service. Does this include the CBL data set within Zen? I would make an educated guess that the size of the CBL data set would be over 100MB alone. 25 million 32bit IP addresses (4 bytes) would be 100MB, if my math is correct. 25 million bot infected hosts around the world seems like a very conservative estimate. > And this with a modest local blacklist loaded in as well. The on-disk > files for all of the lists total just over 100M. We just run the > Spamhaus data on a non-public zone on our general resolvers (running > dnscache) and we have yet to see any latency problems. With fast resolvers and local GigE, performance should be fine for many sites as you state. It's also easier to manage than running rbldnsd on each MX as you have a single update point. I know of one site, coincidentally also in Canada, running two MX hosts. Each receives, IIRC, on an average day, ~50 million connection attempts, 100 million total. This is nowhere near the numbers of a good sized ISP obviously and tiny compared to a gorilla such as Gmail. The OP runs rbldnsd on each MX, with the full Spamhaus zones minus the CBL. Also incorporated into the rbldnsd instances are extensive local block lists, the Enemies List, the CBL data, and some other mirrored dnsbl data. This may be the multi gigabyte setup I was thinking of, which isn't just Spamhaus zones. Interestingly, this site doesn't reject any spam due to any hits against any list. After DATA, a 55x is returned to the client, but the entire message is saved for further anti spam heuristics processing. It's one of the most elaborate setups I've heard of. Then again, some of the most elaborate setups _no one_ will probably hear about. ;) > The biggest sysadmin/network costs for the rsync service are in > configuration (may need extra scripting to distribute the data to > multiple rbldnsd instances, depending on how you want to arrange your > DNS services - otherwise, it's "set up once, let it run") and update > bandwidth - currently they provide a script intended to be called once a > minute to update the zone data source files. Yeah, running the Spamhaus zones on local rbldnsd instances on each MX would require some distribution magic, as you state. Never done this myself. I'd be more inclined to go the route you've taken, if I were ever in a position to manage such a thing. -- Stan