On 2/21/2014 7:25 AM, rge3 wrote:
I havne't found any ideas for this and thought I might ask here.  We have a
fairly straightforward proxy_cache setup with a proxy_pass backend.  We
cache documents for different lengths of time or go the backend for what's
missing.  My problem is we're getting overrun with bot and spider requests.
MSN in particular started hitting us exceptionally hard yesterday and
started bringing our backend servers down.  Because they're crawling the
site from end to end our cache is missing a lot of those pages and nginx has
to pass the request on through.

Are they ignoring your robots.txt?

_______________________________________________
nginx mailing list
nginx@nginx.org
http://mailman.nginx.org/mailman/listinfo/nginx

Reply via email to