So this is one of those issues it is most likely a bad configuration but my
robots.txt file is returning a 404 because of another location because I am
disallowing people to access any text files but I do want to allow only the
robots.txt to be accessed.

location /robots.txt {
root 'location/to/robots/txt/file';
}
#This is to stop people digging into any directories looking for files that
only PHP etc should read or serve
location ~* \.(txt|ini|xml|zip|rar)$ {
return 404;
}

I have not tested but I believe I could fix the above by changing my config
to this.
location /robots.txt {
allow all;
root 'location/to/robots/txt/file';
}
#This is to stop people digging into any directories looking for files that
only PHP etc should read or serve
location ~* \.(txt|ini|xml|zip|rar)$ {
deny all;
}

Is that the only possible way to deny access to all text files apart from
the robots.txt ?

Posted at Nginx Forum: 
https://forum.nginx.org/read.php?2,269960,269960#msg-269960

_______________________________________________
nginx mailing list
[email protected]
http://mailman.nginx.org/mailman/listinfo/nginx

Reply via email to