On 17/02/2016 at 01:39, Jason Morris wrote:
> What would be the proper config of robots.txt to allow only Google to crawl a
> JSPWiki installation and nothing else?
It's all in Google's robots.txt documentation
https://developers.google.com/webmasters/control-crawl-index/docs/robots_txt
However
We developed a plugin which uses the excellent (but commercial) tool
Prince (https://princexml.com) and is made up of a jar file plus a
little bit of a Perl hack. This enables us to produce the cleanest, most
beautiful documents that we can and lets us run varied templates and our
own custom CSS. T
;t hard to write and should be done in 20 minutes or so even with
multiple guesses/web searches for the sed parameters... Does rather
require that this is running on a *nix based system and not Windows...
Shout if you want some more pointers.
Roland Whitehead