Hi Nigel

In such a case I would advise to create a sitemap - unfortunately this Koha feature seems not so well documented, but the following may give you a start:

* https://lists.katipo.co.nz/public/koha/2020-November/055401.html

* https://wiki.koha-community.org/wiki/Commands_provided_by_the_Debian_packages#koha-sitemap

* https://koha-community.org/manual/24.05/en/html/cron_jobs.html#sitemap

Best wishes: Michael
--
Geschäftsführer · Diplombibliothekar BBS, Informatiker eidg. Fachausweis
Admin Kuhn GmbH · Pappelstrasse 20 · 4123 Allschwil · Schweiz
T 0041 (0)61 261 55 61 · E m...@adminkuhn.ch · W www.adminkuhn.ch



Am 25.07.24 um 13:27 schrieb Nigel Titley:
Is anyone else getting problems with the facebook web crawler hammering their OPAC search function?

This has been happening on and off for a couple of months but set in with a vengeance a couple of days ago. The crawler is hitting us with many OPAC search queries, beyond the capacity of our system to respond.

robots.txt is being ignored

I started by blocking facebook's entire IPv6 range as the queries were all coming in over IPv6. They responded by switching to IPv4 and because they have a number of blocks it wasn't practical to block each and every one of them.

I've temporarily switched off OPAC entirely and the system has returned to normal and I can at least perform intranet functions but this is obviously non-ideal.

Does anyone have any thoughts on this?

I'm running 22.05.13.000 on Ubuntu.

Thanks

Nigel
_______________________________________________

Koha mailing list  http://koha-community.org
Koha@lists.katipo.co.nz
Unsubscribe: https://lists.katipo.co.nz/mailman/listinfo/koha


_______________________________________________

Koha mailing list  http://koha-community.org
Koha@lists.katipo.co.nz
Unsubscribe: https://lists.katipo.co.nz/mailman/listinfo/koha

Reply via email to