Hi.

I thought to contribute to the https://github.com/ai-robots-txt/ai.robots.txt repo and would like to here your opinion to the Question.

Should the AI Crawler be
[ ] tarpitted or get directly a
[ ] deny?

Both lines are added in the example config to see that the config is almost the same on HAProxy site.

https://github.com/git001/ai.robots.txt/blob/7169417be76d8f6e8ca69593f626ca24814cf3a2/haproxy-ai-crawler-block.config#L36-L37

I know that HAProxy Enterprise have a great bot-management Solution https://www.haproxy.com/solutions/bot-management

which is mentioned in these Blog Posts

https://www.haproxy.com/blog/how-to-reliably-block-ai-crawlers-using-haproxy-enterprise
https://www.haproxy.com/blog/nearly-90-of-our-ai-crawler-traffic-is-from-tiktok-parent-bytedance-lessons-learned

but maybe the ai.robots.txt can be a poor person solution wich can also be used for the Haproxy Ingress Controller?

Any Answer is highly appreciated.

Regards
Alex


Reply via email to