static: update robots.txt with some additional AI crap

Signed-off-by: Christoph Heiss <christoph@c8h4.io>
This commit is contained in:
Christoph Heiss 2024-02-14 11:34:34 +01:00
parent c03fb0556f
commit ac3e86addb
Signed by: c8h4
GPG key ID: 1538094429952F86

View file

@ -1,5 +1,8 @@
# Based on https://git.sr.ht/~sircmpwn/sr.ht-nginx/tree/master/item/robots.txt
# Primarily based on https://git.sr.ht/~sircmpwn/sr.ht-nginx/tree/master/item/robots.txt
# All credit for collecting to Drew, the sourcehut crew and its contributers!
#
# Also some taken from here, thanks for the idea you AI shills!
# https://github.com/samber/the-great-gpt-firewall
# Too aggressive, marketing/SEO
User-agent: SemrushBot
@ -63,10 +66,22 @@ Disallow: /
User-agent: GPTBot
Disallow: /
# ChatGPT plugins
User-agent: ChatGPT-User
Disallow: /
# Common Crawl, used by e.g. OpenAI .. blargh
User-agent: CCBot
Disallow: /
# Fairly certain that this is an LLM data vacuum
User-agent: ClaudeBot
Disallow: /
# Claude
User-agent: anthropic-ai
Disallow: /
# Same
User-agent: Google-Extended
Disallow: /