Added ai.robots.txt rules for traffic filtering
This commit is contained in:
@@ -119,6 +119,11 @@
|
||||
"name": "InternetMeasurement",
|
||||
"user_agent_regex": "InternetMeasurement",
|
||||
"action": "DENY"
|
||||
},
|
||||
{
|
||||
"name": "OtherAI (ai.robots.txt)",
|
||||
"user_agent_regex": "(AI2Bot|Ai2Bot\\-Dolma|Amazonbot|anthropic\\-ai|Applebot|Applebot\\-Extended|Brightbot\\ 1\\.0|Bytespider|CCBot|ChatGPT\\-User|Claude\\-Web|ClaudeBot|cohere\\-ai|cohere\\-training\\-data\\-crawler|Crawlspace|Diffbot|DuckAssistBot|FacebookBot|FriendlyCrawler|Google\\-Extended|GoogleOther|GoogleOther\\-Image|GoogleOther\\-Video|GPTBot|iaskspider/2\\.0|ICC\\-Crawler|ImagesiftBot|img2dataset|ISSCyberRiskCrawler|Kangaroo\\ Bot|Meta\\-ExternalAgent|Meta\\-ExternalFetcher|OAI\\-SearchBot|omgili|omgilibot|PanguBot|PerplexityBot|PetalBot|Scrapy|SemrushBot\\-OCOB|SemrushBot\\-SWA|Sidetrade\\ indexer\\ bot|Timpibot|VelenPublicWebCrawler|Webzio\\-Extended|YouBot)",
|
||||
"action": "DENY"
|
||||
}
|
||||
]
|
||||
}
|
Reference in New Issue
Block a user