librecmc.org
robots.txt

Robots Exclusion Standard data for librecmc.org

Resource Scan

Scan Details

Site Domain librecmc.org
Base Domain librecmc.org
Scan Status Ok
Last Scan2025-10-07T03:55:28+00:00
Next Scan 2025-11-06T03:55:28+00:00

Last Scan

Scanned2025-10-07T03:55:28+00:00
URL https://librecmc.org/robots.txt
Domain IPs 198.140.141.86
Response IP 198.140.141.86
Found Yes
Hash dfef6ff7e50d32046ae707376ce0c9696b66d43fa09e1825da92c61079fe6206
SimHash 3e178951c6f5

Groups

*

Rule Path
Disallow
Disallow /tarpit

anthropicai
openai
sogou
ahrefsbot
semrushbot
ia_archiver
ai2bot
ai2bot-dolma
amazonbot
anthropic-ai
applebot
applebot-extended
bytespider
ccbot
chatgpt-user
claude-web
claudebot
cohere-ai
cohere-training-data-crawler
crawlspace
diffbot
duckassistbot
facebookbot
friendlycrawler
google-extended
googleother
googleother-image
googleother-video
gptbot
iaskspider/2.0
icc-crawler
imagesiftbot
img2dataset
isscyberriskcrawler
kangaroo bot
meta-externalagent
meta-externalfetcher
oai-searchbot
omgili
omgilibot
pangubot
perplexitybot
petalbot
scrapy
semrushbot-ocob
semrushbot-swa
sidetrade indexer bot
timpibot
velenpublicwebcrawler
webzio-extended
youbot

Rule Path
Disallow /

Comments

  • The following two lines are saying: all user agents are allowed to crawl all of the site
  • Another way of saying the same thing could be:
  • User-agent: *
  • Allow: /
  • You don't really need any of these, if you don't put anything the implicit message is the same: allow all crwalers to the full site
  • The following lines are listing all known (to the date of this file) AI crawlers
  • The last line is saying: all of the above bots are not allowed to crawl any part of the site