xiaozhan.icu
robots.txt

Robots Exclusion Standard data for xiaozhan.icu

Resource Scan

Scan Details

Site Domain xiaozhan.icu
Base Domain xiaozhan.icu
Scan Status Ok
Last Scan2025-09-18T00:35:40+00:00
Next Scan 2025-10-18T00:35:40+00:00

Last Scan

Scanned2025-09-18T00:35:40+00:00
URL https://xiaozhan.icu/robots.txt
Redirect https://archiveofourown.org/robots.txt
Redirect Domain archiveofourown.org
Redirect Base archiveofourown.org
Domain IPs 104.21.69.156, 172.67.209.165, 2606:4700:3034::6815:459c, 2606:4700:3034::ac43:d1a5
Redirect IPs 104.20.8.2, 104.20.9.2, 2606:4700:10::6814:802, 2606:4700:10::6814:902
Response IP 104.20.9.2
Found Yes
Hash 1493972bb388474f4fae4950aca406bd71dd10c17fe4677974c1d8ce989201d3
SimHash ca120d85d1e7

Groups

*

Rule Path Comment
Disallow /works? cruel but efficient
Disallow /autocomplete/ -
Disallow /downloads/ -
Disallow /external_works/ -
Disallow /bookmarks/search? -
Disallow /people/search? -
Disallow /tags/search? -
Disallow /works/search? -

googlebot

Rule Path
Disallow /autocomplete/
Disallow /downloads/
Disallow /external_works/
Disallow /works/*?
Disallow /*search?
Disallow /*?*query=
Disallow /*?*sort_
Disallow /*?*selected_tags
Disallow /*?*view_adult
Disallow /*?*tag_id
Disallow /*?*pseud_id
Disallow /*?*user_id
Disallow /*?*pseud=

ccbot

Rule Path
Disallow /

gptbot

Rule Path
Disallow /

chatgpt-user

Rule Path
Disallow /

slurp

No rules defined. All paths allowed.

Other Records

Field Value
crawl-delay 30

Comments

  • See https://www.robotstxt.org/robotstxt.html for documentation on how to use the robots.txt file
  • disallow indexing of search results
  • Googlebot is smart and knows pattern matching