doujindesu.fun
robots.txt

Robots Exclusion Standard data for doujindesu.fun

Resource Scan

Scan Details

Site Domain doujindesu.fun
Base Domain doujindesu.fun
Scan Status Failed
Failure StageFetching resource.
Failure ReasonCouldn't connect to server.
Last Scan2025-03-06T09:54:12+00:00
Next Scan 2025-06-04T09:54:12+00:00

Last Successful Scan

Scanned2023-07-23T08:14:05+00:00
URL http://doujindesu.fun/robots.txt
Redirect https://212.32.226.234/robots.txt
Redirect Domain 212.32.226.234
Domain IPs 162.255.119.30
Redirect IPs 212.32.226.234
Response IP 212.32.226.234
Found Yes
Hash 61be6e0bcea32a4f62e936671cdf3fc2400899147af5391ae1d6dc79bed61ba4
SimHash bc109d49c574

Groups

*

Rule Path
Disallow /includes/
Disallow /themes/
Disallow /search/
Disallow /memeen/
Allow /themes/front/*/css/
Allow /themes/front/*/images/
Allow /themes/front/*/js/
Allow /themes/front/*/fonts/
Allow /themes/front/*/images/*.jpg
Allow /themes/front/*/images/*.png
Allow /themes/front/*/images/*.gif

Other Records

Field Value
sitemap https://agc.gudangscript.com/sitemap.xml

Comments

  • robots.txt
  • This file is to prevent the crawling and indexing of certain parts
  • of your site by web crawlers and spiders run by sites like Yahoo!
  • and Google. By telling these "robots" where not to go on your site,
  • you save bandwidth and server resources.
  • This file will be ignored unless it is at the root of your host:
  • Used: http://example.com/robots.txt
  • Ignored: http://example.com/site/robots.txt
  • For more information about the robots.txt standard, see:
  • http://www.robotstxt.org/wc/robots.html
  • For syntax checking, see:
  • http://www.sxw.org.uk/computing/robots/check.html
  • Disallow directories
  • Disallow paths
  • Allow themes
  • Allow content images