webcilo.com
robots.txt

Robots Exclusion Standard data for webcilo.com

Resource Scan

Scan Details

Site Domain webcilo.com
Base Domain webcilo.com
Scan Status Failed
Failure StageFetching resource.
Failure ReasonServer returned a server error.
Last Scan2024-06-18T09:01:34+00:00
Next Scan 2024-07-02T09:01:34+00:00

Last Successful Scan

Scanned2024-06-03T08:59:42+00:00
URL https://webcilo.com/robots.txt
Domain IPs 104.21.81.198, 172.67.146.63, 2606:4700:3032::ac43:923f, 2606:4700:3037::6815:51c6
Response IP 172.67.146.63
Found Yes
Hash 09939e82b787f028604dfb6150cb935c76ed6418809c6e828dd7fef921f5cc11
SimHash 3c17115a4775

Groups

*

Rule Path
Disallow /ow_version.xml
Disallow /INSTALL.txt
Disallow /LICENSE.txt
Disallow /README.txt
Disallow /UPDATE.txt
Disallow /CHANGELOG.txt
Disallow /admin/

Comments

  • This file contains rules to prevent the crawling and indexing of certain parts
  • of your web site by spiders of a major search engines likes Google and Yahoo.
  • By managing these rules you can allow or disallow access to specific folders
  • and files for such spyders.
  • The good way to hide private data or save a lot of bandwidth.
  • For more information about the robots.txt standard, see:
  • http://www.robotstxt.org/wc/robots.html
  • For syntax checking, see:
  • http://www.sxw.org.uk/computing/robots/check.html