pango.gnome.org
robots.txt

Robots Exclusion Standard data for pango.gnome.org

Resource Scan

Scan Details

Site Domain pango.gnome.org
Base Domain gnome.org
Scan Status Failed
Failure StageFetching resource.
Failure ReasonCouldn't connect to server.
Last Scan2025-07-09T11:44:58+00:00
Next Scan 2025-10-07T11:44:58+00:00

Last Successful Scan

Scanned2024-11-19T11:42:51+00:00
URL https://pango.gnome.org/robots.txt
Redirect https://static.gnome.org/wiki.gnome.org/robots.txt
Redirect Domain static.gnome.org
Redirect Base gnome.org
Domain IPs 151.101.1.91, 151.101.129.91, 151.101.193.91, 151.101.65.91, 2a04:4e42:200::347, 2a04:4e42:400::347, 2a04:4e42:600::347, 2a04:4e42::347
Redirect IPs 2a02:6ea0:d100::31, 2a02:6ea0:d100::32, 2a02:6ea0:d100::33, 2a02:6ea0:d100::34, 2a02:6ea0:d100::35, 2a02:6ea0:d100::37, 2a02:6ea0:d100::50, 2a02:6ea0:d10c::9, 79.127.213.245, 79.127.235.11, 79.127.235.3, 79.127.235.45, 79.127.235.51, 79.127.235.6, 79.127.235.9, 89.187.162.12
Response IP 207.211.208.183
Found Yes
Hash 2fdc893936595906aa6090f59c015dd563cdf88c09296af7502b91dd46d50de4
SimHash 2c7a5f024f35

Groups

dotbot

Rule Path
Disallow /

*

Rule Path
Disallow /action/

Other Records

Field Value
crawl-delay 20

Comments

  • if you want to add own robot rules, do it BEFORE the final rule matching *
  • This has to match script url + cfg.url_prefix_action - it
  • saves lots of search engine load and traffic by disallowing crawlers
  • to request action related URLs.
  • NOTE - in order to make this have any effect, you have to set
  • url_prefix_action to "action", cf. HelpOnConfiguration