communityfeedback.opengov.com
robots.txt

Robots Exclusion Standard data for communityfeedback.opengov.com

Resource Scan

Scan Details

Site Domain communityfeedback.opengov.com
Base Domain opengov.com
Scan Status Ok
Last Scan2024-10-26T20:10:45+00:00
Next Scan 2024-11-25T20:10:45+00:00

Last Scan

Scanned2024-10-26T20:10:45+00:00
URL https://communityfeedback.opengov.com/robots.txt
Domain IPs 104.19.187.105, 104.19.188.105, 2606:4700::6813:bb69, 2606:4700::6813:bc69
Response IP 104.19.188.105
Found Yes
Hash 509b8c6de704213812e616830a257715550ad0cf350e1de5cec9b6cf0b0203be
SimHash 379cad876e48

Groups

*

Rule Path
Disallow /p/1/*
Disallow /portals/1/*
Disallow /p/211/*
Disallow /portals/211/*
Disallow /p/226/*
Disallow /portals/226/*
Disallow /p/249/*
Disallow /portals/249/*
Disallow /p/289/*
Disallow /portals/289/*
Disallow /p/474/*
Disallow /portals/474/*

Other Records

Field Value
crawl-delay 20

ahrefsbot

Rule Path
Disallow /

aidu

Rule Path
Disallow /

alphabot

Rule Path
Disallow /

amazonbot

Rule Path
Disallow /

baiduspider

Rule Path
Disallow /

bingbot

Rule Path
Disallow /
Disallow /

bubing

Rule Path
Disallow /

claudebot

Rule Path
Disallow /

dotbot

Rule Path
Disallow /

jorgee

Rule Path
Disallow /

ltx71 - (http://ltx71.com/)

Rule Path
Disallow /

mauibot

Rule Path
Disallow /

mj12bot

Rule Path
Disallow /

netestate ne crawler

Rule Path
Disallow /

petalbot

Rule Path
Disallow /

phxbot/0.1 (phxbot@protonmail.com)

Rule Path
Disallow /

riddler (http://riddler.io/about)

Rule Path
Disallow /

scrapy

Rule Path
Disallow /

semrushbot

Rule Path
Disallow /

semrushbot-ba

Rule Path
Disallow /

semrushbot-si

Rule Path
Disallow /

semrushbot-bm

Rule Path
Disallow /

semrushbot-sa

Rule Path
Disallow /

semrushbot-swa

Rule Path
Disallow /

semrushbot-ct

Rule Path
Disallow /

semrushbot-seoab

Rule Path
Disallow /

seznambot

Rule Path
Disallow /

yandex

Rule Path
Disallow /

yandexbot

Rule Path
Disallow /

Comments

  • See http://www.robotstxt.org/wc/norobots.html for documentation on how to use the robots.txt file
  • To ban all spiders from the entire site uncomment the next two lines:
  • User-Agent: *
  • Disallow: /

Warnings

  • `user=agent` is not a known field.