scubadivingintro.com
robots.txt

Robots Exclusion Standard data for scubadivingintro.com

Resource Scan

Scan Details

Site Domain scubadivingintro.com
Base Domain scubadivingintro.com
Scan Status Failed
Failure StageFetching resource.
Failure ReasonServer returned a client error.
Last Scan2024-10-18T18:17:21+00:00
Next Scan 2025-01-16T18:17:21+00:00

Last Successful Scan

Scanned2023-07-04T11:21:45+00:00
URL http://scubadivingintro.com/robots.txt
Redirect https://www.scubadiving.com/robots.txt
Redirect Domain www.scubadiving.com
Redirect Base scubadiving.com
Domain IPs 52.25.184.171, 52.35.201.153
Redirect IPs 52.25.184.171, 52.35.201.153
Response IP 52.35.201.153
Found Yes
Hash 76c44a9928a26dae419a8e3e9676b134a439e2f28cbe4da19ac38b7a0219895d
SimHash 00009f1ac7e1

Groups

*

Rule Path
Allow /*.js
Allow /*.css
Allow /*.jpg
Allow /*.gif
Allow /*.png
Allow /sites/scubadiving.com/
Disallow /includes/
Disallow /misc/
Disallow /modules/
Disallow /profiles/
Disallow /scripts/
Disallow /sites/
Disallow /files/
Disallow /file/
Disallow /themes/
Disallow /rss/all
Disallow /comment/reply
Disallow /statistics
Disallow /customerror/404
Disallow /forward
Disallow /users
Disallow /user
Disallow /find
Disallow /privacy-policy
Disallow /terms-service
Disallow /contact-us
Disallow /abuse
Disallow /newsletter

Other Records

Field Value
crawl-delay 10

twitterbot

Rule Path
Allow *

gigabot

Rule Path
Disallow /

scrubby

Rule Path
Disallow /

nutch

Rule Path
Disallow /

baiduspider

Rule Path
Disallow /

naverbot

Rule Path
Disallow /

yeti

Rule Path
Disallow /

asterias

Rule Path
Disallow /

*

Rule Path
Disallow

Other Records

Field Value
sitemap https://www.scubadiving.com/sitemap.xml

Comments

  • The use of robots or other automated means to access any Bonnier Corporation website
  • without the express permission of Bonnier Corporation is strictly prohibited.
  • Bonnier Corporation may permit automated access to
  • access certain pages but solely for the limited purpose of
  • including content in publicly available search engines. Any other
  • use of robots or failure to obey the robots exclusion standards set
  • forth at http://www.robotstxt.org/wc/exclusion.html is strictly
  • prohibited.
  • Disallow the following spiders