oregonexplorer.info
robots.txt

Robots Exclusion Standard data for oregonexplorer.info

Resource Scan

Scan Details

Site Domain oregonexplorer.info
Base Domain oregonexplorer.info
Scan Status Ok
Last Scan2024-09-21T14:00:15+00:00
Next Scan 2024-10-21T14:00:15+00:00

Last Scan

Scanned2024-09-21T14:00:15+00:00
URL https://oregonexplorer.info/robots.txt
Domain IPs 128.193.164.100, 128.193.164.101
Response IP 128.193.164.100
Found Yes
Hash 3b6e059a52a8aa3f500faf8815015547dc8406b2ca930c71e289a0e90cfc8dfb
SimHash 78945511ce78

Groups

*

Rule Path
Allow /misc/*.css$
Allow /misc/*.css?
Allow /misc/*.js$
Allow /misc/*.js?
Allow /misc/*.gif
Allow /misc/*.jpg
Allow /misc/*.jpeg
Allow /misc/*.png
Allow /modules/*.css$
Allow /modules/*.css?
Allow /modules/*.js$
Allow /modules/*.js?
Allow /modules/*.gif
Allow /modules/*.jpg
Allow /modules/*.jpeg
Allow /modules/*.png
Allow /profiles/*.css$
Allow /profiles/*.css?
Allow /profiles/*.js$
Allow /profiles/*.js?
Allow /profiles/*.gif
Allow /profiles/*.jpg
Allow /profiles/*.jpeg
Allow /profiles/*.png
Allow /themes/*.css$
Allow /themes/*.css?
Allow /themes/*.js$
Allow /themes/*.js?
Allow /themes/*.gif
Allow /themes/*.jpg
Allow /themes/*.jpeg
Allow /themes/*.png
Disallow /includes/
Disallow /misc/
Disallow /modules/
Disallow /profiles/
Disallow /scripts/
Disallow /themes/
Disallow /CHANGELOG.txt
Disallow /cron.php
Disallow /INSTALL.mysql.txt
Disallow /INSTALL.pgsql.txt
Disallow /INSTALL.sqlite.txt
Disallow /install.php
Disallow /INSTALL.txt
Disallow /LICENSE.txt
Disallow /MAINTAINERS.txt
Disallow /update.php
Disallow /UPGRADE.txt
Disallow /xmlrpc.php
Disallow /admin/
Disallow /comment/reply/
Disallow /filter/tips/
Disallow /node/add/
Disallow /search/
Disallow /user/register/
Disallow /user/password/
Disallow /user/login/
Disallow /user/logout/
Disallow /?q=admin%2F
Disallow /?q=comment%2Freply%2F
Disallow /?q=filter%2Ftips%2F
Disallow /?q=node%2Fadd%2F
Disallow /?q=search%2F
Disallow /?q=user%2Fpassword%2F
Disallow /?q=user%2Fregister%2F
Disallow /?q=user%2Flogin%2F
Disallow /?q=user%2Flogout%2F

Other Records

Field Value
crawl-delay 10

gptbot

Rule Path
Disallow /

duckduckbot

Rule Path
Disallow /

facebookexternalhit

Rule Path
Disallow /

fluid

Rule Path
Disallow /

claudebot

Rule Path
Disallow /

dotbot

Rule Path
Disallow /

petalbot

Rule Path
Disallow /

semrushbot

Rule Path
Disallow /

yandexbot

Rule Path
Disallow /

bytespider

Rule Path
Disallow /

amazonbot

Rule Path
Disallow /

rb.gy/oupwis

Rule Path
Disallow /

linkedinbot

Rule Path
Disallow /

magpie-crawler

Rule Path
Disallow /

chatgpt-user

Rule Path
Disallow /

iaskbot

Rule Path
Disallow /

dataforseobot

Rule Path
Disallow /

dubbotbot

Rule Path
Disallow /

imagesiftbot

Rule Path
Disallow /

mail.ru_bot

Rule Path
Disallow /

mj12bot

Rule Path
Disallow /

mojeekbot

Rule Path
Disallow /

monsidobot

Rule Path
Disallow /

seznambot

Rule Path
Disallow /

startmebot

Rule Path
Disallow /

yandexbot

Rule Path
Disallow /

yandexfavicons

Rule Path
Disallow /

yandeximages

Rule Path
Disallow /

yeti

Rule Path
Disallow /

trendictionbot

Rule Path
Disallow /

owler

Rule Path
Disallow /

pubmatic

Rule Path
Disallow /

scrapy

Rule Path
Disallow /

screaming frog seo spider

Rule Path
Disallow /

seobilitybot

Rule Path
Disallow /

serendeputybot

Rule Path
Disallow /

snap url preview service

Rule Path
Disallow /

sogou web spider

Rule Path
Disallow /

turnitin

Rule Path
Disallow /

twitterbot

Rule Path
Disallow /

wp.com

Rule Path
Disallow /

yacybot

Rule Path
Disallow /

yisouspider

Rule Path
Disallow /

zoominfobot

Rule Path
Disallow /

Comments

  • robots.txt
  • This file is to prevent the crawling and indexing of certain parts
  • of your site by web crawlers and spiders run by sites like Yahoo!
  • and Google. By telling these "robots" where not to go on your site,
  • you save bandwidth and server resources.
  • This file will be ignored unless it is at the root of your host:
  • Used: http://example.com/robots.txt
  • Ignored: http://example.com/site/robots.txt
  • For more information about the robots.txt standard, see:
  • http://www.robotstxt.org/robotstxt.html
  • CSS, JS, Images
  • Directories
  • Files
  • Paths (clean URLs)
  • Paths (no clean URLs)
  • Ignore bots

Warnings

  • 2 invalid lines.