trippreserver.com
robots.txt

Robots Exclusion Standard data for trippreserver.com

Resource Scan

Scan Details

Site Domain trippreserver.com
Base Domain trippreserver.com
Scan Status Ok
Last Scan2024-11-06T04:20:22+00:00
Next Scan 2024-12-06T04:20:22+00:00

Last Scan

Scanned2024-11-06T04:20:22+00:00
URL https://trippreserver.com/robots.txt
Redirect https://www.trippreserver.com/robots.txt
Redirect Domain www.trippreserver.com
Redirect Base trippreserver.com
Domain IPs 72.21.91.108
Redirect IPs 152.195.61.170
Response IP 152.195.61.170
Found Yes
Hash fc357b81d85b217afe48e935a1507344b52dd356f1ea20efd9384dea2957202b
SimHash ac1d7de96cd5

Groups

go-http-client/1.1

Rule Path
Disallow /

arquivo-web-crawler
arquivo-web-crawler (compatible; heritrix/1.14.3 +http://arquivo.pt/faq-crawling)

Rule Path
Disallow /

vegebot

Rule Path
Disallow /

panscient.com

Rule Path
Disallow /

mj12bot

Rule Path
Disallow /

easouspider

Rule Path
Disallow /

baiduspider
baiduspider-image
baiduspider-video
baiduspider-news
baiduspider-favo
baiduspider-ads
baiduspider-cpro
baiduspider+
baiduspider+(+http://www.baidu.com/search/spider.htm)
mozilla/5.0 (compatible; baiduspider/2.0; +http://www.baidu.com/search/spider.html)

Rule Path
Disallow /

icarus6j - (contact: phil@icarus6.com)

Rule Path
Disallow /

icarus6j

Rule Path
Disallow /

voltron

Rule Path
Disallow /

wininet test

Rule Path
Disallow /

*

Rule Path
Allow /

ia_archiver

Rule Path
Disallow /

duggmirror

Rule Path
Disallow /

teleport

Rule Path
Disallow /

teleportpro

Rule Path
Disallow /

emailcollector

Rule Path
Disallow /

emailsiphon

Rule Path
Disallow /

webbandit

Rule Path
Disallow /

webzip

Rule Path
Disallow /

webreaper

Rule Path
Disallow /

webstripper

Rule Path
Disallow /

web downloader

Rule Path
Disallow /

webcopier

Rule Path
Disallow /

offline explorer pro

Rule Path
Disallow /

httrack website copier

Rule Path
Disallow /

offline commander

Rule Path
Disallow /

leech

Rule Path
Disallow /

websnake

Rule Path
Disallow /

blackwidow

Rule Path
Disallow /

http weazel

Rule Path
Disallow /

googlebot
*

No rules defined. All paths allowed.

Other Records

Field Value
sitemap https://www.trippreserver.com/sitemap.xml
sitemap https://www.trippreserver.com/sitemap.xml

Comments

  • Baiduspider
  • Google will skip the wildcard entries in full and only take note of what is actually under the user-agent Googlebot. So it's best to try not to use the useragent Googlebot in robots unless you absolutely have to. And if you do then add all the pages/assets that they need to take note of, even if they are duplicated with the wildcard entries.