apotheken-website-vorschau.de
robots.txt

Robots Exclusion Standard data for apotheken-website-vorschau.de

Resource Scan

Scan Details

Site Domain apotheken-website-vorschau.de
Base Domain apotheken-website-vorschau.de
Scan Status Ok
Last Scan2024-06-15T18:23:01+00:00
Next Scan 2024-07-15T18:23:01+00:00

Last Scan

Scanned2024-06-15T18:23:01+00:00
URL http://apotheken-website-vorschau.de/robots.txt
Redirect https://apotheken.de/robots.txt
Redirect Domain apotheken.de
Redirect Base apotheken.de
Domain IPs 62.116.130.8
Redirect IPs 13.226.2.24, 13.226.2.39, 13.226.2.5, 13.226.2.57
Response IP 18.165.171.30
Found Yes
Hash b59553ef210ece0afcdafcde458779f246d44c6942fce715d4a8f4773b81ba0f
SimHash 431f5282cba7

Groups

*

Rule Path
Disallow /intern/
Disallow /include/functions/
Disallow /servicescripts/
Disallow /include/classes/
Disallow /downloads/
Disallow /graphics/
Disallow /highslide/
Disallow /homepages/
Disallow /codelib/
Disallow /backup/
Disallow /typo3conf/
Disallow /typo3temp/
Disallow /typo3/
Disallow /t3lib/
Disallow /thirdparty.php
Disallow /index.php

Other Records

Field Value
crawl-delay 60

blexbot/1.0
aqua_products
ahrefs.com
backdoorbot/1.0
baiduspider
blowfish/1.0
bookmark search tool
botalot
builtbottough
bullseye/1.0
bunnyslippers
cheesebot
cherrypicker
cherrypickerelite/1.0
cherrypickerse/1.0
cligoorobot
copyrightcheck
crescent
crescent internet toolpak http ole control v.1.0
dittospyder
emailcollector
emailsiphon
emailwolf
erocrawler
extractorpro
fairad client
flaming attackbot
foobot
gaisbot
getright/4.2
harvest/1.5
infonavirobot
iron33/1.0.2
jennybot
kenjin spider
keyword density/0.9
lnspiderguy
lexibot
linkscan/8.1a unix
linkwalker
linkextractorpro
miixpc
miixpc/4.2
mj12bot
mlbot
msiecrawler
mata hari
microsoft url control
microsoft url control - 5.01.4511
microsoft url control - 6.00.8169
mister pix
mozilla/4.0 (compatible; bullseye; windows 95)
nicerspro
netants
netmechanic
offline explorer
openbot
openfind
openfind data gathere
oracle ultra search
perman
pixray
pixray-seeker
pixray-seeker/1.1
pixray-seeker/1.1 (pixray-seeker; crawler@pixray.com)
propowerbot/2.14
prowebwalker
python-urllib
queryn metasearch
rma
radiation retriever 1.1
repomonkey
repomonkey bait & tackle/v1.01
seokicks-robot
scoutjet
sistrix
sitesnagger
spankbot
speedy
surveybot
szukacz/1.4
tagoobot
teleport
teleportpro
telesoft
the intraformant
thenomad
true_robot
true_robot/1.0
url control
url_spider_pro
urly warning
vci
vci webviewer vci webviewer win32
www-collector-e
web image collector
webauto
webbandit
webbandit/3.50
webcopier
webenhancer
websauger
webstripper
webzip
webzip/4.0
webmasterworldforumbot
website quester
webster pro
wget
wget/1.5.3
wget/1.6
xenu's
xenu's link sleuth 1.1c
yandex
youdaobot
zeus
zeus 32297 webster pro v2.9 win32
zeus link scout
asterias
b2w/0.1
cosmos
dotbot
grub
grub-client
hloader
httplib
humanlinks
ia_archiver
ia_archiver/1.6
kalooga
larbin
libweb/clshttp
looksmart
lwp-trivial
lwp-trivial/1.34
moget
moget/2.1
netestate ne crawler
pixray
pixray-seeker
plukkie
psbot
searchpreview
sogou
spanner
suzuran
tocrawl/urldispatcher
turingos
twiceler

Rule Path
Disallow /

Other Records

Field Value
sitemap https://www.apotheken.de/sitemap.xml
sitemap https://www.apotheken.de/sitemap-yellowpages.xml

Comments

  • robots.txt
  • http://www.apotheken.de
  • 2012-06-11
  • ************************************************************************************
  • ==============================
  • Bot-Controller: Spider-Verbote
  • ==============================
  • Gesperrte Verzeichnisse für ALLE Spider/Bots
  • ===========================
  • Bot-Detektiv: Spambot-Falle
  • ===========================
  • Allen Robots ist das Spidern des folgenden Ordners verboten:
  • Disallow: /Testordner
  • Hinweise
  • a) Der fiktive Ordner "/Testordner" dient als Falle für Spambots - diese ignorieren
  • die Anweisung "Disallow: /Testordner" und hinterlassen einen leicht zu findenden
  • entsprechenden Eintrag in den Logfiles (> ggf Ausschluß per .htaccess)
  • Nur alle x Sekunden darf eine neue Seite zum Auslesen aufgerufen werden. Gilt für ALLE
  • Folgenden Bots ist das Spidern der gesamten Webseite verboten:

Warnings

  • 2 invalid lines.