tones.be
robots.txt

Robots Exclusion Standard data for tones.be

Resource Scan

Scan Details

Site Domain tones.be
Base Domain tones.be
Scan Status Failed
Failure ReasonScan timed out.
Last Scan2024-06-10T15:19:55+00:00
Next Scan 2024-08-09T15:19:55+00:00

Last Successful Scan

Scanned2024-03-20T14:24:31+00:00
URL https://tones.be/robots.txt
Redirect https://www.tones.be/robots.txt
Redirect Domain www.tones.be
Redirect Base tones.be
Domain IPs 217.19.237.54
Redirect IPs 23.59.168.129, 2600:1413:b000:1d::17d1:2e8d, 2600:1413:b000:1d::17d1:2ea3
Response IP 96.17.180.46
Found Yes
Hash 8ab2a0360f46cee597e0bf89badc133370026f45cc32026ecc91a280d2b02e41
SimHash 2497e8a6cdf9

Groups

*

Rule Path
Disallow *q%3D
Disallow *pageSize%3D
Disallow */budget-card
Disallow */cart
Disallow */checkout
Disallow */instore
Disallow */login
Disallow */myaccount
Disallow */santander
Disallow */sharedlist
Disallow */contentful-preview/*
Disallow /nl/leveranciers
Disallow /fr/fournisseurs
Disallow /nl/lijstjes
Disallow /fr/listes
Disallow /*/c/*q%3D
Disallow /*/c/*pageSize%3D
Disallow /*/c/*reviewPage%3D
Disallow /nl/fr/*
Disallow /nl/en/*
Disallow /fr/nl/*
Disallow /fr/en/*
Disallow /en/nl/*
Disallow /en/fr/*

cazoodlebot

Rule Path
Disallow /

dotbot/1.0

Rule Path
Disallow /

gigabot

Rule Path
Disallow /

yandex

Rule Path
Disallow /

mj12bot

Rule Path
Disallow /

ahrefsbot

Rule Path
Disallow /

baiduspider

Rule Path
Disallow /

nutch

Rule Path
Disallow /

ia_archiver

Rule Path
Disallow /

webcopier

Rule Path
Disallow /

webstripper

Rule Path
Disallow /

teleport

Rule Path
Disallow /

offline explorer

Rule Path
Disallow /

sitesnagger

Rule Path
Disallow /

psbot

Rule Path
Disallow /

teleportpro

Rule Path
Disallow /

emailsiphon

Rule Path
Disallow /

emailcollector

Rule Path
Disallow /

larbin

Rule Path
Disallow /

blexbot

Rule Path
Disallow /

msiecrawler

Rule Path
Disallow /

moget

Rule Path
Disallow /

Other Records

Field Value
sitemap https://www.tones.be/sitemap.xml

Comments

  • For all robots
  • Block access to specific groups of pages
  • Block comnination of multiple locales
  • Sitemap
  • Block CazoodleBot as it does not present correct accept content headers
  • Block dotbot as it cannot parse base urls properly
  • Block Gigabot
  • Yandex bot - A rule breaker, just as Baidu spiders
  • Worst bots according to https://www.benfrederickson.com/robots-txt-analysis/