unicheck.unicum.de
robots.txt

Robots Exclusion Standard data for unicheck.unicum.de

Resource Scan

Scan Details

Site Domain unicheck.unicum.de
Base Domain unicum.de
Scan Status Ok
Last Scan2024-09-09T06:44:35+00:00
Next Scan 2024-10-09T06:44:35+00:00

Last Scan

Scanned2024-09-09T06:44:35+00:00
URL https://unicheck.unicum.de/robots.txt
Domain IPs 157.90.24.236, 2a01:4f8:1c1e:c09e::1
Response IP 157.90.24.236
Found Yes
Hash b20fea8db4c362b2d464099a76826425f910bacb8f9a3a38268f2b758ac1f6cf
SimHash b8109d0b0544

Groups

*

Rule Path
Allow *
Allow /
Disallow /node_modules/
Disallow *.md
Disallow *.sh
Disallow *.json
Disallow /search/
Disallow /passwort
Disallow /passwort-vergessen
Disallow /login
Disallow /logout
Disallow /user
Disallow /thank-you
Disallow /studiengangsuche

Comments

  • robots.txt
  • This file is to prevent the crawling and indexing of certain parts
  • of your site by web crawlers and spiders run by sites like Yahoo!
  • and Google. By telling these "robots" where not to go on your site,
  • you save bandwidth and server resources.
  • This file will be ignored unless it is at the root of your host:
  • Used: http://example.com/robots.txt
  • Ignored: http://example.com/site/robots.txt
  • For more information about the robots.txt standard, see:
  • http://www.robotstxt.org/robotstxt.html
  • Directories
  • Files
  • Paths