onvo.cn
robots.txt

Robots Exclusion Standard data for onvo.cn

Resource Scan

Scan Details

Site Domain onvo.cn
Base Domain onvo.cn
Scan Status Ok
Last Scan2026-01-31T11:56:35+00:00
Next Scan 2026-03-02T11:56:35+00:00

Last Scan

Scanned2026-01-31T11:56:35+00:00
URL https://onvo.cn/robots.txt
Redirect https://www.onvo.cn/robots.txt
Redirect Domain www.onvo.cn
Redirect Base onvo.cn
Domain IPs 101.42.123.192
Redirect IPs 101.42.123.192
Response IP 101.42.123.192
Found Yes
Hash f3ec59b61231689c747ef59452f08757a882e5229b5c925fe5b6f983867f855b
SimHash bc929d0bc174

Groups

feeddemon

Rule Path
Disallow /

bot/0.1 (bot for jce)

Rule Path
Disallow /

crawldaddy

Rule Path
Disallow /

java

Rule Path
Disallow /

jullo

Rule Path
Disallow /

feedly

Rule Path
Disallow /

universalfeedparser

Rule Path
Disallow /

apachebench

Rule Path
Disallow /

swiftbot

Rule Path
Disallow /

yandexbot

Rule Path
Disallow /

jikespider

Rule Path
Disallow /

mj12bot

Rule Path
Disallow /

zmeu phpmyadmin

Rule Path
Disallow /

winhttp

Rule Path
Disallow /

easouspider

Rule Path
Disallow /

httpclient

Rule Path
Disallow /

microsoft url control

Rule Path
Disallow /

yyspider

Rule Path
Disallow /

jaunty

Rule Path
Disallow /

obot

Rule Path
Disallow /

python-urllib

Rule Path
Disallow /

indy library

Rule Path
Disallow /

flightdeckreports bot

Rule Path
Disallow /

linguee bot

Rule Path
Disallow /

*

Rule Path
Disallow /account/
Disallow /zh_CN/

Other Records

Field Value
sitemap https://www.onvo.cn/sitemap.xml

Comments

  • robots.txt
  • This file is to prevent the crawling and indexing of certain parts
  • of your site by web crawlers and spiders run by sites like Yahoo!
  • and Google. By telling these "robots" where not to go on your site,
  • you save bandwidth and server resources.
  • This file will be ignored unless it is at the root of your host:
  • Used: http://example.com/robots.txt
  • Ignored: http://example.com/site/robots.txt
  • For more information about the robots.txt standard, see:
  • http://www.robotstxt.org/robotstxt.html
  • Custom crawlers disable
  • Directories