credenciamento.tse.jus.br
robots.txt

Robots Exclusion Standard data for credenciamento.tse.jus.br

Resource Scan

Scan Details

Site Domain credenciamento.tse.jus.br
Base Domain tse.jus.br
Scan Status Failed
Failure ReasonScan timed out.
Last Scan2025-03-11T06:30:37+00:00
Next Scan 2025-05-10T06:30:37+00:00

Last Successful Scan

Scanned2024-12-19T06:17:23+00:00
URL https://credenciamento.tse.jus.br/robots.txt
Domain IPs 23.213.43.210, 23.213.43.219, 2600:1413:b000:6::17d5:2bd2, 2600:1413:b000:6::17d5:2bdb
Response IP 23.213.43.219
Found Yes
Hash 18c39c85791823cc3e99cbf4099cea503f3696ac3b0007debdae19153cf148ac
SimHash 2d1581554d41

Groups

*

Rule Path
Disallow /*?
Disallow /*atct_album_view$
Disallow /*folder_factories$
Disallow /*folder_summary_view$
Disallow /*login_form$
Disallow /*mail_password_form$
Disallow /*search
Disallow /*search_rss
Disallow /*searchRSS
Disallow /*updated_search
Disallow /*sendto_form$

Other Records

Field Value
sitemap /sitemap.xml.gz

Comments

  • Define access-restrictions for robots/spiders
  • http://www.robotstxt.org/wc/norobots.html
  • By default we allow robots to access all areas of our site accessible to
  • anonymous users, except for search, which burns our CPU for no reason.
  • Block all URLs including query strings (? pattern) - contentish objects expose query string only for actions or status reports which
  • might confuse search results.
  • This will also block ?set_language
  • Add Googlebot-specific syntax extension to exclude forms
  • that are repeated for each piece of content in the site
  • the wildcard is only supported by Googlebot
  • http://www.google.com/support/webmasters/bin/answer.py?answer=40367&ctx=sibling