BeBizzi Wiki

What is the robots.txt file on my website?

The robots.txt file is a simple text file placed in your website’s root directory that tells search engine crawlers (like Google, Bing) which pages or sections of your site they can or cannot access. It’s part of the Robots Exclusion Protocol, a standard used to manage bot traffic.

Why it matters:

  • Control crawling: Prevent bots from indexing sensitive pages (e.g., admin areas, duplicate content, or staging environments)
  • Optimize crawl budget: Guide search engines to focus on your most important pages
  • Protect resources: Reduce server load by blocking unnecessary bot requests

Example use cases:

  • Blocking search engines from indexing /private/ or /admin/ directories
  • Preventing duplicate content issues by disallowing certain URL parameters
  • Directing crawlers to your XML sitemap

Important note: robots.txt is publicly accessible (yoursite.com/robots.txt) and only a guideline—malicious bots may ignore it. It’s not a security measure; use proper authentication for truly sensitive content.

For most businesses, a well-configured robots.txt helps improve SEO performance and ensures search engines focus on your most valuable content.

BACK TO THE WIKI HOME >