Robots.txt Validator

Validate your robots.txt file to ensure it's correctly configured for search engine crawlers.

Best Practices for robots.txt

  • Use the correct syntax for user-agent and disallow directives
  • Be specific with your disallow rules to avoid over-blocking
  • Include a sitemap directive if you have one
  • Test your robots.txt file regularly, especially after site changes
  • Use comments to explain complex rules for future reference

Common robots.txt Directives

  • User-agent: Specifies which web crawler the rules apply to
  • Disallow: Tells crawlers which pages or directories not to access
  • Allow: Explicitly allows crawling of specific pages (used with Disallow)
  • Sitemap: Indicates the location of your XML sitemap