robots.txt Generator

Create valid robots.txt files with an easy-to-use visual interface

Configuration

Quick Presets
User Agent
Allow/Disallow Rules
No rules added yet
Crawl Delay (Optional)
Sitemap URLs

Generated robots.txt

🔒 Privacy First: All processing happens in your browser. No data is sent to any server. Your robots.txt configuration never leaves your device.

Understanding robots.txt Files

The robots.txt file is a text file placed in your website's root directory that tells web crawlers and search engine bots which pages or sections of your site they can or cannot access. It's part of the Robots Exclusion Protocol (REP), a standard used by websites to communicate with web robots.

Why robots.txt is Important

Basic Syntax

The robots.txt file uses a simple syntax:

# Comment
User-agent: *
Disallow: /private/
Allow: /public/
Crawl-delay: 10
Sitemap: https://example.com/sitemap.xml

Common Directives

Blocking AI Crawlers

With the rise of AI systems, many website owners want to prevent their content from being used for AI training. Common AI crawlers include:

Best Practices

  1. Always place robots.txt in your root directory (e.g., example.com/robots.txt)
  2. Use specific rules rather than blocking entire sections when possible
  3. Include your sitemap URL for better indexing
  4. Test your robots.txt using Google Search Console
  5. Remember that robots.txt is publicly accessible and not a security measure
  6. Update regularly as your site structure changes

Related Tools

You might also find these utilities helpful: