Robots.txt Generator

Allow/disallow Google, Bing, Baidu. Block /admin/, add sitemap, crawl delay. Per-robot overrides.

Robots.txt Generator

Override defaults
Google
Google Image
Google Mobile
Bing/MSN
Yahoo
Baidu
Alexa/Archive
Naver
Yahoo MM
Yahoo Blogs
Ask/Teoma
GigaBlast
DMOZ
Nutch
MSN PicSearch

About Robots.txt

What is Robots.txt?
  • Controls crawler access
  • Prevents indexing of private pages
  • Directs to sitemap
Best Practices
  • Block admin folders
  • Include sitemap URL
  • Test before deploying

What is Robots.txt Generator?

Robots.txt Generator is a free online tool that creates robots.txt files for controlling search engine crawler access to your website. You configure global allow/disallow settings, optional crawl delay, sitemap URL, per-robot overrides for major search engines (Google, Bing, Yahoo, Baidu, etc.), and disallow rules for specific folders. The tool outputs a valid robots.txt that you can download or copy and upload to your site root. SEO professionals, webmasters, and developers use it to block private areas, direct crawlers to the sitemap, and manage crawl budget. No account or signup is required.

The interface includes All Robots (Allow/Disallow), Crawl Delay (none, 5, 10, 20, 60, 120 seconds), and Sitemap URL. A section lists search engine robots (Google, Google Image, Google Mobile, Bing, Yahoo, Baidu, Alexa, Naver, and more) with Allow/Disallow/Default per robot. Disallow Folders lets you add paths like /admin/ or /cgi-bin/ with trailing slashes. The generated robots.txt appears in a textarea with download and copy buttons. An info card explains what robots.txt is and best practices.

Who Benefits from This Tool

SEO professionals and webmasters benefit when setting up or updating robots.txt. Block admin, staging, or duplicate content. Add sitemap URL for efficient crawling. Per-robot overrides let you allow Google but restrict others if needed.

Developers benefit when deploying new sites. Generate a sensible default robots.txt quickly. Avoid syntax errors. The tool produces standard-compliant output.

Site owners benefit when they lack technical knowledge. The form-based interface is easier than writing robots.txt by hand. The info card and best practices guide help avoid mistakes.

Key Features

Global Allow/Disallow

Set Allow or Disallow for all robots. Disallow / blocks the entire site. Allow (empty) permits full access unless overridden.

Crawl Delay

Optional delay between requests (5 to 120 seconds). Note: Google ignores Crawl-delay; it is supported by some other crawlers (e.g., Bing historically).

Sitemap URL

Add your sitemap URL (e.g., https://example.com/sitemap.xml). Helps crawlers discover your pages.

Per-Robot Overrides

Override defaults for Google, Google Image, Google Mobile, Bing, Yahoo, Baidu, Alexa, Naver, and more. Each can be Default, Allow, or Disallow.

Disallow Folders

Add paths with trailing slashes (e.g., /admin/, /cgi-bin/, /wp-admin/). Multiple folders supported. Add/delete rows dynamically.

Download and Copy

Download as robots.txt file or copy to clipboard. Instructions explain uploading to site root.

How to Use

  1. Set All Robots. Choose Allow or Disallow for the default.
  2. Set Crawl Delay if needed. Select delay or No Delay.
  3. Enter Sitemap URL. Add your sitemap URL (e.g., https://example.com/sitemap.xml).
  4. Override per-robot if needed. For each search engine, choose Default, Allow, or Disallow.
  5. Add Disallow Folders. Enter paths like /admin/, /wp-admin/, /private/ with trailing slashes. Add more with the Add button.
  6. Click Generate. Complete captcha if required.
  7. Download or copy. Download the file or copy the text. Upload to your site root as robots.txt.

Common Use Cases

  • Blocking admin, staging, or private folders from crawlers
  • Adding sitemap URL for efficient indexing
  • Restricting specific crawlers (e.g., block image crawlers from certain paths)
  • Setting up a new site with sensible defaults
  • Updating robots.txt after site restructuring
  • Creating robots.txt for client sites

Tips & Best Practices

Block admin and sensitive folders. Include your sitemap URL. Test robots.txt in Google Search Console after deployment. Use Disallow for paths you do not want indexed. Allow is usually redundant but can override a parent Disallow in some cases. Avoid blocking important content by mistake. Crawl-delay is ignored by Google; use it only if you target other crawlers that support it.

Limitations & Notes

Robots.txt is a guideline, not enforcement. Malicious crawlers may ignore it. Google and other major engines generally respect it. The tool generates standard syntax; edge cases or non-standard directives may need manual editing. The tool does not validate your sitemap URL. Upload the file to your site root (same level as index.html or index.php).

FAQs

Is the tool free?

Yes. No signup required. A captcha may appear.

Where do I upload robots.txt?

Upload to your website root, e.g., https://yoursite.com/robots.txt. It must be in the root directory.

Does Google support Crawl-delay?

No. Google ignores Crawl-delay. Some other crawlers (e.g., Bing in the past) supported it. You can leave it as No Delay for most sites.

Can I block only certain paths?

Yes. Use Disallow Folders to add paths like /admin/, /private/. Each path should end with /.

What is the difference between Allow and Disallow?

Disallow tells crawlers not to access the specified path. Allow permits access. For "allow all," use Allow (or leave default). For "block all," use Disallow /.

Do I need a sitemap in robots.txt?

It is recommended. Crawlers can discover sitemaps via robots.txt. You can also submit sitemaps in Search Console.

Can I have multiple Disallow lines?

Yes. Add multiple folders. Each Disallow applies to its path.

Why is my robots.txt not working?

Ensure it is in the root. Check syntax. Verify in Search Console. Some crawlers cache robots.txt; changes may take time to apply.

Does the tool support wildcards?

Support depends on the generated format. Standard robots.txt has limited wildcard support. Google supports * and $ in some cases. Check the output.

Can I use it for subdomains?

Each subdomain needs its own robots.txt. Generate one per subdomain if needed.