robots.txt Generator
Control How Search Engines Crawl Your Site
Create robots.txt files visually. Define crawl rules for search engines, specify sitemaps, and control bot access - all without memorizing syntax. 100% client-side. Zero data stored.
robots.txt Generator Tool
User Agents
Common Disallow Patterns
Click to add to the first user-agent:
Sitemap URLs
Crawl Delay
Note: Google ignores Crawl-delay. Use Google Search Console instead.
Presets
Generated robots.txt
Common User Agents Reference
Googlebot
Google's main web crawler for search indexing.
Bingbot
Microsoft Bing's web crawler.
GPTBot
OpenAI's bot for training data collection.
ChatGPT-User
ChatGPT's browsing feature bot.
Claude-Web
Anthropic's Claude AI web crawler.
Baiduspider
Baidu's web crawler (China).
Frequently Asked Questions
Why We Built robots.txt Generator
Writing robots.txt files seems simple, but getting the syntax right matters. A misplaced directive can accidentally block search engines from your entire site, devastating your SEO. We built this tool to help developers create correct robots.txt files visually.
With the rise of AI crawlers like GPTBot and ChatGPT-User, many site owners want to control which bots can access their content. Our generator makes it easy to block AI training bots while still allowing traditional search engines.
robots.txt Generator is part of RJL.io's collection of free developer tools - each designed to do one thing exceptionally well, with no accounts, no tracking, and no data collection. Check out our other tools: .htaccess Generator, Meta Tag Generator, Base64 Encoder, and more.