Robots.txt Generator
Build a valid robots.txt file by adding allow/disallow rules per user-agent. Control which pages and bots can crawl your website.
Frequently Asked Questions
robots.txt file lives at the root of your domain (e.g., https://example.com/robots.txt) and tells crawlers which pages they are allowed to access. It follows the Robots Exclusion Standard. Note: it is advisory, not a security mechanism — malicious bots can and do ignore it.noindex meta tag instead.User-agent: * applies rules to all crawlers. You can target specific crawlers by name — for example, User-agent: Googlebot for Google's main crawler or User-agent: Bingbot for Microsoft Bing. Specific rules take precedence over wildcard rules for that bot.Sitemap: directive in robots.txt makes it easy for crawlers to discover your XML sitemap without it needing to be submitted via Google Search Console. You can include multiple Sitemap: lines for multiple sitemaps.How It Works
Add Rules
Select a user-agent and choose Allow or Disallow for a specific path. Add as many rules as needed, or use a quick preset to start.
Preview Instantly
The robots.txt output updates in real time as you add or modify rules. Rules are grouped by user-agent in the correct format.
Copy or Download
Copy the generated content or download it as robots.txt and upload it to the root directory of your web server.
Common Use Cases
Block Admin Areas
Disallow /admin/, /wp-admin/, and /login/ paths to prevent crawlers from wasting crawl budget on private areas.
Block Staging Sites
Use Disallow: / for all bots on staging environments to prevent duplicate content issues if the staging URL is ever discovered.
Crawl Budget Management
Block paginated pages, filter/sort URLs, and internal search result pages to focus Google's crawl budget on your most important content.
Control Media Indexing
Use Googlebot-Image disallow rules to prevent certain image directories from appearing in Google Image search results.
Block Tag & Category Pages
On blog and CMS sites, disallow tag, category, and archive pages that create shallow content which dilutes the authority of core pages.
Sitemap Discovery
Include the Sitemap: directive to help all search engines discover your XML sitemap automatically without manual submission in each webmaster tool.