Create a properly configured robots.txt file with AI crawler rules, social bot access, and sitemap directives.
# robots.txt generated by Growtika # 2026-03-14 User-agent: * Disallow: /admin/ Disallow: /private/ Disallow: /staging/ # AI Crawlers User-agent: GPTBot Allow: / User-agent: ChatGPT-User Allow: / User-agent: Google-Extended Allow: / User-agent: anthropic-ai Allow: / User-agent: ClaudeBot Allow: / User-agent: PerplexityBot Allow: / User-agent: Applebot-Extended Allow: / User-agent: cohere-ai Allow: / User-agent: CCBot Allow: / User-agent: meta-externalagent Allow: / # Social Media Crawlers User-agent: Twitterbot Allow: / User-agent: facebookexternalhit Allow: / User-agent: LinkedInBot Allow: /Validate this file with our Checker
Writing a robots.txt file manually is error-prone. A misplaced directive or forgotten crawler can make your entire site invisible to search engines or AI platforms. A robots.txt generator creates a properly formatted file based on your preferences, ensuring correct syntax and complete coverage.
This is especially important in 2026, when your robots.txt needs to account for over a dozen AI-specific crawlers beyond traditional search bots. Each AI platform (ChatGPT, Claude, Perplexity, Gemini, Copilot) uses its own user-agent, and each requires explicit rules if you are using a restrictive default policy.
Our generator creates a robots.txt file optimized for both AI search visibility and traditional SEO, with toggles for AI crawlers, social media bots, crawl delays, and sitemap declarations.
yourdomain.com/robots.txt).What we configure for every client and recommend you do the same.
The safest default is Allow: / under User-agent: *. From there, add specific Disallow rules only for paths that genuinely need protection (admin panels, staging, internal APIs). The opposite approach, blocking everything and whitelisting, is fragile and guaranteed to break when someone adds a new page or directory.
GPTBot, ChatGPT-User, ClaudeBot, anthropic-ai, PerplexityBot, Google-Extended. Each one needs its own User-agent section with Allow: /. This is not optional in 2026. If you use a restrictive wildcard policy and forget to add an AI bot, that platform cannot see your content. Our generator handles this automatically.
Twitterbot, facebookexternalhit, and LinkedInBot are not search crawlers. They render Open Graph previews when someone shares your link. If these are blocked, your shared links appear as blank cards with no title, image, or description. That kills click-through rates on social channels.
A Sitemap: directive at the bottom of robots.txt tells every crawler where to find your XML sitemap. It is the simplest line you can add and one of the most impactful. Do not assume crawlers will find it through other means.
Crawl-delay is ignored by Google entirely. Bing respects it but you rarely need it. Setting a high delay slows indexing and does not meaningfully reduce server load on modern infrastructure. If your server struggles under crawler traffic, the problem is your server, not your robots.txt.
Treat robots.txt like code. Put it in Git. Review changes in pull requests. A one-line accidental edit can make your entire site invisible. If someone changes this file, you want a diff, a reviewer, and a deployment log.
We have seen every one of these. Tap to see what went wrong and how to fix it.
A marketing team generates a perfect robots.txt. It sits in a Google Doc. The live site still serves the old file.
Setting User-agent: * Disallow: / as the default, intending to add Allow rules for specific bots. Then shipping it without the bot-specific blocks.
Blocking /api/ seems reasonable. But if your marketing site uses /api-integrations/ or /api-security-guide/ as blog paths, those get blocked too.
Writing Disallow: /admin instead of Disallow: /admin/
Setting a 10-second crawl delay because someone read it was best practice.
A correct robots.txt is the foundation. We help B2B companies become the recommended answer in ChatGPT, Perplexity, and Google AI Overviews.
Get a Free AI Visibility Audit →