sitemap.xml is an XML file listing every canonical URL for search engines. Contains loc, lastmod, changefreq, priority. Limit: 50,000 URLs / 50 MB per file. For large sites — sitemap-index referencing partitioned sitemaps.
sitemap.xml is an XML file listing every canonical URL for search engines. Contains loc, lastmod, changefreq, priority. Limit: 50,000 URLs / 50 MB per file. For large sites — sitemap-index referencing partitioned sitemaps.
robots.txt controls which pages search bots can see. Incorrect directives can accidentally block the entire site from indexing or expose administrative sections.
Parse robots.txt per RFC 9309: all User-agent, Allow/Disallow, Crawl-delay, Sitemap.
Enter a specific URL and User-agent — find out if it's allowed for that bot.
Automatically show status for GPTBot, ClaudeBot, PerplexityBot, Googlebot.
All Sitemap: directives in one place with quick links for verification.
crawl directive audit
post-deploy check
indexation control
block unwanted crawlers
User-agent: * applies to all bots, including AI crawlers.Sitemap: https://example.com/sitemap.xml helps bots find all pages.Robots.txt check history and change monitoring for your site.
Sign up freeIf you work with web infrastructure — yes. See description above.