Robots.txt Tester
Fetch and analyze robots.txt
. See rules per user-agent, Sitemap lines, file size, and test whether a path is Allowed or Blocked for a specific bot.
Fetch and analyze robots.txt
. See rules per user-agent, Sitemap lines, file size, and test whether a path is Allowed or Blocked for a specific bot.
A Robots.txt Tester is an SEO tool that fetches your site’s
/robots.txt
, parses User-agent sections and their
Allow
/Disallow
rules, lists declared Sitemap URLs,
and lets you test any path to see if it’s Allowed or Blocked
for a specific crawler (e.g., Googlebot, Bingbot, DuckDuckBot).
Sitemap:
locations directly in robots.txt
.User-agent
sections (e.g., Googlebot-Image)./robots.txt
)./blog/amp/
or /wp-admin/admin-ajax.php
).Allow
/Disallow
with wildcards (*
) and end anchors ($
).Sitemap:
lines pointing to your XML sitemaps./
).User-agent
if you need bot-specific behavior.Disallow: /
under User-agent: *
prevents crawling of everything.
Fix: Remove or scope to a specific user-agent.noindex
meta (on accessible pages) or remove from sitemaps; consider 410/404./wp-includes/
or /assets/
may hide CSS/JS.
Fix: Allow assets needed to render pages.text/plain
and keep it compact.User-agent:
then rules.User-agent: * Allow: /wp-admin/admin-ajax.php Disallow: /wp-admin/ Disallow: /search/ Sitemap: https://example.com/sitemap.xml
User-agent: * Disallow: /*?sort= Disallow: /*?price= Disallow: /*?color= Allow: /product/ Sitemap: https://example.com/sitemap_index.xml
User-agent: Googlebot-Image Allow: /images/ User-agent: * Disallow: /tmp/ Sitemap: https://example.com/sitemap.xml
We evaluate the selected User-agent section using Google-style pattern rules:
*
matches any sequence; end anchor: $
matches end of URL./robots.txt
and returns 200 OK with text/plain.Sitemap:
declared and reachable (200 OK).User-agent
and use correct patterns.Disallow: /
under User-agent: *
unless intentional.Enter the site URL, select a user-agent, and add your path (e.g., /blog/tag/
). The tester shows Allowed or Blocked and which rule matched.
No. It only controls crawling. To remove indexed pages, allow crawling and use noindex
, remove from sitemaps, or return 410/404.
For staging, use authentication or IP restrictions (best). For public admin paths, you can disallow crawling—but never rely on robots.txt for security.
Add one or more Sitemap:
lines in /robots.txt
pointing to your XML sitemap(s) for faster discovery.