Robots.txt Tester

Fetch and analyze robots.txt. See rules per user-agent, Sitemap lines, file size, and test whether a path is Allowed or Blocked for a specific bot.

What Is a Robots.txt Tester?

A Robots.txt Tester is an SEO tool that fetches your site’s /robots.txt, parses User-agent sections and their Allow/Disallow rules, lists declared Sitemap URLs, and lets you test any path to see if it’s Allowed or Blocked for a specific crawler (e.g., Googlebot, Bingbot, DuckDuckBot).

Why Robots.txt Matters for SEO

How to Use This Robots.txt Tester

  1. Enter any page on your site (we’ll resolve the correct /robots.txt).
  2. Pick a User-agent (Googlebot by default).
  3. (Optional) Enter a path to test (e.g., /blog/amp/ or /wp-admin/admin-ajax.php).
  4. Click Test Robots.txt to see parsed rules, sitemaps, and path result.

Quick Tips for Clean Robots Rules

Common Robots.txt Mistakes (and Fixes)

Recommended Robots.txt Templates

Standard Website

User-agent: *
Allow: /wp-admin/admin-ajax.php
Disallow: /wp-admin/
Disallow: /search/
Sitemap: https://example.com/sitemap.xml

E-commerce (filter parameters)

User-agent: *
Disallow: /*?sort=
Disallow: /*?price=
Disallow: /*?color=
Allow: /product/
Sitemap: https://example.com/sitemap_index.xml

Media-heavy (let Googlebot-Image crawl images)

User-agent: Googlebot-Image
Allow: /images/
User-agent: *
Disallow: /tmp/
Sitemap: https://example.com/sitemap.xml

How Our “Path Test” Logic Works

We evaluate the selected User-agent section using Google-style pattern rules:

Robots.txt vs. Noindex vs. Canonical

Robots.txt Checklist

FAQ: Robots.txt Tester

How do I check if a specific URL is blocked?

Enter the site URL, select a user-agent, and add your path (e.g., /blog/tag/). The tester shows Allowed or Blocked and which rule matched.

Does robots.txt remove pages from Google?

No. It only controls crawling. To remove indexed pages, allow crawling and use noindex, remove from sitemaps, or return 410/404.

Should I block staging or admin areas?

For staging, use authentication or IP restrictions (best). For public admin paths, you can disallow crawling—but never rely on robots.txt for security.

Where should I put my sitemap?

Add one or more Sitemap: lines in /robots.txt pointing to your XML sitemap(s) for faster discovery.