robots.txt Generator
Build a robots.txt with multiple user‑agent blocks, Allow/Disallow, optional Crawl-delay, Host, and Sitemap lines. Live preview updates as you type.
User‑agent blocks
Global
Notes:
Host is honored by some crawlers (e.g., Yandex). Crawl-delay isn’t used by Google but may be used by others. Wildcards * and $ are widely supported (not in the core RFC).
Preview (robots.txt)
URL tester
Reminder:
robots.txt is publicly accessible and is an advisory for well‑behaved crawlers; it’s not an access control mechanism.