User-agent: *
allow
/disallow
/admin/disallow
/private/disallow
/wp-login.phpdisallow
/search?disallow
/*.pdf$Fetch or paste robots.txt rules, test any URL against a chosen crawler, and see exactly which directive allows or blocks the path.
Googlebot can crawl this URL.
//admin//private//wp-login.php/search?/*.pdf$///staging/2https://example.com/sitemap.xmlThe longer and more specific path wins. If Allow and Disallow match equally, Allow typically takes priority.
robots.txt only guides compliant bots. It is not access control and should never protect private or sensitive URLs.
Place robots.txt at the root of the domain, for example https://example.com/robots.txt. Subdirectory robots files are ignored.