🇺🇸 English
🇹🇼 繁體中文
🇨🇳 简体中文
🇯🇵 日本語
🇰🇷 한국어
🇪🇸 Español
🇫🇷 Français
🇩🇪 Deutsch
🇧🇷 Português
🇷🇺 Русский
🇮🇹 Italiano
🇳🇱 Nederlands
🇵🇱 Polski
🇹🇷 Türkçe
🇸🇦 العربية
🇹🇭 ไทย
🇻🇳 Tiếng Việt
🇮🇩 Indonesia
🇲🇾 Melayu
🇮🇳 हिन्दी

Robots.txt Generator

Create robots.txt files for your website

Select Bots

Crawl Rules

Generated robots.txt

About robots.txt

The robots.txt file tells search engine crawlers which URLs they can access on your site. It's placed at the root of your website (e.g., example.com/robots.txt).

User-agent: * applies rules to all bots.

Disallow: /path/ blocks access to a path.

Allow: /path/ explicitly allows access (useful after a broader Disallow).

Sitemap: tells bots where to find your XML sitemap.

Common Patterns

Disallow: / - Block all pages

Disallow: (empty) - Allow all pages

Disallow: /*.pdf$ - Block all PDF files

Disallow: /private/ - Block specific folder