株式会社オブライト
SEO2026-05-17

robots.txt

Also known as: robots.txt / クローラー制御ファイル

A text file placed at the site root that specifies which URLs crawlers may or may not access, using User-agent and Disallow/Allow directives to control crawl scope.


Overview

robots.txt is an industry-standard file based on the Robots Exclusion Protocol. Common uses include blocking admin areas, staging environments, and duplicate content pages. Note that Disallow prevents crawling, not indexing — pages already indexed remain indexed.

Controlling AI Crawlers

AI training crawlers such as GPTBot, ClaudeBot, and Google-Extended can be individually controlled. See the Google-Extended and llms.txt entries, along with our LLMO technical implementation guide.

Related Columns

Related Terms

Feel free to contact us

Contact Us