Robots.txt

The robots.txt file is used to specify which web pages can be indexed by search engines. This file should be placed in the root directory of the website, and each subdomain should have its own robots.txt file.

Instructions in robots.txt follow a particular standard (see examples below). However, these instructions can't be enforced; compliance depends on search engines respecting the file's directives. Some malicious bots may even use the robots.txt file to find sensitive webpages and go directly to them.

Examples

Forbid indexation of the website to all robots.

User-agent: *
Disallow: 

Allow any robots to access all the website.

User-agent: *
Allow: /

Last updated