To exclude all robots from the entire server:
User-agent: * Disallow: /
To exclude all robots from part of the server:
User-agent: * Disallow: /cgi-bin/ Disallow: /tmp/ Disallow: /junk/
To allow a single robot:
User-agent: Google Disallow: User-agent: * Disallow: /
To allow all robots complete access:
User-agent: * Disallow:
See:
Sitemaps protocol allows a webmaster to inform search engines about URLs on a website that are available for crawling.