imgix outputs a default /robots.txt file for every domain with the following:
User-agent: * Disallow:
The "User-agent: *" means that the instructions will apply to all search engine robots, such as Google's Googlebot. The "Disallow: " tells each robot that it has complete access to crawl every URL associated with the domain. This is equivalent to having an empty /robots.txt or none at all. For more information on /robots.txt, please refer here: http://www.robotstxt.org/robotstxt.html
Requests made by robots are counted as regular requests, which means they contribute toward the Origin Image and bandwidth calculations for billing.
Using a custom robots.txt file
It's possible to use your own custom robots.txt on your Source. To do so:
- Upload your
robots.txt
to your origin so that it can be accessed at the root of your imgix domain (ie:subdomain.imgix.net/robots.txt
) - Contact support@imgix.com to enable your custom robots.txt to be loaded and let us know the Source(s) you'd like to enable it for
Note that you will have to use our purge API if any changes are made to your robots.txt
file.