Robots.txt is a text file webmasters create to instruct web robots "how to crawl pages on their website". The robots.txt file is part of the the robots exclusion protocol (REP), a group of web standards that regulate how robots crawl the web, access and index content, and serve that content up to users.