Introduction to robots.txt
A robots.txt file tells search engine crawlers which URLs the crawler can access on your site. This is used mainly to avoid overloading your site with requests;it is not a mechanism for keeping a web page out of Google. To keep a web page out of Google,block indexing withnoindex or password-protect the page.
If you use a CMS, such as Wix or Blogger, you might not need to (or be able to) edit your robots.txt file directly. Instead, your CMS might expose a search settings page or some other mechanism to tell search engines whether or not to crawl your page.
If you want to hide or unhide one of your pages from search engines, search for instructions about modifying your page visibility in search engines on your CMS (for example, search for "wix hide page from search engines").
What is a robots.txt file used for?
A robots.txt file is used primarily to manage crawler traffic to your site, andusually to keep a file off Google, depending on the file type:
Understand the limitations of a robots.txt file
Before you create or edit a robots.txt file, you should know the limits of this URL blocking method. Depending on your goals and situation, you might want to consider other mechanisms to ensure your URLs are not findable on the web.
- robots.txt rules may not be supported by all search engines.
The instructions in robots.txt files cannot enforce crawler behavior to your site; it's up to the crawler to obey them. While Googlebot and other respectable web crawlers obey the instructions in a robots.txt file, other crawlers might not. Therefore, if you want to keep information secure from web crawlers, it's better to use other blocking methods, such aspassword-protecting private files on your server. - Different crawlers interpret syntax differently.
Although respectable web crawlers follow the rules in a robots.txt file, each crawler might interpret the rules differently. You should know theproper syntax for addressing different web crawlers as some might not understand certain instructions. - A page that's disallowed in robots.txt can still be indexed if linked to from other sites.
While Google won't crawl or index the content blocked by a robots.txt file, we might still find and index a disallowed URL if it is linked from other places on the web. As a result, the URL address and, potentially, other publicly available information such as anchor text in links to the page can still appear in Google Search results. To properly prevent your URL from appearing in Google Search results,password-protect the files on your server,use thenoindexmetatag or response header, or remove the page entirely.
Create or update a robots.txt file
If you decided that you need one, learn how tocreate a robots.txt file. Or if you already have one, learn how toupdate it.
Want to learn more? Check out the following resources:
Except as otherwise noted, the content of this page is licensed under theCreative Commons Attribution 4.0 License, and code samples are licensed under theApache 2.0 License. For details, see theGoogle Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.
Last updated 2025-12-10 UTC.