Robots.txt Generator

Robots.txt Generator is a free online SEO tool that creates custom Robots.txt files for your website. Create your custom robots.txt file online in seconds.

Default - All Robots are:  
Sitemap: (leave blank if you don't have) 
Search Robots: Google
  Google Image
  Google Mobile
  MSN Search
  Yahoo MM
  Yahoo Blogs
  DMOZ Checker
  MSN PicSearch
Restricted Directories: The path is relative to root and must contain a trailing slash "/"

Now, Create 'robots.txt' file at your root directory. Copy above text and paste into the text file.

Overview of our online Robots.txt generator

Robots.txt is a text file that tells search robots which pages should be kept private and not seen by other people. It's a text file so don't compare it to an HTML one. Robots.txt is sometimes mistaken for a firewall or some other password protection function. Robots.txt ensures that essential data that web owners want to keep private is kept out.

Our Robots.txt generator tool is designed to help webmasters, SEOs and marketers create their robots.txt files. Please be careful though, as creating your robots.txt file can have a significant impact on Google being able to access your website, whether it's built on WordPress or another CMS.

The Robots.txt generator creates a file that is the exact opposite of a sitemap that conditions pages to cover; Therefore, robots.txt syntax is very important for any site. Every time a search engine crawls a site, it first searches the robots.txt file located at the domain root level. Once detected, the crawler will read the file, and then identify directories and files that may be blocked.

Although our tool is easy to use, we recommend that you familiarize yourself with Google's guidelines before using it. This is because the wrong implementation can cause search engines like Google to be unable to crawl critical pages on your site or even your entire domain, which can very negatively affect your SEO.

What is a Robots.txt file?

The Robots.txt file is a simple text file that tells search engine bots whether or not they can crawl your website. You can also tell search bots about webpages you don't need to crawl, such as areas that contain duplicate content or aren't developed.

A robot.txt file includes the user-agent and under it, you may also be able to enter other directives such as allow, deny or crawl-delay. Writing it manually will take a lot of time. But using this tool you can create your file in seconds.

Search engines check the instructions in the robots.txt file before starting to crawl a website and its content. A robots.txt file is useful if you don't want certain parts of your website to be searchable, such as the thank you page or pages with confidential or legal information.

To check if your website already has a robots.txt file, go to the address bar in your browser and add /robot.txt to your domain name. The URL should be:

If your website already has a robots.txt file, there are a few more additions you can make to optimize your SEO. If you can't find a robots.txt file, you can create one - it's very easy.

How to create a Robots.txt file?

Our Robots.txt generator tool is designed to help you generate robot standard files quickly and without any technical issues. It doesn't matter if it's built on WordPress or another CMS.

The first option you will be presented with is to allow or disallow all web crawlers to access your website. This menu allows you to decide whether you want your website to be crawled; However, there may be reasons why you may choose not to have your website indexed by Google.

Second option allows you to add xml sitemap to your website. Just enter its location in this field. (If you want to create an XML sitemap, you can use our free tool.)

Finally, you are given the option to block certain pages or directories from being indexed by search engines. This is usually done for all pages that do not provide any useful information to Google or other search engines and users (such as login, cart and parameter pages).

Once this is done, you can download the text file.

After you've created your robots.txt file, make sure to upload it to the root directory of your domain For example, your robots.txt file should appear at:

How to manually create a Robots.txt file?

If you are a Windows user, you can use Notepad to create files. For Mac users, the TextEdit program works just fine. We just want to create a blank TXT file. Do not use programs such as MS Word for this task as they may cause encoding problems. Name the file “robots.txt” and save it.

The robots.txt file will now be empty; You need to add the instructions you want - which we're going to see. When you are done with the instructions, upload the robots.txt file to the root of your website using an FTP software like FileZilla or the file manager provided by your hosting provider. Note that if you have subdomains, you need to create robots.txt files for each subdomain.

Now let's see what kind of instructions you can give the robots through your robots.txt file.

If you want all robots to access everything on your website, your robots.txt file should look like this:


Basically, the robots.txt file here doesn't allow anything, or in other words, allows everything to be crawled. An asterisk next to "User-Agent" means that the instruction below applies to all types of robots

On the other hand, if you don't want robots to access anything, just add the forward slash symbol like this:

*Disallow: /

Note that an extra character can invalidate the directive, so be careful when editing your robots.txt file.

If you want to block access to a certain type of GoogleBots, such as searching for images, you can enter:

User-agent: googlebot-images Disallow: /

Or if you want to block access to certain types of files like PDFs, enter this:

User-agent: *
Allow: /
# Disallowed File Types
Disallow: /*.PDF$

If you want to block access to a directory within your website, for example, the admin directory, enter this:

User-agent: *
Disallow: /admin

If you want to block a specific page, just type its URL:

User-agent: *
Disallow: /page-url

And if you don't want Google to index a page, add this directive:

User-agent: *
Noindex: /page-url

If you're not sure what indexing means, it's simply a process that ranks a page in web search.

Finally, for large websites that are frequently updated with new content, it is possible to set up a delay timer to prevent the servers from being overloaded by crawlers coming to check for new content. In a case like this, you can add the following directive:

User-agent: *
Crawl-delay: 120

This way all robots (except Googlebots, which ignore this request) will delay their crawling by 120 seconds, preventing many robots from hitting your server too quickly.

There are other types of instructions you can add, but these are the most important to know.

Importance of Robots.txt for SEO

The reason we recommend using a robots.txt file is that your website may be subject to many third-party crawlers trying to access the content, which can result in slow load times and sometimes server errors. Loading speed affects the experience of website visitors, many of whom will leave your site if it doesn't load quickly. Using a robots.txt file allows you several options:

  • You want to direct search engines to your most important pages
  • You want search engines to ignore duplicate pages, such as pages formatted for print
  • You don't want some content on your website to be searchable (documents, images, etc.)

That's why it's important to know exactly what you put in your robots.txt file so that it improves your SEO optimization rather than compromising it. A robots.txt file with incorrect instructions can cause huge problems and possibly prevent pages from showing up in search results.

Robots.txt files are important and useful if you are not indexing duplicate and broken pages of your website, specific areas of your site and login pages, XML sitemaps. By using the robots.txt file, you can remove pages that don't add any value to your website as search engines focus on the most important pages to crawl.

Search engines can crawl a limited number of pages in a day so it will be beneficial for search engines if you block some unimportant URLs, so they can crawl your pages faster.

If Google finds that crawling your site is annoying then it starts crawling your website slowly. This means that whenever Google sends its crawler to crawl the website, it will only check some pages of your website and it may take time for your latest post to be indexed.

Your website needs robots.txt and a sitemap to overcome this limitation. These files will improve the crawling process by telling which links to pay more attention to.

Our goal is to make search engine optimization (SEO) easier for everyone. We provide simple, professional quality SEO analysis and critical SEO monitoring for websites. By making our tools intuitive and easy to understand, we've helped thousands of small business owners, webmasters and SEO professionals improve their online presence.