Control how search engines crawl your website.



Instruct search engines robots on how to crawl. LEARN MORE

Get generated code:






ABOUT Robots.txt Generator

Before we explain in details about Robots.txt generator tool, let’s brief in few lines about Robots.txt file.

Robots.txt is the first file that is being crawled by Search Engine spiders as it looks for instructions wrt pages that are blocked. Robots.txt can be used to clock a particular Search engine Spider from crawling a particular page or it can also be used to block the entire page from getting crawled by all spiders.

Robots.txt can block a particular page, folder or a bunch of pages.

LXR’s Robots.txt Generator tool makes the task a notch easier.


  • Provides effective feature to instruct whether to crawl or not to crawl a particular page.

  • Allows you to select the Robots/spider/bots that you would like to block from crawling the page.

  • Allows you to also provide sitemap link information which is also a key component to be added in robots.txt page.

  • Most importantly, it helps you generator error free Robots.txt file that you can update in website root.

01 Enter the details
How to use tool image
Enter the URL of the website or page you want to get data for.

Get a comprehensive analysis for the URL you entered, plus much more!
02 Steps to follow
  • The usage of Robots.txt Generator Tool is simple enough. Here are the steps to be followed to create an effective Robots.txt file.

  • Provide the URL of your website.

  • Select whether you would like to allow or disallow certain sets of URLs.

  • Select the Search Engine bots you would like to allow or disallow.

  • Provide the URL of the specific page/folder that you would like to block.

  • You can use the (+) sign and add more number of entries in a similar way.

  • Add the URL destination of the sitemap.

  • Click on the button ‘Generate Robots.txt’.

  • *Please note that while you mention the URL you would like to block, provide only the suffix part of the URL after the domain name and the (/) sign.

Once the Robots.txt file is generated, submit it at the root directory of your website.
As iterated, Robots.txt file is the first file visited by Search Engine Robots before it crawls the entire website. Hence, it is important to provide the URL destination of the sitemap so that Search Engine spiders can access it quickly.
Robots.txt is a text file which is created with the purpose of indicating to search engine spiders the pages that are being blocked from the crawling process.
A Robots.txt file is created in a text file with disallow command, thereby instructing the search engine spiders about the pages that shouldn't be crawled. The same Robots.txt file is placed in the root directory of the website.
The Robots.txt file is located at the root directory of the website. To remove it, one would only need to take down the file from there.
In the Robots.txt file, one should mention the pages that's to be blocked from crawling by search engine robots.
The disallow command helps in blocking pages/folders from getting crawled.
Allow command is occasionally used in case is a certain page(s) are to be still crawled from a blocked folder.
The location of XML sitemap is too mentioned.
Robots.txt is absolutely necessary for search engines - it is the first file accessed by the search engine robots to ascertain the pages that it should or should not crawl.
The Robots.txt can be found at the root directory of the website - you can access the file by typing robots.txt as suffix/extension after the domain.For example, if your domain is abc.com, the robots.txt file can be accessed by typing in www.abc.com/robots.txt.
Yes, Google can still access your website even if you don't have a Robots.txt file at the root directory.
However still, it is recommended to have Robots.txt file as it is necessary for the search engine spiders to understand the pages that are blocked from crawling. Also, it provides sitemap information to the search engine spiders that facilitate indexation.
The only way to ascertain whether the Robots.txt file is working or not, is to directly check on the search engine by tying in the website URL. If the site appears in the search results, the robots.txt is not working or Google has disregarded the Robots.txt file.
You can then check if there are any technical error in the process of creating Robots.txt.
In certain situations, the search engine can disregard the Robots.txt file and still crawl the blocked pages. To ensure that a page is indeed blocked, one can make use of the Robots metatag.
Ask The Expert

Our Google certified experts are ready to answer any questions you might have about your website or digital marketing campaign. How can we help you?