Blogspot is a fruitful platform for every blogger because of its friendly design in all aspects. If the blogger using any platform like, WordPress, and Typepad, they must do some SEO to promote the blog in Search Engines.

Robots.txt is a small code file. It is used by Webmasters to control the search engine web robots. The robots mainly function on directories, Web Pages, blog (or) site.

How to Create and add Robots.txt code in Blogger?

What is Robots.txt?

Robots.txt’ is a text file that contains a few lines of simple code. This file helps to crawl and index your content in the search results. Always remember that search crawlers scan the ‘Robots.txt’ file before crawling any web page.

By default, each website allows the search engine robots, however, if you would like to restrict the robots either to not crawl any bound directory, file (or) the complete website, then you may want the ‘robots. txt’ file in which you are getting to write instructions for search engine bots.

Steps to edit Robots file on Blogger

  1. Login to your blogger dashboard
  2. Go to Settings > Search Preferences > Crawlers and Indexingcrawlers and indexing
  3. In the ‘Crawlers and indexing’ box, there are two options that would offer you the flexibility to customize your ‘robots.txt’ file. Now press the ‘Edit’ button in ‘Custom robots.txt’.add robots code here
  4. Now you can see the text area, to type the content. Copy and paste the code given below then click the Save Changes button.
User-agent: Mediapartners-Google
User-agent: *
Disallow: /search
Allow: /

How to block the link from Search Engines?

For example, If you want to stop robots from crawling this URL

To stop the crawling of this page, use this structure code

User-agent: *Disallow: p/about us

Alternative Method to create Robots.txt code:

Blogger XML Sitemaps
  • Here you will find the “Generate XML Sitemap for Blogger’ field. Paste your blogger URL, then click on “Generate Sitemap”. You will get the XML sitemap code.
ready XML sitemap for submission
  • Copy the code into notepad.
  • Next, Login to your blogger dashboard and go to “Settings › Search preferences › Crawlers and indexing”
paste the xml code here

Then paste it in ‘Custom robots.txt’ box and save the changes.

How to check your Robots code file?

You can check this file on your blog by adding “/robots.txt” at least to your blog URL in the address bar of your browser. Take a look at the below example for a demo.

(Replace with your BlogSpot URL)


Finally, I conclude by saying that ‘Robots.txt’ is used to control the indexing and crawl your blog content for search engines.

But do not try to add this code without any knowledge.

Do not hesitate to ask your questions from the comment box. I will help you to get better functionality to your blog.

Similar Posts


  1. ajay sharma says:

    hello, Satish sir I bookmark your blog for a new post but I want to know about indexing. I make a blog on Blogspot using my custom domain name and now I want to know by using For which domain name I create my robot.txt file for or for please sir help me. I am waiting for your answer.

    1. Hello Ajay, Thank you for landing here. This is the common problem for every Blogspot user. Use your custom domain name with complete form like “http://www. yourdomainname .in” in the This site generates complete XML sitemap code like

      XML sitemap of
      Copy the code and paste at the mentioned place in the blogspot dashboard.

Leave a Reply

Your email address will not be published. Required fields are marked *