Today we are discussing a great and useful post on adding custom robots.txt file in blogger blog.A custom robots.txt file is a simple txt file in which the website owner use to write the commands for the web crawlers to what to crawler or not. That commands are written in different coding which can only be read by web crawlers. We can allow and disallow search crawlers to what area of our blog they have to crawl and what to not. It is one more step to make blog more SEO friendly. In old blogger interface there is not any option to add this txt file but in Blogger's new interface we can add it easily and I will guide you how to do it easily. So lets start this tutorial.
You can Check Your Blog Robots.Txt File by this link:
http://www.yourdomain.com/robots.txt
Enabling Robots.txt File:
So, This process is so easy just follow the simple steps below.
1.) Go To Blogger >> Settings >> Search Preferences
Look For Custom Robots.Txt Section In The Bottom and Edit It.
2.) Now a check box will appear tick "Yes" and a box will appear where you have to write the robots.txt file.
Enter this:
User-agent: Mediapartners-Google
User-agent: *
Disallow: /search?q=*
Disallow: /*?updated-max=*
Allow: /
Sitemap: http://www.yourdomain.com/feeds /posts/default?orderby=updated
Note :The first line "User-agent: Mediapartners-Google"is for Google AdSense. So if you are using Google AdSense in your blog then remain it same otherwise remove it.
3.) Click "Save Changes".
And You are Done! Now Lets take a view on the explanation of each of the line:
User-agent: Mediapartners-Google :This is a first command which is used for Google AdSense enabled blogs if you are not using Google AdSense then remove it. In this command, we're telling to crawl all pages where Adsense Ads are placed!
User-agent: * :Here theUser-agentis calling the robot and*is for all the search engine's robots like Google, Bing etc.
Disallow: /search?q=* :This line tells the search engine's crawler not to crawl the search pages.
Disallow: /*?updated-max=* :This one disallows the search engine's crawler to do not index or crawl label or navigation pages.
Allow: / :This one allows to index the homepage or your blog.
Sitemap :So this last command tells the search engine's crawler to index the every new or updated post.
You can also add your own command to disallow or allow more pages. For Ex: if you want to disallow or allow a page then you can use this commands:
To Allow a Page:
Allow: /p/contact.html
To Disallow:
Disallow: /p/contact.html
Final Words:
I hope you can learn to add robots.txt file in your blogger blog! If you face any difficulty then please let me know in comments. Its your turn to say thanks in comments and keep sharing this post till then Peace, Blessings and Happy Adding.
Like Our Fb Page To Get Updates.
0 comments:
Post a Comment