- Joined
- Aug 28, 2018
- Messages
- 10
- Likes
- 5
- Degree
- 0
Hi guys,
While working on e-commerce websites we block junk/worthless URLs in the robots.txt file so Google won't crawl those pages and instead crawl our important pages(product pages, new content etc).
A client's website had ~600 pages indexed that were nothing but blog tags. We had it blocked adding a "Disallow: /tag/" rule in robots.txt. Big mistake. Turns out it won't be blocked unless the rule is "Disallow: /blog/tag/" . Luckily, I came across this and made the necessary changes.
How can you check if the directive that you have added in robots.txt is working correctly? Use the robots.txt tester in Google Search Console or you may even use a 3rd party tool such as https://technicalseo.com/seo-tools/robots-txt/
TL;DR - Always cross-check any new rule you might be adding to your robots.txt file in GSC's robots.txt tester or a 3rd party tool.
While working on e-commerce websites we block junk/worthless URLs in the robots.txt file so Google won't crawl those pages and instead crawl our important pages(product pages, new content etc).
A client's website had ~600 pages indexed that were nothing but blog tags. We had it blocked adding a "Disallow: /tag/" rule in robots.txt. Big mistake. Turns out it won't be blocked unless the rule is "Disallow: /blog/tag/" . Luckily, I came across this and made the necessary changes.
How can you check if the directive that you have added in robots.txt is working correctly? Use the robots.txt tester in Google Search Console or you may even use a 3rd party tool such as https://technicalseo.com/seo-tools/robots-txt/
TL;DR - Always cross-check any new rule you might be adding to your robots.txt file in GSC's robots.txt tester or a 3rd party tool.