Crawl budget is the time or amount of pages search engines spend crawling your website on a daily basis.
Because of the enormous size of the internet, search engines have to divide their attention across all websites and prioritize their crawl effort. Assigning each website crawl budget helps them with this.
Crawl budget is a common term within the SEO workfield. Often Crawl budget is referred to crawl space or crawl time.
You want search engines to find as much of your indexable pages, and you want them to do that as quickly as possible.
If you’re wasting crawl budget, it’s unlikely search engines will be able to. This means that part of your website is left undiscovered by search engines. This part of your website is therefore inaccessible to potential visitors searching in search engines. Naturally, this hurts your success. Therefore crawl budget optimization should be high on your ToDo-list.
Google’s most transparent about their crawl budget for your website. If you have your website verified in Google Search Console, you can get some insight in your website’s crawl budget (only within Google of course).
Log in to Google Search Console, choose a website and go to
Crawl Stats. There you see the amount of pages crawled per day by Google.
In the example above we see that the average daily crawl budget is 27 pages. So in theory, if this average crawl budget stays the same you would have a monthly crawl budget of 27 pages x 30 days = 810 pages.
Alternatively, you can analyze your webserver’s log files to get insight in search engines’ crawl behaviour of your website.
You first need to identify where crawl budget is being wasted. Below we cover some of the most common reasons for wasted crawl budget.
Common reasons for wasted crawl budget are:
Broken links are a dead end for search engines. The same goes for chains of redirects which are too long, search engines may then break off their crawl process. There’s a limit as to how many redirects in a chain search engines and browsers can take.
Similar to browsers, Google seems to follow a maximum of five chained redirects. It’s unclear how well other search engines deal with subsequent redirects, but we strongly advise you to avoid chained redirects entirely and keep usage of redirects to a minimum.
It’s clear that by fixing broken links and redirecting links you can quickly regain some of your wasted crawl budget. Besides regaining crawl budgets, you’re also significantly improving visitor’s user experience. Redirects, and chains of redirects in particular, cause longer page load time and thereby hurt user experience.
To make finding broken and redirecting links easy, within ContentKing we’ve dedicated a special issue for this. Go to
Links to find out if you are wasting crawl budgets because of fault links. Update all links so they link to an indexable URL.
When pages have high load times, or when pages even time-out, search engines can visit fewer pages within the allotted crawl budget for your website. Besides, high page load times and time-outs hurt the user experience of your visitors significantly, resulting in a lower conversion rate.
Google reports on page load time in both Google Analytics (under
Site Speed) and Google Search Console under
Google Search Console and Bing Webmaster Tools both report on page time-outs. In Google Search Console this can be found under
Crawl Errors and in Bing Webmaster Tools under
Reports & Data >
Check regularly if your pages are loading fast enough and take action immediately if they don’t. Fast loading pages are vital to your online success.
All URLs included in XML sitemaps should be indexable. Especially for large websites, search engines heavily rely on XML sitemaps to find all your pages. If your XML sitemaps are cluttered with pages which for instance don’t exist anymore or redirect, you’re wasting crawl budget. Regularly check your XML sitemap for non-indexable URLs which don’t belong in there. Google Search Console reports on XML sitemap issues under
Sitemaps. Bing Webmaster Tools does the same under
Configure My Site >
One of the crawl budget optimization best practices is to split your XML sitemaps up into smaller sitemaps. You can for instance create XML sitemaps for each of your website section. If you’ve done that, you can quickly determine if there’s any issues going on in certain sections of your website. Say your XML sitemap for section A contains 500 links, and 480 are indexed then you’re doing pretty good. But if your XML sitemap for section B contains 500 links and only 120 are indexed that’s something to look into. You may have included a lot of non-indexable URLs in the XML sitemap for section B.
If your website contains a high amount of non-indexable pages which are accessible to search engines you’re basically keeping search engines busy sifting through irrelevant pages.
In order to find out if you have a high amount of non-indexable pages, look up the total amount of pages crawlers found within your website. You can easily do this using ContentKing.
In ContentKing the total amount of pages crawled is shown on Pages overview at the top.
Compare this amount to the amount of pages indexed using a
In this example there were over 200,000 pages found, but actually there were only 30,000 indexed by Bing. There can be two explanations for this:
How pages within your website link to one another plays a big role in crawl budget optimization. We call this the internal link structure of your website. Backlinks aside, pages that have few internal links get much less attention from search engines than pages that are linked on a lot of pages.
Prevent a very hierarchical link structure with pages in the middle having little links. These pages will not be crawled that often. It’s even worse for pages at the bottom of the hierarchy: becauses of the limited amount of links they may very well be neglected by search engines.
Make sure your most important pages have plenty of internal links. Pages that have recently been crawled typically rank better in search engines. Keep this in mind, and adjust your internal link structure for this.
For example, if you have a blog article dating from 2011 which drives a lot of organic traffic make sure to keep linking to it from other content. Because you produced a lot more blog articles over the years, that article from 2011 is automatically being pushed down in your website’s internal link structure.
There are a few common ways to waste crawl budget that we come across very often. We cover them below, including ways to regain your crawl budget.
The first two are also called ‘crawl traps’. Crawl traps are issues within websites that create a virtually endless amount of URLs which are accessible to search engines. These in particular pose a big issue to your crawl budget.
Within a product filter each filter criterion has at least two values. Combining these criteria enables a visitor to drill down in the product offering. This is very useful for a visitor, but if the product filter is crawlable however, an virtual endless amount of URLs are generated. This forms a crawl trap for search engines.
Generally, you don’t want to have search result pages from your website’s internal search functionality getting crawled and subsequently indexed by search engines. This causes duplicate content easily, and should in most cases be avoided. Assuming you don’t want your internal search result pages to be crawled and indexed, be sure to inform search engines that these internal search result pages shouldn’t be accessed.
Solution: Instructing search engines not to access these internal search result pages can easily be done using your robots.txt. We’ve included an example robots.txt for a WordPress website which takes care this. If this is not possible for you, consider using the URL parameter handling settings in Google Search Console and Bing Webmaster Tools to instruct Google and Bing what pages not to crawl.
A less technical reason to have a high amount of non-indexable pages could be that you used to like working with tags for your blog articles back in 2010. You’re an avid writer, so over time you’ve accumulated thousands of tags. When you read about Google Panda you decided to put a meta robots noindex directive on these tags pages, which fixed only possible indexing issues.
While preventing possible indexing issues, it created crawl issues: search engines are crawling these pages but find time after time that they can’t index these pages, so they ignore them. But they do spend your valuable crawl budget on those tag pages.
Solution: Instructing search engines not to access these tag pages anymore can easily done with your robots.txt file.
Matt Cutts said:
The best way to think about it is that the number of pages that we crawl is roughly proportional to your PageRank. So if you have a lot of incoming links on your root page, we’ll definitely crawl that. Then your root page may link to other pages, and those will get PageRank and we’ll crawl those as well. As you get deeper and deeper in your site, however, PageRank tends to decline.
Even though Google has abandoned updating PageRank values of pages publicly, we think (a form of) PageRank is still apart of their algorithm. Since PageRank is a misunderstood and confusing term, let’s call it page authority. The take-away here is that Matt Cutts basically says: there’s a pretty strong relation between page authority and crawl budget.
So, in order to increase your website’s crawl budget you need to increase the authority of your website. A big part of this is done by earning more links from external websites.
Google has indicated there’s a strong relation between page authority and crawl budget. The more authority a page has, the more crawl budget it has. Simply put, more authority equals more crawl budget.
Yes, and it’s important to understand the differences between indexing issues and crawl issues.
The canonical URL and meta robots tags send a clear signal to search engines what page they should show in their index, but it does not prevent them from crawling those pages.
You can use the robots.txt file and the nofollow link relation for dealing with crawl issues.