What Is Crawl Budget and Why Does It Matter?
Crawl budget determines how much of your site Google actually crawls. Here's when it matters, what wastes it, and how to audit it.
15 May 2026 · 6 min read
Crawl budget is the number of URLs Googlebot will crawl on your site within a given period. It is not a fixed number - it is determined by two things: how quickly your server can respond without being overwhelmed (crawl rate limit), and how much Google thinks your content is worth crawling (crawl demand).
The practical question is: does crawl budget actually matter for your site?
For most small and medium sites: no. If your site has a few thousand pages and publishes content at a normal pace, Google will almost certainly crawl everything that matters. Crawl budget is a real concept, but it is often cited as a concern for sites where it genuinely is not one.
Where it does matter: large sites (tens of thousands of URLs), sites with significant numbers of redirects or errors, heavily paginated sites, and sites that have been through multiple migrations.
The two components of crawl budget
Crawl rate limit
This is how fast Googlebot can crawl without stressing your server. Google monitors response times and backs off if your server is slow or returning errors. A site that responds quickly and reliably gets crawled faster than one returning intermittent 5xx errors.
You can lower the crawl rate in Google Search Console if Googlebot is causing server strain - but raising it above Google's auto-detected limit is not possible.
Crawl demand
Google prioritises URLs it thinks are valuable and likely to change. Popular pages with many inbound links get crawled more often than obscure deep pages. Fresh content gets prioritised over static pages that have not changed in years.
If Google has low crawl demand for your site overall - because it is new, has few inbound links, or has historically had poor content quality - it will visit less frequently regardless of crawl rate limit.
What wastes crawl budget
These are the things that cause Googlebot to spend time on URLs that do not need crawling:
Redirect chains
Redirect chains make Googlebot follow multiple requests to reach a single final URL. Each hop is a separate request. Fix these by collapsing chains to single direct redirects. Use the Redirect Checker to verify individual URLs.
4xx pages being linked internally
A page returning a 404 or 410 that has internal links pointing to it will continue to be crawled until those internal links are removed or redirected. Crawly's Issues tab shows broken pages and every internal link pointing to them.
Non-indexable pages still being crawled
Pages with a noindex tag or a canonical pointing elsewhere will not be indexed, but Googlebot may still crawl them if they have internal links. Review your non-indexable pages and consider whether they need to be linked at all.
Duplicate URLs with parameters
URLs like /products?sort=price and /products?sort=name often return identical or near-identical content. Without proper canonicalisation, Google crawls all variations. Use canonical tags or filter URL parameters in Search Console.
Paginated content without clear signals
Deep pagination (pages 50, 100, 200 of a product listing) gets diminishing crawl attention. Make sure your most important content is linked from shallow pages, not buried at pagination depth 40.
How to audit crawl budget issues with Crawly
Run a full crawl and check these areas:
Response Codes tab - look at the breakdown of 3xx, 4xx, and 5xx responses. A high proportion of non-200 responses is a red flag for wasted crawl budget.
Issues tab - Redirect chains - every redirect chain is a crawl budget drain. Crawly lists all of them with their full hop sequences.
Issues tab - Non-indexable pages - pages that cannot be indexed but are still being crawled. Review whether they need internal links.
Issues tab - Near-duplicate content - exact-duplicate pages that should be consolidated via canonical tags.
Site Structure tab - see how deep your important pages sit. Content at depth 6+ rarely gets crawled as frequently as content at depth 2-3.
When to actually worry about crawl budget
You should investigate crawl budget if:
- Your site has more than 50,000 pages
- A large percentage of your response codes tab is 3xx or 4xx (say, more than 10%)
- You have been through a major migration and have many old redirect chains
- New content is taking weeks to appear in Google Search Console coverage
For most sites under these thresholds, the better investment is ensuring your important pages have strong internal links, load quickly, and have no indexability issues - these improve crawl demand rather than fighting crawl rate limits.
Crawl budget is real, but it is frequently over-cited as a problem for sites where it is not the bottleneck. The right approach is to audit your actual crawl health, fix genuine issues like redirect chains and 4xx errors, and focus on content quality that drives crawl demand.
For a complete walkthrough of auditing crawl health alongside other technical SEO signals, see the technical SEO audit guide.
Download Crawly to audit your site's crawl health today.