404 Errors and SEO: What They Cost You
404 errors hurt SEO, but not always in the way most articles describe. A single broken page rarely tanks a site. The damage accumulates quietly: crawl budget wasted on dead URLs, inbound links pointing nowhere, users bouncing before they convert, and Google drawing slow conclusions about site quality over time.
The question worth asking is not whether 404s matter. They do. The question is which ones matter, how much, and what to do about them in a way that actually moves the needle rather than just filling a spreadsheet.
Key Takeaways
- 404 errors do not directly cause ranking penalties, but they erode crawl efficiency, waste link equity, and signal poor site maintenance to Google over time.
- Not all 404s are equal: a broken page with zero inbound links and no traffic history is largely irrelevant. A broken page that held strong backlinks is a real problem.
- The most commercially damaging 404s are those sitting on URLs that used to convert, rank, or attract links. Prioritise those first.
- 301 redirects are the correct fix for pages that have moved or been replaced. Leaving a 404 in place is only acceptable when the page has no link equity and no search value.
- Crawl budget is finite on large sites. A high volume of 404s forces Googlebot to spend time on dead ends instead of indexing pages that matter.
In This Article
- Why 404 Errors Matter More on Some Sites Than Others
- What Google Actually Does When It Finds a 404
- The Link Equity Problem Is the One Worth Losing Sleep Over
- How to Find 404 Errors Before They Compound
- Soft 404s: The Problem That Is Harder to Spot
- When a 404 Is the Right Answer
- The User Experience Dimension
- Building a Process That Prevents 404s From Accumulating
- The Honest Summary
Why 404 Errors Matter More on Some Sites Than Others
I have audited sites across thirty industries over two decades, and the relationship between 404 errors and ranking performance is never uniform. A ten-page brochure site with two broken links is a housekeeping issue. An e-commerce site with four thousand 404s generated by discontinued product lines is an active drag on crawl efficiency and a graveyard for link equity that once supported rankings.
Site size and link profile are the two variables that determine how seriously you should treat your 404 problem. On a large site, Googlebot has a crawl budget, a rough allocation of how many pages it will process in a given crawl cycle. If a significant proportion of those crawls land on dead pages, Google is spending resources on nothing. Pages that should be indexed, pages that carry commercial value, get crawled less frequently as a result. That is the real cost, not a penalty flag, but a quiet deprioritisation of the pages you actually want Google to find.
On smaller sites, the crawl budget concern largely disappears. But the link equity problem does not. If you have a page that has been deleted and it held backlinks from authoritative domains, those links are now pointing at a dead end. The equity they carried, the ranking signal they provided, is effectively lost until you redirect that URL to something relevant.
If you want to understand how 404s fit into your broader technical and strategic SEO picture, the Complete SEO Strategy hub covers the full landscape, from crawlability to content architecture to link building.
What Google Actually Does When It Finds a 404
Google does not penalise sites for having 404 errors. That is worth stating plainly because a lot of agency pitches and technical audit reports imply otherwise. Google has been clear that 404s are a normal part of the web. Pages get deleted. URLs change. That is expected.
What Google does do is remove 404 pages from its index, eventually. The timeline varies. A page that was previously indexed and ranked will not disappear from search results the moment it returns a 404. Google tends to recrawl it several times before deindexing it, which means there is usually a window to implement a redirect before rankings are fully lost. That window is not infinite, and it is not predictable, so treating it as a safety net is a mistake.
There is also a subtler signal at play. Google uses a range of quality indicators when evaluating a site, and a high proportion of broken pages is one of them. It does not trigger an algorithmic penalty in the way a manual action does, but it contributes to a broader picture of site health. When I was turning around a loss-making agency, one of the first things I did before pitching any new client was a technical audit of their site. 404s were never the headline finding, but they were consistently part of a pattern that included slow load times, thin content, and poor internal linking. Sites with lots of 404s tended to have other problems too. The 404s were a symptom as much as a cause.
The Link Equity Problem Is the One Worth Losing Sleep Over
If there is one aspect of 404 errors that genuinely warrants commercial attention, it is lost link equity. Backlinks remain one of the most influential ranking signals in Google’s algorithm. When a page that holds backlinks returns a 404, those links stop passing value. The referring domains still exist. The links still exist. But the destination is gone, and the equity evaporates.
The fix is straightforward: a 301 redirect from the broken URL to the most relevant live page. Not to the homepage by default, which is a lazy fix that passes less equity and frustrates users, but to the closest equivalent content. If you deleted a product page and replaced it with an updated version at a new URL, redirect the old URL to the new one. If the content is genuinely gone with no equivalent, redirect to the most relevant category or parent page.
The prioritisation question matters here. I have seen teams spend weeks auditing and redirecting hundreds of 404s with no backlinks, no traffic history, and no ranking history. That is resource allocation theatre. The pages worth fixing first are those that can be identified through a backlink audit: find the broken URLs that have inbound links, rank them by the authority of those links, and work through them in order. Everything else is secondary.
Tools like Moz’s SEO resources cover link equity and crawl health in depth and are worth consulting when you are building an audit framework. The mechanics of how link equity flows through a site are well-documented, and understanding them properly changes how you prioritise fixes.
How to Find 404 Errors Before They Compound
There are three reliable methods for surfacing 404 errors, and a thorough audit uses all three because each catches different things.
The first is Google Search Console. The Coverage report shows URLs that returned errors during Google’s crawl, including 404s. This is the most commercially relevant source because it shows you what Google has actually tried to access. If Google is crawling a URL and finding a 404, that is a real crawl budget and indexation issue. Search Console also shows you which 404s were previously indexed, which helps you prioritise by former ranking value.
The second is a crawl tool. Screaming Frog, Sitebulb, or similar tools will crawl your site and follow internal links, flagging any that resolve to a 404. This is particularly useful for catching broken internal links, pages that your own site is pointing to that no longer exist. Internal 404s are often overlooked in favour of external ones, but they create poor user experiences and waste crawl budget in exactly the same way.
The third is a backlink audit tool. Ahrefs, Semrush, or Moz will show you which of your URLs are being linked to from external sites. Cross-reference this list against your 404s to find the broken pages with inbound link equity. This is the list that should drive your redirect priority queue.
When I was growing an agency from twenty to a hundred people, we built a monthly technical audit process that covered all three of these sources. It was not glamorous work, but it was the kind of systematic maintenance that separated sites that held their rankings from sites that gradually drifted down. The clients who saw consistent organic growth were almost always the ones who treated technical hygiene as an ongoing process rather than a one-time project.
Soft 404s: The Problem That Is Harder to Spot
A standard 404 returns an HTTP status code of 404, which tells Google unambiguously that the page does not exist. A soft 404 is more insidious. The page returns a 200 status code, meaning it appears to load successfully, but the content is either empty, near-empty, or a generic error message that does not serve any user intent. Google identifies these as effectively non-existent pages and treats them similarly to hard 404s, but they are harder to catch because your crawl tool sees a 200 and moves on.
Soft 404s are common in e-commerce. A product goes out of stock, the page stays live but displays nothing except “this product is unavailable,” and Google eventually recognises it as thin content with no value. They also appear on sites with search functionality where an empty search result returns a 200 status. Google Search Console flags soft 404s separately in the Coverage report, which is one reason that tool should be your starting point for any 404 audit rather than a crawl tool alone.
The fix for soft 404s depends on the cause. If the page genuinely has no content and no future, return a proper 404 or 301. If the product is temporarily out of stock and will return, keep the page live with meaningful content: related products, a notification signup, information about the item. That approach preserves any ranking equity the page has built and gives users something useful rather than a dead end.
When a 404 Is the Right Answer
Not every 404 needs to be fixed with a redirect. There is a reflexive instinct in SEO to redirect everything, and it leads to redirect chains, irrelevant destination pages, and a lot of wasted effort. Sometimes a 404 is the correct response.
If a page has no backlinks, no traffic history, no ranking history, and no internal links pointing to it, a 404 is fine. Google will deindex it in due course and move on. Creating a redirect to an irrelevant page just to avoid the 404 status is not an improvement. It can actually confuse users and dilute the relevance signal of the destination page.
The same logic applies to pages that were intentionally removed because they contained outdated, inaccurate, or low-quality content. If you deleted a page because it was thin and unhelpful, redirecting it to another page does not improve the situation. It just moves the problem. A clean 404 followed by eventual deindexation is the right outcome.
Where 404s become a problem is when they are accidental rather than intentional. A URL structure change that was not accompanied by redirects. A site migration that broke hundreds of links. A CMS update that altered URL formats without preserving the old ones. These are the scenarios that cause real damage, and they are almost always avoidable with proper planning.
I have seen site migrations go badly wrong more times than I can count. The pattern is always the same: the technical work gets done, the new site launches, and then someone checks Search Console three weeks later and finds hundreds of 404s because the redirect mapping was incomplete. The fix is always more expensive than the prevention would have been. Building a comprehensive redirect map before a migration, and testing it before launch, is not optional work. It is the difference between a migration that maintains rankings and one that costs six months of recovery time.
The User Experience Dimension
SEO discussions about 404 errors tend to focus on crawl budget and link equity, which are legitimate concerns. But there is a user experience dimension that often gets less attention and is arguably more commercially direct.
When a user clicks a link, whether from search results, a social post, an email, or another website, and lands on a 404 page, the default response is to leave. Most 404 pages are unhelpful by design. They confirm that something went wrong and offer little else. The user does not know whether the content has moved, whether the site is broken, or whether they should try searching again. The most common outcome is that they go back to Google and find a competitor.
A well-designed 404 page reduces that damage. It acknowledges the error, offers navigation options, includes a search bar, and ideally suggests related content. It does not recover the situation entirely, but it recovers some of it. Understanding how users behave when they hit a dead end is worth investing in. Tools like Hotjar can help you see where users are dropping off and whether 404 pages are contributing to abandonment patterns you have not yet identified.
The commercial argument for fixing 404s is not just about rankings. It is about not wasting the traffic you already have. If you are spending budget on paid search, content creation, or link building, and some of that traffic is landing on broken pages, you are paying to send people to a dead end. That is a straightforward efficiency problem that does not require any SEO theory to understand.
Building a Process That Prevents 404s From Accumulating
Reactive 404 management, finding and fixing them after the fact, is necessary but insufficient. The goal should be a process that prevents them from accumulating in the first place, or at least catches them quickly enough that the damage is minimal.
That means a few things in practice. First, any time a page is deleted or a URL is changed, a redirect should be created as part of the same workflow, not as a separate task to be done later. Later almost never happens. Building the redirect into the deletion process removes the dependency on someone remembering to do it.
Second, Search Console should be reviewed on a regular cadence, not just when something goes wrong. The Coverage report will surface new 404s as Google finds them. Monthly is the minimum. Weekly is better for large, frequently updated sites.
Third, before any significant site change, whether a redesign, a migration, a CMS upgrade, or a URL restructure, a redirect audit should be part of the project scope. This is not a nice-to-have. It is the kind of technical due diligence that separates agencies and in-house teams that hold rankings through change from those that spend the following year recovering them.
The Moz blog has covered crawl health and technical maintenance in detail over the years, and their frameworks for thinking about site health are worth bookmarking if you manage a site of any complexity. The principles are consistent regardless of site type or industry.
If you are working through the broader mechanics of how technical factors, content, and links interact in a complete SEO strategy, the Complete SEO Strategy hub on The Marketing Juice covers each of these areas in depth and is worth reading alongside the technical specifics here.
The Honest Summary
404 errors are not the catastrophic ranking killer that some technical audits imply, and they are not the harmless non-issue that some SEOs dismiss them as. The reality is more granular than either position.
The 404s that matter commercially are those sitting on URLs with backlinks, those generated by botched migrations, and those that accumulate at scale on large sites where crawl budget is a real constraint. Those deserve time and resource. The 404s that do not matter are orphaned pages with no link equity and no traffic history. Redirecting those is busywork.
The discipline is in telling the difference. That requires a backlink audit, a crawl, and a look at Search Console, not just a count of how many 404s exist. A site with fifty 404s on high-value URLs has a more serious problem than a site with five hundred 404s on pages that never mattered. The number is almost irrelevant. The context is everything.
Most of the time, when I see a team spending disproportionate energy on 404 remediation, it is because someone ran a crawl, saw a red number, and treated it as an emergency. The better question is always: which of these actually affect rankings, traffic, or conversions? Start there, and the priority list gets much shorter and much more useful.
About the Author
Keith Lacy is a marketing strategist and former agency CEO with 20+ years of experience across agency leadership, performance marketing, and commercial strategy. He writes The Marketing Juice to cut through the noise and share what works.
