Web Content Audit: What to Keep, Cut, and Fix

A web content audit is a systematic review of every page on your site, assessed against traffic, engagement, conversions, and strategic fit. Done properly, it tells you which content is earning its place, which is quietly dragging your site down, and which needs work before it can do either.

Most sites have far more content than they think they need, and far less content that actually works. An audit is how you close that gap.

Key Takeaways

  • A content audit is a commercial exercise, not a housekeeping task. Every decision should connect to traffic, leads, or revenue.
  • Most sites have three categories of content: what works, what can be fixed, and what should be removed. The hardest call is the middle one.
  • Thin content and cannibalisation are the two most common audit findings, and both are fixable without starting from scratch.
  • Audit frequency matters. Running one every 18-24 months is the minimum for a site with active content production.
  • The audit is only as useful as the decisions that follow it. A spreadsheet without a prioritised action plan is just documentation.

I have run audits on sites with 50 pages and sites with 50,000. The scale changes the logistics but not the logic. You are always asking the same questions: is this content doing a job? If not, why not? And what is the most commercially sensible thing to do about it?

Why Most Content Audits Fail Before They Start

The most common reason a content audit produces nothing useful is that it starts with the wrong question. Teams ask “what do we have?” when they should be asking “what is working, and what is costing us?”

I picked up this habit during a turnaround I ran at an agency that had been producing content for clients at volume for years. Nobody had ever stopped to ask whether any of it was actually performing. When we finally audited one client’s site, we found that roughly 70% of their indexed pages had received fewer than ten organic visits in the previous twelve months. Not ten per day. Ten total. That content was not neutral. It was creating crawl budget waste, diluting topical authority, and in some cases actively cannibalising better pages.

The fix was not to produce more content. The fix was to stop, assess, and consolidate. Within six months of acting on the audit findings, organic traffic to the pages we kept was up considerably, and the client was spending less on content production, not more.

If you want to understand how content strategy connects to outcomes like this, the broader thinking sits within our Content Strategy and Editorial hub, which covers everything from editorial planning to performance measurement.

The other reason audits fail is scope creep. Teams try to assess everything at once and end up with a spreadsheet so large it becomes unusable. Prioritise by traffic tier first. The top 20% of pages by traffic will almost always account for 80% or more of your organic value. Start there.

What Data Do You Actually Need for a Content Audit?

You do not need a dozen tools. You need clean data from a small number of reliable sources, interpreted with some critical thinking about what the numbers actually mean.

The core data set for any content audit includes organic traffic (sessions and users), average position and impressions from Google Search Console, engagement metrics such as time on page and scroll depth, conversion data tied to specific pages where available, backlink counts for pages with any link equity, and a crawl export to identify technical issues like thin content, duplicate titles, or broken internal links.

Tools like Semrush’s content audit functionality can pull much of this together in one place, which is useful at scale. But I would always cross-reference with Google Search Console directly. Third-party tools give you a useful approximation of search performance. The actual impression and click data comes from Google, and the two rarely match perfectly.

One thing I always flag with teams: engagement metrics are a perspective on user behaviour, not a verdict on content quality. A page with a short average session duration might be performing brilliantly because it answered the question immediately and the user left satisfied. Context matters. Do not let a single metric make the decision for you.

If you are running AI-assisted workflows for content briefs or gap analysis alongside your audit, Moz’s thinking on AI content briefs is worth reviewing for how to structure that process without losing editorial judgement.

How to Categorise Every Page: Keep, Consolidate, or Cut

Once you have your data, the audit becomes a triage exercise. Every page gets one of three labels: keep, consolidate, or cut. The criteria for each should be agreed before you start, not invented page by page.

Keep applies to pages that are generating meaningful organic traffic, ranking in positions where improvement is realistic, earning backlinks, or converting at a rate that justifies their existence. These pages may still need optimisation, but they are doing a job. Leave them largely intact and schedule a review in six to twelve months.

Consolidate applies to pages that cover similar territory, often the result of years of content production without a clear taxonomy. Cannibalisation is the specific problem here: two or more pages competing for the same keyword, splitting authority and confusing search engines about which one to rank. The fix is to merge the weaker into the stronger, redirect the old URL, and update the surviving page to cover the combined ground properly. This is almost always more effective than trying to differentiate pages that were never meaningfully different to begin with.

Cut applies to pages with no traffic, no links, no conversions, and no strategic reason to exist. Thin pages created to target long-tail keywords that never materialised. Old campaign landing pages left live after the campaign ended. Blog posts that were topically irrelevant from day one. These pages do not help users and they do not help search engines understand what your site is about. Remove them and redirect where appropriate, or return a 410 if there is nothing to redirect to.

The middle category is where most of the effort goes, and where most teams lose momentum. Consolidation requires editorial judgement, not just data interpretation. Someone has to decide which version of a topic is the canonical one, what the merged page should say, and whether it is worth the work. That decision should be made by a person who understands both the content and the commercial objective, not delegated entirely to a tool.

The Thin Content Problem Is Bigger Than You Think

Thin content is the single most common finding in any audit I have run, and it is also the most misunderstood. Teams tend to think of thin content as short content. That is not quite right. A 300-word page that answers a specific question clearly and completely is not thin. A 1,500-word page that says nothing useful, repeats itself, and exists primarily to include a keyword is thin by any meaningful definition.

The test I use is simple: if a user landed on this page from a search result, would they find what they came for, or would they leave immediately and try somewhere else? If the honest answer is the latter, the page is thin regardless of its word count.

Thin content is partly a production problem. When agencies or in-house teams are measured on volume, content quality tends to decline over time. I saw this pattern repeatedly when I was running agency teams. Output targets create incentives to publish, not to be useful. An audit is one of the few mechanisms that forces a reckoning with that accumulated mediocrity.

The Content Marketing Institute’s definition of content marketing centres on creating genuinely useful content for a defined audience. That is the standard every page should be held to during an audit. Not “did we publish this?” but “does this serve someone?”

Cannibalisation: The Problem That Compounds Over Time

Keyword cannibalisation happens when multiple pages on your site target the same or closely related search intent. Search engines have to choose between them, and they do not always choose the one you want. The result is that neither page ranks as well as a single consolidated page would.

The longer a site has been producing content without a clear keyword architecture, the worse this problem tends to be. I have audited sites where the same topic had been covered in five or six variations across blog posts, resource pages, and product pages, all fighting each other for the same ranking positions. The fix required consolidating several of those pages into one authoritative piece, redirecting the others, and updating the internal linking structure to signal clearly which page was the primary one.

Identifying cannibalisation starts with a keyword mapping exercise run alongside the audit. Export your Search Console data, filter by query, and look for queries where multiple URLs are appearing in impressions data. That is your cannibalisation list. Cross-reference it with your crawl data to understand the relationship between the competing pages, and then make a decision about which one survives.

One thing worth understanding here is how user-generated content and community pages can sometimes create cannibalisation at scale on larger sites. Search Engine Land’s analysis of UGC and search is a useful reference if your site has forum threads, Q&A sections, or comment-heavy pages in the mix.

How to Build an Audit Action Plan That Actually Gets Executed

The audit findings are not the deliverable. The action plan is the deliverable. A spreadsheet that categorises 400 pages and then sits in a shared drive for six months has produced no value for anyone.

The action plan needs three things: prioritisation, ownership, and deadlines. Prioritisation means ranking actions by expected impact, not by ease of execution. The temptation is always to start with the quick wins because they feel productive. But if your highest-traffic pages are underperforming and your consolidation opportunities are significant, those should come first regardless of how much work they require.

Ownership means a named person responsible for each action, not a team or a function. “Marketing team to update” is not ownership. “Sarah to consolidate the three pricing pages and submit for review by the 15th” is ownership. This sounds basic, but it is the difference between audits that produce change and audits that produce documentation.

Deadlines need to be realistic. I have seen audit action plans with 200 items and a 30-day completion window. That is not a plan, it is a wish list. Tier your actions: high priority in the first 30 days, medium priority in 60 to 90 days, lower priority in the following quarter. Review progress at each milestone and adjust based on what you are seeing in the data.

If you are working with AI tools to accelerate the content update process, Moz’s overview of AI for SEO and content marketing covers where those tools add genuine efficiency and where human editorial judgement remains essential. The short version: AI is useful for research and drafting at scale, but it does not replace the strategic decision-making that an audit requires.

How Often Should You Run a Content Audit?

The answer depends on how actively you are producing content and how competitive your space is. For most B2B sites with consistent content production, a full audit every 18 to 24 months is a reasonable baseline. For sites in fast-moving categories where search intent shifts quickly, or for sites producing content at high volume, annual audits are more appropriate.

Between full audits, a lighter quarterly review of your top 50 pages by traffic is worth building into your editorial calendar. This catches emerging cannibalisation, identifies pages that have dropped in performance and need attention, and keeps the team honest about whether new content is actually filling gaps or just adding to the pile.

One pattern I have seen repeatedly: teams run an audit, act on the findings, see improvement, and then go back to producing content at volume without any ongoing monitoring. Eighteen months later they need another audit to address the same categories of problem. The audit is not a one-time fix. It is a checkpoint in an ongoing editorial discipline.

There is also a case for trigger-based audits outside the regular cycle. A significant algorithm update, a major site migration, a rebranding exercise, or a sustained drop in organic traffic are all reasons to run an audit ahead of schedule rather than waiting for the next planned review.

The Commercial Case for Doing This Properly

I have judged the Effie Awards, which measure marketing effectiveness against commercial outcomes. The discipline that makes Effie-winning work different from ordinary work is the same discipline that makes a content audit useful: a clear connection between what you are doing and what it is supposed to achieve.

Content audits are not exciting. They do not generate the same internal enthusiasm as a new campaign or a redesign. But they are one of the highest-return activities available to a content team, because they extract value from work that has already been done. You are not spending more. You are making what you have spent work harder.

When I grew the agency I ran from 20 to 100 people, one of the disciplines we built early was a regular review of client content performance. Not because clients asked for it, but because it was the most honest way to show that what we were doing was working. If you cannot point to pages that are performing and explain why, and identify pages that are not performing and explain what you are doing about it, you are not managing content. You are publishing it.

A well-executed content audit gives you that clarity. It tells you what you have, what it is doing, and what decisions you need to make. That is the foundation of any content strategy worth the name.

If you want to put the audit findings into a broader editorial framework, the Content Strategy and Editorial hub covers the planning, governance, and measurement disciplines that sit around the audit process and make the findings actionable over time.

About the Author

Keith Lacy is a marketing strategist and former agency CEO with 20+ years of experience across agency leadership, performance marketing, and commercial strategy. He writes The Marketing Juice to cut through the noise and share what works.

Frequently Asked Questions

What is a web content audit?
A web content audit is a structured review of every page on a website, evaluated against traffic, engagement, conversions, and strategic relevance. The goal is to identify which content is performing, which can be improved, and which should be removed or consolidated with other pages.
How long does a content audit take?
For a site with up to 200 pages, a thorough audit typically takes two to four weeks, including data collection, page-level assessment, and action planning. Larger sites take longer, though prioritising by traffic tier means you can complete the most commercially important work in the first phase without waiting for the full audit to finish.
What tools do you need to run a content audit?
The core tools are Google Search Console for organic performance data, Google Analytics or an equivalent for engagement and conversion data, a crawl tool such as Screaming Frog or Sitebulb to inventory pages and identify technical issues, and a backlink tool to assess link equity. Platforms like Semrush can consolidate much of this into one workflow, but always cross-reference with first-party data from Search Console.
Should you delete underperforming content or update it?
It depends on why the content is underperforming. Pages with no traffic, no links, and no strategic relevance are usually better removed or redirected. Pages that cover a relevant topic but are thin, outdated, or cannibalising a stronger page are usually better consolidated or updated. Deletion is not always the right answer, but keeping content purely because it exists is not a content strategy.
How is a content audit different from an SEO audit?
An SEO audit focuses primarily on technical factors: crawlability, indexation, site speed, structured data, and on-page optimisation signals. A content audit focuses on the editorial and strategic value of individual pages. The two overlap, particularly around thin content and cannibalisation, but a content audit asks broader questions about whether each page serves a user need and supports a commercial objective, not just whether it is technically sound.

Similar Posts