Market Validation: Stop Building What Nobody Wants
Market validation is the process of confirming that a real demand exists for a product, service, or idea before committing significant resources to it. Done properly, it answers one question above everything else: will people actually pay for this, or do they just say they will?
The definition sounds simple. The execution is where most teams go wrong, because validation is not a checkbox. It is a discipline of honest inquiry that most organisations are structurally incentivised to skip.
Key Takeaways
- Market validation is about confirming real demand, not collecting reassuring opinions. There is a meaningful difference between the two.
- Most validation fails because teams ask questions designed to confirm what they already believe, not questions designed to challenge it.
- Behavioural signals, search data, and purchase intent are stronger validation evidence than survey responses or interview enthusiasm.
- Validation should happen before you build, not after you have already committed. The later you validate, the more expensive the correction.
- A validated market does not guarantee success. It removes one major category of risk. The commercial work still has to follow.
In This Article
- Why Most Validation Work Produces the Wrong Answer
- What Counts as Evidence and What Does Not
- The ICP Problem: Validating With the Wrong People
- Qualitative Validation: What It Can and Cannot Tell You
- Using Indirect Signals to Validate Without Asking Directly
- The Landing Page Test: Cheap, Fast, and Honest
- Pain Points as Validation Anchors
- Validation Across the Business: Not Just for New Products
- When Validation Says No
If you are building out a research capability or trying to make better strategic decisions with limited budget, the broader market research hub covers the full landscape, from competitive intelligence to qualitative methods to search-based demand analysis.
Why Most Validation Work Produces the Wrong Answer
I have sat in more strategy sessions than I can count where someone presents validation evidence that is, essentially, a curated set of people saying nice things. Friendly customers who did not want to be discouraging. Prospects who were polite but had no intention of buying. Internal stakeholders who had already decided the answer and worked backwards from it.
This is not validation. It is confirmation bias with a research budget attached.
Genuine market validation requires you to design the process specifically to find reasons the idea might fail. If your research method cannot surface a negative result, it is not validation. It is marketing theatre.
The structural problem is that validation often happens inside organisations where someone senior has already championed the idea. Nobody wants to be the person who brings back bad news. So the research gets shaped, consciously or not, to deliver a green light. Teams ask leading questions, select sympathetic respondents, and interpret ambiguous signals optimistically.
Good validation design works against this tendency. It treats scepticism as the default and looks for evidence that changes that position, rather than starting from enthusiasm and looking for confirmation.
What Counts as Evidence and What Does Not
There is a hierarchy of validation evidence, and most teams operate at the bottom of it.
At the weakest end, you have stated intent. Someone says they would buy this product. Someone says they find the concept interesting. Someone says they would switch from their current provider. These signals are almost worthless in isolation. People are terrible at predicting their own behaviour, especially when there is no cost to expressing enthusiasm.
Slightly stronger is qualified stated intent, where you add friction to the question. Not “would you buy this?” but “would you pay £X for this today, given that you already use Y?” Adding specificity, price, and competitive context filters out the polite responses and forces people to actually think about the trade-off.
Stronger still is behavioural evidence. What are people searching for? What are they complaining about in forums and review sites? What are they spending money on already? Behavioural data tells you what the market actually does, not what it claims it would do. Search engine marketing intelligence is one of the most underused validation tools available, because search volume reflects genuine, self-expressed demand at scale, without any researcher bias in the data collection.
The strongest validation evidence is actual purchase behaviour. A pre-order. A deposit. A signed letter of intent with a price attached. A paid pilot. When money changes hands, or someone commits to a specific action with a cost, you have real signal. Everything before that is directional at best.
I ran a paid search campaign early in my career at lastminute.com for a music festival. The brief was simple, the execution was straightforward, and within roughly a day we had generated six figures of revenue. What made it work was not the creative or the targeting sophistication. It was that the demand was already there. People were actively searching for tickets. We validated by selling, not by asking. That distinction matters more than most strategy decks acknowledge.
The ICP Problem: Validating With the Wrong People
Even well-designed validation can produce misleading results if you are testing with the wrong audience. This is one of the most common and most expensive mistakes I see in B2B contexts especially.
Teams validate with whoever is available: existing customers who are already bought in, warm prospects who are already in the funnel, or internal contacts who have a vested interest in the project succeeding. None of these groups represent the actual market you are trying to enter.
Proper validation requires clarity on who your ideal customer actually is before you begin. Without that, you cannot design a representative sample, you cannot interpret the results accurately, and you cannot distinguish between “this segment wants it” and “the market wants it.” Using a rigorous ICP scoring rubric before you run validation research is not a nice-to-have. It is the difference between testing the right hypothesis and testing a comfortable fiction.
In B2B environments this is particularly acute. The person who expresses enthusiasm in a discovery call is rarely the person who controls budget. The person who controls budget has different priorities and different objections. If you only validate with champions and not with economic buyers, you will build something that gets enthusiastic internal advocacy and dies in procurement.
Qualitative Validation: What It Can and Cannot Tell You
Qualitative methods have a legitimate role in validation, but it is a narrower role than most teams assign them.
Interviews and focus groups are good at surfacing language. They tell you how people describe their problems, what vocabulary they use, what they compare your category to, and what objections they have not yet articulated clearly. This is genuinely valuable input for positioning, messaging, and product framing.
What qualitative methods cannot reliably do is confirm market size, predict conversion rates, or validate price sensitivity. The social dynamics of a group setting, the desire to be helpful, and the absence of real financial stakes all distort the outputs. Focus groups as a research method have well-documented limitations when used as demand validation tools, and treating them as the primary evidence base is a risk most teams do not consciously accept.
Used correctly, qualitative research helps you build better quantitative instruments. It surfaces the right questions to ask at scale. It identifies the language that resonates. It uncovers the objections you need to address. But it should inform your validation framework, not constitute it.
For teams that want to understand what customers experience without relying solely on what they say, behavioural analytics tools can bridge the gap between stated preference and actual behaviour on digital properties. Watching what people do on a landing page tells you more than asking them what they think of the concept.
Using Indirect Signals to Validate Without Asking Directly
Some of the most useful validation evidence comes from sources where nobody is being asked anything at all.
Search data shows you what problems people are actively trying to solve, at what volume, and with what urgency. If people are searching for a specific problem and existing solutions are thin or poorly reviewed, that is a validation signal. If the search volume is negligible, that is also a validation signal, and not a positive one.
Community behaviour is another underused source. Reddit, niche forums, LinkedIn groups, and industry communities are full of unprompted conversations about problems, frustrations, and unmet needs. Platforms like Reddit can surface genuine market frustration in ways that no survey ever will, because people are not performing for a researcher. They are venting, asking for help, and comparing notes with peers.
Review data on competing products is particularly valuable. One-star and two-star reviews on competitor products are a direct window into unmet needs. If the same complaints appear repeatedly, that is a gap in the market. If the complaints are about things your product would address, that is validation. This is a form of grey market research: using publicly available signals that most teams walk past without stopping to read.
Competitor behaviour is also informative. If well-funded competitors have tried and abandoned a particular market segment, that is worth understanding before you assume you have found a gap. If new entrants are appearing in a category, that suggests someone else has validated demand. Neither is definitive, but both are data points that a serious validation process should account for.
The Landing Page Test: Cheap, Fast, and Honest
One of the most practical validation methods available to any team with a modest budget is the landing page test. You build a page that describes the product or service as if it exists, drives paid traffic to it, and measures what happens. Click-through rates, time on page, form completions, and email sign-ups all tell you something about real demand before you have built anything.
The discipline required is honesty about what the data means. A high bounce rate on a poorly written page tells you about the page, not the market. A low conversion rate on a page with a confusing value proposition tells you about the messaging, not the demand. Landing page fundamentals matter here: if the test vehicle is broken, the test results are unreliable.
Done well, a landing page test can generate meaningful signal within days and at a fraction of the cost of commissioned research. It is not a complete validation programme on its own, but as a fast, low-cost first filter, it is hard to beat.
Early in my career, when I was told there was no budget for a new website, I taught myself to code and built it myself. The instinct to find a cheaper, faster path to a real answer rather than waiting for perfect conditions has served me well since. Validation does not require a research agency and a six-week timeline. It requires a clear question, an honest method, and the willingness to accept an uncomfortable answer.
Pain Points as Validation Anchors
Markets do not validate around features. They validate around pain. If your product solves a problem that people are actively experiencing and cannot adequately solve with existing options, you have the foundation of a validated market. If your product is a marginal improvement on something that already works well enough, the validation bar is much higher.
This is why pain point research is not a separate activity from market validation. It is central to it. Understanding what problems your target market actually has, how acutely they feel them, and what they are already doing to manage them is the context within which all other validation evidence sits.
A product that addresses a mild inconvenience requires a very different validation standard than one that addresses a costly, urgent, recurring problem. The urgency and cost of the pain determines how much friction the market will tolerate in adopting a new solution. Getting this wrong is how teams build products that get good feedback in testing and poor adoption in market.
Validation Across the Business: Not Just for New Products
Market validation is most commonly discussed in the context of new product launches, but the same discipline applies to a much wider range of business decisions.
Entering a new geographic market. Repositioning an existing product for a different segment. Launching a new pricing tier. Expanding into an adjacent category. All of these carry market risk that validation can reduce before significant investment is committed.
In agency environments, I have seen this play out repeatedly with service line expansion. An agency decides to add a new capability, hires people, builds the infrastructure, and then discovers the market either does not want it from them specifically, or already has adequate alternatives at a lower price point. A few structured conversations with target buyers, combined with a competitive audit and some search intelligence, would have surfaced that risk in a week. Instead, the lesson cost six months and a significant P&L hit.
The same logic applies to technology investment decisions. Before committing to a platform, a build, or a significant operational change, the question of whether the underlying market assumption is sound deserves the same rigour as a new product launch. Aligning technology decisions with business strategy requires validated assumptions about where the market is and where it is heading, not just internal conviction about what would be useful to build.
When Validation Says No
The most valuable outcome of a rigorous validation process is a clear negative. A market that does not exist. A price point that nobody will pay. A problem that is not painful enough to drive behaviour change. A segment that is already adequately served.
Teams that treat a negative validation result as a failure have misunderstood the purpose of the exercise. Finding out early that an idea does not have a viable market is one of the best returns on research investment you can generate. The alternative, finding out after you have built and launched, is considerably more expensive and considerably more demoralising.
A negative result should also prompt a specific question: is the market wrong, or is the framing wrong? Sometimes validation fails not because the opportunity does not exist, but because the product concept, the target segment, or the price architecture is misaligned with real demand. Iterating on the framing and re-testing is a legitimate response. Ignoring the negative and pressing on regardless is not.
I have judged the Effie Awards, where effectiveness is the only criterion that matters. The campaigns that win are almost always built on genuine market insight, not on assumptions that were never tested. The ones that do not win are frequently well-executed ideas that were solving the wrong problem for the wrong audience. Validation would not have guaranteed an Effie. But it would have changed the brief.
For a broader view of how validation fits into a complete research and intelligence programme, the market research hub covers the methods, tools, and strategic frameworks that support better commercial decisions at every stage of planning.
The ecommerce funnel is a useful lens here too, because it makes the relationship between validated demand and commercial performance explicit. Validation does not end at awareness. It runs through the entire purchase experience, and the signals at each stage tell you something different about whether the market is real.
About the Author
Keith Lacy is a marketing strategist and former agency CEO with 20+ years of experience across agency leadership, performance marketing, and commercial strategy. He writes The Marketing Juice to cut through the noise and share what works.
