Central Location Testing: What It Tells You That Surveys Never Will
Central location testing is a form of qualitative and quantitative market research where participants are recruited to a controlled venue, typically a dedicated research facility, shopping centre, or hotel suite, and exposed to a product, concept, or stimulus under supervised conditions. It gives marketers something most digital research methods cannot: observed, in-person reactions to real things, in real time.
Used well, a central location test (CLT) bridges the gap between what consumers say they will do and what they actually do when confronted with a product, a pack, a piece of creative, or a price point. That gap is where most marketing decisions go wrong.
Key Takeaways
- Central location testing captures behavioural and sensory reactions that surveys and online panels cannot replicate, making it the right tool for product, pack, and creative decisions where stimulus matters.
- The quality of a CLT depends almost entirely on the quality of the stimulus and the recruitment brief , both are routinely underinvested in.
- CLTs are expensive relative to online research, but the cost of a bad product launch or a misfiring campaign is almost always higher than the cost of proper testing.
- Combining CLT findings with search intelligence and secondary research produces a more complete picture than any single method alone.
- The most common CLT mistake is treating participant responses as purchase intent predictions , they are directional signals, not forecasts.
In This Article
- What Is a Central Location Test and When Should You Use One?
- How a Central Location Test Actually Works
- The Recruitment Problem Most Researchers Underestimate
- Blind vs. Branded Testing: Choosing the Right Design
- What CLTs Cannot Tell You
- Integrating CLT Data with Broader Research and Intelligence
- The Cost Question: When a CLT Is Worth the Money
- Running a CLT Alongside a Technology or Strategic Review
- Practical Steps for Commissioning a CLT That Delivers
If you are building out a research programme and want to understand where CLT sits relative to other methods, the Market Research and Competitive Intelligence hub covers the full landscape, from primary qualitative methods through to competitive data sourcing and audience profiling.
What Is a Central Location Test and When Should You Use One?
A central location test recruits a defined sample of participants to a single venue, or a small number of venues running simultaneously, where they are exposed to a stimulus and their responses are captured. The stimulus might be a food or drink product, a fragrance, a prototype package, a TV commercial, a shelf display, or a pricing structure. The venue might be a purpose-built research facility with one-way mirrors, a hired room in a shopping centre, or a mobile research unit parked outside a supermarket.
The defining feature is control. You know what participants saw, tasted, smelled, or read. You know the order in which they encountered things. You know what they were asked and when. That level of control is what makes CLT findings interpretable in a way that self-reported, unobserved research often is not.
CLTs are most appropriate when:
- The product or concept requires physical interaction, sensory evaluation, or direct comparison
- You need to control for order effects, context, or competitive framing
- The decision is high-stakes enough to warrant the cost of proper primary research
- You want to observe non-verbal responses alongside stated preferences
- You are testing something that cannot be adequately described or shown via a screen
They are less appropriate when you are exploring broad attitudinal territory, mapping a market, or need large samples at low cost. For that kind of work, online panels, secondary data, or methods like focus group research will often serve you better before you commit to the controlled environment of a CLT.
How a Central Location Test Actually Works
The mechanics of a CLT vary by category and objective, but the core process is consistent. Participants are recruited to quota, screened for relevance, and brought to the venue in waves, usually in small groups or one at a time, to avoid contamination between respondents.
At the venue, they are introduced to the stimulus under controlled conditions. In a product test, this might mean a blind tasting where all branding is removed, a branded test where packaging is visible, or a sequential monadic design where each participant evaluates one product before moving to another. In a creative test, they might watch a piece of video content and be asked to respond immediately, before any discussion or prompting occurs.
Responses are captured through a combination of structured questionnaires, moderator-led probing, and, increasingly, passive observation methods including eye-tracking, facial coding, and biometric measurement. The quantitative component gives you a score. The qualitative component tells you why.
One thing I learned running agency-side research programmes across multiple FMCG and retail clients is that the quality of the debrief conversation matters as much as the questionnaire data. Participants who are asked to explain their reactions in their own words, without being led, will frequently surface objections or associations that the structured instrument never anticipated. Those are often the most commercially useful findings in the entire study.
The Recruitment Problem Most Researchers Underestimate
CLT findings are only as good as the people in the room. This sounds obvious, but recruitment is consistently the most under-scrutinised part of the process. Clients approve the discussion guide and the stimulus design in detail, then sign off on a recruitment brief that says “female, 25-45, buys the category” and leave it at that.
The problem is that category buyers are not a homogeneous group. Someone who buys premium skincare once a year because it was on promotion is not the same research participant as someone who actively seeks out new launches and reads ingredient lists. Putting both in the same CLT and averaging their responses produces a finding that accurately represents neither.
This is where the thinking you would apply to an ICP scoring framework translates directly into research design, even in a B2C context. The discipline of defining precisely who you are trying to understand, and what behaviours and attitudes qualify them, applies as much to a CLT recruitment brief as it does to a sales targeting model. Vague recruitment produces vague findings.
Good recruitment briefs specify: category involvement level, purchase frequency, attitudinal orientation toward the category, exclusions for professional researchers and competitors, and any behavioural or life-stage criteria that are genuinely predictive of how people will respond to the stimulus. This takes more effort upfront. It produces materially better data.
Blind vs. Branded Testing: Choosing the Right Design
One of the most consequential design decisions in a CLT is whether to run a blind test, a branded test, or both. Each answers a different question, and conflating them is a common source of misleading findings.
A blind test strips away all branding and asks: does this product deliver on its core sensory or functional promise? It tells you whether the product itself is competitive. A branded test adds the full brand context, packaging, and framing, and asks: does the brand amplify or diminish the product experience? The gap between blind and branded scores is one of the most informative metrics in consumer goods research. A product that scores well blind but poorly branded has a brand problem. A product that scores poorly blind but well branded has a product problem that marketing is currently masking.
Sequential monadic designs, where the same participant evaluates multiple products in a randomised order, allow direct comparison but introduce order effects and fatigue. Monadic designs, where each participant evaluates only one product, avoid those effects but require larger samples to achieve the same statistical power. Neither is universally superior. The right choice depends on what you are trying to learn and how much budget you have to work with.
What CLTs Cannot Tell You
I have sat in enough research debrief sessions to know that the most dangerous moment in a CLT project is when a client looks at a strong purchase intent score and says “right, let’s launch.” The score is not a forecast. It is a signal.
CLT participants are in an artificial environment. They know they are being researched. They have been recruited because they are category buyers, which means they are more engaged with the category than the average person who will encounter this product in a supermarket aisle while distracted and time-poor. Their stated intent to purchase is systematically inflated relative to actual market behaviour, in almost every category and almost every study.
What CLTs do tell you reliably: relative preference between options, the specific attributes driving positive or negative reactions, the language consumers use to describe the product and its benefits, and the objections that will need to be overcome in positioning or communication. That is genuinely valuable. It is just not a sales forecast.
CLTs also cannot tell you much about competitive context in the real market, about how distribution and shelf placement will affect trial, or about the long-term effects of repeated exposure. Those questions require different research instruments. Understanding the pain points that drive category behaviour over time, rather than in a single controlled exposure, is a separate and complementary research exercise.
Integrating CLT Data with Broader Research and Intelligence
A CLT in isolation is a snapshot. It tells you how a specific group of people responded to a specific stimulus on a specific day. That is useful, but it becomes significantly more useful when set against other data sources.
Search data, for instance, tells you what language people actually use when they are actively looking for solutions in a category. There is often a striking mismatch between the language that performs well in a CLT questionnaire and the language that drives search behaviour. Search engine marketing intelligence can surface the vocabulary of intent in a way that no qualitative research session will, because search queries are unfiltered and unobserved. Triangulating CLT language findings against search data before you finalise positioning is a discipline I would apply to almost any consumer-facing launch.
Similarly, secondary market data and competitive intelligence can contextualise CLT scores. A product that scores 6.8 out of 10 on overall liking means very little without knowing how competitive products score on the same instrument. If the category norm is 6.2, you have something. If the category norm is 7.4, you have a problem, regardless of how encouraging the qualitative feedback sounded.
There is also a category of competitive intelligence that sits in less obvious places. Grey market research, which covers informal channels, resale markets, and unofficial distribution, can reveal genuine unmet needs that CLT participants, drawn from conventional category buyers, will never surface. If a product is being traded informally at a premium, that is a signal worth understanding before you design your CLT stimulus.
The broader point is that no single research method has a monopoly on insight. The marketers who get the most from CLTs are those who treat them as one input into a broader intelligence picture, not as the definitive answer to a commercial question. BCG’s work on portfolio management and value creation makes a related point about strategic decision-making: the quality of a decision is a function of the range of inputs considered, not the confidence of the decision-maker.
The Cost Question: When a CLT Is Worth the Money
CLTs are not cheap. A properly designed study with adequate sample sizes, professional recruitment, venue hire, stimulus production, and analysis will run to tens of thousands of pounds or dollars depending on scale and complexity. For some organisations, that feels prohibitive. For others, it is a rounding error relative to the cost of a failed launch.
The calculation I apply is straightforward. What is the cost of the decision you are trying to de-risk? If you are about to commit to a production run, a retail ranging negotiation, or a media campaign behind a product or concept that has not been properly tested, the downside exposure is almost always larger than the cost of a CLT. The research is not an overhead. It is a risk management instrument.
Early in my career, I had a version of this conversation with a client who was confident enough in their own instincts to skip the pre-launch testing on a product reformulation. The reformulation failed in market. The CLT they eventually commissioned post-launch, to understand what had gone wrong, cost roughly the same as the pre-launch test would have. The difference was that the post-launch version came with a distribution problem, a retailer relationship under strain, and a six-month delay in getting back to where they had started. The research cost was identical. The total cost was not.
That said, CLTs can be made more cost-efficient without sacrificing validity. Smaller, tightly defined samples with rigorous recruitment will outperform large, loosely recruited samples on almost every measure of research quality. Focusing the stimulus on the specific decision at hand, rather than trying to answer every open question in a single study, keeps scope manageable and findings actionable. And combining CLT findings with desk research and external expertise can reduce the number of primary research rounds needed before a decision is ready to be made.
Running a CLT Alongside a Technology or Strategic Review
CLTs are most commonly associated with consumer goods and food and drink categories, but the underlying logic applies across a wider range of commercial decisions than most people assume. Concept testing for a new service proposition, prototype evaluation for a digital product, or creative testing for a B2B campaign all share the same fundamental requirement: you need to understand how real people respond to a real thing before you commit to it at scale.
In a technology or professional services context, the “stimulus” might be a prototype interface, a pricing structure, or a value proposition statement. The controlled environment might be a user research lab rather than a shopping centre facility. The principles are the same: control the exposure, observe the response, probe the reasoning, and treat the findings as directional rather than definitive.
When CLT findings are being used to inform a broader strategic review, the kind of structured analysis covered in a technology consulting SWOT and business strategy alignment framework becomes relevant. Research findings that are not connected to a decision-making structure tend to get noted and then ignored. The value of CLT data is realised when it feeds directly into a strategic question with a named owner and a clear decision timeline.
I have seen this fail repeatedly in large organisations where research is commissioned by one team, presented to a different team, and then filed while the original decision gets made on instinct anyway. The research was not the problem. The process for using it was. Good market research discipline, including CLTs, requires as much attention to how findings will be used as to how data will be collected. Good writing about research has the same requirement: the opening has to earn attention, but what follows has to be structured to drive action.
Practical Steps for Commissioning a CLT That Delivers
If you are commissioning a central location test, these are the decisions that will most directly determine whether the findings are useful.
Define the decision first. Before you brief a research agency, write down the specific commercial decision this research is meant to inform. Not “understand consumer response to the new product” but “decide whether to proceed to full production with formulation A or formulation B.” The decision frame shapes everything: the sample, the stimulus, the measures, and the analysis.
Write a tight recruitment brief. Specify the behavioural and attitudinal criteria that make someone genuinely relevant to the decision. Category purchase frequency, involvement level, and specific exclusions matter more than age and gender alone.
Invest in the stimulus. If you are testing a product, test the product at the quality level it will be produced at, not a rough prototype. If you are testing packaging, test finished artwork, not a sketch. The gap between the stimulus and the final product introduces noise that corrupts your findings.
Choose the right design for the question. Blind or branded, monadic or sequential: these are not default choices. Each has implications for what you can learn and how confident you can be in the findings.
Plan the analysis before you collect the data. Know in advance which measures are primary and which are secondary. Know what score on what instrument would change the decision, and in which direction. This prevents the post-hoc rationalisation of findings that plagues poorly planned research.
Connect findings to the decision process. Identify who owns the decision, when it needs to be made, and how the research findings will be presented to them. Research that arrives after the decision has already been made is not research. It is documentation.
If you want to see how CLT fits within a full market research toolkit, the Market Research and Competitive Intelligence hub covers the range of methods available and how to sequence them for different commercial objectives. CLT is a powerful instrument, but it works best when it sits within a coherent research architecture rather than being commissioned in isolation.
About the Author
Keith Lacy is a marketing strategist and former agency CEO with 20+ years of experience across agency leadership, performance marketing, and commercial strategy. He writes The Marketing Juice to cut through the noise and share what works.
