System 1 Ad Testing: What Your Metrics Are Missing

System 1 ad testing measures how quickly and instinctively people respond to creative, before rational thought kicks in. It works on the premise that most purchase decisions are made emotionally and automatically, and that the ads which win are the ones that land in that first fraction of a second, not the ones that make the most logical argument.

If your creative testing process relies entirely on stated preferences, recall scores, or click-through rates, you are measuring the wrong thing. System 1 testing gives you a read on emotional response, brand recognition, and instinctive appeal, the signals that actually predict whether an ad will shift behaviour at scale.

Key Takeaways

  • System 1 testing measures instinctive emotional response, not rational opinion, which is a fundamentally different signal from standard recall or preference research.
  • Most ad testing asks people what they think. System 1 testing measures what they feel before they have time to construct a considered answer.
  • Strong System 1 performance correlates with long-term brand growth, not just short-term campaign metrics.
  • System 1 and System 2 thinking are not competing frameworks. The best creative testing uses both, for different decisions at different stages.
  • Brands that consistently win in market tend to build strong memory structures, and System 1 testing is one of the clearest ways to evaluate whether your creative is doing that job.

Why Standard Ad Testing Keeps Getting It Wrong

I spent years watching creative get killed in research. A concept would come back from a focus group with a mediocre score, and the client would pull it. The work that survived was usually the work that had been smoothed down to the point of inoffensiveness. Safe. Rational. Forgettable.

The problem is not that research is wrong. The problem is that most ad testing asks people to do something they are not very good at: explain why they like or dislike something, predict whether it would make them buy, and articulate their emotional response in a structured way. Human beings are notoriously unreliable narrators of their own decision-making. We rationalise after the fact. We give socially acceptable answers. We say we would buy the responsible option and then fill our trolley with whatever we actually wanted.

Standard testing methods, particularly those built around claimed purchase intent, have a structural flaw. They measure System 2 thinking: deliberate, considered, verbal. But the decisions that drive most consumer behaviour happen in System 1: fast, automatic, emotional. The ad that scores well in a rational evaluation may still fail to create any instinctive pull. And the one that gets dismissed as “weird” or “too different” may be the one that actually lodges in memory and drives long-term growth.

This is not a theoretical concern. When I was judging the Effie Awards, the work that consistently stood out was not the work that tested cleanly. It was the work that was brave enough to create a genuine emotional reaction, positive or disorienting, but never neutral. Neutral does not sell.

What System 1 Testing Actually Measures

System 1 testing is designed to capture implicit, automatic responses rather than deliberate opinion. The mechanics vary by provider, but the core principle is consistent: expose someone to creative and measure their reaction before they have time to construct a rational response.

That might mean measuring response speed, the assumption being that faster positive responses indicate stronger instinctive appeal. It might mean using facial coding to read micro-expressions. Some approaches use biometric measures like skin conductance or eye-tracking. Others use implicit association tests that measure how strongly a person links a brand or image to specific emotional concepts, without asking them directly.

What all of these approaches have in common is that they are trying to get underneath the stated preference. They are asking: what does this person actually feel, not what do they think they are supposed to say?

The dimensions that System 1 testing typically evaluates include:

  • Emotional intensity: Does the ad generate a strong response at all, positive or negative? Weak emotional response is a warning sign regardless of direction.
  • Emotional valence: Is the response broadly positive? Negative emotional intensity can work in certain contexts but needs to be intentional.
  • Brand recognition: Does the viewer connect the ad to the brand quickly and instinctively? Ads that score well emotionally but fail on brand linkage are a common and expensive problem.
  • Distinctiveness: Does the creative stand out from the category? Ads that blend into the visual wallpaper of a category may score well on rational metrics but fail to build any memory structure.
  • Fame potential: Will people talk about this? Will it spread? This is harder to measure but some System 1 methodologies attempt to capture it through social sharing intent measured implicitly.

If you are building or refining your go-to-market approach and want to understand how creative testing fits into a broader commercial framework, the Go-To-Market and Growth Strategy hub covers the full picture, from audience strategy through to campaign execution.

The Memory Structure Argument

The reason System 1 testing matters commercially is rooted in how brand memory works. Brands grow by being mentally available at the moment of purchase. That availability is built through repeated exposure to distinctive brand assets, visual codes, characters, sounds, and emotional associations that become linked to the brand in long-term memory.

Advertising that generates strong System 1 responses tends to build these memory structures more effectively than advertising that is merely informative. An ad that makes you feel something, even briefly, is more likely to be encoded in memory than one that gives you a rational argument you consciously process and then forget.

Early in my career, I was heavily focused on lower-funnel performance. Click-through rates, conversion rates, cost per acquisition. I thought that was where the real commercial value was. Over time, and after managing hundreds of millions in ad spend across more than 30 industries, I changed my view significantly. A lot of what performance marketing gets credited for was going to happen anyway. The person who typed a branded search term was already on their way. What actually drove them there, in many cases, was the emotional impression left by brand advertising they had seen weeks or months earlier.

System 1 testing is, in part, an attempt to measure whether your advertising is doing that upstream work. Whether it is building the kind of instinctive positive association that makes someone more likely to choose your brand when the moment arrives.

System 1 vs System 2: A False Choice

There is a tendency in some marketing circles to treat System 1 thinking as the only thing that matters and to dismiss rational messaging as a waste of time. That is an overcorrection.

System 2 thinking, the deliberate, analytical kind, plays a real role in high-consideration purchases. Someone buying enterprise software, a car, or a pension is not making a purely instinctive decision. They will research, compare, and evaluate. Rational messaging has a job to do in those contexts. The mistake is applying a System 2 testing framework to creative that is designed to work emotionally, or building advertising that is entirely rational when the category demands emotional resonance.

The better question is not “System 1 or System 2?” but “what is this ad trying to do, and what testing approach is appropriate for that job?” Brand advertising designed to build broad emotional salience should be evaluated on System 1 dimensions. Direct response creative designed to drive immediate action can reasonably be evaluated on rational and behavioural metrics. The failure mode is using the wrong test for the wrong job.

I remember a brainstorm early in my agency career where we were developing creative for a major drinks brand. The instinct in the room was to lead with product attributes: taste, heritage, quality. Rational claims. The work that ended up being most powerful was the piece that abandoned the product argument entirely and went for pure emotional territory. It felt risky at the time. It performed. The lesson stayed with me.

How System 1 Testing Fits Into the Creative Development Process

System 1 testing is most useful at specific points in the creative process, not as a blanket replacement for all other forms of evaluation.

Early concept screening: Before significant production investment, System 1 approaches can help identify which creative territories generate the strongest instinctive response. This is faster and cheaper than full copy testing and can prevent resources being committed to directions that will not land emotionally.

Pre-launch validation: Once creative is developed but before it goes live, System 1 testing can flag whether the emotional response is strong enough and positive enough to justify the media investment. It can also identify whether brand linkage is strong, a common failure point in emotionally driven creative.

Post-launch learning: System 1 measures can be used to track whether emotional associations are building over time. This is particularly useful for brand campaigns with long time horizons where behavioural metrics alone will not tell you whether the advertising is working until much later.

Competitive benchmarking: Some organisations use System 1 testing to evaluate competitor advertising alongside their own, building a picture of how the emotional landscape of a category is shifting and where there may be white space.

The practical challenge is that System 1 testing requires specialist methodology and, in many cases, specialist providers. It is not something you can approximate with a standard survey. The response time measurements, implicit association protocols, and biometric approaches require proper research design. This adds cost and time, which means it tends to be used selectively rather than routinely. That is a reasonable trade-off. The question is whether you are being selective about the right decisions.

Common Mistakes Brands Make With Creative Testing

Having worked across a wide range of categories and seen a lot of research processes from the inside, a few failure patterns repeat themselves.

Testing too late: Research commissioned after the creative is finished and the production budget is spent is not really testing, it is validation-seeking. The findings rarely change anything because the sunk cost is too high. Testing needs to happen early enough that the results can actually influence decisions.

Optimising for the wrong metric: Brands that use click-through rate or recall as their primary creative quality signal are measuring something real, but not necessarily the thing that drives long-term brand growth. Recall without emotional association does not build purchasing preference. Being remembered is not the same as being chosen.

Averaging out the response: Aggregate scores hide important information. An ad that gets a strong positive response from 40% of viewers and a neutral or negative response from the rest may outperform an ad that gets a moderate positive response from everyone. Depending on your targeting strategy, the concentrated response may be more valuable. Looking only at mean scores misses this.

Ignoring brand linkage: Emotionally powerful creative that does not link back to the brand is one of the most common and expensive mistakes in advertising. It builds positive feelings that get attributed to the category or, worse, to a competitor. System 1 testing should always include a brand linkage measure, not just an emotional response measure.

Using research to make the decision rather than inform it: Testing should sharpen judgment, not replace it. I have seen research used to avoid accountability: if the ad tests well and fails in market, the research is blamed. If it tests badly and someone champions it anyway, that person is exposed. The result is a culture where research is used defensively rather than constructively. The best creative decisions I have seen made were ones where experienced people used research as one input among several, not as the final word.

What Good Looks Like in Practice

Brands that use System 1 testing well tend to share a few characteristics. They have a clear brief that specifies the emotional territory the creative is trying to own, not just the rational message it needs to communicate. They test at multiple stages rather than once at the end. They use System 1 measures alongside other signals rather than in isolation. And they have the organisational confidence to act on the findings, including killing work that scores poorly even if it cost a lot to develop.

The brands that struggle with creative testing are usually the ones where the research process is disconnected from the creative process. Testing happens in a separate lane, run by a separate team, and the findings land too late to be useful. Or the testing methodology is mismatched to the creative objective, applying rational evaluation criteria to emotional creative or vice versa.

There is also a category-specific dimension to this. In fast-moving consumer goods, where purchase decisions are genuinely low-involvement and instinctive, System 1 performance is probably the most important creative quality signal you have. In B2B, where the purchase process involves multiple stakeholders and extended evaluation periods, the balance shifts. But even in B2B, the emotional impression created by brand advertising affects which vendors get invited into the consideration set in the first place. The instinctive response still matters. It just operates at a different point in the process.

Understanding where creative testing sits within a broader go-to-market framework matters more than most brands acknowledge. The decisions you make about how to test creative are downstream of the decisions you make about what the advertising is trying to do and who it is trying to reach. If those upstream decisions are unclear, no testing methodology will save you. The Go-To-Market and Growth Strategy hub at The Marketing Juice addresses how these decisions connect, from market definition through to campaign measurement.

The Honest Limitation

System 1 testing is a better predictor of emotional impact than most alternatives. It is not a guarantee of commercial success. An ad can generate strong instinctive positive responses and still fail to shift market share if the media plan is wrong, the distribution is limited, the price point is off, or the competitive context has shifted. Testing tells you about the creative. It does not tell you about everything else.

There is also a ceiling effect in well-established categories. If the emotional territory of a category is already saturated, generating a strong positive response to advertising in that territory may not differentiate you from competitors who are doing the same thing. Distinctiveness and emotional resonance are both necessary. Neither is sufficient on its own.

The honest position is that System 1 testing gives you better information about a specific dimension of creative quality than most other approaches. That is worth having. It is not a silver bullet, and treating it as one will lead to the same disappointments as treating any other single metric as the definitive measure of marketing effectiveness.

What it does do, when used well, is push creative development in a more productive direction. It creates accountability for emotional impact, not just rational clarity. It shifts the conversation from “does this communicate the message?” to “does this make people feel something?” That is a more honest conversation about what advertising actually does. And having that conversation more often, earlier in the process, tends to produce better work.

About the Author

Keith Lacy is a marketing strategist and former agency CEO with 20+ years of experience across agency leadership, performance marketing, and commercial strategy. He writes The Marketing Juice to cut through the noise and share what works.

Frequently Asked Questions

What is System 1 ad testing?
System 1 ad testing measures instinctive, emotional responses to advertising before rational thought takes over. It uses methods like response time measurement, facial coding, and implicit association tests to capture how people feel about creative, rather than what they say they think about it.
How is System 1 testing different from traditional ad research?
Traditional ad research typically asks people to rate, recall, or evaluate advertising consciously. System 1 testing bypasses deliberate opinion by measuring automatic responses, which are faster and less subject to social desirability bias. The result is a different, and often more predictive, signal of how advertising will perform in the real world.
When should brands use System 1 testing?
System 1 testing is most valuable at the concept screening stage, before production investment is committed, and again at pre-launch validation. It can also be used post-launch to track whether emotional brand associations are building over time. It is less useful as a post-hoc justification for work that is already finished.
Does System 1 testing work for B2B advertising?
Yes, though the application differs from consumer categories. In B2B, emotional impressions influence which vendors make the consideration set, even if the final decision involves rational evaluation. System 1 testing can help B2B brands understand whether their advertising is creating the instinctive positive associations that get them into the room in the first place.
What are the limitations of System 1 ad testing?
System 1 testing measures emotional impact, not commercial outcomes. Strong emotional response to an ad does not guarantee sales growth if the media plan, pricing, distribution, or competitive context is misaligned. It also requires specialist methodology and tends to cost more than standard survey research. Used well, it is a valuable input. Used in isolation, it is incomplete.

Similar Posts