Misleading Correlations Are Costing You Better Decisions
Misleading correlations occur when two variables move together in a way that looks meaningful but shares no causal relationship. In marketing, they are everywhere, and they are expensive. When you build strategy around a correlation that does not hold up under scrutiny, you are not making data-driven decisions. You are making confident mistakes.
The problem is not that marketers lack data. It is that the data tells stories we want to hear, and we rarely stop to question whether the story is true.
Key Takeaways
- Correlation between two marketing variables does not confirm that one is driving the other. A third variable, or pure coincidence, is almost always worth investigating first.
- Attribution models assign credit to touchpoints based on rules, not proof. The model you choose shapes what you believe is working, not necessarily what is.
- Absolute performance metrics are consistently misleading without market context. Growing 15% while your category grows 30% is a loss dressed up as a win.
- The most dangerous correlations are the ones that align with what you already believed. Confirmation bias makes them feel like evidence.
- Good analytical practice means building the habit of asking “what else could explain this?” before drawing conclusions from any data pattern.
In This Article
- Why Marketers Are Particularly Vulnerable to False Correlations
- The Attribution Problem Is a Correlation Problem in Disguise
- Absolute Metrics Versus Relative Performance: The Context Problem
- Confirmation Bias Makes Misleading Correlations Invisible
- Seasonality, External Events, and the Third Variable Problem
- Vanity Metrics and the Correlation Between Activity and Outcomes
- How to Build the Habit of Questioning Correlations
- When Correlations Are Useful Without Being Causal
If you are working through how data and evidence should sit inside a broader commercial framework, the articles in the Go-To-Market and Growth Strategy hub cover the strategic thinking that sits behind decisions like these.
Why Marketers Are Particularly Vulnerable to False Correlations
Marketing data is almost perfectly designed to produce misleading correlations. You have multiple channels running simultaneously. You have long and variable purchase cycles. You have seasonal effects, economic conditions, competitor activity, and product changes all operating at the same time. And on top of all of that, you have attribution models that take credit assignment and present it as causal proof.
Early in my career, I watched a client become genuinely convinced that a particular display campaign was driving their sales uplift. The timing lined up. The numbers moved together. The agency presenting the data was confident. What nobody had properly accounted for was that the campaign had run during a period when a major competitor had pulled back their own spend significantly. The sales lift was real. The cause was not what anyone thought it was.
That experience stuck with me. Not because it was unusual, but because it was so completely normal. This kind of thing happens in marketing constantly. The correlation is genuine. The interpretation is wrong. And decisions get made, budgets get allocated, and strategies get built on a foundation that was never solid.
Part of the vulnerability comes from how we are wired. We are pattern-seeking by nature. When two lines on a chart move in the same direction, the brain wants to connect them. It takes deliberate effort to resist that instinct and ask whether the connection is real or coincidental.
The Attribution Problem Is a Correlation Problem in Disguise
Attribution modelling is the most institutionalised form of misleading correlation in marketing. Every attribution model, whether last-click, first-click, linear, or data-driven, is a set of rules about which touchpoints get credit for a conversion. Those rules are not derived from causal proof. They are assumptions baked into a framework that then produces numbers that look authoritative.
Last-click attribution is the most obvious offender. It assigns 100% of the credit for a conversion to the final touchpoint before purchase. If someone clicks a branded search ad after seeing six weeks of social content, watching a YouTube pre-roll, and reading a comparison article, the search ad gets all the credit. The correlation between the final click and the conversion is real. The conclusion that the search ad caused the purchase is almost certainly wrong.
When I was running performance accounts at scale, managing significant budgets across multiple channels, we could demonstrate clearly that switching attribution models changed the apparent performance of every channel in the mix. Paid social looked weak under last-click and strong under first-click. Branded search looked dominant under last-click and modest under linear. The underlying reality had not changed. Only the lens had changed. But that lens determined where budget went.
The honest position is that attribution models show you correlation between touchpoints and conversions. They do not show you causation. Treating them as if they do leads to systematic misallocation of budget, and it does so with complete confidence because the numbers look precise.
Tools like those covered in SEMrush’s breakdown of growth analysis tools can give you more angles on channel performance, but even the best tooling cannot resolve the fundamental attribution problem. It can only give you more data to misinterpret if you are not thinking carefully about causation.
Absolute Metrics Versus Relative Performance: The Context Problem
One of the cleanest examples of misleading correlation in marketing is the relationship between absolute performance metrics and business health. Revenue went up. Conversions increased. Cost per acquisition improved. All of these things can be true simultaneously with the business losing competitive ground.
I have used this framing many times when working with clients who were proud of their growth numbers: if your business grew 10% last year but the market grew 20%, you lost ground. You did not grow. You shrank relative to the opportunity. The correlation between your revenue line and success is misleading because it ignores the context in which that revenue was generated.
This is not a niche observation. It is one of the most common ways that marketing teams and boards alike get comfortable with performance that should be making them uncomfortable. The absolute number goes up, confidence goes up with it, and nobody asks what the number would look like if you benchmarked it against category growth, competitor performance, or market share movement.
BCG’s work on go-to-market strategy and product launch planning makes a related point about the importance of market context in reading performance signals. Absolute numbers without market benchmarks are not performance data. They are activity data.
When I joined an agency that was loss-making and worked to turn it around, one of the first things I had to do was separate the metrics that felt good from the metrics that actually indicated health. Revenue was not the problem. Margin was. The top line looked fine. The business was bleeding. Correlating revenue growth with business health had masked the real issue for long enough that it had become a structural problem.
Confirmation Bias Makes Misleading Correlations Invisible
The most dangerous misleading correlations are not the ones that surprise you. They are the ones that confirm what you already believed. When data appears to validate an existing assumption, the scrutiny drops. The questions stop. The correlation becomes evidence.
This is confirmation bias operating at the level of data interpretation, and it is endemic in marketing. Teams run a campaign, believe it worked, find data that supports that belief, and stop looking. The possibility that the data is misleading, that the correlation has an alternative explanation, rarely gets investigated because the conclusion feels satisfying.
I judged the Effie Awards, which are specifically designed to evaluate marketing effectiveness rather than creative quality. One of the things that process taught me was how often case studies are constructed around correlations that have been selected because they support the narrative, not because they represent the most rigorous reading of the evidence. The work might genuinely have been effective. But the proof offered is frequently the correlation that looks best, not the analysis that is most defensible.
That is not a criticism of the Effies or the teams submitting. It reflects a broader industry habit of treating correlation as proof when the correlation happens to align with what we wanted to show. The antidote is a simple discipline: before you accept a data pattern as evidence, ask what else could explain it. Not as a rhetorical device. As a genuine analytical step.
Seasonality, External Events, and the Third Variable Problem
In statistics, the third variable problem refers to situations where two variables appear correlated because both are being driven by a third variable that nobody is measuring. In marketing, third variables are everywhere, and they are usually hiding in plain sight.
Seasonality is the most common one. If you launch a campaign in October and your sales increase through November and December, the correlation between the campaign and the sales uplift looks compelling. But if your category always grows in Q4, you have not demonstrated that the campaign caused the growth. You have demonstrated that your sales moved with the season, which they were going to do anyway.
External events create the same problem. Economic conditions, news cycles, competitor behaviour, platform algorithm changes, and distribution shifts can all move your numbers in ways that look like the result of your own activity. Without proper controls or at minimum a serious attempt to account for these factors, the correlation between your marketing activity and your results is not evidence of effectiveness. It is noise that happens to look like signal.
Forrester’s analysis of go-to-market challenges in complex markets touches on exactly this difficulty. In markets with long sales cycles, multiple stakeholders, and significant external variables, attributing outcomes to specific marketing activities requires a level of analytical rigour that most teams simply do not apply. The correlation is easy to find. The honest interpretation is much harder.
The practical implication is that you need comparison data. Year-on-year performance adjusted for market conditions. Control groups where you can create them. Periods of activity versus inactivity that allow you to isolate the effect of what you are doing. None of this is perfect. But it is significantly more honest than treating a raw correlation as proof.
Vanity Metrics and the Correlation Between Activity and Outcomes
There is a specific category of misleading correlation that deserves its own section: the correlation between marketing activity and business outcomes when the connection is assumed rather than demonstrated.
Impressions correlate with brand awareness. Clicks correlate with interest. Followers correlate with reach. Email open rates correlate with engagement. All of these correlations are real in a narrow sense. The problem is that the jump from these activity metrics to revenue, margin, or market share is not a short step. It is a leap across a gap that is rarely measured.
I have sat in enough board meetings to know how this plays out. The marketing team presents impressive activity numbers. The board nods. Nobody asks whether any of it connected to commercial outcomes because the correlation between looking busy and being effective has been assumed for so long that it has become invisible.
The teams that avoid this trap are the ones who build their measurement frameworks backwards from commercial outcomes, not forwards from activity. What does a conversion actually require? What behaviour precedes purchase? What does the data show about the path from awareness to revenue? Platforms like Vidyard’s research into pipeline and revenue potential for go-to-market teams highlight how much untapped signal sits between engagement metrics and actual revenue, and how rarely that gap gets properly examined.
When I grew an agency from 20 people to 100 and moved it from the bottom of the market to the top five in its category, the discipline that mattered most was not the quality of the work, though that mattered. It was the rigour around connecting what we did to what clients actually cared about commercially. Vanity metrics were not tolerated, not because they were philosophically wrong, but because they consistently led to decisions that did not hold up.
How to Build the Habit of Questioning Correlations
None of this is an argument against using data. It is an argument for using data more honestly. The practical shift is not complicated, but it does require discipline, particularly in environments where confident-sounding data is rewarded regardless of its quality.
The first habit is to ask “what else could explain this?” as a standard step in any analysis. Not occasionally. Every time. If your email campaign drove a 20% uplift in site visits, what else happened that week? Was there a PR mention? Did a competitor have an outage? Did you send the email to a segment that was already more engaged than your average list? The question is not designed to dismiss the data. It is designed to test whether the interpretation holds up.
The second habit is to separate the measurement of activity from the measurement of outcomes, and to be explicit about the assumptions connecting them. If you are measuring impressions as a proxy for brand awareness, say so. If you are measuring brand awareness as a proxy for purchase intent, say so. The chain of assumptions is not automatically wrong. But making it visible means it can be challenged, which is where the analytical value sits.
The third habit is to build in market context as a default. Your performance metrics should always be read against category benchmarks, competitor data where available, and macroeconomic conditions. An absolute number without context is not a performance metric. It is a starting point for a conversation that has not happened yet.
User behaviour tools like Hotjar’s feedback and growth tools can help bridge the gap between surface-level metrics and actual user intent, which is one practical way to add context to data that would otherwise be read in isolation. Behavioural signals are not a substitute for causal analysis, but they add texture that pure conversion data does not provide.
The fourth habit is to treat your attribution model as a hypothesis, not a fact. Whichever model you use, it is making assumptions about how credit should be distributed. Those assumptions may be reasonable. They are not proven. Running multiple attribution views in parallel and looking for where they agree and disagree tells you far more than committing to a single model and treating its output as ground truth.
When Correlations Are Useful Without Being Causal
It would be an overcorrection to conclude that correlations are only useful when causation can be proven. In practice, causal proof is extremely difficult to establish in marketing, and waiting for it before making decisions is not a viable operating model.
The honest position is that correlations are useful for generating hypotheses, prioritising where to look, and making probabilistic decisions under uncertainty. They become problematic when they are treated as confirmation rather than as evidence worth investigating further.
If your content investment correlates consistently with pipeline growth across multiple periods and multiple segments, that is worth taking seriously even if you cannot prove causation definitively. The correlation is not proof, but it is a signal that warrants continued investment and more rigorous testing. The difference between useful correlation and misleading correlation is often not the data itself. It is the confidence with which conclusions are drawn from it.
Creator-led campaigns are a good example of this complexity. The Later webinar on go-to-market with creators gets into the mechanics of how creator content connects to conversion, which is a genuinely difficult attribution problem. The correlation between creator reach and sales is real in many cases. Whether the creator caused the sale or was simply present at the right moment in a purchase experience that was already underway is a much harder question to answer.
Honest marketing practice means holding both things at once: using correlations as useful signals while maintaining the analytical discipline to question what they actually prove. That tension is not comfortable. But it produces better decisions than either ignoring the data or treating it as more certain than it is.
If you want to see how this kind of thinking applies across the full range of go-to-market decisions, from channel selection to measurement frameworks to growth planning, the Go-To-Market and Growth Strategy hub covers the strategic layer that sits behind the analytical choices.
About the Author
Keith Lacy is a marketing strategist and former agency CEO with 20+ years of experience across agency leadership, performance marketing, and commercial strategy. He writes The Marketing Juice to cut through the noise and share what works.
