Everyone Is Talking about MMM. That is the Problem.

Marketing Mix Modeling is having a moment. Open any marketing newsletter, attend any growth conference, and you will hear the same thing: MMM is back, it is affordable now, it is the answer to the post-cookie measurement problem.

Here is what those conversations keep leaving out. MMM alone is not the answer. It never was. Brands that invest in MMM as their single measurement solution are trading one blind spot for a slightly different one.

The real answer is triangulation. A measurement stack that uses five methods together, each designed to answer a different question, each covering what the others cannot see. No single tool does this on its own. The brands getting measurement right know exactly which method to reach for at which moment, and they have the platform to run all five from the same data foundation.

Here is why that matters, and what it looks like in practice.


Why MMM alone leaves you with blind spots

MMM is a macro tool. It looks at historical spend data over time and estimates how much each channel contributed to revenue. The modern versions, Google's open-source Meridian and Meta's Robyn, have brought the cost down so that mid-market brands can now run credible models. That is genuinely good news.

But MMM has hard limits that the hype cycle glosses over:

  • It cannot optimize at the placement level. MMM will tell you that TV is working. It will not tell you that your primetime NBC buy outperformed your late-night cable rotation by three to one. It cannot tell you which podcast show drove results versus which one generated noise. It sees channels, not individual placements.
  • It needs your spending to vary over time. If you run TV and social at the same budget every single week, the model cannot separate their effects. It needs on and off patterns to work. This is a real operational constraint that most media plans do not naturally create.
  • It was built for scheduled, concentrated media. A TV spot airs at 9pm on Tuesday and the whole market sees it at once. That is what MMM was designed to detect. Programmatic ads dripped across millions of auctions around the clock, podcast impressions scattered across a six-month back catalog, algorithmically-delivered social content, none of these have a concentrated moment of exposure for MMM to pick up cleanly.
  • It measures correlation, not cause. MMM can show you that spend and revenue moved together. It cannot prove that the spend caused the revenue. Without testing to validate it, you are working with a sophisticated hypothesis.
MMM is the right tool for setting strategy and allocating budget across channels. It is the wrong tool for placement optimization, proving incrementality, measuring channels with no trackable exposure, and creative-level insight. Using it as your only measurement tool is like navigating with just a compass when you need a compass, a map, a GPS, and someone who knows the local roads.

The five methods, and what each one is actually for

A complete measurement stack has five components. They are not competing methods. Each one answers a question the others cannot.

  1. MMM: Strategic budget allocation Best for annual and quarterly decisions about how much to spend across channels. Also the only method that can pick up signals from channels where people never click anything, like linear TV, radio, and billboards. Honest limit: Cannot optimize within a channel. Cannot prove causality on its own.
  2. MTA / CAPI: Daily campaign performance Multi-touch attribution tracks the individual user journey across touchpoints. Server-side implementations like Meta's Conversions API (CAPI) and Google's Enhanced Conversions have recovered a lot of the signal lost after Apple's iOS privacy changes. Best for paid search, email, and retargeting, channels where users click something you can track. Honest limit: Over-credits the platform running the attribution. Cannot answer whether the sale was incremental.
  3. Geo holdout testing: Proving incrementality for contained media Splits comparable markets into exposed and holdout groups, runs media in the exposed markets only, and measures the true revenue difference. This is the closest thing to a controlled experiment in marketing. It answers the question MMM cannot: would those customers have bought anyway? Honest limit: Defeated by social and programmatic advertising, where the algorithm delivers across cookie pools that do not respect geographic lines. Spillover and shared audiences contaminate the test.
  4. Spike lift analysis: TV and streaming placement optimization Matches individual ad airings to branded search spikes in the 15 to 20 minute window right after the spot airs. When someone sees your ad and immediately searches for your brand, that search is the cleanest signal you have that the creative landed. Spike lift tells you which programs, which time slots, and which networks drive the strongest response, at a granularity MMM will never reach. CoreMedia Systems, now part of Simpli.fi, is the specialist tool here, providing direct response analytics and spot-level attribution for linear TV and CTV. Honest limit: Only works when there is a known, specific moment of exposure. Programmatic, social, and podcast DAI (dynamic ad insertion) do not have that, so spike lift does not apply.
  5. HDYHAU: What the customer actually remembers A post-purchase survey asking "How did you hear about us?" This is underrated by performance marketers, but it captures something no model can: the consumer's own memory of what moved them to buy. It is the most reliable signal available for channels that leave no trackable footprint, podcast host reads, influencer content, word of mouth, PR. These channels (sometimes called "dark channels" because they are invisible to click-based tracking) drive real behavior but generate nothing for your attribution platform to detect. Honest limit: Only motivated customers complete surveys. Response rates vary. Needs to run consistently over time to be statistically useful. Fairing is the specialist platform here, with native Shopify integration and compatibility with the major attribution platforms including Rockerbox and Northbeam, making it easy to get HDYHAU data flowing into whichever measurement stack you are running.

The channels where the measurement problem is most serious

When you look at where the money actually goes and map it against these five methods, the picture is uncomfortable. The fastest-growing, highest-spend channels are the hardest to measure independently.

  • Paid social, Meta, TikTok, Snap ($83B in 2025, up 20% year over year): Geo holdout is defeated by algorithmic delivery and content that spreads across networks regardless of geography. Platform-native lift tests are methodologically sound, but the platform is grading its own homework. There is no reliable independent measurement method for prospecting on these channels. That is $83 billion largely operating on platform-reported numbers.
  • Retail media, Amazon and Walmart ($54B, up 20%): Amazon shows you a closed loop: someone saw your sponsored product, searched, and bought. The attribution chain within the platform is real. What it cannot tell you is whether that customer would have found you through organic search anyway, or whether you are now paying for placement you used to earn on merit. Brands that run controlled tests, pausing spend on their strongest organic products for a defined period, consistently find they were cannibalizing themselves. The platform dashboard looks like accountability. The incrementality question is untouched.
  • Programmatic display and digital video ($85B combined): Geo holdout is defeated by national cookie pools. Continuous drip delivery defeats spike lift. MMM signal is weak because there is no on and off pattern. This is $85 billion with no clean independent path to understanding true impact.
The brands winning on measurement are not the ones who found the best single tool. They are the ones who stopped looking for one.

Every channel mapped against all five methods

2025 US ad spend, measurement signal quality, and honest best method for each channel

High signal Medium Low None Best method Defeated by spillover No reliable method Walled garden
Channel 2025 Spend YoY MMM MTA / CAPI Geo holdout HDYHAU Spike lift Honest best method
Performance digital: search well measured, rest problematic
Paid search
(Google / Bing)
$124B
+14% Med High None None None MTA / CAPI. Intent is deterministic, attribution is clean
Retail media
(AMZN / WMT)
$54B
+20% Low Walled garden None None None No reliable method. Cannibalization of organic placement unknown
Display / DSP
(DV360 / Trade Desk)
$54B
+8% Low Med Defeated None None No reliable method. Cookie pools defeat geo, delivery is continuous
Digital video
(YouTube / online video)
$31B
+17% Low Med Defeated Low None No reliable method. Algorithmic delivery defeats geo isolation
Paid social: largest measurement blind spot
Social media
(Meta / TikTok / Snap)
$83B
+20% Med Walled garden Defeated Med None No reliable independent method. Platforms hide incrementality by design
Television: best measured traditional medium
Linear TV
(NBC / CBS / Cable)
$59B
-19% High None High Med High MMM (budget) + geo holdout (incrementality) + spike lift (placement)
CTV / streaming TV
(Hulu / Peacock / Disney+)
$29B
+18% Med Low Med Low Med Spike lift via ACR data + MMM for budget allocation
Direct mail: underrated measurement story
Direct mail
(catalogs / inserts / EDDM)
$37B
flat Med None High Med None Geo holdout. Hard physical boundary, no spillover possible
Audio: mixed picture
Linear radio
(iHeart / Cumulus / local)
$11B
-5% High None Med Med Med MMM for budget + spike lift for daypart optimization
Podcast
(Spotify / Apple / iHeart)
$2.5B
+16% Med Med High High None Geo holdout (baked-in) + HDYHAU + Podscribe MTA (DAI)
Out-of-home: geo is the only answer
OOH / DOOH
(Lamar / Clear Channel / Outfront)
$11B
+5% Med None High Med None Geo holdout. Location-fixed media, clean market isolation

Sources: Winterberry Group, PwC, eMarketer, Marketing Charts 2025. "Walled garden" = platform reports attribution within its own system but cannot answer whether the sale was incremental. "Defeated" = geographic or spillover contamination makes geo holdout structurally unreliable for that channel. Linear TV trend reflects continued structural decline ex-political spend.

How a triangulated stack works in practice

Running five methods independently across disconnected tools is not practical for most marketing teams. The stack only works when the data foundation is unified, consistent inputs feeding all five methods from a single source. Here is how the methods divide the workload:

  • MMM runs quarterly for strategic budget allocation. It tells you how much to spend on TV versus social versus search versus podcast at the channel level.
  • Spike lift runs continuously for TV and premium streaming. It tells you which programs and time slots drive the strongest response, giving your media buyer the signal to optimize the actual buy.
  • Geo holdout runs periodically for linear TV, direct mail, and OOH. One well-designed holdout test per channel per year also gives your MMM model a real calibration anchor, making both methods more reliable.
  • MTA and CAPI run daily for paid search, email, and retargeting, the channels where the user journey is trackable and deterministic attribution works.
  • HDYHAU runs at checkout continuously, weighted against conversion volume and tracked over time to detect when podcast, influencer, or word of mouth is moving the needle in ways no platform can see.

The real value shows up when methods agree and when they disagree. When MMM says TV is working and spike lift confirms the specific placements, you have both strategic and tactical validation. When MMM says social is efficient but almost no one mentions it in post-purchase surveys, you have a hypothesis worth testing. Disagreement between methods is not a problem. It is a signal.


Platforms building toward this

A handful of platforms have built toward unified measurement. None covers all five methods perfectly, but the best ones are closing the gap fast.

MTA + MMM + Incrementality + Spike Lift

The most complete unified platform for mid-market and enterprise brands. Combines MTA, MMM, and incrementality testing on a single data foundation with 100+ integrations covering digital and offline channels including TV, direct mail, and podcast. Importantly, Rockerbox will accept spike lift attribution data from a third-party supplier and incorporate it directly into the model. In practice this means pairing Rockerbox with a TV attribution platform like CoreMedia Systems (now part of Simpli.fi), which provides direct response analytics and spot-level attribution for linear TV and CTV, giving you a genuinely unified view across all five measurement methods in one place. Built to show where methods agree and where they diverge. Acquired by DoubleVerify in 2024, which raises questions about future direction worth monitoring.

Incrementality-first

Built around proving incrementality through structured experiments, with MMM and attribution layered in. Purpose-built for brands where proving causal impact to the CFO is the primary goal. Strong enterprise customer ratings. Less suited for teams that need fast daily tactical optimization.

MTA + MMM for DTC

Strong for DTC and ecommerce brands with substantial paid digital budgets. Creative-level attribution granularity that media buyers actually use day to day. Expanded into MMM in 2025 and continues to mature its offline measurement capabilities. Less equipped for brands with heavy offline media mixes that include TV, OOH, and podcast.

Spike lift specialist

The specialist tool for TV and CTV spike lift. CoreMedia, now part of Simpli.fi, provides direct response analytics and spot-level attribution for linear TV, CTV, and radio, matching individual airings to response data in near real time. Gives media buyers the placement-level signal to move beyond ratings-based buying toward outcome-based decisions. Works directly with Rockerbox, feeding spot-level attribution data into the unified model so spike lift does not sit in a separate silo.

HDYHAU specialist

The best purpose-built post-purchase survey platform available and the natural choice for the HDYHAU layer of the stack. Fairing integrates natively with Shopify and delivers category-leading response rates of 40 to 80 percent. Survey data syncs to your data warehouse, Klaviyo, and the major attribution platforms including Rockerbox and Northbeam, making it compatible with whichever unified measurement stack you are running. For DTC and ecommerce brands, it is the cleanest way to capture what customers actually remember about how they found you, covering the dark channels that no attribution model can see on its own.

Podcast attribution specialist

The leading podcast attribution platform, using IP-address matching to connect podcast ad exposures to website conversions. Podscribe captures a meaningful share of podcast-driven conversions that would otherwise appear as direct or organic traffic and be invisible to standard MTA platforms. Works across both baked-in host reads and DAI programmatic placements, and integrates with unified measurement platforms so podcast attribution does not sit in a separate silo. Match rates typically run between 40 and 60 percent of true impact due to VPN usage and cross-device journeys, so Podscribe works best when combined with HDYHAU surveys and branded search lift monitoring for a fuller picture of podcast performance.

Google Meridian / Meta Robyn
Open-source MMM

For brands with in-house data science resources. Both are Bayesian MMM frameworks that can be run without a vendor. Meridian in particular is designed to ingest geo holdout results as calibration inputs, making the model more accurate over time. The tradeoff is the analyst resources required to build, run, and interpret them.


The bottom line

MMM is a great tool. It is one instrument in an orchestra, not the whole band.

The measurement problem in 2026 is not that good tools do not exist. It is that the channels eating the most budget are structurally resistant to independent measurement, and the tools best suited to answer the real question, did this spend actually cause incremental revenue, are being applied to the wrong channels or used in isolation.

The brands solving this build a stack where:

  • MMM sets strategic allocation and picks up signals from channels where people never click anything
  • Spike lift optimizes placement within TV and streaming at a granularity MMM cannot reach
  • Geo holdout provides causal validation for the channels where it works cleanly
  • MTA and CAPI handle the deterministic, click-based channels with daily precision
  • HDYHAU captures what no model can, the consumer's own account of what moved them

And they run it all from a unified data foundation. Rockerbox is worth calling out specifically here: it will accept spike lift attribution data from a supplier like CoreMedia Systems (now part of Simpli.fi) and incorporate it directly into the model, which means you can get close to all five methods feeding a single platform rather than managing five disconnected dashboards producing five different answers.

For pre-IPO brands, the measurement story you tell investors matters as much as your growth numbers. A CMO who can speak to triangulated incrementality, here is what MMM shows, here is what our geo holdout confirmed, here is what customers tell us they remember, is a fundamentally different conversation than one showing platform screenshots.

Everyone is talking about MMM. The brands getting ahead are building the stack that makes MMM mean something.


How AI makes the five-method stack a practical reality

There is an obvious objection to the triangulated measurement stack we have described. Running five methods is one thing. But five methods means five data sources, five separate outputs, and five different answers sitting in five different places. A human analyst still has to make sense of all of it. That synthesis work is where most measurement programs break down in practice.

This is where AI changes the equation. Not by replacing the measurement methods, and not by building better models. By doing the interpretation work that currently falls on a person who may not have the time, the context, or the statistical fluency to do it well.

Think about what that synthesis actually requires. Your MMM says TV is working. Your geo holdout confirms social is not driving the incremental revenue Meta's dashboard claims. Spike lift shows Monday night cable outperforming primetime. Your Fairing surveys show one in three customers first heard about the brand through a podcast. Your MTA dashboard shows paid search as the top converter. Every one of those signals is true. None of them tells the complete story on its own. Turning them into a coherent budget recommendation that the CMO can defend to a CFO and a board, is a substantial analytical task that most marketing teams are not resourced to do consistently.

Connect your measurement data to an AI layer, including Rockerbox exports to a data warehouse, Fairing survey responses, MMM outputs, and geo holdout results, and that synthesis becomes a conversation rather than a project. You ask what is working and what the data suggests you should do differently. The AI reads across all five sources, identifies where the methods agree, surfaces the divergences that warrant investigation, and drafts a narrative ready for a leadership meeting. Work that previously took a senior analyst two days now takes ten minutes.

The human still makes the decisions. That is the point. AI handles the aggregation and pattern recognition so the marketer can focus on judgment: which channels to trust, which divergences matter, and which budget moves make strategic sense given things the data cannot see, like an upcoming product launch or a competitive shift. The measurement stack provides the evidence. AI assembles it into a brief. The human acts on it.

For pre-IPO brands this matters in a concrete way. Investors and boards want marketing accountability that goes beyond platform dashboards. A CMO who can walk into a board meeting with a synthesis across five independent measurement methods, covering what each one shows, where they agree, and what is being done about the gaps, is telling a fundamentally more credible story than one presenting Meta's reported ROAS. AI is what makes that level of synthesis operationally possible without a dedicated data science team.

The measurement stack gives you the truth. AI turns it into the story your organization can act on.

The brands that figure this out first, and the platforms that build the connectors to make AI and measurement data talk to each other natively, will have a durable advantage over everyone still interpreting five dashboards manually and hoping the numbers tell a consistent story.


Glossary: key measurement terms explained

These are the terms that come up most in any serious measurement conversation. Each one is worth understanding on its own before trying to triangulate across all five methods.

What is Media Mix Modeling (MMM)?
Media Mix Modeling is a statistical technique that analyzes historical media spend data over time to estimate how much each channel contributed to revenue. It works at an aggregate level, looking at patterns across weeks or months, rather than tracking individual users. MMM is privacy-safe because it uses aggregated data, not user-level signals, which is one reason it has resurged as cookies and mobile tracking have degraded. Modern Bayesian versions like Google's Meridian and Meta's Robyn have made it accessible to mid-market brands that previously could not afford it. MMM is best used for strategic budget allocation across channels. Not for daily optimization or placement-level decisions.
What is Multi-Touch Attribution (MTA)?
Multi-Touch Attribution tracks the individual user journey across marketing touchpoints and assigns credit to each interaction that preceded a conversion. Unlike last-click attribution which credits only the final touchpoint, MTA distributes credit across the full path, from first awareness ad through to purchase. MTA requires user-level tracking data, which is why it has been weakened by Apple's iOS privacy changes and the deprecation of third-party cookies. Modern implementations use server-side tracking and platform Conversions APIs (CAPI) to recover signal. MTA is best suited for digital channels where clicks and sessions are trackable, including paid search, email, and retargeting, and less reliable for offline or view-through channels.
What is CAPI (Conversions API)?
CAPI stands for Conversions API. It is a server-side tracking implementation that sends conversion event data directly from a brand's server to an advertising platform, most commonly Meta's Conversions API or Google's Enhanced Conversions, rather than relying on a browser-based pixel. Because the data travels server to server rather than through a browser, it is not affected by ad blockers, iOS tracking restrictions, or cookie deprecation. CAPI significantly improves the signal quality available to platform attribution models and ad delivery algorithms, particularly for Meta advertising. It does not solve the incrementality question. CAPI tells you what happened, not whether your ad caused it.
What is a geo holdout test?
A geo holdout test is a controlled marketing experiment that splits comparable geographic markets into two groups: exposed and holdout. Media runs in the exposed markets only, while the holdout markets receive no advertising. By comparing revenue or conversion rates between the two groups during the test period, you can isolate the true incremental lift driven by the campaign. It is the closest thing to a randomized controlled trial in marketing, and it answers the question that MMM and MTA cannot: would those customers have bought anyway? Geo holdout works best for media that is physically contained within geographic boundaries, including linear TV, direct mail, OOH, and radio. It breaks down for algorithmically-delivered media like social and programmatic, where spillover and shared cookie pools contaminate the holdout group.
What is spike lift analysis?
Spike lift analysis is a measurement technique that matches individual ad airings to branded search volume spikes in the 15 to 20 minute window immediately following the spot. When a TV ad airs at 9pm, a measurable increase in branded search queries in the minutes after is one of the cleanest available signals that the creative reached and engaged the audience. Spike lift works at the placement level, comparing programs, dayparts, and networks, which gives media buyers the signal to optimize actual buying decisions in a way that MMM cannot. It requires a known, discrete exposure timestamp, which is why it works for linear TV and premium CTV with ACR data, but not for programmatic, social, or podcast DAI where impressions are dripped continuously with no single moment of concentrated delivery.
What is HDYHAU?
HDYHAU stands for "How Did You Hear About Us?" This is a post-purchase survey question shown to customers immediately after they complete a transaction. Despite being one of the oldest measurement tools in marketing, it remains one of the most underrated. It captures self-reported attribution that no tracking model can replicate. It captures the consumer's own memory of what first made them aware of a brand. HDYHAU is particularly valuable for dark channels including podcast host reads, influencer content, word of mouth, and PR, that leave no trackable digital footprint. Fairing is the leading specialist platform for HDYHAU surveys, delivering 40 to 80 percent response rates for Shopify brands and integrating with major attribution platforms including Rockerbox and Northbeam.
What are dark marketing channels?
Dark channels is a term used in marketing measurement to describe advertising touchpoints that are invisible to click-based tracking systems. When someone hears a podcast host read, sees a billboard, watches a linear TV spot, or discovers a brand through word of mouth, they may convert days later through a direct website visit or a branded search query, with no trackable link between the exposure and the purchase. Attribution platforms record the conversion but cannot see the cause. Dark channels are often the most powerful drivers of brand awareness precisely because they operate outside the algorithmic feeds where most performance marketing lives. Measuring them requires a combination of geo holdout testing, HDYHAU post-purchase surveys, and MMM, none of which rely on click tracking.
What is OOH advertising?
OOH stands for Out-of-Home advertising, meaning any paid media that reaches consumers outside their homes. This includes billboards, transit advertising, street furniture, place-based screens in airports and gyms, and digital out-of-home (DOOH) displays. OOH is a physically location-fixed medium, which makes it one of the most cleanly measurable traditional channels through geo holdout testing. Because the ad can only be seen in the specific location where it is placed, you can construct exposed and holdout markets with genuine geographic separation. DOOH in particular has grown significantly, with digital screens now accounting for nearly half of total OOH spend in the US, enabling dynamic creative and daypart targeting that was impossible with static print.
What is ACR (Automatic Content Recognition)?
ACR stands for Automatic Content Recognition, a technology embedded in smart TVs and streaming devices that identifies what content is being watched by analyzing audio or visual fingerprints. For advertising measurement, ACR data is valuable because it creates a device-level record of ad exposure including which specific ads were seen on which device at what time. This exposure data can be matched against conversion events to build view-through attribution for CTV advertising, which is not possible with traditional linear TV measurement. ACR data enables spike lift analysis on premium streaming platforms where a discrete exposure timestamp can be established, making CTV more measurable than programmatic video formats where impressions are dripped continuously.
What is incrementality testing?
Incrementality testing is a method of measuring the true causal impact of advertising by comparing outcomes between a group that was exposed to an ad and a control group that was not. Unlike attribution, which observes who converted after seeing an ad, incrementality testing isolates whether the ad actually caused the conversion, or whether those customers would have purchased anyway. Geo holdout tests and audience-based holdout tests are the two main forms. Incrementality is considered the most accurate available measure of advertising effectiveness, but it is operationally expensive, takes time to produce results, and is structurally defeated for algorithmically-delivered media like paid social and programmatic where control groups cannot be cleanly isolated from exposed groups.
What is Podscribe and how does podcast attribution work?
Podscribe is a podcast attribution platform that uses IP-address matching to connect podcast ad exposures to website conversions. When a podcast ad is served, Podscribe logs the listener's IP address. When that same IP address later visits the advertiser's website and converts, Podscribe matches the two events and attributes the conversion to the podcast impression. This server-side approach captures a significant portion of podcast-driven conversions that would otherwise appear as direct or organic traffic, invisible to standard MTA platforms. The honest limitation is that IP matching is probabilistic, not deterministic. VPN usage, carrier-grade NAT where thousands of users share a single IP, and cross-device journeys where someone listens on a phone and converts on a laptop all reduce match rates, typically to somewhere between 40 and 60 percent of true impact. Podscribe works best for baked-in host reads on shows with measurable branded search lift, and its data can be fed into unified measurement platforms to improve podcast visibility within a broader attribution model.
Marc Viale is the founder of LAB415, a growth acceleration consultancy for pre-IPO companies in the San Francisco Bay Area. LAB415 specializes in full-funnel marketing strategy with deep expertise in TV, CTV, performance measurement, and building the marketing infrastructure that supports a successful IPO. lab415.com
Previous
Previous

Your reviews are your reputation. Own them.

Next
Next

Why would Quince sell a $50 cashmere sweater?