By 2026, marketing measurement is no longer about colourful dashboards or last-click reports that make every campaign look profitable. Privacy restrictions, signal loss and fragmented customer journeys have exposed a simple truth: correlation is not proof. For small businesses in particular, the key question is not “Which channel got the conversion?” but “Would this sale have happened anyway?” This article explains, in practical terms, the difference between attribution, Marketing Mix Modelling (MMM) and incrementality. It then outlines realistic experiment designs – geo-split, holdout audiences and time-based tests – and shows how to calculate minimum budgets, sensible durations and how to read results without convincing yourself you have found growth where there is none.
Attribution models, including last-click, first-click or data-driven approaches inside ad platforms, attempt to assign credit for a conversion to touchpoints along the user journey. They answer the question: “Which interactions were present before the sale?” What they do not reliably answer is whether those interactions caused the sale. In a privacy-constrained environment, attribution is often based on modelled or partial data. It is useful for optimisation within a channel, but it is not designed to prove causal lift.
Marketing Mix Modelling (MMM) works at a higher level. It uses aggregated time-series data – spend, impressions, seasonality, promotions, pricing, macro factors – to estimate how different channels contribute to overall sales. Modern MMM approaches in 2026 are more privacy-safe because they rely on aggregated data rather than user-level tracking. MMM is particularly helpful for understanding budget allocation across channels over months or quarters. However, it requires sufficient data volume and statistical discipline; for very small businesses with limited variation in spend, results can be unstable.
Incrementality, by contrast, is about causality. It asks: “What is the additional revenue generated because we ran this campaign, compared to not running it?” Instead of assigning credit retrospectively, incrementality uses experiments. By creating a control group that does not see the treatment and comparing outcomes, you can estimate true lift. For a small business, incrementality does not require complex modelling. It requires discipline in experimental design and a willingness to accept uncomfortable results.
If you are managing daily bids and creatives in a paid search account, attribution data is still useful. It helps you identify which keywords or audiences are associated with conversions and where to shift budget in the short term. Treat it as an operational tool, not as a strategic proof of impact.
If you are allocating budget between paid search, paid social, email, affiliates and offline media over a financial year, MMM becomes relevant. Even a simplified regression using monthly data can reveal diminishing returns or saturation effects. For small businesses, a lightweight MMM built in a spreadsheet or with open-source tools can be sufficient, provided the data is clean and spans at least 18–24 months.
If you are considering increasing spend in a specific channel and want to know whether it truly drives additional sales, incrementality testing is the most direct method. It is especially valuable for channels that often capture existing demand, such as branded search or retargeting. Here, the risk of “paying for what you would have received anyway” is high, and only a controlled test can give you confidence.
The geo-split experiment is one of the most accessible designs. You divide your market into comparable geographic areas – for example, regions, cities or postcode clusters. In test regions, you increase or introduce spend in a channel. In control regions, you keep activity at business-as-usual levels. After a defined period, you compare sales trends between the two groups. The key requirement is similarity: historical sales patterns in test and control regions should move in parallel before the experiment.
Holdout audience testing is common in digital channels. You intentionally exclude a random portion of your target audience from seeing ads. For example, 10% of your CRM list receives no paid social ads, while 90% does. Over several weeks, you compare purchase rates. The difference between exposed and non-exposed groups represents incremental lift. This design is particularly effective for email, remarketing and customer reactivation campaigns.
Time-based experiments are simpler but require caution. You run a campaign during a defined “on” period and pause it during an “off” period, ideally multiple times. By comparing performance in on versus off windows, while controlling for seasonality and promotions, you estimate incremental impact. This approach works best when demand is relatively stable and when no major external events distort results.
For geo-split tests, ensure pre-test validation. Analyse at least 8–12 weeks of historical data to confirm that test and control regions have similar baseline trends. If one region is growing faster than another before the experiment, your result will be biased. In more advanced setups, synthetic control methods can be used to construct a weighted control group that mirrors the test region’s historical behaviour.
For holdout tests, randomisation is critical. The control group must be randomly assigned, not manually selected. Avoid excluding only “low-value” users from exposure, as this will artificially inflate lift. Maintain identical pricing, offers and messaging across groups; the only difference should be media exposure.
For time-based experiments, use multiple cycles where possible: for example, two weeks on, two weeks off, repeated three times. This reduces the chance that one-off fluctuations drive your outcome. Always annotate external factors such as public holidays, price changes or stock issues. Without this discipline, you risk attributing normal volatility to marketing impact.

A common mistake is running tests that are too small to detect meaningful differences. Before launching, define your minimum detectable effect (MDE). If your baseline weekly sales are £50,000 and you would only consider the campaign successful if it drives at least a 5% uplift, your MDE is £2,500 per week. Using historical variance in weekly sales, you can estimate how many weeks are required to detect that uplift with acceptable statistical confidence.
As a rule of thumb for small businesses, aim for tests that can generate at least several hundred conversions in both test and control groups combined. If your channel only produces 20 conversions per week, a two-week test will not provide reliable evidence. Either extend the duration or increase the budget so the signal is large enough relative to natural variability.
Budget should be proportional to the expected lift and revenue impact. If your gross margin on incremental sales is 40%, and you expect an additional £10,000 in revenue from the test period, spending £15,000 to prove it makes little financial sense. Measurement should not cost more than the decision it informs. In many cases, a modest but well-designed experiment is more valuable than a large, uncontrolled campaign.
First, focus on absolute incremental revenue, not on platform-reported return on ad spend. If your test region generated £120,000 and your control region £110,000, but historically they differ by £8,000 on average, your true incremental lift may be only £2,000. Always adjust for baseline differences observed before the experiment.
Second, consider confidence intervals rather than single-point estimates. If the estimated lift is 6% but the confidence interval ranges from -2% to +14%, you do not have strong evidence of positive impact. Accepting uncertainty is part of responsible measurement. Scaling a channel based on statistically weak evidence increases financial risk.
Finally, separate statistical significance from business significance. A 1% uplift on a very large revenue base may be statistically robust but commercially irrelevant once media costs are deducted. Conversely, a 7% uplift with moderate uncertainty might justify a phased scale-up if the potential upside is meaningful. The discipline is not in finding positive numbers; it is in deciding whether those numbers represent real, incremental growth that improves profit.