EP7: No Interference, No Ambiguity: The SUTVA Assumption

causal inference
identification
assumption
SUTVA
Author

Lin Jia

Published

April 3, 2026

Randomization guarantees balanced, comparable groups. It does not guarantee that what you are comparing is well defined.

Consider an experiment on a referral programme. Users are randomly assigned to receive a referral link or not. The experiment looks clean — randomized groups, clear outcome. But referral programmes do not stay confined to individual users. User A gets a link, shares it with user B, and suddenly B’s outcome depends on A’s treatment. The experiment is contaminated.

The natural instinct is to fix this after the fact — adjust the model, add covariates, filter out the contaminated users. But as Fisher famously warned (Fisher 1938):

To consult the statistician after an experiment is finished is often merely to ask him to conduct a post mortem examination.

No amount of statistical adjustment will rescue a broken design. This is a violation of SUTVA — the Stable Unit Treatment Value Assumption — and it is the one I see violated most often in practice. When it breaks, you get confident numbers that mean nothing — you scale campaigns that never worked, or kill features that actually did.

The two components of SUTVA — no interference and consistency.

Listen to This Episode

Want to dive deeper into SUTVA? This episode of the podcast walks through the intuition and practical implications of the assumption.

1. What SUTVA Requires

The causal effect for user \(i\) is \(\tau_i = Y_i(1) - Y_i(0)\) — the difference between what would happen with and without the referral link (for a full introduction to potential outcomes, see EP4).

But writing \(Y_i(1)\) quietly assumes each user has exactly one potential outcome under treatment. For that to hold, two things must be true: user \(i\)’s outcome cannot depend on what treatment other users received (no interference), and “receiving the referral link” must mean the same thing for every user (consistency).

That is what SUTVA ensures (Rubin 1980). When either component breaks, \(Y_i(1)\) is no longer a single well-defined number — and neither is the causal effect.

2. No Interference

No interference requires that user \(i\)’s potential outcome \(Y_i(t_i)\) depends only on their own treatment assignment. In our referral example, B’s potential outcome can no longer be written as \(Y_B(t_B)\) — it becomes \(Y_B(t_A, t_B)\).

In practice, interference arises through three distinct pathways. Understanding which one applies determines how you fix it.

The three mechanisms of interference — direct contact, indirect competition, and behavioral redirection.

Direct Contact: Network Effects and Contagion

User A receives a referral link, shares it with friend B, and B’s behaviour changes — even if B was in the control group. The treatment spreads through social connections, not through the experimental design. The same happens in public health: vaccinating part of a population protects unvaccinated neighbours through herd immunity (Aronow and Samii 2017; Halloran and Struchiner 1995).

When one unit’s treatment directly changes another unit’s outcome through a network, the standard fix is to randomize at a higher level — cities or regions instead of individuals — so network interactions stay within the randomization unit (Toulis and Kao 2013; Hudgens and Halloran 2008). The trade-off: higher aggregation reduces statistical power, making it harder to detect small effects.

Direct contact interference — treatment leaks through networks. Fix: cluster randomization.

Indirect Competition: Shared Resources and Market Competition

Suppose our referral programme runs during a promotion with limited sign-up slots. A referred user claims a slot, fewer remain for non-referred users — even though the two never interact. The same pattern appears in two-sided marketplaces: riders competing for drivers, buyers competing for limited inventory. Treatment shifts demand, the market rebalances, and control users are affected through changes in availability, wait times, or prices (Blake and Coey 2014; Johari et al. 2022; Holtz et al. 2024).

When units compete for finite supply rather than interacting directly, a common fix is switchback experiments — alternating treatments over time within a unit, so each unit serves as its own control (Crépon et al. 2013; Bojinov, Simchi-Levi, and Zhao 2023). The caveat: carryover effects between time windows can introduce bias, requiring wash-out periods in the design.

Marketplace competition — units compete for finite supply. Fix: switchback experiments.

Behavioral Redirection: Spatial Displacement and Systemic Shifts

If our referral programme scales widely enough, word spreads that bonuses exist — and users who would have signed up organically start waiting for a referral link instead. The baseline shifts for everyone, not because of direct contact with treated users, but because the programme has changed expectations across the whole market (Sobel 2006; Manski 2013).

Now imagine a different promotion: a chain offers in-store discounts at randomly selected locations. Customers near non-discounted stores don’t stop shopping — some simply drive to a discounted store nearby. Demand spills over geographically, inflating outcomes at discounted locations and deflating them at control locations (Verbitsky-Savitz and Raudenbush 2012).

When the intervention doesn’t eliminate a behaviour but moves it somewhere else, the fix depends on the type. For systemic shifts, a traditional A/B test fails by construction — the intervention changes the baseline for everyone. Instead, launch to an entire market and measure against a synthetic control: a weighted combination of similar unaffected markets that serves as the counterfactual. For spatial displacement, randomize at the level of naturally separated markets — e.g., different cities or countries — so geographic spillover between treatment and control is negligible. Where that’s not possible, exclude border areas from measurement to create a buffer between groups.

Behavioral redirection — demand is moved, not created. Fix: buffer zones and synthetic controls.

3. Consistency

The second component of SUTVA is consistency: each treatment level must correspond to a single well-defined intervention.

At first glance this seems obvious. In practice, it is often violated.

Consider again the referral example. The treatment is defined as “sent a referral link.” But referrals may vary:

  • some include a monetary bonus
  • some offer a free trial
  • others contain no incentive

Two users assigned to treatment may therefore receive different interventions. If these differences affect outcomes, the potential outcome \(Y(1)\) is no longer uniquely defined. The treatment label hides multiple versions of the intervention.

As Hernan and Robins (2024) put it: “the more precisely we define the meaning of treatment \(T = 1\) and \(T = 0\), the more precise our causal questions are. We only need sufficiently well-defined interventions for which no meaningful vagueness remains.”

Consistency breaks in two ways: the treatment varies across users (some get a bonus, others get a free trial) or drifts over time (the incentive changes mid-experiment). Both are preventable — align with product and engineering on a precise treatment definition before the experiment is built, and verify the implementation matches the spec.

A quick self-check before sign-off: can you explain the treatment to a colleague in one sentence without using “it depends”? If the answer requires qualifications, those qualifications are likely hidden treatment versions.

The illusion of consistency — a single treatment label can hide multiple interventions.

4. Before You Run the Experiment

Before running an experiment, ask two questions:

  • Can one unit’s treatment affect another unit’s outcome?
  • Is the treatment precisely defined?

If the answer to either is uncertain, the experiment needs redesign — not a larger sample.

Here is a quick-reference summary of SUTVA violations — what breaks, what it costs, and how to fix it.

The SUTVA diagnostic matrix — recognizing and fixing violations in practice.

Keep Going

Up next: EP8: Identification by Conditioning

With exchangeability (EP5), positivity (EP6), and SUTVA (EP7), we have now covered all three identifiability assumptions. In the next episode, we bring them together — connecting these assumptions to the conditioning strategy that underpins observational causal inference and its link to randomized experiments and A/B testing.

Recommended background (optional)

References

Aronow, Peter M., and Cyrus Samii. 2017. “Estimating Average Causal Effects Under General Interference, with Application to a Social Network Experiment.” The Annals of Applied Statistics 11 (4): 1912–47.
Blake, Thomas, and Dominic Coey. 2014. “Why Marketplace Experimentation Is Harder Than It Seems: The Role of Test-Control Interference.” In Proceedings of the Fifteenth ACM Conference on Economics and Computation, 567–82.
Bojinov, Iavor, David Simchi-Levi, and Jinglong Zhao. 2023. “Design and Analysis of Switchback Experiments.” Management Science 69 (7): 3759–77. https://doi.org/10.1287/mnsc.2022.4583.
Crépon, Bruno, Esther Duflo, Marc Gurgand, Roland Rathelot, and Philippe Zamora. 2013. “Do Labor Market Policies Have Displacement Effects? Evidence from a Clustered Randomized Experiment.” The Quarterly Journal of Economics 128 (2): 531–80.
Fisher, Ronald A. 1938. “Presidential Address to the First Indian Statistical Congress.” Sankhyā: The Indian Journal of Statistics 4 (1): 14–17.
Halloran, M. Elizabeth, and Claudio J. Struchiner. 1995. “Causal Inference in Infectious Diseases.” Epidemiology 6 (2): 142–51.
Hernan, Miguel A, and James M Robins. 2024. Causal Inference: What If. Boca Raton: Chapman & Hall/CRC.
Holtz, David, Fradel Lobel, Ruben Lobel, Inessa Liskovich, and Sinan Aral. 2024. “Reducing Interference Bias in Online Marketplace Experiments Using Cluster Randomization: Evidence from a Pricing Meta-Experiment on Airbnb.” Management Science 71 (1): 390–406.
Hudgens, Michael G., and M. Elizabeth Halloran. 2008. “Toward Causal Inference with Interference.” Journal of the American Statistical Association 103 (482): 832–42.
Johari, Ramesh, Hannah Li, Inessa Liskovich, and Gabriel Y. Weintraub. 2022. “Experimental Design in Two-Sided Platforms: An Analysis of Bias.” Management Science 68 (10): 7069–89.
Manski, Charles F. 2013. “Identification of Treatment Response with Social Interactions.” The Econometrics Journal 16 (1): S1–23.
Rubin, Donald B. 1980. “Comment on Randomization Analysis of Experimental Data in the Fisher Randomization Test’ by Basu.” Journal of the American Statistical Association 75 (371): 591–93. https://doi.org/10.1080/01621459.1980.10477517.
Sobel, Michael E. 2006. “What Do Randomized Studies of Housing Mobility Demonstrate? Causal Inference in the Face of Interference.” Journal of the American Statistical Association 101 (476): 1398–1407.
Toulis, Panos, and Edward Kao. 2013. “Estimation of Causal Peer Influence Effects.” In Proceedings of the 30th International Conference on Machine Learning, 28:1489–97.
Verbitsky-Savitz, Natalya, and Stephen W. Raudenbush. 2012. “Causal Inference Under Interference in Spatial Settings: A Case Study Evaluating Community Policing Program in Chicago.” Epidemiological Methods 1 (1): 106–30.