Regression Discontinuity Designs (RDD) offer a powerful quasi-experimental approach to estimate causal effects in situations where treatment assignment isn't random but depends deterministically or probabilistically on whether an observed continuous variable (the "running" or "forcing" variable) crosses a specific threshold. This design is particularly valuable when we suspect unobserved confounders might bias naive comparisons, as RDD leverages the assignment rule itself to isolate the treatment effect locally around the cutoff.
Imagine a scenario where students are assigned to a remedial program (T=1) if their score (R) on an entrance exam falls below a certain cutoff (c), and receive the standard curriculum (T=0) otherwise. Comparing the average outcomes (Y) of all students in the remedial program versus all those in the standard program would likely be misleading due to pre-existing differences (confounding). Students with lower scores might differ in motivation, prior knowledge, or socioeconomic background, all of which could affect their final outcomes independently of the remedial program.
RDD exploits the discontinuity in treatment assignment at the cutoff c. The core idea is that individuals just below the cutoff (R=c−ϵ) and just above the cutoff (R=c+ϵ) are likely very similar in all relevant characteristics, both observed and unobserved, except for their treatment status. Any abrupt change, or "jump," in the average outcome observed precisely at the cutoff can then be attributed to the treatment itself.
There are two primary types of RDD:
The validity of RDD hinges on several assumptions:
Discontinuity in Treatment Assignment: The probability of treatment must change discontinuously at the cutoff c. This is definitional for RDD. In Sharp RDD, this jump is from 0 to 1. In Fuzzy RDD, the jump must be non-zero.
Continuity of Conditional Expectations: This is the most significant identifying assumption. It states that the average potential outcomes, E[Y(0)∣R=r] and E[Y(1)∣R=r], must be continuous functions of the running variable R around the cutoff c. In simpler terms, if the treatment had no effect, the plot of the average outcome against the running variable would show no jump at the cutoff. This assumption allows us to attribute any observed discontinuity in the actual outcome Y at c to the causal effect of the treatment T.
No Precise Manipulation of the Running Variable: Individuals should not be able to perfectly control their score R to strategically place themselves just above or below the cutoff. If manipulation occurs, individuals just below and just above the cutoff might differ systematically, violating the principle that they are comparable except for treatment status. A common diagnostic is the McCrary density test, which checks for a discontinuity in the density of the running variable R at the cutoff.
A smooth density plot around the cutoff suggests no manipulation. A significant jump or drop exactly at the cutoff would raise concerns.
Exclusion Restriction (Fuzzy RDD only): The cutoff c should only affect the outcome Y through its effect on the treatment T. This is analogous to the exclusion restriction in Instrumental Variables.
The goal is to estimate the magnitude of the jump in the outcome Y at the cutoff c. Since we typically don't have data exactly at the cutoff, estimation involves comparing outcomes for units just below and just above it.
Sharp RDD Estimation: The causal effect in Sharp RDD is the difference in the expected outcome just above and just below the cutoff:
τSRD=E[Y∣R=c,T=1]−E[Y∣R=c,T=0]=r↓climE[Y∣R=r]−r↑climE[Y∣R=r]The standard approach is local polynomial regression:
Fuzzy RDD Estimation: In Fuzzy RDD, the jump in the outcome needs to be scaled by the jump in the probability of treatment. It resembles an Instrumental Variable (IV) approach where the discontinuity assignment (being above vs. below the cutoff) serves as the instrument for the actual treatment received. The effect is estimated as the ratio of the jump in the outcome to the jump in the treatment probability:
τFRD=limr↓cE[T∣R=r]−limr↑cE[T∣R=r]limr↓cE[Y∣R=r]−limr↑cE[Y∣R=r]This is often estimated using a local IV framework (like two-stage least squares, 2SLS) within a bandwidth h around the cutoff. The first stage models the treatment T as a function of the cutoff indicator and polynomial terms of (R−c). The second stage models the outcome Y as a function of the predicted treatment from the first stage and polynomial terms of (R−c).
Bandwidth Selection: Choosing the bandwidth h is important. A smaller bandwidth reduces bias (by using observations closer to the cutoff, strengthening the continuity assumption) but increases variance (fewer data points). A larger bandwidth increases statistical power but risks introducing bias if the relationship between R and Y(0) or Y(1) is highly non-linear. Data-driven methods like Imbens-Kalyanaraman (IK) or Calonico-Cattaneo-Titiunik (CCT) are commonly used.
Plotting the data is essential for assessing RDD validity.
Outcome vs. Running Variable: Plot the average outcome Y within bins of the running variable R. Look for a clear jump at the cutoff c. Superimposing the fitted local polynomial lines helps visualize the estimated discontinuity.
Visualization of Sharp RDD. The jump in the fitted lines at the cutoff represents the estimated treatment effect. Data points show the underlying relationship.
Treatment vs. Running Variable (Fuzzy RDD): Plot the proportion of units receiving treatment within bins of R. Look for a jump at c to confirm the first stage in Fuzzy RDD.
Covariates vs. Running Variable: Plot pre-treatment covariates against R. These should not show a discontinuity at the cutoff. If they do, it casts doubt on the assumption that units just above and below the cutoff are comparable.
RDD estimates a Local Average Treatment Effect (LATE), specifically the average causal effect for the subpopulation whose running variable values are near the cutoff c. This effect might not generalize to individuals far from the cutoff, who could have different characteristics and potentially different treatment effect magnitudes.
RDD requires sufficient data density around the cutoff for reliable estimation. It's also sensitive to the choice of bandwidth and polynomial degree. Violations, such as manipulation of the running variable or the presence of other treatments or policies changing at the same cutoff, can invalidate the results.
While often seen in econometrics and policy evaluation, RDD concepts are relevant in ML:
RDD provides a rigorous framework for causal inference when assignment follows a threshold rule, offering a way to address unobserved confounding locally without needing to measure all confounders directly. Careful implementation and thorough diagnostic checks are essential for obtaining credible estimates.
© 2025 ApX Machine Learning