Study Notes on Bounding OVB 🌀 in Causal ML

Motivation

In empirical research, one of the challenges to causal inference is the potential presence of unobserved confounding. Even when we adjust for a wide range of observed covariates, there’s often a lingering concern: what if there are important variables we’ve failed to measure or include? This challenge necessitates a careful approach to sensitivity analysis, where we assess how strong unobserved confounders would need to be to meaningfully alter our conclusions.

image of Mafūba technique
Mafūba: A technique designed to seal demons by drawing them into a container, which is then secured with a special "Demon Seal" ofuda. Source: Image from Dragon Ball Fandom

Questions1

  1. How “strong” would a particular confounder (or group of confounders) have to be to change the conclusions of a study?

  2. In a worst case scenario how vulnerable is the study’s result to many or all unobserved confounders acting together, possibly nonlinearly?

  3. Are these confounders or scenarios plausible? How strong would they have to be relative to observed covariates (e.g female), in order to be problematic?

  4. How can we present these sensitivity results concisely for easy routine reporting?

TL;DR

Chernozhukov et al. introduces a framework for bounding omitted variable bias (OVB) in causal ML (specificially, DML) setting. They develop a general theory applicable to a broad class of causal parameters that can be expressed as linear functionals of the conditional expectation function. The key insight is to characterize the bias using the Riesz representation. Specifically, they express the OVB as the covariance of two error terms:

  1. The error in the outcome regression

  2. The error in the Riesz representer (RR)

This formulation leads to a bound on the squared bias that is the product of two terms:

  1. The MSE of the outcome regression
  2. The MSE of the RR

How to understand the Riesz representer (RR)?

  • We can derive it using Frisch–Waugh–Lovell (FWL) theorem

  • In a special case, i.e. the average treatment effect (ATE) with binary treatment, the Riesz representer (RR) “looks like” the weights in Inverse Propensity Score weighting (IPW)

Theoretical Details

Recall that:

For classical linear regression, “Short equals long plus the effect of omitted times the regression of omitted on included.”

– Angrist and Pischke, Mostly Harmless Econometrics.

In Linear Regression Setting

  • Linear model: Yi=α+βDi+γXi+δUi+ϵi, where Ui is an unobserved (scalar) confounding. Note that, if there are multiple confounders, we can regard Ui as a function of them, which are summarized as a single confounder.

  • Recall the omitted variable bias formula: β^pβ+δ×Cov(UiX,DiX)V(DiX), where

    • DiX is the residual from lm(DX), i.e. DiX=Dilm(DX)
    • Similarly, UiX is the residual from lm(UX)
    • The fraction term is the coefficient of residual-on-residual regression. It means how much variation of U can be explained by D condition on X. In other words, the association between U and D condition on X
    • δ is the association between Y and U condition on {D,X}
  • How to bound the bias?

    • image-20250129075223109
    • Partial R2 represents how much variation of Y can explained by U condition on {T,X}

In Causal ML Setting

  1. In the ideal case:

    • “long” regression: Y=θD+f(X,A)+ϵ

    • Define W:=(D,X,A) as the “long” list of regressors

    • A is unobserved vector of covariates

    • Assume E(ϵD,X,A)=0

  2. In the practical case:

    • “short” regression: Y=θsD+fs(X)+ϵs
    • Define Ws:=(D,X) as the “short” list of observed regressors because A is unmeasured/unobservable
    • Use θs to approximate θ; need to bound θsθ
  3. Conditional Expectation (outcome regression function)

    • “long” CET, g(W):=E(YD,X,A)=θD+f(X,A)

    • “short” CET, gs(W):=θsD+fs(X)

  4. Riesz Representers (RR):

    • “long” RR: α:=α(W):=DE[DX,A]E(DE[DX,A])2

      By FWL Theorem, we have θ=Cov(YD~)Var(D~)=E(yα), where D~=DE[DX,A],

    • “short” RR αs:=αs(WS):=DE[DX]E(DE[DX])2

    • In special case, one can show: α(W)=DP(D=1X,A)1DP(D=0X,A),

    αs(W)=DP(D=1X)1DP(D=0X),

    The RR “looks like” the weights in Inverse Propensity Score weighting (IPW). For example,

    αs(W)=DP(D=1X)1DP(D=0X)=De(X)1D1e(X), where e(X)=P(D=1X)

    Recall the property of IPW estimator, we have:

    θs=ipwE[YDe(X)Y(1D)1e(X)]=E(Yαs)
  5. Note that, θ=E(gα). Why? E(gα)=E{E(YD,X,A)DE[DX,A]E(DE[DX,A])2}=E{[θD+f(X,A)]DE[DX,A]E(DE[DX,A])2}=θ

    For the third equation, we use the fact that: E[DX,A]  DE[DX,A] and f(X,A)  DE[DX,A].

Short Summary:

  • g is the outcome regression function

  • α is the Riesz Representer

  • We have θ=E(Yα)=E(gα) and θs=E(Yαs)=E(gsαs).

Key Results

image-20240913170504351

Key observation:

  1. OVB is bounded by E(outcome regression errorRR error)

  2. Square bias is bounded by MSE(regression)MSE(RR)

Idea of Proof:

  1. Write θ=E(gα) and θs=E(gsαs)

  2. Use the fact that αs  ggs and gs  ααs

Now, we know the omitted variable bias,

θsθ=E(gsg)(αsα),

is the covariance between the regression error,

gsg=E(Y|Ws)E(Y|W),

and RR error,

αsα,

that is, the “weights in IPW” calculated using the short list of regressors, minus the “weights in IPW” using the long list of regressors.

As we can see, this is not easy to interpret, so the following Corollary provides a more intuitive understanding using the familiar R2.

image-20240913170858030

Note that,

image-20250118105142237

The bias is thus bounded by the additional variation that omitted variables cause in the regression function and Riezs representers.

We can make this easier to interpret with further algebra,

image-20250118105610894

Short Summary:

  1. S is the scale of the bias, which can be identified from the observed data.

  2. The confounding strength CY and CD have to be restricted by the analyst.

  3. CY2 measures the proportion of residual variation of the outcome explained by latent confounders; in short, the strength of unmeasured confounding in outcome equation

  4. CD2 measures the proportion of residual variation of the treatment explained by latent confounders; in short, the strength of unmeasured confounding in treatment equation

image-20240913193118914
Source: Victor’s tutorial at the Chamberlain Seminar

Empirical Challenge

How do we determine the plausible strength of these unobserved confounders?

How to set the strength of these unobserved confounders? CY2=? and CD2=?

This is a critical question, as setting these values arbitrarily could lead to overly conservative or overly optimistic interpretations of our results. We need a principled approach to guide our choices.

Benchmarking

One solution is benchmarking. This approach leverages the observed data to inform our judgments about unobserved confounders. The basic idea is the following: we can use the impact of observed covariates as a reference point for the potential impact of unobserved ones. For instance, if we’ve measured income, the “most” important factor, and found it explains 15% of the variation in our outcome, we might reason that an unobserved confounder is unlikely to have an even larger effect.

In practice, benchmarking can take several forms. We might purposely omit a known important covariate, refit our model, and observe the change. This gives us a concrete example of how omitting an important variable affects our estimates. Alternatively, we could express the strength of unobserved confounders relative to observed ones. For example, we might consider scenarios where an unobserved confounder is as strong as income, or perhaps 25% as strong.

Recent research (Chernozhukov et al.,2024) has demonstrated the utility of this approach. In a study on 401(k) eligibility, they used the observed impact of income, IRA participation, and two-earner status as benchmarks for potential unobserved firm characteristics. Similarly, in a study on gasoline demand, the known impact of income brackets informed the choice of sensitivity parameters for potential remnant income effects.

Robustness Value

Another useful concept is the Robustness Value (RV) which represents the minimum strength of confounding required to change a study’s conclusions. This provides a clear threshold for evaluating the robustness of results.

image-20250119083624670

The idea of robustness values is to quickly communicate how robust the short estimate is to systematic errors due to residual confounding. For example, RVθ=0,α=0.05 measures the minimal strength on both confounding factors such that the estimated confidence bound for the ATE would include zero, at the 5% significance level.

401(k) example in the paper

image-20250119090152561

“Table 1 illustrates our proposal for a minimal sensitivity reporting of causal effect estimates. Beyond the usual estimates under the assumption of conditional ignorability, it reports the robustness values of the short estimate.
Starting with the PLM, the RVθ=0,α=0.05=5.4% means that unobserved confounders that explain less than 5.4% of the residual variation, both of the treatment, and of the outcome, are not sufficiently strong to bring the lower limit of the confidence bound to zero, at the 5% significance level.”

Interpretation of RV:

  • A higher RV suggests that the causal estimate is more robust to omitted variable bias because stronger confounding would be needed to alter the conclusions.

  • A lower RV suggests that a relatively small amount of confounding could change the conclusions.

In summary, the robustness value (RV) provides a concise way to communicate how sensitive the results of a causal analysis are to potential omitted variable bias. It is the minimum strength of confounding that would be required to change the conclusions. A larger RV indicates a more robust analysis.

Software to Implement

References

Chernozhukov, Victor, Carlos Cinelli, Whitney Newey, Amit Sharma, and Vasilis Syrgkanis. “Long Story Short: Omitted Variable Bias in Causal Machine Learning.” arXiv, May 26, 2024. https://doi.org/10.48550/arXiv.2112.13398.

10. Sensitivity analysis in DoubleML package

Angrist, Joshua D., and Jörn-Steffen Pischke. Mostly Harmless Econometrics: An Empiricist’s Companion. Princeton University Press, 2009.

sensemakr: Sensitivity Analysis Tools for OLS in R and Stata

Carlos Cinelli and Chad Hazlett. ‘Making sense of sensitivity: Extending omitted variable bias’. In: Journal of the Royal Statistical Society: Series B (Statistical Methodology) 82.1 (2020), pp. 39–67 (cited on pages 23, 26).

sensitivity analysis lecture


  1. These questions are in the slides of this tutorial ↩︎

Chen Xing
Chen Xing
Founder & Data Scientist

Enjoy Life & Enjoy Work!

Related