STATS 413

Outcome regression

In this post, we consider the task of estimating treatment effects. This is the basic problem in causal inference, and it arises in a many areas of science and engineering. As a running example, we consider the task of estimating the efficacy of a vaccine booster. We begin by mathematically defining treatment effects using the potential outcomes framework.

To keep things simple, we focus on estimating the effect of a binary treatment (e.g. booster vs no booster). We define two potential outcomes Yi(1) and Yi(0) for each subject in the study. In the running example, Yi(1) is the viral load in the i-subject if the subject got the booster, and Yi(0) is the viral load if the subject did not get the booster. The effect of the treatment on the i-th subject is

ΔiYi(1)Yi(0).

The fundamental challenge in causal inference is only one treatment can be assigned to a subject, so only one of Yi(1) and Yi(0) can be observed. Thus Δi is never observed. Nevertheless, it is possible (as we shall see) to estimate the average treatment effect (ATE)

τE[Δi]=E[Yi(1)]E[Yi(0)] by performing randomized experiments.

In a randomized experiment, we randomly assign treatments to the subjects and record the outcomes. Let Wi{0,1} and Yi be the treatment assignment and observed outcome of the i-th subject. In the running example, Wi indicates whether the i-th subject got the booster and Yi is the (observed) viral load in the i-th subject. Mathematically, in a randomized experiment, we have

Yi=Yi(Wi)(SUTVA),(Yi(1),Yi(0))Wi(random treatment assignment).

The first condition (SUTVA) relates the observed outcomes to the potential outcomes: the observed outcome of the i-subject Yi is Yi(1) (resp Yi(0)) if Wi=1 (resp Wi=0). The second condition says treatments are assigned in a way that does not depend on the potential outcomes. It implies the distribution of potential outcomes in the treated and untreated groups are identical:

(Yi(1),Yi(0)){Wi=1}=d(Yi(1),Yi(0)){Wi=0}.

In practice, treatments are often assigned randomly (e.g. by flipping a coin) to satisfy this condition.

Difference-in-means

A simple estimate of the ATE in a randomized experiment is the difference between the (sample) mean outcomes in treated and untreated subjects:

τ^DM=1n1i=1nYi1{Wi=1}1n0i=1nYi1{Wi=0},

where nwi=1n1{Wi=w} is the number of subjects assigned treatment $w\in{0,1}$. This is called the difference-in-means estimator, and it is motivated by the observation that the (sample) mean outcome in a treatment group is an unbiased estimate of the expected potential outcome in a randomized experiment:

E[1nwi=1nYi1{Wi=w}]=E[YiWi=w]=E[Yi(w)Wi=w](SUTVA)=E[Yi(w)](random treatment assignment).

In light of this observation, it is not hard to see that the difference-in-means estimator is unbiased:

E[τ^DM]=E[1n1i=1nYi1{Wi=1}]E[1n0i=1nYi1{Wi=0}]=E[Yi(1)]E[Yi(0)]=τ.

We leave as an exercise to show that τ^DM is asymptotically normal:

n1+n0(τ^DMτ)dN(0,σ12π1+σ02π0),

where σw2var[Yi(w)] and πwP{Wi=w} for w{0,1}. This result allows us to form confidence intervals and test hypothesis regarding the ATE.

Posted on December 11, 2021 from Ann Arbor, MI