PSY 652: Research Methods in Psychology I

Causal Inference

Kimberly L. Henry: kim.henry@colostate.edu

What is Causal Inference?

  • Causal inference is about more than finding patterns in our data — it’s about understanding what causes what. It asks: “If we change one thing, what happens to another?”

  • To do this, we build scientific models that explain the relationships between variables, using logic and theory to guide us.

  • Why start with the scientific model? Because if our understanding of how things are connected is wrong, no statistical method can fix it.

Correlation Does Not Imply Causation

An Example: Countries with faster internet speeds (X) have stronger economies (Y). But, correlation is likely spurious because wealthier countries are more likely to invest in advanced infrastructure, including high-speed internet. Economic strength and internet speed are both outcomes of higher national wealth.

But Also, Causation Does Not Imply Correlation

An Example: When cruise control is set on a car, changes in road incline (X) causally affect the engine’s power output (Z) to maintain the set speed (Y). Despite the causal effect of road incline on engine power, the vehicle speed Y remains constant due to the cruise control. Data collected on vehicle speed (Y) and road incline (X) will show no correlation, despite X causing Y. The cruise control is an example of a feedback control mechanism.

Potential Outcomes Framework

Last week, you made a decision to either:

  • Read the Module 18 material (\(X = 1\))

  • Not read the Module 18 material (\(X = 0\))

Which of these actions will be better for your comprehension of today’s lecture?

Definition of Variables

  • We’ll label the treatment/exposure, reading the material, X.
  • We’ll label the outcome, comprehension of the lecture, Y.

The Potential Outcomes

We have two potential outcomes in this scenario:

  • One represents your comprehension if you read the material: \(Y^{X = 1}\)

  • The other represents your comprehension if you didn’t read the material: \(Y^{X = 0}\)

If you read the Module 18 material prior to class — then we will observe \(Y^{X = 1}\). The other potential outcome, \(Y^{X = 0}\), won’t be observed and will remain the counterfactual (i.e., since X = 0 didn’t occur – this is the outcome that can’t be observed).

If you didn’t read the Module 18 material prior to class — then we will observe \(Y^{X = 0}\). The other potential outcome, \(Y^{X = 1}\), won’t be observed and will remain the counterfactual (i.e., since X = 1 didn’t occur – this is the outcome that can’t be observed).

The Individual-Level Causal Effect

  • The individual‐level causal effect is \(Y^{X = 1} - Y^{X = 0}\).

  • That is — the individual-level causal effect is the difference in your comprehension of the lecture if you read the Module 18 material versus if you didn’t read the material.

  • This difference is unknowable!

  • This is the fundamental problem of causal inference (Holland, 1986).

Random Assignment to Condition

Let’s imagine that I randomly assign half of you to read the Module 18 material and the other half to not read the material.

  • If you’re assigned to read the material, and you read it — then we’ll observe \(Y^{X = 1}\).

  • If you’re not assigned to read the material, and you don’t read it — then we’ll observe \(Y^{X = 0}\).

Therefore, we can compute the Average Causal Effect as:

\[ E\left(Y_i^{X=1}\right) - E\left(Y_i^{X=0}\right) \]

… where \(E\) represents expectation (i.e., the average or mean value of a random variable if the process were repeated infinitely many times).

The incredible power or randomization

Randomization is a tremendously powerful procedure for examining causal effects. However, it’s not always perfect in practice:

  • Non-compliance: Sometimes participants don’t follow instructions, which can affect the treatment assignment.

  • Drop-out: Participants may fail to provide outcome data, leading to missing information.

  • Unmanipulable causes: In some cases, the cause of interest cannot be directly manipulated (e.g., the causal effect of emotional responses).

  • Ethical and practical constraints: Randomization might be unethical (e.g., studying the causal effect of drug use) or unfeasible (e.g., studying the effects of socioeconomic status).

What to do Absent a Perfectly Executed Experiment?

In these scenarios, it’s helpful to think about the entire causal network underlying the data. Graphical notation can make it easier to understand and analyze these complex relationships.

Directed Acyclic Graphs (DAGS)

DAGS are a powerful and intuitive tool for representing causal relationships.

  • Heuristic Models: DAGs are simplified representations of a causal model, focusing on the structure of relationships between variables rather than the exact mechanisms or mathematical details.

  • Directed: Each arrow shows the direction of influence between two variables. For example, an arrow from X → Y means X is assumed to cause Y (but not the other way around).

  • Acyclic: DAGs do not allow loops or cycles—no variable can directly or indirectly cause itself. This ensures clarity in identifying causal pathways.

  • Anonymous Influence: DAGs don’t specify how variables influence each other (e.g., linear or nonlinear effects); they only indicate that one variable influences another.

An Example DAG

Start by drawing the exposure (e.g., reading the material) and the outcome (e.g., comprehension of the lecture).

To the Basic DAG …

  • Draw all common causes of X and Y – including measured and unmeasured variables.

  • As you add new variables, include any common causes of any pair of variables in your DAG. This ensures that the DAG accounts for potential bias introduced by these shared causes.

  • If there are variables through which X affects Y (i.e., mediators), draw arrows from X to the mediator(s), and from the mediator(s) to Y.

  • Include all selection variables – these are variables that represent selection processes that occur as part of the study – e.g., drop out, death, non-compliance.

Drawing a DAG requires expert knowledge and must be complete! NOT including an arrow implies an ASSUMPTION that there is NO CAUSAL EFFECT.

Consider an Observational Version of the Study

If instead of randomly assigning students to read/not read — I simply assessed whether students chose to read the material — will the difference in comprehension between those who chose to read vs. not read still represent an average causal effect?

A DAG

Paths of the Graph

  • Path: A path is a sequence of arrows between two variables in a Directed Acyclic Graph (DAG). The path must not pass through any variable more than once. The direction of the arrows is irrelevant when defining a path.

  • Causal Paths: A causal path is a path in which all arrows point in the same direction, starting from the cause (X) and ending at the effect (Y). This path represents a direct or indirect causal relationship from X to Y.

  • Non-Causal Paths: A non-causal path includes at least one arrow that points toward X (or away from Y when tracing the path). These paths may introduce confounding, colliding, or other non-causal associations between X and Y.

Identify the Paths Connecting X to Y

  1. READ -> COMPREHEND (causal path)

  2. READ <- MOTIVATION -> COMPREHEND (non-causal path/back-door)

Block the Back Door Path(s)

  • A back-door path connects the exposure (X) to the outcome (Y) through a common cause — also called a confounder.

  • Conditioning on the confounder (e.g., adjusting for MOTIVATION) blocks the back door path.

  • This breaks the spurious connection and isolates the causal effect.

MOTIVATION (in this example) makes a Fork

READ <- MOTIVATION -> COMPREHEND

Simulate a Fork Part 1

Simulate a Fork Part 2

Simulate the Fork Part 3

Consider an Experiment

READ -> COMPREHEND (causal path)

Simulate an Experiment Part 1

Simulate an Experiment Part 2

Simulate an Experiment Part 3

Consider an Experiment with Non-Compliance

Identify the Paths Connecting X to Y

  1. READ -> COMPREHEND (causal path)

  2. READ -> ACTUAL_READ -> COMPREHEND (causal path)

  3. READ -> ACTUAL_READ <- MOTIVATION -> COMPREHEND (non-causal path)

ACTUAL_READ makes a pipe

READ -> ACTUAL_READ -> COMPREHEND

ACTUAL_READ + MOTIVATION makes a collider

READ -> ACTUAL_READ <- MOTIVATION -> COMPREHEND

Proper Handling of Pipes

  • If You Want the Total Effect:

    • Do not adjust for the mediator (i.e., ACTUAL_READ)

    • Simply estimate the overall effect of READ -> COMPREHEND

    • Otherwise, you will “explain away” part of the effect of interest.

Simulate a Pipe Part 1

Simulate a Pipe Part 2

Simulate a Pipe Part 3

Proper Handling of Colliders

  • By default, collider paths are blocked, meaning they do not create a spurious association between the exposure and the outcome.

  • But, beware, if you condition on a collider — you will open a back door path and create a spurious association.

Simulate a Collider Part 1

Simulate a Collider Part 2

Simulate a Collider Part 3

What’s Happening with this Collider?

  • When we condition on Z, we are essentially selecting a subset of the data where Z takes specific values (e.g., Z=1, Z=0). Because Z is influenced by both X and Y, conditioning on Z artificially links X and Y through Z, inducing a spurious dependency, even if X and Y were originally independent.

An Intuitive Example of a Collider

Lawn Simulation

What Model Should be Fit for the Experiment?

  • By conditioning on ACTUAL_READ (e.g., analyzing only those who complied or controlling for compliance in a model), you open a non-causal path between READ and MOTIVATION that would otherwise remain blocked.

  • This creates a spurious association between READ and COMPREHEND, undermining the causal interpretation.

  • Revel in the beauty of your experiment, and don’t include any post-treatment variables.

Take Home Message for Experiments

Let’s Try another Example

What are all paths that link Smoking and Cancer in this DAG?

The Paths

SMK -> CAN
SMK <- SES -> CAN
SMK <- SES -> DTH <- CAN
SMK <- SES <- CI -> DTH <- CAN

Let’s Set it Up in DAGitty