Course Lesson Plan

Image by manfredsteger via Pixabay.

This section of the final deliverable presents a concise thirteen-week lesson plan for a fourth-year undergraduate course in causal inference. The lesson plan is limited in scope due to the following reasons:

A note on weeks 10, 11, 12, and 13!

The final four weeks of the course will not be used to introduce students to additional modelling techniques in causal inference (as outlined in stages 4 and 5 of the roadmap) or to discuss general guidelines for interpreting, reporting, and storytelling in causal studies (as detailed in stages 6, 7, and 8 of the roadmap). Instead, students are expected to focus on the following major deliverables:

  • Weeks 10 and 11 will be dedicated to a team-based appraisal presentation of an already published causal study, utilizing the roadmap covered throughout the course. There will not be a written report; only an oral presentation is required.
  • Weeks 12 and 13 will focus on a team-based final project where students will conduct their final causal study using a provided dataset and following the causal roadmap. Instead of a final oral presentation, a written report will be submitted to replace the in-class final exam.

Therefore, for the above week, there will not be mapped scholarly materials.

1 Lecture Breakdown

Below is a weekly breakdown of the learning goals per lecture. Additionally, we provide a brief summary of each learning week and outline the expectations at the end of the respective two lectures.

1.1 Know your Research Question! (stage 1 in the roadmap)

During week 1, the instructor is expected to introduce students to a foundational mindset in causal inference, which will be linked to the course’s inferential roadmap. The primary goal for this week is to help students develop the skill to distinguish between statistical inquiries involving association or causation in scientific research. Additionally, students will be introduced to the causal roadmap, which consists of eight stages: (1) research question, (2) causal model representing knowledge, (3) counterfactuals and causal parameter definition, (4) statistical model definition, (5) model fitting, (6) result interpretation, (7) data analysis report. This roadmap will serve as the main pillar of the course’s learning strategy.

Image by Manfred Stege via Pixabay.

Furthermore, students will explore real-world examples to understand that causal inference is not merely an isolated statistical technique for modelling data (such as solely fitting ordinary-least squares or binary logistic regression to a dataset). Instead, it represents a comprehensive thought process about counterfactual outcomes (i.e., potential values of \(Y\)) and potential treatment interventions (i.e., potential values of \(X\)). By the end of this week, students should be able to comprehend the philosophical and practical differences between asking:

Is \(X\) associated with \(Y\)?

and

Does \(X\) cause \(Y\)?

Important

Recognizing the statistical difference between the above two statements is crucial across all scientific disciplines in the practice of causal inference. You can check Cummiskey et al. (2020) and Lübke et al. (2020) for further elaboration on this matter.

1.1.1 Lecture 1

By the end of this lecture, students should be able to:

  1. Describe what causal inference is along with its usefulness in scientific research (remember level).
  2. Explain how causation and association differ via practical examples (understand level).
  3. Classify different types of research questions in terms causation and association (understand level).
  4. Identify the components of a causal questions, e.g., treatment, outcome, and population.

The above goals can be mapped to the following materials:

  1. Chapter 2 (Research Questions) from Huntington-Klein (2021) and Chapter 1 (A definition of causal effect) from Hernán (2024) and Roadmap Overview & Roadmap Step 0 - Research Question (Lecture 1a - Why Bother with Causal Inference?) from Petersen and Balzer (2014).
  2. Chapter 1 (Correlation, Association, and the Yule-Simpson Paradox) from Ding (2024) and Roadmap Overview & Roadmap Step 0 - Research Question (Lecture 1b - Causal vs. Statistical inference) from Petersen and Balzer (2014).
  3. Chapter 1 (Introduction - Do Not Confuse Correlation with Causality) from Cunningham (2021) and Roadmap Overview & Roadmap Step 0 - Research Question (Lecture 1c: The Causal Roadmap Steps 0-3) from Petersen and Balzer (2014).
  4. Chapters 2 (Randomized Experiments) and 3 (Observational Studies) from Hernán (2024) and Chapter 2 (Potential Outcomes) from Ding (2024).

1.1.2 Lecture 2

By the end of this lecture, students should be able to:

  1. Explain the steps of the causal inference roadmap (understand level).
  2. Examine real-life study descriptions to identify whether they pose a causal or association question (analyze level).
  3. Formulate well-structured causal questions based on a given observational or experimental context. (create level).
  4. Assess the validity and clarity of causal questions in peer-reviewed studies (evaluate level).

The above goals can be mapped to the following materials:

  1. Roadmap Overview & Roadmap Step 0 - Research Question (Lecture 1c - The Causal Roadmap) from Petersen and Balzer (2014). Recall that the roadmap for this course is inspired by this reference. Hence, this material can be used as a baseline to introduce our roadmap.
  2. For this learning goal, the instructor could provide a summary of the following four published papers (two of which pose causal questions, whereas the other two pose association questions):
  • Causal questions:
    • Does Compulsory School Attendance Affect Schooling and Earnings?” from Angrist and Krueger (1991).
    • Using Geographic Variation in College Proximity to Estimate the Return to Schooling” from Card (1995).
  • Association questions:
    • Association of Animal and Plant Protein Intake With All-Cause and Cause-Specific Mortality” from Song et al. (2016).
    • Association between Sleep Duration and Mortality among Adults with Type 2 Diabetes: A Prospective Cohort Study” from Wang et al. (2020).
  1. Roadmap Overview & Roadmap Step 0 - Research Question (Lecture 1c - The Causal Roadmap) from Petersen and Balzer (2014).
  2. Morgan and Winship (2007) provide a fair list of referenced examples in Chapter 1 (Introduction), specifically in sections 1.3.1 (Broad Examples from Sociology, Economics, and Political Science) and 1.3.2 (Narrow and Specific Examples).

1.2 Using Graphical Models to Depict Causality (stage 2 in the roadmap)

During Week 2, students will have their first exposure to directed acyclic graphs (DAGs). According to the course roadmap, DAGs are graphical models that represent knowledge about causal inquiries, which we are expected to develop in collaboration with subject-matter experts. We will view DAGs as essential tools for visualizing and reasoning about causal relationships.

Image by Manfred Stege via Pixabay.

The week will begin with a formal introduction to DAG syntax and terminology. After that, students will engage in hands-on practice constructing DAGs based on real-world scenarios from various scientific fields. Additionally, we will differentiate between experimental, quasi-experimental, and observational studies to demonstrate how DAGs can help identify assumptions in different study designs. By the end of the week, students should understand why DAGs offer valuable insights into causal logic, which will be beneficial for the subsequent stages of the course roadmap.

1.2.1 Lecture 3

By the end of this lecture, students should be able to:

  1. Explain key components of DAGs, including nodes, directed edges, and acyclic structure (understand level).
  2. Illustrate how DAGs represent causal assumptions and the concept of d-separation (understand level).
  3. Compare DAGs derived from experimental, quasi-experimental, and observational designs to assess differences in identifiability (analyze level).
  4. Construct simple DAGs to represent real-life study scenarios (analyze level).

The above goals can be mapped to the following materials:

  1. Chapter 6 (Graphical Representation of Causal Effects) from Hernán (2024) and Chapter 6 (Causal Diagrams) from Huntington-Klein (2021).
  2. Chapter 6 (Graphical Representation of Causal Effects) from Hernán (2024).
  3. The following academic papers can be used to develop material to address this goal:
  • How to develop causal directed acyclic graphs for observational health research: a scoping review” by Poppe et al. (2025).
  • Reducing bias in experimental ecology through directed acyclic graphs” by Arif and Massey (2023)
  • Utilizing Causal Diagrams across Quasi-Experimental Approaches” by Arif and MacNeil (2022).
  1. Chapter 7 (Drawing Causal Diagrams) from Huntington-Klein (2021).

1.2.2 Lecture 4

By the end of this lecture, students should be able to:

  1. Create DAGs to represent hypothetical causal mechanisms based on subject-matter expertise (create level).
  2. Identify common structures in DAGs such as confounding, mediating, backdoor criterion, and colliding paths (analyze level),
  3. Evaluate the adequacy of DAGs constructed by peers or drawn from published studies, focusing on clarity and underlying assumptions (evaluate level).
  4. Differentiate between experimental, quasi-experimental, and observational study designs using DAGs to illustrate assumptions and limitations (understand level).

The above goals can be mapped to the following materials:

  1. To address this goal, the students could have hands-on practice with the cases depicted by Morgan and Winship (2007) provide a fair list of referenced examples in Chapter 1 (Introduction), specifically in sections 1.3.1 (Broad Examples from Sociology, Economics, and Political Science).
  2. Chapter 8 (Causal Paths and Closing Back Doors) from Huntington-Klein (2021), Chapter 7 (Confounding) from Hernán (2024), and Chapter 1 (Causal Graphs, Identification, and Models of Causal Exposure) from Morgan and Winship (2007).
  3. To address this goal, students are expected to assess the DAGs delivered in goal (1) while applying the concepts addressed in goal (2).
  4. The developed material from lecture 3, to address the respective goal (3), can be used to fulfil this goal.

1.3 Introduction to Counterfactuals (stage 3 in the roadmap)

Week 3 focuses on both the theoretical and practical aspects of counterfactuals. As outlined in the course roadmap, counterfactuals play a crucial role in the entire causal inference process. They are an essential component of the structural causal model (SCM), which is a mathematical framework for illustrating the causal relationships between variables within a given system. This SCM framework will also provide students with computational tools to connect DAGs, discussed in week 2, with counterfactual reasoning.

The workflow for this week begins with defining the DAG in the context of a causal inference inquiry within a specific system. Students will then derive a SCM composed of equations that represent the data-generating process for each variable. These equations explicitly illustrate the mechanisms that produce counterfactual outcomes, such as:

What would happen to an outcome \(Y\) if treatment \(X\) were set to a different value?

This type of inquiry is essential for formulating and interpreting what we refer to as causal parameters, such as the average treatment effect (ATE).

Image by Manfred Stege via Pixabay.

By the end of the week, students are expected to learn how to use code-based simulations in R (R Core Team 2024) or Python (Van Rossum and Drake 2009) to apply the logic associated with DAGs, SCMs, and counterfactuals through simple scenarios. They will simulate interventions, explore counterfactuals, and compute causal parameters to address research questions effectively. This process specifically involves examining how potential outcomes arise from different interventions, which allows us to connect our findings back to the DAG and ultimately enables us to produce valid causal claims. It is essential to emphasize to students that our simulated causal parameters must align with the research question(s) established in stage 1 of the roadmap.

1.3.1 Lecture 5

By the end of this lecture, students should be able to:

  1. Explain the structure and components of a SCM derived from a DAG (understand level).
  2. Explain the conceptual link between structural equations and counterfactual outcomes (understand level).
  3. Implement simple SCMs using code to simulate counterfactual scenarios (apply level).
  4. Classify the types of causal questions that we can answer from a given SCM and DAG (understand level).

The above goals can be mapped to the following materials:

  1. Chapters 3 (Observational studies) and 5 (Interaction) from Hernán (2024). These chapters offer a comprehensive theoretical content. For in-class material, Roadmap Step 1 - Causal Model (Lecture 2a - Intro to Structural Causal Models), from Petersen and Balzer (2014), is a fair starting point.
  2. Chapter 2 (The Counterfactual Model) from Morgan and Winship (2007); this chapter offers a comprehensive theoretical content. For in-class material, Roadmap Step 1 - Causal Model (Lecture 2b - How SCMs encode causal assumptionss), from Petersen and Balzer (2014), is a fair starting point.
  3. Chapter 4 (Potential Outcomes Causal Model), from Cunningham (2021), provides simple coding examples via R and Python that the instructor could use in their in-class material.
  4. For in-class material, Roadmap Step 1 - Causal Model (Lecture 2c - SCMs & Directed Acyclic Graphs), from Petersen and Balzer (2014), is a fair starting point. Moreover, students can continue with the DAGs constructed to address goal (4) from lecture 4 to deliver the corresponding SCMs for the corresponding experimental, quasi-experimental, and observational designs.

1.3.2 Lecture 6

By the end of this lecture, students should be able to:

  1. Interpret counterfactual quantities such as \(Y^a\), the potential outcome under treatment level \(a\) (understand level).
  2. Implement causal parameters (e.g.; ATE or average treatment effect on the treated, ATT) from simulated SCMs using counterfactual logic (apply level).
  3. Identify causal parameters with specific scientific research questions (apply level).
  4. Design small simulation exercises to test assumptions underlying counterfactual inference (create level).

The above goals can be mapped to the following materials:

  1. Chapters 3 (Observational studies) and 5 (Interaction) from Hernán (2024). These chapters offer a comprehensive theoretical content.
  2. There are two potential sources of material we can use to fulfill this goal:
  1. Students are expected to fulfill this learning goal by using the SCMs delivered in lecture 5 for its corresponding goal (4).
  2. Students are expected to continue working with the SCMs from the above goal (3). Now, it is time to code the SCMs delivered from the corresponding experimental, quasi-experimental, and observational designs.

1.4 Defining the Statistical Model: Observed Data and Identifiability Conditions (stage 4 in the roadmap)

During week 4, the instructor is expected to connect the causal model represented by DAGs and SCMs with the corresponding statistical model, with a focus on the identifiability of causal parameters. We start by recalling DAGs as formal representations of causal assumptions. From these DAGs, we construct SCMs, specifying structural equations and endogenous variables that determine counterfactual outcomes. Once the SCM is defined, we analyze how these counterfactuals relate to observed data, emphasizing that only a subset of potential outcomes can be observed.

A key focus this week is on the spectrum of assumptions in statistical modelling. Parametric models specify a complete probability distribution (for example, ordinary least-squares regression with normally distributed errors), while semi-parametric models relax some of these specifications (like Cox models). Non-parametric models impose minimal assumptions on the functional form, relying solely on the assumptions encoded in the DAG.

Image by Manfred Stege via Pixabay.

In the second half of the week, we define identifiability as the condition under which a causal parameter (such as the ATE) can be expressed as a function of the observed data distribution. Students will learn to utilize d-separation and the back-door criterion as tools to determine when identifiability holds, given the assumptions of no unmeasured confounding and positivity.

We will evaluate how each type of statistical model (parametric, semi-parametric, and non-parametric) impacts identifiability. The week will conclude with a comparative discussion across disciplines—such as epidemiology, economics, and statistics—of the various terminologies and conceptual frameworks used to express identifiability conditions. For example, students will explore the relationship between the stable Unit treatment value assumption (SUTVA), ignorability in econometrics, and conditional exchangeability in epidemiology. This reflection will help students understand how each domain approaches the identifiability problem through its unique perspective.

1.4.1 Lecture 7

By the end of this lecture, students should be able to:

  1. Define how DAGs and SCMs formalize causal assumptions and determine potential outcomes (remember level).
  2. Construct a SCM from a DAG using structural equations and endogenous variables (apply level).
  3. Examine the relationship between SCMs and observed data, identifying which counterfactuals are observed versus unobserved (analyze level).
  4. Distinguish between parametric, semi-parametric, and non-parametric statistical models, and their implications for inference (analyze level).

The above goals can be mapped to the following materials:

  1. Roadmap Step 2 - Counterfactuals & Causal Effects (Lecture 3a - Defining Counterfactuals) from Petersen and Balzer (2014).
  2. Roadmap Step 2 - Counterfactuals & Causal Effects (Lecture 3b - Deriving Counterfactuals from the Causal Model) and Lecture 3c (Using Counterfactuals to Define Causal Effects) from Petersen and Balzer (2014).
  3. Roadmap Step 3 - Observed Data (Lecture 5a - Specify the observed data & their link to the causal model and Lecture 5b - Causal models & (in)dependence in the observed data) from Petersen and Balzer (2014).
  4. Roadmap Step 3 - Observed Data (Lecture 5c - Defining the Statistical Model) from Petersen and Balzer (2014).

1.4.2 Lecture 8

By the end of this lecture, students should be able to:

  1. Determine whether a causal parameter is identifiable using d-separation and the back-door criterion (evaluate level).
  2. Assess whether key identifiability conditions—no unmeasured confounding and positivity—hold in study designs (evaluate level).
  3. Compare how different statistical modelling assumptions (parametric, semi, non-parametric) affect identifiability (understand level).
  4. Build a visual map that translates vocabulary across disciplines: SUTVA, ignorability, independence, and exchangeability (create level).
A smooth introduction to identifiability

Chapter 5 - Identification, from Huntington-Klein (2021), offers a smooth introduction to identifiability. It could be used as a pre-reading.

The above goals can be mapped to the following materials:

  1. Chapter 3 (Observational Studies) from Hernán (2024). Also, Roadmap Step 4 - Identifiability & Step 5 - Estimation Problem (Lecture 6a - Overview & Intuition for Identifiability and Lecture 6c - The Backdoor Criterion) from Petersen and Balzer (2014).
  2. Roadmap Step 4 - Identifiability & Step 5 - Estimation Problem (Lecture 6d - The Positivity Assumption) from Petersen and Balzer (2014).
  3. Chapter 11 (Why Model?) from Hernán (2024)
  4. Section 4.1.5, from Cunningham (2021), provides a detailed elaboration on SUTVA. On the other hand, Kamath, Graham, and Naylor (2023) provide a chapter elaborating on the rest of the terms.

1.5 Inverse Probability Weights (stages 4 and 5 in the roadmap)

Week 5 introduces inverse probability weighting (IPW) as the first method in the course for estimating causal effects from observational data. Building on the identifiability assumptions discussed in week 4, we will now focus on the statistical estimation of causal parameters using weights derived from the treatment assignment mechanism. Students will be guided through both the conceptual rationale and the mathematical foundations of IPW. Specifically, we will explore how IPW constructs a pseudo-population in which treatment is independent of confounders, allowing us to recover the causal estimand from observed data. We will emphasize the critical conditions under which this estimator is valid, particularly the no unmeasured confounding and positivity assumptions.

Image by Manfred Stege via Pixabay.

The week is divided into two sessions. The first session focuses on developing theoretical insights into the IPW framework, while the second session emphasizes practical implementation using R. Students will learn how to estimate weights, assess their quality, and use diagnostic tools to detect violations of key assumptions, such as positivity. We will work with both real and simulated datasets to visualize weight distributions, evaluate performance, and interpret the causal estimates obtained through IPW. This week also provides an opportunity to connect the theoretical framework to concrete software tools, such as the ipw (van der Wal and Geskus 2011) and WeightIt (Greifer 2025) R packages.

1.5.1 Lecture 9

By the end of this lecture, students should be able to:

  1. Define IPWs and describe their purpose in causal effect estimation (remember level).
  2. Explain how IPW adjusts for measured confounding by reweighting observations to achieve exchangeability (understand level).
  3. Examine the assumptions required for valid IPW estimation, including positivity and no unmeasured confounding, and discuss what happens when these assumptions are violated (analyze level).

The above goals can be mapped to the following materials:

  1. Chapter 2 (Randomized Experiments), from Hernán (2024), where Section 2.4 specifically introduces IPW in a non-parametric way. Then, Chapter 12 (IP Weighting and Marginal Estructural Models) elaborates on IPW in a parametric way. Also, Roadmap Step 6B - Estimation with IPW (Lecture 9b - Intro to the IPW Estimator), from Petersen and Balzer (2014), offers a fair starting point for in-class material.
  2. Chapter 12 (IP Weighting and Marginal Estructural Models) provides a complete technical elaboration. Moreover, Roadmap Step 6B - Estimation with IPW (Lecture 9b - Intuition and Implementation of IPW), from Petersen and Balzer (2014), offers a fair starting point for in-class material.
  3. Chapter 12 (IP Weighting and Marginal Estructural Models) provides a complete technical elaboration. Furthermore, Roadmap Step 6B - Estimation with IPW (Lecture 9c - Impact of Positivity Violations on Estimator Performance), from Petersen and Balzer (2014), offers a fair starting point for in-class material.

1.5.2 Lecture 10

By the end of this lecture, students should be able to:

  1. Utilize IPW to estimate average treatment effects using regression models in R (apply level).
  2. Create a simulated dataset with known treatment effects and counterfactuals to evaluate the performance of IPW under controlled conditions (create level).
  3. Assess the quality of estimated IP weights by visualizing their distribution, identifying extreme weights, and interpreting ATE sensitivity to such weights (evaluate level).

The above goals can be mapped to the following materials:

  1. van der Wal and Geskus (2011) provide an overview of the ipw package with general theoretical background on IPW along with examples via real data. Heiss (2020) provides an example via real data to illustrate both the use of ipw and WeightIt
  2. van der Wal and Geskus (2011) provide an simulated example via the package ipw.
  3. Roadmap Step 6B - Estimation with IPW (Lecture 9c - Impact of Positivity Violations on Estimator Performance), from Petersen and Balzer (2014), offers a fair starting point for in-class material. Moreover, they provide a full lab exercise on IPW.

1.6 G-estimation for Causal Parameters (stages 4 and 5 in the roadmap)

Week 6 introduces students to G-estimation methods for identifying and estimating causal parameters. This builds directly on the counterfactual framework and the potential outcomes approach developed in previous weeks. G-estimation enables us to estimate causal effects in the presence of time-varying confounders. Students will learn the intuition behind G-estimation and its conceptual connection to structural nested models (SNMs).

Image by Manfred Stege via Pixabay.

This week’s lectures will integrate statistical theory, identification strategies, and applied estimation using R. First, students will learn about the formulation of G-estimation and how it differs from IPW. Next, students will simulate longitudinal data, implement G-estimation in R, and assess the performance of these models in accurately estimating causal parameters. The week will conclude with a discussion on model diagnostics for G-estimation.

1.6.1 Lecture 11

By the end of this lecture, students should be able to:

  1. Define the assumptions required for G-estimation (remember level).
  2. Explain the differences between G-estimation and IPW (understand level).
  3. Derive the estimating equations used in G-estimation (apply level).
  4. Assess when G-estimation is preferable to IPW in terms of finite sample efficiency and bias (evaluate level).
A smooth introduction to G-estimation

Sterne and Tilling (2002) provide a smooth introduction to G-estimation in sections 1 and 2, while the remaining content focuses on its application in Stata. These two initial sections can be used as pre-reading material.

The above goals can be mapped to the following materials:

  1. Chapter 14 (G-estimation of structural nested models) from Hernán (2024).
  2. Daniel et al. (2013) provide a fair biostatistical tutorial that describes both methods, highlighting their relationships and differences. Additionally, they guide the reader on how to choose between them.
  3. Chapter 14 (G-estimation of structural nested models) from Hernán (2024).
  4. As in the case of learning goal (2), Daniel et al. (2013) is a fair starting point.

1.6.2 Lecture 12

By the end of this lecture, students should be able to:

  1. Implement G-estimation using different R packages (apply level).
  2. Simulate longitudinal data under a known data-generating process with time-varying confounding (apply level).
  3. Execute model diagnostics on G-estimation modelling (apply level).

The above goals can be mapped to the following materials:

  1. Loh and Ren (2023) provide a tutorial on G-estimation using simulated data via the lavaan package. In contrast, M. Wallace et al. (2025) offered a tutorial using the DTRreg package, employing both totally simulated data and simulated data inspired by real data.
  2. For this learning goal, we can use the work by Loh and Ren (2023) and M. Wallace et al. (2025) Moreover, M. P. Wallace, Moodie, and Stephens (2017) provide an additional simulated example via DTRreg.
  3. Rich et al. (2010) provide a fair overview on residual diagnostic plots for G-estimation via simulated and real data.

1.7 Week 7: Causal mediation analysis (STEPS 4 and 5 in the roadmap)

  • Introduction to causal mediation analysis.
  • Parametric and non-parametric.

1.8 Week 8: Interpretation and report (STEPS 6 and 7 in the roadmap)

1.9 Week 9: Storytelling (STEP 8 in the roadmap)

1.10 Week 10: Appraisal of causal analysis using the roadmap

  • Review an existing causal analysis (i.e., a published research paper) using the causal roadmap covered in class.
  • This is a team-based work.

1.11 Week 11: Presentation of appraisals

Team presentations.

2 Analysis of a Publicly Available Dataset via the Causal Roadmap

In the last two weeks of this course (i.e., weeks 12 and 13), students will take the lead in executing a team-based project that encompasses every stage of the causal roadmap. This means the instructor will primarily support the students throughout the process, rather than delivering traditional lectures. Note this analysis will serve as a final capstone project, summarizing all the content covered in the course from weeks 1 to 9. The project aims to incorporate every stage of the roadmap introduced during the term, which consists of the following eight stages: (1) research question, (2) causal model representing knowledge, (3) counterfactuals and causal parameter definition, (4) statistical model definition, (5) model fitting, (6) result interpretation, (7) data analysis report. The student teams will remain the same as those formed during weeks 10 and 11 for the previous appraisal exercise of a causal analysis.

Image by manfredsteger via Pixabay.

2.1 Learning Goals

The learning goals outlined below ensure that this final project demonstrates students’ mastery of the complete causal inference process. Additionally, it evaluates their ability to justify and communicate modelling assumptions, as well as their capacity to collaborate effectively in designing, executing, and defending a real-world causal analysis.

2.1.1 Remember Level

  • Recall the eight stages of the causal roadmap
  • Define the role of the above eight stages when conducting a causal analysis.

2.1.2 Understand Level

  • Summarize the background of the provided dataset and describe how it supports causal inference.
  • Interpret the assumptions (e.g., exchangeability, positivity, etc.) underlying the chosen estimation approach for the causal analysis.

2.1.3 Apply Level

  • Apply the causal roadmap to formulate a clearly defined causal research question based on the provided dataset.
  • Construct a comprehensive DAG that represents the assumed causal structure and justifies the variables to condition on.
  • Utilize estimation procedures via R (e.g., IPW, G-estimation, or causal mediation analysis), which should be consistent with the outlined causal model.

2.1.4 Analyze Level

  • Examine the links between underlying assumptions, model choices, and causal conclusions to assess coherence and validity.

2.1.5 Evaluate Level

  • Assess the quality of model diagnostics used to evaluate the robustness of causal estimates.
  • Criticize the strengths and limitations of the team’s analytic approach and provide a concise assessment of the causal claim.

2.1.6 Create Level

  • Develop a comprehensive and coherent written report that includes the causal question, methodology, assumptions, and findings in an accessible way to both technical and non-technical audiences.
  • Compile the roadmap stages into a reproducible causal workflow, including code, visualizations, and documentation, to support transparency and reproduciblity.
  • Design an executive summary or policy brief from the report tailored to a relevant stakeholder or decision-maker.

2.2 Weekly Breakdown

At the beginning of week 12 (specifically, the first in-class session), teams will be introduced to a publicly available dataset, accompanied by a detailed background presentation from the instructor. This presentation will include relevant contextual, scientific, and structural information necessary to support a wide range of causal inquiries. Teams will be tasked with formulating their own causal research questions—rather than being provided with predefined inquiries—and must justify their formulations based on the data and background information provided. This approach allows teams to leverage the first stage of the roadmap effectively.

Starting with the second session of week 12, teams will conduct their causal analyses using the dataset in conjunction with the eight-stage causal roadmap. They will be expected to iteratively apply concepts such as graphical modeling with DAGs, SCMs, counterfactual definition, identifiability assessment, and appropriate estimation methods, along with interpretation and storytelling. Note that we will use the same dataset for all teams, allowing instructors to evaluate different methodological approaches and assumptions reflected in each team’s application of the roadmap.

The last two sessions of the course (i.e., week 13) will be structured as open lab-style working sessions, where teams can receive feedback from the instructor. This feedback will provide an opportunity for teams to refine their analyses and coordinate roles and responsibilities. The final deliverable for this causal analysis will be a comprehensive team report that documents each step of the roadmap, the rationale behind the team’s decisions, diagnostic evaluations, limitations, and their final estimated causal effects.

A note on the lack of final exam in this course

It is important to note that the above final deliverable will replace a traditional final exam. Additionally, we want to emphasize that this capstone project is a team-based effort that highlights collaborative causal analysis, as is commonly practiced in both academic and applied research settings.

References

Anderson, Lorin W., and David R. Krathwohl, eds. 2001. A Taxonomy for Learning, Teaching, and Assessing. A Revision of Bloom’s Taxonomy of Educational Objectives. 2nd ed. New York: Allyn & Bacon.
Angrist, Joshua D., and Alan B. Krueger. 1991. “Does Compulsory School Attendance Affect Schooling and Earnings?” The Quarterly Journal of Economics 106 (4): 979–1014. https://doi.org/10.2307/2937954.
Arif, Suchinta, and M. Aaron MacNeil. 2022. “Utilizing Causal Diagrams Across Quasi-Experimental Approaches.” Ecosphere 13 (4): e4009. https://doi.org/https://doi.org/10.1002/ecs2.4009.
Arif, Suchinta, and Melanie Duc Bo Massey. 2023. “Reducing Bias in Experimental Ecology Through Directed Acyclic Graphs.” Ecology and Evolution 13 (3): e9947. https://doi.org/https://doi.org/10.1002/ece3.9947.
Bloom, B. S., M. B. Engelhart, E. J. Furst, W. H. Hill, and D. R. Krathwohl. 1956. Taxonomy of Educational Objectives. The Classification of Educational Goals. Handbook 1: Cognitive Domain. New York: Longmans Green.
Card, David. 1995. “Using Geographic Variation in College Proximity to Estimate the Return to Schooling.” Aspects of Labor Market Behaviour: Essays in Honour of John Vanderkamp, 201–23.
Cummiskey, Kevin, Bryan Adams, James Pleuss, Dusty Turner, Nicholas Clark, and Krista Watts and. 2020. “Causal Inference in Introductory Statistics Courses.” Journal of Statistics Education 28 (1): 2–8. https://doi.org/10.1080/10691898.2020.1713936.
Cunningham, Scott. 2021. Causal Inference: The Mixtape. Yale University Press. https://mixtape.scunning.com/.
Daniel, Rhian M., Simon N. Cousens, Bianca L. De Stavola, Michael G. Kenward, and Jonathan A. C. Sterne. 2013. “Methods for Dealing with Time-Dependent Confounding.” Statistics in Medicine 32 (9): 1584–1618. https://doi.org/10.1002/sim.5686.
Ding, P. 2024. A First Course in Causal Inference. Chapman; Hall/CRC. https://doi.org/10.1201/9781003484080.
Greifer, Noah. 2025. WeightIt: Weighting for Covariate Balance in Observational Studies. https://CRAN.R-project.org/package=WeightIt.
Heiss, Andrew. 2020. “Generating Inverse Probability Weights for Both Binary and Continuous Treatments.” December 1, 2020. https://doi.org/10.59350/1svkc-rkv91.
Hernán, Miguel A. 2024. Causal Inference: What If. Edited by James M. Robins. Boca Raton.
Huntington-Klein, N. 2021. The Effect: An Introduction to Research Design and Causality. CRC Press. https://theeffectbook.net/index.html.
Kamath, U., K. Graham, and M. Naylor. 2023. Applied Causal Inference. Amazon Digital Services LLC - Kdp. https://appliedcausalinference.github.io/aci_book/.
Loh, Wen Wei, and Dongning Ren. 2023. “A Tutorial on Causal Inference in Longitudinal Data with Time-Varying Confounding Using g-Estimation.” Advances in Methods and Practices in Psychological Science 6 (3). https://doi.org/10.1177/25152459231174029.
Lübke, Karsten, Matthias Gehrke, Jörg Horst, and Gero Szepannek and. 2020. “Why We Should Teach Causal Inference: Examples in Linear Regression with Simulated Data.” Journal of Statistics Education 28 (2): 133–39. https://doi.org/10.1080/10691898.2020.1752859.
Morgan, Stephen L., and Christopher Winship. 2007. Counterfactuals and Causal Inference: Methods and Principles for Social Research. 1st ed. New York: Cambridge University Press.
Petersen, Maya, and Laura Balzer. 2014. “Introduction to Causal Inference.” https://ctml.berkeley.edu/introduction-causal-inference.
Poppe, Louise, Johan Steen, Wen Wei Loh, Geert Crombez, Fien De Block, Noortje Jacobs, Peter W. G. Tennant, Jelle Van Cauwenberg, and Annick L. De Paepe. 2025. “How to Develop Causal Directed Acyclic Graphs for Observational Health Research: A Scoping Review.” Health Psychology Review 19 (1): 45–65. https://doi.org/10.1080/17437199.2024.2402809.
R Core Team. 2024. “R: A Language and Environment for Statistical Computing.” Vienna, Austria: R Foundation for Statistical Computing. https://www.R-project.org/.
Rich, Ben, Erica E. M. Moodie, David A. Stephens, and Robert W. Platt. 2010. “Model Checking with Residuals for g-Estimation of Optimal Dynamic Treatment Regimes.” The International Journal of Biostatistics 6 (2). https://doi.org/10.2202/1557-4679.1210.
Song, Mingyang, Teresa T. Fung, Frank B. Hu, Walter C. Willett, Valter D. Longo, Andrew T. Chan, and Edward L. Giovannucci. 2016. “Association of Animal and Plant Protein Intake with All-Cause and Cause-Specific Mortality.” JAMA Internal Medicine 176 (10): 1453.
Sterne, Jonathan A. C., and Kate Tilling. 2002. “G-Estimation of Causal Effects, Allowing for Time-Varying Confounding.” The Stata Journal 2 (2): 164–82. https://doi.org/10.1177/1536867X0200200205.
van der Wal, Willem M., and Ronald B. Geskus. 2011. ipw: An R Package for Inverse Probability Weighting.” Journal of Statistical Software 43 (13): 1–23. https://doi.org/10.18637/jss.v043.i13.
Van Rossum, Guido, and Fred L. Drake. 2009. Python 3 Reference Manual. Scotts Valley, CA: CreateSpace.
Wallace, Michael P., Erica E. M. Moodie, and David A. Stephens. 2017. “An r Package for g-Estimation of Structural Nested Mean Models.” Epidemiology 28 (2): e18–20. https://doi.org/10.1097/EDE.0000000000000586.
Wallace, Michael, Erica E M Moodie, David A Stephens, Gabrielle Simoneau, Shannon T. Holloway, and Juliana Schulz. 2025. DTRreg: DTR Estimation and Inference via g-Estimation, Dynamic WOLS, q-Learning, and Dynamic Weighted Survival Modeling (DWSurv). https://CRAN.R-project.org/package=DTRreg.
Wang, Yafeng, Wentao Huang, Adrienne O’Neil, Yutao Lan, Dagfinn Aune, Wei Wang, Chuanhua Yu, and Xiong Chen. 2020. “Association Between Sleep Duration and Mortality Risk Among Adults with Type 2 Diabetes: A Prospective Cohort Study.” Diabetologia 63 (11): 2292–2304.