Skip to content

A Python package for causal inference in quasi-experimental settings

License

Notifications You must be signed in to change notification settings

westford14/CausalPy

 
 

Repository files navigation


Build Ruff PyPI version GitHub Repo stars Read the Docs PyPI - Downloads Interrogate codecov

CausalPy

A Python package focussing on causal inference in quasi-experimental settings. The package allows for sophisticated Bayesian model fitting methods to be used in addition to traditional OLS.

Installation

To get the latest release:

pip install CausalPy

Alternatively, if you want the very latest version of the package you can install from GitHub:

pip install git+https://github.com/pymc-labs/CausalPy.git

Quickstart

import causalpy as cp
import matplotlib.pyplot as plt

# Import and process data
df = (
    cp.load_data("drinking")
    .rename(columns={"agecell": "age"})
    .assign(treated=lambda df_: df_.age > 21)
    )

# Run the analysis
result = cp.RegressionDiscontinuity(
    df,
    formula="all ~ 1 + age + treated",
    running_variable_name="age",
    model=cp.pymc_models.LinearRegression(),
    treatment_threshold=21,
    )

# Visualize outputs
fig, ax = result.plot();

# Get a results summary
result.summary()

plt.show()

Roadmap

Plans for the repository can be seen in the Issues.

Videos

Click on the thumbnail below to watch a video about CausalPy on YouTube. Youtube video thumbnail image

Overview of package capabilities

Synthetic control

This is appropriate when you have multiple units, one of which is treated. You build a synthetic control as a weighted combination of the untreated units.

Time Outcome Control 1 Control 2 Control 3
0 $y_0$ $x_{1,0}$ $x_{2,0}$ $x_{3,0}$
1 $y_1$ $x_{1,1}$ $x_{2,1}$ $x_{3,1}$
$\ldots$ $\ldots$ $\ldots$ $\ldots$ $\ldots$
T $y_T$ $x_{1,T}$ $x_{2,T}$ $x_{3,T}$
Frequentist Bayesian

The data (treated and untreated units), pre-treatment model fit, and counterfactual (i.e. the synthetic control) are plotted (top). The causal impact is shown as a blue shaded region. The Bayesian analysis shows shaded Bayesian credible regions of the model fit and counterfactual. Also shown is the causal impact (middle) and cumulative causal impact (bottom).

Geographical lift (Geolift)

We can also use synthetic control methods to analyse data from geographical lift studies. For example, we can try to evaluate the causal impact of an intervention (e.g. a marketing campaign) run in one geographical area by using control geographical areas which are similar to the intervention area but which did not recieve the specific marketing intervention.

ANCOVA

This is appropriate for non-equivalent group designs when you have a single pre and post intervention measurement and have a treament and a control group.

Group pre post
0 $x_1$ $y_1$
0 $x_2$ $y_2$
1 $x_3$ $y_3$
1 $x_4$ $y_4$
Frequentist Bayesian
coming soon

The data from the control and treatment group are plotted, along with posterior predictive 94% credible intervals. The lower panel shows the estimated treatment effect.

Difference in Differences

This is appropriate for non-equivalent group designs when you have pre and post intervention measurement and have a treament and a control group. Unlike the ANCOVA approach, difference in differences is appropriate when there are multiple pre and/or post treatment measurements.

Data is expected to be in the following form. Shown are just two units - one in the treated group (group=1) and one in the untreated group (group=0), but there can of course be multiple units per group. This is panel data (also known as repeated measures) where each unit is measured at 2 time points.

Unit Time Group Outcome
0 0 0 $y_{0,0}$
0 1 0 $y_{0,0}$
1 0 1 $y_{1,0}$
1 1 1 $y_{1,1}$
Frequentist Bayesian

The data, model fit, and counterfactual are plotted. Frequentist model fits result in points estimates, but the Bayesian analysis results in posterior distributions, represented by the violin plots. The causal impact is the difference between the counterfactual prediction (treated group, post treatment) and the observed values for the treated group, post treatment.

Regression discontinuity designs

Regression discontinuity designs are used when treatment is applied to units according to a cutoff on the running variable (e.g. $x$) which is typically not time. By looking for the presence of a discontinuity at the precise point of the treatment cutoff then we can make causal claims about the potential impact of the treatment.

Running variable Outcome Treated
$x_0$ $y_0$ False
$x_1$ $y_0$ False
$\ldots$ $\ldots$ $\ldots$
$x_{N-1}$ $y_{N-1}$ True
$x_N$ $y_N$ True
Frequentist Bayesian

The data, model fit, and counterfactual are plotted (top). Frequentist analysis shows the causal impact with the blue shaded region, but this is not shown in the Bayesian analysis to avoid a cluttered chart. Instead, the Bayesian analysis shows shaded Bayesian credible regions of the model fits. The Frequentist analysis visualises the point estimate of the causal impact, but the Bayesian analysis also plots the posterior distribution of the regression discontinuity effect (bottom).

Regression kink designs

Regression discontinuity designs are used when treatment is applied to units according to a cutoff on a running variable, which is typically not time. By looking for the presence of a discontinuity at the precise point of the treatment cutoff then we can make causal claims about the potential impact of the treatment.

Running variable Outcome
$x_0$ $y_0$
$x_1$ $y_0$
$\ldots$ $\ldots$
$x_{N-1}$ $y_{N-1}$
$x_N$ $y_N$
Frequentist Bayesian
coming soon

The data and model fit. The Bayesian analysis shows the posterior mean with credible intervals (shaded regions). We also report the Bayesian $R^2$ on the data along with the posterior mean and credible intervals of the change in gradient at the kink point.

Interrupted time series

Interrupted time series analysis is appropriate when you have a time series of observations which undergo treatment at a particular point in time. This kind of analysis has no control group and looks for the presence of a change in the outcome measure at or soon after the treatment time. Multiple predictors can be included.

Time Outcome Treated Predictor
$t_0$ $y_0$ False $x_0$
$t_1$ $y_0$ False $x_1$
$\ldots$ $\ldots$ $\ldots$ $\ldots$
$t_{N-1}$ $y_{N-1}$ True $x_{N-1}$
$t_N$ $y_N$ True $x_N$
Frequentist Bayesian

The data, pre-treatment model fit, and counterfactual are plotted (top). The causal impact is shown as a blue shaded region. The Bayesian analysis shows shaded Bayesian credible regions of the model fit and counterfactual. Also shown is the causal impact (middle) and cumulative causal impact (bottom).

Instrumental Variable Regression

Instrumental Variable regression is an appropriate technique when you wish to estimate the treatment effect of some variable on another, but are concerned that the treatment variable is endogenous in the system of interest i.e. correlated with the errors. In this case an “instrument” variable can be used in a regression context to disentangle treatment effect due to the threat of confounding due to endogeneity.

Inverse Propensity Score Weighting

Propensity scores are often used to address the risks of bias or confounding introduced in an observational study by selection effects into the treatment condition. Propensity scores can be used in a number of ways, but here we demonstrate their usage within corrective weighting schemes aimed to recover as-if random allocation of subjects to the treatment condition. The technique "up-weights" or "down-weights" individual observations to better estimate a causal estimand such as the average treatment effect.

Learning resources

Here are some general resources about causal inference:

  • The official PyMC examples gallery has a set of examples specifically relating to causal inference.
  • Angrist, J. D., & Pischke, J. S. (2009). Mostly harmless econometrics: An empiricist's companion. Princeton university press.
  • Angrist, J. D., & Pischke, J. S. (2014). Mastering'metrics: The path from cause to effect. Princeton university press.
  • Cunningham, S. (2021). Causal inference: The Mixtape. Yale University Press.
  • Huntington-Klein, N. (2021). The effect: An introduction to research design and causality. Chapman and Hall/CRC.
  • Reichardt, C. S. (2019). Quasi-experimentation: A guide to design and analysis. Guilford Publications.

License

Apache License 2.0


Support

This repository is supported by PyMC Labs.

If you are interested in seeing what PyMC Labs can do for you, then please email [email protected]. We work with companies at a variety of scales and with varying levels of existing modeling capacity. We also run corporate workshop training events and can provide sessions ranging from introduction to Bayes to more advanced topics.

About

A Python package for causal inference in quasi-experimental settings

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 99.8%
  • Makefile 0.2%