CausalPy logo

CausalPy - causal inference for quasi-experiments

A Python package focussing on causal inference for quasi-experiments. The package allows users to use different model types. Sophisticated Bayesian methods can be used, harnessing the power of PyMC and ArviZ. But users can also use more traditional Ordinary Least Squares estimation methods via scikit-learn models.

Installation

To get the latest release:

pip install CausalPy

Alternatively, if you want the very latest version of the package you can install from GitHub:

pip install git+https://github.com/pymc-labs/CausalPy.git

Quickstart

import causalpy as cp


# Import and process data
df = (
   cp.load_data("drinking")
   .rename(columns={"agecell": "age"})
   .assign(treated=lambda df_: df_.age > 21)
   )

# Run the analysis
result = cp.pymc_experiments.RegressionDiscontinuity(
   df,
   formula="all ~ 1 + age + treated",
   running_variable_name="age",
   model=cp.pymc_models.LinearRegression(),
   treatment_threshold=21,
   )

# Visualize outputs
fig, ax = result.plot();

# Get a results summary
result.summary()

Videos

Features

Different quasi-experimental methods

Rather than focussing on one particular quasi-experimental setting, this package aims to have broad applicability. We can analyse data from the following quasi-experimental methods:

Synthetic control

This is appropriate when you have multiple units, one of which is treated. You build a synthetic control as a weighted combination of the untreated units.

_images/synthetic_control_pymc.svg

Geographical Lift / Geolift

We can also use synthetic control methods to analyse data from geographical lift studies. For example, we can try to evaluate the causal impact of an intervention (e.g. a marketing campaign) run in one geographical area by using control geographical areas which are similar to the intervention area but which did not recieve the specific marketing intervention.

ANCOVA

This is appropriate when you have a single pre and post intervention measurement and have a treament and a control group.

_images/anova_pymc.svg

Difference in differences

This is appropriate when you have pre and post intervention measurement(s) and have a treament and a control group.

_images/difference_in_differences_pymc.svg

Regression discontinuity

Regression discontinuity designs are used when treatment is applied to units according to a cutoff on a running variable, which is typically not time. By looking for the presence of a discontinuity at the precise point of the treatment cutoff then we can make causal claims about the potential impact of the treatment.

_images/regression_discontinuity_pymc.svg

Regression kink designs

Regression kink designs are used when there is a change in the level of treatment at a “kink point” on a running variable, which is typically not time. By looking for the presence of a discontinuity in the gradient at the kink point then we can make causal claims about the potential impact of changes in the treatment.

_images/regression_kink_pymc.svg

Interrupted time series

Interrupted time series analysis is appropriate when you have a time series of observations which undergo treatment at a particular point in time. This kind of analysis has no control group and looks for the presence of a change in the outcome measure at or soon after the treatment time.

_images/interrupted_time_series_pymc.svg

Instrumental Variable Regression

Instrumental Variable regression is an appropriate technique when you wish to estimate the treatment effect of some variable on another, but are concerned that the treatment variable is endogenous in the system of interest i.e. correlated with the errors. In this case an “instrument” variable can be used in a regression context to disentangle treatment effect due to the threat of confounding due to endogeneity.

_images/iv_reg2.png

Support

This repository is supported by PyMC Labs.

_images/pymc-labs-log.png

Documentation outline

Index