Dor Leventer
Fall 2022
A causal inference framework is made up of the following components:
Today we will focus on validation tests
All our causal inference frameworks have assumptions
Question: how to perform these validation tests?
We usually call such a test pre-trend testing.
We now turn to the formal statistical test
This has several problems, which we will now go over.
Say, for a second, we are testing whether some treatment has some effect.
Thats how we construct the test statistic
Going back to the PT example
Point 1: convential testing assumes identification holds, when we want to assume it doesn’t.
When testing for treatment effects, we (want to?) control for rate of type I error.
| PT true | PT false | |
|---|---|---|
| Reject PT | \(\mathbb{P}(\text{PT holds and we reject it})=\alpha\) | \(\mathbb{P}(\text{PT doesn't hold and we reject it})=1-\alpha\) |
| Accept PT | \(\mathbb{P}(\text{PT holds and we accept it})=1-\beta\) | \(\mathbb{P}(\text{PT doesn't hold and we accept it})=\beta\) |
Point 2: convential testing controls for type I when we want to control for type II.
Point 1: convential testing assumes identification holds, when we want to assume it doesn’t.
Point 2: convential testing controls for type I when we want to control for type II.
\(\rightarrow\) next approch please
Point 1: convential testing assumes identification holds, when we want to assume it doesn’t.
Point 2: convential testing controls for type I when we want to control for type II.
Point 1: convential testing assumes identification holds, when we want to assume it doesn’t.
Point 2: convential testing controls for type I when we want to control for type II.
We say that the estimator and truth are equivalent if they are less then \(k\) apart.
Hence the first point above is solved. What about controlling for the correct error type?
But first, a statistical test.
We want to test \(H_{0}:\left|\Delta Y_{1,2}-\Delta Y_{0,2}\right|>k\)
\(\rightarrow\) \(\alpha\) now controls for type II error
Hartman and Hidalgo (2018) show that we can do the above test
This was the plot from before, using \(\alpha = 0.1\)
Lets focus on the period -5. The estimated difference is \(-1.1\).
If we set \(k=2.5\), both one sided tests are not rejected, and so deemed equivalent
If we set \(k=1\), one test is rejected, and hence deemed not equivalent (PT fails)
To build more intuition, consider \(t=-6\). Zero is not rejected.
If we set \(k = 0.5\), zero is accepted (not different) but equivalence is rejected (yes different)
Hartman and Hidalgo (2018) discuss these tests in the context of balance tables for RCTs
In most of our context, seems hard to argue for correct range
Lets again look at a tentative graph
Lets focus on the pre-trends
Get rid of conventional error bars
Add equivalence range - minimal \(k\) that rejects \(H_0\) of not equivalent at \(\alpha=0.05\)
Add equivalence interval: 0.36 standard deviation in control group
There is some interesting stuff going on in this area
Main pro:
Main con:
But, maybe not a con? This is more work for researchers (arg, again with the econometricians producing work for applied…)
Thanks for listening.
All code (and hence slides) is available at Git repo https://github.com/dorlev3/equiv_test_and_identification_talk
Dor Leventer, Equivalence Tests, Fall 2022