Chapter 12 Resampling methods
12.1 Randomization-Based Methods
Athey and Imbens are prominent economists who have contributed significantly to the development and application of randomization-based methods in econometrics. These methods are particularly valuable for inferring the probability that an estimated coefficient is not simply a result of chance. Here’s an explanation of the context and importance of their work:
12.1.1 Traditional Methods vs. Randomization-Based Methods
Traditional Methods: Economists often use parametric tests (like t-tests or F-tests) which rely on assumptions about the distribution of the error terms (e.g., normality, homoscedasticity). These methods use asymptotic approximations to calculate p-values, which may not always be accurate, especially in small samples or when assumptions are violated.
Randomization-Based Methods: These methods, such as permutation tests, do not rely on these distributional assumptions. Instead, they use the actual data to construct the distribution of the test statistic under the null hypothesis. This approach involves reshuffling the treatment labels or data points multiple times to create a distribution of the test statistic that purely reflects random chance.
12.1.2 Why Use Randomization-Based Methods?
Exact P-Values: Randomization methods can construct exact p-values, which provide a more precise measure of the probability that the observed effect could have arisen by chance. This is especially important in cases where traditional assumptions do not hold.
Robustness: These methods are robust to violations of common statistical assumptions (e.g., non-normality, heteroscedasticity). They rely on the randomization process used in the experiment or observational study, ensuring that the inferences are valid regardless of the underlying data distribution.
Small Samples: In small sample sizes, randomization-based methods are particularly advantageous because traditional asymptotic approximations may be unreliable.
12.1.3 How Randomization-Based Methods Work
Calculate the Observed Statistic: First, calculate the test statistic (e.g., the difference in means, regression coefficient) from the observed data.
Randomize Data: Randomly reassign the treatment labels or shuffle the data points many times (typically thousands). Each randomization should maintain the structure of the data but break any systematic association between treatment and outcome.
Calculate Test Statistics for Randomizations: For each randomization, calculate the test statistic. This generates a distribution of the test statistic under the null hypothesis of no treatment effect.
Compare Observed Statistic: Compare the observed test statistic to this distribution. The p-value is the proportion of randomized test statistics that are as extreme as or more extreme than the observed statistic.
12.1.4 Example
Suppose Athey and Imbens are evaluating the effect of a training program on income:
Observed Difference in Means: The difference in mean income between the treatment (received training) and control (no training) groups is calculated.
Randomization: The treatment labels (received training or not) are randomly shuffled, and the difference in means is recalculated for each shuffle. This process is repeated many times.
Distribution of Differences: The randomization process generates a distribution of differences in means under the null hypothesis that training has no effect.
Calculate P-Value: The p-value is the proportion of differences from the randomization distribution that are as extreme or more extreme than the observed difference.
12.1.5 Contribution of Athey and Imbens
Athey and Imbens’ work emphasizes the use of these robust, non-parametric methods to provide more reliable and exact p-values. Their approach is part of a broader trend in econometrics and other social sciences toward more reliable inference methods that do not heavily rely on restrictive assumptions. By focusing on randomization-based methods, they help ensure that the conclusions drawn from empirical studies are more credible and less sensitive to potential violations of traditional model assumptions.
This methodology is particularly valuable in experimental economics, field experiments, and observational studies where treatment effects are being evaluated. By using randomization-based methods, researchers can make stronger causal inferences and provide more robust evidence for policy and decision-making.
12.1.5.1 Jackknife
The jackknife method is a resampling technique where each observation (in this case, each treated unit) is systematically omitted from the dataset, and the analysis is repeated each time to estimate the variance of the treatment effect. This provides a robust estimate of the standard errors that accounts for the potential variability across different treated units.
12.2 Bootstrapping
bootstrapping is a procedure used to estimate the variance of an estimator. In the context of inverse probability weighting, we would repeatedly draw (“with replacement”) a random sample of our original data and then use that smaller sample to calculate the sample analogs of the ATE or ATT.
standard bootstrapping
wild bootstrapping