🤖 AI Summary
This paper addresses instrument variable (IV) design bias arising from exogenous leniency policy changes in settings with multiple decision-makers or control variables. To mitigate this, we propose a systematic leniency design framework. Our core methodological contribution is the unbiased Jackknife IV estimator (UJIVE), which rigorously tests key identification assumptions—including quasi-random assignment and first-stage monotonicity—and assesses the external validity of treatment effects. The framework integrates IV regression, formal hypothesis testing, and non-clustered standard error inference to avoid conventional clustering misspecification. Empirically, we successfully replicate Farre-Mensa et al. (2020), precisely identifying startup patent value and demonstrating the framework’s robustness and generalizability in complex institutional environments.
📝 Abstract
We develop a step-by-step guide to leniency (a.k.a. judge or examiner instrument) designs, drawing on recent econometric literatures. The unbiased jackknife instrumental variables estimator (UJIVE) is purpose-built for leveraging exogenous leniency variation, avoiding subtle biases even in the presence of many decision-makers or controls. We show how UJIVE can also be used to assess key assumptions underlying leniency designs, including quasi-random assignment and average first-stage monotonicity, and to probe the external validity of treatment effect estimates. We further discuss statistical inference, arguing that non-clustered standard errors are often appropriate. A reanalysis of Farre-Mensa et al. (2020), using quasi-random examiner assignment to estimate the value of patents to startups, illustrates our checklist.