🤖 AI Summary
Hartigan’s Dip test for multimodality detection in empirical distributions suffers from sample-size sensitivity and reliance on lookup tables, limiting its practical utility. To address these limitations, we propose the Z-Dip statistic: a standardized version of the Dip value calibrated via simulation-based null distribution estimation and bootstrap resampling, augmented with a downsampling strategy to suppress residual sample-size bias in large datasets—thereby enabling a sample-size-invariant, unified decision threshold. This eliminates dependence on lookup tables and yields a fully reproducible, adaptive multimodality test. An open-source toolkit provides ready-to-use implementation and precomputed calibration tables. Experiments demonstrate that Z-Dip maintains stable Type-I error control across diverse sample sizes while achieving superior statistical power, significantly enhancing the accuracy, robustness, and generalizability of dip-based multimodality assessment in real-world data analysis.
📝 Abstract
Detecting multimodality in empirical distributions is a fundamental problem in statistics and data analysis, with applications ranging from clustering to social science. Hartigan's Dip Test is a classical nonparametric procedure for testing unimodality versus multimodality, but its interpretation is hindered by strong dependence on sample size and the need for lookup tables. We introduce the Z-Dip, a standardized extension of the Dip Test that removes sample-size dependence by comparing observed Dip values to simulated null distributions. We calibrate a universal decision threshold for Z-Dip via simulation and bootstrap resampling, providing a unified criterion for multimodality detection. In the final section, we also propose a downsampling-based approach to further mitigate residual sample-size effects in very large datasets. Lookup tables and software implementations are made available for efficient use in practice.