🤖 AI Summary
Traditional conformal prediction (CP) struggles with the structural complexity and dynamic nature of multimodal, streaming, and large-scale data. To address this, we reconceptualize CP from a data-centric perspective, proposing a novel methodological framework tailored to modern data science. Our approach designs calibration and ensemble construction strategies adaptable to structured, unstructured, and dynamically evolving data, integrating permutation tests, quantile regression, online learning, and adaptive reweighting—enabling distribution-free uncertainty quantification even for black-box models. It uniformly supports diverse modalities—including images, text, and time series—and introduces a new evaluation criterion that jointly optimizes validity and computational efficiency in large-model and big-data settings. Empirically, our framework significantly enhances CP’s applicability, scalability, and practical utility in real-world complex scenarios.
📝 Abstract
Conformal prediction (CP), a distribution-free uncertainty quantification (UQ) framework, reliably provides valid predictive inference for black-box models. CP constructs prediction sets that contain the true output with a specified probability. However, modern data science diverse modalities, along with increasing data and model complexity, challenge traditional CP methods. These developments have spurred novel approaches to address evolving scenarios. This survey reviews the foundational concepts of CP and recent advancements from a data-centric perspective, including applications to structured, unstructured, and dynamic data. We also discuss the challenges and opportunities CP faces in large-scale data and models.