🤖 AI Summary
This work proposes a general, model-agnostic post-processing framework that systematically integrates ensemble learning into fairness optimization. By aggregating predictions from multiple base models without requiring access to their internal architectures, the approach uniformly supports diverse predictive tasks—including classification, regression, and survival analysis—and accommodates a wide range of fairness definitions. The method addresses the inherent trade-off between predictive performance and fairness in machine learning models. Experimental results demonstrate that it significantly enhances fairness while maintaining or only marginally compromising prediction accuracy, thereby validating its effectiveness and generalizability across multiple scenarios.
📝 Abstract
Striking an optimal balance between predictive performance and fairness continues to be a fundamental challenge in machine learning. In this work, we propose a post-processing framework that facilitates fairness-aware prediction by leveraging model ensembling. Designed to operate independently of any specific model internals, our approach is widely applicable across various learning tasks, model architectures, and fairness definitions. Through extensive experiments spanning classification, regression, and survival analysis, we demonstrate that the framework effectively enhances fairness while maintaining, or only minimally affecting, predictive accuracy.