🤖 AI Summary
Face verification systems suffer from compounded demographic bias and privacy risks due to skewed training data. Method: This paper proposes a controllable and interpretable fair generative framework. Building upon DCFaces, it introduces the first end-to-end tunable synthetic face generation pipeline enabling demographically balanced sampling across gender and race. It further innovates a joint deep fairness statistical analysis method—integrating logit regression with ANOVA—to quantify bias sources and guide generative adjustments. Contribution/Results: Experiments demonstrate that, while preserving privacy, the framework reduces the equal opportunity (EO) gap across gender and race subgroups by 37%, improves overall verification accuracy by 0.8%, and significantly outperforms existing debiasing approaches.
📝 Abstract
Face recognition and verification are two computer vision tasks whose performances have advanced with the introduction of deep representations. However, ethical, legal, and technical challenges due to the sensitive nature of face data and biases in real-world training datasets hinder their development. Generative AI addresses privacy by creating fictitious identities, but fairness problems remain. Using the existing DCFace SOTA framework, we introduce a new controlled generation pipeline that improves fairness. Through classical fairness metrics and a proposed indepth statistical analysis based on logit models and ANOVA, we show that our generation pipeline improves fairness more than other bias mitigation approaches while slightly improving raw performance.