🤖 AI Summary
Generative AI exhibits context-conditioned fairness gaps—e.g., disparate representation quality for “low-income individuals” versus “business leaders”—posing critical challenges for equitable deployment across sociocultural contexts.
Method: We propose the first two-tier conditional fairness framework tailored to generative models. First, we formalize context-aware conditional fairness criteria. Second, we introduce a worst-case distance metric and an intrinsic fairness evaluation mechanism under neutral prompts. Third, we design combinatorial testing to assess intersectional fairness and develop a proxy-driven, lightweight prompt injection strategy.
Results: Extensive experiments across multiple state-of-the-art text-to-image models demonstrate that our approach achieves significant suppression of group-specific representational bias with minimal intervention, substantially enhancing contextual fairness assurance. The framework offers a verifiable, deployable paradigm for fairness research in generative AI—bridging theoretical rigor with practical applicability.
📝 Abstract
The deployment of generative AI (GenAI) models raises significant fairness concerns, addressed in this paper through novel characterization and enforcement techniques specific to GenAI. Unlike standard AI performing specific tasks, GenAI's broad functionality requires"conditional fairness"tailored to the context being generated, such as demographic fairness in generating images of poor people versus successful business leaders. We define two fairness levels: the first evaluates fairness in generated outputs, independent of prompts and models; the second assesses inherent fairness with neutral prompts. Given the complexity of GenAI and challenges in fairness specifications, we focus on bounding the worst case, considering a GenAI system unfair if the distance between appearances of a specific group exceeds preset thresholds. We also explore combinatorial testing for accessing relative completeness in intersectional fairness. By bounding the worst case, we develop a prompt injection scheme within an agent-based framework to enforce conditional fairness with minimal intervention, validated on state-of-the-art GenAI systems.