BiasMap: Leveraging Cross-Attentions to Discover and Mitigate Hidden Social Biases in Text-to-Image Generation

📅 2025-09-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses implicit concept-level social biases—e.g., gender or racial stereotypes—in text-to-image diffusion models (e.g., Stable Diffusion). We propose BiasMap, a model-agnostic framework for bias discovery and mitigation. Unlike prior methods that analyze only output distributions, BiasMap introduces cross-attention attribution maps to quantify the entanglement between sensitive attributes (e.g., gender, race) and semantic concepts (e.g., occupation) within the latent representation space, and defines SoftIoU as an interpretable coupling metric. Furthermore, we integrate energy-guided diffusion sampling to directly optimize the latent noise space during denoising, enforcing attribute–concept disentanglement. Experiments demonstrate that BiasMap significantly improves statistical fairness in generated images while substantially reducing spatial coupling between sensitive attributes and semantic concepts at the representation level—enhancing both transparency and controllability of the generation process.

Technology Category

Application Category

📝 Abstract
Bias discovery is critical for black-box generative models, especiall text-to-image (TTI) models. Existing works predominantly focus on output-level demographic distributions, which do not neces- sarily guarantee concept representations to be disentangled post- mitigation. We propose BiasMap, a model-agnostic framework for uncovering latent concept-level representational biases in stable dif- fusion models. BiasMap leverages cross-attention attribution maps to reveal structural entanglements between demographics (e.g., gender, race) and semantics (e.g., professions), going deeper into representational bias during the image generation. Using attribu- tion maps of these concepts, we quantify the spatial demographics- semantics concept entanglement via Intersection over Union (IoU), offering a lens into bias that remains hidden in existing fairness dis- covery approaches. In addition, we further utilize BiasMap for bias mitigation through energy-guided diffusion sampling that directly modifies latent noise space and minimizes the expected SoftIoU dur- ing the denoising process. Our findings show that existing fairness interventions may reduce the output distributional gap but often fail to disentangle concept-level coupling, whereas our mitigation method can mitigate concept entanglement in image generation while complementing distributional bias mitigation.
Problem

Research questions and friction points this paper is trying to address.

Discover hidden social biases in text-to-image generation models
Quantify demographic-semantic entanglement using cross-attention maps
Mitigate concept-level bias through energy-guided diffusion sampling
Innovation

Methods, ideas, or system contributions that make the work stand out.

Cross-attention attribution maps reveal biases
Intersection over Union quantifies concept entanglement
Energy-guided diffusion sampling minimizes bias
🔎 Similar Papers
No similar papers found.