SurrogateSHAP: Training-Free Contributor Attribution for Text-to-Image (T2I) Models

📅 2026-01-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the lack of efficient and fair attribution mechanisms for data contributors in current text-to-image (T2I) generation models, which hinders equitable compensation and sustainable data markets. The authors propose the first attribution framework that operates without retraining: leveraging the inference process of pretrained T2I models, it employs gradient-boosted trees to model a multidimensional utility function—encompassing image quality, aesthetic scores, and diversity—and efficiently approximates Shapley values via an analytical approach. This method substantially reduces computational complexity while accurately identifying key data contributors. In experiments, it successfully pinpoints sources of spurious correlations in clinical images, outperforming existing attribution techniques in both accuracy and efficiency.

Technology Category

Application Category

📝 Abstract
As Text-to-Image (T2I) diffusion models are increasingly used in real-world creative workflows, a principled framework for valuing contributors who provide a collection of data is essential for fair compensation and sustainable data marketplaces. While the Shapley value offers a theoretically grounded approach to attribution, it faces a dual computational bottleneck: (i) the prohibitive cost of exhaustive model retraining for each sampled subset of players (i.e., data contributors) and (ii) the combinatorial number of subsets needed to estimate marginal contributions due to contributor interactions. To this end, we propose SurrogateSHAP, a retraining-free framework that approximates the expensive retraining game through inference from a pretrained model. To further improve efficiency, we employ a gradient-boosted tree to approximate the utility function and derive Shapley values analytically from the tree-based model. We evaluate SurrogateSHAP across three diverse attribution tasks: (i) image quality for DDPM-CFG on CIFAR-20, (ii) aesthetics for Stable Diffusion on Post-Impressionist artworks, and (iii) product diversity for FLUX.1 on Fashion-Product data. Across settings, SurrogateSHAP outperforms prior methods while substantially reducing computational overhead, consistently identifying influential contributors across multiple utility metrics. Finally, we demonstrate that SurrogateSHAP effectively localizes data sources responsible for spurious correlations in clinical images, providing a scalable path toward auditing safety-critical generative models.
Problem

Research questions and friction points this paper is trying to address.

Text-to-Image
contributor attribution
Shapley value
data valuation
fair compensation
Innovation

Methods, ideas, or system contributions that make the work stand out.

SurrogateSHAP
Shapley value
training-free attribution
text-to-image models
data valuation
🔎 Similar Papers
No similar papers found.
M
Mingyu Lu
Paul G. Allen School of Computer Science & Engineering, University of Washington
S
Soham U. Gadgil
Paul G. Allen School of Computer Science & Engineering, University of Washington
C
Chris Lin
Paul G. Allen School of Computer Science & Engineering, University of Washington
C
Chanwoo Kim
Paul G. Allen School of Computer Science & Engineering, University of Washington
Su-In Lee
Su-In Lee
Computer Science & Engineering, University of Washington
AIMLComputational biology & medicine