Modeling human reputation-seeking behavior in a spatio-temporally complex public good provision game

📅 2025-06-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates reputation-driven cooperation mechanisms for public goods provision in spatiotemporal complex environments. We propose a multi-agent reinforcement learning (MARL)-based reputation modeling framework and validate it—within the Clean Up spatiotemporal social dilemma—using real human behavioral data, marking the first such empirical validation of reputation-guided cooperation in MARL. Results show that under identifiable-identity conditions, both humans and agents spontaneously evolve turn-taking strategies, achieving high cooperation success rates (>85%); under anonymity, cooperation rapidly collapses (<20%). This emergent turn-taking mechanism transcends classical game-theoretic predictions, revealing that identity identifiability sustains cooperation not primarily through punishment, but by enabling strategic reciprocity. Our findings provide critical theoretical foundations for modeling socially situated agents and designing institutional mechanisms that leverage reputational incentives.

Technology Category

Application Category

📝 Abstract
Multi-agent reinforcement learning algorithms are useful for simulating social behavior in settings that are too complex for other theoretical approaches like game theory. However, they have not yet been empirically supported by laboratory experiments with real human participants. In this work we demonstrate how multi-agent reinforcement learning can model group behavior in a spatially and temporally complex public good provision game called Clean Up. We show that human groups succeed in Clean Up when they can see who is who and track reputations over time but fail under conditions of anonymity. A new multi-agent reinforcement learning model of reputation-based cooperation demonstrates the same difference between identifiable and anonymous conditions. Furthermore, both human groups and artificial agent groups solve the problem via turn-taking despite other options being available. Our results highlight the benefits of using multi-agent reinforcement learning to model human social behavior in complex environments.
Problem

Research questions and friction points this paper is trying to address.

Modeling human reputation-seeking in complex public good games
Validating multi-agent reinforcement learning with human experiments
Comparing cooperation strategies in identifiable vs anonymous conditions
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multi-agent reinforcement learning models reputation-based cooperation
Identifiable vs anonymous conditions affect cooperation success
Turn-taking strategy solves complex public good provision
🔎 Similar Papers
No similar papers found.
Edward Hughes
Edward Hughes
Staff Research Scientist, DeepMind
reinforcement learningmulti-agent systemsopen-endedness
T
Tina Zhu
Google DeepMind, London, UK
M
Martin J. Chadwick
Google DeepMind, London, UK
R
Raphael Koster
Google DeepMind, London, UK
A
Antonio García Castañeda
Google DeepMind, London, UK
C
Charles Beattie
Google DeepMind, London, UK
Thore Graepel
Thore Graepel
Global Lead Computational Science, AI & ML at Altos Labs and Chair of Machine Learning, UCL
Machine LearningReinforcement LearningDeep LearningMulti-Agent LearningComputational Biology
M
Matthew M. Botvinick
Gatsby Computational Neuroscience Unit, UCL, London, UK
J
Joel Z. Leibo
Google DeepMind, London, UK