Multi-Agent Reinforcement Learning for Language-Based Social Deduction

Authors

  • David R. Miller Department of Computer Science, University of Toronto, Toronto, ON M5S 2E4, Canada Author
  • Sophie L. Gagnon Department of Computer Science, University of Toronto, Toronto, ON M5S 2E4, Canada Author
  • James P. Wilson Department of Computer Science, University of Toronto, Toronto, ON M5S 2E4, Canada Author

DOI:

https://doi.org/10.71465/fair719

Keywords:

Social deduction, natural language communication, multi-agent RL

Abstract

This study applies MARL to train LLM agents for language-based social deduction, where communication directly influences multi-agent outcomes. Speaking policies are optimized using rewards that capture the causal impact of messages on other agents’ beliefs, while listening models predict hidden state information from dialogue. Training on 12,000 simulated game episodes results in a 2.1× increase in win rate over standard RL baselines and demonstrates emergent strategic behaviors such as coordinated accusations and evidence sharing.

Downloads

Download data is not yet available.

Downloads

Published

2026-03-10