Applying multi-agent reinforcement learning (MARL) to real-world scenarios is challenging because agents often need to adapt quickly to unexpected situations, including those rarely or never encountered in training. Recent methods for out-of-distribution generalization are unsuitable for applications on out-of-distribution tasks with limited communication, because they are typically restricted to centralized training or some specialized instances of distribution shifts. To address this limitation, we introduce the Unexpectedness Encoding Scheme, a new decentralized MARL algorithm in which agents communicate “unexpectedness,” the surprising aspects of the environment. In addition to sending their usual reward-driven messages, each agent predicts the next observation based on past experience and then compares this prediction with the actual outcome. The discrepancy between the two is encoded as a message, enabling agents to adapt more effectively to sudden or extreme changes. Experimental results on multi-agent cooperative tasks demonstrate that our method adapts robustly to both dynamically changing training environments and previously unseen out-of-distribution scenarios.
Communicating Unexpectedness for Out-of-Distribution Multi-Agent Reinforcement Learning
Min Whoo Lee,Yunsu Lee,Kibeom Kim,Soo Wung Shin,Jun Ki Lee,Min Whoo Lee,Byoung-Tak Zhang
Published 2026 in IEEE Access
ABSTRACT
PUBLICATION RECORD
- Publication year
2026
- Venue
IEEE Access
- Publication date
Unknown publication date
- Fields of study
Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-31 of 31 references · Page 1 of 1
CITED BY
- No citing papers are available for this paper.
Showing 0-0 of 0 citing papers · Page 1 of 1