Addressing multi-agent scenarios is a challenging task for traditional reinforcement learning algorithms due to the environments' dynamic nature from each agent's perspective. In this research, we delve into the performance of multi-agent deep deterministic policy gradients (MAD-DPG), a tailored algorithm for multi-agent scenarios, in the context of a cooperative navigation OpenAI environment involving three agents. Through extensive experimentation, we investigate the impact of various hyperparameters and action exploration strategies on MADDPG's performance. Moreover, we propose enhancements to MAD-DPG to overcome its main weakness, which lies in its limited scalability to more significant numbers of agents. Our findings demonstrate that MAD-DPG outperforms singleagent policy gradient approaches significantly. Furthermore, our innovative improvements to MAD-DPG effectively tackle its primary limitation, allowing the algorithm to scale more efficiently to accommodate larger numbers of agents while only minimally impacting its overall performance.
Scaling MADDPG for Enhanced Multi-Agent Reinforcement Learning
Arth Singh,Dhruv Dixit,A. Verma
Published 2025 in 2025 Second International Conference on Pioneering Developments in Computer Science & Digital Technologies (IC2SDT)
ABSTRACT
PUBLICATION RECORD
- Publication year
2025
- Venue
2025 Second International Conference on Pioneering Developments in Computer Science & Digital Technologies (IC2SDT)
- Publication date
2025-12-04
- Fields of study
Not labeled
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-23 of 23 references · Page 1 of 1
CITED BY
- No citing papers are available for this paper.
Showing 0-0 of 0 citing papers · Page 1 of 1