Scaling MADDPG for Enhanced Multi-Agent Reinforcement Learning

Arth Singh,Dhruv Dixit,A. Verma

Published 2025 in 2025 Second International Conference on Pioneering Developments in Computer Science & Digital Technologies (IC2SDT)

ABSTRACT

Addressing multi-agent scenarios is a challenging task for traditional reinforcement learning algorithms due to the environments' dynamic nature from each agent's perspective. In this research, we delve into the performance of multi-agent deep deterministic policy gradients (MAD-DPG), a tailored algorithm for multi-agent scenarios, in the context of a cooperative navigation OpenAI environment involving three agents. Through extensive experimentation, we investigate the impact of various hyperparameters and action exploration strategies on MADDPG's performance. Moreover, we propose enhancements to MAD-DPG to overcome its main weakness, which lies in its limited scalability to more significant numbers of agents. Our findings demonstrate that MAD-DPG outperforms singleagent policy gradient approaches significantly. Furthermore, our innovative improvements to MAD-DPG effectively tackle its primary limitation, allowing the algorithm to scale more efficiently to accommodate larger numbers of agents while only minimally impacting its overall performance.

PUBLICATION RECORD

  • Publication year

    2025

  • Venue

    2025 Second International Conference on Pioneering Developments in Computer Science & Digital Technologies (IC2SDT)

  • Publication date

    2025-12-04

  • Fields of study

    Not labeled

  • Identifiers
  • External record

    Open on Semantic Scholar

  • Source metadata

    Semantic Scholar

CITATION MAP

EXTRACTION MAP

CLAIMS

  • No claims are published for this paper.

CONCEPTS

  • No concepts are published for this paper.

REFERENCES

Showing 1-23 of 23 references · Page 1 of 1

CITED BY

  • No citing papers are available for this paper.

Showing 0-0 of 0 citing papers · Page 1 of 1