Decentralized_Multi_Agent_Advantage_Actor_Critic_Abstract.pdf (251.67 kB)
Download fileDecentralized Multi-Agent Advantage Actor-Critic
We present a decentralized advantage actor-critic algorithm that utilizes learning agents in parallel environments with synchronous gradient descent. This approach decorrelates agents’ experiences, stabilizing observations and eliminating the need for a replay buffer, requires no knowledge of the other agents’ internal state during training or execution, and runs on a single multi-core CPU.
History
Email Address of Submitting Author
scottgbarnes@gwu.eduSubmitting Author's Institution
The George Washington UniversitySubmitting Author's Country
- United States of America