Have a personal or library account? Click to login
Multi Agent Deep Learning with Cooperative Communication Cover
Open Access
|May 2020

Abstract

We consider the problem of multi agents cooperating in a partially-observable environment. Agents must learn to coordinate and share relevant information to solve the tasks successfully. This article describes Asynchronous Advantage Actor-Critic with Communication (A3C2), an end-to-end differentiable approach where agents learn policies and communication protocols simultaneously. A3C2 uses a centralized learning, distributed execution paradigm, supports independent agents, dynamic team sizes, partially-observable environments, and noisy communications. We compare and show that A3C2 outperforms other state-of-the-art proposals in multiple environments.

Language: English
Page range: 189 - 207
Submitted on: Nov 1, 2019
|
Accepted on: Mar 26, 2020
|
Published on: May 23, 2020
Published by: SAN University
In partnership with: Paradigm Publishing Services
Publication frequency: 4 issues per year

© 2020 David Simões, Nuno Lau, Luís Paulo Reis, published by SAN University
This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License.