Sie befinden Sich nicht im Netzwerk der Universität Paderborn. Der Zugriff auf elektronische Ressourcen ist gegebenenfalls nur via VPN oder Shibboleth (DFN-AAI) möglich. mehr Informationen...
Ergebnis 4 von 4
2022 International Conference on Computer Engineering and Artificial Intelligence (ICCEAI), 2022, p.706-710
2022
Volltextzugriff (PDF)

Details

Autor(en) / Beteiligte
Titel
Neighborhood Focused Critic Policy Gradients for Multi-agent Reinforcement Learning
Ist Teil von
  • 2022 International Conference on Computer Engineering and Artificial Intelligence (ICCEAI), 2022, p.706-710
Ort / Verlag
IEEE
Erscheinungsjahr
2022
Quelle
IEEE Electronic Library (IEL)
Beschreibungen/Notizen
  • Centralized Training and Decentralized Exucetion (CTDE) is typical training pattern in multi-agent reinforcement learning. In actor-critic method, the actor updates based on the centralized critic which takes global state as criteria for individual contribution. While in cooperative multi-agent tasks, estimating individual contribution in the range of all agents leads to the overlook of local coordination, and thus intensify the mislead in credit assignment. This paper propose Neighbourhood Focused Critic (NFC) Policy Gradients for multi-agent reinforcement learning, parameterizing critic by neighbourhood focused graph neural network in Actor-critic method. NFC allows the centralized critic focus on local coordination in each agent's neighbourhood, and decentralized actors optimizes policies according to the critic estimation, thus alleviating the credit assignment problem. We test our NFC in StarCraft Multi-Agent Challenge (SMAC) environment, the results show that it improves performance and convergence speed significantly comparing to methods with globally focused critic.
Sprache
Englisch
Identifikatoren
DOI: 10.1109/ICCEAI55464.2022.00150
Titel-ID: cdi_ieee_primary_9853370

Weiterführende Literatur

Empfehlungen zum selben Thema automatisch vorgeschlagen von bX