Access the full text.
Sign up today, get DeepDyve free for 14 days.
D. Simões, N. Lau, L. Reis (2020)
Multi-agent actor centralized-critic with communicationNeurocomputing, 390
Yuchen Xiao, Joshua Hoffman, Tian Xia, Chris Amato (2019)
Learning Multi-Robot Decentralized Macro-Action-Based Policies via a Centralized Q-Net2020 IEEE International Conference on Robotics and Automation (ICRA)
Pablo Hernandez-Leal, Bilal Kartal, Matthew Taylor (2018)
A survey and critique of multiagent deep reinforcement learningAutonomous Agents and Multi-Agent Systems, 33
Sainbayar Sukhbaatar, Arthur Szlam, R. Fergus (2016)
Learning Multiagent Communication with Backpropagation
Yuchen Xiao, Xueguang Lyu, Chris Amato (2021)
Local Advantage Actor-Critic for Robust Multi-Agent Deep Reinforcement Learning2021 International Symposium on Multi-Robot and Multi-Agent Systems (MRS)
K Arulkumaran (2017)
26IEEE Signal Proc Mag, 34
(2019)
Individualized controlled continuous communication model for multiagent cooperative and competitive tasks
N Mahdoui (2020)
325J Intell Rob Syst, 98
R. Sutton, A. Barto (2005)
Reinforcement Learning: An IntroductionIEEE Transactions on Neural Networks, 16
Ronald Williams (1992)
Simple statistical gradient-following algorithms for connectionist reinforcement learningMachine Learning, 8
F. D’Souza, João Costa, J. Pires (2020)
Development of a solution for adding a collaborative robot to an industrial AGVInd. Robot, 47
Kai Arulkumaran, M. Deisenroth, Miles Brundage, A. Bharath (2017)
Deep Reinforcement Learning: A Brief SurveyIEEE Signal Processing Magazine, 34
Fei Chen, W. Ren (2019)
On the Control of Multi-Agent Systems: A SurveyFound. Trends Syst. Control., 6
Wei Du, Shifei Ding (2020)
A survey on multi-agent deep reinforcement learning: from the perspective of challenges and applicationsArtificial Intelligence Review, 54
B. Kiran, Ibrahim Sobh, V. Talpaert, P. Mannion, A. Sallab, S. Yogamani, P. P'erez (2020)
Deep Reinforcement Learning for Autonomous Driving: A SurveyIEEE Transactions on Intelligent Transportation Systems, 23
Nesrine Chedly (2018)
Communicating multi-UAV system for cooperative SLAM-based exploration. (Système multi-UAV communicant pour l'exploration coopérative basée sur le SLAM)Journal of Intelligent & Robotic Systems
Y LeCun (2015)
436Nature, 521
Weiwei Liu, Shanqi Liu, Junjie Cao, Qi Wang, Xiaolei Lang, Yong Liu (2021)
Learning Communication for Cooperation in Dynamic Agent-Number EnvironmentIEEE/ASME Transactions on Mechatronics, 26
Yann LeCun, Yoshua Bengio, Geoffrey Hinton (2015)
Deep LearningNature, 521
Ke Xu, Jimmy Ba, Ryan Kiros, Kyunghyun Cho, Aaron Courville, R. Salakhutdinov, R. Zemel, Yoshua Bengio (2015)
Show, Attend and Tell: Neural Image Caption Generation with Visual Attention
T. Nguyen, Ngoc Nguyen, S. Nahavandi (2018)
Deep Reinforcement Learning for Multiagent Systems: A Review of Challenges, Solutions, and ApplicationsIEEE Transactions on Cybernetics, 50
F D’Souza, J Costa, JN Pires (2020)
Development of a solution for adding a collaborative robot to an industrial agvInd Rob:, Int J Rob Res Appl, 47
R Lowe, Y WU, A Tamar, J Harb, O Pieter Abbeel, I Mordatch (2017)
Multi-agent actor-critic for mixed cooperative-competitive environmentsAdv Neural Inf Process Syst, 30
Jakob Foerster, Gregory Farquhar, Triantafyllos Afouras, Nantas Nardelli, Shimon Whiteson (2017)
Counterfactual Multi-Agent Policy GradientsArXiv, abs/1705.08926
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan Gomez, Lukasz Kaiser, Illia Polosukhin (2017)
Attention is All you Need
C. Mu, Ke Wang, Z. Ni (2021)
Adaptive Learning and Sampled-Control for Nonlinear Game Systems Using Dynamic Event-Triggering StrategyIEEE Transactions on Neural Networks and Learning Systems, 33
Tingxiang Fan, Pinxin Long, Wenxi Liu, Jia Pan (2020)
Distributed multi-robot collision avoidance via deep reinforcement learning for navigation in complex scenariosThe International Journal of Robotics Research, 39
Abhishek Das, Théophile Gervet, Joshua Romoff, Dhruv Batra, Devi Parikh, Michael Rabbat, Joelle Pineau (2018)
TarMAC: Targeted Multi-Agent Communication
Ryan Lowe, Yi Wu, Aviv Tamar, J. Harb, P. Abbeel, Igor Mordatch (2017)
Multi-Agent Actor-Critic for Mixed Cooperative-Competitive EnvironmentsArXiv, abs/1706.02275
Shariq Iqbal, Fei Sha (2018)
Actor-Attention-Critic for Multi-Agent Reinforcement Learning
S. Hochreiter, J. Schmidhuber (1997)
Long Short-Term MemoryNeural Computation, 9
N Mahdoui, V Frémont, E Natalizio (2020)
Communicating multi-uav system for cooperative slam-based explorationJ Intell Rob Syst, 98
F Chen (2019)
339Found Trends® Syst Control, 6
F D’Souza (2020)
723Ind Rob:, Int J Rob Res Appl, 47
(2019)
Individualized controlled continuous communication model for multiagent cooperative and competitive tasks. In: International conference on learning representations
Derya Soydaner (2022)
Attention mechanism in neural networks: where it comes and where it goesNeural Computing and Applications, 34
Jiechuan Jiang, Zongqing Lu (2018)
Learning Attentional Communication for Multi-Agent Cooperation
C. Mu, Ke Wang, Changyin Sun (2021)
Learning Control Supported by Dynamic Event Communication Applying to Industrial SystemsIEEE Transactions on Industrial Informatics, 17
Oriol Vinyals, I. Babuschkin, Wojciech Czarnecki, Michaël Mathieu, Andrew Dudzik, Junyoung Chung, David Choi, Richard Powell, Timo Ewalds, Petko Georgiev, Junhyuk Oh, Dan Horgan, M. Kroiss, Ivo Danihelka, Aja Huang, L. Sifre, Trevor Cai, J. Agapiou, Max Jaderberg, A. Vezhnevets, Rémi Leblond, Tobias Pohlen, Valentin Dalibard, D. Budden, Yury Sulsky, James Molloy, T. Paine, Caglar Gulcehre, Ziyun Wang, T. Pfaff, Yuhuai Wu, Roman Ring, Dani Yogatama, Dario Wünsch, Katrina McKinney, Oliver Smith, T. Schaul, T. Lillicrap, K. Kavukcuoglu, D. Hassabis, C. Apps, David Silver (2019)
Grandmaster level in StarCraft II using multi-agent reinforcement learningNature, 575
Multi-agent deep reinforcement learning (MDRL) has attracted attention for solving complex tasks. Two main challenges of MDRL are non-stationarity and partial observability from the perspective of agents, impacting the performance of agents’ learning cooperative policies. In this study, Controlled and Targeted Communication with the Centralized Critic (COTAC) is proposed, thereby constructing the paradigm of centralized learning and decentralized execution with partial communication. It is capable of decoupling how the MAS obtains environmental information during training and execution. Specifically, COTAC can make the environment faced by agents to be stationarity in the training phase and learn partial communication to overcome the limitation of partial observability in the execution phase. Based on this, decentralized actors learn controlled and targeted communication and policies optimized by centralized critics during training. As a result, agents comprehensively learn when to communicate during the sending and how to target information aggregation during the receiving. Apart from that, COTAC is evaluated on two multi-agent scenarios with continuous space. Experimental results demonstrated that partial agents with important information choose to send messages and targeted aggregate received information by identifying the relevant important information, which can still have better cooperation performance while reducing the communication traffic of the system.
Applied Intelligence – Springer Journals
Published: Jun 1, 2023
Keywords: Reinforcement learning; Centralized critic; Communication; Cooperation; Multi-agent system
Read and print from thousands of top scholarly journals.
Already have an account? Log in
Bookmark this article. You can see your Bookmarks on your DeepDyve Library.
To save an article, log in first, or sign up for a DeepDyve account if you don’t already have one.
Copy and paste the desired citation format or use the link below to download a file formatted for EndNote
Access the full text.
Sign up today, get DeepDyve free for 14 days.
All DeepDyve websites use cookies to improve your online experience. They were placed on your computer when you launched this website. You can change your cookie settings through your browser.