2019
DOI: 10.1613/jair.1.11396
|View full text |Cite
|
Sign up to set email alerts
|

A Survey on Transfer Learning for Multiagent Reinforcement Learning Systems

Abstract: Multiagent Reinforcement Learning (RL) solves complex tasks that require coordination with other agents through autonomous exploration of the environment. However, learning a complex task from scratch is impractical due to the huge sample complexity of RL algorithms. For this reason, reusing knowledge that can come from previous experience or other agents is indispensable to scale up multiagent RL algorithms. This survey provides a unifying view of the literature on knowledge reuse in mult… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
119
0
8

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 211 publications
(127 citation statements)
references
References 112 publications
0
119
0
8
Order By: Relevance
“…Since then, the number of published MAL works continues to steadily rise, which led to different surveys on the area, ranging from analyzing the basics of MAL and their challenges [3,4,5], to addressing specific subareas: game theory and MAL [2,6], cooperative scenarios [7,8], and evolutionary dynamics of MAL [9]. In just the last couple of years, three surveys related to MAL have been published: learning in non-stationary environments [10], agents modeling agents [11], and transfer learning in multiagent RL [12].…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…Since then, the number of published MAL works continues to steadily rise, which led to different surveys on the area, ranging from analyzing the basics of MAL and their challenges [3,4,5], to addressing specific subareas: game theory and MAL [2,6], cooperative scenarios [7,8], and evolutionary dynamics of MAL [9]. In just the last couple of years, three surveys related to MAL have been published: learning in non-stationary environments [10], agents modeling agents [11], and transfer learning in multiagent RL [12].…”
Section: Introductionmentioning
confidence: 99%
“…Despite this complexity, top AI conferences like AAAI, ICML, ICLR, IJCAI and NeurIPS, and specialized conferences such as AAMAS, have published works reporting successes in MDRL. In light of these works, we believe it is pertinent to first, have an overview of the recent MDRL works, and second, understand how these recent works relate to the existing literature.This article contributes to the state of the art with a brief survey of the current works in MDRL in an effort to complement existing surveys on multiagent learning [36,10], cooperative learning [7,8], agents modeling agents [11], knowledge reuse in multiagent RL [12], and (singleagent) deep reinforcement learning [23,37].First, we provide a short review of key algorithms in RL such as Q-learning and REINFORCE (see Section 2.1). Second, we review DRL highlighting the challenges in this setting and reviewing recent works (see Section 2.2).…”
mentioning
confidence: 99%
“…The effectiveness of learning in CNN models can be improved even further. There are many important factors to consider, such as improving model weight initialization by transfer learning or using data augmentation and dropout as methods of regularization to combat overfitting during model training [ 16 , 17 , 18 ]. In training CNN models, a large dataset is needed for the model to learn the patterns of features that are complex in detail so that the CNN model can classify those features, achieving an appropriate classification performance [ 19 , 20 ].…”
Section: Introductionmentioning
confidence: 99%
“…The automatic generation of curricula [Da Silva and Costa, 2019] has been divided into two sub-problems: task generation [Narvekar et al, 2016;Da Silva and Costa, 2018], that is the problem of creating a set of tasks such that transferring from them is most likely beneficial for the final task; and task sequencing [Svetlik et al, 2017;Narvekar et al, 2017;Da Silva and Costa, 2018;Foglino et al, 2019], whereby previously generated tasks are optimally selected and ordered. Current methods for task sequencing attempt to determine the optimal order of tasks either with [Narvekar et al, 2017;Baranes and Oudeyer, 2013] or without [Svetlik et al, 2017;Da Silva and Costa, 2018] executing the tasks.…”
Section: Related Workmentioning
confidence: 99%