2019 IEEE International Conference on Big Data (Big Data) 2019
DOI: 10.1109/bigdata47090.2019.9006027
|View full text |Cite
|
Sign up to set email alerts
|

Multi-task Deep Reinforcement Learning for Scalable Parallel Task Scheduling

Abstract: The ability to leverage shared behaviors between tasks is critical for sample-efficient multi-task reinforcement learning (MTRL). While prior methods have primarily explored parameter and data sharing, direct behavior-sharing has been limited to task families requiring only similar behaviors. Our goal is to extend the efficacy of behaviorsharing to more general task families that could require a mix of shareable and conflicting behaviors. Our key insight is an agent's behavior across tasks can be used for mutu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
6
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(6 citation statements)
references
References 45 publications
0
6
0
Order By: Relevance
“…Afterwards it is compared to frequently used algorithms in the field of real time underwater rescue assignments like a neural network based improved self-organizing map (ISOM) and an improved ant colony optimization (IACO). In [65,66] a Multi-DRL approach for scalable parallel Task Scheduling (MDTS) in autonomous driving is introduced. They evaluate the performance of MDTS by comparing it to other popular schedulers (e.g., PSO, least-connection, A3C) through simulation experiments, showing improvement on various parameters.…”
Section: Dispatching and Schedulingmentioning
confidence: 99%
“…Afterwards it is compared to frequently used algorithms in the field of real time underwater rescue assignments like a neural network based improved self-organizing map (ISOM) and an improved ant colony optimization (IACO). In [65,66] a Multi-DRL approach for scalable parallel Task Scheduling (MDTS) in autonomous driving is introduced. They evaluate the performance of MDTS by comparing it to other popular schedulers (e.g., PSO, least-connection, A3C) through simulation experiments, showing improvement on various parameters.…”
Section: Dispatching and Schedulingmentioning
confidence: 99%
“… Ran et al (2019) used the Deep Determining Policy Gradient (DDPG) algorithm to find the optimal task assignment scheme meeting the requirements of the Service Level Agreement (SLA). Zhang et al (2019) proposed a parallel execution multi-task scheduling algorithm based on deep reinforcement learning. And compared with least connection and particle swarm optimization, this algorithm significantly reduces the completion time of the job.…”
Section: Related Workmentioning
confidence: 99%
“…References ( Cui and Xiaoqing, 2018 ; Xiaoqing et al, 2018 ) gave the corresponding weight coefficients of each resource through subjective experience. References ( Zhang et al, 2019 ; Dong et al, 2020 ) did not take into account the transmission cost between resource nodes of the execution results of services in the actual scheduling process of composite services. In the actual environment, the data transmission time between sub-services affects the completion time and operation cost of composite services to some extent.…”
Section: Related Workmentioning
confidence: 99%
“…DRL combines the perception of Deep Learning and the decision making of Reinforcement Learning [1]. Therefore, DRL can implement a variety of tasks requiring both precious perceptions of high dimensional raw inputs and policy control [1,13]. AlphaGO, which defeated the world Go champion, Sedol Lee is a program that is developed by Google DeepMind following the DRL approach.…”
Section: Deep Learning Reinforcement Learning and Deep Reinforcementmentioning
confidence: 99%
“…Uber is also trying to use DRL to teach Grand Theft Auto to handle real cars on real roads. In one of recent research, a Multi-Task Deep reinforcement learning approach for scalable parallel Task Scheduling (MDTS) has been developed by a group of researchers [13]. But, when interacting with complex concurrent computing environments and tasks with distinct features, DLR experiences the curse of high dimensionality for decision making [13].…”
Section: Deep Learning Reinforcement Learning and Deep Reinforcementmentioning
confidence: 99%