2018 IEEE International Conference on Systems, Man, and Cybernetics (SMC) 2018
DOI: 10.1109/smc.2018.00717
|View full text |Cite
|
Sign up to set email alerts
|

Deep Reinforcement Learning with Fully Convolutional Neural Network to Solve an Earthwork Scheduling Problem

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 19 publications
0
3
0
Order By: Relevance
“…The MLP used in this work is a fully-connected layer which is deployed with fine regression. As a MLP is frequently used in reinforcement learning [26], [27], it is more likely to show feasible results when the curriculum settings from reinforcement learning are followed, where curriculum learning proceeds during all training epochs.…”
Section: Curriculum Learning For Multi-layer Perceptronmentioning
confidence: 99%
“…The MLP used in this work is a fully-connected layer which is deployed with fine regression. As a MLP is frequently used in reinforcement learning [26], [27], it is more likely to show feasible results when the curriculum settings from reinforcement learning are followed, where curriculum learning proceeds during all training epochs.…”
Section: Curriculum Learning For Multi-layer Perceptronmentioning
confidence: 99%
“…In PixelRL approach, reinforcement learning (RL) is combined with state-of-theart image processing techniques like Convolutional Neural Networks (CNN) to solve real-time complex computer vision problems. Scheduling of important tasks or finding a shortest path between two points in images are few more examples where CNN extracts the features from images and RL learns the optimized way to proceed and perform scheduled task [337], [336] and [338].…”
Section: F Computer Visionmentioning
confidence: 99%
“…The RL has been applied in many fields, such as in robotics, control, multiagent systems and optimization (Gambardella and Dorigo 2000;Kober et al 2013;Shao et al 2014;Bianchi et al 2015;Yliniemi and Tumer 2016;Da Silva et al 2019;Mnih et al 2015;Asiain et al 2019;Alipour et al 2018;Carvalho et al 2019;Li et al 2019;Low et al 2019;Bazzan 2019;Da Silva et al 2019). A growing interesting to apply the RL can be seen in combinatorial optimization (Gambardella and Dorigo 1995;Likas et al 1995;Miagkikh and Punch 1999;Mariano and Morales 2000;Sun et al 2001;Ma et al 2008;Liu and Zeng 2009;Lima Júnior et al 2010;Santos et al 2014;Alipour and Razavi 2015;Alipour et al 2018;Ottoni et al 2018;Woo et al 2018;Miki et al 2018;Chhabra and Warn 2019), such as the travelling salesman problem (TSP) (Gambardella and Dorigo 1995;Alipour et al 2018), Job-Shop Problem (Zhang and Dietterich 1995;Cunha et al 2020), the K-Server Problem (Costa et al 2016) and the multidimensional knapsack problem (MKP) (Arin and Rabadi 2017;Ottoni et al 2017). Although, it seems evident that a great number of works have been devoted to solving combinatorial optimization, less attention has been paid to the sequential ordering problem (SOP)…”
Section: Introductionmentioning
confidence: 99%