2018
DOI: 10.19153/cleiej.21.2.1
|View full text |Cite
|
Sign up to set email alerts
|

Towards Autonomous Reinforcement Learning: Automatic Setting of Hyper-parameters using Bayesian Optimization

Abstract: With the increase of machine learning usage by industries and scientific communities in a variety of tasks such as text mining, image recognition and self-driving cars, automatic setting of hyper-parameter in learning algorithms is a key factor for obtaining good performances regardless of user expertise in the inner workings of the techniques and methodologies. In particular, for a reinforcement learning algorithm, the efficiency of an agent learning a control policy in an uncertain environment is heavily dep… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
12
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
4
2

Relationship

2
4

Authors

Journals

citations
Cited by 11 publications
(12 citation statements)
references
References 25 publications
0
12
0
Order By: Relevance
“…Other studies, on the other hand, focused on resorting to Bayesian optimization to optimize several RL hyper-parameters at the same time. In this setting, a preceding work is Barsce et al (2017) 44 , where a Bayesian optimization framework was proposed to optimize RL hyper-parameters. However, in such work, Bayesian optimization and the RL algorithm are decoupled in such a way that the meta-learning does not make specific assumptions with respect to an RL algorithm, making the method also inefficient because the learned tuples of experience ( , , , ′ ) are all aggregated in the metric selected for the objective function and, therefore, cannot be used directly to improve the meta-learning layer.…”
Section: Related Work In Hyper-parameter Optimizationmentioning
confidence: 99%
See 1 more Smart Citation
“…Other studies, on the other hand, focused on resorting to Bayesian optimization to optimize several RL hyper-parameters at the same time. In this setting, a preceding work is Barsce et al (2017) 44 , where a Bayesian optimization framework was proposed to optimize RL hyper-parameters. However, in such work, Bayesian optimization and the RL algorithm are decoupled in such a way that the meta-learning does not make specific assumptions with respect to an RL algorithm, making the method also inefficient because the learned tuples of experience ( , , , ′ ) are all aggregated in the metric selected for the objective function and, therefore, cannot be used directly to improve the meta-learning layer.…”
Section: Related Work In Hyper-parameter Optimizationmentioning
confidence: 99%
“…The reasoning behind integrating Bayesian optimization with an RL algorithm is that using a black-box approach is well suited for an expensive optimization task such as RL, by making the most of past queries in order to maximize the gain of selecting the next query of hyper-parameter setting through the acquisition function. As it employs Bayesian optimization for optimizing RL algorithms, this architecture has its foundations in RLOpt 44 .…”
Section: Bayesian Optimization Of An Rl Agentmentioning
confidence: 99%
“…In RL, an agent learns from rewards and penalties in interacting with an environment [68]. One of the main topics of investigation in RL is the estimation of learning parameters, like learning rate (α) and discount factor (γ ), -greedy and reinforcement function [6,17,23,24,40,54,63]. In fact, parameter definition can directly influence a good route learning [5,12,52,54].…”
Section: Introductionmentioning
confidence: 99%
“…Automated hyperparameter selection is a very similar problem to meta-learning since it often uses a higher level learning procedure to "train" the hyperparameters of the lower level algorithm. These automated methods use a variety of intelligent approaches such as evolutionary computation (Schweighofer and Doya 2003;Young et al 2015) and Bayesian optimisation methods (Barsce et al 2017;Bergstra et al 2011).…”
Section: Introductionmentioning
confidence: 99%