2017 XLIII Latin American Computer Conference (CLEI) 2017
DOI: 10.1109/clei.2017.8226439
|View full text |Cite
|
Sign up to set email alerts
|

Towards autonomous reinforcement learning: Automatic setting of hyper-parameters using Bayesian optimization

Abstract: With the increase of machine learning usage by industries and scientific communities in a variety of tasks such as text mining, image recognition and self-driving cars, automatic setting of hyper-parameter in learning algorithms is a key factor for achieving satisfactory performance regardless of user expertise in the inner workings of the techniques and methodologies. In particular, for a reinforcement learning algorithm, the efficiency of an agent learning a control policy in an uncertain environment is heav… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
3
0
2

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 17 publications
(5 citation statements)
references
References 12 publications
0
3
0
2
Order By: Relevance
“…One popular way of automating algorithm selection is modeling the problem as MAB problem and assigning an action to each algorithm [26]. BO, which works well for automated machine learning, has been used for hyperparameters tuning for RL algorithms [27] and for adjusting weights of different objectives in the reward function [28]. In [29], hyper-parameters of the RL algorithm and the network structure are jointly optimized using the Genetic algorithm in which each individual is a DRL agent.…”
Section: B Automated Reinforcement Learning (Autorl)mentioning
confidence: 99%
“…One popular way of automating algorithm selection is modeling the problem as MAB problem and assigning an action to each algorithm [26]. BO, which works well for automated machine learning, has been used for hyperparameters tuning for RL algorithms [27] and for adjusting weights of different objectives in the reward function [28]. In [29], hyper-parameters of the RL algorithm and the network structure are jointly optimized using the Genetic algorithm in which each individual is a DRL agent.…”
Section: B Automated Reinforcement Learning (Autorl)mentioning
confidence: 99%
“…The algorithms focus on establishing a method to define the relationship between a set of variables (which represent the characteristics) and a continuous target variable. Examples of such algorithms being applied in self-driving systems include Bayesian regression [71], neural network regression [72] and decision forest regression [73].…”
Section: Regressionmentioning
confidence: 99%
“…aprendizado (0 < α ≤ 1) regula a velocidade em que as novas informações sobrepõem-se sobre o aprendizado já armazenado na matriz Q. Já o fator de desconto tem o papel de controlar a influência das recompensas futuras: se γ = 0, o reforço imediato tem grande influência; se 0 < γ ≤ 1, as recompensas futuras são descontadas; se γ = 1, as recompensas não são descontadas. Nesse aspecto, um desafio do ARé configurar os parâmetros de aprendizado de forma a maximizar o desempenho no domínio em estudo, pois , α e γ podem assumir diferentes combinações de valores (Schweighofer and Doya, 2003;Even-Dar and Mansour, 2003;Barsce et al, 2017;Ottoni et al, 2018).…”
Section: Aprendizado Por Reforçounclassified
“…De fato, um dos principais aspectos do ML e também do ARé a estimação de parâmetros, como taxa de aprendi-zado, fator de desconto, − greedy e função de reforço (Schweighofer and Doya, 2003;Even-Dar and Mansour, 2003;Barsce et al, 2017;Ottoni et al, 2018;Liessner et al, 2019;Hutter et al, 2019). Nesse aspecto, o desafioé propor métodos para a otimização e recomendação de parâmetros, de modo a otimizar o desempenho do aprendizado (Hutter et al, 2019).…”
Section: Introductionunclassified