Proceedings of the 2016 on Genetic and Evolutionary Computation Conference Companion 2016
DOI: 10.1145/2908961.2930952
|View full text |Cite
|
Sign up to set email alerts
|

Model-Based Relative Entropy Stochastic Search

Abstract: Stochastic search algorithms are general black-box optimizers. Due to their ease of use and their generality, they have recently also gained a lot of attention in operations research, machine learning and policy search. Yet, these algorithms require a lot of evaluations of the objective, scale poorly with the problem dimension, are affected by highly noisy objective functions and may converge prematurely. To alleviate these problems, we introduce a new surrogate-based stochastic search approach. We learn simpl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
88
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
3
2
2

Relationship

3
4

Authors

Journals

citations
Cited by 44 publications
(89 citation statements)
references
References 11 publications
1
88
0
Order By: Relevance
“…Both layers are connected by regression problems that are used to estimate the corresponding reward models. While we implemented learning algorithms that are based on the Model-Based Relative Entropy Stochastic Search (MORE) [14] algorithm for the individual layers, the whole framework is more general and other algorithms could be used that allow for the use of sample reweighting with importance weights.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Both layers are connected by regression problems that are used to estimate the corresponding reward models. While we implemented learning algorithms that are based on the Model-Based Relative Entropy Stochastic Search (MORE) [14] algorithm for the individual layers, the whole framework is more general and other algorithms could be used that allow for the use of sample reweighting with importance weights.…”
Section: Discussionmentioning
confidence: 99%
“…The sub-policy optimization is an extension of ModelBased Relative Entropy Stochastic Search (MORE) [14]. Both algorithms consist of the two steps: first estimating a reward model and second updating the (sub-)policy.…”
Section: B Learning On the Sub-policy Layermentioning
confidence: 99%
“…The use of trust regions is a common approach in policy search to obtain a stable policy update and to avoid premature convergence [1,17,18]. The model-based relative entropy stochastic search algorithm (MORE) [1] uses a local quadratic surrogate to update it's Gaussian search distribution.…”
Section: Related Workmentioning
confidence: 99%
“…The model-based relative entropy stochastic search algorithm (MORE) [1] uses a local quadratic surrogate to update it's Gaussian search distribution. As these surrogates can be inaccurate, MORE doesn't exploit the surrogate greedily, but uses a trust region in form of a KL-bound.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation