Proceedings of the 23rd International Conference on Machine Learning - ICML '06 2006
DOI: 10.1145/1143844.1143872
|View full text |Cite
|
Sign up to set email alerts
|

Dealing with non-stationary environments using context detection

Abstract: In this paper we introduce RL-CD, a method for solving reinforcement learning problems in non-stationary environments. The method is based on a mechanism for creating, updating and selecting one among several partial models of the environment. The partial models are incrementally built according to the system's capability of making predictions regarding a given sequence of observations. We propose, formalize and show the efficiency of this method both in a simple non-stationary environment and in a noisy scena… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
84
0
1

Year Published

2008
2008
2023
2023

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 112 publications
(85 citation statements)
references
References 7 publications
0
84
0
1
Order By: Relevance
“…Choi et al developed a method called Multiple-Model Reinforcement Learning (MMRL) [8] which assumes a fixed number of dynamics and known by the designer. This assumption is not practical and unpretentious for real world problems.…”
Section: A Reinforcement Learningmentioning
confidence: 99%
See 2 more Smart Citations
“…Choi et al developed a method called Multiple-Model Reinforcement Learning (MMRL) [8] which assumes a fixed number of dynamics and known by the designer. This assumption is not practical and unpretentious for real world problems.…”
Section: A Reinforcement Learningmentioning
confidence: 99%
“…The agent uses the model with the highest quality, E max . When there is no model that has a higher E than a pre-specified minimum match quality threshold, i.e., E min is less than the minimum quality threshold that means a new model must be created by the agent [8].…”
Section: F Reinforcement Learning With Context Detectionmentioning
confidence: 99%
See 1 more Smart Citation
“…In [8] a set of techniques were tried in order to improve the learning ability of the agents in a simple scenario. Performance of reinforcement learning approaches such as Q-learning and Prioritized Sweeping in non-stationary environments are compared in [13]. Co-learning is discussed in [19] (detailed here in Section 2.3).…”
Section: Real-time Optimization Of Traffic Lightsmentioning
confidence: 99%
“…We have expected Q-learning to perform bad because it is already known that it does not have a good performance in noisy and non-stationary traffic scenarios [13]. In order to test this, we have implemented a Q-learning mechanism in the traffic lights.…”
Section: Q-learning Traffic Lightsmentioning
confidence: 99%