1998
DOI: 10.1109/5326.704563
|View full text |Cite
|
Sign up to set email alerts
|

Fuzzy inference system learning by reinforcement methods

Abstract: Fuzzy Actor-Critic Learning (FACL) and Fuzzy Q-Learning (FQL) are reinforcement learning methods based on Dynamic Programming (DP) principles. In this paper, they are used to tune online the conclusion part of Fuzzy Inference Systems (FIS). The only information available for learning is the system feedback, which describes in terms of reward and punishment the task the fuzzy agent has to realize. At each time step, the agent receives a reinforcement signal according to the last action it has performed in the p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
216
0
7

Year Published

2002
2002
2023
2023

Publication Types

Select...
7
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 327 publications
(224 citation statements)
references
References 39 publications
1
216
0
7
Order By: Relevance
“…This section presents the FQL algorithm as given in [25]. Let the state vector s t = s t 1 , ...., s t j , ....., s t J , where j is the j th element of state vector before fuzzification.…”
Section: Fql Algorithm Descriptionmentioning
confidence: 99%
“…This section presents the FQL algorithm as given in [25]. Let the state vector s t = s t 1 , ...., s t j , ....., s t J , where j is the j th element of state vector before fuzzification.…”
Section: Fql Algorithm Descriptionmentioning
confidence: 99%
“…Fuzzy approximators have typically been used in modelfree (RL) techniques such as Q-learning [13,15,17] and actor-critic algorithms [2,20]. Most of these approaches are heuristic in nature, and their theoretical properties have not been investigated yet.…”
Section: Related Workmentioning
confidence: 99%
“…The first term in the sum depends linearly on ε ′ Q , which is related to the accuracy of the fuzzy approximator, and is more difficult to control. This ε ′ Q -dependent term also contributes to the suboptimality of the asymptotic solutions (17), (19). Ideally, one can find ε ′ Q = min Q∈ Q Q * − Q ∞ , which provides the smallest upper bounds in (17)-(20).…”
Section: Theorem 3 (Near-optimality) Denote the Set Of Qfunctions Repmentioning
confidence: 99%
“…2). The Fuzzy net has an RBF-like architecture and powerful ability to classify continuous input and give continuous output (Jouffe, 1998). It has been successfully adopted in cart-pole balance control (Wang, X. S., et al, 2007), adaptive behavior learning of autonomous robots (Samejima and Omori, 1999), (Perez- Fig.1.…”
Section: Fuzzy Netmentioning
confidence: 99%