2009
DOI: 10.3724/sp.j.1004.2009.00682
|View full text |Cite
|
Sign up to set email alerts
|

Data-based Optimal Control for Discrete-time Zero-sum Games of 2-D Systems Using Adaptive Critic Designs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
10
0

Year Published

2011
2011
2016
2016

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 20 publications
(10 citation statements)
references
References 25 publications
0
10
0
Order By: Relevance
“…Hence, some studies have attempted to solve the optimal control solution based on ADP technique without an a priori system model [15]- [18]. For linear discrete-time systems, Q-learning was introduced to relax some of the exact model-matching restrictions in [15] and [16], which allows model-free tuning of the action and critic networks. For linear continuous-time systems, Vrabie et al proposed a new formulation of the proportional algorithm which converges to the optimal control solution without using internal dynamics of the system in [17].…”
Section: Introductionmentioning
confidence: 99%
“…Hence, some studies have attempted to solve the optimal control solution based on ADP technique without an a priori system model [15]- [18]. For linear discrete-time systems, Q-learning was introduced to relax some of the exact model-matching restrictions in [15] and [16], which allows model-free tuning of the action and critic networks. For linear continuous-time systems, Vrabie et al proposed a new formulation of the proportional algorithm which converges to the optimal control solution without using internal dynamics of the system in [17].…”
Section: Introductionmentioning
confidence: 99%
“…Game theory [22][23][24][25][26] has been applied in modeling strategic behavior, where the performance of each player depends on the action of himself and all the other players. In practice, a large class of practical systems such as wireless communication systems is controlled by more players (controllers) with every player trying to optimize individual performance objective.…”
Section: Introductionmentioning
confidence: 99%
“…Since proposed by Werbos [1], ACD has been used to solve optimal control problem forward in time in many works, such as [2], [4], [5], [7], [6], [17], [18], [27], [15]. In recent years, ACD has also been utilized to solve for zero-sum games [21], [16], [20]. Some recent surveys in [10] and [19] on ACD techniques present excellent overview of the state-of-the-art developments.…”
Section: Introductionmentioning
confidence: 99%
“…In [3], the optimal strategies for discrete-time linear system quadratic zero-sum games without knowing the system dynamical matrices by using Q-learning method was proposed. In [16], an iterative ADP algorithm was proposed to solve a class of discrete-time two-person zero-sum for Roesser type 2-D system based on Q-learning method. It is important to note that the controller designed based on Q-learning method is still a state feedback controller which demand the state of the system to be measurable.…”
Section: Introductionmentioning
confidence: 99%