2018
DOI: 10.1007/s10489-018-1331-y
|View full text |Cite
|
Sign up to set email alerts
|

Inverse discounted-based LQR algorithm for learning human movement behaviors

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 17 publications
(10 citation statements)
references
References 33 publications
0
10
0
Order By: Relevance
“…The IOC/IRL [82][83][84] problems seek the weights of the utility function n x; u ð Þ that are responsible for obtaining an optimal or expert decision making function. These methods are considered as a striatum cognitive models because they take cognitive models of both the neocortex and hippocampus for complementary learning and experience transference.…”
Section: Inverse Optimal Control/inverse Reinforcement Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…The IOC/IRL [82][83][84] problems seek the weights of the utility function n x; u ð Þ that are responsible for obtaining an optimal or expert decision making function. These methods are considered as a striatum cognitive models because they take cognitive models of both the neocortex and hippocampus for complementary learning and experience transference.…”
Section: Inverse Optimal Control/inverse Reinforcement Learningmentioning
confidence: 99%
“…However, there exists multiple solutions that derive the same performance and we require to add some constraints. The most common approaches to solve optimization problems under constraints use a convex optimization algorithm [84] based on linear or quadratic programming methods [83,50] to obtain the optimal variable or variables of interest.…”
Section: Optimization Techniquesmentioning
confidence: 99%
“…with f : R + × R n × R p h → R n and where x ∈ R n and u (h) ∈ R p h are the system states and the inputs of the human, respectively. A quadratic cost function for modelling human actions (see [23], [24])…”
Section: A Modelling Shared Controllers By Differential Gamesmentioning
confidence: 99%
“…Among them, p−q represents the Euclidean distance between the pixel point p and the pixel point q, and their size determines the actual application effect of the bilateral filter. From the aforementioned formulas of Gaussian filter and bilateral filter, it can be seen that the weight coefficient of the Gaussian filter depends on the spatial distance, which is a fixed value in the filter window [31,32], whereas the weight coefficient of the bilateral filter depends on the space difference and pixel difference. e final filtering effect is determined by both, and hence, the weight coefficient in the filtering window is not a fixed value.…”
Section: Bilateral Filteringmentioning
confidence: 99%