2019
DOI: 10.1109/lra.2019.2891311
|View full text |Cite
|
Sign up to set email alerts
|

Robot-Assisted Training in Laparoscopy Using Deep Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
14
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 45 publications
(14 citation statements)
references
References 14 publications
0
14
0
Order By: Relevance
“…whereḟ ,f , ... f , f (4) and f (5) denote the first-order, secondorder, third-order, fourth-order and fifth-order derivatives, respectively; ξ 1 , ξ 2 , ξ 3 , ξ 4 , ξ 5 , ξ 6 and ξ 7 lie in the interval (kτ, (k + 1)τ ), ((k − 1)τ, kτ ), ((k − 2)τ, kτ ), ((k − 3)τ, kτ ), ((k −4)τ, kτ ), ((k −5)τ, kτ ), ((k −6)τ, kτ ) and ((k −7)τ, kτ ), correspondingly; symbol ! denotes the factorial operator.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…whereḟ ,f , ... f , f (4) and f (5) denote the first-order, secondorder, third-order, fourth-order and fifth-order derivatives, respectively; ξ 1 , ξ 2 , ξ 3 , ξ 4 , ξ 5 , ξ 6 and ξ 7 lie in the interval (kτ, (k + 1)τ ), ((k − 1)τ, kτ ), ((k − 2)τ, kτ ), ((k − 3)τ, kτ ), ((k −4)τ, kτ ), ((k −5)τ, kτ ), ((k −6)τ, kτ ) and ((k −7)τ, kτ ), correspondingly; symbol ! denotes the factorial operator.…”
Section: Discussionmentioning
confidence: 99%
“…Moreover, to eliminate the termsf (kτ ), ... f (kτ ) and f (4) Then, according to Routh stability criterion [33], [34], 7is finally obtained. The proof is thus completed.…”
Section: Discussionmentioning
confidence: 99%
“…At present, there are mainly two research directions, methods of value-based and policy-based. The representative algorithm based on value function is Q-learning in which the selection policy of action is greedy [ 54 ]. The algorithm updates the action-value function in accordance with the following formula: …”
Section: Deep Reinforcement Learningmentioning
confidence: 99%
“…In [50], the Restricted Boltzmann machine with the hidden variables has been adopted to approximate the action and state of the RL, and the hidden variables can be estimated by state-action-reward-state-action (SARSA) algorithm. In [51], the HMM is used with RL to predict the zone in the 5G system. Similar work can be found in [52].…”
Section: Related Workmentioning
confidence: 99%