2019 International Conference on Robotics and Automation (ICRA) 2019
DOI: 10.1109/icra.2019.8793696
|View full text |Cite
|
Sign up to set email alerts
|

Weakly Supervised Recognition of Surgical Gestures

Abstract: Kinematic trajectories recorded from surgical robots contain information about surgical gestures and potentially encode cues about surgeon's skill levels. Automatic segmentation of these trajectories into meaningful action units could help to develop new metrics for surgical skill assessment as well as to simplify surgical automation. State-of-the-art methods for action recognition relied on manual labelling of large datasets, which is time consuming and error prone. Unsupervised methods have been developed to… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
15
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
8
1

Relationship

2
7

Authors

Journals

citations
Cited by 28 publications
(15 citation statements)
references
References 33 publications
0
15
0
Order By: Relevance
“…While unsupervised approaches mitigate significantly labelling costs and issues, as labels are only necessary for model testing, they still show a significant gap in performance with respect to the supervised methods. A midway solution is represented by semi-supervised learning [73,74,75,76], where a small pool of annotated demonstrations is used for model training. Such demonstrations can be used to transfer gesture labels to a large set of unlabelled observations previously aligned with DTW [73], or to initialize clustering of unlabelled data and avoid tedious parameter tuning [74].…”
Section: Semi-supervised Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…While unsupervised approaches mitigate significantly labelling costs and issues, as labels are only necessary for model testing, they still show a significant gap in performance with respect to the supervised methods. A midway solution is represented by semi-supervised learning [73,74,75,76], where a small pool of annotated demonstrations is used for model training. Such demonstrations can be used to transfer gesture labels to a large set of unlabelled observations previously aligned with DTW [73], or to initialize clustering of unlabelled data and avoid tedious parameter tuning [74].…”
Section: Semi-supervised Learningmentioning
confidence: 99%
“…A midway solution is represented by semi-supervised learning [73,74,75,76], where a small pool of annotated demonstrations is used for model training. Such demonstrations can be used to transfer gesture labels to a large set of unlabelled observations previously aligned with DTW [73], or to initialize clustering of unlabelled data and avoid tedious parameter tuning [74]. These systems are however highly sensitive to large data variability, showing rapid drop in recognition accuracy with increased heterogeneity in terms of gesture ordering or surgical skill level respectively.…”
Section: Semi-supervised Learningmentioning
confidence: 99%
“…The JIG-SAWS dataset has extended annotations at a sub-task level. AI techniques learn patterns and temporal interconnections of the sub-task sequences from combinations of robot kinematics and surgical video and detect and temporally localize each sub-task [19][20][21][22][23][24]. Recently, AI models for activity recognition have been developed and tested on annotated datasets from real cases of robotic-assisted radical prostatectomy and ocular microsurgery [18][19][20].…”
Section: Surgical Phase Recognitionmentioning
confidence: 99%
“…Zappella et al proposed methods of modeling surgical video clips for single-action classification [16]. The Transition State Clustering (TSC) and Gaussian Mixture Model methods provide unsupervised or weakly-supervised methods for surgical trajectory segmentation [17,18]. More recently, deep learning methods have come to define the state-of-the-art, such as Temporal Convolutional Networks (TCN) [19], Time Delay Neural Network (TDNN) [7], and Long-Short Term Memory (LSTM) [6,20].…”
Section: Introductionmentioning
confidence: 99%