2018
DOI: 10.1007/978-3-030-00937-3_31
|View full text |Cite
|
Sign up to set email alerts
|

DeepPhase: Surgical Phase Recognition in CATARACTS Videos

Abstract: Automated surgical workflow analysis and understanding can assist surgeons to standardize procedures and enhance post-surgical assessment and indexing, as well as, interventional monitoring. Computerassisted interventional (CAI) systems based on video can perform workflow estimation through surgical instruments' recognition while linking them to an ontology of procedural phases. In this work, we adopt a deep learning paradigm to detect surgical instruments in cataract surgery videos which in turn feed a surgic… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
83
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
7
2

Relationship

1
8

Authors

Journals

citations
Cited by 108 publications
(84 citation statements)
references
References 15 publications
1
83
0
Order By: Relevance
“…The multi-task learning of tool and phase recognition requires the simultaneous annotations for both tasks on the same dataset, which restricts the development to a certain extent. Fortunately, most works regard it as a worthy trade-off: some label the two tasks with utilizing binary tool usage to address phase recognition task (Padoy et al (2012); Yu et al (2019)); others are dedicated to establishing more advanced multi-task strategies (Zisimopoulos et al (2018); Nakawala et al (2019)). In addition, more relevant datasets begin to be released for public usage, which alleviate the annotation problem to a great extent (Nakawala et al (2019); Stefanie et al (2018)).…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…The multi-task learning of tool and phase recognition requires the simultaneous annotations for both tasks on the same dataset, which restricts the development to a certain extent. Fortunately, most works regard it as a worthy trade-off: some label the two tasks with utilizing binary tool usage to address phase recognition task (Padoy et al (2012); Yu et al (2019)); others are dedicated to establishing more advanced multi-task strategies (Zisimopoulos et al (2018); Nakawala et al (2019)). In addition, more relevant datasets begin to be released for public usage, which alleviate the annotation problem to a great extent (Nakawala et al (2019); Stefanie et al (2018)).…”
Section: Discussionmentioning
confidence: 99%
“…Although this work has achieved outstanding performance, temporal dependencies, which are crucial for phase analysis, are detached from the unified framework. Zisimopoulos et al (2018) proposed to first train a ResNet to recognize tool presence and then combine the tool binary predictions and tool features from the last layer to train a RNN for phase recognition, which achieved promising results in cataract video analysis. Very recently, Nakawala et al (2019) present a Deep-Onto network which integrates deep models with ontology and production rules.…”
Section: Multi-task Learningmentioning
confidence: 99%
“…To extract APMs from video, a number of building block algorithmic capabilities are important to develop, for example: detection of surgical instrument presence [50], delineation of surgical tools' position and motion [51], segmentation of surgical site into objects [52] or the video into key surgical steps [53,54], activity or significant event detection [55] as well as others like the detection of critical structures [56]. Marked progress in each of these building blocks of surgical process understanding has taken place in recent years, but a significant challenge is still the availability of large, well annotated datasets that can be used to evaluate systems in a fair and comparable manner.…”
Section: Ai and Machine Learning (Ml)mentioning
confidence: 99%
“…The associate editor coordinating the review of this manuscript and approving it for publication was Ting Li . scheduling, and offline video indexing for educational purposes [7]. Hence, in this study, we focus on real-time surgical tool detection in videos.…”
Section: Introductionmentioning
confidence: 99%