2018
DOI: 10.1101/316349
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Multilabel multiclass classification of OCT images augmented with age, gender and visual acuity data

Abstract: Optical Coherence Tomography (OCT) imaging of the retina is in widespread clinical use to diagnose a wide range of retinal pathologies and several previous studies have used deep learning to create systems that can accurately classify retinal OCT as indicative of one of these pathologies. However, patients often exhibit multiple pathologies concurrently. Here, we designed a novel neural network algorithm that performs multiclass and multilabel classification of retinal images from OCT images in four common ret… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
1
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(9 citation statements)
references
References 8 publications
0
9
0
Order By: Relevance
“…Then, all weights were unfrozen and the network was retrained for 20 epochs with the same callbacks and learning rate. As suggested in the literature, 22 training from scratch on OCT images may be preferable as many of the low-level filters in networks pretrained on natural images are tuned to colors and OCT images are monochromatic. However, retraining the entire network preinitialized on ImageNet provided us with better performance than training from uninitialized weights, likely due to our significantly smaller dataset.…”
Section: Experimental Settingsmentioning
confidence: 99%
See 1 more Smart Citation
“…Then, all weights were unfrozen and the network was retrained for 20 epochs with the same callbacks and learning rate. As suggested in the literature, 22 training from scratch on OCT images may be preferable as many of the low-level filters in networks pretrained on natural images are tuned to colors and OCT images are monochromatic. However, retraining the entire network preinitialized on ImageNet provided us with better performance than training from uninitialized weights, likely due to our significantly smaller dataset.…”
Section: Experimental Settingsmentioning
confidence: 99%
“…Class activation maps (CAMs) 20 are a common method where a heat map is generated by projecting the class specific weights of the output classification layer back to the feature maps of the last convolutional layer, thereby highlighting important regions for predicting a particular class. This method has been used in ophthalmic application previously to confirm CNN decision was based off the anterior chamber angle in categorizing angle closure, 21 areas of OCT B-scans associated with various diagnoses 22 , 23 and areas of segmentation error, 24 and area of OCT enface images associated with the diagnosis of glaucoma. 25 There exists several variants of this method that build off of the original CAM paper, 20 including: Grad-Cam, 26 Guided Grad-Cam, 26 Guided Grad-Cam++, 27 and GAIN.…”
Section: Introductionmentioning
confidence: 99%
“…Marzieh Mokhtari et al [20] calculate local cup to disc ratio by fusing fundus images and OCT B-scans to get the symmetry of two eyes, which can detect early ocular diseases better. Mehta et al [19] proposed a OCT images system for multi-class, multi-label classification, which augments data by using patient information, such as age, gender and visual acuity data.…”
Section: Oct Images Recognitionmentioning
confidence: 99%
“…Since the choice for the label is from a set of multiple descriptors (classes), the term multiclass is used [10]. Oftentimes, multilabels are treated as subsets of multiclass labeling [3], [4], [5].…”
Section: Related Workmentioning
confidence: 99%
“…In the context of videos, our multilabel vector represents distinct spatial regions (i.e. regions of interest) of an image frame, unlike common uses of the multilabel vector as a representation of the presence or absence of multiple descriptors of interest [3], [4], [5]. Each region is assigned one of three possible levels of activity to train and validate our deep learning model.…”
Section: Introductionmentioning
confidence: 99%