2021
DOI: 10.48550/arxiv.2107.13741
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Self-Paced Contrastive Learning for Semi-supervised Medical Image Segmentation with Meta-labels

Abstract: Pre-training a recognition model with contrastive learning on a large dataset of unlabeled data has shown great potential to boost the performance of a downstream task, e.g., image classification. However, in domains such as medical imaging, collecting unlabeled data can be challenging and expensive. In this work, we propose to adapt contrastive learning to work with meta-label annotations, for improving the model's performance in medical image segmentation even when no additional unlabeled data is available. … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
1
1

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(7 citation statements)
references
References 46 publications
0
7
0
Order By: Relevance
“…In this approach, a network is pre-trained to bring closer the feature embeddings of an image under different transformations (positive pairs), while pushing away those from different images (negative pairs). This idea was used to learn a global representation at the end of the network's encoder, using meta-labels on anatomical similarity or subject ID to define the positive pairs (Chaitanya et al, 2020;Peng et al, 2021b;Zeng et al, 2021). To also pre-train the decoder, the method in (Chaitanya et al, 2020) defined positive or negative embedding pairs based on their spatial distance in a feature map, those with a large distance considered as negative while those at the same spatial position but coming from different transformations as positives.…”
Section: B Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…In this approach, a network is pre-trained to bring closer the feature embeddings of an image under different transformations (positive pairs), while pushing away those from different images (negative pairs). This idea was used to learn a global representation at the end of the network's encoder, using meta-labels on anatomical similarity or subject ID to define the positive pairs (Chaitanya et al, 2020;Peng et al, 2021b;Zeng et al, 2021). To also pre-train the decoder, the method in (Chaitanya et al, 2020) defined positive or negative embedding pairs based on their spatial distance in a feature map, those with a large distance considered as negative while those at the same spatial position but coming from different transformations as positives.…”
Section: B Related Workmentioning
confidence: 99%
“…We consider the 3D-MRI scans as 2D images throughplane due to the high anisotropic acquisition resolution, and re-sample them to a fix space ranging of 1.0 × 1.0 mm. Following (Peng et al, 2021b), we normalize the pixel intensities based on the 1% and 99% percentile of the intensity histogram for each scan. Normalized slices are then cropped to 384 × 384 pixels, coarsely centered based on the foreground delineation of the ground truth.…”
Section: Acdc Datasetmentioning
confidence: 99%
See 3 more Smart Citations