2022
DOI: 10.1007/978-3-031-16443-9_61
|View full text |Cite
|
Sign up to set email alerts
|

Discrepancy and Gradient-Guided Multi-modal Knowledge Distillation for Pathological Glioma Grading

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 16 publications
(5 citation statements)
references
References 24 publications
0
5
0
Order By: Relevance
“…Hu et al [53] proposed to use knowledge distillation to transfer knowledge from a trained multi-modal network to a mono-modal one for medical image segmentation. Xing et al [54] proposed a discrepancy and gradient-guided knowledge distillation framework to transfer privileged knowledge from a multi-modal teacher network to a student network for pathological glioma grading. Yang et al [55] proposed an affinity-guided dense tumor-region knowledge distillation mechanism to align features for brain tumor segmentation with missing modalities.…”
Section: Knowledge Distillation On Medical Image Analysismentioning
confidence: 99%
“…Hu et al [53] proposed to use knowledge distillation to transfer knowledge from a trained multi-modal network to a mono-modal one for medical image segmentation. Xing et al [54] proposed a discrepancy and gradient-guided knowledge distillation framework to transfer privileged knowledge from a multi-modal teacher network to a student network for pathological glioma grading. Yang et al [55] proposed an affinity-guided dense tumor-region knowledge distillation mechanism to align features for brain tumor segmentation with missing modalities.…”
Section: Knowledge Distillation On Medical Image Analysismentioning
confidence: 99%
“…Transfer knowledge for missing modality robustness Recent studies (Zhou et al 2021b;Shen and Gao 2019;Wang et al 2021;Azad, Khosravi, and Merhof 2021) have introduced Knowledge Distillation (Gou et al 2021;Xing et al 2022) or co-training (Nigam and Ghani 2000) for improving missing-modality robustness. Despite the computational and memory cost of additional networks in their method, the choice of similarity measurements significantly impacts the final performance (Azad, Khosravi, and Merhof 2021).…”
Section: Related Workmentioning
confidence: 99%
“…(2) reconstruction-based modal-completion; (3) missingmodality robust architecture. For distillation or co-training approaches (Xing et al 2022;Wang et al 2021;Poklukar et al 2022), Although they marginally improve the missingmodality robustness, their methods are highly dependent on the choice of similarity measurements and require modalincomplete input to be filled with masking values (Wang et al 2021;Tsai et al 2019a), which may cause unexpected behavior in the models, leading to degraded performance (Shen and Gao 2019). Reconstruction-based methods (Khat-…”
Section: Introductionmentioning
confidence: 99%
“…Although coordinated representations have traditionally tended to be more challenging to implement, the convenience of neural network architectural and loss adjustments have resulted in increased traction in publications embodying coordinated representations (Xing et al, 2022;Chauhan et al, 2020;Radford et al, 2021;Bhalodia et al, 2021). One of the most notable of these in recent AI approaches is OpenAI's Contrastive Language-Image Pre-Training (CLIP) model, which forms representations for OpenAI's DALL•E 2 (Radford et al, 2021;Ramesh et al, 2022) and uses a contrastive-learning approach to shape both image embeddings of entire images to match text embeddings of entire captions describing those images.…”
Section: Tablementioning
confidence: 99%
“…In privileged models based on traditional approaches (before deep neural networks), privileged information can be embedded in the model either through an alteration of allowable error ("slack variables" from SVM+) (Vapnik and Vashist, 2009), or through decision trees constructed with non-privileged features to mimic the discriminative ability of privileged features (Random Forest+) (Warner et al, 2022;Moradi et al, 2016). In a deep learning model, privileged learning is often achieved through the use of additional loss functions which attempt to constrain latent and output vectors from the non-privileged modality to mimic those from the combined privileged and non-privileged models Xing et al, 2022). For example, in Chauhan et al (2020), encoders for each modality are compared and cross entropy loss is calculated for each modality separately.…”
Section: Privileged Learningmentioning
confidence: 99%