2019
DOI: 10.1007/978-3-030-32251-9_57
|View full text |Cite
|
Sign up to set email alerts
|

Overcoming Data Limitation in Medical Visual Question Answering

Abstract: Traditional approaches for Visual Question Answering (VQA) require large amount of labeled data for training. Unfortunately, such large scale data is usually not available for medical domain. In this paper, we propose a novel medical VQA framework that overcomes the labeled data limitation. The proposed framework explores the use of the unsupervised Denoising Auto-Encoder (DAE) and the supervised Meta-Learning. The advantage of DAE is to leverage the large amount of unlabeled images while the advantage of Meta… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
66
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 107 publications
(66 citation statements)
references
References 13 publications
0
66
0
Order By: Relevance
“…Slow progress in the medical field Medical image processing [54] Drug discovery [56] Cancer detection [57] Medical vision question answering [58] Skin lesion segmentation tasks [60] Predicting the specific behavior of molecules Ability to learn to weigh support samples Computational Intelligence and Neuroscience (DDA) method proposed by Sun and Saenko [67] and Rozantsev et al [68] and the hybrid heterogeneous transfer learning (HHTL) algorithm proposed by Zhou et al [69] based on deep learning.…”
Section: Intelligent Medicinementioning
confidence: 99%
See 2 more Smart Citations
“…Slow progress in the medical field Medical image processing [54] Drug discovery [56] Cancer detection [57] Medical vision question answering [58] Skin lesion segmentation tasks [60] Predicting the specific behavior of molecules Ability to learn to weigh support samples Computational Intelligence and Neuroscience (DDA) method proposed by Sun and Saenko [67] and Rozantsev et al [68] and the hybrid heterogeneous transfer learning (HHTL) algorithm proposed by Zhou et al [69] based on deep learning.…”
Section: Intelligent Medicinementioning
confidence: 99%
“…In [ 57 ], MAML can be adapted to weakly supervised breast cancer detection tasks, and the order of the tasks is selected according to the course, rather than randomly selected. MAML is also combined with a denoising autoencoder for medical vision question answering [ 58 ], and learning to weigh support samples as done in [ 59 ] can be applied to pixel weighting to handle skin lesion segmentation tasks with noisy labels [ 60 ].…”
Section: Specific Analysis Of the Application Of Metalearning In The Field Of Artificial Intelligencementioning
confidence: 99%
See 1 more Smart Citation
“…In [275] MAML is adapted to weaklysupervised breast cancer detection tasks, and the order of tasks are selected according to a curriculum. MAML is also combined with denoising autoencoders to do medical visual question answering [276], while learning to weigh support samples [218] is adapted to pixel wise weighting for skin lesion segmentation tasks that have noisy labels [277].…”
Section: Emerging Topicsmentioning
confidence: 99%
“…This technology effectively improves the training style and training time, and has strong adaptability and robustness to the unknown scenarios. For example, Nguyen et al [34] used a combination of unsupervised denoising auto encoder (DAE) and supervised meta-learning in order to overcome the data limitation in medical visual Q&A with only a small sample set, and the performance of this method is better than the existing medical visual question answering method; Wang et al [35] proposed a metalearning face recognition (meta face recognition, MFR) method, and the experimental results proved that the method is optimal in cross-racial and cross-scene tests, which greatly improves the generalization performance of the model. Meta-learning also has broad application prospects in the field of robotics learning [36].…”
Section: A Related Work In Meta-learningmentioning
confidence: 99%