2024
DOI: 10.1038/s41746-024-01101-z
|View full text |Cite
|
Sign up to set email alerts
|

FFA-GPT: an automated pipeline for fundus fluorescein angiography interpretation and question-answer

Xiaolan Chen,
Weiyi Zhang,
Pusheng Xu
et al.

Abstract: Fundus fluorescein angiography (FFA) is a crucial diagnostic tool for chorioretinal diseases, but its interpretation requires significant expertise and time. Prior studies have used Artificial Intelligence (AI)-based systems to assist FFA interpretation, but these systems lack user interaction and comprehensive evaluation by ophthalmologists. Here, we used large language models (LLMs) to develop an automated interpretation pipeline for both report generation and medical question-answering (QA) for FFA images. … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
3
2

Relationship

2
3

Authors

Journals

citations
Cited by 10 publications
(4 citation statements)
references
References 40 publications
0
4
0
Order By: Relevance
“… 18 These models can be directly integrated into the cross-modal model to overcome the limitations in model output and provide further interactive explanations. 19 , 20 Additionally, they can serve as external sources of knowledge and semantic refiners to optimize the fine-tuning process of cross-modal models. This enables the models to accurately capture details and semantic information in images, facilitating generalization to unseen objects or concepts and performing more complex downstream tasks.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“… 18 These models can be directly integrated into the cross-modal model to overcome the limitations in model output and provide further interactive explanations. 19 , 20 Additionally, they can serve as external sources of knowledge and semantic refiners to optimize the fine-tuning process of cross-modal models. This enables the models to accurately capture details and semantic information in images, facilitating generalization to unseen objects or concepts and performing more complex downstream tasks.…”
Section: Discussionmentioning
confidence: 99%
“…FFA images were retrospectively collected from a tertiary center between 2016 and 2019. 20 After data cleaning by removing duplicates, handling missing values, and correcting erroneous data, a total of 654,343 FFA images from 9,392 patients were used for model development. All patient data were anonymized and de-identified, as well as stored in a secure data center, managed and analyzed by authorized personnel.…”
Section: Methodsmentioning
confidence: 99%
“…[24] Another way to improve the model is to align the ophthalmic image features with medical texts. [11 25] Additionally, researchers may improve their models by applying transfer learning, fine-tuning, and reinforcement learning techniques, or by incorporating knowledge enrichment methods like Retrieval-Augmented Generation (RAG). [26 27] However, before applying any model to clinical practice, detailed clinical trials should be conducted to assess its safety and efficacy.…”
Section: Discussionmentioning
confidence: 99%
“…[10] However, ophthalmology was not included in the study, leaving the practical application capabilities of GPT-4V in addressing image-related concerns in ophthalmology uncertain. Meanwhile, previous work on ophthalmic VQA focuses on a specific modality,[11 12] leaving multimodal ocular VQA unexplored.…”
Section: Introductionmentioning
confidence: 99%