2023
DOI: 10.1007/s10489-023-04949-y
|View full text |Cite
|
Sign up to set email alerts
|

Joint ordinal regression and multiclass classification for diabetic retinopathy grading with transformers and CNNs fusion network

Lei Ma,
Qihang Xu,
Hanyu Hong
et al.
Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 58 publications
0
3
0
Order By: Relevance
“…In the domain of diabetic retinopathy classification, Karthika and Durgadevi (2024) and Bala et al (2023) employed a fusion of deep learning and attention mechanisms, yielding commendable results. Ma et al (2023) approached diabetic retinopathy classification as a dual challenge, addressing both ordered regression and multi-class classification. Leveraging the transformer's feature extraction capabilities, the integration of category and ordinal supervision information led to enhanced classification accuracy and kappa values.…”
Section: Relate Workmentioning
confidence: 99%
“…In the domain of diabetic retinopathy classification, Karthika and Durgadevi (2024) and Bala et al (2023) employed a fusion of deep learning and attention mechanisms, yielding commendable results. Ma et al (2023) approached diabetic retinopathy classification as a dual challenge, addressing both ordered regression and multi-class classification. Leveraging the transformer's feature extraction capabilities, the integration of category and ordinal supervision information led to enhanced classification accuracy and kappa values.…”
Section: Relate Workmentioning
confidence: 99%
“…The hybrid structure also utilizes images with a resolution of 256×256 as input, achieving an accuracy increase to 94%. In addition, the combination of ViT with CNN was also proposed and applied [ 29 , 30 ]. Sadeghzadeh et al [ 29 ] fused EfficientNet-B0 with Transformer and achieved state-of-the-art classification results using 224×224 images as input on EyePACS, APTOS, DDR, Messidor-1, and Messidor-2 datasets.…”
Section: Related Workmentioning
confidence: 99%
“…Sadeghzadeh et al [ 29 ] fused EfficientNet-B0 with Transformer and achieved state-of-the-art classification results using 224×224 images as input on EyePACS, APTOS, DDR, Messidor-1, and Messidor-2 datasets. Ma et al [ 30 ] proposed a fusion network based on Transformer and CNN for the grading of DR, where DR grading was treated as a joint ordinal regression and multi-classification problem. They utilized images with a slightly higher resolution of 384×384 and demonstrated superior performance on the DeepDR and IDRiD datasets.…”
Section: Related Workmentioning
confidence: 99%