2023
DOI: 10.1109/tpami.2022.3155612
|View full text |Cite
|
Sign up to set email alerts
|

Real-Time Scene Text Detection With Differentiable Binarization and Adaptive Scale Fusion

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
86
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 208 publications
(87 citation statements)
references
References 61 publications
1
86
0
Order By: Relevance
“…On the ICDAR2015 dataset, compared to the base model, ADFM improved P by 1.4% to 90.2%, R by 6.4% and F by 4.4%, and performed better than most of the more advanced algorithms in recent years. Compared with the test results of DBNet trained with ResNet-18, the P improved by 3.4%, R by 0.3% and F by 1.8%; compared with the test results of the more recent algorithm DBNet++ [ 36 ] trained on the ResNet-18 backbone, the P is improved by 0.1%, the R is improved by 1.5% and the F is improved by 1%. The results of the comparison with other methods are shown in Table 2 .…”
Section: Methodsmentioning
confidence: 97%
“…On the ICDAR2015 dataset, compared to the base model, ADFM improved P by 1.4% to 90.2%, R by 6.4% and F by 4.4%, and performed better than most of the more advanced algorithms in recent years. Compared with the test results of DBNet trained with ResNet-18, the P improved by 3.4%, R by 0.3% and F by 1.8%; compared with the test results of the more recent algorithm DBNet++ [ 36 ] trained on the ResNet-18 backbone, the P is improved by 0.1%, the R is improved by 1.5% and the F is improved by 1%. The results of the comparison with other methods are shown in Table 2 .…”
Section: Methodsmentioning
confidence: 97%
“…Table 4 shows the quantitative comparison on the HSText-1000 dataset, we compare our method with some state-of-the-art methods. 1,2,17,20,[33][34][35] Through the experimental comparison, we can find that our method can achieve the best detection performance on the HSText-1000 dataset, which is 10.3% higher than CTPN, 1 43.5% higher than EAST, 2 25.4% higher than CRAFT, 33 31.3% higher than shape robust text detection with progressive scale expansion network (PSE-Net), 34 10.6% higher than ContourNet, 17 3.7% higher than Pan++, 35 and 1.6% higher than DBNet++. 20 This proves that the method proposed in this paper is more competitive than other recent methods for text detection in haze scenes.…”
Section: Comparison With Other Methodsmentioning
confidence: 99%
“…Other detection methods are usually based on image segmentation, for example, CRAFT [13] proposed detected text area by exploring each character and the affinity between characters., Liao et al. [14] proposed a Differentiable Binarization (DB) module that integrates the binarization process and an efficient Adaptive Scale Fusion (ASF) module to improve the scale robustness by fusing features of different scales adaptively. Text recognition task includes Connectionist Temporal Classification (CTC) based methods [15] and encoder‐decoder methods [16].…”
Section: Related Workmentioning
confidence: 99%
“…In the detection task of the deep learning era, some methods are focused on the bounding boxes: Deng et al [11] detected multi-oriented text with corner-based region proposals by the precise setting of the anchor, and Zhong et al [12] proposed a novel progressive region prediction network (PRPN) with directional pooling for predicting text regions. Other detection methods are usually based on image segmentation, for example, CRAFT [13] proposed detected text area by exploring each character and the affinity between characters., Liao et al [14] proposed a Differentiable Binarization (DB) module that integrates the binarization process and an efficient Adaptive Scale Fusion (ASF) module to improve the scale robustness by fusing features of different scales adaptively. Text recognition task includes Connectionist Temporal Classification (CTC) based methods [15] and encoder-decoder methods [16].…”
Section: Text Detection and Recognitionmentioning
confidence: 99%