2022
DOI: 10.3390/electronics11132028
|View full text |Cite
|
Sign up to set email alerts
|

SV-FPN: Small Object Feature Enhancement and Variance-Guided RoI Fusion for Feature Pyramid Networks

Abstract: Small object detection is one of the research difficulties in object detection, and Feature Pyramid Networks (FPN) is a common feature extractor in deep learning; thus, improving the results of small object detection based on FPN is of great significance in this field. In this paper, SV-FPN is proposed for a small object detection task, which consists of Small Object Feature Enhancement (SOFE) and Variance-guided Region of Interest Fusion (VRoIF). When using FPN as a feature extractor, an SOFE module is design… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 35 publications
0
3
0
Order By: Relevance
“…An FPN is a multiscale pyramid network that detects objects of varying sizes in images by using feature maps of varying resolutions [ 84 ]. An FPN has been used to recognize and segment small objects in aerial photos and to segment buildings in satellite images [ 85 , 86 , 87 ].…”
Section: Literature Reviewmentioning
confidence: 99%
“…An FPN is a multiscale pyramid network that detects objects of varying sizes in images by using feature maps of varying resolutions [ 84 ]. An FPN has been used to recognize and segment small objects in aerial photos and to segment buildings in satellite images [ 85 , 86 , 87 ].…”
Section: Literature Reviewmentioning
confidence: 99%
“…FPN [18] is an important part of the Mask R-CNN in order to improve the robustness of Mask R-CNN for detecting targets of different sizes. FPN adopts a multi-scale feature fusion method, which can significantly improve the scale robustness of feature expression without greatly increasing the amount of computation [19].…”
Section: Fpnmentioning
confidence: 99%
“…First,A 4 is sent into the mixed hole convolution, which is composed of multi-path dilated convolution layers with different rates to ensure that the final receiving field fully covers the entire area. The rate is set to, rate = [3,6,12,18]. Use a different rate for each layer, and add a GroupNorm operation after each layer.…”
Section: Scbam-fpn Architecturementioning
confidence: 99%