2017
DOI: 10.48550/arxiv.1709.00179
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Effective Use of Dilated Convolutions for Segmenting Small Object Instances in Remote Sensing Imagery

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
24
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(25 citation statements)
references
References 15 publications
1
24
0
Order By: Relevance
“…Key point classification can benefit from larger receptive field kernels which can provide better context aware feature extraction. This can be achieved efficiently using dilated convolutions [21] which increase receptive field of filters without increasing the number of parameters. Average pooling is better suited to preserve the overall context features compared to max pooling which picks the most dominant feature in the receptive field.…”
Section: B Network Architecturementioning
confidence: 99%
See 1 more Smart Citation
“…Key point classification can benefit from larger receptive field kernels which can provide better context aware feature extraction. This can be achieved efficiently using dilated convolutions [21] which increase receptive field of filters without increasing the number of parameters. Average pooling is better suited to preserve the overall context features compared to max pooling which picks the most dominant feature in the receptive field.…”
Section: B Network Architecturementioning
confidence: 99%
“…To further improve the receptive fields of the filters without increasing parameter size, we use dilated convolutions throughout the downsampling blocks. Hamaguchi et al [21] showed that dilated convolutions are useful in feature extraction for small and crowded objects. This feature suits our problem of object detection in BEV images where objects are relatively small.…”
Section: B Network Architecturementioning
confidence: 99%
“…where RF i , K i , R i and Stride i are the size of receptive field, size of kernel, dilation rate and stride of the ith layer respectively. Inspired by [41] and other novel networks [20], [43], the dilation rate is usually set to sequence, we deployed dilated convolutions in stage2 demonstrated in Table I, and the sequence of dilation rate is 1, 2, 1, 4, 1, 8, 1, 16. Using (7) and ( 8), we can calculate the receptive field of our ESFNet is 599, and if we remove the dilated convolutions in stage2, the receptive field is only 183 that is not enough to cover the whole image.…”
Section: A Our Core Modulementioning
confidence: 99%
“…D ILATED convolutions, also known as atrous convolutions, have been widely explored in deep convolutional neural networks (DCNNs) for various tasks, including semantic image segmentation [1], [2], [3], [4], [5], [6], [7], [8], [9], [10], object detection [11], [12], [13], [14], audio generation [15], video modeling [16], and machine translation [17]. The idea of dilated filters was developed in the algorithm à trous for efficient wavelet decomposition in [18] and has been used in image pixel-wise prediction tasks to allow efficient computation [1], [2], [11], [12].…”
Section: Introductionmentioning
confidence: 99%
“…Dilation upsamples convolutional filters by inserting zeros between weights, as illustrated in Figure 1. It enlarges the receptive field, or field of view [5], [6], [8], but does not require training extra parameters in DCNNs. Dilated convolutions can be used in cascade to build multi-layer networks [15], [16], [17].…”
Section: Introductionmentioning
confidence: 99%