2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017
DOI: 10.1109/cvpr.2017.75
|View full text |Cite
|
Sign up to set email alerts
|

Dilated Residual Networks

Abstract: Convolutional networks for image classification progressively reduce resolution until the image is represented by tiny feature maps in which the spatial structure of the scene is no longer discernible. Such loss of spatial acuity can limit image classification accuracy and complicate the transfer of the model to downstream applications that require detailed scene understanding. These problems can be alleviated by dilation, which increases the resolution of output feature maps without reducing the receptive fie… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

3
1,190
0
1

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 1,733 publications
(1,194 citation statements)
references
References 19 publications
3
1,190
0
1
Order By: Relevance
“…Architectures and Training: We implement three architectures: an LSTM [48], a Transformer [49], and a dilated residual network (ResNet) [50]. We use a 12-layer Transformer with a hidden size of 512 units and 8 attention heads, leading to a 38M-parameter model.…”
Section: Models and Experimental Setupmentioning
confidence: 99%
“…Architectures and Training: We implement three architectures: an LSTM [48], a Transformer [49], and a dilated residual network (ResNet) [50]. We use a 12-layer Transformer with a hidden size of 512 units and 8 attention heads, leading to a 38M-parameter model.…”
Section: Models and Experimental Setupmentioning
confidence: 99%
“…The encoder E uses residual connections and dilated convolutions (dilation rate = 2) to enlarge the size of receptive field while preserving the spatial resolution for dense predictions [46]. Let {Ck, Rk, Dk} denote a convolutional layer, a residual block and a dilated residual block with k channels, respectively.…”
Section: Network Configurations and Implementation Detailsmentioning
confidence: 99%
“…7 is inspired by the refinement procedures proposed in CRL [17], iResNet [12], StereoNet [8], and ActiveStere-oNet [31]. We adopted the basic architecture for refinement as described in StereoNet [8] with dilated residual blocks [28] to increase the receptive field of filtering without compromising resolution. This technique was also adopted in recent work on optical flow prediction Pwc-net [23].…”
Section: Dilation In Cost Filteringmentioning
confidence: 99%