2016
DOI: 10.1007/978-3-319-46493-0_12
|View full text |Cite
|
Sign up to set email alerts
|

A Multi-scale CNN for Affordance Segmentation in RGB Images

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
78
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
4
2

Relationship

0
10

Authors

Journals

citations
Cited by 127 publications
(78 citation statements)
references
References 37 publications
0
78
0
Order By: Relevance
“…23. http://www.semantic3d.net/ [43] GoogLeNet(FCN) Patchwise CNN, Standalone CRF CRFasRNN [70] FCN-8s CRF reformulated as RNN Dilation [71] VGG-16 Dilated convolutions ENet [72] ENet bottleneck Bottleneck module for efficiency Multi-scale-CNN-Raj [73] VGG-16(FCN) Multi-scale architecture Multi-scale-CNN-Eigen [74] Custom Multi-scale sequential refinement Multi-scale-CNN-Roy [75] Multi-scale-CNN-Eigen Multi-scale coarse-to-fine refinement Multi-scale-CNN-Bian [76] FCN Independently trained multi-scale FCNs ParseNet [77] VGG-16 Global context feature fusion ReSeg [78] VGG-16 + ReNet Extension of ReNet to semantic segmentation LSTM-CF [79] Fast R-CNN + DeepMask Fusion of contextual information from multiple sources 2D-LSTM [80] MDRNN Image context modelling rCNN [81] MDRNN Different input sizes, image context DAG-RNN [82] Elman network Graph image structure for context modelling SDS [10] R-CNN + Box CNN Simultaneous detection and segmentation DeepMask [83] VGG-A Proposals generation for segmentation SharpMask [84] DeepMask Top-down refinement module MultiPathNet [85] Fast R-CNN + DeepMask Multi path information flow through network Huang-3DCNN [86] Own 3DCNN 3DCNN for voxelized point clouds PointNet [87] Own MLP-based Segmentation of unordered point sets Clockwork Convnet [88] FCN Clockwork scheduling for sequences 3DCNN-Zhang…”
Section: Methodsmentioning
confidence: 99%
“…23. http://www.semantic3d.net/ [43] GoogLeNet(FCN) Patchwise CNN, Standalone CRF CRFasRNN [70] FCN-8s CRF reformulated as RNN Dilation [71] VGG-16 Dilated convolutions ENet [72] ENet bottleneck Bottleneck module for efficiency Multi-scale-CNN-Raj [73] VGG-16(FCN) Multi-scale architecture Multi-scale-CNN-Eigen [74] Custom Multi-scale sequential refinement Multi-scale-CNN-Roy [75] Multi-scale-CNN-Eigen Multi-scale coarse-to-fine refinement Multi-scale-CNN-Bian [76] FCN Independently trained multi-scale FCNs ParseNet [77] VGG-16 Global context feature fusion ReSeg [78] VGG-16 + ReNet Extension of ReNet to semantic segmentation LSTM-CF [79] Fast R-CNN + DeepMask Fusion of contextual information from multiple sources 2D-LSTM [80] MDRNN Image context modelling rCNN [81] MDRNN Different input sizes, image context DAG-RNN [82] Elman network Graph image structure for context modelling SDS [10] R-CNN + Box CNN Simultaneous detection and segmentation DeepMask [83] VGG-A Proposals generation for segmentation SharpMask [84] DeepMask Top-down refinement module MultiPathNet [85] Fast R-CNN + DeepMask Multi path information flow through network Huang-3DCNN [86] Own 3DCNN 3DCNN for voxelized point clouds PointNet [87] Own MLP-based Segmentation of unordered point sets Clockwork Convnet [88] FCN Clockwork scheduling for sequences 3DCNN-Zhang…”
Section: Methodsmentioning
confidence: 99%
“…Scale variation typically refers to large variations in scale of the objects being counted (in this case heads) (i) within image and (ii) across images in a dataset. Several other related tasks like object detection [6,16,23,30,37,45] and visual saliency detection [10,14,41,73] are also affected by such effects. However, these effects are more evident especially in crowd counting in congested scenes.…”
Section: Introductionmentioning
confidence: 99%
“…The network input was encoded as a HHA encoding [28] of depth along with the color image. A more generic affordance prediction framework was presented in [185], which used a multi-scale CNN to provide affordance segmentations for indoor scenes (see Figure 21). Their architecture explicitly used mid-level geometric and semantic representations such as labels, surface normals and depth maps at coarse and fine levels to effectively aggregate information.…”
Section: Methods Overviewmentioning
confidence: 99%