2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018
DOI: 10.1109/cvpr.2018.00256
|View full text |Cite
|
Sign up to set email alerts
|

Convolutional Neural Networks with Alternately Updated Clique

Abstract: Improving information flow in deep networks helps to ease the training difficulties and utilize parameters more efficiently. Here we propose a new convolutional neural network architecture with alternately updated clique (CliqueNet). In contrast to prior networks, there are both forward and backward connections between any two layers in the same block. The layers are constructed as a loop and are updated alternately. The CliqueNet has some unique properties. For each layer, it is both the input and output of a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
111
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 134 publications
(111 citation statements)
references
References 29 publications
0
111
0
Order By: Relevance
“…Semantic segmentation. Fully convolutional network (FCN) [22] based methods have made great progress in image semantic segmentation by leveraging the powerful convolutional features of classification networks [14,15,33] pre-trained on large-scale data [28]. Several model variants are proposed to enhance the multi-scale contextual aggregation.…”
Section: Related Workmentioning
confidence: 99%
“…Semantic segmentation. Fully convolutional network (FCN) [22] based methods have made great progress in image semantic segmentation by leveraging the powerful convolutional features of classification networks [14,15,33] pre-trained on large-scale data [28]. Several model variants are proposed to enhance the multi-scale contextual aggregation.…”
Section: Related Workmentioning
confidence: 99%
“…The initial learning rate is usually the single most important hyper-parameter, we should always make sure that it is tuned [32]. There are two structure parameters to set first [26], T is the sum of the layers of all the blocks, k represents the number of filters per layer in each block. The kernel size of convolution layers in all blocks are 3 × 3, and we use one-pixel padding to keep the size of the matrix the same before and after the convolution.…”
Section: Training Detailsmentioning
confidence: 99%
“…Batch normalization [31] can accelerate deep network training by reducing internal covariate shift. It allows us to use much higher learning rates and be less careful about initialization.…”
Section: Octonion Batch Normalization Modulementioning
confidence: 99%