2018
DOI: 10.48550/arxiv.1811.12152
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Efficient Coarse-to-Fine Non-Local Module for the Detection of Small Objects

Abstract: a) Can you see the swimming people? (b) Where is the ball?Figure 1: Using a relational non-local module directly on the feature maps in a coarse-to-fine manner enables the detection of small objects, based on (i) repeating instances of the same class and (ii) the existence of larger related objects, allowing us to: (a) pay attention to the tiny swimmers in the sea and (b) locate the ball. Cyan -NL, Redours, ENL. Best viewed in color.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
9
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(9 citation statements)
references
References 33 publications
0
9
0
Order By: Relevance
“…Comparing with the standard NL block, the proposed LightNL block is about 400× computationally cheaper (6.2G vs. 15M) with comparable performance (75.2% vs. 75.0%). Comparing with Levi et al[22] which optimized the matrix multiplication with the associative law, the proposed LightNL block is still 10× computationally cheaper. Compared with a very recent work proposed by Zhu et al[49] which leverages the pyramid pooling to reduce the complexity, LightNL is around 7× computationally cheaper.…”
mentioning
confidence: 90%
See 3 more Smart Citations
“…Comparing with the standard NL block, the proposed LightNL block is about 400× computationally cheaper (6.2G vs. 15M) with comparable performance (75.2% vs. 75.0%). Comparing with Levi et al[22] which optimized the matrix multiplication with the associative law, the proposed LightNL block is still 10× computationally cheaper. Compared with a very recent work proposed by Zhu et al[49] which leverages the pyramid pooling to reduce the complexity, LightNL is around 7× computationally cheaper.…”
mentioning
confidence: 90%
“…Before matrix multiplications, the outputs of 1 × 1 convolution are reshaped to (H × W, C). Levi et al [22] discover that for NL blocks instantiated in the form of Eqn. (3), employing the associative law of matrix multiplication can largely reduce the computation overhead.…”
Section: Lightweight Non-local Blocksmentioning
confidence: 99%
See 2 more Smart Citations
“…Computational efficiency There are multiple ways to control the computational cost of deep neural networks. We categorize them into four groups: i) compression methods that aim to remove redundancy from already trained models [34]; ii) lightweight design strategies used to replace network components with computationally lighter counterparts [35]; iii) partial computation methods selectively utilize units of a network, thus creating forward-propagation paths with different computational costs [36]; and iv) attention mechanisms that can be used to selectively process subsets of the input, based on their importance for the task of interest [8,37,38]. The latter being the strategy we consider in the proposed Zoom-In architecture.…”
Section: Tiny Object Classificationmentioning
confidence: 99%