2021
DOI: 10.1587/transinf.2020edp7111
|View full text |Cite
|
Sign up to set email alerts
|

Neural Architecture Search for Convolutional Neural Networks with Attention

Abstract: The recent development of neural architecture search (NAS) has enabled us to automatically discover architectures of neural networks with high performance within a few days. Convolutional neural networks extract fruitful features by repeatedly applying standard operations (convolutions and poolings). However, these operations also extract useless or even disturbing features. Attention mechanisms enable neural networks to discard information of no interest, having achieved the state-of-the-art performance. Whil… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2023
2023
2025
2025

Publication Types

Select...
3
1
1

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 25 publications
0
5
0
Order By: Relevance
“…The downside of this algorithm is its running time; it takes more than 50 GPU days to optimize the model. Moreover, there are different methods, except the mentioned algorithm, in the literature like binarized neural networks [147], swarm intelligence [148], greedy optimizers [47,149], novelty search strategy [150], attention-based search [151], slow-fast learning [152], enhanced RL mixed with a new reward function [60], etc.…”
Section: Adaptive Trainingmentioning
confidence: 99%
“…The downside of this algorithm is its running time; it takes more than 50 GPU days to optimize the model. Moreover, there are different methods, except the mentioned algorithm, in the literature like binarized neural networks [147], swarm intelligence [148], greedy optimizers [47,149], novelty search strategy [150], attention-based search [151], slow-fast learning [152], enhanced RL mixed with a new reward function [60], etc.…”
Section: Adaptive Trainingmentioning
confidence: 99%
“…Hao et al [47] proposed to introduce an attention mechanism in the network architecture to help the information interaction between candidate architectures, enabling the search process to focus on selecting better network architectures. Weng et al [48] and Nakai et al [49] both added attention mechanism modules as an operation to the search space. The former added the attention module directly into the search space, while the latter introduced a new attention search space containing multiple attention operations and concatenated the operations searched in the attention search space with the operations searched in the original search space, and both approaches improved the performance of the network architecture.…”
Section: Related Workmentioning
confidence: 99%
“…Other gradient-based neural network architecture search methods have explicit and implicit approaches to deal with the problem of poor stability in architecture search, respectively. P-DARTS [45], ASM-NAS [47], Att-DARTS [49], DARTS+ [51], DARTS- [52], MileNAS [54], R-DARTS [59],…”
Section: Relationship With Previous Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In addition, the attention mechanism can help the neural network select useful features and discard the less-useful ones. Attention mechanism modules have been introduced to enrich the search space 24 , 25 to improve the architecture search performance. Some other methods have also used different search strategies 26 30 to try to alleviate the above problems.…”
Section: Introductionmentioning
confidence: 99%