2021
DOI: 10.1016/j.neucom.2020.08.069
|View full text |Cite
|
Sign up to set email alerts
|

Attention based convolutional recurrent neural network for environmental sound classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
54
0
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 84 publications
(55 citation statements)
references
References 16 publications
0
54
0
1
Order By: Relevance
“…In the work of Zhang et al [16], a convolutional RNN architecture with an attention mechanism, namely ACRNN, was proposed. Attention for both CNN and RNN layers was investigated.…”
Section: Attention Mechanismsmentioning
confidence: 99%
See 1 more Smart Citation
“…In the work of Zhang et al [16], a convolutional RNN architecture with an attention mechanism, namely ACRNN, was proposed. Attention for both CNN and RNN layers was investigated.…”
Section: Attention Mechanismsmentioning
confidence: 99%
“…As shown in Tables I, the two model architectures of A-BiLSTM and A-BiGRU have the same structures, with the only differing aspect being the type of the RNN network, i.e., LSTM or GRU, implemented in layers 1 and 3. The choice of implementing the attention mechanism in the second layer of each model was influenced by the suggestion in [16], which demonstrated that the attention mechanism was best suited to increase the model accuracy by being on layer 2 or 10 of the network. With the choice of a dense layer being the final connected layer of the two architectures, therefore the only remaining option was to implement the mechanism on layer 2.…”
Section: B the Proposed Model Architecturesmentioning
confidence: 99%
“…Traditional machine learning algorithms for audio pattern recognition include K-nearest neighbors, support vector machines, and Gaussian mixture models, etc. But with the support of more labelled datasets, neural networks based methods including convolutional neural networks [4,5], recurrent neural networks [6,7], and their combination [8,9], have achieved superior performance over the traditional approaches.…”
Section: Introductionmentioning
confidence: 99%
“…Li et al [15] proposed a multi-stream network with temporal attention in which the structure is composed of three streams, each containing a single temporal attention vector. Zhang et al [9] integrated temporal attention into its CRNN architecture and the same authors [16] proposed a model that combines channel attention and temporal attention together.…”
Section: Introductionmentioning
confidence: 99%
“…Recent studies have shown that recurrent neural networks (RNN) produce excellent results for variable-length sound sequences. Zhang et al [23] proposed a CNN architecture to learn spectro-temporal features and a bidirectional gated recurrent unit (Bi-GRU) with a frame-level attention mechanism for sound classification. Wang et al [24] proposed a CNN architecture with a parallel temporal-spectral attention mechanism to capture certain frames where sound events occur and pay attention to varying frequency bands.…”
Section: Introductionmentioning
confidence: 99%