2020
DOI: 10.1121/10.0000514
|View full text |Cite
|
Sign up to set email alerts
|

Classification of odontocete echolocation clicks using convolutional neural network

Abstract: A method based on a convolutional neural network for the automatic classification of odontocete echolocation clicks is presented. The proposed convolutional neural network comprises six layers: three one-dimensional convolutional layers, two fully connected layers, and a softmax classification layer. Rectified linear units were chosen as the activation function for each convolutional layer. The input to the first convolutional layer is the raw time signal of an echolocation click. Species prediction was perfor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 12 publications
(7 citation statements)
references
References 28 publications
0
7
0
Order By: Relevance
“…In contrast to our work which encompasses a large spectral input for the soundscape as a whole, Ibrahim et al (2020) focuses on the frequency range 10 -400 Hz, narrowing the focus of the spectral input to the signals of interest. As the bandwidth of interest (Liu et al, 2018) for echolocation clicks, 93% for both whistles and clicks of a single species (Bergler et al, 2019) and outperform existing general mixed model efforts (Roch et al, 2011a) achieving high accuracies at multi-species click classification (Yang et al, 2020), making use of larger architectures and labelled training sets. Using an open-source 'light-weight' architecture and a small annotated training set we demonstrate similar overall accuracies of our model and in-depth exploration of seasonal variation on model performance to present researchers with an insight into the reliability of CNNs across annual cycles.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…In contrast to our work which encompasses a large spectral input for the soundscape as a whole, Ibrahim et al (2020) focuses on the frequency range 10 -400 Hz, narrowing the focus of the spectral input to the signals of interest. As the bandwidth of interest (Liu et al, 2018) for echolocation clicks, 93% for both whistles and clicks of a single species (Bergler et al, 2019) and outperform existing general mixed model efforts (Roch et al, 2011a) achieving high accuracies at multi-species click classification (Yang et al, 2020), making use of larger architectures and labelled training sets. Using an open-source 'light-weight' architecture and a small annotated training set we demonstrate similar overall accuracies of our model and in-depth exploration of seasonal variation on model performance to present researchers with an insight into the reliability of CNNs across annual cycles.…”
Section: Discussionmentioning
confidence: 99%
“…CNNs learn to discriminate spectrotemporal information directly from a labelled spectrogram used as an image input, removing the dependence on human experts for manual feature extraction, and improving the robustness to variation in signal structure, caller distance and signal-to-noiseratio (SNR) conditions (Gibb et al, 2019). The success of CNNs has been demonstrated by many studies in the marine domain for binary species detection and multi-class species classification (Belgith et al, 2018;Harvey, 2018;Liu et al, 2018;Bergler et al, 2019;Bermant et al, 2019;Shiu et al, 2020;Yang et al, 2020;Zhong et al, 2020;Allen et al, 2021) advancing the capabilities of mining large PAM datasets for detecting species of interest. Existing work tends to make use of spectrogram representations across a limited bandwidth, which is selected according to the species (or signal) of interest.…”
Section: Introductionmentioning
confidence: 99%
“…Recent studies in acoustics also showed that CNNs work well with acoustic data to identify and classify sounds in both time-(1D) and time-frequency domains (2D). In these works, the marine mammal species were classified using the sounds produced by animals using CNNs [21,22]. Our goal here is to solve a related regression problem by training CNNs directly on the time-domain echo data to learn the edge-diffraction induced patterns of temporal echoes (see Figure 1B) and map them to the material parameter tuple (ρ, K, G).…”
Section: Materials Parameter Retrieval Using Cnnsmentioning
confidence: 99%
“…In classifying animal sounds, deep neural network (DNN) methods have progressed tremendously with accessibility to large training data and increasing computational power. Using spectrograms generated from raw audio recordings as input, researchers have applied convolutional neural networks (CNN), either by training the model from scratch or using transfer learning with pre-trained model weights, to classify calls from different species (Bergler et al, 2019;Yang et al, 2020, Zhong et al, 2020, Kirsebom et al, 2020. Another approach is the use of recurrent neural networks (RNN), which utilize temporal information of animal calls for classification tasks (Ibrahim et al, 2018;Shiu et al, 2020).…”
Section: B Motivation For the Workmentioning
confidence: 99%