2022
DOI: 10.3389/fmars.2022.879145
|View full text |Cite
|
Sign up to set email alerts
|

More than a whistle: Automated detection of marine sound sources with a convolutional neural network

Abstract: The effective analysis of Passive Acoustic Monitoring (PAM) data has the potential to determine spatial and temporal variations in ecosystem health and species presence if automated detection and classification algorithms are capable of discrimination between marine species and the presence of anthropogenic and environmental noise. Extracting more than a single sound source or call type will enrich our understanding of the interaction between biological, anthropogenic and geophonic soundscape components in the… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
6
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 13 publications
(6 citation statements)
references
References 71 publications
0
6
0
Order By: Relevance
“…A whistle label could include all other labels, click labels excluded whistles, vessel labels excluded clicks and whistles, and ambient labels excluded the other three. The labelling processes was designed for a CNN model 23 , with 3 second labelling windows chosen as a compromise between the time periods of milliseconds for clicks, seconds for whistles and minutes for a vessel passage. Data were labelled both visually using spectrograms, and aurally using Audacity Software (Audacity version 3.0.02, 2021).…”
Section: Datamentioning
confidence: 99%
See 1 more Smart Citation
“…A whistle label could include all other labels, click labels excluded whistles, vessel labels excluded clicks and whistles, and ambient labels excluded the other three. The labelling processes was designed for a CNN model 23 , with 3 second labelling windows chosen as a compromise between the time periods of milliseconds for clicks, seconds for whistles and minutes for a vessel passage. Data were labelled both visually using spectrograms, and aurally using Audacity Software (Audacity version 3.0.02, 2021).…”
Section: Datamentioning
confidence: 99%
“…Data was labelled between the hours of 18:00 and 09:00 for each day to maximise the detection of delphinid activity. Mean error rate in the labelling was estimated to be 3.3%, with the greatest error occurring between Ambient (6.3%) and Vessel (3.7%) classes (see 23 for full details).…”
Section: Datamentioning
confidence: 99%
“…Large-scale studies often employ Convolutional Neural Network (CNN) architectures to carry out supervised detection or classification tasks across diverse species. These include detection of humpback whale vocalizations in passive acoustic monitoring datasets 5 ; detection, classification, and censusing of blue whale sounds 6 ; avian species monitoring based on a CNN trained to predict the species class label given input audio 7 ; presence indication of Hainanese gibbons using a ResNet-based CNN 8 ; and, recently, detection and classification of marine sound sources using an image-based approach to spectrogram classification 9 . However, such supervised learning approaches remain limited in their scope, hindering their capacity to be deployed in real-time data processing pipelines.…”
Section: Introductionmentioning
confidence: 99%
“…Automation of the detection of these acoustic signals is therefore an essential component of the analysis pipeline for gathering data on delphinid occurrence.The application of Deep Learning (DL) algorithms is increasingly prevalent within the field of bioacoustics. Convolutional Neural Networks (CNNs), which use image processing techniques to read and classify acoustic data into predetermined classes, have been shown to excel at detecting signals which vary in respect to time, frequency, and amplitude 13,14,15,16 . Using broadband acoustic data as input, CNNs have achieved high accuracies at the task of detecting dolphin vocalisations within variable marine environments 17,18,19,20 .…”
mentioning
confidence: 99%
“…Proprietary software is provided which contains a custom classifier (KERNO) to classify click trains as originating from either dolphins, porpoises or other cetaceans based on their intensity, duration, frequency content and inter-click intervals (ICI) 31,32 . In this work we evaluate the performance of the C-POD compared to a newly available multi-sound source CNN 16 for the task of monitoring hourly dolphin presence in the waters off the west coast of Scotland (Figure 1), using PAM data collected within the COMPASS array (EU INTERREG COMPASS project). We present the first empirical comparison between a CNN and the C-POD for dolphin detection, highlighting the efficiency of each method as a monitoring tool in diverse soundscape conditions.…”
mentioning
confidence: 99%