2020 International Joint Conference on Neural Networks (IJCNN) 2020
DOI: 10.1109/ijcnn48605.2020.9206992
|View full text |Cite
|
Sign up to set email alerts
|

Learning Deep Models from Synthetic Data for Extracting Dolphin Whistle Contours

Abstract: We present a learning-based method for extracting whistles of toothed whales (Odontoceti) in hydrophone recordings. Our method represents audio signals as time-frequency spectrograms and decomposes each spectrogram into a set of time-frequency patches. A deep neural network learns archetypical patterns (e.g., crossings, frequency modulated sweeps) from the spectrogram patches and predicts time-frequency peaks that are associated with whistles. We also developed a comprehensive method to synthesize training sam… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
10
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(10 citation statements)
references
References 84 publications
0
10
0
Order By: Relevance
“…This would involve an accurate identification of each whistle. However, this task is complex, and detection techniques developed so far have not been able to solve this problem without errors [69][70][71][72][73][74][75][76][77][78][79][80][81][82]84,85]. If whistles are first identified, then a classification of whistles using unsupervised classification techniques such as UMAP [67] could be applied, also using a metric to determine the optimal number of clusters [95].…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…This would involve an accurate identification of each whistle. However, this task is complex, and detection techniques developed so far have not been able to solve this problem without errors [69][70][71][72][73][74][75][76][77][78][79][80][81][82]84,85]. If whistles are first identified, then a classification of whistles using unsupervised classification techniques such as UMAP [67] could be applied, also using a metric to determine the optimal number of clusters [95].…”
Section: Discussionmentioning
confidence: 99%
“…Thus, to detect and identify whistles, a vocalisation tracking method is necessary. Selecting vocalisations in an audio track is a complex task, and several techniques have already been developed to achieve it: statistical modelling of whistles [69][70][71][72][73][74], tracking algorithms based on hand-picked parameters [75][76][77], image processing approaches [78][79][80][81][82][83] or deep learning models associated with clustering methods [84,85]. For our dataset, we chose to adapt a tracking algorithm developed during the DECAV project [86].…”
Section: Vocalisation Detectionmentioning
confidence: 99%
“…In the future, larger training datasets should be considered, either by the hand annotation of the field data or, potentially, by constructing them from existing data with the help of the deep learning procedures. 56 The parameters reported in this paper were trained based on the whistle cross-correlograms and used for tracking all of the measurement types, including echolocation clicks and combined clicks and whistles. The parameters related to the evolution of the TDOA tracks are expected to be similar between the whistle or click cross-correlograms.…”
Section: Discussionmentioning
confidence: 99%
“…In particular, Convolutional Neural Networks (CNN) (LeCun et al, 1995 ) have been applied to detection of whales vocalizations, producing false-positive rates that are orders of magnitude lower than traditional algorithms, while substantially increasing the ability to detect calls (Jiang et al, 2019 ; Shiu et al, 2020 ). Deep learning has also been used to automatically classify dolphin whistles into specific categories (Li et al, 2021 ) and to extract whistle contours by exploiting peak tracking algorithms (Li et al, 2020 ) or by training CNN-based semantic segmentation models (Jin et al, 2022 ).…”
Section: Introductionmentioning
confidence: 99%