2021
DOI: 10.1101/2021.03.26.436927
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Fast and accurate annotation of acoustic signals with deep neural networks

Abstract: Acoustic signals serve communication within and across species throughout the animal kingdom. Studying the genetics, evolution, and neurobiology of acoustic communication requires annotating acoustic signals: segmenting and identifying individual acoustic elements like syllables or sound pulses. To be useful, annotations need to be accurate, robust to noise, fast. We introduce DeepSS, a method that annotates acoustic signals across species based on a deep-learning derived hierarchical presentation of sound. We… Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(2 citation statements)
references
References 84 publications
(261 reference statements)
0
2
0
Order By: Relevance
“…A more recent tool is Deep Song Segmenter (DeepSS), which has been used for annotation of songs of mice, birds and flies (Steinfath et al, 2021). DeepSS learns a representation of sounds features directly from raw audio recordings using temporal convolutional networks (TCNs), based on dilated convolutions.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…A more recent tool is Deep Song Segmenter (DeepSS), which has been used for annotation of songs of mice, birds and flies (Steinfath et al, 2021). DeepSS learns a representation of sounds features directly from raw audio recordings using temporal convolutional networks (TCNs), based on dilated convolutions.…”
Section: Introductionmentioning
confidence: 99%
“…Over the past two decades there has been a growing interest in the usage and signaling of vocalizations in mice, with large efforts going into studying the underlying neurobiological mechanisms for auditory processing (Pomerantz et al, 1983;Liu et al, 2003;Neilans et al, 2014;Perrodin et al, 2020;Holy and Guo, 2005), and the production of vocalizations (Arriaga et al, 2012;Chabout et al, 2016;Okobi et al, 2019;Zimmer et al, 2019;Gao et al, 2019;Tschida et al, 2019;Michael et al, 2020). The tools available for experiments in mice provide a promising model for studying the neural basis of vocalizations, as well as the effects of genes on the origin and development of vocal and neural anatomy (Grimsley et al, 2011;Bowers et al, 2013;Chabout et al, 2016;Tabler et al, The copyright holder for this preprint this version posted August 13, 2021. ; https://doi.org/10.1101/2021.08.13.456283 doi: bioRxiv preprint (Steinfath et al, 2021). DeepSS learns a representation of sounds features directly from raw audio recordings using temporal convolutional networks (TCNs), based on dilated convolutions.…”
Section: Introductionmentioning
confidence: 99%