Animal acoustic communication often takes the form of complex sequences, made up of multiple distinct acoustic units. Apart from the well-known example of birdsong, other animals such as insects, amphibians, and mammals (including bats, rodents, primates, and cetaceans) also generate complex acoustic sequences. Occasionally, such as with birdsong, the adaptive role of these sequences seems clear (e.g. mate attraction and territorial defence). More often however, researchers have only begun to characterise – let alone understand – the significance and meaning of acoustic sequences. Hypotheses abound, but there is little agreement as to how sequences should be defined and analysed. Our review aims to outline suitable methods for testing these hypotheses, and to describe the major limitations to our current and near-future knowledge on questions of acoustic sequences. This review and prospectus is the result of a collaborative effort between 43 scientists from the fields of animal behaviour, ecology and evolution, signal processing, machine learning, quantitative linguistics, and information theory, who gathered for a 2013 workshop entitled, “Analysing vocal sequences in animals”. Our goal is to present not just a review of the state of the art, but to propose a methodological framework that summarises what we suggest are the best practices for research in this field, across taxa and across disciplines. We also provide a tutorial-style introduction to some of the most promising algorithmic approaches for analysing sequences. We divide our review into three sections: identifying the distinct units of an acoustic sequence, describing the different ways that information can be contained within a sequence, and analysing the structure of that sequence. Each of these sections is further subdivided to address the key questions and approaches in that area. We propose a uniform, systematic, and comprehensive approach to studying sequences, with the goal of clarifying research terms used in different fields, and facilitating collaboration and comparative studies. Allowing greater interdisciplinary collaboration will facilitate the investigation of many important questions in the evolution of communication and sociality.
The spectral and temporal properties of echolocation clicks and the use of clicks for species classification are investigated for five species of free-ranging dolphins found offshore of southern California: short-beaked common (Delphinus delphis), long-beaked common (D. capensis), Risso's (Grampus griseus), Pacific white-sided (Lagenorhynchus obliquidens), and bottlenose (Tursiops truncatus) dolphins. Spectral properties are compared among the five species and unique spectral peak and notch patterns are described for two species. The spectral peak mean values from Pacific white-sided dolphin clicks are 22.2, 26.6, 33.7, and 37.3 kHz and from Risso's dolphins are 22.4, 25.5, 30.5, and 38.8 kHz. The spectral notch mean values from Pacific white-sided dolphin clicks are 19.0, 24.5, and 29.7 kHz and from Risso's dolphins are 19.6, 27.7, and 35.9 kHz. Analysis of variance analyses indicate that spectral peaks and notches within the frequency band 24-35 kHz are distinct between the two species and exhibit low variation within each species. Post hoc tests divide Pacific white-sided dolphin recordings into two distinct subsets containing different click types, which are hypothesized to represent the different populations that occur within the region. Bottlenose and common dolphin clicks do not show consistent patterns of spectral peaks or notches within the frequency band examined (1-100 kHz).
Beaked whale echolocation signals are mostly frequency-modulated (FM) upsweep pulses and appear to be species specific. Evolutionary processes of niche separation may have driven differentiation of beaked whale signals used for spatial orientation and foraging. FM pulses of eight species of beaked whales were identified, as well as five distinct pulse types of unknown species, but presumed to be from beaked whales. Current evidence suggests these five distinct but unidentified FM pulse types are also species-specific and are each produced by a separate species. There may be a relationship between adult body length and center frequency with smaller whales producing higher frequency signals. This could be due to anatomical and physiological restraints or it could be an evolutionary adaption for detection of smaller prey for smaller whales with higher resolution using higher frequencies. The disadvantage of higher frequencies is a shorter detection range. Whales echolocating with the highest frequencies, or broadband, likely lower source level signals also use a higher repetition rate, which might compensate for the shorter detection range. Habitat modeling with acoustic detections should give further insights into how niches and prey may have shaped species-specific FM pulse types.
Acoustic data provide scientific and engineering insights in fields ranging from biology and communications to ocean and Earth science. We survey the recent advances and transformative potential of machine learning (ML), including deep learning, in the field of acoustics. ML is a broad family of statistical techniques for automatically detecting and utilizing patterns in data. Relative to conventional acoustics and signal processing, ML is data-driven. Given sufficient training data, ML can discover complex relationships between features and desired labels or actions, or between features themselves. With large volumes of training data, ML can discover models describing complex acoustic phenomena such as human speech and reverberation. ML in acoustics is rapidly developing with compelling results and significant future promise. We first introduce ML, then highlight ML developments in five acoustics research areas: source localization in speech processing, source localization in ocean acoustics, bioacoustics, seismic exploration, and environmental sounds in everyday scenes.
Deep neural networks have advanced the field of detection and classification and allowed for effective identification of signals in challenging data sets. Numerous time-critical conservation needs may benefit from these methods. We developed and empirically studied a variety of deep neural networks to detect the vocalizations of endangered North Atlantic right whales (Eubalaena glacialis). We compared the performance of these deep architectures to that of traditional detection algorithms for the primary vocalization produced by this species, the upcall. We show that deep-learning architectures are capable of producing false-positive rates that are orders of magnitude lower than alternative algorithms while substantially increasing the ability to detect calls. We demonstrate that a deep neural network trained with recordings from a single geographic region recorded over a span of days is capable of generalizing well to data from multiple years and across the species' range, and that the low false positives make the output of the algorithm amenable to quality control for verification. The deep neural networks we developed are relatively easy to implement with existing software, and may provide new insights applicable to the conservation of endangered species.
scite is a Brooklyn-based organization that helps researchers better discover and understand research articles through Smart Citations–citations that display the context of the citation and describe whether the article provides supporting or contrasting evidence. scite is used by students and researchers from around the world and is funded in part by the National Science Foundation and the National Institute on Drug Abuse of the National Institutes of Health.
customersupport@researchsolutions.com
10624 S. Eastern Ave., Ste. A-614
Henderson, NV 89052, USA
This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
Copyright © 2024 scite LLC. All rights reserved.
Made with 💙 for researchers
Part of the Research Solutions Family.