1
|
Ajji M. J, Lang JW. Gharial acoustic signaling: Novel underwater pops are temporally based, context-dependent, seasonally stable, male-specific, and individually distinctive. J Anat 2025; 246:415-443. [PMID: 39887971 PMCID: PMC11828749 DOI: 10.1111/joa.14171] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2024] [Revised: 09/26/2024] [Accepted: 10/22/2024] [Indexed: 02/01/2025] Open
Abstract
Gharials (Gavialis gangeticus) produce a sudden, high amplitude, pulsatile, underwater sound called a POP. In this study, gharial POPs ranged from 9 to 55 ms, and were clearly audible on land and water, at ≥500 m. POPs were only performed underwater by adult males possessing a sex-specific, cartilaginous narial excrescence, termed the ghara. We recorded 130 POP events of seven wild adult males in 115 km stretch of the Chambal River during 2017-2019, using hydrophones and aerial mics. A POP event occurs when a male produces a single or double or triple POP, each with a specific duration and timing. A POP event was incorporated into a complex, multi-modal breathing display, typically performed by each male during the breeding season. Key features of this novel gharial POP signal are documented here for the first time. These include its incorporation into a complex breathing display, its reliance on temporal rather than spectral elements, its dependence on a specific social context, its stability within an individual, and its individually distinctive patterning specific to a particular male. The breathing display consisted of sub-audible vibrations (SAV) preceding each POP, then a stereotyped exhalation-inhalation-exhalation sequence, concluding with bubbling and submergence. In our study, 96% of the variation in POP signal parameters was explained by POP signal timings (92%) and number of POPs (4%), and only 2% was related to spectral features. Each POP event was performed in a specific social setting. Two behavioral contexts were examined: ALERT and PATROL. In each context, male identities were examined using Discriminant Function Analysis (DFA). Within each context, each of the seven males exhibited distinctive POP patterns that were context-specific and denoted a male's identity and his location. POP signal features were stable for individual males, from 1 year to the next. Overall, the seven males showed POP patterns that were individually specific, with minimal overlap amongst males, yet these were remarkably diverse. The stereotypy of POP patterns, based on temporal versus frequency difference was best characterized statistically using DFA metrics, rather than Beecher's Information Statistic, MANOVA, or Discriminant Score computations. Our field observations indicated that audiences of gharial, located nearby, and/or in the distance, responded immediately to POPs by orienting in the signal direction. Extensive auditory studies of crocodylians indicate that their capacity for auditory temporal discrimination and neural processing in relation to locating a sound target is on par with that of birds. How the POP sound is produced and broadcast loudly in both water and air has received little study to date. We briefly summarize existing reports on ghara anatomy, ontogeny, and paleontology. Finally, preliminary observations made in a clear underwater zoo enclosure indicate that jaw claps performed entirely underwater produce POP sounds. Simultaneous bubble clouds emanating from the base of the ghara are suggestive of cavitation phenomena associated with loud high volume sounds such as shrimp snaps and seal/walrus claps. We discuss the likelihood that the adult male's ghara plays an essential role in the production of the non-vocal underwater POP, a sexually dimorphic acoustic signal unique to gharial.
Collapse
Affiliation(s)
- Jailabdeen Ajji M.
- Gharial Ecology ProjectMadras Crocodile Bank TrustMamallapuram, Tamil NaduIndia
| | - Jeffrey W. Lang
- Gharial Ecology ProjectMadras Crocodile Bank TrustMamallapuram, Tamil NaduIndia
| |
Collapse
|
2
|
Contina A, Abelson E, Allison B, Stokes B, Sanchez KF, Hernandez HM, Kepple AM, Tran Q, Kazen I, Brown KA, Powell JH, Keitt TH. BioSense: An automated sensing node for organismal and environmental biology. HARDWAREX 2024; 20:e00584. [PMID: 39314536 PMCID: PMC11417332 DOI: 10.1016/j.ohx.2024.e00584] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/10/2024] [Revised: 08/28/2024] [Accepted: 09/06/2024] [Indexed: 09/25/2024]
Abstract
Automated remote sensing has revolutionized the fields of wildlife ecology and environmental science. Yet, a cost-effective and flexible approach for large scale monitoring has not been fully developed, resulting in a limited collection of high-resolution data. Here, we describe BioSense, a low-cost and fully programmable automated sensing platform for applications in bioacoustics and environmental studies. Our design offers customization and flexibility to address a broad array of research goals and field conditions. Each BioSense is programmed through an integrated Raspberry Pi computer board and designed to collect and analyze avian vocalizations while simultaneously collecting temperature, humidity, and soil moisture data. We illustrate the different steps involved in manufacturing this sensor including hardware and software design and present the results of our laboratory and field testing in southwestern United States.
Collapse
Affiliation(s)
- Andrea Contina
- School of Integrative Biological and Chemical Sciences, The University of Texas Rio Grande Valley, Brownsville, TX 78520, USA
- Department of Integrative Biology, The University of Texas at Austin, Austin, TX 78703, USA
| | - Eric Abelson
- Department of Integrative Biology, The University of Texas at Austin, Austin, TX 78703, USA
| | - Brendan Allison
- Department of Integrative Biology, The University of Texas at Austin, Austin, TX 78703, USA
| | - Brian Stokes
- Department of Integrative Biology, The University of Texas at Austin, Austin, TX 78703, USA
| | | | - Henry M. Hernandez
- Department of Physics, The University of Texas at Austin, Austin, TX 78712, USA
| | - Anna M. Kepple
- Department of Integrative Biology, The University of Texas at Austin, Austin, TX 78703, USA
| | - Quynhmai Tran
- Department of Integrative Biology, The University of Texas at Austin, Austin, TX 78703, USA
| | - Isabella Kazen
- Department of Physics, The University of Texas at Austin, Austin, TX 78712, USA
| | - Katherine A. Brown
- The Oden Institute for Computational Engineering and Sciences, The University of Texas at Austin, Austin, TX 78712, USA
- Cavendish Laboratory, University of Cambridge, Cambridge CB3 0HE, UK
| | - Je’aime H. Powell
- Texas Advanced Computing Center, The University of Texas at Austin, Austin, TX 78758, USA
| | - Timothy H. Keitt
- Department of Integrative Biology, The University of Texas at Austin, Austin, TX 78703, USA
| |
Collapse
|
3
|
Kershenbaum A, Akçay Ç, Babu-Saheer L, Barnhill A, Best P, Cauzinille J, Clink D, Dassow A, Dufourq E, Growcott J, Markham A, Marti-Domken B, Marxer R, Muir J, Reynolds S, Root-Gutteridge H, Sadhukhan S, Schindler L, Smith BR, Stowell D, Wascher CAF, Dunn JC. Automatic detection for bioacoustic research: a practical guide from and for biologists and computer scientists. Biol Rev Camb Philos Soc 2024. [PMID: 39417330 DOI: 10.1111/brv.13155] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2023] [Revised: 09/30/2024] [Accepted: 10/04/2024] [Indexed: 10/19/2024]
Abstract
Recent years have seen a dramatic rise in the use of passive acoustic monitoring (PAM) for biological and ecological applications, and a corresponding increase in the volume of data generated. However, data sets are often becoming so sizable that analysing them manually is increasingly burdensome and unrealistic. Fortunately, we have also seen a corresponding rise in computing power and the capability of machine learning algorithms, which offer the possibility of performing some of the analysis required for PAM automatically. Nonetheless, the field of automatic detection of acoustic events is still in its infancy in biology and ecology. In this review, we examine the trends in bioacoustic PAM applications, and their implications for the burgeoning amount of data that needs to be analysed. We explore the different methods of machine learning and other tools for scanning, analysing, and extracting acoustic events automatically from large volumes of recordings. We then provide a step-by-step practical guide for using automatic detection in bioacoustics. One of the biggest challenges for the greater use of automatic detection in bioacoustics is that there is often a gulf in expertise between the biological sciences and the field of machine learning and computer science. Therefore, this review first presents an overview of the requirements for automatic detection in bioacoustics, intended to familiarise those from a computer science background with the needs of the bioacoustics community, followed by an introduction to the key elements of machine learning and artificial intelligence that a biologist needs to understand to incorporate automatic detection into their research. We then provide a practical guide to building an automatic detection pipeline for bioacoustic data, and conclude with a discussion of possible future directions in this field.
Collapse
Affiliation(s)
- Arik Kershenbaum
- Girton College and Department of Zoology, University of Cambridge, Huntingdon Road, Cambridge, CB3 0JG, UK
| | - Çağlar Akçay
- Behavioural Ecology Research Group, School of Life Sciences, Anglia Ruskin University, East Road, Cambridge, CB1 1PT, UK
| | - Lakshmi Babu-Saheer
- Computing Informatics and Applications Research Group, School of Computing and Information Sciences, Anglia Ruskin University, East Road, Cambridge, CB1 1PT, UK
| | - Alex Barnhill
- Pattern Recognition Lab, Department of Computer Science, Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen, 91058, Germany
| | - Paul Best
- Université de Toulon, Aix Marseille Univ, CNRS, LIS, ILCB, CS 60584, Toulon, 83041 CEDEX 9, France
| | - Jules Cauzinille
- Université de Toulon, Aix Marseille Univ, CNRS, LIS, ILCB, CS 60584, Toulon, 83041 CEDEX 9, France
| | - Dena Clink
- K. Lisa Yang Center for Conservation Bioacoustics, Cornell Lab of Ornithology, Cornell University, 159 Sapsucker Woods Road, Ithaca, New York, 14850, USA
| | - Angela Dassow
- Biology Department, Carthage College, 2001 Alford Park Dr, 68 David A Straz Jr, Kenosha, Wisconsin, 53140, USA
| | - Emmanuel Dufourq
- African Institute for Mathematical Sciences, 7 Melrose Road, Muizenberg, Cape Town, 7441, South Africa
- Stellenbosch University, Jan Celliers Road, Stellenbosch, 7600, South Africa
- African Institute for Mathematical Sciences - Research and Innovation Centre, District Gasabo, Secteur Kacyiru, Cellule Kamatamu, Rue KG590 ST No 1, Kigali, Rwanda
| | - Jonathan Growcott
- Centre of Ecology and Conservation, College of Life and Environmental Sciences, University of Exeter, Cornwall Campus, Exeter, TR10 9FE, UK
- Wildlife Conservation Research Unit, Recanati-Kaplan Centre, Tubney House, Abingdon Road Tubney, Abingdon, OX13 5QL, UK
| | - Andrew Markham
- Department of Computer Science, University of Oxford, Parks Road, Oxford, OX1 3QD, UK
| | | | - Ricard Marxer
- Université de Toulon, Aix Marseille Univ, CNRS, LIS, ILCB, CS 60584, Toulon, 83041 CEDEX 9, France
| | - Jen Muir
- Behavioural Ecology Research Group, School of Life Sciences, Anglia Ruskin University, East Road, Cambridge, CB1 1PT, UK
| | - Sam Reynolds
- Behavioural Ecology Research Group, School of Life Sciences, Anglia Ruskin University, East Road, Cambridge, CB1 1PT, UK
| | - Holly Root-Gutteridge
- School of Natural Sciences, University of Lincoln, Joseph Banks Laboratories, Beevor Street, Lincoln, Lincolnshire, LN5 7TS, UK
| | - Sougata Sadhukhan
- Institute of Environment Education and Research, Pune Bharati Vidyapeeth Educational Campus, Satara Road, Pune, Maharashtra, 411 043, India
| | - Loretta Schindler
- Department of Zoology, Faculty of Science, Charles University, Prague, 128 44, Czech Republic
| | - Bethany R Smith
- Institute of Zoology, Zoological Society of London, Outer Circle, London, NW1 4RY, UK
| | - Dan Stowell
- Tilburg University, Tilburg, The Netherlands
- Naturalis Biodiversity Center, Darwinweg 2, Leiden, 2333 CR, The Netherlands
| | - Claudia A F Wascher
- Behavioural Ecology Research Group, School of Life Sciences, Anglia Ruskin University, East Road, Cambridge, CB1 1PT, UK
| | - Jacob C Dunn
- Behavioural Ecology Research Group, School of Life Sciences, Anglia Ruskin University, East Road, Cambridge, CB1 1PT, UK
- Department of Archaeology, University of Cambridge, Downing Street, Cambridge, CB2 3DZ, UK
- Department of Behavioral and Cognitive Biology, University of Vienna, University Biology Building (UBB), Djerassiplatiz 1, Vienna, 1030, Austria
| |
Collapse
|
4
|
Jancovich BA, Rogers TL. BASSA: New software tool reveals hidden details in visualisation of low-frequency animal sounds. Ecol Evol 2024; 14:e11636. [PMID: 38962019 PMCID: PMC11220835 DOI: 10.1002/ece3.11636] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2023] [Revised: 06/06/2024] [Accepted: 06/12/2024] [Indexed: 07/05/2024] Open
Abstract
The study of animal sounds in biology and ecology relies heavily upon time-frequency (TF) visualisation, most commonly using the short-time Fourier transform (STFT) spectrogram. This method, however, has inherent bias towards either temporal or spectral details that can lead to misinterpretation of complex animal sounds. An ideal TF visualisation should accurately convey the structure of the sound in terms of both frequency and time, however, the STFT often cannot meet this requirement. We evaluate the accuracy of four TF visualisation methods (superlet transform [SLT], continuous wavelet transform [CWT] and two STFTs) using a synthetic test signal. We then apply these methods to visualise sounds of the Chagos blue whale, Asian elephant, southern cassowary, eastern whipbird, mulloway fish and the American crocodile. We show that the SLT visualises the test signal with 18.48%-28.08% less error than the other methods. A comparison between our visualisations of animal sounds and their literature descriptions indicates that the STFT's bias may have caused misinterpretations in describing pygmy blue whale songs and elephant rumbles. We suggest that use of the SLT to visualise low-frequency animal sounds may prevent such misinterpretations. Finally, we employ the SLT to develop 'BASSA', an open-source, GUI software application that offers a no-code, user-friendly tool for analysing short-duration recordings of low-frequency animal sounds for the Windows platform. The SLT visualises low-frequency animal sounds with improved accuracy, in a user-friendly format, minimising the risk of misinterpretation while requiring less technical expertise than the STFT. Using this method could propel advances in acoustics-driven studies of animal communication, vocal production methods, phonation and species identification.
Collapse
Affiliation(s)
- Benjamin A. Jancovich
- Centre for Marine Science and Innovation, School of Biological, Earth and Environmental SciencesUniversity of New South WalesKensingtonNew South WalesAustralia
| | - Tracey L. Rogers
- Centre for Marine Science and Innovation, School of Biological, Earth and Environmental SciencesUniversity of New South WalesKensingtonNew South WalesAustralia
| |
Collapse
|
5
|
Erb WM, Ross W, Kazanecki H, Mitra Setia T, Madhusudhana S, Clink DJ. Vocal complexity in the long calls of Bornean orangutans. PeerJ 2024; 12:e17320. [PMID: 38766489 PMCID: PMC11100477 DOI: 10.7717/peerj.17320] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2023] [Accepted: 04/09/2024] [Indexed: 05/22/2024] Open
Abstract
Vocal complexity is central to many evolutionary hypotheses about animal communication. Yet, quantifying and comparing complexity remains a challenge, particularly when vocal types are highly graded. Male Bornean orangutans (Pongo pygmaeus wurmbii) produce complex and variable "long call" vocalizations comprising multiple sound types that vary within and among individuals. Previous studies described six distinct call (or pulse) types within these complex vocalizations, but none quantified their discreteness or the ability of human observers to reliably classify them. We studied the long calls of 13 individuals to: (1) evaluate and quantify the reliability of audio-visual classification by three well-trained observers, (2) distinguish among call types using supervised classification and unsupervised clustering, and (3) compare the performance of different feature sets. Using 46 acoustic features, we used machine learning (i.e., support vector machines, affinity propagation, and fuzzy c-means) to identify call types and assess their discreteness. We additionally used Uniform Manifold Approximation and Projection (UMAP) to visualize the separation of pulses using both extracted features and spectrogram representations. Supervised approaches showed low inter-observer reliability and poor classification accuracy, indicating that pulse types were not discrete. We propose an updated pulse classification approach that is highly reproducible across observers and exhibits strong classification accuracy using support vector machines. Although the low number of call types suggests long calls are fairly simple, the continuous gradation of sounds seems to greatly boost the complexity of this system. This work responds to calls for more quantitative research to define call types and quantify gradedness in animal vocal systems and highlights the need for a more comprehensive framework for studying vocal complexity vis-à-vis graded repertoires.
Collapse
Affiliation(s)
- Wendy M. Erb
- K. Lisa Yang Center for Conservation Bioacoustics, Cornell Lab of Ornithology, Cornell University, Ithaca, NY, United States of America
- Department of Anthropology, Rutgers, The State University of New Jersey, New Brunswick, United States of America
| | - Whitney Ross
- K. Lisa Yang Center for Conservation Bioacoustics, Cornell Lab of Ornithology, Cornell University, Ithaca, NY, United States of America
| | - Haley Kazanecki
- K. Lisa Yang Center for Conservation Bioacoustics, Cornell Lab of Ornithology, Cornell University, Ithaca, NY, United States of America
| | - Tatang Mitra Setia
- Primate Research Center, Universitas Nasional Jakarta, Jakarta, Indonesia
- Department of Biology, Faculty of Biology and Agriculture, Universitas Nasional Jakarta, Jakarta, Indonesia
| | - Shyam Madhusudhana
- K. Lisa Yang Center for Conservation Bioacoustics, Cornell Lab of Ornithology, Cornell University, Ithaca, NY, United States of America
- Centre for Marine Science and Technology, Curtin University, Perth, Australia
| | - Dena J. Clink
- K. Lisa Yang Center for Conservation Bioacoustics, Cornell Lab of Ornithology, Cornell University, Ithaca, NY, United States of America
| |
Collapse
|
6
|
Colligan T, Irish K, Emlen DJ, Wheeler TJ. DISCO: A deep learning ensemble for uncertainty-aware segmentation of acoustic signals. PLoS One 2023; 18:e0288172. [PMID: 37494341 PMCID: PMC10370718 DOI: 10.1371/journal.pone.0288172] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/10/2023] [Accepted: 06/21/2023] [Indexed: 07/28/2023] Open
Abstract
Recordings of animal sounds enable a wide range of observational inquiries into animal communication, behavior, and diversity. Automated labeling of sound events in such recordings can improve both throughput and reproducibility of analysis. Here, we describe our software package for labeling elements in recordings of animal sounds, and demonstrate its utility on recordings of beetle courtships and whale songs. The software, DISCO, computes sensible confidence estimates and produces labels with high precision and accuracy. In addition to the core labeling software, it provides a simple tool for labeling training data, and a visual system for analysis of resulting labels. DISCO is open-source and easy to install, it works with standard file formats, and it presents a low barrier of entry to use.
Collapse
Affiliation(s)
- Thomas Colligan
- College of Pharmacy, University of Arizona, Tucson, AZ, United States of America
- Department of Computer Science, University of Montana, Missoula, MT, United States of America
| | - Kayla Irish
- Department of Computer Science, University of Montana, Missoula, MT, United States of America
- Department of Statistics, University of Washington, Seattle, WA, United States of America
| | - Douglas J Emlen
- Division of Biological Sciences, University of Montana, Missoula, MT, United States of America
| | - Travis J Wheeler
- College of Pharmacy, University of Arizona, Tucson, AZ, United States of America
- Department of Computer Science, University of Montana, Missoula, MT, United States of America
| |
Collapse
|
7
|
Smith-Vidaurre G, Pérez-Marrufo V, Hobson EA, Salinas-Melgoza A, Wright TF. Individual identity information persists in learned calls of introduced parrot populations. PLoS Comput Biol 2023; 19:e1011231. [PMID: 37498847 PMCID: PMC10374045 DOI: 10.1371/journal.pcbi.1011231] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/23/2022] [Accepted: 06/01/2023] [Indexed: 07/29/2023] Open
Abstract
Animals can actively encode different types of identity information in learned communication signals, such as group membership or individual identity. The social environments in which animals interact may favor different types of information, but whether identity information conveyed in learned signals is robust or responsive to social disruption over short evolutionary timescales is not well understood. We inferred the type of identity information that was most salient in vocal signals by combining computational tools, including supervised machine learning, with a conceptual framework of "hierarchical mapping", or patterns of relative acoustic convergence across social scales. We used populations of a vocal learning species as a natural experiment to test whether the type of identity information emphasized in learned vocalizations changed in populations that experienced the social disruption of introduction into new parts of the world. We compared the social scales with the most salient identity information among native and introduced range monk parakeet (Myiopsitta monachus) calls recorded in Uruguay and the United States, respectively. We also evaluated whether the identity information emphasized in introduced range calls changed over time. To place our findings in an evolutionary context, we compared our results with another parrot species that exhibits well-established and distinctive regional vocal dialects that are consistent with signaling group identity. We found that both native and introduced range monk parakeet calls displayed the strongest convergence at the individual scale and minimal convergence within sites. We did not identify changes in the strength of acoustic convergence within sites over time in the introduced range calls. These results indicate that the individual identity information in learned vocalizations did not change over short evolutionary timescales in populations that experienced the social disruption of introduction. Our findings point to exciting new research directions about the robustness or responsiveness of communication systems over different evolutionary timescales.
Collapse
Affiliation(s)
- Grace Smith-Vidaurre
- Department of Biology, New Mexico State University, Las Cruces, New Mexico, United States of America
- Laboratory of Neurogenetics of Language, Rockefeller University, New York, New York, United States of America
- Rockefeller University Field Research Center, Millbrook, New York, United States of America
- Department of Biological Sciences, University of Cincinnati, Cincinnati, Ohio, United States of America
| | - Valeria Pérez-Marrufo
- Department of Biology, New Mexico State University, Las Cruces, New Mexico, United States of America
- Department of Biology, Syracuse University, Syracuse, New York, United States of America
| | - Elizabeth A. Hobson
- Department of Biological Sciences, University of Cincinnati, Cincinnati, Ohio, United States of America
| | | | - Timothy F. Wright
- Department of Biology, New Mexico State University, Las Cruces, New Mexico, United States of America
| |
Collapse
|
8
|
Jourjine N, Woolfolk ML, Sanguinetti-Scheck JI, Sabatini JE, McFadden S, Lindholm AK, Hoekstra HE. Two pup vocalization types are genetically and functionally separable in deer mice. Curr Biol 2023; 33:1237-1248.e4. [PMID: 36893759 DOI: 10.1016/j.cub.2023.02.045] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/26/2022] [Revised: 02/11/2023] [Accepted: 02/14/2023] [Indexed: 03/10/2023]
Abstract
Vocalization is a widespread social behavior in vertebrates that can affect fitness in the wild. Although many vocal behaviors are highly conserved, heritable features of specific vocalization types can vary both within and between species, raising the questions of why and how some vocal behaviors evolve. Here, using new computational tools to automatically detect and cluster vocalizations into distinct acoustic categories, we compare pup isolation calls across neonatal development in eight taxa of deer mice (genus Peromyscus) and compare them with laboratory mice (C57BL6/J strain) and free-living, wild house mice (Mus musculus domesticus). Whereas both Peromyscus and Mus pups produce ultrasonic vocalizations (USVs), Peromyscus pups also produce a second call type with acoustic features, temporal rhythms, and developmental trajectories that are distinct from those of USVs. In deer mice, these lower frequency "cries" are predominantly emitted in postnatal days one through nine, whereas USVs are primarily made after day 9. Using playback assays, we show that cries result in a more rapid approach by Peromyscus mothers than USVs, suggesting a role for cries in eliciting parental care early in neonatal development. Using a genetic cross between two sister species of deer mice exhibiting large, innate differences in the acoustic structure of cries and USVs, we find that variation in vocalization rate, duration, and pitch displays different degrees of genetic dominance and that cry and USV features can be uncoupled in second-generation hybrids. Taken together, this work shows that vocal behavior can evolve quickly between closely related rodent species in which vocalization types, likely serving distinct functions in communication, are controlled by distinct genetic loci.
Collapse
Affiliation(s)
- Nicholas Jourjine
- Department of Molecular & Cellular Biology, Department of Organismic & Evolutionary Biology, Center for Brain Science, Museum of Comparative Zoology, Harvard University and the Howard Hughes Medical Institute, 16 Divinity Avenue, Cambridge, MA 02138, USA
| | - Maya L Woolfolk
- Department of Molecular & Cellular Biology, Department of Organismic & Evolutionary Biology, Center for Brain Science, Museum of Comparative Zoology, Harvard University and the Howard Hughes Medical Institute, 16 Divinity Avenue, Cambridge, MA 02138, USA
| | - Juan I Sanguinetti-Scheck
- Department of Molecular & Cellular Biology, Department of Organismic & Evolutionary Biology, Center for Brain Science, Museum of Comparative Zoology, Harvard University and the Howard Hughes Medical Institute, 16 Divinity Avenue, Cambridge, MA 02138, USA
| | - John E Sabatini
- Department of Molecular & Cellular Biology, Department of Organismic & Evolutionary Biology, Center for Brain Science, Museum of Comparative Zoology, Harvard University and the Howard Hughes Medical Institute, 16 Divinity Avenue, Cambridge, MA 02138, USA
| | - Sade McFadden
- Department of Molecular & Cellular Biology, Department of Organismic & Evolutionary Biology, Center for Brain Science, Museum of Comparative Zoology, Harvard University and the Howard Hughes Medical Institute, 16 Divinity Avenue, Cambridge, MA 02138, USA
| | - Anna K Lindholm
- Department of Evolutionary Biology & Environmental Studies, University of Zürich, Winterthurerstrasse, 190 8057 Zürich, Switzerland
| | - Hopi E Hoekstra
- Department of Molecular & Cellular Biology, Department of Organismic & Evolutionary Biology, Center for Brain Science, Museum of Comparative Zoology, Harvard University and the Howard Hughes Medical Institute, 16 Divinity Avenue, Cambridge, MA 02138, USA.
| |
Collapse
|
9
|
Colligan T, Irish K, Emlen DJ, Wheeler TJ. DISCO: A deep learning ensemble for uncertainty-aware segmentation of acoustic signals. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.01.24.525459. [PMID: 36747788 PMCID: PMC9900853 DOI: 10.1101/2023.01.24.525459] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 01/26/2023]
Abstract
Recordings of animal sounds enable a wide range of observational inquiries into animal communication, behavior, and diversity. Automated labeling of sound events in such recordings can improve both throughput and reproducibility of analysis. Here, we describe our software package for labeling sound elements in recordings of animal sounds and demonstrate its utility on recordings of beetle courtships and whale songs. The software, DISCO, computes sensible confidence estimates and produces labels with high precision and accuracy. In addition to the core labeling software, it provides a simple tool for labeling training data, and a visual system for analysis of resulting labels. DISCO is open-source and easy to install, it works with standard file formats, and it presents a low barrier of entry to use.
Collapse
Affiliation(s)
- Thomas Colligan
- Department of Pharmacy Practice & Science, University of Arizona, Tucson, AZ, USA,Department of Computer Science, University of Montana, Missoula, MT, USA
| | - Kayla Irish
- Department of Computer Science, University of Montana, Missoula, MT, USA,Department of Statistics, University of Washington, Seattle, WA, USA
| | - Douglas J. Emlen
- Division of Biological Sciences, University of Montana, Missoula, MT, USA
| | - Travis J. Wheeler
- Department of Pharmacy Practice & Science, University of Arizona, Tucson, AZ, USA,Department of Computer Science, University of Montana, Missoula, MT, USA
| |
Collapse
|
10
|
Jermyn AS, Stevenson DJ, Levitin DJ. 1/f laws found in non-human music. Sci Rep 2023; 13:1324. [PMID: 36694022 PMCID: PMC9873655 DOI: 10.1038/s41598-023-28444-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2022] [Accepted: 01/18/2023] [Indexed: 01/25/2023] Open
Abstract
A compelling question at the intersection of physics, neuroscience, and evolutionary biology concerns the extent to which the brains of various species evolved to encode regularities of the physical world. It would be parsimonious and adaptive, for example, for brains to evolve an innate understanding of gravity and the laws of motion, and to be able to detect, auditorily, those patterns of noises that ambulatory creatures make when moving about the world. One such physical regularity of the world is fractal structure, generally characterized by power-law correlations or 1/f β spectral distributions. Such laws are found broadly in nature and human artifacts, from noise in physical systems, to coastline topography (e.g., the Richardson effect), to neuronal spike patterns. These distributions have also been found to hold for the rhythm and power spectral density of a wide array of human music, suggesting that human music incorporates regularities of the physical world that our species evolved to recognize and produce. Here we show for the first time that 1/fβ laws also govern the spectral density of a wide range of animal vocalizations (music), from songbirds, to whales, to howling wolves. We discovered this 1/fβ power-law distribution in the vocalizations within all of the 17 diverse species examined. Our results demonstrate that such power laws are prevalent in the animal kingdom, evidence that their brains have evolved a sensitivity to them as an aid in processing sensory features of the natural world.
Collapse
Affiliation(s)
- Adam S Jermyn
- Kavli Institute for Theoretical Physics, University of California at Santa Barbara, Santa Barbara, CA, 93106, USA
| | - David J Stevenson
- Division of Geology and Planetary Science, California Institute of Technology, Pasadena, CA, 91125, USA
| | - Daniel J Levitin
- Department of Psychology, School of Computer Science, and Schulich School of Music, McGill University, Montreal, QC, H3A 1B1, Canada.
| |
Collapse
|
11
|
McGinn K, Kahl S, Peery MZ, Klinck H, Wood CM. Feature embeddings from the BirdNET algorithm provide insights into avian ecology. ECOL INFORM 2023. [DOI: 10.1016/j.ecoinf.2023.101995] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/19/2023]
|
12
|
Provost KL, Yang J, Carstens BC. The impacts of fine-tuning, phylogenetic distance, and sample size on big-data bioacoustics. PLoS One 2022; 17:e0278522. [PMID: 36477744 PMCID: PMC9728902 DOI: 10.1371/journal.pone.0278522] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/11/2022] [Accepted: 11/17/2022] [Indexed: 12/12/2022] Open
Abstract
Vocalizations in animals, particularly birds, are critically important behaviors that influence their reproductive fitness. While recordings of bioacoustic data have been captured and stored in collections for decades, the automated extraction of data from these recordings has only recently been facilitated by artificial intelligence methods. These have yet to be evaluated with respect to accuracy of different automation strategies and features. Here, we use a recently published machine learning framework to extract syllables from ten bird species ranging in their phylogenetic relatedness from 1 to 85 million years, to compare how phylogenetic relatedness influences accuracy. We also evaluate the utility of applying trained models to novel species. Our results indicate that model performance is best on conspecifics, with accuracy progressively decreasing as phylogenetic distance increases between taxa. However, we also find that the application of models trained on multiple distantly related species can improve the overall accuracy to levels near that of training and analyzing a model on the same species. When planning big-data bioacoustics studies, care must be taken in sample design to maximize sample size and minimize human labor without sacrificing accuracy.
Collapse
Affiliation(s)
- Kaiya L. Provost
- Department of Evolution, Ecology and Organismal Biology, The Ohio State University, Columbus, Ohio, United States of America
| | - Jiaying Yang
- Department of Evolution, Ecology and Organismal Biology, The Ohio State University, Columbus, Ohio, United States of America
| | - Bryan C. Carstens
- Department of Evolution, Ecology and Organismal Biology, The Ohio State University, Columbus, Ohio, United States of America
| |
Collapse
|
13
|
Escalante I, Kellner JR, Rodríguez RL, Desjonquères C. A female mimic signal type in the vibrational repertoire of male Enchenopa treehoppers. BEHAVIOUR 2022. [DOI: 10.1163/1568539x-bja10181] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
Abstract
Abstract
Animals vary in the complexity and size of the signal repertoire used in communication. Often, these behavioural repertoires include multiple signal types for the same process, for instance, courtship. In Enchenopa treehoppers (Hemiptera: Membracidae) mate-searching males produce plant-borne vibrational advertisement signals. Receptive females then respond to males with their own signals. Here we describe an additional signal type in the repertoire of these males. We collected nymphs in Wisconsin, USA, and recorded the spontaneous signalling bouts of adult males and duetting signals of females using laser vibrometry. Two-thirds of males produced the additional signal type, which differed in temporal and spectral features from the main male advertisement signals, whilst resembling female duetting signals in placement and acoustic features. Our findings suggest that this might be a female mimic signal. Overall, our findings highlight the diversity in the behavioural repertoire that animals may use for reproduction.
Collapse
Affiliation(s)
- Ignacio Escalante
- Behavioral and Molecular Ecology Group, Department of Biological Sciences, University of Wisconsin, 3209 N Maryland Avenue, Milwaukee, WI 53211, USA
| | - Jerald R. Kellner
- Behavioral and Molecular Ecology Group, Department of Biological Sciences, University of Wisconsin, 3209 N Maryland Avenue, Milwaukee, WI 53211, USA
| | - Rafael L. Rodríguez
- Behavioral and Molecular Ecology Group, Department of Biological Sciences, University of Wisconsin, 3209 N Maryland Avenue, Milwaukee, WI 53211, USA
| | - Camille Desjonquères
- Behavioral and Molecular Ecology Group, Department of Biological Sciences, University of Wisconsin, 3209 N Maryland Avenue, Milwaukee, WI 53211, USA
| |
Collapse
|
14
|
Introducing the Software CASE (Cluster and Analyze Sound Events) by Comparing Different Clustering Methods and Audio Transformation Techniques Using Animal Vocalizations. Animals (Basel) 2022; 12:ani12162020. [PMID: 36009611 PMCID: PMC9404437 DOI: 10.3390/ani12162020] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2022] [Revised: 07/28/2022] [Accepted: 08/04/2022] [Indexed: 11/17/2022] Open
Abstract
Simple Summary Unsupervised clustering algorithms are widely used in ecology and conservation to classify animal vocalizations, but also offer various advantages in basic research, contributing to the understanding of acoustic communication. Nevertheless, there are still some challenges to overcome. For instance, the quality of the clustering result depends on the audio transformation technique previously used to adjust the audio data. Moreover, it is difficult to verify the reliability of the clustering result. To analyze bioacoustic data using a clustering algorithm, it is, therefore, essential to select a reasonable algorithm from the many existing algorithms and prepare the recorded vocalizations so that the resulting values characterize a vocalization as accurately as possible. Frequency-modulated vocalizations, whose frequencies change over time, pose a particular problem. In this paper, we present the software CASE, which includes various clustering methods and provides an overview of their strengths and weaknesses concerning the classification of bioacoustic data. This software uses a multidimensional feature-extraction method to achieve better clustering results, especially for frequency-modulated vocalizations. Abstract Unsupervised clustering algorithms are widely used in ecology and conservation to classify animal sounds, but also offer several advantages in basic bioacoustics research. Consequently, it is important to overcome the existing challenges. A common practice is extracting the acoustic features of vocalizations one-dimensionally, only extracting an average value for a given feature for the entire vocalization. With frequency-modulated vocalizations, whose acoustic features can change over time, this can lead to insufficient characterization. Whether the necessary parameters have been set correctly and the obtained clustering result reliably classifies the vocalizations subsequently often remains unclear. The presented software, CASE, is intended to overcome these challenges. Established and new unsupervised clustering methods (community detection, affinity propagation, HDBSCAN, and fuzzy clustering) are tested in combination with various classifiers (k-nearest neighbor, dynamic time-warping, and cross-correlation) using differently transformed animal vocalizations. These methods are compared with predefined clusters to determine their strengths and weaknesses. In addition, a multidimensional data transformation procedure is presented that better represents the course of multiple acoustic features. The results suggest that, especially with frequency-modulated vocalizations, clustering is more applicable with multidimensional feature extraction compared with one-dimensional feature extraction. The characterization and clustering of vocalizations in multidimensional space offer great potential for future bioacoustic studies. The software CASE includes the developed method of multidimensional feature extraction, as well as all used clustering methods. It allows quickly applying several clustering algorithms to one data set to compare their results and to verify their reliability based on their consistency. Moreover, the software CASE determines the optimal values of most of the necessary parameters automatically. To take advantage of these benefits, the software CASE is provided for free download.
Collapse
|
15
|
Backhouse F, Dalziell AH, Magrath RD, Welbergen JA. Higher-order sequences of vocal mimicry performed by male Albert's lyrebirds are socially transmitted and enhance acoustic contrast. Proc Biol Sci 2022; 289:20212498. [PMID: 35259987 PMCID: PMC8905160 DOI: 10.1098/rspb.2021.2498] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
Abstract
Most studies of acoustic communication focus on short units of vocalization such as songs, yet these units are often hierarchically organized into higher-order sequences and, outside human language, little is known about the drivers of sequence structure. Here, we investigate the organization, transmission and function of vocal sequences sung by male Albert's lyrebirds (Menura alberti), a species renowned for vocal imitations of other species. We quantified the organization of mimetic units into sequences, and examined the extent to which these sequences are repeated within and between individuals and shared among populations. We found that individual males organized their mimetic units into stereotyped sequences. Sequence structures were shared within and to a lesser extent among populations, implying that sequences were socially transmitted. Across the entire species range, mimetic units were sung with immediate variety and a high acoustic contrast between consecutive units, suggesting that sequence structure is a means to enhance receiver perceptions of repertoire complexity. Our results provide evidence that higher-order sequences of vocalizations can be socially transmitted, and that the order of vocal units can be functionally significant. We conclude that, to fully understand vocal behaviours, we must study both the individual vocal units and their higher-order temporal organization.
Collapse
Affiliation(s)
- Fiona Backhouse
- The Hawkesbury Institute for the Environment, Western Sydney University, Richmond, NSW, Australia
| | - Anastasia H Dalziell
- The Hawkesbury Institute for the Environment, Western Sydney University, Richmond, NSW, Australia.,Centre for Sustainable Ecosystem Solutions, School of Earth, Atmospheric and Life Sciences, University of Wollongong, Wollongong, NSW, Australia.,Fuller Evolutionary Biology Program, Cornell Lab of Ornithology, Cornell University, Ithaca, NY, USA
| | - Robert D Magrath
- Research School of Biology, The Australian National University, Canberra, ACT, Australia
| | - Justin A Welbergen
- The Hawkesbury Institute for the Environment, Western Sydney University, Richmond, NSW, Australia
| |
Collapse
|
16
|
Rose EM, Prior NH, Ball GF. The singing question: re-conceptualizing birdsong. Biol Rev Camb Philos Soc 2021; 97:326-342. [PMID: 34609054 DOI: 10.1111/brv.12800] [Citation(s) in RCA: 19] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2021] [Revised: 09/21/2021] [Accepted: 09/22/2021] [Indexed: 01/31/2023]
Abstract
Birdsong has been the subject of broad research from a variety of sub-disciplines and has taught us much about the evolution, function, and mechanisms driving animal communication and cognition. Typically, birdsong refers to the specialized vocalizations produced by oscines. Historically, much of the research on birdsong was conducted in north temperate regions (specifically in Europe and North America) leading to multiple biases. Due to these historic biases these vocalizations are generally considered to be highly sexually dimorphic, heavily shaped by sexual selection and essential for courtship and territoriality. Song is also typically defined as a learned trait shaped by cultural evolution. Together, this framework focuses research specifically on males, particularly during the north temperate breeding season - reflecting and thereby reinforcing this framework. The physiological underpinnings of song often emphasize the role of the hypothalamic-pituitary-gonadal axis (associated with breeding changes) and the song control system (underlying vocal learning). Over the years there has been great debate over which features of song are essential to the definition of birdsong, which features apply broadly to contexts outside males in the north temperate region, and over the importance of having a definition at all. Importantly, the definitions we use can both guide and limit the progress of research. Here, we describe the history of these definitions, and how these definitions have directed and restricted research to focus on male song in sexually selected contexts. Additionally, we highlight the gaps in our scientific knowledge, especially with respect to the function and physiological mechanisms underlying song in females and in winter, as well as in non-seasonally breeding species. Furthermore, we highlight the problems with using complexity and learning as dichotomous variables to categorize songs and calls. Across species, no one characteristic of song - sexual dimorphism, seasonality, complexity, sexual selection, learning - consistently delineates song from other songbird vocal communication. We provide recommendations for next steps to build an inclusive information framework that will allow researchers to explore nuances in animal communication and promote comparative research. Specifically, we recommend that researchers should operationalize the axis of variation most relevant to their study/species by identifying their specific question and the variable(s) of focus (e.g. seasonality). Researchers should also identify the axis (axes) of variation (e.g. degree of control by testosterone) most relevant to their study and use language consistent with the question and axis (axes) of variation (e.g. control by testosterone in the seasonal vocal production of birds).
Collapse
Affiliation(s)
- Evangeline M Rose
- Department of Psychology, University of Maryland, College Park, 4094 Campus Dr., College Park, MD, 20742, U.S.A.,Program in Neuroscience and Cognitive Science, University of Maryland, College Park, 0219 Cole Student Activities Building, 4090 Union Drive, College Park, MD, 20742, U.S.A
| | - Nora H Prior
- Department of Psychology, University of Maryland, College Park, 4094 Campus Dr., College Park, MD, 20742, U.S.A.,Program in Neuroscience and Cognitive Science, University of Maryland, College Park, 0219 Cole Student Activities Building, 4090 Union Drive, College Park, MD, 20742, U.S.A
| | - Gregory F Ball
- Department of Psychology, University of Maryland, College Park, 4094 Campus Dr., College Park, MD, 20742, U.S.A.,Program in Neuroscience and Cognitive Science, University of Maryland, College Park, 0219 Cole Student Activities Building, 4090 Union Drive, College Park, MD, 20742, U.S.A
| |
Collapse
|