1
|
Wituszynski D, Hayford D, Poesel A, Apte G, Matthews SN, Martin J. Effects of a large-scale bioretention installation on the species composition of an urban bird community as determined by passive acoustic monitoring. ENVIRONMENTAL MONITORING AND ASSESSMENT 2024; 196:1037. [PMID: 39382737 DOI: 10.1007/s10661-024-13143-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/29/2023] [Accepted: 09/13/2024] [Indexed: 10/10/2024]
Abstract
As urbanization accelerates worldwide, municipalities are attempting to construct new green spaces within their borders. The perceived ecological value of these places is frequently tied to their ability to attract urban wildlife, such as birds, which can easily be observed and enjoyed. As one strategy, stormwater is now frequently managed with green infrastructure: planted areas that retain and treat stormwater rather than merely directing it to surface waters. While these practices have the potential to provide habitat for urban wildlife, the ecological effects of these systems are largely unknown. To assess whether one green infrastructure project increases habitat value, we used passive acoustic monitoring to survey urban bird communities in and near a large green infrastructure project in Columbus, Ohio (USA). Bird communities near bioretention cells (rain gardens) were compared to those at nearby lawns and remnant or restored natural areas. We found that recently installed bioretention cells tended to support more omnivores, lower-canopy foraging species, and species from a higher diversity of feeding guilds than did nearby lawn control sites. We were unable to detect effects of nearby bioretention installations on bird species richness at other sites. The observed differences in species richness were fairly small, and we urge caution when anticipating the habitat value of bioretention cells, at least for bird species. However, the results that we observed suggest that bioretention cells could have a more positive impact on bird communities in different contexts or using different design strategies. The bioretention cells surveyed in this study were small and only planted in grasses and forbs, potentially limiting their ability to offer complex habitat. They were also relatively young, and future work is needed to determine their long-term effect on avian communities and biodiversity of other taxa.
Collapse
Affiliation(s)
- David Wituszynski
- Department of Food Agricultural and Biological Engineering, The Ohio State University, OH, Columbus, USA.
- Department of Research and Development, Engineering Ministries International, Kajjansi, Uganda.
| | | | - Angelika Poesel
- Borror Laboratory of Bioacoustics, The Ohio State University, Columbus, USA, OH
| | - Gautam Apte
- School of Environment and Natural Resources, The Ohio State University, Columbus, USA, OH
| | - Stephen N Matthews
- School of Environment and Natural Resources, The Ohio State University, Columbus, USA, OH
| | - Jay Martin
- Department of Food Agricultural and Biological Engineering, The Ohio State University, OH, Columbus, USA
| |
Collapse
|
2
|
Michaud F, Sueur J, Le Cesne M, Haupert S. Unsupervised classification to improve the quality of a bird song recording dataset. ECOL INFORM 2022. [DOI: 10.1016/j.ecoinf.2022.101952] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
|
3
|
Sun Y, Yen S, Lin T. soundscape_IR: A source separation toolbox for exploring acoustic diversity in soundscapes. Methods Ecol Evol 2022. [DOI: 10.1111/2041-210x.13960] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
Affiliation(s)
- Yi‐Jen Sun
- Biodiversity Research Center Academia Sinica Taipei Taiwan (R.O.C)
| | - Shih‐Ching Yen
- Center for General Education National Tsing Hua University Hsinchu Taiwan (R.O.C)
| | - Tzu‐Hao Lin
- Biodiversity Research Center Academia Sinica Taipei Taiwan (R.O.C)
| |
Collapse
|
4
|
Symes LB, Kittelberger KD, Stone SM, Holmes RT, Jones JS, Castaneda Ruvalcaba IP, Webster MS, Ayres M. Analytical approaches for evaluating passive acoustic monitoring data: A case study of avian vocalizations. Ecol Evol 2022; 12:e8797. [PMID: 35475182 PMCID: PMC9022445 DOI: 10.1002/ece3.8797] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/06/2021] [Revised: 03/04/2022] [Accepted: 03/16/2022] [Indexed: 11/28/2022] Open
Abstract
The interface between field biology and technology is energizing the collection of vast quantities of environmental data. Passive acoustic monitoring, the use of unattended recording devices to capture environmental sound, is an example where technological advances have facilitated an influx of data that routinely exceeds the capacity for analysis. Computational advances, particularly the integration of machine learning approaches, will support data extraction efforts. However, the analysis and interpretation of these data will require parallel growth in conceptual and technical approaches for data analysis. Here, we use a large hand‐annotated dataset to showcase analysis approaches that will become increasingly useful as datasets grow and data extraction can be partially automated. We propose and demonstrate seven technical approaches for analyzing bioacoustic data. These include the following: (1) generating species lists and descriptions of vocal variation, (2) assessing how abiotic factors (e.g., rain and wind) impact vocalization rates, (3) testing for differences in community vocalization activity across sites and habitat types, (4) quantifying the phenology of vocal activity, (5) testing for spatiotemporal correlations in vocalizations within species, (6) among species, and (7) using rarefaction analysis to quantify diversity and optimize bioacoustic sampling. To demonstrate these approaches, we sampled in 2016 and 2018 and used hand annotations of 129,866 bird vocalizations from two forests in New Hampshire, USA, including sites in the Hubbard Brook Experiment Forest where bioacoustic data could be integrated with more than 50 years of observer‐based avian studies. Acoustic monitoring revealed differences in community patterns in vocalization activity between forests of different ages, as well as between nearby similar watersheds. Of numerous environmental variables that were evaluated, background noise was most clearly related to vocalization rates. The songbird community included one cluster of species where vocalization rates declined as ambient noise increased and another cluster where vocalization rates declined over the nesting season. In some common species, the number of vocalizations produced per day was correlated at scales of up to 15 km. Rarefaction analyses showed that adding sampling sites increased species detections more than adding sampling days. Although our analyses used hand‐annotated data, the methods will extend readily to large‐scale automated detection of vocalization events. Such data are likely to become increasingly available as autonomous recording units become more advanced, affordable, and power efficient. Passive acoustic monitoring with human or automated identification at the species level offers growing potential to complement observer‐based studies of avian ecology.
Collapse
Affiliation(s)
- Laurel B. Symes
- K. Lisa Yang Center for Conservation Bioacoustics Cornell Lab of Ornithology Cornell University Ithaca New York USA
- Department of Biological Sciences Dartmouth College Hanover New Hampshire USA
- Smithsonian Tropical Research Institute Panama City Republic of Panama
| | - Kyle D. Kittelberger
- Department of Biological Sciences Dartmouth College Hanover New Hampshire USA
- School of Biological Sciences University of Utah Salt Lake City Utah USA
| | - Sophia M. Stone
- Department of Biological Sciences Dartmouth College Hanover New Hampshire USA
| | - Richard T. Holmes
- Department of Biological Sciences Dartmouth College Hanover New Hampshire USA
| | - Jessica S. Jones
- Department of Biological Sciences Dartmouth College Hanover New Hampshire USA
| | | | - Michael S. Webster
- Macaulay Library Cornell Lab of Ornithology Cornell University Ithaca New York USA
| | - Matthew P. Ayres
- Department of Biological Sciences Dartmouth College Hanover New Hampshire USA
| |
Collapse
|
5
|
Diversity Monitoring of Coexisting Birds in Urban Forests by Integrating Spectrograms and Object-Based Image Analysis. FORESTS 2022. [DOI: 10.3390/f13020264] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
In the context of rapid urbanization, urban foresters are actively seeking management monitoring programs that address the challenges of urban biodiversity loss. Passive acoustic monitoring (PAM) has attracted attention because it allows for the collection of data passively, objectively, and continuously across large areas and for extended periods. However, it continues to be a difficult subject due to the massive amount of information that audio recordings contain. Most existing automated analysis methods have limitations in their application in urban areas, with unclear ecological relevance and efficacy. To better support urban forest biodiversity monitoring, we present a novel methodology for automatically extracting bird vocalizations from spectrograms of field audio recordings, integrating object-based classification. We applied this approach to acoustic data from an urban forest in Beijing and achieved an accuracy of 93.55% (±4.78%) in vocalization recognition while requiring less than ⅛ of the time needed for traditional inspection. The difference in efficiency would become more significant as the data size increases because object-based classification allows for batch processing of spectrograms. Using the extracted vocalizations, a series of acoustic and morphological features of bird-vocalization syllables (syllable feature metrics, SFMs) could be calculated to better quantify acoustic events and describe the soundscape. A significant correlation between the SFMs and biodiversity indices was found, with 57% of the variance in species richness, 41% in Shannon’s diversity index and 38% in Simpson’s diversity index being explained by SFMs. Therefore, our proposed method provides an effective complementary tool to existing automated methods for long-term urban forest biodiversity monitoring and conservation.
Collapse
|
6
|
Parsons MJG, Lin TH, Mooney TA, Erbe C, Juanes F, Lammers M, Li S, Linke S, Looby A, Nedelec SL, Van Opzeeland I, Radford C, Rice AN, Sayigh L, Stanley J, Urban E, Di Iorio L. Sounding the Call for a Global Library of Underwater Biological Sounds. Front Ecol Evol 2022. [DOI: 10.3389/fevo.2022.810156] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/12/2022] Open
Abstract
Aquatic environments encompass the world’s most extensive habitats, rich with sounds produced by a diversity of animals. Passive acoustic monitoring (PAM) is an increasingly accessible remote sensing technology that uses hydrophones to listen to the underwater world and represents an unprecedented, non-invasive method to monitor underwater environments. This information can assist in the delineation of biologically important areas via detection of sound-producing species or characterization of ecosystem type and condition, inferred from the acoustic properties of the local soundscape. At a time when worldwide biodiversity is in significant decline and underwater soundscapes are being altered as a result of anthropogenic impacts, there is a need to document, quantify, and understand biotic sound sources–potentially before they disappear. A significant step toward these goals is the development of a web-based, open-access platform that provides: (1) a reference library of known and unknown biological sound sources (by integrating and expanding existing libraries around the world); (2) a data repository portal for annotated and unannotated audio recordings of single sources and of soundscapes; (3) a training platform for artificial intelligence algorithms for signal detection and classification; and (4) a citizen science-based application for public users. Although individually, these resources are often met on regional and taxa-specific scales, many are not sustained and, collectively, an enduring global database with an integrated platform has not been realized. We discuss the benefits such a program can provide, previous calls for global data-sharing and reference libraries, and the challenges that need to be overcome to bring together bio- and ecoacousticians, bioinformaticians, propagation experts, web engineers, and signal processing specialists (e.g., artificial intelligence) with the necessary support and funding to build a sustainable and scalable platform that could address the needs of all contributors and stakeholders into the future.
Collapse
|
7
|
|
8
|
Rivera-Correa M, Ospina-L AM, Rojas-Montoya M, Venegas-Valencia K, Rueda-Solano LA, Gutiérrez-Cárdenas PDA, Vargas-Salinas F. Cantos de las ranas y los sapos de Colombia: estado actual del conocimiento y perspectivas de investigación en ecoacústica. NEOTROPICAL BIODIVERSITY 2021. [DOI: 10.1080/23766808.2021.1957651] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022] Open
Affiliation(s)
- Mauricio Rivera-Correa
- Grupo Herpetológico de Antioquia (GHA), Instituto de Biología, Universidad de Antioquia, Medellín, Colombia
- Semillero de Investigación en Biodiversidad de Anfibios (BIO), Seccional Oriente, Universidad de Antioquia, El Carmen de Viboral, Colombia
| | - Ana María Ospina-L
- Grupo de Investigación en Evolución, Ecología y Conservación (EECO), Programa de Biología, Universidad del Quindío, Armenia, Colombia
- Department of Biological Sciences, Purdue University, West Lafayette, IN, USA
| | - Maribel Rojas-Montoya
- Grupo de Investigación en Evolución, Ecología y Conservación (EECO), Programa de Biología, Universidad del Quindío, Armenia, Colombia
| | - Khristian Venegas-Valencia
- Grupo Herpetológico de Antioquia (GHA), Instituto de Biología, Universidad de Antioquia, Medellín, Colombia
- Semillero de Investigación en Biodiversidad de Anfibios (BIO), Seccional Oriente, Universidad de Antioquia, El Carmen de Viboral, Colombia
| | - Luis Alberto Rueda-Solano
- Grupo Biomis, Facultad de Ciencias, Universidad de los Andes, Bogotá, Colombia
- Grupo de Investigación en Biodiversidad y Ecología Aplicada, Facultad de Ciencias Básicas, Universidad del Magdalena, Santa Marta, Colombia
| | - Paul David Alfonso Gutiérrez-Cárdenas
- Grupo de Ecología y Diversidad de Anfibios y Reptiles (GEDAR), Facultad de Ciencias Exactas y Naturales, Universidad de Caldas, Manizales, Colombia
- Grupo de Investigación en Ecologia de Vertebrados Tropicais, Instituto de Biologia Roberto Alcantara Gomes, Universidade do Estado do Rio de Janeiro, RJ, Brasil
| | - Fernando Vargas-Salinas
- Grupo de Investigación en Evolución, Ecología y Conservación (EECO), Programa de Biología, Universidad del Quindío, Armenia, Colombia
| |
Collapse
|
9
|
Ärje J, Melvad C, Jeppesen MR, Madsen SA, Raitoharju J, Rasmussen MS, Iosifidis A, Tirronen V, Gabbouj M, Meissner K, Høye TT. Automatic image‐based identification and biomass estimation of invertebrates. Methods Ecol Evol 2020. [DOI: 10.1111/2041-210x.13428] [Citation(s) in RCA: 36] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
Affiliation(s)
- Johanna Ärje
- Department of Bioscience and Arctic Research Centre Aarhus University Ronde Denmark
- Unit of Computing Sciences Tampere University Tampere Finland
- Department of Mathematics and Statistics University of Jyvaskyla Jyvaskyla Finland
| | - Claus Melvad
- Aarhus School of Engineering and Arctic Research Centre Aarhus University Ronde Denmark
| | | | | | | | | | | | - Ville Tirronen
- Department of Information Technology University of Jyvaskyla Jyvaskyla Finland
| | - Moncef Gabbouj
- Unit of Computing Sciences Tampere University Tampere Finland
| | | | - Toke Thomas Høye
- Department of Bioscience and Arctic Research Centre Aarhus University Ronde Denmark
| |
Collapse
|
10
|
Stehle M, Lasseck M, Khorramshahi O, Sturm U. Evaluation of acoustic pattern recognition of nightingale (Luscinia megarhynchos) recordings by citizens. RESEARCH IDEAS AND OUTCOMES 2020. [DOI: 10.3897/rio.6.e50233] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022] Open
Abstract
Acoustic pattern recognition methods introduce new perspectives for species identification, biodiversity monitoring and data validation in citizen science but are rarely evaluated in real world scenarios. In this case study we analysed the performance of a machine learning algorithm for automated bird identification to reliably identify common nightingales (Luscinia megarhynchos) in field recordings taken by users of the smartphone app Naturblick. We found that the performance of the automated identification tool was overall robust in our selected recordings. Although most of the recordings had a relatively low confidence score, a large proportion of the recordings were identified correctly.
Collapse
|
11
|
de Oliveira AG, Ventura TM, Ganchev TD, Silva LNS, Marques MI, Schuchmann KL. Speeding up training of automated bird recognizers by data reduction of audio features. PeerJ 2020; 8:e8407. [PMID: 32025373 PMCID: PMC6991130 DOI: 10.7717/peerj.8407] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2019] [Accepted: 12/16/2019] [Indexed: 11/20/2022] Open
Abstract
Automated acoustic recognition of birds is considered an important technology in support of biodiversity monitoring and biodiversity conservation activities. These activities require processing large amounts of soundscape recordings. Typically, recordings are transformed to a number of acoustic features, and a machine learning method is used to build models and recognize the sound events of interest. The main problem is the scalability of data processing, either for developing models or for processing recordings made over long time periods. In those cases, the processing time and resources required might become prohibitive for the average user. To address this problem, we evaluated the applicability of three data reduction methods. These methods were applied to a series of acoustic feature vectors as an additional postprocessing step, which aims to reduce the computational demand during training. The experimental results obtained using Mel-frequency cepstral coefficients (MFCCs) and hidden Markov models (HMMs) support the finding that a reduction in training data by a factor of 10 does not significantly affect the recognition performance.
Collapse
Affiliation(s)
- Allan G de Oliveira
- Computational Bioacoustics Research Unit (CO.BRA), National Institute for Science and Technology in Wetlands (INAU), Universidade Federal de Mato Grosso, Cuiabá, Mato Grosso, Brazil.,Institute of Computing, Universidade Federal de Mato Grosso, Cuiabá, Mato Grosso, Brazil
| | - Thiago M Ventura
- Computational Bioacoustics Research Unit (CO.BRA), National Institute for Science and Technology in Wetlands (INAU), Universidade Federal de Mato Grosso, Cuiabá, Mato Grosso, Brazil.,Institute of Computing, Universidade Federal de Mato Grosso, Cuiabá, Mato Grosso, Brazil
| | - Todor D Ganchev
- Computational Bioacoustics Research Unit (CO.BRA), National Institute for Science and Technology in Wetlands (INAU), Universidade Federal de Mato Grosso, Cuiabá, Mato Grosso, Brazil.,Faculty of Computing and Automation, Technical University of Varna, Varna, Bulgaria
| | - Lucas N S Silva
- Computational Bioacoustics Research Unit (CO.BRA), National Institute for Science and Technology in Wetlands (INAU), Universidade Federal de Mato Grosso, Cuiabá, Mato Grosso, Brazil.,Institute of Computing, Universidade Federal de Mato Grosso, Cuiabá, Mato Grosso, Brazil
| | - Marinêz I Marques
- Computational Bioacoustics Research Unit (CO.BRA), National Institute for Science and Technology in Wetlands (INAU), Universidade Federal de Mato Grosso, Cuiabá, Mato Grosso, Brazil.,Institute of Bioscienses, Universidade Federal de Mato Grosso, Cuiabá, Mato Grosso, Brazil.,Postgraduate Program in Ecology and Biodiversity Conservation, Institute of Biosciences, Universidade Federal de Mato Grosso, Cuiabá, Mato Grosso, Brazil.,Postgraduate Program in Zoology, Institute of Biosciences, Universidade Federal de Mato Grosso, Cuiabá, Mato Grosso, Brazil
| | - Karl-L Schuchmann
- Computational Bioacoustics Research Unit (CO.BRA), National Institute for Science and Technology in Wetlands (INAU), Universidade Federal de Mato Grosso, Cuiabá, Mato Grosso, Brazil.,Institute of Bioscienses, Universidade Federal de Mato Grosso, Cuiabá, Mato Grosso, Brazil.,Postgraduate Program in Zoology, Institute of Biosciences, Universidade Federal de Mato Grosso, Cuiabá, Mato Grosso, Brazil.,Zoological Research Museum Alexander Koenig and University of Bonn, Bonn, Germany
| |
Collapse
|
12
|
Florentin J, Dutoit T, Verlinden O. Detection and identification of European woodpeckers with deep convolutional neural networks. ECOL INFORM 2020. [DOI: 10.1016/j.ecoinf.2019.101023] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
13
|
Ntalampiras SA, Ludovico LA, Presti G, Prato Previde EP, Battini M, Cannas S, Palestrini C, Mattiello S. Automatic Classification of Cat VocalizationsEmitted in Different Contexts. Animals (Basel) 2019; 9:ani9080543. [PMID: 31405018 PMCID: PMC6719916 DOI: 10.3390/ani9080543] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/04/2019] [Revised: 07/25/2019] [Accepted: 08/08/2019] [Indexed: 02/07/2023] Open
Abstract
Simple Summary Cat vocalizations are their basic means of communication. They are particularly important in assessing their welfare status since they are indicative of information associated with the environment they were produced, the animal’s emotional state, etc. As such, this work proposes a fully automatic framework with the ability to process such vocalizations and reveal the context in which they were produced. To this end, we used suitable audio signal processing and pattern recognition algorithms. We recorded vocalizations from Maine Coon and European Shorthair breeds emitted in three different contexts, namely waiting for food, isolation in unfamiliar environment, and brushing. The obtained results are excellent, rendering the proposed framework particularly useful towards a better understanding of the acoustic communication between humans and cats. Abstract Cats employ vocalizations for communicating information, thus their sounds can carry a wide range of meanings. Concerning vocalization, an aspect of increasing relevance directly connected with the welfare of such animals is its emotional interpretation and the recognition of the production context. To this end, this work presents a proof of concept facilitating the automatic analysis of cat vocalizations based on signal processing and pattern recognition techniques, aimed at demonstrating if the emission context can be identified by meowing vocalizations, even if recorded in sub-optimal conditions. We rely on a dataset including vocalizations of Maine Coon and European Shorthair breeds emitted in three different contexts: waiting for food, isolation in unfamiliar environment, and brushing. Towards capturing the emission context, we extract two sets of acoustic parameters, i.e., mel-frequency cepstral coefficients and temporal modulation features. Subsequently, these are modeled using a classification scheme based on a directed acyclic graph dividing the problem space. The experiments we conducted demonstrate the superiority of such a scheme over a series of generative and discriminative classification solutions. These results open up new perspectives for deepening our knowledge of acoustic communication between humans and cats and, in general, between humans and animals.
Collapse
Affiliation(s)
| | | | - Giorgio Presti
- Department of Computer Science, University of Milan, 20133 Milan, Italy.
| | | | - Monica Battini
- Department of Veterinary Medicine, University of Milan, 20133 Milan, Italy.
| | - Simona Cannas
- Department of Veterinary Medicine, University of Milan, 20133 Milan, Italy.
| | - Clara Palestrini
- Department of Veterinary Medicine, University of Milan, 20133 Milan, Italy.
| | - Silvana Mattiello
- Department of Veterinary Medicine, University of Milan, 20133 Milan, Italy.
| |
Collapse
|
14
|
Buxton RT, Lendrum PE, Crooks KR, Wittemyer G. Pairing camera traps and acoustic recorders to monitor the ecological impact of human disturbance. Glob Ecol Conserv 2018. [DOI: 10.1016/j.gecco.2018.e00493] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022] Open
|
15
|
Kholghi M, Phillips Y, Towsey M, Sitbon L, Roe P. Active learning for classifying long‐duration audio recordings of the environment. Methods Ecol Evol 2018. [DOI: 10.1111/2041-210x.13042] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
Affiliation(s)
| | | | - Michael Towsey
- Queensland University of Technology Brisbane Qld Australia
| | | | - Paul Roe
- Queensland University of Technology Brisbane Qld Australia
| |
Collapse
|
16
|
Ovaskainen O, Moliterno de Camargo U, Somervuo P. Animal Sound Identifier (ASI): software for automated identification of vocal animals. Ecol Lett 2018; 21:1244-1254. [PMID: 29938881 DOI: 10.1111/ele.13092] [Citation(s) in RCA: 20] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2018] [Revised: 04/01/2018] [Accepted: 05/03/2018] [Indexed: 01/22/2023]
Abstract
Automated audio recording offers a powerful tool for acoustic monitoring schemes of bird, bat, frog and other vocal organisms, but the lack of automated species identification methods has made it difficult to fully utilise such data. We developed Animal Sound Identifier (ASI), a MATLAB software that performs probabilistic classification of species occurrences from field recordings. Unlike most previous approaches, ASI locates training data directly from the field recordings and thus avoids the need of pre-defined reference libraries. We apply ASI to a case study on Amazonian birds, in which we classify the vocalisations of 14 species in 194 504 one-minute audio segments using in total two weeks of expert time to construct, parameterise, and validate the classification models. We compare the classification performance of ASI (with training templates extracted automatically from field data) to that of monitoR (with training templates extracted manually from the Xeno-Canto database), the results showing ASI to have substantially higher recall and precision rates.
Collapse
Affiliation(s)
- Otso Ovaskainen
- Faculty of Biological and Environmental Sciences, University of Helsinki, P.O. Box 65, Helsinki, FI-00014, Finland.,Centre for Biodiversity Dynamics, Department of Biology, Norwegian University of Science and Technology, N-7491, Trondheim, Norway
| | - Ulisses Moliterno de Camargo
- Faculty of Biological and Environmental Sciences, University of Helsinki, P.O. Box 65, Helsinki, FI-00014, Finland.,The Helsinki Lab of Ornithology, The Finnish Museum of Natural History, University of Helsinki, Helsinki, Finland
| | - Panu Somervuo
- Faculty of Biological and Environmental Sciences, University of Helsinki, P.O. Box 65, Helsinki, FI-00014, Finland
| |
Collapse
|
17
|
|
18
|
Khan AA, Qureshi IZ. Vocalizations of adult male Asian koels (Eudynamys scolopacea) in the breeding season. PLoS One 2017; 12:e0186604. [PMID: 29053720 PMCID: PMC5650150 DOI: 10.1371/journal.pone.0186604] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.9] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/23/2017] [Accepted: 10/04/2017] [Indexed: 11/18/2022] Open
Abstract
Defining the vocal repertoire provides a basis for understanding the role of acoustic signals in sexual and social interactions of an animal. The Asian koel (Eudynamys scolopacea) is a migratory bird which spends its summer breeding season in the plains of Pakistan. The bird is typically wary and secretive but produces loud and distinct calls, making it easily detected when unseen. Like the other birds in the wild, presumably Asian koels use their calls for social cohesion and coordination of different behaviors. To date, the description of vocal repertoire of the male Asian koel has been lacking. Presently we analyzed and described for the first time the vocalizations of the adult male Asian koel, recorded in two consecutive breeding seasons. Using 10 call parameters, we categorized the vocalization type into six different categories on the basis of spectrogram and statistical analyses, namely the; “type 1 cooee call”, “type 2 cooee call”, “type 1 coegh call”, “type 2 coegh call”, “wurroo call” and “coe call”. These names were assigned not on the basis of functional analysis and were therefore onomatopoeic. Stepwise cross validated discriminant function analysis classified the vocalization correctly (100%) into the predicted vocal categories that we initially classified on the basis of spectrographic examination. Our findings enrich the biological knowledge about vocalizations of the adult male Asian koel and provide a foundation for future acoustic monitoring of the species, as well as for comparative studies with vocalizations of other bird species of the cuckoo family. Further studies on the vocalizations of the Asian koel are required to unravel their functions in sexual selection and individual recognition.
Collapse
Affiliation(s)
- Abdul Aziz Khan
- Department of Animal Sciences, Faculty of Biological Sciences, Quaid-i-Azam University, Islamabad, Pakistan
| | - Irfan Zia Qureshi
- Department of Animal Sciences, Faculty of Biological Sciences, Quaid-i-Azam University, Islamabad, Pakistan
- * E-mail:
| |
Collapse
|
19
|
de Camargo UM, Somervuo P, Ovaskainen O. PROTAX-Sound: A probabilistic framework for automated animal sound identification. PLoS One 2017; 12:e0184048. [PMID: 28863178 PMCID: PMC5581177 DOI: 10.1371/journal.pone.0184048] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2017] [Accepted: 08/17/2017] [Indexed: 11/19/2022] Open
Abstract
Autonomous audio recording is stimulating new field in bioacoustics, with a great promise for conducting cost-effective species surveys. One major current challenge is the lack of reliable classifiers capable of multi-species identification. We present PROTAX-Sound, a statistical framework to perform probabilistic classification of animal sounds. PROTAX-Sound is based on a multinomial regression model, and it can utilize as predictors any kind of sound features or classifications produced by other existing algorithms. PROTAX-Sound combines audio and image processing techniques to scan environmental audio files. It identifies regions of interest (a segment of the audio file that contains a vocalization to be classified), extracts acoustic features from them and compares with samples in a reference database. The output of PROTAX-Sound is the probabilistic classification of each vocalization, including the possibility that it represents species not present in the reference database. We demonstrate the performance of PROTAX-Sound by classifying audio from a species-rich case study of tropical birds. The best performing classifier achieved 68% classification accuracy for 200 bird species. PROTAX-Sound improves the classification power of current techniques by combining information from multiple classifiers in a manner that yields calibrated classification probabilities.
Collapse
Affiliation(s)
| | - Panu Somervuo
- Department of Biosciences, University of Helsinki, Helsinki, Finland
| | - Otso Ovaskainen
- Department of Biosciences, University of Helsinki, Helsinki, Finland
- Department of Biology, Norwegian University of Science and Technology, Trondheim, Norway
| |
Collapse
|
20
|
|
21
|
An FPGA-Based WASN for Remote Real-Time Monitoring of Endangered Species: A Case Study on the Birdsong Recognition of Botaurus stellaris. SENSORS 2017; 17:s17061331. [PMID: 28594373 PMCID: PMC5492858 DOI: 10.3390/s17061331] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/31/2017] [Revised: 06/01/2017] [Accepted: 06/06/2017] [Indexed: 11/19/2022]
Abstract
Fast environmental variations due to climate change can cause mass decline or even extinctions of species, having a dramatic impact on the future of biodiversity. During the last decade, different approaches have been proposed to track and monitor endangered species, generally based on costly semi-automatic systems that require human supervision adding limitations in coverage and time. However, the recent emergence of Wireless Acoustic Sensor Networks (WASN) has allowed non-intrusive remote monitoring of endangered species in real time through the automatic identification of the sound they emit. In this work, an FPGA-based WASN centralized architecture is proposed and validated on a simulated operation environment. The feasibility of the architecture is evaluated in a case study designed to detect the threatened Botaurus stellaris among other 19 cohabiting birds species in The Parc Natural dels Aiguamolls de l’Empordà, showing an averaged recognition accuracy of 91% over 2h 55’ of representative data. The FPGA-based feature extraction implementation allows the system to process data from 30 acoustic sensors in real time with an affordable cost. Finally, several open questions derived from this research are discussed to be considered for future works.
Collapse
|
22
|
|
23
|
Zhao Z, Zhang SH, Xu ZY, Bellisario K, Dai NH, Omrani H, Pijanowski BC. Automated bird acoustic event detection and robust species classification. ECOL INFORM 2017. [DOI: 10.1016/j.ecoinf.2017.04.003] [Citation(s) in RCA: 57] [Impact Index Per Article: 8.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
24
|
Jahn O, Ganchev TD, Marques MI, Schuchmann KL. Automated Sound Recognition Provides Insights into the Behavioral Ecology of a Tropical Bird. PLoS One 2017; 12:e0169041. [PMID: 28085893 PMCID: PMC5235375 DOI: 10.1371/journal.pone.0169041] [Citation(s) in RCA: 21] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2016] [Accepted: 12/09/2016] [Indexed: 11/28/2022] Open
Abstract
Computer-assisted species recognition facilitates the analysis of relevant biological information in continuous audio recordings. In the present study, we assess the suitability of this approach for determining distinct life-cycle phases of the Southern Lapwing Vanellus chilensis lampronotus based on adult vocal activity. For this purpose we use passive 14-min and 30-min soundscape recordings (n = 33 201) collected in 24/7 mode between November 2012 and October 2013 in Brazil’s Pantanal wetlands. Time-stamped detections of V. chilensis call events (n = 62 292) were obtained with a species-specific sound recognizer. We demonstrate that the breeding season fell in a three-month period from mid-May to early August 2013, between the end of the flood cycle and the height of the dry season. Several phases of the lapwing’s life history were identified with presumed error margins of a few days: pre-breeding, territory establishment and egg-laying, incubation, hatching, parental defense of chicks, and post-breeding. Diurnal time budgets confirm high acoustic activity levels during midday hours in June and July, indicative of adults defending young. By August, activity patterns had reverted to nonbreeding mode, with peaks around dawn and dusk and low call frequency during midday heat. We assess the current technological limitations of the V. chilensis recognizer through a comprehensive performance assessment and scrutinize the usefulness of automated acoustic recognizers in studies on the distribution pattern, ecology, life history, and conservation status of sound-producing animal species.
Collapse
Affiliation(s)
- Olaf Jahn
- National Institute for Science and Technology in Wetlands (INAU), Science without Borders Program, Federal University of Mato Grosso (UFMT), Cuiabá, Mato Grosso, Brazil
- Zoological Research Museum A. Koenig (ZFMK), Bonn, North Rhine-Westphalia, Germany
- * E-mail:
| | - Todor D. Ganchev
- National Institute for Science and Technology in Wetlands (INAU), Science without Borders Program, Federal University of Mato Grosso (UFMT), Cuiabá, Mato Grosso, Brazil
- Department of Computer Science and Engineering, Technical University of Varna, Varna, Varna, Bulgaria
| | - Marinez I. Marques
- National Institute for Science and Technology in Wetlands (INAU), Science without Borders Program, Federal University of Mato Grosso (UFMT), Cuiabá, Mato Grosso, Brazil
- Institute of Biosciences, UFMT, Cuiabá, Mato Grosso, Brazil
| | - Karl-L. Schuchmann
- National Institute for Science and Technology in Wetlands (INAU), Science without Borders Program, Federal University of Mato Grosso (UFMT), Cuiabá, Mato Grosso, Brazil
- Zoological Research Museum A. Koenig (ZFMK), Bonn, North Rhine-Westphalia, Germany
- Institute of Biosciences, UFMT, Cuiabá, Mato Grosso, Brazil
- University of Bonn, Bonn, North Rhine-Westphalia, Germany
| |
Collapse
|
25
|
Florentin J, Dutoit T, Verlinden O. Identification of European woodpecker species in audio recordings from their drumming rolls. ECOL INFORM 2016. [DOI: 10.1016/j.ecoinf.2016.08.006] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
26
|
Ruiz-Muñoz JF, Castellanos-Dominguez G, Orozco-Alzate M. Enhancing the dissimilarity-based classification of birdsong recordings. ECOL INFORM 2016. [DOI: 10.1016/j.ecoinf.2016.04.001] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
27
|
Salimi N, Loh KH, Kaur Dhillon S, Chong VC. Fully-automated identification of fish species based on otolith contour: using short-time Fourier transform and discriminant analysis (STFT-DA). PeerJ 2016; 4:e1664. [PMID: 26925315 PMCID: PMC4768690 DOI: 10.7717/peerj.1664] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2015] [Accepted: 01/16/2016] [Indexed: 11/20/2022] Open
Abstract
Background. Fish species may be identified based on their unique otolith shape or contour. Several pattern recognition methods have been proposed to classify fish species through morphological features of the otolith contours. However, there has been no fully-automated species identification model with the accuracy higher than 80%. The purpose of the current study is to develop a fully-automated model, based on the otolith contours, to identify the fish species with the high classification accuracy. Methods. Images of the right sagittal otoliths of 14 fish species from three families namely Sciaenidae, Ariidae, and Engraulidae were used to develop the proposed identification model. Short-time Fourier transform (STFT) was used, for the first time in the area of otolith shape analysis, to extract important features of the otolith contours. Discriminant Analysis (DA), as a classification technique, was used to train and test the model based on the extracted features. Results. Performance of the model was demonstrated using species from three families separately, as well as all species combined. Overall classification accuracy of the model was greater than 90% for all cases. In addition, effects of STFT variables on the performance of the identification model were explored in this study. Conclusions. Short-time Fourier transform could determine important features of the otolith outlines. The fully-automated model proposed in this study (STFT-DA) could predict species of an unknown specimen with acceptable identification accuracy. The model codes can be accessed at http://mybiodiversityontologies.um.edu.my/Otolith/ and https://peerj.com/preprints/1517/. The current model has flexibility to be used for more species and families in future studies.
Collapse
Affiliation(s)
- Nima Salimi
- Institute of Biological Sciences, University of Malaya , Kuala Lumpur , Malaysia
| | - Kar Hoe Loh
- Institute of Ocean & Earth Sciences, University of Malaya , Kuala Lumpur , Malaysia
| | | | - Ving Ching Chong
- Institute of Biological Sciences, University of Malaya, Kuala Lumpur, Malaysia; Institute of Ocean & Earth Sciences, University of Malaya, Kuala Lumpur, Malaysia
| |
Collapse
|
28
|
Ulloa JS, Gasc A, Gaucher P, Aubin T, Réjou-Méchain M, Sueur J. Screening large audio datasets to determine the time and space distribution of Screaming Piha birds in a tropical forest. ECOL INFORM 2016. [DOI: 10.1016/j.ecoinf.2015.11.012] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|
29
|
Potamitis I. Unsupervised dictionary extraction of bird vocalisations and new tools on assessing and visualising bird activity. ECOL INFORM 2015. [DOI: 10.1016/j.ecoinf.2015.01.002] [Citation(s) in RCA: 16] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
30
|
Lasseck M. Towards Automatic Large-Scale Identification of Birds in Audio Recordings. LECTURE NOTES IN COMPUTER SCIENCE 2015:364-375. [DOI: 10.1007/978-3-319-24027-5_39] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/02/2023]
|
31
|
Stowell D, Plumbley MD. Automatic large-scale classification of bird sounds is strongly improved by unsupervised feature learning. PeerJ 2014; 2:e488. [PMID: 25083350 PMCID: PMC4106198 DOI: 10.7717/peerj.488] [Citation(s) in RCA: 72] [Impact Index Per Article: 7.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2014] [Accepted: 06/26/2014] [Indexed: 11/20/2022] Open
Abstract
Automatic species classification of birds from their sound is a computational tool of increasing importance in ecology, conservation monitoring and vocal communication studies. To make classification useful in practice, it is crucial to improve its accuracy while ensuring that it can run at big data scales. Many approaches use acoustic measures based on spectrogram-type data, such as the Mel-frequency cepstral coefficient (MFCC) features which represent a manually-designed summary of spectral information. However, recent work in machine learning has demonstrated that features learnt automatically from data can often outperform manually-designed feature transforms. Feature learning can be performed at large scale and “unsupervised”, meaning it requires no manual data labelling, yet it can improve performance on “supervised” tasks such as classification. In this work we introduce a technique for feature learning from large volumes of bird sound recordings, inspired by techniques that have proven useful in other domains. We experimentally compare twelve different feature representations derived from the Mel spectrum (of which six use this technique), using four large and diverse databases of bird vocalisations, classified using a random forest classifier. We demonstrate that in our classification tasks, MFCCs can often lead to worse performance than the raw Mel spectral data from which they are derived. Conversely, we demonstrate that unsupervised feature learning provides a substantial boost over MFCCs and Mel spectra without adding computational complexity after the model has been trained. The boost is particularly notable for single-label classification tasks at large scale. The spectro-temporal activations learned through our procedure resemble spectro-temporal receptive fields calculated from avian primary auditory forebrain. However, for one of our datasets, which contains substantial audio data but few annotations, increased performance is not discernible. We study the interaction between dataset characteristics and choice of feature representation through further empirical analysis.
Collapse
Affiliation(s)
- Dan Stowell
- Centre for Digital Music, Queen Mary University of London , UK
| | - Mark D Plumbley
- Centre for Digital Music, Queen Mary University of London , UK
| |
Collapse
|