Loading...
Thumbnail Image
Publication

Bioacoustic and Ecoacoustic Data in Audiovisual Core

Citations
Altmetric:
Advisors
Editors
Other Contributors
Affiliation
EPub Date
Issue Date
2024-11-19
Submitted Date
2024-11-18
Subject Terms
bioacoustics
ecoacoustics
Collections
Research Projects
Organizational Units
Journal Issue
Other Titles
Abstract
Audiovisual Core (Audiovisual Core Maintenance Group 2023), is the TDWG standard for metadata related to biodiversity multimedia. The Audiovisual Core Maintenance Group has been working to expand the standard to provide the terms necessary for handling sound recordings. Audiovisual Core can now handle acoustic metadata related to biodiversity from single species (bioacoustics) to the ecosystem scale (ecoacoustics). Bioacoustics The Natural History Museum, London has a significant collection of recorded insect sounds (Ragge and Reynolds 1998) that are often directly linked to museum specimens (Fig. 1). The sound collection has previously been digitised and made available electronically through the BioAcoustica project (Baker et al. 2015). The BioAcoustica platform allows for annotation of audio files with tags including "Call" for deliberate sound made by an organism, "Voice Introduction" for metadata, and "Extraneous Noise." These boundaries are defined by their start and end times (in seconds) relative to the start of the file (Fig. 2). Ecoacoustics Ecoacoustics deals with the sounds present within an entire soundscape or ecosystem. The calls of individual species form the biological part of the soundscape (biophony) alongside sounds produced by non-living natural sources (geophony) and humans (anthropophony). Individual components are often defined by date and time boundaries, and sometimes by upper and lower frequency limits (Fig. 3). Regions of Interest The recently added concept of a "Region of Interest" (ROI) allows for the annotation of sound files, identifying multiple regions within a single recording with time and/or frequency bounds. However, the vocabulary of ROI*1 is not just intended for sounds. Equivalent terms also allow for regions to be specified with images and videos. The use of well-defined annotations has the potential to generate large amounts of training data for machine learning models and provide a standard for generating observation records from these models (e.g., BirdNet, see Kahl et al. 2021), which can be verified by linking them to audio segments within a much larger recording. The development of a metadata standard for regions of interest has several interesting possibilities, including linking multiple observation records to a single soundscape recording (the recording acts similarly to a voucher specimen) and aggregating regions across multiple datasets to create larger corpora for training machine learning models.
Citation
Baker E (2024) Bioacoustic and Ecoacoustic Data in Audiovisual Core. Biodiversity Information Science and Standards 8: e142073. https://doi.org/10.3897/biss.8.142073
Research Unit
PubMed ID
PubMed Central ID
Embedded videos
Type
Conference Proceedings
Item Description
Copyright © Baker E. This is an open access article distributed under the terms of the Creative Commons Attribution License (CC BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. The attached file is the published version of the article.
NHM Repository
Series/Report no.
ISSN
EISSN
2535-0897
ISBN
ISMN
GovDoc
Test Link
Sponsors