Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (40)

Search Parameters:
Keywords = bioacoustic monitoring

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
13 pages, 1305 KiB  
Article
Fine-Tuning BirdNET for the Automatic Ecoacoustic Monitoring of Bird Species in the Italian Alpine Forests
by Giacomo Schiavo, Alessia Portaccio and Alberto Testolin
Information 2025, 16(8), 628; https://doi.org/10.3390/info16080628 - 23 Jul 2025
Viewed by 301
Abstract
The ongoing decline in global biodiversity constitutes a critical challenge for environmental science, necessitating the prompt development of effective monitoring frameworks and conservation protocols to safeguard the structure and function of natural ecosystems. Recent progress in ecoacoustic monitoring, supported by advances in artificial [...] Read more.
The ongoing decline in global biodiversity constitutes a critical challenge for environmental science, necessitating the prompt development of effective monitoring frameworks and conservation protocols to safeguard the structure and function of natural ecosystems. Recent progress in ecoacoustic monitoring, supported by advances in artificial intelligence, might finally offer scalable tools for systematic biodiversity assessment. In this study, we evaluate the performance of BirdNET, a state-of-the-art deep learning model for avian sound recognition, in the context of selected bird species characteristic of the Italian Alpine region. To this end, we assemble a comprehensive, manually annotated audio dataset targeting key regional species, and we investigate a variety of strategies for model adaptation, including fine-tuning with data augmentation techniques to enhance recognition under challenging recording conditions. As a baseline, we also develop and evaluate a simple Convolutional Neural Network (CNN) trained exclusively on our domain-specific dataset. Our findings indicate that BirdNET performance can be greatly improved by fine-tuning the pre-trained network with data collected within the specific regional soundscape, outperforming both the original BirdNET and the baseline CNN by a significant margin. These findings underscore the importance of environmental adaptation and data variability for the development of automated ecoacoustic monitoring devices while highlighting the potential of deep learning methods in supporting conservation efforts and informing soundscape management in protected areas. Full article
(This article belongs to the Special Issue Signal Processing Based on Machine Learning Techniques)
Show Figures

Graphical abstract

28 pages, 1634 KiB  
Review
AI-Powered Vocalization Analysis in Poultry: Systematic Review of Health, Behavior, and Welfare Monitoring
by Venkatraman Manikandan and Suresh Neethirajan
Sensors 2025, 25(13), 4058; https://doi.org/10.3390/s25134058 - 29 Jun 2025
Viewed by 989
Abstract
Artificial intelligence and bioacoustics represent a paradigm shift in non-invasive poultry welfare monitoring through advanced vocalization analysis. This comprehensive systematic review critically examines the transformative evolution from traditional acoustic feature extraction—including Mel-Frequency Cepstral Coefficients (MFCCs), spectral entropy, and spectrograms—to cutting-edge deep learning architectures [...] Read more.
Artificial intelligence and bioacoustics represent a paradigm shift in non-invasive poultry welfare monitoring through advanced vocalization analysis. This comprehensive systematic review critically examines the transformative evolution from traditional acoustic feature extraction—including Mel-Frequency Cepstral Coefficients (MFCCs), spectral entropy, and spectrograms—to cutting-edge deep learning architectures encompassing Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM) networks, attention mechanisms, and groundbreaking self-supervised models such as wav2vec2 and Whisper. The investigation reveals compelling evidence for edge computing deployment via TinyML frameworks, addressing critical scalability challenges in commercial poultry environments characterized by acoustic complexity and computational constraints. Advanced applications spanning emotion recognition, disease detection, and behavioral phenotyping demonstrate unprecedented potential for real-time welfare assessment. Through rigorous bibliometric co-occurrence mapping and thematic clustering analysis, this review exposes persistent methodological bottlenecks: dataset standardization deficits, evaluation protocol inconsistencies, and algorithmic interpretability limitations. Critical knowledge gaps emerge in cross-species domain generalization and contextual acoustic adaptation, demanding urgent research prioritization. The findings underscore explainable AI integration as essential for establishing stakeholder trust and regulatory compliance in automated welfare monitoring systems. This synthesis positions acoustic AI as a cornerstone technology enabling ethical, transparent, and scientifically robust precision livestock farming, bridging computational innovation with biological relevance for sustainable poultry production systems. Future research directions emphasize multi-modal sensor integration, standardized evaluation frameworks, and domain-adaptive models capable of generalizing across diverse poultry breeds, housing conditions, and environmental contexts while maintaining interpretability for practical farm deployment. Full article
(This article belongs to the Special Issue Feature Papers in Smart Agriculture 2025)
Show Figures

Figure 1

25 pages, 6794 KiB  
Article
Animal-Borne Adaptive Acoustic Monitoring
by Devin Jean, Jesse Turner, Will Hedgecock, György Kalmár, George Wittemyer and Ákos Lédeczi
J. Sens. Actuator Netw. 2025, 14(4), 66; https://doi.org/10.3390/jsan14040066 - 24 Jun 2025
Viewed by 792
Abstract
Animal-borne acoustic sensors provide valuable insights into wildlife behavior and environments but face significant power and storage constraints that limit deployment duration. We present a novel adaptive acoustic monitoring system designed for long-term, real-time observation of wildlife. Our approach combines low-power hardware, configurable [...] Read more.
Animal-borne acoustic sensors provide valuable insights into wildlife behavior and environments but face significant power and storage constraints that limit deployment duration. We present a novel adaptive acoustic monitoring system designed for long-term, real-time observation of wildlife. Our approach combines low-power hardware, configurable firmware, and an unsupervised machine learning algorithm that intelligently filters acoustic data to prioritize novel or rare sounds while reducing redundant storage. The system employs a variational autoencoder to project audio features into a low-dimensional space, followed by adaptive clustering to identify events of interest. Simulation results demonstrate the system’s ability to normalize the collection of acoustic events across varying abundance levels, with rare events retained at rates of 80–85% while frequent sounds are reduced to 3–10% retention. Initial field deployments on caribou, African elephants, and bighorn sheep show promising application across diverse species and ecological contexts. Power consumption analysis indicates the need for additional optimization to achieve multi-month deployments. This technology enables the creation of novel wilderness datasets while addressing the limitations of traditional static acoustic monitoring approaches, offering new possibilities for wildlife research, ecosystem monitoring, and conservation efforts. Full article
Show Figures

Figure 1

14 pages, 7505 KiB  
Article
Audio-Based Automatic Giant Panda Behavior Recognition Using Competitive Fusion Learning
by Yuancheng Li, Yong Luo, Qijun Zhao, Mingchun Zhang, Yue Yang and Desheng Li
Sensors 2025, 25(13), 3878; https://doi.org/10.3390/s25133878 - 21 Jun 2025
Viewed by 755
Abstract
Automated giant panda (Ailuropoda melanoleuca) behavior recognition (GPBR) systems are highly beneficial for efficiently monitoring giant pandas in wildlife conservation missions. While video-based behavior recognition attracts a lot of attention, few studies have focused on audio-based methods. In this paper, we [...] Read more.
Automated giant panda (Ailuropoda melanoleuca) behavior recognition (GPBR) systems are highly beneficial for efficiently monitoring giant pandas in wildlife conservation missions. While video-based behavior recognition attracts a lot of attention, few studies have focused on audio-based methods. In this paper, we propose the exploitation of the audio data recorded by collar-mounted devices on giant pandas for the purpose of GPBR. We construct a new benchmark audio dataset of giant pandas named abPanda-5 for GPBR, which consists of 18,930 samples from five giant panda individuals with five main behaviors. To fully explore the bioacoustic features, we propose an audio-based method for automatic GPBR using competitive fusion learning. The method improves behavior recognition accuracy and robustness, without additional computational overhead in the inference stage. Experiments performed on the abPanda-5 dataset demonstrate the feasibility and effectiveness of our proposed method. Full article
Show Figures

Figure 1

25 pages, 5011 KiB  
Review
Mapping Soundscape Research: Authors, Institutions, and Collaboration Networks
by Andy W. L. Chung and Wai Ming To
Acoustics 2025, 7(2), 38; https://doi.org/10.3390/acoustics7020038 - 19 Jun 2025
Viewed by 961
Abstract
Soundscape is the sonic environment that a living being, like a human or animal, experiences in a certain setting. It affects how a space functions and how the being perceives its quality. Consequently, the soundscape is crucial in ecosystems globally. In recent decades, [...] Read more.
Soundscape is the sonic environment that a living being, like a human or animal, experiences in a certain setting. It affects how a space functions and how the being perceives its quality. Consequently, the soundscape is crucial in ecosystems globally. In recent decades, researchers have explored soundscapes using various methodologies across different disciplines. This study aims to provide a brief overview of the soundscape research history, pinpoint key authors, institutions, and collaboration networks, and identify trends and main themes through a bibliometric analysis. A search in the Scopus database on 26 February 2025 found 5825 articles, reviews, and conference papers on soundscape published from 1985 to 2024. The analysis indicated a significant increase in soundscape publications, rising from 1 in 1985 to 19 in 2002, and reaching 586 in 2024. J. Kang was the most prolific author with 265 publications, while University College London emerged as the most productive institution. Co-citation analysis revealed three research groups: one focused on urban soundscapes, another on aquatic soundscapes, and a third on soundscapes in landscape ecology. The keyword co-occurrence analysis identified three themes: “soundscape(s), acoustic environment, and urban planning”, “noise, animal(s), bioacoustics, biodiversity, passive acoustic monitoring, fish, and bird(s)”, and “human(s), sound, perception, and physiology”. Full article
Show Figures

Figure 1

18 pages, 4885 KiB  
Article
Decoding Poultry Welfare from Sound—A Machine Learning Framework for Non-Invasive Acoustic Monitoring
by Venkatraman Manikandan and Suresh Neethirajan
Sensors 2025, 25(9), 2912; https://doi.org/10.3390/s25092912 - 5 May 2025
Cited by 2 | Viewed by 1436
Abstract
Acoustic monitoring presents a promising, non-invasive modality for assessing animal welfare in precision livestock farming. In poultry, vocalizations encode biologically relevant cues linked to health status, behavioral states, and environmental stress. This study proposes an integrated analytical framework that combines signal-level statistical analysis [...] Read more.
Acoustic monitoring presents a promising, non-invasive modality for assessing animal welfare in precision livestock farming. In poultry, vocalizations encode biologically relevant cues linked to health status, behavioral states, and environmental stress. This study proposes an integrated analytical framework that combines signal-level statistical analysis with machine learning and deep learning classifiers to interpret chicken vocalizations in a welfare assessment context. The framework was evaluated using three complementary datasets encompassing health-related vocalizations, behavioral call types, and stress-induced acoustic responses. The pipeline employs a multistage process comprising high-fidelity signal acquisition, feature extraction (e.g., mel-frequency cepstral coefficients, spectral contrast, zero-crossing rate), and classification using models including Random Forest, HistGradientBoosting, CatBoost, TabNet, and LSTM. Feature importance analysis and statistical tests (e.g., t-tests, correlation metrics) confirmed that specific MFCC bands and spectral descriptors were significantly associated with welfare indicators. LSTM-based temporal modeling revealed distinct acoustic trajectories under visual and auditory stress, supporting the presence of habituation and stressor-specific vocal adaptations over time. Model performance, validated through stratified cross-validation and multiple statistical metrics (e.g., F1-score, Matthews correlation coefficient), demonstrated high classification accuracy and generalizability. Importantly, the approach emphasizes model interpretability, facilitating alignment with known physiological and behavioral processes in poultry. The findings underscore the potential of acoustic sensing and interpretable AI as scalable, biologically grounded tools for real-time poultry welfare monitoring, contributing to the advancement of sustainable and ethical livestock production systems. Full article
(This article belongs to the Special Issue Sensors in 2025)
Show Figures

Figure 1

38 pages, 2098 KiB  
Review
Rethinking Poultry Welfare—Integrating Behavioral Science and Digital Innovations for Enhanced Animal Well-Being
by Suresh Neethirajan
Poultry 2025, 4(2), 20; https://doi.org/10.3390/poultry4020020 - 29 Apr 2025
Viewed by 2268
Abstract
The relentless drive to meet global demand for poultry products has pushed for rapid intensification in chicken farming, dramatically boosting efficiency and yield. Yet, these gains have exposed a host of complex welfare challenges that have prompted scientific scrutiny and ethical reflection. In [...] Read more.
The relentless drive to meet global demand for poultry products has pushed for rapid intensification in chicken farming, dramatically boosting efficiency and yield. Yet, these gains have exposed a host of complex welfare challenges that have prompted scientific scrutiny and ethical reflection. In this review, I critically evaluate recent innovations aimed at mitigating such concerns by drawing on advances in behavioral science and digital monitoring and insights into biological adaptations. Specifically, I focus on four interconnected themes: First, I spotlight the complexity of avian sensory perception—encompassing vision, auditory capabilities, olfaction, and tactile faculties—to underscore how lighting design, housing configurations, and enrichment strategies can better align with birds’ unique sensory worlds. Second, I explore novel tools for gauging emotional states and cognition, ranging from cognitive bias tests to developing protocols for identifying pain or distress based on facial cues. Third, I examine the transformative potential of computer vision, bioacoustics, and sensor-based technologies for the continuous, automated tracking of behavior and physiological indicators in commercial flocks. Fourth, I assess how data-driven management platforms, underpinned by precision livestock farming, can deploy real-time insights to optimize welfare on a broad scale. Recognizing that climate change and evolving production environments intensify these challenges, I also investigate how breeds resilient to extreme conditions might open new avenues for welfare-centered genetic and management approaches. While the adoption of cutting-edge techniques has shown promise, significant hurdles persist regarding validation, standardization, and commercial acceptance. I conclude that truly sustainable progress hinges on an interdisciplinary convergence of ethology, neuroscience, engineering, data analytics, and evolutionary biology—an integrative path that not only refines welfare assessment but also reimagines poultry production in ethically and scientifically robust ways. Full article
Show Figures

Figure 1

18 pages, 4389 KiB  
Article
How Vegetation Structure Shapes the Soundscape: Acoustic Community Partitioning and Its Implications for Urban Forestry Management
by Yilin Zhao, Zhenkai Sun, Zitong Bai, Jiali Jin and Cheng Wang
Forests 2025, 16(4), 669; https://doi.org/10.3390/f16040669 - 11 Apr 2025
Viewed by 463
Abstract
Urban green spaces are critical yet understudied areas where anthropogenic and biological sounds interact. This study investigates how vegetation structure mediates the acoustic partitioning of urban soundscapes and informs sustainable forestry management. Through the principal component analysis (PCA) of 1–11 kHz frequency bands, [...] Read more.
Urban green spaces are critical yet understudied areas where anthropogenic and biological sounds interact. This study investigates how vegetation structure mediates the acoustic partitioning of urban soundscapes and informs sustainable forestry management. Through the principal component analysis (PCA) of 1–11 kHz frequency bands, we identified anthropogenic sounds (1–2 kHz) and biological sounds (2–11 kHz). Within bio-acoustic communities, PCA further revealed three positively correlated sub-clusters (2–4 kHz, 5–6 kHz, and 6–11 kHz), suggesting cooperative niche partitioning among avian, amphibian, and insect vocalizations. Linear mixed models highlighted vegetation’s dual role: mature tree stands (explaining 19.9% variance) and complex vertical structures (leaf-height diversity: 12.2%) significantly enhanced biological soundscapes (R2m = 0.43) while suppressing anthropogenic noise through canopy stratification (32.3% variance explained). Based on our findings, we suggest that an acoustic data-driven framework—comprising (1) the preservation of mature stands with multi-layered canopies to enhance bioacoustic resilience, (2) strategic planting of mid-story vegetation to disrupt low-frequency noise propagation, and (3) real-time soundscape monitoring to balance biophony and anthropophony allocation—can contribute to promoting sustainable urban forestry management. Full article
(This article belongs to the Section Urban Forestry)
Show Figures

Figure 1

21 pages, 12814 KiB  
Article
Multi-Scale Deep Feature Fusion with Machine Learning Classifier for Birdsong Classification
by Wei Li, Danju Lv, Yueyun Yu, Yan Zhang, Lianglian Gu, Ziqian Wang and Zhicheng Zhu
Appl. Sci. 2025, 15(4), 1885; https://doi.org/10.3390/app15041885 - 12 Feb 2025
Cited by 1 | Viewed by 1329
Abstract
Birds are significant bioindicators in the assessment of habitat biodiversity, ecological impacts and ecosystem health. Against the backdrop of easier bird vocalization data acquisition, and with deep learning and machine learning technologies as the technical support, exploring recognition and classification networks suitable for [...] Read more.
Birds are significant bioindicators in the assessment of habitat biodiversity, ecological impacts and ecosystem health. Against the backdrop of easier bird vocalization data acquisition, and with deep learning and machine learning technologies as the technical support, exploring recognition and classification networks suitable for bird calls has become the focus of bioacoustics research. Due to the fact that the spectral differences among various bird calls are much greater than the differences between human languages, constructing birdsong classification networks based on human speech recognition networks does not yield satisfactory results. Effectively capturing the differences in birdsong across species is a crucial factor in improving recognition accuracy. To address the differences in features, this study proposes multi-scale deep features. At the same time, we separate the classification part from the deep network by using machine learning to adapt to classification with distinct feature differences in birdsong. We validate the effectiveness of multi-scale deep features on a publicly available dataset of 20 bird species. The experimental results show that the accuracy of the multi-scale deep features on a log-wavelet spectrum, log-Mel spectrum and log-power spectrum reaches 94.04%, 97.81% and 95.89%, respectively, achieving an improvement over single-scale deep features on these three spectrograms. Comparative experimental results show that the proposed multi-scale deep feature method is superior to five state-of-the-art birdsong identification methods, which provides new perspectives and tools for birdsong identification research, and is of great significance for ecological monitoring, biodiversity conservation and forest research. Full article
Show Figures

Figure 1

20 pages, 2495 KiB  
Article
Monitoring Postfire Biodiversity Dynamics in Mediterranean Pine Forests Using Acoustic Indices
by Dimitrios Spatharis, Aggelos Tsaligopoulos, Yiannis G. Matsinos, Ilias Karmiris, Magdalini Pleniou, Elisabeth Navarrete, Eleni Boikou and Christos Astaras
Environments 2024, 11(12), 277; https://doi.org/10.3390/environments11120277 - 4 Dec 2024
Viewed by 2365
Abstract
In recent decades, climate change has significantly influenced the frequency and intensity of wildfires across Mediterranean pine forests. The loss of forest cover can bring long-term ecological changes that impact the overall biodiversity and alter species composition. Understanding the long-term impact of wildfires [...] Read more.
In recent decades, climate change has significantly influenced the frequency and intensity of wildfires across Mediterranean pine forests. The loss of forest cover can bring long-term ecological changes that impact the overall biodiversity and alter species composition. Understanding the long-term impact of wildfires requires effective and cost-efficient methods for monitoring the postfire ecosystem dynamics. Passive acoustic monitoring (PAM) has been increasingly used to monitor the biodiversity of vocal species at large spatial and temporal scales. Using acoustic indices, where the biodiversity of an area is inferred from the overall structure of the soundscape, rather than the more labor-intensive identification of individual species, has yielded mixed results, emphasizing the importance of testing their efficacy at the regional level. In this study, we examined whether widely used acoustic indicators were effective at capturing changes in the avifauna diversity in Pinus halepensis forest stands with different fire burning histories (burnt in 2001, 2009, and 2018 and unburnt for >20 years) on the Sithonia Peninsula, Greece. We recorded the soundscape of each stand using two–three sensors across 11 days of each season from March 2022 to January 2023. We calculated for each site and season the following five acoustic indices: the Acoustic Complexity Index (ACI), Acoustic Diversity Index (ADI), Acoustic Evenness Index (AEI), Normalized Difference Soundscape Index (NDSI), and Bioacoustic Index (BI). Each acoustic index was then assessed in terms of its efficacy at predicting the local avifauna diversity, as estimated via two proxies—the species richness (SR) and the Shannon Diversity Index (SDI) of vocal bird calls. Both the SR and SDI were calculated by having an expert review the species identification of calls detected within the same acoustic dataset by the BirdNET convolutional neural network algorithm. A total of 53 bird species were identified. Our analysis shows that the BI and NDSI have the highest potential for monitoring the postfire biodiversity dynamics in Mediterranean pine forests. We propose the development of regional-scale acoustic observatories at pine and other fire-prone Mediterranean habitats, which will further improve our understanding of how to make the best use of acoustic indices as a tool for rapid biodiversity assessments. Full article
(This article belongs to the Special Issue Interdisciplinary Noise Research)
Show Figures

Figure 1

11 pages, 3672 KiB  
Article
Aquariums as Research Platforms: Characterizing Fish Sounds in Controlled Settings with Preliminary Insights from the Blackbar Soldierfish Myripristis jacobus
by Javier Almunia, María Fernández-Maquieira and Melvin Flores
J. Zool. Bot. Gard. 2024, 5(4), 630-640; https://doi.org/10.3390/jzbg5040042 - 29 Oct 2024
Viewed by 1499
Abstract
This study highlights the potential of aquariums as research platforms for bioacoustic research. Aquariums provide access to a wide variety of fish species, offering unique opportunities to characterize their acoustic features in controlled settings. In particular, we present a preliminary description of the [...] Read more.
This study highlights the potential of aquariums as research platforms for bioacoustic research. Aquariums provide access to a wide variety of fish species, offering unique opportunities to characterize their acoustic features in controlled settings. In particular, we present a preliminary description of the acoustic characteristics of Myripristis jacobus, a soniferous species in the Holocentridae family, within a controlled environment at a zoological facility in the Canary Islands, Spain. Using two HydroMoth 1.0 hydrophones, we recorded vocalizations of the blackbar soldierfish in a glass tank, revealing a pulsed sound type with a peak frequency around 355 Hz (DS 64), offering a more precise characterization than previously available. The vocalizations exhibit two distinct patterns: short sequences with long pulse intervals and fast pulse trains with short inter-pulse intervals. Despite some limitations, this experimental setup highlights the efficacy of cost-effective methodologies in public aquariums for initial bioacoustic research. These findings contribute to the early stages of acoustic characterization of coastal fishes in the western central Atlantic, emphasizing the value of passive acoustic monitoring for ecological assessments and conservation efforts. Moreover, this study opens new avenues for considering the acoustic environment as a crucial factor in the welfare of captive fish, an aspect that has largely been overlooked in aquarium management. Full article
Show Figures

Graphical abstract

19 pages, 5934 KiB  
Article
Detection of Typical Transient Signals in Water by XGBoost Classifier Based on Shape Statistical Features: Application to the Call of Southern Right Whale
by Zemin Zhou, Yanrui Qu, Boqing Zhu and Bingbing Zhang
J. Mar. Sci. Eng. 2024, 12(9), 1596; https://doi.org/10.3390/jmse12091596 - 9 Sep 2024
Viewed by 1521
Abstract
Whale sound is a typical transient signal. The escalating demands of ecological research and marine conservation necessitate advanced technologies for the automatic detection and classification of underwater acoustic signals. Traditional energy detection methods, which focus primarily on amplitude, often perform poorly in the [...] Read more.
Whale sound is a typical transient signal. The escalating demands of ecological research and marine conservation necessitate advanced technologies for the automatic detection and classification of underwater acoustic signals. Traditional energy detection methods, which focus primarily on amplitude, often perform poorly in the non-Gaussian noise conditions typical of oceanic environments. This study introduces a classified-before-detect approach that overcomes the limitations of amplitude-focused techniques. We also address the challenges posed by deep learning models, such as high data labeling costs and extensive computational requirements. By extracting shape statistical features from audio and using the XGBoost classifier, our method not only outperforms the traditional convolutional neural network (CNN) method in accuracy but also reduces the dependence on labeled data, thus improving the detection efficiency. The integration of these features significantly enhances model performance, promoting the broader application of marine acoustic remote sensing technologies. This research contributes to the advancement of marine bioacoustic monitoring, offering a reliable, rapid, and training-efficient method suitable for practical deployment. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

20 pages, 21445 KiB  
Article
Using Deep Learning to Classify Environmental Sounds in the Habitat of Western Black-Crested Gibbons
by Ruiqi Hu, Kunrong Hu, Leiguang Wang, Zhenhua Guan, Xiaotao Zhou, Ning Wang and Longjia Ye
Diversity 2024, 16(8), 509; https://doi.org/10.3390/d16080509 - 22 Aug 2024
Viewed by 1784
Abstract
The western black-crested gibbon (Nomascus concolor) is a rare and endangered primate that inhabits southern China and northern Vietnam, and has become a key conservation target due to its distinctive call and highly endangered status, making its identification and monitoring particularly [...] Read more.
The western black-crested gibbon (Nomascus concolor) is a rare and endangered primate that inhabits southern China and northern Vietnam, and has become a key conservation target due to its distinctive call and highly endangered status, making its identification and monitoring particularly urgent. Identifying calls of the western black-crested gibbon using passive acoustic monitoring data is a crucial method for studying and analyzing these gibbons; however, traditional call recognition models often overlook the temporal information in audio features and fail to adapt to channel-feature weights. To address these issues, we propose an innovative deep learning model, VBSNet, designed to recognize and classify a variety of biological calls, including those of endangered western black-crested gibbons and certain bird species. The model incorporates the image feature extraction capability of the VGG16 convolutional network, the sequence modeling capability of bi-directional LSTM, and the feature selection capability of the SE attention module, realizing the multimodal fusion of image, sequence and attention information. In the constructed dataset, the VBSNet model achieved the best performance in the evaluation metrics of accuracy, precision, recall, and F1-score, realizing an accuracy of 98.35%, demonstrating high accuracy and generalization ability. This study provides an effective deep learning method in the field of automated bioacoustic monitoring, which is of great theoretical and practical significance for supporting wildlife conservation and maintaining biodiversity. Full article
Show Figures

Figure 1

11 pages, 297 KiB  
Article
Efficient Speech Detection in Environmental Audio Using Acoustic Recognition and Knowledge Distillation
by Drew Priebe, Burooj Ghani and Dan Stowell
Sensors 2024, 24(7), 2046; https://doi.org/10.3390/s24072046 - 22 Mar 2024
Cited by 4 | Viewed by 2089
Abstract
The ongoing biodiversity crisis, driven by factors such as land-use change and global warming, emphasizes the need for effective ecological monitoring methods. Acoustic monitoring of biodiversity has emerged as an important monitoring tool. Detecting human voices in soundscape monitoring projects is useful both [...] Read more.
The ongoing biodiversity crisis, driven by factors such as land-use change and global warming, emphasizes the need for effective ecological monitoring methods. Acoustic monitoring of biodiversity has emerged as an important monitoring tool. Detecting human voices in soundscape monitoring projects is useful both for analyzing human disturbance and for privacy filtering. Despite significant strides in deep learning in recent years, the deployment of large neural networks on compact devices poses challenges due to memory and latency constraints. Our approach focuses on leveraging knowledge distillation techniques to design efficient, lightweight student models for speech detection in bioacoustics. In particular, we employed the MobileNetV3-Small-Pi model to create compact yet effective student architectures to compare against the larger EcoVAD teacher model, a well-regarded voice detection architecture in eco-acoustic monitoring. The comparative analysis included examining various configurations of the MobileNetV3-Small-Pi-derived student models to identify optimal performance. Additionally, a thorough evaluation of different distillation techniques was conducted to ascertain the most effective method for model selection. Our findings revealed that the distilled models exhibited comparable performance to the EcoVAD teacher model, indicating a promising approach to overcoming computational barriers for real-time ecological monitoring. Full article
(This article belongs to the Special Issue Acoustic Sensing and Monitoring in Urban and Natural Environments)
Show Figures

Figure 1

15 pages, 4031 KiB  
Article
Bioacoustic IoT Sensors as Next-Generation Tools for Monitoring: Counting Flying Insects through Buzz
by Simona Alberti, Gianluca Stasolla, Simone Mazzola, Luca Pietro Casacci and Francesca Barbero
Insects 2023, 14(12), 924; https://doi.org/10.3390/insects14120924 - 5 Dec 2023
Cited by 2 | Viewed by 4188
Abstract
The global loss of biodiversity is an urgent concern requiring the implementation of effective monitoring. Flying insects, such as pollinators, are vital for ecosystems, and establishing their population dynamics has become essential in conservation biology. Traditional monitoring methods are labour-intensive and show time [...] Read more.
The global loss of biodiversity is an urgent concern requiring the implementation of effective monitoring. Flying insects, such as pollinators, are vital for ecosystems, and establishing their population dynamics has become essential in conservation biology. Traditional monitoring methods are labour-intensive and show time constraints. In this work, we explore the use of bioacoustic sensors for monitoring flying insects. Data collected at four Italian farms using traditional monitoring methods, such as hand netting and pan traps, and bioacoustic sensors were compared. The results showed a positive correlation between the average number of buzzes per hour and insect abundance measured by traditional methods, primarily by pan traps. Intraday and long-term analysis performed on buzzes revealed temperature-related patterns of insect activity. Passive acoustic monitoring proved to be effective in estimating flying insect abundance, while further development of the algorithm is required to correctly identify insect taxa. Overall, innovative technologies, such as bioacoustic sensors, do not replace the expertise and data quality provided by professionals, but they offer unprecedented opportunities to ease insect monitoring to support conservation biodiversity efforts. Full article
(This article belongs to the Section Insect Ecology, Diversity and Conservation)
Show Figures

Graphical abstract

Back to TopTop