Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (265)

Search Parameters:
Keywords = infrared videos

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
27 pages, 4051 KB  
Article
Lossless Compression of Large Field-of-View Infrared Video Based on Transform Domain Hybrid Prediction
by Ya Liu, Rui Zhang, Yong Zhang and Yuwei Chen
Sensors 2026, 26(3), 868; https://doi.org/10.3390/s26030868 - 28 Jan 2026
Abstract
Large field-of-view (FOV) infrared imaging, widely utilized in applications including target detection and remote sensing, generates massive datasets that pose significant challenges for transmission and storage. To address this issue, we propose an efficient lossless compression method for large FOV infrared video. Our [...] Read more.
Large field-of-view (FOV) infrared imaging, widely utilized in applications including target detection and remote sensing, generates massive datasets that pose significant challenges for transmission and storage. To address this issue, we propose an efficient lossless compression method for large FOV infrared video. Our approach employs a hybrid prediction strategy within the transform domain. The video frames are first decomposed into low- and high-frequency components via the discrete wavelet transform. For the low-frequency subbands, an improved low-latency Multi-view High-Efficiency Video Coding (MV-HEVC) encoder is adopted, where the background reference frames are treated as one view to enable more accurate inter-frame prediction. For high-frequency components, pixel-wise clustered edge prediction is applied. Furthermore, the prediction residuals are reduced by optimal direction prediction, according to the principle of minimizing residual energy. Experimental results demonstrate that our method significantly outperforms mainstream video compression techniques. While maintaining compression performance comparable to MV-HEVC, the proposed method exhibits a 19.3-fold improvement in computational efficiency. Full article
(This article belongs to the Section Sensing and Imaging)
23 pages, 4261 KB  
Article
Efficient Drone Detection Using Temporal Anomalies and Small Spatio-Temporal Networks
by Abhijit Mahalanobis and Amadou Tall
Sensors 2026, 26(1), 170; https://doi.org/10.3390/s26010170 - 26 Dec 2025
Viewed by 367
Abstract
Detecting small drones in Infrared (IR) sequences poses significant challenges due to their low visibility, low resolution, and complex cluttered backgrounds. These factors often lead to high false alarm and missed detection rates. This paper frames drone detection as a spatio-temporal anomaly detection [...] Read more.
Detecting small drones in Infrared (IR) sequences poses significant challenges due to their low visibility, low resolution, and complex cluttered backgrounds. These factors often lead to high false alarm and missed detection rates. This paper frames drone detection as a spatio-temporal anomaly detection problem and proposes a remarkably lightweight pipeline solution (well-suited for edge applications), by employing a statistical temporal anomaly detector (known as the temporal Reed Xiaoli (TRX) algorithm), in parallel with a light-weight convolutional neural network known as the TCRNet. While the TRX detector is unsupervised, the TCRNet is trained to discriminate between drones and clutter using spatio-temporal patches (or chips). The confidence maps from both modules are additively fused to localize drones in video imagery. We compare our method, dubbed TRX-TCRnet, to other state-of-the-art drone detection techniques using the Detection of Aircraft Under Background (DAUB) dataset. Our approach achieves exceptional computational efficiency with only 0.17 GFLOPs with 0.83 M parameters, outperforming methods that require 145–795 times more computational resources. At the same time, the TRX–TCRNet achieves one of the highest detection accuracies (mAP50 of 97.40) while requiring orders of magnitude fewer computational resources than competing methods, demonstrating unprecedented efficiency–performance trade-offs for real-time applications. Experimental results, including ROC and PR curves, confirm the framework’s exceptional suitability for resource-constrained environments and embedded systems. Full article
(This article belongs to the Special Issue Signal Processing and Machine Learning for Sensor Systems)
Show Figures

Figure 1

19 pages, 24785 KB  
Article
Capsicum Counting Algorithm Using Infrared Imaging and YOLO11
by Enrico Mendez, Jesús Arturo Escobedo Cabello, Alfonso Gómez-Espinosa, Jose Antonio Cantoral-Ceballos and Oscar Ochoa
Agriculture 2025, 15(24), 2574; https://doi.org/10.3390/agriculture15242574 - 12 Dec 2025
Viewed by 489
Abstract
Fruit detection and counting is a key component of data-driven resource management and yield estimation in greenhouses. This work presents a novel infrared-based approach to capsicum counting in greenhouses that takes advantage of the light penetration of infrared (IR) imaging to enhance detection [...] Read more.
Fruit detection and counting is a key component of data-driven resource management and yield estimation in greenhouses. This work presents a novel infrared-based approach to capsicum counting in greenhouses that takes advantage of the light penetration of infrared (IR) imaging to enhance detection under challenging lighting conditions. The proposed capsicum counting pipeline integrates the YOLO11 detection model for capsicum identification and the BoT-SORT multi-object tracker to track detections across a video stream, enabling accurate fruit counting. The detector model is trained on a dataset of 1000 images, with 11,916 labeled capsicums, captured with an OAK-D pro camera mounted on a mobile robot inside a capsicum greenhouse. On the IR test set, the YOLO11m model achieved an F1-score of 0.82, while the tracker obtained a multiple object tracking accuracy (MOTA) of 0.85, correctly counting 67 of 70 capsicums in a representative greenhouse row. The results demonstrate the effectiveness of this IR-based approach in automating fruit counting in greenhouse environments, offering potential applications in yield estimation. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

17 pages, 1732 KB  
Article
Enhancing Endangered Feline Conservation in Asia via a Pose-Guided Deep Learning Framework for Individual Identification
by Weiwei Xiao, Wei Zhang and Haiyan Liu
Diversity 2025, 17(12), 853; https://doi.org/10.3390/d17120853 - 12 Dec 2025
Viewed by 435
Abstract
The re-identification of endangered felines is critical for species conservation and biodiversity assessment. This paper proposes the Pose-Guided Network with the Adaptive L2 Regularization (PGNet-AL2) framework to overcome key challenges in wild feline re-identification, such as extensive pose variations, small sample sizes, and [...] Read more.
The re-identification of endangered felines is critical for species conservation and biodiversity assessment. This paper proposes the Pose-Guided Network with the Adaptive L2 Regularization (PGNet-AL2) framework to overcome key challenges in wild feline re-identification, such as extensive pose variations, small sample sizes, and inconsistent image quality. This framework employs a dual-branch architecture for multi-level feature extraction and incorporates an adaptive L2 regularization mechanism to optimize parameter learning, effectively mitigating overfitting in small-sample scenarios. Applying the proposed method to the Amur Tiger Re-identification in the Wild (ATRW) dataset, we achieve a mean Average Precision (mAP) of 91.3% in single-camera settings, outperforming the baseline PPbM-b (Pose Part-based Model) by 18.5 percentage points. To further evaluate its generalization, we apply it to a more challenging task, snow leopard re-identification, using a dataset of 388 infrared videos obtained from the Wildlife Conservation Society (WCS). Despite the poor quality of infrared videos, our method achieves a mAP of 94.5%. The consistent high performance on both the ATRW and snow leopard datasets collectively demonstrates the method’s strong generalization capability and practical utility. Full article
Show Figures

Graphical abstract

12 pages, 1048 KB  
Article
Fluorescence-Guided Thoracoscopic Surgery Using Indocyanine Green (ICG) in Canine Cadavers: A Descriptive Evaluation of Video-Assisted (VATS) and Robot-Assisted (RATS) Approaches
by Francisco M. Sánchez-Margallo, Lucía Salazar-Carrasco, Manuel J. Pérez-Salazar and Juan A. Sánchez-Margallo
Animals 2025, 15(24), 3519; https://doi.org/10.3390/ani15243519 - 5 Dec 2025
Viewed by 384
Abstract
Precise intraoperative identification of the canine thoracic duct remains challenging due to anatomical variability and limited visualization. This exploratory cadaveric feasibility study aimed to describe the technical applicability of fluorescence-guided thoracic duct mapping using video-assisted thoracoscopy (VATS) and robot-assisted thoracoscopy (Versius™ system). Four [...] Read more.
Precise intraoperative identification of the canine thoracic duct remains challenging due to anatomical variability and limited visualization. This exploratory cadaveric feasibility study aimed to describe the technical applicability of fluorescence-guided thoracic duct mapping using video-assisted thoracoscopy (VATS) and robot-assisted thoracoscopy (Versius™ system). Four adult Beagle cadavers underwent bilateral thoracoscopic exploration after intranodal injection of indocyanine green (ICG, Verdye®, 0.05 mg/kg; 0.5 mL). Near-infrared (NIR) fluorescence imaging enabled real-time visualization of the thoracic duct and its branches. Fluorescence quality was quantitatively characterized using signal-to-noise ratio (SNR), contrast-to-noise ratio (CNR), and contrast resolution (CR) calculated from standardized image frames. Both approaches achieved successful duct identification in all cadavers. VATS provided brighter overall fluorescence, whereas the robotic-assisted approach offered stable imaging, enhanced instrument dexterity, and improved duct-to-background discrimination. These findings confirm the feasibility of fluorescence-guided thoracic duct identification using both minimally invasive modalities in canine cadavers. The standardized assessment of optical parameters proposed here may support future in vivo studies to optimize imaging protocols and evaluate the clinical impact of fluorescence-guided thoracic duct surgery in dogs. Full article
(This article belongs to the Section Companion Animals)
Show Figures

Figure 1

10 pages, 2402 KB  
Article
Novel System for Measuring Tension Force in Eyeball Movement
by Jae Yun Sung, Ju Mi Kim, Il Doh and Yeon-Hee Lee
Biosensors 2025, 15(12), 769; https://doi.org/10.3390/bios15120769 - 25 Nov 2025
Viewed by 425
Abstract
Accurate assessment of extraocular muscle mechanics is crucial for diagnosing and treating ocular motility disorders, yet current methods, such as the forced duction test, rely on subjective tactile sensation and gross visual observation. To overcome the limitations of subjectivity and the impracticality of [...] Read more.
Accurate assessment of extraocular muscle mechanics is crucial for diagnosing and treating ocular motility disorders, yet current methods, such as the forced duction test, rely on subjective tactile sensation and gross visual observation. To overcome the limitations of subjectivity and the impracticality of previous quantitative devices, we developed a novel biosensing system capable of simultaneously and objectively measuring passive ocular tension and rotation angle during forced duction. The system integrates custom-engineered surgical forceps equipped with dual strain gauges and an infrared video camera that precisely tracks pupil displacement to calculate real-time rotation angle. We clinically validated this system in a prospective study involving 10 patients (20 eyes) with intermittent exotropia, with measurements performed under general anesthesia. Reliable tension–angle curves were successfully obtained in all cases without complications. Passive tension increased progressively with ocular rotation, following a linear-parabolic trajectory up to 40°. The mean duction force of the medial and lateral rectus muscles showed comparable symmetry. This lightweight, practical, and objective biosensing system offers a reliable tool for quantifying ocular mechanics, with the potential to enhance diagnostic accuracy, enable individualized surgical planning, and support fundamental research in ocular motility disorders. Full article
(This article belongs to the Section Biosensors and Healthcare)
Show Figures

Figure 1

24 pages, 24641 KB  
Article
Memory-Based Temporal Transformer U-Net for Multi-Frame Infrared Small Target Detection
by Zicheng Feng, Wenlong Zhang, Donghui Liu, Xingfu Tao, Ang Su and Yixin Yang
Remote Sens. 2025, 17(23), 3801; https://doi.org/10.3390/rs17233801 - 23 Nov 2025
Viewed by 653
Abstract
In the field of infrared small target detection (ISTD), single-frame ISTD (SISTD), using only spatial features, cannot deal well with dim targets in cluttered backgrounds. In contrast, multi-frame ISTD (MISTD), utilizing spatio-temporal information from videos, can significantly enhance moving target features and effectively [...] Read more.
In the field of infrared small target detection (ISTD), single-frame ISTD (SISTD), using only spatial features, cannot deal well with dim targets in cluttered backgrounds. In contrast, multi-frame ISTD (MISTD), utilizing spatio-temporal information from videos, can significantly enhance moving target features and effectively suppress background interference. However, current MISTD algorithms are limited by fixed-size time windows, resulting in an inability to adaptively adjust the input amount of spatio-temporal information for different detection scenarios. Moreover, utilizing spatio-temporal features remains a significant challenge in MISTD, particularly in scenarios involving slow-moving targets and fast-moving backgrounds. To address the above problems, we propose a memory-based temporal Transformer U-Net (MTTU-Net), which integrates a memory-based temporal Transformer module (MTTM) into U-Net. Specifically, MTTM utilizes the proposed D-ConvLSTM to sequentially transmit the temporal information in the form of memory, breaking through the limitation of the time window paradigm. And we propose a Transformer-based interactive fusion approach, which is dominated by spatial features of the to-be-detected frame and supplemented by temporal features in the memory, thereby effectively dealing with targets and backgrounds with various motion states. In addition, MTTM is divided into a temporal channel-cross Transformer module (TCTM) and a temporal space-cross Transformer module (TSTM), which achieve target feature enhancement and global background perception through feature interactive fusion in the channel and space dimensions, respectively. Extensive experiments on IRDST and IDSMT datasets demonstrate that our MTTU-Net outperforms existing MISTD algorithms, and they verify the effectiveness of the proposed modules. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Figure 1

24 pages, 2454 KB  
Article
Low-Cost Eye-Tracking Fixation Analysis for Driver Monitoring Systems Using Kalman Filtering and OPTICS Clustering
by Jonas Brandstetter, Eva-Maria Knoch and Frank Gauterin
Sensors 2025, 25(22), 7028; https://doi.org/10.3390/s25227028 - 17 Nov 2025
Viewed by 771
Abstract
Driver monitoring systems benefit from fixation-related eye-tracking features, yet dedicated eye-tracking hardware is costly and difficult to integrate at scale. This study presents a practical software pipeline that extracts fixation-related features from conventional RGB video. Facial and pupil landmarks obtained with MediaPipe are [...] Read more.
Driver monitoring systems benefit from fixation-related eye-tracking features, yet dedicated eye-tracking hardware is costly and difficult to integrate at scale. This study presents a practical software pipeline that extracts fixation-related features from conventional RGB video. Facial and pupil landmarks obtained with MediaPipe are denoised using a Kalman filter, fixation centers are identified with the OPTICS algorithm within a sliding window, and an affine normalization compensates for head motion and camera geometry. Fixation segments are derived from smoothed velocity profiles based on a moving average. Experiments with laptop camera recordings show that the combined Kalman and OPTICS pipeline reduces landmark jitter and yields more stable fixation centroids, while the affine normalization further improves referential pupil stability. The pipeline operates with minimal computational overhead and can be implemented as a software update in existing driver monitoring or advanced driver assistance systems. This work is a proof of concept that demonstrates feasibility in a low-cost RGB setting with a limited evaluation scope. Remaining challenges include sensitivity to lighting conditions and head motion that future work may address through near-infrared sensing, adaptive calibration, and broader validation across subjects, environments, and cameras. The extracted features are relevant for future studies on cognitive load and attention, although cognitive state inference is not validated here. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

19 pages, 2109 KB  
Article
SF6 Leak Detection in Infrared Video via Multichannel Fusion and Spatiotemporal Features
by Zhiwei Li, Xiaohui Zhang, Zhilei Xu, Yubo Liu and Fengjuan Zhang
Appl. Sci. 2025, 15(20), 11141; https://doi.org/10.3390/app152011141 - 17 Oct 2025
Viewed by 655
Abstract
With the development of infrared imaging technology and the integration of intelligent algorithms, the realization of non-contact, dynamic and real-time detection of SF6 gas leakage based on infrared video has been a significant research direction. However, the existing real-time detection algorithms exhibit low [...] Read more.
With the development of infrared imaging technology and the integration of intelligent algorithms, the realization of non-contact, dynamic and real-time detection of SF6 gas leakage based on infrared video has been a significant research direction. However, the existing real-time detection algorithms exhibit low accuracy in detecting SF6 leakage and are susceptible to noise, which makes it difficult to meet the actual needs of engineering. To address this problem, this paper proposes a real-time SF6 leakage detection method, VGEC-Net, based on multi-channel fusion and spatiotemporal feature extraction. The proposed method first employs the ViBe-GMM algorithm to extract foreground masks, which are then fused with infrared images to construct a dual-channel input. In the backbone network, a CE-Net structure—integrating CBAM and ECA-Net—is combined with the P3D network to achieve efficient spatiotemporal feature extraction. A Feature Pyramid Network (FPN) and a temporal Transformer module are further integrated to enhance multi-scale feature representation and temporal modeling, thereby significantly improving the detection performance for small-scale targets. Experimental results demonstrate that VGEC-Net achieves a mean average precision (mAP) of 61.7% on the dataset used in this study, with a mAP@50 of 87.3%, which represents a significant improvement over existing methods. These results validate the effectiveness and advancement of the proposed method for infrared video-based gas leakage detection. Furthermore, the model achieves 78.2 frames per second (FPS) during inference, demonstrating good real-time processing capability while maintaining high detection accuracy, exhibiting strong application potential. Full article
Show Figures

Figure 1

20 pages, 1288 KB  
Article
Spatio-Temporal Residual Attention Network for Satellite-Based Infrared Small Target Detection
by Yan Chang, Decao Ma, Qisong Yang, Shaopeng Li and Daqiao Zhang
Remote Sens. 2025, 17(20), 3457; https://doi.org/10.3390/rs17203457 - 16 Oct 2025
Viewed by 828
Abstract
With the development of infrared remote sensing technology and the deployment of satellite constellations, infrared video from orbital platforms is playing an increasingly important role in airborne target surveillance. However, due to the limitations of remote sensing imaging, the aerial targets in such [...] Read more.
With the development of infrared remote sensing technology and the deployment of satellite constellations, infrared video from orbital platforms is playing an increasingly important role in airborne target surveillance. However, due to the limitations of remote sensing imaging, the aerial targets in such videos are often small in scale, low in contrast, and slow in movement, making them difficult to detect in complex backgrounds. In this paper, we propose a novel detection network that integrates inter-frame residual guidance with spatio-temporal feature enhancement to address the challenge of small object detection in infrared satellite video. This method first extracts residual features to highlight motion-sensitive regions, then uses a dual-branch structure to encode spatial semantics and temporal evolution, and then fuses them deeply through a multi-scale feature enhancement module. Extensive experiments show that this method outperforms mainstream methods in terms on various infrared small target video datasets, and has good robustness under low-signal-to-noise-ratio conditions. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Figure 1

19 pages, 762 KB  
Article
TMRGBT-D2D: A Temporal Misaligned RGB-Thermal Dataset for Drone-to-Drone Target Detection
by Hexiang Hao, Yueping Peng, Zecong Ye, Baixuan Han, Wei Tang, Wenchao Kang, Xuekai Zhang, Qilong Li and Wenchao Liu
Drones 2025, 9(10), 694; https://doi.org/10.3390/drones9100694 - 10 Oct 2025
Viewed by 1937
Abstract
In the field of drone-to-drone detection tasks, the issue of fusing temporal information with infrared and visible light data for detection has been rarely studied. This paper presents the first temporal misaligned rgb-thermal dataset for drone-to-drone target detection, named TMRGBT-D2D. The dataset covers [...] Read more.
In the field of drone-to-drone detection tasks, the issue of fusing temporal information with infrared and visible light data for detection has been rarely studied. This paper presents the first temporal misaligned rgb-thermal dataset for drone-to-drone target detection, named TMRGBT-D2D. The dataset covers various lighting conditions (i.e., high-light scenes captured during the day, medium-light and low-light scenes captured at night, with night scenes accounting for 38.8% of all data), different scenes (sky, forests, buildings, construction sites, playgrounds, roads, etc.), different seasons, and different locations, consisting of a total of 42,624 images organized into sequential frames extracted from 19 RGB-T video pairs. Each frame in the dataset has been meticulously annotated, with a total of 94,323 annotations. Except for drones that cannot be identified under extreme conditions, infrared and visible light annotations are one-to-one corresponding. This dataset presents various challenges, including small object detection (the average size of objects in visible light images is approximately 0.02% of the image area), motion blur caused by fast movement, and detection issues arising from imaging differences between different modalities. To our knowledge, this is the first temporal misaligned rgb-thermal dataset for drone-to-drone target detection, providing convenience for research into rgb-thermal image fusion and the development of drone target detection. Full article
(This article belongs to the Special Issue Detection, Identification and Tracking of UAVs and Drones)
Show Figures

Figure 1

16 pages, 13443 KB  
Article
NIR Indocyanine–White Light Overlay Visualization for Neuro-Oto-Vascular Preservation During Anterior Transpetrosal Approaches: A Technical Note
by Leonardo Tariciotti, Alejandra Rodas, Erion De Andrade, Juan Manuel Revuelta Barbero, Youssef M. Zohdy, Roberto Soriano, Jackson R. Vuncannon, Justin Maldonado, Samir Lohana, Francesco DiMeco, Tomas Garzon-Muvdi, Camilo Reyes, C. Arturo Solares and Gustavo Pradilla
J. Clin. Med. 2025, 14(19), 6954; https://doi.org/10.3390/jcm14196954 - 1 Oct 2025
Viewed by 636
Abstract
Objectives: Anterior petrosectomy is a challenging neurosurgical procedure requiring precise identification and preservation of multiple critical structures. This technical note explores the feasibility of using real-time near-infrared indocyanine green (NIR-ICG) fluorescence with white light overlay to enhance visualization of the petrous internal [...] Read more.
Objectives: Anterior petrosectomy is a challenging neurosurgical procedure requiring precise identification and preservation of multiple critical structures. This technical note explores the feasibility of using real-time near-infrared indocyanine green (NIR-ICG) fluorescence with white light overlay to enhance visualization of the petrous internal carotid artery (ICA) during transpetrosal drilling. We aimed to assess its utility for planning and performing modified Dolenc–Kawase drilling. Methods: We integrated NIR-ICG and white light overlay using a robotic microscope with simultaneous visualization capabilities. This technique was applied to improve neurovascular preservation and skull base landmark identification. Intraoperative video frames and images were captured during an anterior transpetrosal approach for a petroclival meningioma, with technical details, surgical time, and feedback documented. Results: Real-time NIR-ICG with white light overlay successfully identified the posterior genu, horizontal petrosal segment, anterior genu, and superior petrosal sinus. It facilitated precise localization of cochlear landmarks, enabling tailored drilling of the Dolenc–Kawase rhomboid according to patient anatomy and accommodating potential anatomical variants. Conclusions: This approach could enhance intraoperative safety and improve exposure, possibly reducing neurovascular risks without extending operative time. It may serve as a valuable adjunct for complex skull base surgeries. Full article
(This article belongs to the Section Clinical Neurology)
Show Figures

Graphical abstract

4 pages, 2857 KB  
Abstract
Can Transfer Learning Overcome the Challenge of Identifying Lemming Species in Images Taken in the near Infrared Spectrum?
by Davood Kalhor, Mathilde Poirier, Xavier Maldague and Gilles Gauthier
Proceedings 2025, 129(1), 65; https://doi.org/10.3390/proceedings2025129065 - 12 Sep 2025
Viewed by 435
Abstract
Using a camera system developed earlier for monitoring the behavior of lemmings under the snow, we are now able to record a large number of short image sequences from this rodent which plays a central role in the Arctic food web. Identifying lemming [...] Read more.
Using a camera system developed earlier for monitoring the behavior of lemmings under the snow, we are now able to record a large number of short image sequences from this rodent which plays a central role in the Arctic food web. Identifying lemming species in these images manually is wearisome and time-consuming. To perform this task, we present a deep neural network which has several million parameters to configure. Training a network of such an immense size with conventional methods requires a huge amount of data but a sufficiently large labeled dataset of lemming images is currently lacking. Another challenge is that images are obtained in darkness in the near infrared spectrum, causing the loss of some image texture information. We investigate whether these challenges can be tackled by a transfer learning approach in which a network is pretrained on a dataset of visible spectrum images that does not include lemmings. We believe this work provides a basis for moving toward developing intelligent software programs that can facilitate the analysis of videos by biologists. Full article
Show Figures

Figure 1

18 pages, 5522 KB  
Article
Automated Detection of Methane Leaks by Combining Infrared Imaging and a Gas-Faster Region-Based Convolutional Neural Network Technique
by Jinhui Zuo, Zhengqiang Li, Wenbin Xu, Jinxin Zuo and Zhipeng Rong
Sensors 2025, 25(18), 5714; https://doi.org/10.3390/s25185714 - 12 Sep 2025
Cited by 1 | Viewed by 1599
Abstract
Gas leaks threaten ecological and social safety. Non-contact infrared imaging enables large-scale, real-time measurements; however, in complex environments, weak signals from small leaks can hinder reliable detection. This study proposes a novel automated methane leak detection method based on infrared imaging and a [...] Read more.
Gas leaks threaten ecological and social safety. Non-contact infrared imaging enables large-scale, real-time measurements; however, in complex environments, weak signals from small leaks can hinder reliable detection. This study proposes a novel automated methane leak detection method based on infrared imaging and a Gas-Faster Region-based convolutional neural network (Gas R-CNN) to classify leakage amounts (≥30 mL/min). An uncooled infrared imaging system was employed to capture gas leak images containing leak volume features. We developed the Gas R-CNN model for gas leakage detection. This model introduces a multiscale feature network to improve leak feature extraction and enhancement, and it incorporates region-of-interest alignment to address the mismatch caused by double quantization. Feature extraction was enhanced by integrating ResNet50 with an efficient channel attention mechanism. Image enhancement techniques were applied to expand the dataset diversity. Leak detection capabilities were validated using the IOD-Video dataset, while the constructed gas dataset enabled the first quantitative leak assessment. The experimental results demonstrated that the model can accurately detect the leakage area and classify leakage amounts, enabling the quantitative analysis of infrared images. The proposed method achieved average precisions of 0.9599, 0.9647, and 0.9833 for leak rates of 30, 100, and 300 mL/min, respectively. Full article
(This article belongs to the Section Optical Sensors)
Show Figures

Figure 1

32 pages, 4331 KB  
Article
Deep Learning for Wildlife Monitoring: Near-Infrared Bat Detection Using YOLO Frameworks
by José-Joel González-Barbosa, Israel Cruz Rangel, Alfonso Ramírez-Pedraza, Raymundo Ramírez-Pedraza, Isabel Bárcenas-Reyes, Erick-Alejandro González-Barbosa and Miguel Razo-Razo
Signals 2025, 6(3), 46; https://doi.org/10.3390/signals6030046 - 4 Sep 2025
Cited by 2 | Viewed by 1763
Abstract
Bats are ecologically vital mammals, serving as pollinators, seed dispersers, and bioindicators of ecosystem health. Many species inhabit natural caves, which offer optimal conditions for survival but present challenges for direct ecological monitoring due to their dark, complex, and inaccessible environments. Traditional monitoring [...] Read more.
Bats are ecologically vital mammals, serving as pollinators, seed dispersers, and bioindicators of ecosystem health. Many species inhabit natural caves, which offer optimal conditions for survival but present challenges for direct ecological monitoring due to their dark, complex, and inaccessible environments. Traditional monitoring methods, such as mist-netting, are invasive and limited in scope, highlighting the need for non-intrusive alternatives. In this work, we present a portable multisensor platform designed to operate in underground habitats. The system captures multimodal data, including near-infrared (NIR) imagery, ultrasonic audio, 3D structural data, and RGB video. Focusing on NIR imagery, we evaluate the effectiveness of the YOLO object detection framework for automated bat detection and counting. Experiments were conducted using a dataset of NIR images collected in natural shelters. Three YOLO variants (v10, v11, and v12) were trained and tested on this dataset. The models achieved high detection accuracy, with YOLO v12m reaching a mean average precision (mAP) of 0.981. These results demonstrate that combining NIR imaging with deep learning enables accurate and non-invasive monitoring of bats in challenging environments. The proposed approach offers a scalable tool for ecological research and conservation, supporting population assessment and behavioral studies without disturbing bat colonies. Full article
Show Figures

Figure 1

Back to TopTop