Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,605)

Search Parameters:
Keywords = multi-modality sensors

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 20117 KB  
Article
Intelligent Corrosion Diagnosis of High-Strength Bolts Based on Multi-Modal Feature Fusion and APO-XGBoost
by Hanyue Zhang, Yin Wu, Bo Sun, Yanyi Liu and Wenbo Liu
Sensors 2026, 26(8), 2520; https://doi.org/10.3390/s26082520 (registering DOI) - 19 Apr 2026
Abstract
High-strength bolts are critical structural components that are highly susceptible to corrosion in complex environments, posing significant threats to structural safety and reliability. Although acoustic emission (AE) technology has been widely applied in structural health monitoring, existing studies mainly focus on damage mode [...] Read more.
High-strength bolts are critical structural components that are highly susceptible to corrosion in complex environments, posing significant threats to structural safety and reliability. Although acoustic emission (AE) technology has been widely applied in structural health monitoring, existing studies mainly focus on damage mode identification or source localization, while the identification of corrosion evolution stages based on AE signals remains insufficient. This study develops an intelligent corrosion diagnosis framework for high-strength bolts by integrating multimodal feature fusion and optimized machine learning. AE signals are first collected from the near-end and far-end of bolts using a wireless sensor network and then transformed into time–frequency representations via continuous wavelet transform (CWT). The resulting time–frequency images are fed into a modified ResNet-18 network to extract deep features, while statistical features are simultaneously extracted from the raw signals to preserve global information. These heterogeneous features are subsequently fused to form a comprehensive representation of corrosion characteristics. Furthermore, an artificial protozoa optimizer (APO) is introduced to adaptively optimize the hyperparameters of the XGBoost model. The results demonstrate that AE signals generated by hammering bolts with different corrosion levels can be successfully distinguished. The proposed method achieves high accuracy in corrosion stage classification and outperforms conventional approaches. Even when evaluated on an additional M30 bolt dataset, the proposed method maintains robust performance, demonstrating excellent generalization capability across different bolt sizes. These results demonstrate the practical potential of the proposed method for intelligent bolt corrosion diagnosis. Full article
(This article belongs to the Section Fault Diagnosis & Sensors)
Show Figures

Figure 1

19 pages, 4121 KB  
Technical Note
drone2report: A Configuration-Driven Multi-Sensor Batch-Processing Engine for UAV-Based Plot Analysis in Precision Agriculture
by Nelson Nazzicari, Giulia Moscatelli, Agostino Fricano, Elisabetta Frascaroli, Roshan Paudel, Eder Groli, Paolo De Franceschi, Giorgia Carletti, Nicolò Franguelli and Filippo Biscarini
Drones 2026, 10(4), 301; https://doi.org/10.3390/drones10040301 (registering DOI) - 18 Apr 2026
Abstract
Unmanned aerial vehicles (UAVs) have become indispensable tools in precision agriculture and plant phenotyping, enabling the rapid, non-destructive assessment of crop traits across space and time. Equipped with RGB, multispectral, thermal, and other sensors, UAVs provide detailed information on canopy structure, physiology, and [...] Read more.
Unmanned aerial vehicles (UAVs) have become indispensable tools in precision agriculture and plant phenotyping, enabling the rapid, non-destructive assessment of crop traits across space and time. Equipped with RGB, multispectral, thermal, and other sensors, UAVs provide detailed information on canopy structure, physiology, and stress responses that can guide management decisions and accelerate breeding programs. Despite these advances, the downstream processing of UAV imagery remains technically demanding. Converting orthomosaics into standardized, biologically meaningful data often requires a combination of photogrammetry, geospatial analysis, and custom scripting, which can limit reproducibility and accessibility across research groups. We present drone2report, an open-source python-based software that processes orthomosaics from UAV flights to generate vegetation indices, summary statistics, derived subimages, and text (html) reports, supporting both research and applied crop breeding needs. Alongside the basic structure and functioning of drone2report, we also present five case studies that illustrate practical applications common in UAV-/drone-phenotyping of plants: (i) thresholding to remove background noise and highlight regions of interest; (ii) monitoring plant phenotypes over time; (iii) extracting information on plant height to detect events like lodging or the falling over of spikes; (iv) integrating multiple sensors (cameras) to construct and optimize new synthetic indices; (v) integrate a trained deep learning network to implement a classification task. These examples demonstrate the tool’s ability to automate analysis, integrate heterogeneous data and models, and support reproducible computation of agronomically relevant traits. drone2report streamlines orthorectified UAV-image processing for precision agriculture by linking orthomosaics to standardized, plot-level outputs. Its modular, configuration-driven design allows transparent workflows, easy customization, and integration of multiple sensors within a unified analytical framework. By facilitating reproducible, multi-modal image analysis, drone2report lowers technical barriers to UAV-based phenotyping and opens the way to robust, data-driven crop monitoring and breeding applications. Full article
(This article belongs to the Special Issue Advances in UAV-Based Remote Sensing for Climate-Smart Agriculture)
Show Figures

Figure 1

29 pages, 24864 KB  
Article
Improving the Robustness of Odour Recognition with Odour-Image Data Fusion in Open-Air Settings
by Fanny Monori and Alin Tisan
Sensors 2026, 26(8), 2493; https://doi.org/10.3390/s26082493 - 17 Apr 2026
Abstract
Odour recognition with low-cost gas sensors is challenging in open-air settings due to the non-specificity of the sensors and environmental variability. This can be mitigated by incorporating additional information into the classification process. This paper investigates odour-image multimodality in two case-studies of increasing [...] Read more.
Odour recognition with low-cost gas sensors is challenging in open-air settings due to the non-specificity of the sensors and environmental variability. This can be mitigated by incorporating additional information into the classification process. This paper investigates odour-image multimodality in two case-studies of increasing complexity: banana ripening in open-air environment and strawberry ripening in a glasshouse environment. Data were collected using custom acquisition platforms equipped with cameras and MOX gas sensors operated with temperature modulation. For the visual modality, image classification (MobileNetV3) and object detection (YoloV5) models are trained. For the odour modality, established classical machine learning methods (Random Forest, XGBoost, SVM and Logistic Regression) and neural networks (1D-CNN, LSTM, MLP, and ELM) are employed. Each modality is analysed independently and together to critically assess scenarios in which combining modalities provides a clear advantage over using either modality alone. Results show that models trained on odour data achieve high accuracy in controlled environments but underperform in more dynamic open-air settings. Image-based models are sensitive to the image quality in all environments; however, they are more robust when deployed in different environments. Lastly, it is demonstrated that decision fusion consistently increases the accuracy, by as much as +12.36% in the banana ripening and +3.63% in the strawberry ripening scenario. Where decision fusion does not improve classification accuracy significantly, it is shown that the multimodal approach can still be leveraged to identify high-confidence predictions by selecting samples where both modalities agree on the label. Full article
(This article belongs to the Special Issue Recent Advances in Gas Sensors)
25 pages, 1098 KB  
Review
Applications of Heart Rate Variability Metrics in Wearable Sensor Technologies: A Comprehensive Review
by Emi Yuda
Electronics 2026, 15(8), 1707; https://doi.org/10.3390/electronics15081707 - 17 Apr 2026
Abstract
Heart rate variability (HRV) has emerged as a key biomarker for assessing autonomic nervous system activity, stress, fatigue, and emotional states. With the rapid development of wearable sensor technologies, HRV analysis has expanded from clinical environments to real-world, continuous monitoring. This review summarizes [...] Read more.
Heart rate variability (HRV) has emerged as a key biomarker for assessing autonomic nervous system activity, stress, fatigue, and emotional states. With the rapid development of wearable sensor technologies, HRV analysis has expanded from clinical environments to real-world, continuous monitoring. This review summarizes current applications of HRV metrics in wearable devices, including fitness tracking, mental stress assessment, sleep quality evaluation, and early detection of physiological or psychological disorders. Recent advances in photoplethysmography (PPG)-based HRV estimation have enabled noninvasive and user-friendly measurement, though challenges remain in accuracy under motion and variable environmental conditions. We also discuss methodological considerations, such as artifact correction, data segmentation, and the integration of HRV with other biosignals for multimodal analysis. Emerging research suggests that combining HRV with metrics such as respiration rate, skin conductance, and accelerometry can enhance robustness and interpretability in dynamic settings. Finally, future directions are proposed toward personalized health analytics, emotion-aware computing, and real-time adaptive feedback systems. This review highlights the growing potential of wearable HRV analysis as a foundation for preventive healthcare and human–machine symbiosis. Full article
(This article belongs to the Special Issue Smart Devices and Wearable Sensors: Recent Advances and Prospects)
36 pages, 1788 KB  
Article
A Blockchain-Integrated IoT–BIM Platform for Real-Time Carbon Monitoring in Modular Integrated Construction
by Yiyu Zhao, Yaning Zhang, Xiaohan Wu, Xinping Wen, Chen Chen, Yue Teng and Man Piu Ben Lau
Buildings 2026, 16(8), 1587; https://doi.org/10.3390/buildings16081587 - 17 Apr 2026
Abstract
Modular integrated construction (MiC) is an innovative construction method that shifts on-site activities to a controlled factory environment, thereby offering sustainability benefits. However, current carbon management relies on labor-intensive manual data collection, causing delayed and inaccurate carbon accounting that increases greenwashing risks. Existing [...] Read more.
Modular integrated construction (MiC) is an innovative construction method that shifts on-site activities to a controlled factory environment, thereby offering sustainability benefits. However, current carbon management relies on labor-intensive manual data collection, causing delayed and inaccurate carbon accounting that increases greenwashing risks. Existing approaches lack real-time, automated, and trustworthy carbon tracking capabilities across fragmented supply chains. This study develops and validates the Blockchain-enabled IoT-BIM Platform (BIBP), which combines Internet of Things (IoT), Building Information Modeling (BIM), and blockchain for real-time carbon monitoring. IoT sensors automate data capture from construction equipment and BIM provides spatial visualization of carbon at the module and building levels. A Hyperledger Fabric blockchain ensures the authenticity, immutability, and traceability of carbon records. Validated on a 15-story MiC project in Hong Kong, BIBP established a cradle-to-end-of-construction baseline of 949.84 kgCO2e/m2, identifying steel and concrete as the primary hotspots (80% of material emissions). Real-time analytics demonstrated that combining high-volume ground granulated blast furnace slag (GGBS) concrete substitution, new energy sea–land multimodal transport, and 10% steel waste reduction achieves over 20% carbon savings. Furthermore, the BIBP automated data acquisition and calculation, improving assessment efficiency by 92.4%. The platform demonstrates the potential to transform carbon management from a static, retrospective evaluation into a proactive, data-driven monitoring process, equipping stakeholders with a tool to dynamically track emissions and make timely interventions toward carbon reduction targets. Full article
29 pages, 3416 KB  
Article
Enhancing Collaborative AI Learning: A Blockchain-Secured, Edge-Enabled Platform for Multimodal Education in IIoT Environments
by Ahsan Rafiq, Eduard Melnik, Alexey Samoylov, Alexander Kozlovskiy and Irina Safronenkova
Big Data Cogn. Comput. 2026, 10(4), 123; https://doi.org/10.3390/bdcc10040123 - 17 Apr 2026
Abstract
As industries deploy more connected devices in factories, warehouses, and smart facilities, the need for artificial intelligence (AI) systems that can operate securely in distributed, data-intensive environments is growing. Traditional centralized learning and online education platforms struggle when students and systems have to [...] Read more.
As industries deploy more connected devices in factories, warehouses, and smart facilities, the need for artificial intelligence (AI) systems that can operate securely in distributed, data-intensive environments is growing. Traditional centralized learning and online education platforms struggle when students and systems have to process real-time streams (sensors, video, text) with strict latency and privacy requirements. To address this challenge, a blockchain-secured, edge-enabled multimodal federated learning framework tailored for Industrial IoT (IIoT) environments is proposed. The model integrates four key layers: (i) a blockchain layer that provides credentialing, transparency, and token-based incentives; (ii) a multimodal community layer that supports group formation, peer consensus, and cross-modal learning across text, images, audio, and sensor data; (iii) an edge computing layer that enables low-latency task offloading and secure training within Intel SGX enclaves; and (iv) a data layer that applies pre-processing, differential privacy, and synthetic augmentation to safeguard sensitive information. Experiments on industrial multimodal datasets demonstrate 42% faster model aggregation, 78.9% multimodal accuracy, and 1.9% accuracy loss under ε = 1.0 differential privacy. This shows a scalable and practical path for decentralized AI training in next-generation IIoT systems, confirming the possibility of technical support for educational processes. However, the conducted research requires a validation of pedagogical effectiveness. Full article
Show Figures

Figure 1

35 pages, 6272 KB  
Article
AI-Enhanced Thermal–Visual–Inertial Odometry and Autonomous Planning for GPS-Denied Search-and- Rescue Robotics
by Islam T. Almalkawi, Sabya Shtaiwi, Alaa Alhowaide and Manel Guerrero Zapata
Sensors 2026, 26(8), 2462; https://doi.org/10.3390/s26082462 - 16 Apr 2026
Viewed by 146
Abstract
Search and rescue (SAR) missions in collapsed or underground environments remain challenging due to GPS unavailability, which hinders localization and autonomous navigation. Systems that rely on single-sensor inputs or structured settings often degrade under smoke, dust, or dynamic clutter. This paper presents an [...] Read more.
Search and rescue (SAR) missions in collapsed or underground environments remain challenging due to GPS unavailability, which hinders localization and autonomous navigation. Systems that rely on single-sensor inputs or structured settings often degrade under smoke, dust, or dynamic clutter. This paper presents an autonomous ground robot for GPS-denied SAR that integrates low-cost thermal, visual, inertial, and acoustic cues within a unified, computation-efficient architecture. The stack combines Thermal–Visual Odometry (TV–VO) with Zero-Velocity Updates (ZUPT) for drift-resistant localization, RescueGraph for multimodal survivor detection, and a Proximal Policy Optimization (PPO) planner for adaptive navigation under uncertainty. Across simulated disaster scenarios and benchmark corridor runs, the system shows embedded-feasible runtime behavior and supports return to base without external beacons under the evaluated conditions. Quantitatively, TV–VO+ZUPT reduces drift in short internal evaluations, while RescueGraph attains an F1-score of 0.6923 and an area under the ROC curve (AUC) of 0.976 for survivor detection. At the system level, the integrated navigation stack achieves full mission completion in the reported SAR-style trials, while the separate A*/PPO comparison highlights a trade-off between completion rate, traversal time, and collisions. Overall, the results support the practical promise of a low-cost sensor-fusion and learning-assisted navigation framework for GPS-denied SAR robotics. Full article
(This article belongs to the Section Sensors and Robotics)
Show Figures

Graphical abstract

58 pages, 4676 KB  
Review
Vision-Based Artificial Intelligence for Adaptive Peen Forming: Sensing Architectures, Learning Models, and Closed-Loop Smart Manufacturing
by Sehar Shahzad Farooq, Abdul Rehman, Fuad Ali Mohammed Al-Yarimi, Sejoon Park, Jaehyun Baik and Hosu Lee
Sensors 2026, 26(8), 2460; https://doi.org/10.3390/s26082460 - 16 Apr 2026
Viewed by 166
Abstract
Peen forming is a dieless manufacturing process used to shape large, thin aerospace panels through controlled shot impacts that induce residual stresses and curvature. Despite long-standing industrial use, process monitoring still depends largely on indirect proxies such as Almen intensity and coverage, limiting [...] Read more.
Peen forming is a dieless manufacturing process used to shape large, thin aerospace panels through controlled shot impacts that induce residual stresses and curvature. Despite long-standing industrial use, process monitoring still depends largely on indirect proxies such as Almen intensity and coverage, limiting spatially resolved deformation assessment and hindering closed-loop control. In parallel, vision-based artificial intelligence (AI) has enabled adaptive monitoring and feedback in smart-manufacturing domains such as welding, additive manufacturing, and sheet forming. This review examines how such sensing and learning strategies can be transferred to adaptive peening forming. We compare six vision sensing modalities and assess major AI model families for surface mapping, temporal prediction, robustness, and deployment maturity. The synthesis shows that progress is primarily constrained by limited validated datasets, harsh in-cabinet sensing conditions, scarce closed-loop demonstrations, and weak validation on curved aerospace geometries. We conclude that the sensing and AI foundations for adaptive peen forming are already emerging, but industrial translation now depends on stronger experimental validation, standardized benchmarking, robust multi-sensor integration, and edge-capable feedback pipelines. Full article
(This article belongs to the Special Issue Artificial Intelligence and Sensing Technology in Smart Manufacturing)
26 pages, 956 KB  
Article
Environment-Guided Multimodal Pest Detection and Risk Assessment in Fruit and Vegetable Production Systems
by Jiapeng Sun, Yucheng Peng, Zhimeng Zhang, Wenrui Xu, Boyuan Xi, Yuanying Zhang and Yihong Song
Horticulturae 2026, 12(4), 486; https://doi.org/10.3390/horticulturae12040486 - 16 Apr 2026
Viewed by 82
Abstract
Aimed at the practical challenge that pest occurrence in fruit and vegetable horticultural production exhibits strong environmental dependency, pronounced stage characteristics, and high sensitivity to control decision-making, a multimodal pest recognition and occurrence risk joint modeling method is proposed to address the limitation [...] Read more.
Aimed at the practical challenge that pest occurrence in fruit and vegetable horticultural production exhibits strong environmental dependency, pronounced stage characteristics, and high sensitivity to control decision-making, a multimodal pest recognition and occurrence risk joint modeling method is proposed to address the limitation that conventional intelligent plant protection systems focus primarily on pest identification while lacking risk discrimination capability. Within a unified network framework, pest visual information and environmental temporal data are integrated through the construction of an environment-guided representation learning mechanism, a recognition–risk joint optimization strategy, and a risk-aware decision representation modeling structure. In this manner, pest category recognition and occurrence risk evaluation are conducted simultaneously, thereby providing direct decision support for precision prevention and control in fruit and vegetable production. Systematic experimental evaluation is conducted based on multi-crop and multi-year field data collected from Wuyuan County, Bayannur City, Inner Mongolia. Overall comparative results demonstrate that an identification accuracy of 0.947, a precision of 0.936, and a recall of 0.924 are achieved on the test set, all of which significantly outperform mainstream visual detection models such as YOLOv8, DETR, and Mask R-CNN. In terms of detection performance, mAP@50 and mAP@75 reach 0.962 and 0.821, respectively, indicating stable localization and discrimination capability under complex backgrounds and dense small-target conditions. For the occurrence risk discrimination task, a risk accuracy of 0.887 is obtained, representing an improvement of approximately 4.5 percentage points compared with the simple multimodal feature concatenation method. Cross-crop, cross-site, and cross-year generalization experiments further show that risk accuracy remains above 0.84 with stable recognition performance under significant distribution shifts. Ablation studies verify the synergistic contributions of the proposed core modules to overall performance improvement. The results indicate that the proposed framework enables the transition from single recognition to risk-driven plant protection decision-making, providing a technically viable pathway for pest diagnosis and control strategy optimization in fruit and vegetable horticulture. Full article
30 pages, 2314 KB  
Article
Confidence-Aware Gated Multimodal Fusion for Robust Temporal Action Localization in Occluded Environments
by Masato Takami and Tomohiro Fukuda
Sensors 2026, 26(8), 2454; https://doi.org/10.3390/s26082454 - 16 Apr 2026
Viewed by 154
Abstract
In industrial environments, robust Temporal Action Localization (TAL) is essential; however, frequent occlusions often compromise the reliability of skeletal data, leading to negative transfer in multimodal fusion. To address this challenge, we propose a Gated Skeleton Refinement Module (Gated SRM), a universal front-end [...] Read more.
In industrial environments, robust Temporal Action Localization (TAL) is essential; however, frequent occlusions often compromise the reliability of skeletal data, leading to negative transfer in multimodal fusion. To address this challenge, we propose a Gated Skeleton Refinement Module (Gated SRM), a universal front-end preprocessing module that explicitly incorporates OpenPose confidence scores into the network architecture. By applying these scores as a logarithmic bias within a self-attention mechanism, our method achieves soft suppression—dynamically attenuating the attention weights assigned to unreliable joints—before adaptively fusing the refined skeletal features with RGB representations through a learnable gating network. Extensive experiments on the heavily occluded IKEA ASM dataset demonstrate that our approach effectively prevents the catastrophic accuracy degradation typical of naive and established multimodal fusion strategies, improving the mean Average Precision (mAP) to 21.77%, maintaining parity with the RGB-only baseline while demonstrating superior robustness. Furthermore, the system maintains a practical end-to-end inference speed of approximately 9.2 frames per second (FPS), which is sufficient for monitoring macro-level industrial workflows. By prioritizing confidence-based data selection over data restoration, this sensor-metadata-driven architecture offers a robust and principled approach acting as a critical fail-safe and safety-net for real-world action recognition under occlusion. Full article
26 pages, 6550 KB  
Article
Clinical Thermography of the Diabetic Foot Using a Low-Cost Thermal Camera: Processing and Instrumental Framework
by Vanéva Chingan-Martino, Mériem Allali, Stéphane Henri, El Hadji Mama Guène, Dominique Gibert and Antoine Chéret
Sensors 2026, 26(8), 2438; https://doi.org/10.3390/s26082438 - 16 Apr 2026
Viewed by 252
Abstract
Infrared thermography is a non-contact tool for monitoring inflammatory processes in the diabetic foot, but quantitative bedside use remains challenging with low-cost thermal infrared cameras due to radiometric drift, non-uniformity (vignetting), geometric distortions, and visible–thermal parallax. This paper presents an end-to-end clinical and [...] Read more.
Infrared thermography is a non-contact tool for monitoring inflammatory processes in the diabetic foot, but quantitative bedside use remains challenging with low-cost thermal infrared cameras due to radiometric drift, non-uniformity (vignetting), geometric distortions, and visible–thermal parallax. This paper presents an end-to-end clinical and instrumental framework built around a cheap thermal camera to ensure reproducible acquisition and physically consistent temperature estimation. The approach combines a standardized mobile acquisition setup and measurement protocol, extraction of embedded radiometric data from raw images, radiometric inversion with atmospheric correction, vignette correction performed in the radiometric domain, and geometric calibration of both visible and infrared sensors using dedicated (thermal) calibration targets. Accurate visible–infrared registration is obtained from hybrid heated markers, enabling reliable overlay and downstream analysis. The full processing chain yields quantitative thermograms with radiometric errors below 0.15 °C and sub-pixel multimodal alignment, supporting the detection of clinically relevant plantar temperature asymmetries and paving the way for routine calibrated low-cost thermography in diabetic foot care. Full article
(This article belongs to the Collection Biomedical Imaging and Sensing)
Show Figures

Figure 1

30 pages, 711 KB  
Article
Artificial Intelligence-Driven Multimodal Sensor Fusion for Complex Market Systems via Federated Transformer-Based Learning
by Lei Shi, Mingran Tian, Yinfei Yi, Xinyi Hu, Xiaoya Wang, Yating Yang and Manzhou Li
Sensors 2026, 26(8), 2418; https://doi.org/10.3390/s26082418 - 15 Apr 2026
Viewed by 136
Abstract
In highly digitalized and networked modern trading systems, large volumes of heterogeneous data are continuously generated from multiple sources during market operations. However, due to the complexity of data structures, significant differences in temporal scales, and constraints imposed by data privacy protection, traditional [...] Read more.
In highly digitalized and networked modern trading systems, large volumes of heterogeneous data are continuously generated from multiple sources during market operations. However, due to the complexity of data structures, significant differences in temporal scales, and constraints imposed by data privacy protection, traditional single-source modeling approaches are unable to fully exploit multisource information. To address this issue, a federated multimodal prediction framework for complex market systems, termed Federated Market-Sensor Transformer (FMST), is proposed. In this framework, data originating from different information sources are uniformly modeled as multimodal time series. A multimodal market-sensor representation module is constructed to perform unified feature encoding, and a cross-modal Transformer fusion architecture is employed to characterize dynamic interaction relationships among different information sources. Meanwhile, a federated collaborative learning mechanism is introduced during the training phase, enabling multiple data nodes to perform collaborative model optimization without sharing raw data. In this manner, data privacy can be preserved while improving the cross-region generalization capability of the model. Systematic experimental evaluation is conducted on the constructed multimodal market-sensor dataset. The experimental results demonstrate that the proposed method consistently outperforms traditional statistical models and deep learning approaches across multiple evaluation metrics. In the main prediction experiment, FMST achieves a root mean square error (RMSE) of 0.1136, a mean absolute error (MAE) of 0.0832, and a coefficient of determination R2 of 0.8517, while the direction prediction accuracy reaches 74.56%, clearly outperforming baseline models including ARIMA, LSTM, Temporal CNN, Transformer, and FedAvg-LSTM. In the cross-region generalization experiment, FMST maintains strong performance, achieving an RMSE of 0.1242, an MAE of 0.0908, an R2 value of 0.8261, and a direction prediction accuracy of 72.48%. The ablation study further indicates that the three core components—multimodal market-sensor representation, cross-modal Transformer fusion, and federated collaborative learning—each make important contributions to the overall model performance. These experimental findings demonstrate that the proposed method can effectively integrate multisource market information and significantly enhance the prediction capability for complex market dynamics, providing a new technical pathway for the application of artificial intelligence-driven multimodal sensing systems in economic data analysis. Full article
(This article belongs to the Special Issue Artificial Intelligence-Driven Sensing)
18 pages, 1217 KB  
Article
Detect and Repair: Robust Self-Supervised Wearable Sensing Under Missing Modalities
by Aboul Hassane Cisse and Shoya Ishimaru
Sensors 2026, 26(8), 2419; https://doi.org/10.3390/s26082419 - 15 Apr 2026
Viewed by 181
Abstract
Wearable sensor systems are being increasingly deployed in real-world environments to monitor human activities and cognitive states. However, such systems frequently suffer from degraded or missing sensor modalities due to occlusions, energy constraints, or hardware failures. In this work, we introduce CognifySSL v2.0, [...] Read more.
Wearable sensor systems are being increasingly deployed in real-world environments to monitor human activities and cognitive states. However, such systems frequently suffer from degraded or missing sensor modalities due to occlusions, energy constraints, or hardware failures. In this work, we introduce CognifySSL v2.0, a self-supervised learning framework designed to detect and repair missing modalities in real time under simulated real-world missing-modality conditions. The model combines contrastive and masked modeling objectives across multiple physiological and motion signals (e.g., IMU, ECG, EDA) using a fusion architecture with dropout simulation. Evaluation on WESAD demonstrated effective multimodal detection and reconstruction under missing-modality conditions, while experiments on MobiAct validated unimodal robustness and representation learning under sensor dropout. We released our code and interactive visualization dashboard to support reproducibility and future research on robust multimodal fusion. Full article
35 pages, 1113 KB  
Article
Intelligent UAV-UGV-SN Systems for Monitoring and Avoiding Wildfires in Context of Sustainable Development of Smart Regions
by Dmytro Korniienko, Nazar Serhiichuk, Vyacheslav Kharchenko, Herman Fesenko, Jose Borges and Nikolaos Bardis
Sustainability 2026, 18(8), 3908; https://doi.org/10.3390/su18083908 - 15 Apr 2026
Viewed by 212
Abstract
Advancing environmental monitoring through coordinated autonomous systems is central to sustainable smart region governance and data-driven territorial management. The article presents an engineering-oriented architecture and deployment methodology for an integrated wildfire monitoring and response system that combines unmanned aerial vehicles (UAVs), unmanned ground [...] Read more.
Advancing environmental monitoring through coordinated autonomous systems is central to sustainable smart region governance and data-driven territorial management. The article presents an engineering-oriented architecture and deployment methodology for an integrated wildfire monitoring and response system that combines unmanned aerial vehicles (UAVs), unmanned ground vehicles (UGVs), and stationary sensor networks (SNs). We formalise hub-and-spoke infrastructure placement as a mixed-integer optimisation problem that accounts for platform types, endurance, travel times and logistical constraints, and propose a practical pre-processing pipeline (confidence scoring, resampling, Kalman/median filtering, strategy fusion) for heterogeneous telemetry and imagery. The system couples multimodal neural network processing (image backbones, clustering and time-series models) with online resource-allocation and mission-planning mechanisms to prioritise UAV/UGV sorties and dynamically select launch sites. The article describes scenario-driven operational modes (early warning, alarm verification, autonomous local extinguishing, post-fire recovery, sensor-gap compensation, and inter-hub reinforcement), defines validation protocols (synthetic experiments, precision/recall/F1, and hardware-in-the-loop testing), and proposes KPIs to assess environmental, social, and economic impacts for smart regions. The contribution is a reproducible, deployment-focused blueprint that bridges conceptual UAV–UGV–SN research and practical implementation, highlighting trade-offs in reliability, communication redundancy, and sustainability, and outlining directions for simulation, field pilots and algorithmic refinement. Full article
23 pages, 9927 KB  
Article
A Relative Orbital Motion-Guided Framework for Generating Multimodal Visual Data of Spacecraft
by Wanyun Li, Yurong Huo, Qinyu Zhu, Yao Lu, Yuqiang Fang and Yasheng Zhang
Remote Sens. 2026, 18(8), 1177; https://doi.org/10.3390/rs18081177 - 15 Apr 2026
Viewed by 194
Abstract
The advancement of on-orbit servicing and space debris removal missions has established high-precision visual perception for non-cooperative spacecraft as a key research focus. However, the availability of high-quality, diverse spacecraft image datasets is severely limited due to extreme on-orbit imaging conditions, data confidentiality, [...] Read more.
The advancement of on-orbit servicing and space debris removal missions has established high-precision visual perception for non-cooperative spacecraft as a key research focus. However, the availability of high-quality, diverse spacecraft image datasets is severely limited due to extreme on-orbit imaging conditions, data confidentiality, and morphological diversity of targets, significantly constraining the advancement of data-driven algorithms in this domain. To address this challenge, we propose a relative orbital motion-guided framework for generating multimodal visual data of spacecraft. The proposed method integrates an orbital dynamics model into the synthetic data generation pipeline to simulate typical relative motion patterns between the camera and the target in a realistic orbital environment, thereby generating image sequences characterized by continuous spatiotemporal evolution. Targeting four representative spacecraft—Tiangong, Spacedragon, ICESat, and Cassini—this work simultaneously generates a dataset comprising 8000 samples, each containing four strictly aligned modalities: RGB images, instance segmentation masks, depth maps, and surface normal maps, along with precise 6-degree-of-freedom (6-DoF) pose ground truth. Furthermore, an end-to-end physical image degradation model is developed to accurately simulate the complete imaging chain—from optical diffraction and aberrations to sensor sampling and noise—thereby effectively narrowing the domain gap between synthetic and real data. By addressing three key aspects—physical motion modeling, synchronous multimodal ground truth, and imaging degradation simulation—this work provides a crucial data foundation for training, testing, and validating data-driven on-orbit perception algorithms. Full article
Show Figures

Figure 1

Back to TopTop