Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (474)

Search Parameters:
Keywords = spatiotemporal representation

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 5744 KB  
Article
Decoupling Rainfall and Surface Runoff Effects Based on Spatio-Temporal Spectra of Wireless Channel State Information
by Hao Li, Yin Long and Tehseen Zia
Electronics 2025, 14(20), 4102; https://doi.org/10.3390/electronics14204102 - 20 Oct 2025
Abstract
Leveraging ubiquitous wireless signals for environmental sensing provides a highly promising pathway toward constructing low-cost and high-density flood monitoring systems. However, in real-world flood scenarios, the wireless channel is simultaneously affected by rainfall-induced signal attenuation and complex multipath effects caused by surface runoff [...] Read more.
Leveraging ubiquitous wireless signals for environmental sensing provides a highly promising pathway toward constructing low-cost and high-density flood monitoring systems. However, in real-world flood scenarios, the wireless channel is simultaneously affected by rainfall-induced signal attenuation and complex multipath effects caused by surface runoff (water accumulation). These two physical phenomena become intertwined in the received signals, resulting in severe feature ambiguity. This not only greatly limits the accuracy of environmental sensing but also hinders communication systems from performing effective channel compensation. How to disentangle these combined effects from a single wireless link represents a fundamental scientific challenge for achieving high-precision wireless environmental sensing and ensuring communication reliability under harsh conditions. To address this challenge, we propose a novel signal processing framework that aims to effectively decouple the effects of rainfall and surface runoff from Channel State Information (CSI) collected using commercial Wi-Fi devices. The core idea of our method lies in first constructing a two-dimensional CSI spatiotemporal spectrogram from continuously captured multicarrier CSI data. This spectrogram enables high-resolution visualization of the unique “fingerprints” of different physical effects—rainfall manifests as smooth background attenuation, whereas surface runoff appears as sparse high-frequency textures. Building upon this representation, we design and implement a Dual-Decoder Convolutional Autoencoder deep learning model. The model employs a shared encoder to learn the mixed CSI features, while two distinct decoder branches are responsible for reconstructing the global background component attributed to rainfall and the local texture component associated with surface runoff, respectively. Based on the decoupled signal components, we achieve simultaneous and highly accurate estimation of rainfall intensity (mean absolute error below 1.5 mm/h) and surface water accumulation (detection accuracy of 98%). Furthermore, when the decoupled and refined channel estimates are applied to a communication receiver for channel equalization, the Bit Error Rate (BER) is reduced by more than one order of magnitude compared to conventional equalization methods. Full article
Show Figures

Figure 1

20 pages, 6483 KB  
Article
Loop-MapNet: A Multi-Modal HDMap Perception Framework with SDMap Dynamic Evolution and Priors
by Yuxuan Tang, Jie Hu, Daode Zhang, Wencai Xu, Feiyu Zhao and Xinghao Cheng
Appl. Sci. 2025, 15(20), 11160; https://doi.org/10.3390/app152011160 - 17 Oct 2025
Viewed by 171
Abstract
High-definition maps (HDMaps) are critical for safe autonomy on structured roads. Yet traditional production—relying on dedicated mapping fleets and manual quality control—is costly and slow, impeding large-scale, frequent updates. Recently, standard-definition maps (SDMaps) derived from remote sensing have been adopted as priors to [...] Read more.
High-definition maps (HDMaps) are critical for safe autonomy on structured roads. Yet traditional production—relying on dedicated mapping fleets and manual quality control—is costly and slow, impeding large-scale, frequent updates. Recently, standard-definition maps (SDMaps) derived from remote sensing have been adopted as priors to support HDMap perception, lowering cost but struggling with subtle urban changes and localization drift. We propose Loop-MapNet, a self-evolving, multimodal, closed-loop mapping framework. Loop-MapNet effectively leverages surround-view images, LiDAR point clouds, and SDMaps; it fuses multi-scale vision via a weighted BiFPN, and couples PointPillars BEV and SDMap topology encoders for cross-modal sensing. A Transformer-based bidirectional adaptive cross-attention aligns SDMap with online perception, enabling robust fusion under heterogeneity. We further introduce a confidence-guided masked autoencoder (CG-MAE) that leverages confidence and probabilistic distillation to both capture implicit SDMap priors and enhance the detailed representation of low-confidence HDMap regions. With spatiotemporal consistency checks, Loop-MapNet incrementally updates SDMaps to form a perception–mapping–update loop, compensating remote-sensing latency and enabling online map optimization. On nuScenes, within 120 m, Loop-MapNet attains 61.05% mIoU, surpassing the best baseline by 0.77%. Under extreme localization errors, it maintains 60.46% mIoU, improving robustness by 2.77%; CG-MAE pre-training raises accuracy in low-confidence regions by 1.72%. These results demonstrate advantages in fusion and robustness, moving beyond one-way prior injection and enabling HDMap–SDMap co-evolution for closed-loop autonomy and rapid SDMap refresh from remote sensing. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

19 pages, 2109 KB  
Article
SF6 Leak Detection in Infrared Video via Multichannel Fusion and Spatiotemporal Features
by Zhiwei Li, Xiaohui Zhang, Zhilei Xu, Yubo Liu and Fengjuan Zhang
Appl. Sci. 2025, 15(20), 11141; https://doi.org/10.3390/app152011141 - 17 Oct 2025
Viewed by 95
Abstract
With the development of infrared imaging technology and the integration of intelligent algorithms, the realization of non-contact, dynamic and real-time detection of SF6 gas leakage based on infrared video has been a significant research direction. However, the existing real-time detection algorithms exhibit low [...] Read more.
With the development of infrared imaging technology and the integration of intelligent algorithms, the realization of non-contact, dynamic and real-time detection of SF6 gas leakage based on infrared video has been a significant research direction. However, the existing real-time detection algorithms exhibit low accuracy in detecting SF6 leakage and are susceptible to noise, which makes it difficult to meet the actual needs of engineering. To address this problem, this paper proposes a real-time SF6 leakage detection method, VGEC-Net, based on multi-channel fusion and spatiotemporal feature extraction. The proposed method first employs the ViBe-GMM algorithm to extract foreground masks, which are then fused with infrared images to construct a dual-channel input. In the backbone network, a CE-Net structure—integrating CBAM and ECA-Net—is combined with the P3D network to achieve efficient spatiotemporal feature extraction. A Feature Pyramid Network (FPN) and a temporal Transformer module are further integrated to enhance multi-scale feature representation and temporal modeling, thereby significantly improving the detection performance for small-scale targets. Experimental results demonstrate that VGEC-Net achieves a mean average precision (mAP) of 61.7% on the dataset used in this study, with a mAP@50 of 87.3%, which represents a significant improvement over existing methods. These results validate the effectiveness and advancement of the proposed method for infrared video-based gas leakage detection. Furthermore, the model achieves 78.2 frames per second (FPS) during inference, demonstrating good real-time processing capability while maintaining high detection accuracy, exhibiting strong application potential. Full article
Show Figures

Figure 1

23 pages, 2593 KB  
Article
Robust Offline Reinforcement Learning Through Causal Feature Disentanglement
by Ao Ma, Peng Li and Xiaolong Su
Electronics 2025, 14(20), 4064; https://doi.org/10.3390/electronics14204064 - 16 Oct 2025
Viewed by 154
Abstract
Offline reinforcement learning suffers from critical vulnerability to data corruption from sensor noise or adversarial attacks. Recent research has achieved a lot by downweighting corrupted samples and fixing the corrupted data, while data corruption induces feature entanglement that undermines policy robustness. Existing methods [...] Read more.
Offline reinforcement learning suffers from critical vulnerability to data corruption from sensor noise or adversarial attacks. Recent research has achieved a lot by downweighting corrupted samples and fixing the corrupted data, while data corruption induces feature entanglement that undermines policy robustness. Existing methods fail to identify causal features behind performance degradation caused by corruption. To analyze causal relationships in corrupted data, we propose a method, Robust Causal Feature Disentanglement(RCFD). Our method introduces a learnable causal feature disentanglement mechanism specifically designed for reinforcement learning scenarios, integrating the CausalVAE framework to disentangle causal features governing environmental dynamics from corruption-sensitive non-causal features. Theoretically, this disentanglement confers a robustness advantage under data corruption conditions. Concurrently, causality-preserving perturbation training injects Gaussian noise solely into non-causal features to generate counterfactual samples and is enhanced by dual-path feature alignment and contrastive learning for representation invariance. A dynamic graph diagnostic module further employs graph convolutional attention networks to model spatiotemporal relationships and identify corrupted edges through structural consistency analysis, enabling precise data repair. The results exhibit highly robust performance across D4rl benchmarks under diverse data corruption conditions. This confirms that causal feature invariance helps bridge distributional gaps, promoting reliable deployment in complex real-world settings. Full article
Show Figures

Figure 1

18 pages, 1960 KB  
Article
CasDacGCN: A Dynamic Attention-Calibrated Graph Convolutional Network for Information Popularity Prediction
by Bofeng Zhang, Yanlin Zhu, Zhirong Zhang, Kaili Liao, Sen Niu, Bingchun Li and Haiyan Li
Entropy 2025, 27(10), 1064; https://doi.org/10.3390/e27101064 - 14 Oct 2025
Viewed by 307
Abstract
Information popularity prediction is a critical problem in social network analysis. With the increasing prevalence of social platforms, accurate prediction of the diffusion process has become increasingly important. Existing methods mainly rely on graph neural networks to model structural relationships, but they are [...] Read more.
Information popularity prediction is a critical problem in social network analysis. With the increasing prevalence of social platforms, accurate prediction of the diffusion process has become increasingly important. Existing methods mainly rely on graph neural networks to model structural relationships, but they are often insufficient in capturing the complex interplay between temporal evolution and local cascade structures, especially in real-world scenarios involving sparse or rapidly changing cascades. To address this issue, we propose the Cascading Dynamic attention-calibrated Graph Convolutional Network, named CasDacGCN. It enhances prediction performance through spatiotemporal feature fusion and adaptive representation learning. The model integrates snapshot-level local encoding, global temporal modeling, cross-attention mechanisms, and a hypernetwork-based sample-wise calibration strategy, enabling flexible modeling of multi-scale diffusion patterns. Results from experiments demonstrate that the proposed model consistently surpasses existing approaches on two real-world datasets, validating its effectiveness in popularity prediction tasks. Full article
Show Figures

Figure 1

21 pages, 2372 KB  
Article
IDG-ViolenceNet: A Video Violence Detection Model Integrating Identity-Aware Graphs and 3D-CNN
by Hong Huang and Qingping Jiang
Sensors 2025, 25(20), 6272; https://doi.org/10.3390/s25206272 - 10 Oct 2025
Viewed by 319
Abstract
Video violence detection plays a crucial role in intelligent surveillance and public safety, yet existing methods still face challenges in modeling complex multi-person interactions. To address this, we propose IDG-ViolenceNet, a dual-stream video violence detection model that integrates identity-aware spatiotemporal graphs with three-dimensional [...] Read more.
Video violence detection plays a crucial role in intelligent surveillance and public safety, yet existing methods still face challenges in modeling complex multi-person interactions. To address this, we propose IDG-ViolenceNet, a dual-stream video violence detection model that integrates identity-aware spatiotemporal graphs with three-dimensional convolutional neural networks (3D-CNN). Specifically, the model utilizes YOLOv11 for high-precision person detection and cross-frame identity tracking, constructing a dynamic spatiotemporal graph that encodes spatial proximity, temporal continuity, and individual identity information. On this basis, a GINEConv branch extracts structured interaction features, while an R3D-18 branch models local spatiotemporal patterns. The two representations are fused in a dedicated module for cross-modal feature integration. Experimental results show that IDG-ViolenceNet achieves accuracies of 97.5%, 99.5%, and 89.4% on the Hockey Fight, Movies Fight, and RWF-2000 datasets, respectively, significantly outperforming state-of-the-art methods. Additionally, ablation studies validate the contributions of key components in improving detection accuracy and robustness. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

22 pages, 724 KB  
Article
State of Health Estimation for Batteries Based on a Dynamic Graph Pruning Neural Network with a Self-Attention Mechanism
by Xuanyuan Gu, Mu Liu and Jilun Tian
Energies 2025, 18(20), 5333; https://doi.org/10.3390/en18205333 - 10 Oct 2025
Viewed by 409
Abstract
The accurate estimation of the state of health (SOH) of lithium-ion batteries is critical for ensuring the safety, reliability, and efficiency of modern energy storage systems. Traditional model-based and data-driven methods often struggle to capture complex spatiotemporal degradation patterns, leading to reduced accuracy [...] Read more.
The accurate estimation of the state of health (SOH) of lithium-ion batteries is critical for ensuring the safety, reliability, and efficiency of modern energy storage systems. Traditional model-based and data-driven methods often struggle to capture complex spatiotemporal degradation patterns, leading to reduced accuracy and robustness. To address these limitations, this paper proposes a novel dynamic graph pruning neural network with self-attention mechanism (DynaGPNN-SAM) for SOH estimation. The method transforms sequential battery features into graph-structured representations, enabling the explicit modeling of spatial dependencies among operational variables. A self-attention-guided pruning strategy is introduced to dynamically preserve informative nodes while filtering redundant ones, thereby enhancing interpretability and computational efficiency. The framework is validated on the NASA lithium-ion battery dataset, with extensive experiments and ablation studies demonstrating superior performance compared to conventional approaches. Results show that DynaGPNN-SAM achieves lower root mean square error (RMSE) and mean absolute error (MAE) values across multiple batteries, particularly excelling during rapid degradation phases. Overall, the proposed approach provides an accurate, robust, and scalable solution for real-world battery management systems. Full article
Show Figures

Figure 1

20 pages, 4466 KB  
Article
SA-STGCN: A Spectral-Attentive Spatio-Temporal Graph Convolutional Network for Wind Power Forecasting with Wavelet-Enhanced Multi-Scale Learning
by Yakai Yang, Zhenqing Liu and Zhongze Yu
Energies 2025, 18(19), 5315; https://doi.org/10.3390/en18195315 - 9 Oct 2025
Viewed by 415
Abstract
Wind power forecasting remains a major challenge for renewable energy integration, as conventional models often perform poorly when confronted with complex atmospheric dynamics. This study addresses the problem by developing a Spectral-Attentive Spatio-Temporal Graph Convolutional Network (SA-STGCN) designed to capture the intricate temporal [...] Read more.
Wind power forecasting remains a major challenge for renewable energy integration, as conventional models often perform poorly when confronted with complex atmospheric dynamics. This study addresses the problem by developing a Spectral-Attentive Spatio-Temporal Graph Convolutional Network (SA-STGCN) designed to capture the intricate temporal and spatial dependencies of wind systems. The approach first applies wavelet transform decomposition to separate volatile wind signals into distinct frequency components, enabling more interpretable representation of rapidly changing conditions. A dynamic temporal attention mechanism is then employed to adaptively identify historical patterns that are most relevant for prediction, moving beyond the fixed temporal windows used in many existing methods. In addition, spectral graph convolution is conducted in the frequency domain to capture farm-wide spatial correlations, thereby modeling long-range atmospheric interactions that conventional localized methods overlook. Although this design increases computational complexity, it proves critical for representing wind variability. Evaluation on real-world datasets demonstrates that SA-STGCN achieves substantial accuracy improvements, with a mean absolute error of 1.52 and a root mean square error of 2.31. These results suggest that embracing more expressive architectures can yield reliable forecasting performance, supporting the stable integration of wind power into modern energy systems. Full article
(This article belongs to the Section A3: Wind, Wave and Tidal Energy)
Show Figures

Figure 1

19 pages, 2549 KB  
Article
STAE-BiSSSM: A Traffic Flow Forecasting Model with High Parameter Effectiveness
by Duoliang Liu, Qiang Qu and Xuebo Chen
ISPRS Int. J. Geo-Inf. 2025, 14(10), 388; https://doi.org/10.3390/ijgi14100388 - 4 Oct 2025
Viewed by 415
Abstract
Traffic flow forecasting plays a significant role in intelligent transportation systems (ITSs) and is instructive for traffic planning, management and control. Increasingly complex traffic conditions pose further challenges to the traffic flow forecasting. While improving the accuracy of model forecasting, the parameter effectiveness [...] Read more.
Traffic flow forecasting plays a significant role in intelligent transportation systems (ITSs) and is instructive for traffic planning, management and control. Increasingly complex traffic conditions pose further challenges to the traffic flow forecasting. While improving the accuracy of model forecasting, the parameter effectiveness of the model is also an issue that cannot be ignored. In addition, existing traffic prediction models have failed to organically integrate data with well-designed model architectures. Therefore, to address the above two issues, we propose the STAE-BiSSSM model as a solution. STAE-BiSSSM consists of Spatio-Temporal Adaptive Embedding (STAE) and Bidirectional Selective State Space Model (BiSSSM), where STAE aims to process features to obtain richer spatio-temporal feature representations. BiSSSM is a novel structural design serving as an alternative to Transformer, capable of extracting patterns of traffic flow changes from both the forward and backward directions of time series with much fewer parameters. Comparative tests between baseline models and STAE-BiSSSM on five real-world datasets illustrates the advance performance of STAE-BiSSSM. This is especially so on METRLA and PeMSBAY datasets, compared with the SOTA model STAEformer. In the short-term forecasting task (horizon: 15 min), MAE, RMSE and MAPE of STAE-BiSSSM decrease by 1.89%/13.74%, 3.72%/16.19% and 1.46%/17.39%, respectively. In the long-term forecasting task (horizon: 60 min), MAE, RMSE and MAPE of STAE-BiSSSM decrease by 3.59%/13.83%, 7.26%/16.36% and 2.16%/15.65%, respectively. Full article
Show Figures

Figure 1

25 pages, 12200 KB  
Article
BIM-Based Integration and Visualization Management of Construction Risks in Water Pumping Station Projects
by Yanyan Xu, Meiru Li, Guiping Huang, Qi Liu, Xueyan Zou, Xin Xu, Zhengyu Guo, Cong Li and Gang Lai
Buildings 2025, 15(19), 3573; https://doi.org/10.3390/buildings15193573 - 3 Oct 2025
Viewed by 448
Abstract
Water pumping stations are essential components of national water infrastructure, yet their construction involves complex, high-risk processes, and traditional risk management approaches often show significant limitations in practice. To address this challenge, this study proposes a Building Information Modeling (BIM)-based approach that integrates [...] Read more.
Water pumping stations are essential components of national water infrastructure, yet their construction involves complex, high-risk processes, and traditional risk management approaches often show significant limitations in practice. To address this challenge, this study proposes a Building Information Modeling (BIM)-based approach that integrates structured risk information into an interactive nD BIM environment. We first developed an extended Risk Breakdown Matrix (eRBM), which systematically organizes risk factors, assessment levels, and causal relationships. This is linked to the BIM model through a customized BIM–risk integration framework. Subsequently, the framework is further implemented and quantitatively validated via a Navisworks plug-in. The system incorporates three core components: (1) a structured risk information model, (2) a visualization mechanism for dynamic, spatiotemporal risk representation and (3) risk influence path analysis using the Decision-Making Trial and Evaluation Laboratory–Interpretive Structural Modeling (DEMATEL–ISM) method. The plug-in allows users to access risk information on demand and monitor its evolution over time and space during the construction process. This study makes contributions by innovatively integrating risk information with BIM and developing a data-driven visualization tool for decision support, thereby enhancing project managers’ ability to anticipate, prioritize, and mitigate risks throughout the construction lifecycle of water pumping station projects. Full article
(This article belongs to the Section Construction Management, and Computers & Digitization)
Show Figures

Figure 1

21 pages, 2248 KB  
Article
TSFNet: Temporal-Spatial Fusion Network for Hybrid Brain-Computer Interface
by Yan Zhang, Bo Yin and Xiaoyang Yuan
Sensors 2025, 25(19), 6111; https://doi.org/10.3390/s25196111 - 3 Oct 2025
Viewed by 418
Abstract
Unimodal brain–computer interfaces (BCIs) often suffer from inherent limitations due to the characteristic of using single modalities. While hybrid BCIs combining electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) offer complementary advantages, effectively integrating their spatiotemporal features remains a challenge due to inherent signal [...] Read more.
Unimodal brain–computer interfaces (BCIs) often suffer from inherent limitations due to the characteristic of using single modalities. While hybrid BCIs combining electroencephalography (EEG) and functional near-infrared spectroscopy (fNIRS) offer complementary advantages, effectively integrating their spatiotemporal features remains a challenge due to inherent signal asynchrony. This study aims to develop a novel deep fusion network to achieve synergistic integration of EEG and fNIRS signals for improved classification performance across different tasks. We propose a novel Temporal-Spatial Fusion Network (TSFNet), which consists of two key sublayers: the EEG-fNIRS-guided Fusion (EFGF) layer and the Cross-Attention-based Feature Enhancement (CAFÉ) layer. The EFGF layer extracts temporal features from EEG and spatial features from fNIRS to generate a hybrid attention map, which is utilized to achieve more effective and complementary integration of spatiotemporal information. The CAFÉ layer enables bidirectional interaction between fNIRS and fusion features via a cross-attention mechanism, which enhances the fusion features and selectively filters informative fNIRS representations. Through the two sublayers, TSFNet achieves deep fusion of multimodal features. Finally, TSFNet is evaluated on motor imagery (MI), mental arithmetic (MA), and word generation (WG) classification tasks. Experimental results demonstrate that TSFNet achieves superior classification performance, with average accuracies of 70.18% for MI, 86.26% for MA, and 81.13% for WG, outperforming existing state-of-the-art multimodal algorithms. These findings suggest that TSFNet provides an effective solution for spatiotemporal feature fusion in hybrid BCIs, with potential applications in real-world BCI systems. Full article
Show Figures

Figure 1

34 pages, 3611 KB  
Review
A Review of Multi-Sensor Fusion in Autonomous Driving
by Hui Qian, Mingchen Wang, Maotao Zhu and Hai Wang
Sensors 2025, 25(19), 6033; https://doi.org/10.3390/s25196033 - 1 Oct 2025
Viewed by 1626
Abstract
Multi-modal sensor fusion has become a cornerstone of robust autonomous driving systems, enabling perception models to integrate complementary cues from cameras, LiDARs, radars, and other modalities. This survey provides a structured overview of recent advances in deep learning-based fusion methods, categorizing them by [...] Read more.
Multi-modal sensor fusion has become a cornerstone of robust autonomous driving systems, enabling perception models to integrate complementary cues from cameras, LiDARs, radars, and other modalities. This survey provides a structured overview of recent advances in deep learning-based fusion methods, categorizing them by architectural paradigms (e.g., BEV-centric fusion and cross-modal attention), learning strategies, and task adaptations. We highlight two dominant architectural trends: unified BEV representation and token-level cross-modal alignment, analyzing their design trade-offs and integration challenges. Furthermore, we review a wide range of applications, from object detection and semantic segmentation to behavior prediction and planning. Despite considerable progress, real-world deployment is hindered by issues such as spatio-temporal misalignment, domain shifts, and limited interpretability. We discuss how recent developments, such as diffusion models for generative fusion, Mamba-style recurrent architectures, and large vision–language models, may unlock future directions for scalable and trustworthy perception systems. Extensive comparisons, benchmark analyses, and design insights are provided to guide future research in this rapidly evolving field. Full article
(This article belongs to the Section Vehicular Sensing)
Show Figures

Figure 1

19 pages, 7875 KB  
Article
SATSN: A Spatial-Adaptive Two-Stream Network for Automatic Detection of Giraffe Daily Behaviors
by Haiming Gan, Xiongwei Wu, Jianlu Chen, Jingling Wang, Yuxin Fang, Yuqing Xue, Tian Jiang, Huanzhen Chen, Peng Zhang, Guixin Dong and Yueju Xue
Animals 2025, 15(19), 2833; https://doi.org/10.3390/ani15192833 - 28 Sep 2025
Viewed by 238
Abstract
The daily behavioral patterns of giraffes reflect their health status and well-being. Behaviors such as licking, walking, standing, and eating are not only essential components of giraffes’ routine activities but also serve as potential indicators of their mental and physiological conditions. This is [...] Read more.
The daily behavioral patterns of giraffes reflect their health status and well-being. Behaviors such as licking, walking, standing, and eating are not only essential components of giraffes’ routine activities but also serve as potential indicators of their mental and physiological conditions. This is particularly relevant in captive environments such as zoos, where certain repetitive behaviors may signal underlying well-being concerns. Therefore, developing an efficient and accurate automated behavior detection system is of great importance for scientific management and welfare improvement. This study proposes a multi-behavior automatic detection method for giraffes based on YOLO11-Pose and the spatial-adaptive two-stream network (SATSN). Firstly, YOLO11-Pose is employed to detect giraffes and estimate the keypoints of their mouths. Observation-Centric SORT (OC-SORT) is then used to track individual giraffes across frames, ensuring temporal identity consistency based on the keypoint positions estimated by YOLO11-Pose. In the SATSN, we propose a region-of-interest extraction strategy for licking behavior to extract local motion features and perform daily behavior classification. In this network, the original 3D ResNet backbone in the slow pathway is replaced with a video transformer encoder to enhance global spatiotemporal modeling, while a Temporal Attention (TA) module is embedded in the fast pathway to improve the representation of fast motion features. To validate the effectiveness of the proposed method, a giraffe behavior dataset consisting of 420 video clips (10 s per clip) was constructed, with 336 clips used for training and 84 for validation. Experimental results show that for the detection tasks of licking, walking, standing, and eating behaviors, the proposed method achieves a mean average precision (mAP) of 93.99%. This demonstrates the strong detection performance and generalization capability of the approach, providing robust support for automated multi-behavior detection and well-being assessment of giraffes. It also lays a technical foundation for building intelligent behavioral monitoring systems in zoos. Full article
Show Figures

Figure 1

23 pages, 6010 KB  
Review
A Review and Design of Semantic-Level Feature Spatial Representation and Resource Spatiotemporal Mapping for Socialized Service Resources in Rural Characteristic Industries
by Yuansheng Wang, Huarui Wu, Cheng Chen and Gongming Wang
Sustainability 2025, 17(19), 8534; https://doi.org/10.3390/su17198534 - 23 Sep 2025
Viewed by 455
Abstract
Socialized services for rural characteristic industries are becoming a key support for promoting rural industries’ transformation and upgrading. They are permeating the development process of modern agricultural service technologies, achieving significant progress in specialized fields such as mechanized operations and plant-protection services. However, [...] Read more.
Socialized services for rural characteristic industries are becoming a key support for promoting rural industries’ transformation and upgrading. They are permeating the development process of modern agricultural service technologies, achieving significant progress in specialized fields such as mechanized operations and plant-protection services. However, challenges remain, including low efficiency in matching service resources and limited spatiotemporal coordination capabilities. With the deep integration of spatiotemporal information technology and knowledge graph technology, the enormous potential of semantic-level feature spatial representation in intelligent scheduling of service resources has been fully demonstrated, providing a new technical pathway to solve the above problem. This paper systematically analyzes the technological evolution trends of socialized services for rural characteristic industries and proposes a collaborative scheduling framework based on semantic feature space and spatiotemporal maps for characteristic industry service resources. At the technical architecture level, the paper aims to construct a spatiotemporal graph model integrating geographic knowledge graphs and temporal tree technology to achieve semantic-level feature matching between service demand and supply. Regarding implementation pathways, the model significantly improves the spatiotemporal allocation efficiency of service resources through cloud service platforms that integrate spatial semantic matching algorithms and dynamic optimization technologies. This paper conducts in-depth discussions and analyses on technical details such as agricultural semantic feature extraction, dynamic updates of rural service resources, and the collaboration of semantic matching and spatio-temporal matching of supply and demand relationships. It also presents relevant implementation methods to enhance technical integrity and logic, which is conducive to the engineering implementation of the proposed methods. The effectiveness of the proposed collaborative scheduling framework for service resources is proved by the synthesis of principal analysis, logical deduction and case comparison. We have proposed a practical “three-step” implementation path conducive to realizing the proposed method. Regarding application paradigms, this technical system will promote the transformation of rural industry services from traditional mechanical operations to an intelligent service model of “demand perception–intelligent matching–precise scheduling”. In the field of socialized services for rural characteristic industries, it is suggested that relevant institutions promote this technical framework and pay attention to the development trends of new technologies such as knowledge services, spatio-temporal services, the Internet of Things, and unmanned farms so as to promote the sustainable development of rural characteristic industries. Full article
Show Figures

Figure 1

24 pages, 9770 KB  
Article
TransMambaCNN: A Spatiotemporal Transformer Network Fusing State-Space Models and CNNs for Short-Term Precipitation Forecasting
by Kai Zhang, Guojing Zhang and Xiaoying Wang
Remote Sens. 2025, 17(18), 3200; https://doi.org/10.3390/rs17183200 - 16 Sep 2025
Viewed by 424
Abstract
Deep learning for precipitation forecasting remains constrained by complex meteorological factors affecting accuracy. To address this issue, this paper proposes TransMambaCNN, which is a spatiotemporal transformer network fusing state-space models and CNNs for short-term precipitation forecasting. The core of the model employs a [...] Read more.
Deep learning for precipitation forecasting remains constrained by complex meteorological factors affecting accuracy. To address this issue, this paper proposes TransMambaCNN, which is a spatiotemporal transformer network fusing state-space models and CNNs for short-term precipitation forecasting. The core of the model employs a Convolutional State-Space Module (C-SSM), which efficiently extracts spatiotemporal features from multi-source meteorological variables by replacing the self-attention mechanism in the Vision Transformer (ViT) with an Attentive State-Space Module (ASSM) and augmenting its feature extraction capacity with integrated depthwise convolution. Its dual-branch architecture consists of a global branch, where C-SSM captures long-range dependencies and global spatiotemporal patterns, and a local branch, which leverages multi-scale convolutions based on SimVP’s Inception structure to extract fine-grained local features. The deep fusion of these dual branches significantly enhances spatiotemporal feature representation.Experiments demonstrate that in southeastern China and adjacent marine areas (period of high precipitation: April–September), TransMambaCNN achieves a 13.38% and 47.67% improvement in Threat Score (TS) over PredRNN at thresholds of ≥25 mm and ≥50 mm, respectively. In the Qinghai Sanjiangyuan region of western China (a precipitation-scarce area), TransMambaCNN’s TS score surpasses SimVP by 11.86 times at the ≥25 mm threshold. Full article
Show Figures

Figure 1

Back to TopTop