Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (3,512)

Search Parameters:
Keywords = fast extraction

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
16 pages, 2859 KiB  
Article
Maritime Spectrum Sensing Based on Cyclostationary Features and Convolutional Neural Networks
by Xuan Geng and Boyu Hu
Entropy 2025, 27(8), 809; https://doi.org/10.3390/e27080809 - 28 Jul 2025
Abstract
For maritime cognitive radio networks (MCRN), spectrum sensing (SS) is challenging due to the movement of the sea, channel interference, and unstable link quality. Besides the basic sensing capabilities that are required, SS in MCRN also requires the ability to adapt to complex [...] Read more.
For maritime cognitive radio networks (MCRN), spectrum sensing (SS) is challenging due to the movement of the sea, channel interference, and unstable link quality. Besides the basic sensing capabilities that are required, SS in MCRN also requires the ability to adapt to complex and dynamic environments. By transforming spectrum sensing into a classification problem and leveraging cyclostationary features and Convolutional Neural Networks (CNN), This paper proposes a classification-guided TC2NND (Transfer Cyclostationary- feature and Convolutional Neural Networks Detection) spectrum sensing algorithm, which regards the maritime spectrum sensing as a classification problem. The TC2NND algorithm first classifies the received signal features by extracting cycle power spectrum (CPS) features using the FFT (Fast Fourier Transform) Accumulation Method (FAM), and then makes a judgment using a variety of C2NND decision models. The experimental results demonstrate that the proposed TC2NND algorithm could achieve a detection probability of 91.5% with a false-alarm probability of 5% at SNR = −10 dB, which significantly outperforms the conventional methods. Full article
(This article belongs to the Special Issue Space-Air-Ground-Sea Integrated Communication Networks)
21 pages, 15647 KiB  
Article
Research on Oriented Object Detection in Aerial Images Based on Architecture Search with Decoupled Detection Heads
by Yuzhe Kang, Bohao Zheng and Wei Shen
Appl. Sci. 2025, 15(15), 8370; https://doi.org/10.3390/app15158370 - 28 Jul 2025
Abstract
Object detection in aerial images can provide great support in traffic planning, national defense reconnaissance, hydrographic surveys, infrastructure construction, and other fields. Objects in aerial images are characterized by small pixel–area ratios, dense arrangements between objects, and arbitrary inclination angles. In response to [...] Read more.
Object detection in aerial images can provide great support in traffic planning, national defense reconnaissance, hydrographic surveys, infrastructure construction, and other fields. Objects in aerial images are characterized by small pixel–area ratios, dense arrangements between objects, and arbitrary inclination angles. In response to these characteristics and problems, we improved the feature extraction network Inception-ResNet using the Fast Architecture Search (FAS) module and proposed a one-stage anchor-free rotation object detector. The structure of the object detector is simple and only consists of convolution layers, which reduces the number of model parameters. At the same time, the label sampling strategy in the training process is optimized to resolve the problem of insufficient sampling. Finally, a decoupled object detection head is used to separate the bounding box regression task from the object classification task. The experimental results show that the proposed method achieves mean average precision (mAP) of 82.6%, 79.5%, and 89.1% on the DOTA1.0, DOTA1.5, and HRSC2016 datasets, respectively, and the detection speed reaches 24.4 FPS, which can meet the needs of real-time detection. Full article
(This article belongs to the Special Issue Innovative Applications of Artificial Intelligence in Engineering)
Show Figures

Figure 1

23 pages, 3847 KiB  
Article
Optimizing Sentiment Analysis in Multilingual Balanced Datasets: A New Comparative Approach to Enhancing Feature Extraction Performance with ML and DL Classifiers
by Hamza Jakha, Souad El Houssaini, Mohammed-Alamine El Houssaini, Souad Ajjaj and Abdelali Hadir
Appl. Syst. Innov. 2025, 8(4), 104; https://doi.org/10.3390/asi8040104 - 28 Jul 2025
Abstract
Social network platforms have a big impact on the development of companies by influencing clients’ behaviors and sentiments, which directly affect corporate reputations. Analyzing this feedback has become an essential component of business intelligence, supporting the improvement of long-term marketing strategies on a [...] Read more.
Social network platforms have a big impact on the development of companies by influencing clients’ behaviors and sentiments, which directly affect corporate reputations. Analyzing this feedback has become an essential component of business intelligence, supporting the improvement of long-term marketing strategies on a larger scale. The implementation of powerful sentiment analysis models requires a comprehensive and in-depth examination of each stage of the process. In this study, we present a new comparative approach for several feature extraction techniques, including TF-IDF, Word2Vec, FastText, and BERT embeddings. These methods are applied to three multilingual datasets collected from hotel review platforms in the tourism sector in English, French, and Arabic languages. Those datasets were preprocessed through cleaning, normalization, labeling, and balancing before being trained on various machine learning and deep learning algorithms. The effectiveness of each feature extraction method was evaluated using metrics such as accuracy, F1-score, precision, recall, ROC AUC curve, and a new metric that measures the execution time for generating word representations. Our extensive experiments demonstrate significant and excellent results, achieving accuracy rates of approximately 99% for the English dataset, 94% for the Arabic dataset, and 89% for the French dataset. These findings confirm the important impact of vectorization techniques on the performance of sentiment analysis models. They also highlight the important relationship between balanced datasets, effective feature extraction methods, and the choice of classification algorithms. So, this study aims to simplify the selection of feature extraction methods and appropriate classifiers for each language, thereby contributing to advancements in sentiment analysis. Full article
(This article belongs to the Topic Social Sciences and Intelligence Management, 2nd Volume)
Show Figures

Figure 1

18 pages, 4836 KiB  
Article
Deep Learning to Analyze Spatter and Melt Pool Behavior During Additive Manufacturing
by Deepak Gadde, Alaa Elwany and Yang Du
Metals 2025, 15(8), 840; https://doi.org/10.3390/met15080840 - 28 Jul 2025
Abstract
To capture the complex metallic spatter and melt pool behavior during the rapid interaction between the laser and metal material, high-speed cameras are applied to record the laser powder bed fusion process and generate a large volume of image data. In this study, [...] Read more.
To capture the complex metallic spatter and melt pool behavior during the rapid interaction between the laser and metal material, high-speed cameras are applied to record the laser powder bed fusion process and generate a large volume of image data. In this study, four deep learning algorithms are applied: YOLOv5, Fast R-CNN, RetinaNet, and EfficientDet. They are trained by the recorded videos to learn and extract information on spatter and melt pool behavior during the laser powder bed fusion process. The well-trained models achieved high accuracy and low loss, demonstrating strong capability in accurately detecting and tracking spatter and melt pool dynamics. A stability index is proposed and calculated based on the melt pool length change rate. Greater index value reflects a more stable melt pool. We found that more spatters were detected for the unstable melt pool, while fewer spatters were found for the stable melt pool. The spatter’s size can affect its initial ejection speed, and large spatters are ejected slowly while small spatters are ejected rapidly. In addition, more than 58% of detected spatters have their initial ejection angle in the range of 60–120°. These findings provide a better understanding of spatter and melt pool dynamics and behavior, uncover the influence of melt pool stability on spatter formation, and demonstrate the correlation between the spatter size and its initial ejection speed. This work will contribute to the extraction of important information from high-speed recorded videos for additive manufacturing to reduce waste, lower cost, enhance part quality, and increase process reliability. Full article
(This article belongs to the Special Issue Machine Learning in Metal Additive Manufacturing)
Show Figures

Figure 1

26 pages, 6806 KiB  
Article
Fine Recognition of MEO SAR Ship Targets Based on a Multi-Level Focusing-Classification Strategy
by Zhaohong Li, Wei Yang, Can Su, Hongcheng Zeng, Yamin Wang, Jiayi Guo and Huaping Xu
Remote Sens. 2025, 17(15), 2599; https://doi.org/10.3390/rs17152599 - 26 Jul 2025
Viewed by 162
Abstract
The Medium Earth Orbit (MEO) spaceborne Synthetic Aperture Radar (SAR) has great coverage ability, which can improve maritime ship target surveillance performance significantly. However, due to the huge computational load required for imaging processing and the severe defocusing caused by ship motions, traditional [...] Read more.
The Medium Earth Orbit (MEO) spaceborne Synthetic Aperture Radar (SAR) has great coverage ability, which can improve maritime ship target surveillance performance significantly. However, due to the huge computational load required for imaging processing and the severe defocusing caused by ship motions, traditional ship recognition conducted in focused image domains cannot process MEO SAR data efficiently. To address this issue, a multi-level focusing-classification strategy for MEO SAR ship recognition is proposed, which is applied to the range-compressed ship data domain. Firstly, global fast coarse-focusing is conducted to compensate for sailing motion errors. Then, a coarse-classification network is designed to realize major target category classification, based on which local region image slices are extracted. Next, fine-focusing is performed to correct high-order motion errors, followed by applying fine-classification applied to the image slices to realize final ship classification. Equivalent MEO SAR ship images generated by real LEO SAR data are utilized to construct training and testing datasets. Simulated MEO SAR ship data are also used to evaluate the generalization of the whole method. The experimental results demonstrate that the proposed method can achieve high classification precision. Since only local region slices are used during the second-level processing step, the complex computations induced by fine-focusing for the full image can be avoided, thereby significantly improving overall efficiency. Full article
(This article belongs to the Special Issue Advances in Remote Sensing Image Target Detection and Recognition)
Show Figures

Figure 1

21 pages, 4863 KiB  
Article
Detection Model for Cotton Picker Fire Recognition Based on Lightweight Improved YOLOv11
by Zhai Shi, Fangwei Wu, Changjie Han, Dongdong Song and Yi Wu
Agriculture 2025, 15(15), 1608; https://doi.org/10.3390/agriculture15151608 - 25 Jul 2025
Viewed by 172
Abstract
In response to the limited research on fire detection in cotton pickers and the issue of low detection accuracy in visual inspection, this paper proposes a computer vision-based detection method. The method is optimized according to the structural characteristics of cotton pickers, and [...] Read more.
In response to the limited research on fire detection in cotton pickers and the issue of low detection accuracy in visual inspection, this paper proposes a computer vision-based detection method. The method is optimized according to the structural characteristics of cotton pickers, and a lightweight improved YOLOv11 algorithm is designed for cotton fire detection in cotton pickers. The backbone of the model is replaced with the MobileNetV2 network to achieve effective model lightweighting. In addition, the convolutional layers in the original C3k2 block are optimized using partial convolutions to reduce computational redundancy and improve inference efficiency. Furthermore, a visual attention mechanism named CBAM-ECA (Convolutional Block Attention Module-Efficient Channel Attention) is designed to suit the complex working conditions of cotton pickers. This mechanism aims to enhance the model’s feature extraction capability under challenging environmental conditions, thereby improving overall detection accuracy. To further improve localization performance and accelerate convergence, the loss function is also modified. These improvements enable the model to achieve higher precision in fire detection while ensuring fast and accurate localization. Experimental results demonstrate that the improved model reduces the number of parameters by 38%, increases the frame processing speed (FPS) by 13.2%, and decreases the computational complexity (GFLOPs) by 42.8%, compared to the original model. The detection accuracy for flaming combustion, smoldering combustion, and overall detection is improved by 1.4%, 3%, and 1.9%, respectively, with an increase of 2.4% in mAP (mean average precision). Compared to other models—YOLOv3-tiny, YOLOv5, YOLOv8, and YOLOv10—the proposed method achieves higher detection accuracy by 5.9%, 7%, 5.9%, and 5.3%, respectively, and shows improvements in mAP by 5.4%, 5%, 4.8%, and 6.3%. The improved detection algorithm maintains high accuracy while achieving faster inference speed and fewer model parameters. These improvements lay a solid foundation for fire prevention and suppression in cotton collection boxes on cotton pickers. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

26 pages, 6528 KiB  
Article
Lightweight Sheep Face Recognition Model Combining Grouped Convolution and Parameter Fusion
by Gaochao Liu, Lijun Kang and Yongqiang Dai
Sensors 2025, 25(15), 4610; https://doi.org/10.3390/s25154610 - 25 Jul 2025
Viewed by 80
Abstract
Sheep face recognition technology is critical in key areas such as individual sheep identification and behavior monitoring. Existing sheep face recognition models typically require high computational resources. When these models are deployed on mobile or embedded devices, problems such as reduced model recognition [...] Read more.
Sheep face recognition technology is critical in key areas such as individual sheep identification and behavior monitoring. Existing sheep face recognition models typically require high computational resources. When these models are deployed on mobile or embedded devices, problems such as reduced model recognition accuracy and increased recognition time arise. To address these problems, an improved Parameter Fusion Lightweight You Only Look Once (PFL-YOLO) sheep face recognition model based on YOLOv8n is proposed. In this study, the Efficient Hybrid Conv (EHConv) module is first integrated to enhance the extraction capability of the model for sheep face features. At the same time, the Residual C2f (RC2f) module is introduced to facilitate the effective fusion of multi-scale feature information and improve the information processing capability of the model; furthermore, the Efficient Spatial Pyramid Pooling Fast (ESPPF) module was used to fuse features of different scales. Finally, parameter fusion optimization work was carried out for the detection head, and the construction of the Parameter Fusion Detection (PFDetect) module was achieved, which significantly reduced the number of model parameters and computational complexity. The experimental results show that the PFL-YOLO model exhibits an excellent performance–efficiency balance in sheep face recognition tasks: mAP@50 and mAP@50:95 reach 99.5% and 87.4%, respectively, and the accuracy is close to or equal to the mainstream benchmark model. At the same time, the number of parameters is only 1.01 M, which is reduced by 45.1%, 83.7%, 66.6%, 71.4%, and 61.2% compared to YOLOv5n, YOLOv7-tiny, YOLOv8n, YOLOv9-t, and YOLO11n, respectively. The size of the model was compressed to 2.1 MB, which was reduced by 44.7%, 82.5%, 65%, 72%, and 59.6%, respectively, compared to similar lightweight models. The experimental results confirm that the PFL-YOLO model maintains high accuracy recognition performance while being lightweight and can provide a new solution for sheep face recognition models on resource-constrained devices. Full article
(This article belongs to the Section Smart Agriculture)
Show Figures

Figure 1

22 pages, 16984 KiB  
Article
Small Ship Detection Based on Improved Neural Network Algorithm and SAR Images
by Jiaqi Li, Hongyuan Huo, Li Guo, De Zhang, Wei Feng, Yi Lian and Long He
Remote Sens. 2025, 17(15), 2586; https://doi.org/10.3390/rs17152586 - 24 Jul 2025
Viewed by 182
Abstract
Synthetic aperture radar images can be used for ship target detection. However, due to the unclear ship outline in SAR images, noise and land background factors affect the difficulty and accuracy of ship (especially small target ship) detection. Therefore, based on the YOLOv5s [...] Read more.
Synthetic aperture radar images can be used for ship target detection. However, due to the unclear ship outline in SAR images, noise and land background factors affect the difficulty and accuracy of ship (especially small target ship) detection. Therefore, based on the YOLOv5s model, this paper improves its backbone network and feature fusion network algorithm to improve the accuracy of ship detection target recognition. First, the LSKModule is used to improve the backbone network of YOLOv5s. By adaptively aggregating the features extracted by large-size convolution kernels to fully obtain context information, at the same time, key features are enhanced and noise interference is suppressed. Secondly, multiple Depthwise Separable Convolution layers are added to the SPPF (Spatial Pyramid Pooling-Fast) structure. Although a small number of parameters and calculations are introduced, features of different receptive fields can be extracted. Third, the feature fusion network of YOLOv5s is improved based on BIFPN, and the shallow feature map is used to optimize the small target detection performance. Finally, the CoordConv module is added before the detect head of YOLOv5, and two coordinate channels are added during the convolution operation to further improve the accuracy of target detection. The map50 of this method for the SSDD dataset and HRSID dataset reached 97.6% and 91.7%, respectively, and was compared with a variety of advanced target detection models. The results show that the detection accuracy of this method is higher than other similar target detection algorithms. Full article
Show Figures

Figure 1

23 pages, 9603 KiB  
Article
Label-Efficient Fine-Tuning for Remote Sensing Imagery Segmentation with Diffusion Models
by Yiyun Luo, Jinnian Wang, Jean Sequeira, Xiankun Yang, Dakang Wang, Jiabin Liu, Grekou Yao and Sébastien Mavromatis
Remote Sens. 2025, 17(15), 2579; https://doi.org/10.3390/rs17152579 - 24 Jul 2025
Viewed by 113
Abstract
High-resolution remote sensing imagery plays an essential role in urban management and environmental monitoring, providing detailed insights for applications ranging from land cover mapping to disaster response. Semantic segmentation methods are among the most effective techniques for comprehensive land cover mapping, and they [...] Read more.
High-resolution remote sensing imagery plays an essential role in urban management and environmental monitoring, providing detailed insights for applications ranging from land cover mapping to disaster response. Semantic segmentation methods are among the most effective techniques for comprehensive land cover mapping, and they commonly employ ImageNet-based pre-training semantics. However, traditional fine-tuning processes exhibit poor transferability across different downstream tasks and require large amounts of labeled data. To address these challenges, we introduce Denoising Diffusion Probabilistic Models (DDPMs) as a generative pre-training approach for semantic features extraction in remote sensing imagery. We pre-trained a DDPM on extensive unlabeled imagery, obtaining features at multiple noise levels and resolutions. In order to integrate and optimize these features efficiently, we designed a multi-layer perceptron module with residual connections. It performs channel-wise optimization to suppress feature redundancy and refine representations. Additionally, we froze the feature extractor during fine-tuning. This strategy significantly reduces computational consumption and facilitates fast transfer and deployment across various interpretation tasks on homogeneous imagery. Our comprehensive evaluation on the sparsely labeled dataset MiniFrance-S and the fully labeled Gaofen Image Dataset achieved mean intersection over union scores of 42.7% and 66.5%, respectively, outperforming previous works. This demonstrates that our approach effectively reduces reliance on labeled imagery and increases transferability to downstream remote sensing tasks. Full article
(This article belongs to the Special Issue AI-Driven Mapping Using Remote Sensing Data)
Show Figures

Figure 1

21 pages, 1936 KiB  
Article
FFT-RDNet: A Time–Frequency-Domain-Based Intrusion Detection Model for IoT Security
by Bingjie Xiang, Renguang Zheng, Kunsan Zhang, Chaopeng Li and Jiachun Zheng
Sensors 2025, 25(15), 4584; https://doi.org/10.3390/s25154584 - 24 Jul 2025
Viewed by 157
Abstract
Resource-constrained Internet of Things (IoT) devices demand efficient and robust intrusion detection systems (IDSs) to counter evolving cyber threats. The traditional IDS models, however, struggle with high computational complexity and inadequate feature extraction, limiting their accuracy and generalizability in IoT environments. To address [...] Read more.
Resource-constrained Internet of Things (IoT) devices demand efficient and robust intrusion detection systems (IDSs) to counter evolving cyber threats. The traditional IDS models, however, struggle with high computational complexity and inadequate feature extraction, limiting their accuracy and generalizability in IoT environments. To address this, we propose FFT-RDNet, a lightweight IDS framework leveraging depthwise separable convolution and frequency-domain feature fusion. An ADASYN-Tomek Links hybrid strategy first addresses class imbalances. The core innovation of FFT-RDNet lies in its novel two-dimensional spatial feature modeling approach, realized through a dedicated dual-path feature embedding module. One branch extracts discriminative statistical features in the time domain, while the other branch transforms the data into the frequency domain via Fast Fourier Transform (FFT) to capture the essential energy distribution characteristics. These time–frequency domain features are fused to construct a two-dimensional feature space, which is then processed by a streamlined residual network using depthwise separable convolution. This network effectively captures complex periodic attack patterns with minimal computational overhead. Comprehensive evaluation on the NSL-KDD and CIC-IDS2018 datasets shows that FFT-RDNet outperforms state-of-the-art neural network IDSs across accuracy, precision, recall, and F1 score (improvements: 0.22–1%). Crucially, it achieves superior accuracy with a significantly reduced computational complexity, demonstrating high efficiency for resource-constrained IoT security deployments. Full article
(This article belongs to the Section Internet of Things)
Show Figures

Figure 1

20 pages, 28281 KiB  
Article
Infrared-Guided Thermal Cycles in FEM Simulation of Laser Welding of Thin Aluminium Alloy Sheets
by Pasquale Russo Spena, Manuela De Maddis, Valentino Razza, Luca Santoro, Husniddin Mamarayimov and Dario Basile
Metals 2025, 15(8), 830; https://doi.org/10.3390/met15080830 - 24 Jul 2025
Viewed by 186
Abstract
Climate concerns are driving the automotive industry to adopt advanced manufacturing technologies that aim to improve energy efficiency and reduce vehicle weight. In this context, lightweight structural materials such as aluminium alloys have gained significant attention due to their favorable strength-to-weight ratio. Laser [...] Read more.
Climate concerns are driving the automotive industry to adopt advanced manufacturing technologies that aim to improve energy efficiency and reduce vehicle weight. In this context, lightweight structural materials such as aluminium alloys have gained significant attention due to their favorable strength-to-weight ratio. Laser welding plays a crucial role in assembling such materials, offering high flexibility and fast joining capabilities for thin aluminium sheets. However, welding these materials presents specific challenges, particularly in controlling heat input to minimize distortions and ensure consistent weld quality. As a result, numerical simulations based on the Finite Element Method (FEM) are essential for predicting weld-induced phenomena and optimizing process performance. This study investigates welding-induced distortions in laser butt welding of 1.5 mm-thick Al 6061 samples through FEM simulations performed in the SYSWELD 2024.0 environment. The methodology provided by the software is based on the Moving Heat Source (MHS) model, which simulates the physical movement of the heat source and typically requires extensive calibration through destructive metallographic testing. This transient approach enables the detailed prediction of thermal, metallurgical, and mechanical behavior, but it is computationally demanding. To improve efficiency, the Imposed Thermal Cycle (ITC) model is often used. In this technique, a thermal cycle, extracted from an MHS simulation or experimental data, is imposed on predefined subregions of the model, allowing only mechanical behavior to be simulated while reducing computation time. To avoid MHS-based calibration, this work proposes using thermal cycles acquired in-line during welding via infrared thermography as direct input for the ITC model. The method was validated experimentally and numerically, showing good agreement in the prediction of distortions and a significant reduction in workflow time. The distortion values from simulations differ from the real experiment by less than 0.3%. Our method exhibits a slight decrease in performance, resulting in an increase in estimation error of 0.03% compared to classic approaches, but more than 85% saving in computation time. The integration of real process data into the simulation enables a virtual representation of the process, supporting future developments toward Digital Twin applications. Full article
(This article belongs to the Special Issue Manufacturing Processes of Metallic Materials)
Show Figures

Figure 1

23 pages, 13739 KiB  
Article
Traffic Accident Rescue Action Recognition Method Based on Real-Time UAV Video
by Bo Yang, Jianan Lu, Tao Liu, Bixing Zhang, Chen Geng, Yan Tian and Siyu Zhang
Drones 2025, 9(8), 519; https://doi.org/10.3390/drones9080519 - 24 Jul 2025
Viewed by 227
Abstract
Low-altitude drones, which are unimpeded by traffic congestion or urban terrain, have become a critical asset in emergency rescue missions. To address the current lack of emergency rescue data, UAV aerial videos were collected to create an experimental dataset for action classification and [...] Read more.
Low-altitude drones, which are unimpeded by traffic congestion or urban terrain, have become a critical asset in emergency rescue missions. To address the current lack of emergency rescue data, UAV aerial videos were collected to create an experimental dataset for action classification and localization annotation. A total of 5082 keyframes were labeled with 1–5 targets each, and 14,412 instances of data were prepared (including flight altitude and camera angles) for action classification and position annotation. To mitigate the challenges posed by high-resolution drone footage with excessive redundant information, we propose the SlowFast-Traffic (SF-T) framework, a spatio-temporal sequence-based algorithm for recognizing traffic accident rescue actions. For more efficient extraction of target–background correlation features, we introduce the Actor-Centric Relation Network (ACRN) module, which employs temporal max pooling to enhance the time-dimensional features of static backgrounds, significantly reducing redundancy-induced interference. Additionally, smaller ROI feature map outputs are adopted to boost computational speed. To tackle class imbalance in incident samples, we integrate a Class-Balanced Focal Loss (CB-Focal Loss) function, effectively resolving rare-action recognition in specific rescue scenarios. We replace the original Faster R-CNN with YOLOX-s to improve the target detection rate. On our proposed dataset, the SF-T model achieves a mean average precision (mAP) of 83.9%, which is 8.5% higher than that of the standard SlowFast architecture while maintaining a processing speed of 34.9 tasks/s. Both accuracy-related metrics and computational efficiency are substantially improved. The proposed method demonstrates strong robustness and real-time analysis capabilities for modern traffic rescue action recognition. Full article
(This article belongs to the Special Issue Cooperative Perception for Modern Transportation)
Show Figures

Figure 1

26 pages, 6051 KiB  
Article
A Novel Sound Coding Strategy for Cochlear Implants Based on Spectral Feature and Temporal Event Extraction
by Behnam Molaee-Ardekani, Rafael Attili Chiea, Yue Zhang, Julian Felding, Aswin Adris Wijetillake, Peter T. Johannesen, Enrique A. Lopez-Poveda and Manuel Segovia-Martínez
Technologies 2025, 13(8), 318; https://doi.org/10.3390/technologies13080318 - 23 Jul 2025
Viewed by 164
Abstract
This paper presents a novel cochlear implant (CI) sound coding strategy called Spectral Feature Extraction (SFE). The SFE is a novel Fast Fourier Transform (FFT)-based Continuous Interleaved Sampling (CIS) strategy that provides less-smeared spectral cues to CI patients compared to Crystalis, a predecessor [...] Read more.
This paper presents a novel cochlear implant (CI) sound coding strategy called Spectral Feature Extraction (SFE). The SFE is a novel Fast Fourier Transform (FFT)-based Continuous Interleaved Sampling (CIS) strategy that provides less-smeared spectral cues to CI patients compared to Crystalis, a predecessor strategy used in Oticon Medical devices. The study also explores how the SFE can be enhanced into a Temporal Fine Structure (TFS)-based strategy named Spectral Event Extraction (SEE), combining spectral sharpness with temporal cues. Background/Objectives: Many CI recipients understand speech in quiet settings but struggle with music and complex environments, increasing cognitive effort. De-smearing the power spectrum and extracting spectral peak features can reduce this load. The SFE targets feature extraction from spectral peaks, while the SEE enhances TFS-based coding by tracking these features across frames. Methods: The SFE strategy extracts spectral peaks and models them with synthetic pure tone spectra characterized by instantaneous frequency, phase, energy, and peak resemblance. This deblurs input peaks by estimating their center frequency. In SEE, synthetic peaks are tracked across frames to yield reliable temporal cues (e.g., zero-crossings) aligned with stimulation pulses. Strategy characteristics are analyzed using electrodograms. Results: A flexible Frequency Allocation Map (FAM) can be applied to both SFE and SEE strategies without being limited by FFT bandwidth constraints. Electrodograms of Crystalis and SFE strategies showed that SFE reduces spectral blurring and provides detailed temporal information of harmonics in speech and music. Conclusions: SFE and SEE are expected to enhance speech understanding, lower listening effort, and improve temporal feature coding. These strategies could benefit CI users, especially in challenging acoustic environments. Full article
(This article belongs to the Special Issue The Challenges and Prospects in Cochlear Implantation)
Show Figures

Figure 1

38 pages, 6851 KiB  
Article
FGFNet: Fourier Gated Feature-Fusion Network with Fractal Dimension Estimation for Robust Palm-Vein Spoof Detection
by Seung Gu Kim, Jung Soo Kim and Kang Ryoung Park
Fractal Fract. 2025, 9(8), 478; https://doi.org/10.3390/fractalfract9080478 - 22 Jul 2025
Viewed by 176
Abstract
The palm-vein recognition system has garnered attention as a biometric technology due to its resilience to external environmental factors, protection of personal privacy, and low risk of external exposure. However, with recent advancements in deep learning-based generative models for image synthesis, the quality [...] Read more.
The palm-vein recognition system has garnered attention as a biometric technology due to its resilience to external environmental factors, protection of personal privacy, and low risk of external exposure. However, with recent advancements in deep learning-based generative models for image synthesis, the quality and sophistication of fake images have improved, leading to an increased security threat from counterfeit images. In particular, palm-vein images acquired through near-infrared illumination exhibit low resolution and blurred characteristics, making it even more challenging to detect fake images. Furthermore, spoof detection specifically targeting palm-vein images has not been studied in detail. To address these challenges, this study proposes the Fourier-gated feature-fusion network (FGFNet) as a novel spoof detector for palm-vein recognition systems. The proposed network integrates masked fast Fourier transform, a map-based gated feature fusion block, and a fast Fourier convolution (FFC) attention block with global contrastive loss to effectively detect distortion patterns caused by generative models. These components enable the efficient extraction of critical information required to determine the authenticity of palm-vein images. In addition, fractal dimension estimation (FDE) was employed for two purposes in this study. In the spoof attack procedure, FDE was used to evaluate how closely the generated fake images approximate the structural complexity of real palm-vein images, confirming that the generative model produced highly realistic spoof samples. In the spoof detection procedure, the FDE results further demonstrated that the proposed FGFNet effectively distinguishes between real and fake images, validating its capability to capture subtle structural differences induced by generative manipulation. To evaluate the spoof detection performance of FGFNet, experiments were conducted using real palm-vein images from two publicly available palm-vein datasets—VERA Spoofing PalmVein (VERA dataset) and PLUSVein-contactless (PLUS dataset)—as well as fake palm-vein images generated based on these datasets using a cycle-consistent generative adversarial network. The results showed that, based on the average classification error rate, FGFNet achieved 0.3% and 0.3% on the VERA and PLUS datasets, respectively, demonstrating superior performance compared to existing state-of-the-art spoof detection methods. Full article
Show Figures

Figure 1

24 pages, 3714 KiB  
Article
DTCMMA: Efficient Wind-Power Forecasting Based on Dimensional Transformation Combined with Multidimensional and Multiscale Convolutional Attention Mechanism
by Wenhan Song, Enguang Zuo, Junyu Zhu, Chen Chen, Cheng Chen, Ziwei Yan and Xiaoyi Lv
Sensors 2025, 25(15), 4530; https://doi.org/10.3390/s25154530 - 22 Jul 2025
Viewed by 214
Abstract
With the growing global demand for clean energy, the accuracy of wind-power forecasting plays a vital role in ensuring the stable operation of power systems. However, wind-power generation is significantly influenced by meteorological conditions and is characterized by high uncertainty and multiscale fluctuations. [...] Read more.
With the growing global demand for clean energy, the accuracy of wind-power forecasting plays a vital role in ensuring the stable operation of power systems. However, wind-power generation is significantly influenced by meteorological conditions and is characterized by high uncertainty and multiscale fluctuations. Traditional recurrent neural network (RNN) and long short-term memory (LSTM) models, although capable of handling sequential data, struggle with modeling long-term temporal dependencies due to the vanishing gradient problem; thus, they are now rarely used. Recently, Transformer models have made notable progress in sequence modeling compared to RNNs and LSTM models. Nevertheless, when dealing with long wind-power sequences, their quadratic computational complexity (O(L2)) leads to low efficiency, and their global attention mechanism often fails to capture local periodic features accurately, tending to overemphasize redundant information while overlooking key temporal patterns. To address these challenges, this paper proposes a wind-power forecasting method based on dimension-transformed collaborative multidimensional multiscale attention (DTCMMA). This method first employs fast Fourier transform (FFT) to automatically identify the main periodic components in wind-power data, reconstructing the one-dimensional time series as a two-dimensional spatiotemporal representation, thereby explicitly encoding periodic features. Based on this, a collaborative multidimensional multiscale attention (CMMA) mechanism is designed, which hierarchically integrates channel, spatial, and pixel attention to adaptively capture complex spatiotemporal dependencies. Considering the geometric characteristics of the reconstructed data, asymmetric convolution kernels are adopted to enhance feature extraction efficiency. Experiments on multiple wind-farm datasets and energy-related datasets demonstrate that DTCMMA outperforms mainstream methods such as Transformer, iTransformer, and TimeMixer in long-sequence forecasting tasks, achieving improvements in MSE performance by 34.22%, 2.57%, and 0.51%, respectively. The model’s training speed also surpasses that of the fastest baseline by 300%, significantly improving both prediction accuracy and computational efficiency. This provides an efficient and accurate solution for wind-power forecasting and contributes to the further development and application of wind energy in the global energy mix. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

Back to TopTop