Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (46)

Search Parameters:
Keywords = multi abnormal scenarios

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 9038 KiB  
Article
Deep Learning-Based Detection and Digital Twin Implementation of Beak Deformities in Caged Layer Chickens
by Hengtai Li, Hongfei Chen, Jinlin Liu, Qiuhong Zhang, Tao Liu, Xinyu Zhang, Yuhua Li, Yan Qian and Xiuguo Zou
Agriculture 2025, 15(11), 1170; https://doi.org/10.3390/agriculture15111170 - 29 May 2025
Viewed by 778
Abstract
With the increasing urgency for digital transformation in large-scale caged layer farms, traditional methods for monitoring the environment and chicken health, which often rely on human experience, face challenges related to low efficiency and poor real-time performance. In this study, we focused on [...] Read more.
With the increasing urgency for digital transformation in large-scale caged layer farms, traditional methods for monitoring the environment and chicken health, which often rely on human experience, face challenges related to low efficiency and poor real-time performance. In this study, we focused on caged layer chickens and proposed an improved abnormal beak detection model based on the You Only Look Once v8 (YOLOv8) framework. Data collection was conducted using an inspection robot, enhancing automation and consistency. To address the interference caused by chicken cages, an Efficient Multi-Scale Attention (EMA) mechanism was integrated into the Spatial Pyramid Pooling-Fast (SPPF) module within the backbone network, significantly improving the model’s ability to capture fine-grained beak features. Additionally, the standard convolutional blocks in the neck of the original model were replaced with Grouped Shuffle Convolution (GSConv) modules, effectively reducing information loss during feature extraction. The model was deployed on edge computing devices for the real-time detection of abnormal beak features in layer chickens. Beyond local detection, a digital twin remote monitoring system was developed, combining three-dimensional (3D) modeling, the Internet of Things (IoT), and cloud-edge collaboration to create a dynamic, real-time mapping of physical layer farms to their virtual counterparts. This innovative approach not only improves the extraction of subtle features but also addresses occlusion challenges commonly encountered in small target detection. Experimental results demonstrate that the improved model achieved a detection accuracy of 92.7%. In terms of the comprehensive evaluation metric (mAP), it surpassed the baseline model and YOLOv5 by 2.4% and 3.2%, respectively. The digital twin system also proved stable in real-world scenarios, effectively mapping physical conditions to virtual environments. Overall, this study integrates deep learning and digital twin technology into a smart farming system, presenting a novel solution for the digital transformation of poultry farming. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

24 pages, 1264 KiB  
Review
Indoor Abnormal Behavior Detection for the Elderly: A Review
by Tianxiao Gu and Min Tang
Sensors 2025, 25(11), 3313; https://doi.org/10.3390/s25113313 - 24 May 2025
Viewed by 857
Abstract
Due to the increased age of the global population, the proportion of the elderly population continues to rise. The safety of the elderly living alone is becoming an increasingly prominent area of concern. They often miss timely treatment due to undetected falls or [...] Read more.
Due to the increased age of the global population, the proportion of the elderly population continues to rise. The safety of the elderly living alone is becoming an increasingly prominent area of concern. They often miss timely treatment due to undetected falls or illnesses, which pose risks to their lives. In order to address this challenge, the technology of indoor abnormal behavior detection has become a research hotspot. This paper systematically reviews detection methods based on sensors, video, infrared, WIFI, radar, depth, and multimodal fusion. It analyzes the technical principles, advantages, and limitations of various methods. This paper further explores the characteristics of relevant datasets and their applicable scenarios and summarizes the challenges facing current research, including multimodal data scarcity, risk of privacy leakage, insufficient adaptability of complex environments, and human adoption of wearable devices. Finally, this paper proposes future research directions, such as combining generative models, federated learning to protect privacy, multi-sensor fusion for robustness, and abnormal behavior detection on the Internet of Things environment. This paper aims to provide a systematic reference for academic research and practical application in the field of indoor abnormal behavior detection. Full article
(This article belongs to the Section Wearables)
Show Figures

Figure 1

14 pages, 2185 KiB  
Review
Ten Questions on Using Lung Ultrasonography to Diagnose and Manage Pneumonia in Hospital-at-Home Model: Part II—Confounders and Mimickers
by Nin-Chieh Hsu, Yu-Feng Lin, Hung-Bin Tsai, Charles Liao and Chia-Hao Hsu
Diagnostics 2025, 15(10), 1200; https://doi.org/10.3390/diagnostics15101200 - 9 May 2025
Viewed by 748
Abstract
The hospital-at-home (HaH) model offers hospital-level care within patients’ homes and has proven effective for managing conditions such as pneumonia. The point-of-care ultrasonography (PoCUS) is a key diagnostic tool in this model, especially when traditional imaging modalities are unavailable. This review explores how [...] Read more.
The hospital-at-home (HaH) model offers hospital-level care within patients’ homes and has proven effective for managing conditions such as pneumonia. The point-of-care ultrasonography (PoCUS) is a key diagnostic tool in this model, especially when traditional imaging modalities are unavailable. This review explores how PoCUS can be optimized to manage pneumonia in HaH settings, focusing on its diagnostic accuracy in patients with comorbidities, differentiation from mimickers, and role in assessing disease severity. Pulmonary comorbidities, such as heart failure and interstitial lung disease (ILD), can complicate lung ultrasound (LUS) interpretation. In heart failure, combining lung, cardiac, and venous assessments (e.g., IVC collapsibility, VExUS score) improves diagnostic clarity. In ILD, distinguishing chronic changes from acute infections requires attention to B-line patterns and pleural abnormalities. PoCUS must differentiate pneumonia from conditions such as atelectasis, lung contusion, cryptogenic organizing pneumonia, eosinophilic pneumonia, and neoplastic lesions—many of which present with similar sonographic features. Serial LUS scoring provides useful information on pneumonia severity and disease progression. Studies, particularly during the COVID-19 pandemic, show correlations between worsening LUS scores and poor outcomes, including increased ventilator dependency and mortality. Furthermore, LUS scores correlate with inflammatory markers and gas exchange metrics, supporting their prognostic value. In conclusion, PoCUS in HaH care requires clinicians to integrate multi-organ ultrasound findings, clinical context, and serial monitoring to enhance diagnostic accuracy and patient outcomes. Mastery of LUS interpretation in complex scenarios is crucial to delivering personalized, high-quality care in the home setting. Full article
(This article belongs to the Special Issue Clinical Diagnosis and Management in Emergency and Hospital Medicine)
Show Figures

Figure 1

25 pages, 5307 KiB  
Article
A Transformer–VAE Approach for Detecting Ship Trajectory Anomalies in Cross-Sea Bridge Areas
by Jiawei Hou, Hongzhu Zhou, Manel Grifoll, Yusheng Zhou, Jiao Liu, Yun Ye and Pengjun Zheng
J. Mar. Sci. Eng. 2025, 13(5), 849; https://doi.org/10.3390/jmse13050849 - 25 Apr 2025
Viewed by 888
Abstract
Abnormal ship navigation behaviors in cross-sea bridge waters pose significant threats to maritime safety, creating a critical need for accurate anomaly detection methods. Ship AIS trajectory data contain complex temporal features but often lack explicit labels. Most existing anomaly detection methods heavily rely [...] Read more.
Abnormal ship navigation behaviors in cross-sea bridge waters pose significant threats to maritime safety, creating a critical need for accurate anomaly detection methods. Ship AIS trajectory data contain complex temporal features but often lack explicit labels. Most existing anomaly detection methods heavily rely on labeled or semi-supervised data, thus limiting their applicability in scenarios involving completely unlabeled ship trajectory data. Furthermore, these methods struggle to capture long-term temporal dependencies inherent in trajectory data. To address these limitations, this study proposes an unsupervised trajectory anomaly detection model combining a transformer architecture with a variational autoencoder (transformer–VAE). By training on large volumes of unlabeled normal trajectory data, the transformer–VAE employs a multi-head self-attention mechanism to model both local and global temporal relationships within the latent feature space. This approach significantly enhances the model’s ability to learn and reconstruct normal trajectory patterns, with reconstruction errors serving as the criterion for anomaly detection. Experimental results show that the transformer–VAE outperforms conventional VAE and LSTM–VAE in reconstruction accuracy and achieves better detection balance and robustness compared to LSTM–-VAE and transformer–GAN in anomaly detection. The model effectively identifies abnormal behaviors such as sudden changes in speed, heading, and trajectory deviation under fully unsupervised conditions. Preliminary experiments using the POT method validate the feasibility of dynamic thresholding, enhancing the model’s adaptability in complex maritime environments. Overall, the proposed approach enables early identification and proactive warning of potential risks, contributing to improved maritime traffic safety. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

17 pages, 11684 KiB  
Article
Wi-FiAG: Fine-Grained Abnormal Gait Recognition via CNN-BiGRU with Attention Mechanism from Wi-Fi CSI
by Anming Dong, Jiahao Zhang, Wendong Xu, Jia Jia, Shanshan Yun and Jiguo Yu
Mathematics 2025, 13(8), 1227; https://doi.org/10.3390/math13081227 - 9 Apr 2025
Viewed by 551
Abstract
Abnormal gait recognition, which aims to detect and identify deviations from normal walking patterns indicative of various health conditions or impairments, holds promising applications in healthcare and many other related fields. Currently, Wi-Fi-based abnormal gait recognition methods in the literature mainly distinguish the [...] Read more.
Abnormal gait recognition, which aims to detect and identify deviations from normal walking patterns indicative of various health conditions or impairments, holds promising applications in healthcare and many other related fields. Currently, Wi-Fi-based abnormal gait recognition methods in the literature mainly distinguish the normal and abnormal gaits, which belongs to coarse-grained classification. In this work, we explore fine-grained gait rectification methods for distinguishing multiple classes of abnormal gaits. Specifically, we propose a deep learning-based framework for multi-class abnormal gait recognition, comprising three key modules: data collection, data preprocessing, and gait classification. For the gait classification module, we design a hybrid deep learning architecture that integrates convolutional neural networks (CNNs), bidirectional gated recurrent units (BiGRUs), and an attention mechanism to enhance performance. Compared to traditional CNNs, which rely solely on spatial features, or recurrent neural networks like long short-term memory (LSTM) and gated recurrent units (GRUs), which primarily capture temporal dependencies, the proposed CNN-BiGRU network integrates both spatial and temporal features concurrently. This dual-feature extraction capability positions the proposed CNN-BiGRU architecture as a promising approach for enhancing classification accuracy in scenarios involving multiple gaits with subtle differences in their characteristics. Moreover, the attention mechanism is employed to selectively focus on critical spatiotemporal features for fine-grained abnormal gait detection, enhancing the model’s sensitivity to subtle anomalies. We construct an abnormal gait dataset comprising seven distinct gait classes to train and evaluate the proposed network. Experimental results demonstrate that the proposed method achieves an average recognition accuracy of 95%, surpassing classical baseline models by at least 2%. Full article
(This article belongs to the Special Issue Data-Driven Decentralized Learning for Future Communication Networks)
Show Figures

Figure 1

24 pages, 3963 KiB  
Article
Development of a Bayesian Network-Based Parallel Mechanism for Lower Limb Gait Rehabilitation
by Huiguo Ma, Yuqi Bao, Chao Jia, Guoqiang Chen, Jingfu Lan, Mingxi Shi, He Li, Qihan Guo, Lei Guan, Shuang Li and Peng Zhang
Biomimetics 2025, 10(4), 230; https://doi.org/10.3390/biomimetics10040230 - 8 Apr 2025
Viewed by 573
Abstract
This study aims to address the clinical needs of hemiplegic and stroke patients with lower limb motor impairments, including gait abnormalities, muscle weakness, and loss of motor coordination during rehabilitation. To achieve this, it proposes an innovative design method for a lower limb [...] Read more.
This study aims to address the clinical needs of hemiplegic and stroke patients with lower limb motor impairments, including gait abnormalities, muscle weakness, and loss of motor coordination during rehabilitation. To achieve this, it proposes an innovative design method for a lower limb rehabilitation training system based on Bayesian networks and parallel mechanisms. A Bayesian network model is constructed based on expert knowledge and structural mechanics analysis, considering key factors such as rehabilitation scenarios, motion trajectory deviations, and rehabilitation goals. By utilizing the motion characteristics of parallel mechanisms, we designed a rehabilitation training device that supports multidimensional gait correction. A three-dimensional digital model is developed, and multi-posture ergonomic simulations are conducted. The study focuses on quantitatively assessing the kinematic characteristics of the hip, knee, and ankle joints while wearing the device, establishing a comprehensive evaluation system that includes range of motion (ROM), dynamic load, and optimization matching of motion trajectories. Kinematic analysis verifies that the structural design of the device is reasonable, aiding in improving patients’ gait, enhancing strength, and restoring flexibility. The Bayesian network model achieves personalized rehabilitation goal optimization through dynamic probability updates. The design of parallel mechanisms significantly expands the range of joint motion, such as enhancing hip sagittal plane mobility and reducing dynamic load, thereby validating the notable optimization effect of parallel mechanisms on gait rehabilitation. Full article
(This article belongs to the Special Issue Advanced Service Robots: Exoskeleton Robots 2025)
Show Figures

Figure 1

22 pages, 3107 KiB  
Article
A Deep Learning Approach for Identifying Intentional AIS Signal Tampering in Maritime Trajectories
by Xiangdong Lv, Ruhao Jiang, Chao Chang, Nina Shu and Tao Wu
J. Mar. Sci. Eng. 2025, 13(4), 660; https://doi.org/10.3390/jmse13040660 - 26 Mar 2025
Viewed by 861
Abstract
In the field of maritime safety research, ship behavior analysis is usually based on data provided by automatic identification systems (AISs). Prevailing studies predominantly focus on detecting the behaviors of vessels that may affect maritime safety, especially the abnormal disappearance of ship AIS [...] Read more.
In the field of maritime safety research, ship behavior analysis is usually based on data provided by automatic identification systems (AISs). Prevailing studies predominantly focus on detecting the behaviors of vessels that may affect maritime safety, especially the abnormal disappearance of ship AIS signals, neglecting subsequent measures to trace these illegal ships. To fill this gap, we propose a deep learning model named multi-dimensional convolutional long short-term memory (MConLSTM) to tackle the challenge of recognizing ship trajectories in cases where AIS signals are intentionally altered. By employing a self-supervised approach, the model is trained using historical real-world data. Extensive experiments show that MConLSTM exhibits superior analytical capabilities when it comes to processing and analyzing AIS data. Notably, even in scenarios with scant training data, the model exhibits exceptional performance, with an average accuracy 22.74% higher than the general model. Finally, we validated the practical significance and feasibility of the proposed method by simulating real-world scenarios. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

23 pages, 2120 KiB  
Article
Urban Road Anomaly Monitoring Using Vision–Language Models for Enhanced Safety Management
by Hanyu Ding, Yawei Du and Zhengyu Xia
Appl. Sci. 2025, 15(5), 2517; https://doi.org/10.3390/app15052517 - 26 Feb 2025
Viewed by 1988
Abstract
Abnormal phenomena on urban roads, including uneven surfaces, garbage, traffic congestion, floods, fallen trees, fires, and traffic accidents, present significant risks to public safety and infrastructure, necessitating real-time monitoring and early warning systems. This study develops Urban Road Anomaly Visual Large Language Models [...] Read more.
Abnormal phenomena on urban roads, including uneven surfaces, garbage, traffic congestion, floods, fallen trees, fires, and traffic accidents, present significant risks to public safety and infrastructure, necessitating real-time monitoring and early warning systems. This study develops Urban Road Anomaly Visual Large Language Models (URA-VLMs), a generative AI-based framework designed for the monitoring of diverse urban road anomalies. The InternVL was selected as a foundational model due to its adaptability for this monitoring purpose. The URA-VLMs framework features dedicated modules for anomaly detection, flood depth estimation, and safety level assessment, utilizing multi-step prompting and retrieval-augmented generation (RAG) for precise and adaptive analysis. A comprehensive dataset of 3034 annotated images depicting various urban road scenarios was developed to evaluate the models. Experimental results demonstrate the system’s effectiveness, achieving an overall anomaly detection accuracy of 93.20%, outperforming state-of-the-art models such as InternVL2.5 and ResNet34. By facilitating early detection and real-time decision-making, this generative AI approach offers a scalable and robust solution that contributes to a smarter, safer road environment. Full article
Show Figures

Figure 1

19 pages, 13043 KiB  
Article
Anomaly-Aware Tropical Cyclone Track Prediction Using Multi-Scale Generative Adversarial Networks
by He Huang, Difei Deng, Liang Hu and Nan Sun
Remote Sens. 2025, 17(4), 583; https://doi.org/10.3390/rs17040583 - 8 Feb 2025
Viewed by 965
Abstract
Tropical cyclones (TCs) frequently encompass multiple hazards, including extreme winds, intense rainfall, storm surges, flooding, lightning, and tornadoes. Accurate methods for forecasting TC tracks are essential to mitigate the loss of life and property associated with these hazards. Despite significant advancements, accurately forecasting [...] Read more.
Tropical cyclones (TCs) frequently encompass multiple hazards, including extreme winds, intense rainfall, storm surges, flooding, lightning, and tornadoes. Accurate methods for forecasting TC tracks are essential to mitigate the loss of life and property associated with these hazards. Despite significant advancements, accurately forecasting the paths of TCs remains a challenge, particularly when they interact with complex land features, weaken into remnants after landfall, or are influenced by abnormal satellite observations. To address these challenges, we propose a generative adversarial network (GAN) model with a multi-scale architecture that processes input data at four distinct resolution levels. The model is designed to handle diverse inputs, including satellite cloud imagery, vorticity, wind speed, and geopotential height, and it features an advanced center detection algorithm to ensure precise TC center identification. Our model demonstrates robustness during testing, accurately predicting TC paths over both ocean and land while also identifying weak TC remnants. Compared to other deep learning approaches, our method achieves superior detection accuracy with an average error of 41.0 km for all landfalling TCs in Australia from 2015 to 2020. Notably, for five TCs with abnormal satellite observations, our model maintains high accuracy with a prediction error of 35.2 km, which is a scenario often overlooked by other approaches. Full article
Show Figures

Graphical abstract

20 pages, 7855 KiB  
Article
Adaptive Ultra-Wideband/Pedestrian Dead Reckoning Localization Algorithm Based on Maximum Point-by-Point Distance
by Minglin Li and Songlin Liu
Electronics 2024, 13(24), 4987; https://doi.org/10.3390/electronics13244987 - 18 Dec 2024
Viewed by 1144
Abstract
Positioning using ultra-wideband (UWB) signals can be used to achieve centimeter-level indoor positioning. UWB has been widely used in indoor localization, vehicle networking, industrial IoT, etc. However, due to non-line-of-sight (NLOS) and multipath interference problems, UWB cannot provide adequate position information, which affects [...] Read more.
Positioning using ultra-wideband (UWB) signals can be used to achieve centimeter-level indoor positioning. UWB has been widely used in indoor localization, vehicle networking, industrial IoT, etc. However, due to non-line-of-sight (NLOS) and multipath interference problems, UWB cannot provide adequate position information, which affects the final positioning accuracy. This paper proposes an adaptive UWB/PDR localization algorithm based on the maximum point-by-point distance to solve the problems of poor UWB performance and the error accumulation of the pedestrian dead reckoning (PDR) algorithm in NLOS scenarios that is used to enhance the robustness and accuracy of indoor positioning. Specifically, firstly, the cumulative distribution function (CDF) map of localization under normal conditions is obtained through offline pretraining and then compared with the CDF obtained when pedestrians are moving on the line. Then, the maximum point-by-point distance algorithm is used to identify the abnormal base stations. Then, the standard base stations are filtered out for localization. To further improve the localization accuracy, this paper proposes a UWB/PDR algorithm based on an improved adaptive extended Kalman filtering (EKF), which dynamically adjusts the position information through the adaptive factor, eliminates the influence of significant errors on the current position information and realizes multi-sensor fusion positioning. The realization results show that the algorithm in this paper has a solid ability to identify abnormal base stations and that the adaptive extended Kalman filtering (AEKF) algorithm is improved by 81.27%, 58.50%, 29.76%, and 18.06% compared to the PDR, UWB, EKF, and AEKF algorithms, respectively. Full article
Show Figures

Figure 1

21 pages, 4785 KiB  
Article
A Unified Multi-Task Learning Model with Joint Reverse Optimization for Simultaneous Skin Lesion Segmentation and Diagnosis
by Mohammed A. Al-masni, Abobakr Khalil Al-Shamiri, Dildar Hussain and Yeong Hyeon Gu
Bioengineering 2024, 11(11), 1173; https://doi.org/10.3390/bioengineering11111173 - 20 Nov 2024
Cited by 2 | Viewed by 1516
Abstract
Classifying and segmenting skin cancer represent pivotal objectives for automated diagnostic systems that utilize dermoscopy images. However, these tasks present significant challenges due to the diverse shape variations of skin lesions and the inherently fuzzy nature of dermoscopy images, including low contrast and [...] Read more.
Classifying and segmenting skin cancer represent pivotal objectives for automated diagnostic systems that utilize dermoscopy images. However, these tasks present significant challenges due to the diverse shape variations of skin lesions and the inherently fuzzy nature of dermoscopy images, including low contrast and the presence of artifacts. Given the robust correlation between the classification of skin lesions and their segmentation, we propose that employing a combined learning method holds the promise of considerably enhancing the performance of both tasks. In this paper, we present a unified multi-task learning strategy that concurrently classifies abnormalities of skin lesions and allows for the joint segmentation of lesion boundaries. This approach integrates an optimization technique known as joint reverse learning, which fosters mutual enhancement through extracting shared features and limiting task dominance across the two tasks. The effectiveness of the proposed method was assessed using two publicly available datasets, ISIC 2016 and PH2, which included melanoma and benign skin cancers. In contrast to the single-task learning strategy, which solely focuses on either classification or segmentation, the experimental findings demonstrated that the proposed network improves the diagnostic capability of skin tumor screening and analysis. The proposed method achieves a significant segmentation performance on skin lesion boundaries, with Dice Similarity Coefficients (DSC) of 89.48% and 88.81% on the ISIC 2016 and PH2 datasets, respectively. Additionally, our multi-task learning approach enhances classification, increasing the F1 score from 78.26% (baseline ResNet50) to 82.07% on ISIC 2016 and from 82.38% to 85.50% on PH2. This work showcases its potential applicability across varied clinical scenarios. Full article
(This article belongs to the Section Biosignal Processing)
Show Figures

Graphical abstract

17 pages, 5063 KiB  
Article
Enhancing Recovery of Structural Health Monitoring Data Using CNN Combined with GRU
by Nguyen Thi Cam Nhung, Hoang Nguyen Bui and Tran Quang Minh
Infrastructures 2024, 9(11), 205; https://doi.org/10.3390/infrastructures9110205 - 16 Nov 2024
Cited by 2 | Viewed by 1355
Abstract
Structural health monitoring (SHM) plays a crucial role in ensuring the safety of infrastructure in general, especially critical infrastructure such as bridges. SHM systems allow the real-time monitoring of structural conditions and early detection of abnormalities. This enables managers to make accurate decisions [...] Read more.
Structural health monitoring (SHM) plays a crucial role in ensuring the safety of infrastructure in general, especially critical infrastructure such as bridges. SHM systems allow the real-time monitoring of structural conditions and early detection of abnormalities. This enables managers to make accurate decisions during the operation of the infrastructure. However, for various reasons, data from SHM systems may be interrupted or faulty, leading to serious consequences. This study proposes using a Convolutional Neural Network (CNN) combined with Gated Recurrent Units (GRUs) to recover lost data from accelerometer sensors in SHM systems. CNNs are adept at capturing spatial patterns in data, making them highly effective for recognizing localized features in sensor signals. At the same time, GRUs are designed to model sequential dependencies over time, making the combined architecture particularly suited for time-series data. A dataset collected from a real bridge structure will be used to validate the proposed method. Different cases of data loss are considered to demonstrate the feasibility and potential of the CNN-GRU approach. The results show that the CNN-GRU hybrid network effectively recovers data in both single-channel and multi-channel data loss scenarios. Full article
Show Figures

Figure 1

21 pages, 8944 KiB  
Article
Industrial Image Anomaly Detection via Self-Supervised Learning with Feature Enhancement Assistance
by Bin Wu and Xiaoqi Wang
Appl. Sci. 2024, 14(16), 7301; https://doi.org/10.3390/app14167301 - 19 Aug 2024
Cited by 2 | Viewed by 3702
Abstract
Industrial anomaly detection is constrained by the scarcity of anomaly samples, limiting the applicability of supervised learning methods. Many studies have focused on anomaly detection by generating anomaly images and adopting self-supervised learning approaches. Leveraging pre-trained networks on ImageNet has been explored to [...] Read more.
Industrial anomaly detection is constrained by the scarcity of anomaly samples, limiting the applicability of supervised learning methods. Many studies have focused on anomaly detection by generating anomaly images and adopting self-supervised learning approaches. Leveraging pre-trained networks on ImageNet has been explored to assist in this training process. However, achieving accurate anomaly detection remains time-consuming due to the network’s depth and parameter count not being reduced. In this paper, we propose a self-supervised learning method based on Feature Enhancement Patch Distribution Modeling (FEPDM), which generates simulated anomalies. Unlike direct training on the original feature extraction network, our approach utilizes a pre-trained network to extract multi-scale features. By aggregating these multi-scale features, we are able to train at the feature level, thereby adapting more efficiently to various network structures and reducing domain bias with respect to natural image classification. Additionally, it significantly reduces the number of parameters in the training process. Introducing this approach not only enhances the model’s generalization ability but also significantly improves the efficiency of anomaly detection. The method was evaluated on MVTec AD and BTAD datasets, and (image-level, pixel-level) AUROC scores of (95.7%, 96.2%), (93.4%, 97.6%) were obtained, respectively. The experimental results have convincingly demonstrated the efficacy of our method in tackling the scarcity of abnormal samples in industrial scenarios, while simultaneously highlighting its broad generalizability. Full article
(This article belongs to the Special Issue State-of-the-Art of Computer Vision and Pattern Recognition)
Show Figures

Figure 1

24 pages, 3230 KiB  
Article
Augmented Millimeter Wave Radar and Vision Fusion Simulator for Roadside Perception
by Haodong Liu, Jian Wan, Peng Zhou, Shanshan Ding and Wei Huang
Electronics 2024, 13(14), 2729; https://doi.org/10.3390/electronics13142729 - 11 Jul 2024
Cited by 2 | Viewed by 2646
Abstract
Millimeter-wave radar has the advantages of strong penetration, high-precision speed detection and low power consumption. It can be used to conduct robust object detection in abnormal lighting and severe weather conditions. The emerging 4D millimeter-wave radar has improved the quality and quantity of [...] Read more.
Millimeter-wave radar has the advantages of strong penetration, high-precision speed detection and low power consumption. It can be used to conduct robust object detection in abnormal lighting and severe weather conditions. The emerging 4D millimeter-wave radar has improved the quality and quantity of generated point clouds. Adding radar–camera fusion enhances the tracking reliability of transportation system operation. However, it is challenging due to the absence of standardized testing methods. Hence, this paper proposes a radar–camera fusion algorithm testing framework in a highway roadside scenario using SUMO and CARLA simulators. First, we propose a 4D millimeter-wave radar simulation method. A roadside multi-sensor perception dataset is generated in a 3D environment through co-simulation. Then, deep-learning object detection models are trained under different weather and lighting conditions. Finally, we propose a baseline fusion method for the algorithm testing framework. This framework provides a realistic virtual environment for device selection, algorithm testing and parameter tuning for millimeter-wave radar–camera fusion algorithms. Solutions show that the method proposed in this paper can provide a realistic virtual environment for radar–camera fusion algorithm testing for roadside traffic perception. Compared to the camera-only tracking method, the radar–vision fusion method proposed significantly improves tracking performance in rainy night scenarios. The trajectory RMSE is improved by 68.61% in expressway scenarios and 67.45% in urban scenarios. This method can also be applied to improve the detection of stop-and-go waves on congested expressways. Full article
Show Figures

Figure 1

17 pages, 1626 KiB  
Article
Modeling Autonomous Vehicle Responses to Novel Observations Using Hierarchical Cognitive Representations Inspired Active Inference
by Sheida Nozari, Ali Krayani, Pablo Marin, Lucio Marcenaro, David Martin Gomez and Carlo Regazzoni
Computers 2024, 13(7), 161; https://doi.org/10.3390/computers13070161 - 28 Jun 2024
Cited by 3 | Viewed by 1813
Abstract
Equipping autonomous agents for dynamic interaction and navigation is a significant challenge in intelligent transportation systems. This study aims to address this by implementing a brain-inspired model for decision making in autonomous vehicles. We employ active inference, a Bayesian approach that models decision-making [...] Read more.
Equipping autonomous agents for dynamic interaction and navigation is a significant challenge in intelligent transportation systems. This study aims to address this by implementing a brain-inspired model for decision making in autonomous vehicles. We employ active inference, a Bayesian approach that models decision-making processes similar to the human brain, focusing on the agent’s preferences and the principle of free energy. This approach is combined with imitation learning to enhance the vehicle’s ability to adapt to new observations and make human-like decisions. The research involved developing a multi-modal self-awareness architecture for autonomous driving systems and testing this model in driving scenarios, including abnormal observations. The results demonstrated the model’s effectiveness in enabling the vehicle to make safe decisions, particularly in unobserved or dynamic environments. The study concludes that the integration of active inference with imitation learning significantly improves the performance of autonomous vehicles, offering a promising direction for future developments in intelligent transportation systems. Full article
(This article belongs to the Special Issue System-Integrated Intelligence and Intelligent Systems 2023)
Show Figures

Figure 1

Back to TopTop