Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,782)

Search Parameters:
Keywords = machining feature recognition

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
26 pages, 4563 KB  
Article
Personalized Smart Home Automation Using Machine Learning: Predicting User Activities
by Mark M. Gad, Walaa Gad, Tamer Abdelkader and Kshirasagar Naik
Sensors 2025, 25(19), 6082; https://doi.org/10.3390/s25196082 - 2 Oct 2025
Abstract
A personalized framework for smart home automation is introduced, utilizing machine learning to predict user activities and allow for the context-aware control of living spaces. Predicting user activities, such as ‘Watch_TV’, ‘Sleep’, ‘Work_On_Computer’, and ‘Cook_Dinner’, is essential for improving occupant comfort, optimizing energy [...] Read more.
A personalized framework for smart home automation is introduced, utilizing machine learning to predict user activities and allow for the context-aware control of living spaces. Predicting user activities, such as ‘Watch_TV’, ‘Sleep’, ‘Work_On_Computer’, and ‘Cook_Dinner’, is essential for improving occupant comfort, optimizing energy consumption, and offering proactive support in smart home settings. The Edge Light Human Activity Recognition Predictor, or EL-HARP, is the main prediction model used in this framework to predict user behavior. The system combines open-source software for real-time sensing, facial recognition, and appliance control with affordable hardware, including the Raspberry Pi 5, ESP32-CAM, Tuya smart switches, NFC (Near Field Communication), and ultrasonic sensors. In order to predict daily user activities, three gradient-boosting models—XGBoost, CatBoost, and LightGBM (Gradient Boosting Models)—are trained for each household using engineered features and past behaviour patterns. Using extended temporal features, LightGBM in particular achieves strong predictive performance within EL-HARP. The framework is optimized for edge deployment with efficient training, regularization, and class imbalance handling. A fully functional prototype demonstrates real-time performance and adaptability to individual behavior patterns. This work contributes a scalable, privacy-preserving, and user-centric approach to intelligent home automation. Full article
(This article belongs to the Special Issue Sensor-Based Human Activity Recognition)
Show Figures

Graphical abstract

18 pages, 24741 KB  
Article
Cross-Domain Residual Learning for Shared Representation Discovery
by Baoqi Zhao, Jie Pan, Zhijie Zhang and Fang Yang
Information 2025, 16(10), 852; https://doi.org/10.3390/info16100852 - 2 Oct 2025
Abstract
In order to solve the problem of inconsistent data distribution in machine learning, domain adaptation based on feature representation methods extracts features from the source domain, and transfers them to the target domain for classification. The existing feature representation-based methods mainly solve the [...] Read more.
In order to solve the problem of inconsistent data distribution in machine learning, domain adaptation based on feature representation methods extracts features from the source domain, and transfers them to the target domain for classification. The existing feature representation-based methods mainly solve the problem of inconsistent feature distribution between the source domain data and the target domain data, but only a few methods analyze the correlation of cross-domain features between the original space and shared latent space, which reduces the performance of domain adaptation. To this end, we propose a domain adaptation method with a residual module, the main ideas of which are as follows: (1) transfer the source domain data features to the target domain data through the shared latent space to achieve features sharing; (2) build a cross-domain residual learning model using the latent feature space as the residual connection of the original feature space, which improves the propagation efficiency of features; (3) use a regular feature space to sparse feature representation, which can improve the robustness of the model; and (4) give an optimization algorithm, and the experiments on the public visual datasets (Office31, Office-Caltech, Office-Home, PIE, MNIST-UPS, COIL20) results show that our method achieved 92.7% accuracy on Office-Caltech and 83.2% on PIE and achieved the highest recognition accuracy in three datasets, which verify the effectiveness of the method. Full article
(This article belongs to the Special Issue Machine Learning in Image Processing and Computer Vision)
Show Figures

Figure 1

43 pages, 3035 KB  
Article
Real-Time Recognition of NZ Sign Language Alphabets by Optimal Use of Machine Learning
by Seyed Ebrahim Hosseini, Mubashir Ali, Shahbaz Pervez and Muneer Ahmad
Bioengineering 2025, 12(10), 1068; https://doi.org/10.3390/bioengineering12101068 - 30 Sep 2025
Abstract
The acquisition of a person’s first language is one of their greatest accomplishments. Nevertheless, being fluent in sign language presents challenges for many deaf students who rely on it for communication. Effective communication is essential for both personal and professional interactions and is [...] Read more.
The acquisition of a person’s first language is one of their greatest accomplishments. Nevertheless, being fluent in sign language presents challenges for many deaf students who rely on it for communication. Effective communication is essential for both personal and professional interactions and is critical for community engagement. However, the lack of a mutually understood language can be a significant barrier. Estimates indicate that a large portion of New Zealand’s disability population is deaf, with an educational approach predominantly focused on oralism, emphasizing spoken language. This makes it essential to bridge the communication gap between the general public and individuals with speech difficulties. The aim of this project is to develop an application that systematically cycles through each letter and number in New Zealand Sign Language (NZSL), assessing the user’s proficiency. This research investigates various machine learning methods for hand gesture recognition, with a focus on landmark detection. In computer vision, identifying specific points on an object—such as distinct hand landmarks—is a standard approach for feature extraction. Evaluation of this system has been performed using machine learning techniques, including Random Forest (RF) Classifier, k-Nearest Neighbours (KNN), AdaBoost (AB), Naïve Bayes (NB), Support Vector Machine (SVM), Decision Trees (DT), and Logistic Regression (LR). The dataset used for model training and testing consists of approximately 100,000 hand gesture expressions, formatted into a CSV dataset for model training. Full article
(This article belongs to the Special Issue AI and Data Science in Bioengineering: Innovations and Applications)
31 pages, 1529 KB  
Review
Artificial Intelligence-Enhanced Liquid Biopsy and Radiomics in Early-Stage Lung Cancer Detection: A Precision Oncology Paradigm
by Swathi Priya Cherukuri, Anmolpreet Kaur, Bipasha Goyal, Hanisha Reddy Kukunoor, Areesh Fatima Sahito, Pratyush Sachdeva, Gayathri Yerrapragada, Poonguzhali Elangovan, Mohammed Naveed Shariff, Thangeswaran Natarajan, Jayarajasekaran Janarthanan, Samuel Richard, Shakthidevi Pallikaranai Venkatesaprasath, Shiva Sankari Karuppiah, Vivek N. Iyer, Scott A. Helgeson and Shivaram P. Arunachalam
Cancers 2025, 17(19), 3165; https://doi.org/10.3390/cancers17193165 - 29 Sep 2025
Cited by 1
Abstract
Background: Lung cancer remains the leading cause of cancer-related mortality globally, largely due to delayed diagnosis in its early stages. While conventional diagnostic tools like low-dose CT and tissue biopsy are routinely used, they suffer from limitations including invasiveness, radiation exposure, cost, and [...] Read more.
Background: Lung cancer remains the leading cause of cancer-related mortality globally, largely due to delayed diagnosis in its early stages. While conventional diagnostic tools like low-dose CT and tissue biopsy are routinely used, they suffer from limitations including invasiveness, radiation exposure, cost, and limited sensitivity for early-stage detection. Liquid biopsy, a minimally invasive alternative that captures circulating tumor-derived biomarkers such as ctDNA, cfRNA, and exosomes from body fluids, offers promising diagnostic potential—yet its sensitivity in early disease remains suboptimal. Recent advances in Artificial Intelligence (AI) and radiomics are poised to bridge this gap. Objective: This review aims to explore how AI, in combination with radiomics, enhances the diagnostic capabilities of liquid biopsy for early detection of lung cancer and facilitates personalized monitoring strategies. Content Overview: We begin by outlining the molecular heterogeneity of lung cancer, emphasizing the need for earlier, more accurate detection strategies. The discussion then transitions into liquid biopsy and its key analytes, followed by an in-depth overview of AI techniques—including machine learning (e.g., SVMs, Random Forest) and deep learning models (e.g., CNNs, RNNs, GANs)—that enable robust pattern recognition across multi-omics datasets. The role of radiomics, which quantitatively extracts spatial and morphological features from imaging modalities such as CT and PET, is explored in conjunction with AI to provide an integrative, multimodal approach. This convergence supports the broader vision of precision medicine by integrating omics data, imaging, and electronic health records. Discussion: The synergy between AI, liquid biopsy, and radiomics signifies a shift from traditional diagnostics toward dynamic, patient-specific decision-making. Radiomics contributes spatial information, while AI improves pattern detection and predictive modeling. Despite these advancements, challenges remain—including data standardization, limited annotated datasets, the interpretability of deep learning models, and ethical considerations. A push toward rigorous validation and multimodal AI frameworks is necessary to facilitate clinical adoption. Conclusion: The integration of AI with liquid biopsy and radiomics holds transformative potential for early lung cancer detection. This non-invasive, scalable, and individualized diagnostic paradigm could significantly reduce lung cancer mortality through timely and targeted interventions. As technology and regulatory pathways mature, collaborative research is crucial to standardize methodologies and translate this innovation into routine clinical practice. Full article
(This article belongs to the Special Issue The Genetic Analysis and Clinical Therapy in Lung Cancer: 2nd Edition)
Show Figures

Figure 1

18 pages, 449 KB  
Review
Decoding Emotions from fNIRS: A Survey on Tensor-Based Approaches in Affective Computing and Medical Applications
by Aleksandra Kawala-Sterniuk, Michal Podpora, Dariusz Mikolajewski, Maciej Piasecki, Ewa Rudnicka, Adrian Luckiewicz, Adam Sudol and Mariusz Pelc
Appl. Sci. 2025, 15(19), 10525; https://doi.org/10.3390/app151910525 - 29 Sep 2025
Abstract
Understanding and interpreting human emotions through neurophysiological signals has become a central goal in affective computing. This paper presents a focused survey of recent advances in emotion recognition using tensor factorization techniques specifically applied to functional Near-Infrared Spectroscopy (fNIRS) data. We examine how [...] Read more.
Understanding and interpreting human emotions through neurophysiological signals has become a central goal in affective computing. This paper presents a focused survey of recent advances in emotion recognition using tensor factorization techniques specifically applied to functional Near-Infrared Spectroscopy (fNIRS) data. We examine how tensor-based frameworks have been leveraged to capture the temporal, spatial, and spectral characteristics of fNIRS brain signals, enabling effective dimensionality reduction and latent pattern extraction. Focusing on third-order tensor constructions (trials × channels × time), we compare the use of Canonical Polyadic (CP) and Tucker decompositions in isolating components representative of emotional states. The review further evaluates the performance of extracted features when classified by conventional machine learning models such as Random Forests and Support Vector Machines. Emphasis is placed on comparative accuracy, interpretability, and the advantages of tensor methods over traditional approaches for distinguishing arousal and valence levels. We conclude by discussing the relevance of these methods for the development of real-time, explainable, emotion-aware systems in wearable neurotechnology, with a particular focus on medical applications such as mental health monitoring, early diagnosis of affective disorders, and personalized neurorehabilitation. Full article
Show Figures

Figure 1

68 pages, 8643 KB  
Article
From Sensors to Insights: Interpretable Audio-Based Machine Learning for Real-Time Vehicle Fault and Emergency Sound Classification
by Mahmoud Badawy, Amr Rashed, Amna Bamaqa, Hanaa A. Sayed, Rasha Elagamy, Malik Almaliki, Tamer Ahmed Farrag and Mostafa A. Elhosseini
Machines 2025, 13(10), 888; https://doi.org/10.3390/machines13100888 - 28 Sep 2025
Abstract
Unrecognized mechanical faults and emergency sounds in vehicles can compromise safety, particularly for individuals with hearing impairments and in sound-insulated or autonomous driving environments. As intelligent transportation systems (ITSs) evolve, there is a growing need for inclusive, non-intrusive, and real-time diagnostic solutions that [...] Read more.
Unrecognized mechanical faults and emergency sounds in vehicles can compromise safety, particularly for individuals with hearing impairments and in sound-insulated or autonomous driving environments. As intelligent transportation systems (ITSs) evolve, there is a growing need for inclusive, non-intrusive, and real-time diagnostic solutions that enhance situational awareness and accessibility. This study introduces an interpretable, sound-based machine learning framework to detect vehicle faults and emergency sound events using acoustic signals as a scalable diagnostic source. Three purpose-built datasets were developed: one for vehicular fault detection, another for emergency and environmental sounds, and a third integrating both to reflect real-world ITS acoustic scenarios. Audio data were preprocessed through normalization, resampling, and segmentation and transformed into numerical vectors using Mel-Frequency Cepstral Coefficients (MFCCs), Mel spectrograms, and Chroma features. To ensure performance and interpretability, feature selection was conducted using SHAP (explainability), Boruta (relevance), and ANOVA (statistical significance). A two-phase experimental workflow was implemented: Phase 1 evaluated 15 classical models, identifying ensemble classifiers and multi-layer perceptrons (MLPs) as top performers; Phase 2 applied advanced feature selection to refine model accuracy and transparency. Ensemble models such as Extra Trees, LightGBM, and XGBoost achieved over 91% accuracy and AUC scores exceeding 0.99. SHAP provided model transparency without performance loss, while ANOVA achieved high accuracy with fewer features. The proposed framework enhances accessibility by translating auditory alarms into visual/haptic alerts for hearing-impaired drivers and can be integrated into smart city ITS platforms via roadside monitoring systems. Full article
(This article belongs to the Section Vehicle Engineering)
Show Figures

Figure 1

12 pages, 4847 KB  
Article
Surformer v1: Transformer-Based Surface Classification Using Tactile and Vision Features
by Manish Kansana, Elias Hossain, Shahram Rahimi and Noorbakhsh Amiri Golilarz
Information 2025, 16(10), 839; https://doi.org/10.3390/info16100839 - 27 Sep 2025
Abstract
Surface material recognition is a key component in robotic perception and physical interaction, particularly when leveraging both tactile and visual sensory inputs. In this work, we propose Surformer v1, a transformer-based architecture designed for surface classification using structured tactile features and Principal Component [...] Read more.
Surface material recognition is a key component in robotic perception and physical interaction, particularly when leveraging both tactile and visual sensory inputs. In this work, we propose Surformer v1, a transformer-based architecture designed for surface classification using structured tactile features and Principal Component Analysis (PCA)-reduced visual embeddings extracted via ResNet 50. The model integrates modality-specific encoders with cross-modal attention layers, enabling rich interactions between vision and touch. Currently, state-of-the-art deep learning models for vision tasks have achieved remarkable performance. With this in mind, our first set of experiments focused exclusively on tactile-only surface classification. Using feature engineering, we trained and evaluated multiple machine learning models, assessing their accuracy and inference time. We then implemented an encoder-only Transformer model tailored for tactile features. This model not only achieves the highest accuracy, but also demonstrated significantly faster inference time compared to other evaluated models, highlighting its potential for real-time applications. To extend this investigation, we introduced a multimodal fusion setup by combining vision and tactile inputs. We trained both Surformer v1 (using structured features) and a Multimodal CNN (using raw images) to examine the impact of feature-based versus image-based multimodal learning on classification accuracy and computational efficiency. The results showed that Surformer v1 achieved 99.4% accuracy with an inference time of 0.7271 ms, while the Multimodal CNN achieved slightly higher accuracy but required significantly more inference time. These findings suggest that Surformer v1 offers a compelling balance between accuracy, efficiency, and computational cost for surface material recognition. The results also underscore the effectiveness of integrating feature learning, cross-modal attention and transformer-based fusion in capturing the complementary strengths of tactile and visual modalities. Full article
(This article belongs to the Special Issue AI-Based Image Processing and Computer Vision)
Show Figures

Figure 1

19 pages, 1025 KB  
Article
Research on Trade Credit Risk Assessment for Foreign Trade Enterprises Based on Explainable Machine Learning
by Mengjie Liao, Wanying Jiao and Jian Zhang
Information 2025, 16(10), 831; https://doi.org/10.3390/info16100831 - 26 Sep 2025
Abstract
As global economic integration deepens, import and export trade plays an increasingly vital role in China’s economy. To enhance regulatory efficiency and achieve scientific, transparent credit supervision, this study proposes a trade credit risk evaluation model based on interpretable machine learning, incorporating loss [...] Read more.
As global economic integration deepens, import and export trade plays an increasingly vital role in China’s economy. To enhance regulatory efficiency and achieve scientific, transparent credit supervision, this study proposes a trade credit risk evaluation model based on interpretable machine learning, incorporating loss preferences. Key risk features are identified through a comprehensive interpretability framework combining SHapley Additive exPlanations (SHAP) and Local Interpretable Model-agnostic Explanations (LIME), forming an optimal feature subset. Using Light Gradient Boosting Machine (LightGBM) as the base model, a weight adjustment strategy is introduced to reduce costly misclassification of high-risk enterprises, effectively improving their recognition rate. However, this adjustment leads to a decline in overall accuracy. To address this trade-off, a Bagging ensemble framework is applied, which restores and slightly improves accuracy while maintaining low misclassification costs. Experimental results demonstrate that the interpretability framework improves transparency and business applicability, the weight adjustment strategy enhances high-risk enterprise detection, and Bagging balances the overall classification performance. The proposed method ensures reliable identification of high-risk enterprises while preserving overall model robustness, thereby providing strong practical value for enterprise credit risk assessment and decision-making. Full article
Show Figures

Figure 1

20 pages, 1348 KB  
Article
Joint Learning for Mask-Aware Facial Expression Recognition Based on Exposed Feature Analysis and Occlusion Feature Enhancement
by Huanyu Hou and Xiaoming Sun
Appl. Sci. 2025, 15(19), 10433; https://doi.org/10.3390/app151910433 - 26 Sep 2025
Abstract
Facial expression recognition (FER), applied in fields such as interaction and intelligent security, has seen widespread development with the advancement of machine vision technology. However, in natural environments, faces are often obscured by masks, posture, and body parts, leading to incomplete features, which [...] Read more.
Facial expression recognition (FER), applied in fields such as interaction and intelligent security, has seen widespread development with the advancement of machine vision technology. However, in natural environments, faces are often obscured by masks, posture, and body parts, leading to incomplete features, which results in poor accuracy of existing facial expression recognition algorithms. Apart from extreme scenarios where facial features are completely blocked, the key information of facial expression features is mostly preserved in most cases, yet insufficient parsing of these features leads to poor recognition results. To address this, we propose a novel joint learning framework that integrates explicit occlusion parsing and feature enhancement. Our model consists of three core modules: a Facial Occlusion Parsing Module (FOPM) for real-time occlusion estimation, an Expression Feature Fusion Module (EFFM) for integrating appearance and geometric features, and a Facial Expression Recognition Module (FERM) for final classification. Extensive experiments under a rigorous and reproducible protocol demonstrate significant improvements of our approach. On the masked facial expression datasets RAF-DB and FER+, our model achieves accuracies of 91.24% and 90.18%, surpassing previous state-of-the-art methods by 2.62% and 0.96%, respectively. Additional evaluation on a real-world masked dataset with diverse mask types further confirms the robustness and generalizability of our method, where it attains an accuracy of 89.75%. Moreover, the model maintains high computational efficiency with an inference time of 12.4 ms per image. By effectively parsing and integrating partially obscured facial features, our approach enables more accurate and robust expression recognition, which is essential for real-world applications in interaction and intelligent security systems. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

16 pages, 3708 KB  
Article
Myoelectric and Inertial Data Fusion Through a Novel Attention-Based Spatiotemporal Feature Extraction for Transhumeral Prosthetic Control: An Offline Analysis
by Andrea Tigrini, Alessandro Mengarelli, Ali H. Al-Timemy, Rami N. Khushaba, Rami Mobarak, Mara Scattolini, Gaith K. Sharba, Federica Verdini, Ennio Gambi and Laura Burattini
Sensors 2025, 25(18), 5920; https://doi.org/10.3390/s25185920 - 22 Sep 2025
Viewed by 124
Abstract
This study proposes a feature extraction scheme that fuses accelerometric (ACC) and electromyographic (EMG) data to improve shoulder movement identification in individuals with transhumeral amputation, in whom the clinical need for intuitive control strategies enabling reliable activation of full-arm prostheses is underinvestigated. A [...] Read more.
This study proposes a feature extraction scheme that fuses accelerometric (ACC) and electromyographic (EMG) data to improve shoulder movement identification in individuals with transhumeral amputation, in whom the clinical need for intuitive control strategies enabling reliable activation of full-arm prostheses is underinvestigated. A novel spatiotemporal warping feature extraction architecture was employed to realize EMG and ACC information fusion at the feature level. EMG and ACC data were collected from six participants with intact limbs and four participants with transhumeral amputation using an NI USB-6009 device at 1000 Hz to support the proposed feature extraction scheme. For each participant, a leave-one-trial-out (LOTO) training and testing approach was used for developing pattern recognition models for both the intact-limb (IL) and amputee (AMP) groups. The analysis revealed that the introduction of ACC information has a positive impact when using windows of length (WLs) lower than 150 ms. A linear discriminant analysis (LDA) classifier was able to exceed the accuracy of 90% in each WL condition and for each group. Similar results were observed for an extreme learning machine (ELM), whereas k-nearest neighbors (kNN) and an autonomous learning multi-model classifier showed a mean accuracy of less than 87% for both IL and AMP groups at different WLs, guaranteeing applicability over a large set of shallow pattern-recognition models that can be used in real scenarios. The present work lays the groundwork for future studies involving real-time validation of the proposed methodology on a larger population, acknowledging the current limitation of offline analysis. Full article
(This article belongs to the Special Issue Advanced Sensors and AI Integration for Human–Robot Teaming)
Show Figures

Figure 1

26 pages, 13551 KB  
Article
Hybrid Cloud–Edge Architecture for Real-Time Cryptocurrency Market Forecasting: A Distributed Machine Learning Approach with Blockchain Integration
by Mohammed M. Alenazi and Fawwad Hassan Jaskani
Mathematics 2025, 13(18), 3044; https://doi.org/10.3390/math13183044 - 22 Sep 2025
Viewed by 344
Abstract
The volatile nature of cryptocurrency markets demands real-time analytical capabilities that traditional centralized computing architectures struggle to provide. This paper presents a novel hybrid cloud–edge computing framework for cryptocurrency market forecasting, leveraging distributed systems to enable low-latency prediction models. Our approach integrates machine [...] Read more.
The volatile nature of cryptocurrency markets demands real-time analytical capabilities that traditional centralized computing architectures struggle to provide. This paper presents a novel hybrid cloud–edge computing framework for cryptocurrency market forecasting, leveraging distributed systems to enable low-latency prediction models. Our approach integrates machine learning algorithms across a distributed network: edge nodes perform real-time data preprocessing and feature extraction, while the cloud infrastructure handles deep learning model training and global pattern recognition. The proposed architecture uses a three-tier system comprising edge nodes for immediate data capture, fog layers for intermediate processing and local inference, and cloud servers for comprehensive model training on historical blockchain data. A federated learning mechanism allows edge nodes to contribute to a global prediction model while preserving data locality and reducing network latency. The experimental results show a 40% reduction in prediction latency compared to cloud-only solutions while maintaining comparable accuracy in forecasting Bitcoin and Ethereum price movements. The system processes over 10,000 transactions per second and delivers real-time insights with sub-second response times. Integration with blockchain ensures data integrity and provides transparent audit trails for all predictions. Full article
(This article belongs to the Special Issue Recent Computational Techniques to Forecast Cryptocurrency Markets)
Show Figures

Figure 1

17 pages, 2566 KB  
Article
Secure and Decentralized Hybrid Multi-Face Recognition for IoT Applications
by Erëza Abdullahu, Holger Wache and Marco Piangerelli
Sensors 2025, 25(18), 5880; https://doi.org/10.3390/s25185880 - 19 Sep 2025
Viewed by 369
Abstract
The proliferation of smart environments and Internet of Things (IoT) applications has intensified the demand for efficient, privacy-preserving multi-face recognition systems. Conventional centralized systems suffer from latency, scalability, and security vulnerabilities. This paper presents a practical hybrid multi-face recognition framework designed for decentralized [...] Read more.
The proliferation of smart environments and Internet of Things (IoT) applications has intensified the demand for efficient, privacy-preserving multi-face recognition systems. Conventional centralized systems suffer from latency, scalability, and security vulnerabilities. This paper presents a practical hybrid multi-face recognition framework designed for decentralized IoT deployments. Our approach leverages a pre-trained Convolutional Neural Network (VGG16) for robust feature extraction and a Support Vector Machine (SVM) for lightweight classification, enabling real-time recognition on resource-constrained devices such as IoT cameras and Raspberry Pi boards. The purpose of this work is to demonstrate the feasibility and effectiveness of a lightweight hybrid system for decentralized multi-face recognition, specifically tailored to the constraints and requirements of IoT applications. The system is validated on a custom dataset of 20 subjects collected under varied lighting conditions and facial expressions, achieving an average accuracy exceeding 95% while simultaneously recognizing multiple faces. Experimental results demonstrate the system’s potential for real-world applications in surveillance, access control, and smart home environments. The proposed architecture minimizes computational load, reduces dependency on centralized servers, and enhances privacy, offering a promising step toward scalable edge AI solutions. Full article
(This article belongs to the Special Issue Secure and Decentralised IoT Systems)
Show Figures

Figure 1

25 pages, 12760 KB  
Article
Intelligent Face Recognition: Comprehensive Feature Extraction Methods for Holistic Face Analysis and Modalities
by Thoalfeqar G. Jarullah, Ahmad Saeed Mohammad, Musab T. S. Al-Kaltakchi and Jabir Alshehabi Al-Ani
Signals 2025, 6(3), 49; https://doi.org/10.3390/signals6030049 - 19 Sep 2025
Viewed by 428
Abstract
Face recognition technology utilizes unique facial features to analyze and compare individuals for identification and verification purposes. This technology is crucial for several reasons, such as improving security and authentication, effectively verifying identities, providing personalized user experiences, and automating various operations, including attendance [...] Read more.
Face recognition technology utilizes unique facial features to analyze and compare individuals for identification and verification purposes. This technology is crucial for several reasons, such as improving security and authentication, effectively verifying identities, providing personalized user experiences, and automating various operations, including attendance monitoring, access management, and law enforcement activities. In this paper, comprehensive evaluations are conducted using different face detection and modality segmentation methods, feature extraction methods, and classifiers to improve system performance. As for face detection, four methods are proposed: OpenCV’s Haar Cascade classifier, Dlib’s HOG + SVM frontal face detector, Dlib’s CNN face detector, and Mediapipe’s face detector. Additionally, two types of feature extraction techniques are proposed: hand-crafted features (traditional methods: global local features) and deep learning features. Three global features were extracted, Scale-Invariant Feature Transform (SIFT), Speeded Robust Features (SURF), and Global Image Structure (GIST). Likewise, the following local feature methods are utilized: Local Binary Pattern (LBP), Weber local descriptor (WLD), and Histogram of Oriented Gradients (HOG). On the other hand, the deep learning-based features fall into two categories: convolutional neural networks (CNNs), including VGG16, VGG19, and VGG-Face, and Siamese neural networks (SNNs), which generate face embeddings. For classification, three methods are employed: Support Vector Machine (SVM), a one-class SVM variant, and Multilayer Perceptron (MLP). The system is evaluated on three datasets: in-house, Labelled Faces in the Wild (LFW), and the Pins dataset (sourced from Pinterest) providing comprehensive benchmark comparisons for facial recognition research. The best performance accuracy for the proposed ten-feature extraction methods applied to the in-house database in the context of the facial recognition task achieved 99.8% accuracy by using the VGG16 model combined with the SVM classifier. Full article
Show Figures

Figure 1

21 pages, 3742 KB  
Article
Research on Monitoring and Intelligent Identification of Typical Defects in Small and Medium-Sized Bridges Based on Ultra-Weak FBG Sensing Array
by Xinyan Lin, Yichan Zhang, Yinglong Kang, Sheng Li, Qiuming Nan, Lina Yue, Yan Yang and Min Zhou
Optics 2025, 6(3), 43; https://doi.org/10.3390/opt6030043 - 19 Sep 2025
Viewed by 287
Abstract
To address the challenge of efficiently identifying and providing early warnings for typical structural damages in small and medium-sized bridges during long-term service, this paper proposes an intelligent monitoring and recognition method based on ultra-weak fiber Bragg grating (UWFBG) array sensing. By deploying [...] Read more.
To address the challenge of efficiently identifying and providing early warnings for typical structural damages in small and medium-sized bridges during long-term service, this paper proposes an intelligent monitoring and recognition method based on ultra-weak fiber Bragg grating (UWFBG) array sensing. By deploying UWFBG strain-sensing cables across the bridge, the system enables continuous acquisition and spatial analysis of multi-point strain data. Based on this, a series of experimental scenarios simulating typical structural damages—such as single-slab loading, eccentric loading, and bearing detachment—are designed to systematically analyze strain evolution patterns before and after damage occurrence. While strain distribution maps allow for visual identification of some typical damages, the approach remains limited by reliance on manual interpretation, low recognition efficiency, and weak detection capability for atypical damages. To overcome these limitations, machine learning algorithms are further introduced to extract features from strain data and perform pattern recognition, enabling the construction of an automated damage identification model. This approach enhances both the accuracy and robustness of damage recognition, achieving rapid classification and intelligent diagnosis of structural conditions. The results demonstrate that the integration of the monitoring system with intelligent recognition algorithms effectively distinguishes different types of damage and shows promising potential for engineering applications. Full article
Show Figures

Figure 1

23 pages, 670 KB  
Article
DPIBP: Dining Philosophers Problem-Inspired Binary Patterns for Facial Expression Recognition
by Archana Pallakonda, Rama Muni Reddy Yanamala, Rayappa David Amar Raj, Christian Napoli and Cristian Randieri
Technologies 2025, 13(9), 420; https://doi.org/10.3390/technologies13090420 - 18 Sep 2025
Cited by 1 | Viewed by 351
Abstract
Emotion recognition plays a crucial role in our day-to-day communication, and detecting emotions is one of the most formidable tasks in the field of human–computer Interaction (HCI). Facial expressions are the most straightforward and efficient way to identify emotions. With so many real-time [...] Read more.
Emotion recognition plays a crucial role in our day-to-day communication, and detecting emotions is one of the most formidable tasks in the field of human–computer Interaction (HCI). Facial expressions are the most straightforward and efficient way to identify emotions. With so many real-time applications, although automatic facial expression recognition (FER) is essential for numerous real-world applications in computer vision, developing a feature descriptor that accurately captures the subtle variations in facial expressions remains a significant challenge. Towards addressing this issue, a novel feature extraction technique inspired by Dining Philosophers Problem, named Dining Philosophers Problem Inspired Binary Patterns (DPIBP), has been proposed in this work. The proposed DPIBP methods extract three features in a local 5 × 5 neighborhood by considering the impact of both neighboring pixels and the adjacent pixels on the current pixel. To categorize facial expressions, the system used a multi-class Support Vector Machine (SVM) classifier. Reflecting real-world use, researchers tested the method on JAFFE, MUG, CK+, and TFEID benchmark datasets using a person-independent protocol. The proposed method, DPIBP, achieved superior performance compared to existing techniques that rely on manually crafted features for extraction. Full article
Show Figures

Graphical abstract

Back to TopTop