Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (449)

Search Parameters:
Keywords = human face recognition

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 7561 KB  
Article
Fine-Grained Image Recognition with Bio-Inspired Gradient-Aware Attention
by Bing Ma, Junyi Li, Zhengbei Jin, Wei Zhang, Xiaohui Song and Beibei Jin
Biomimetics 2025, 10(12), 834; https://doi.org/10.3390/biomimetics10120834 - 12 Dec 2025
Viewed by 153
Abstract
Fine-grained image recognition is one of the key tasks in the field of computer vision. However, due to subtle inter-class differences and significant intra-class differences, it still faces severe challenges. Conventional approaches often struggle with background interference and feature degradation. To address these [...] Read more.
Fine-grained image recognition is one of the key tasks in the field of computer vision. However, due to subtle inter-class differences and significant intra-class differences, it still faces severe challenges. Conventional approaches often struggle with background interference and feature degradation. To address these issues, we draw inspiration from the human visual system, which adeptly focuses on discriminative regions, to propose a bio-inspired gradient-aware attention mechanism. Our method explicitly models gradient information to guide the attention, mimicking biological edge sensitivity, thereby enhancing the discrimination between global structures and local details. Experiments on the CUB-200-2011, iNaturalist2018, nabbirds and Stanford Cars datasets demonstrated the superiority of our method, achieving Top-1 accuracy rates of 92.9%, 90.5%, 93.1% and 95.1%, respectively. Full article
(This article belongs to the Special Issue Biologically Inspired Vision and Image Processing 2025)
Show Figures

Figure 1

22 pages, 2905 KB  
Article
Image Captioning with Object Detection and Facial Expression Recognition for Smart Industry
by Abdul Saboor Khan, Abdul Haseeb Khan, Muhammad Jamshed Abbass and Imran Shafi
Bioengineering 2025, 12(12), 1325; https://doi.org/10.3390/bioengineering12121325 - 5 Dec 2025
Viewed by 436
Abstract
This paper presents a new image captioning system which contains facial expression recognition as a way to provide better emotional and contextual comprehension of the captions generated. A combination of affective cues and visual features is made, which enables semantically full and emotionally [...] Read more.
This paper presents a new image captioning system which contains facial expression recognition as a way to provide better emotional and contextual comprehension of the captions generated. A combination of affective cues and visual features is made, which enables semantically full and emotionally conscious descriptions. Experiments were carried out on two created datasets, FlickrFace11k and COCOFace15k, with standard benchmarks such as BLEU, METEOR, ROUGE-L, CIDEr, and SPICE to analyze their effectiveness. The suggested model produced better results in all metrics as compared to baselines, like Show-Attend-Tell and Up-Down, remaining consistently better on all the scores. Remarkably, it has reached gains of 2.5 points on CIDEr and 1.0 on SPICE, which means a closer correlation to the prompt captions made by people. A 5-fold cross-validation confirmed the model’s robustness, with minimal standard deviation across folds (<±0.2). Qualitative results further demonstrated its ability to capture fine-grained emotional expressions often missed by conventional models. These findings underscore the model’s potential in affective computing, assistive technologies, and human-centric AI applications. The pipeline is designed for on-prem/edge deployment with lightweight interfaces to IoT middleware (MQTT/OPC UA), enabling smart-factory integration. These characteristics align the method with Industry 4.0 sensor networks and human-centric analytics. Full article
(This article belongs to the Special Issue AI-Driven Imaging and Analysis for Biomedical Applications)
Show Figures

Graphical abstract

24 pages, 1367 KB  
Article
Algorithmic Empowerment and Its Impact on Circular Economy Participation: An Empirical Study Based on Human–Machine Collaborative Decision-Making Mechanisms
by Xingjun Ru, Le Liu and Min Chen
J. Theor. Appl. Electron. Commer. Res. 2025, 20(4), 353; https://doi.org/10.3390/jtaer20040353 - 5 Dec 2025
Viewed by 318
Abstract
At the intersection of the circular economy and artificial intelligence (AI), high-value secondhand trading faces a “triple decision dilemma” of cognitive overload, trust risk, and emotional attachment. To address the limits of traditional human-centered theories, this study develops and empirically tests a novel [...] Read more.
At the intersection of the circular economy and artificial intelligence (AI), high-value secondhand trading faces a “triple decision dilemma” of cognitive overload, trust risk, and emotional attachment. To address the limits of traditional human-centered theories, this study develops and empirically tests a novel framework of Algorithmic Empowerment. Drawing on data from 1396 users of Chinese secondhand luxury platforms and analyzed using Partial Least Squares Structural Equation Modeling (PLS-SEM), the findings reveal that users’ empowerment perception arises from three dimensions—Algorithmic Connectivity (AC), Human–Agent Symbiotic Trust (HAST), and Algorithmic Value Alignment (AVA). This perceived empowerment affects participation willingness through two parallel pathways: the social pathway, where algorithmic curation shapes social norms and recognition, and the cognitive pathway, where AI enhances decision fluency and reduces cognitive friction. The results confirm the dual mediating effects of these mechanisms. This study advances understanding of human–AI collaboration in sustainable consumption by conceptualizing empowerment as the bridge linking algorithmic functions to user engagement, and provides actionable implications for designing AI systems that both enhance efficiency and foster user trust and identification. Full article
Show Figures

Figure 1

23 pages, 3296 KB  
Article
Enhancing the Effectiveness of Juvenile Protection: Deep Learning-Based Facial Age Estimation via JPSD Dataset Construction and YOLO-ResNet50
by Yuqiang Wu, Qingyang Gao, Yichen Lin, Zhanhai Yang and Xinmeng Wang
Appl. Syst. Innov. 2025, 8(6), 185; https://doi.org/10.3390/asi8060185 - 29 Nov 2025
Viewed by 327
Abstract
An increasing number of juveniles are accessing adult-oriented venues, such as bars and nightclubs, where supervision is frequently inadequate, thereby elevating their risk of both offline harm and unmonitored exposure to harmful online content. Existing facial age estimation systems, which are primarily designed [...] Read more.
An increasing number of juveniles are accessing adult-oriented venues, such as bars and nightclubs, where supervision is frequently inadequate, thereby elevating their risk of both offline harm and unmonitored exposure to harmful online content. Existing facial age estimation systems, which are primarily designed for adults, have significant limitations when it comes to protecting juveniles, hindering the efficiency of supervising them in key venues. To address these challenges, this study proposes a facial age estimation solution for juvenile protection. First, we have designed a ‘detection–cropping–classification’ framework comprising three stages. This first detects facial regions using a detection algorithm, then crops the image before inputting the results into a classification model for age estimation. Secondly, we constructed the the Juvenile Protection Surveillance and Detection (JPSD) Dataset by integrating five public datasets: UTKface, AgeDB, APPA-REAL, MegaAge and FG-NET. This dataset contains 14,260 images categorised into four age groups: 0–8 years, 8–14 years, 14–18 years and over 18 years. Thirdly, we conducted baseline model comparisons. In the object detection phase, three YOLO algorithms were selected for face recognition. In the age estimation phase, traditional convolutional neural networks (CNNs), such as ResNet50 and VGG16, were contrasted with vision transformer (ViT)-based models, such as ViT and BiFormer. Gradient-weighted Class Activation Mapping (Grad-CAM) was used for visual analysis to highlight differences in the models’ decision-making processes. Experiments revealed that YOLOv11 is the optimal detector for accurate facial localisation, and that ResNet50 is the best base classifier for enhancing age-sensitive feature extraction, outperforming BiFormer. The results show that the framework achieves Recall of 89.17% for the 0–8 age group and 95.17% for the over-18 age group. However, we have found that the current model has low Recall rates for the 8–14 and 14–18 age groups. Therefore, in the near term, we emphasise that this technology should only be used as a decision-support tool under strict human-in-the-loop supervision. This study provides an essential dataset and technical framework for juvenile facial age estimation, offering support for juvenile online protection, smart policing and venue supervision. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

21 pages, 998 KB  
Article
Attention-Based CNN-BiGRU-Transformer Model for Human Activity Recognition
by Mingda Miao, Weijie Yan, Xueshan Gao, Le Yang, Jiaqi Zhou and Wenyi Zhang
Appl. Sci. 2025, 15(23), 12592; https://doi.org/10.3390/app152312592 - 27 Nov 2025
Viewed by 340
Abstract
Human activity recognition (HAR) based on wearable sensors is a key technology in the fields of smart sensing and health monitoring. With the rapid development of deep learning, its powerful feature extraction capabilities have significantly enhanced recognition performance and reduced reliance on traditional [...] Read more.
Human activity recognition (HAR) based on wearable sensors is a key technology in the fields of smart sensing and health monitoring. With the rapid development of deep learning, its powerful feature extraction capabilities have significantly enhanced recognition performance and reduced reliance on traditional handcrafted feature engineering. However, current deep learning models still face challenges in effectively capturing complex temporal dependencies in long-term time-series sensor data and addressing information redundancy, which affect model recognition accuracy and generalization ability. To address these issues, this paper proposes an innovative CNN-BiGRU–Transformer hybrid deep learning model aimed at improving the accuracy and robustness of human activity recognition. The proposed model integrates a multi-branch Convolutional Neural Network (CNN) to effectively extract multi-scale local spatial features, and combines a Bidirectional Gated Recurrent Unit (BiGRU) with a Transformer hybrid module for modeling temporal dependencies and extracting temporal features in long-term time-series data. Additionally, an attention mechanism is incorporated to dynamically allocate weights, suppress redundant information, and enhance key features, further improving recognition performance. To demonstrate the capability of the proposed model, evaluations are performed on three public datasets: WISDM, PAMAP2, and UCI-HAR. The model achieved recognition accuracies of 98.41%, 95.62%, and 96.74% on the three datasets, respectively, outperforming several state-of-the-art methods. These results confirm that the proposed approach effectively addresses feature extraction and redundancy challenges in long-term sensor time-series data and provides a robust solution for wearable sensor-based human activity recognition. Full article
Show Figures

Figure 1

23 pages, 5957 KB  
Article
TeaPickingNet: Towards Robust Recognition of Fine-Grained Picking Actions in Tea Gardens Using an Attention-Enhanced Framework
by Ru Han, Ye Zheng, Lei Shu and Grzegorz Cielniak
Agriculture 2025, 15(23), 2441; https://doi.org/10.3390/agriculture15232441 - 26 Nov 2025
Viewed by 192
Abstract
With the emergence of smart agriculture, precise behavior recognition in tea gardens has become increasingly important for operational monitoring and intelligent management. However, existing behavior detection systems face limitations when deployed in real-world agricultural environments, particularly due to dense vegetation, variable lighting, and [...] Read more.
With the emergence of smart agriculture, precise behavior recognition in tea gardens has become increasingly important for operational monitoring and intelligent management. However, existing behavior detection systems face limitations when deployed in real-world agricultural environments, particularly due to dense vegetation, variable lighting, and diverse human–machine interactions. This paper proposes a novel deep learning framework for picking behavior recognition tailored to complex tea plantation environments. We first construct a large-scale, annotated dataset comprising 12,195 images across 7 behavior categories, collected from both field and web sources, capturing a diverse range of geographic, temporal, and environmental conditions. To address occlusion and multi-scale detection challenges, we enhance YOLOv5 by integrating an Exponential Moving Average (EMA) attention mechanism, Complete Intersection over Union (CIoU) loss, and Atrous Spatial Pyramid Pooling (ASPP), achieving a 73.6% mAP@0.5, representing an 11.6% relative improvement over the baseline model, which indicates a notable enhancement in detection accuracy under complex tea garden conditions. Furthermore, we propose an SE-Faster R-CNN model by embedding Squeeze-and-Excitation (SE) channel attention modules and anchor box optimization strategies, which significantly boosts performance in complex scenarios. A lightweight visual interface for real-time image and video-based detection is also developed to enhance the practical deployability of the system. Experimental results demonstrate the effectiveness, robustness, and real-time potential of the proposed system in recognizing tea garden behaviors under real field conditions. Full article
(This article belongs to the Section Artificial Intelligence and Digital Agriculture)
Show Figures

Figure 1

30 pages, 2818 KB  
Article
LAViTSPose: A Lightweight Cascaded Framework for Robust Sitting Posture Recognition via Detection– Segmentation–Classification
by Shu Wang, Adriano Tavares, Carlos Lima, Tiago Gomes, Yicong Zhang, Jiyu Zhao and Yanchun Liang
Entropy 2025, 27(12), 1196; https://doi.org/10.3390/e27121196 - 25 Nov 2025
Viewed by 290
Abstract
Sitting posture recognition, defined as automatically localizing and categorizing seated human postures, has become essential for large-scale ergonomics assessment and longitudinal health-risk monitoring in classrooms and offices. However, in real-world multi-person scenes, pervasive occlusions and overlaps induce keypoint misalignment, causing global-attention backbones to [...] Read more.
Sitting posture recognition, defined as automatically localizing and categorizing seated human postures, has become essential for large-scale ergonomics assessment and longitudinal health-risk monitoring in classrooms and offices. However, in real-world multi-person scenes, pervasive occlusions and overlaps induce keypoint misalignment, causing global-attention backbones to fail to localize critical local structures. Moreover, annotation scarcity makes small-sample training commonplace, leaving models insufficiently robust to misalignment perturbations and thereby limiting cross-domain generalization. To address these challenges, we propose LAViTSPose, a lightweight cascaded framework for sitting posture recognition. Concretely, a YOLOR-based detector trained with a Range-aware IoU (RaIoU) loss yields tight person crops under partial visibility; ESBody suppresses cross-person leakage and estimates occlusion/head-orientation cues; a compact ViT head (MLiT) with Spatial Displacement Contact (SDC) and a learnable temperature (LT) mechanism performs skeleton-only classification with a local structural-consistency regularizer. From an information-theoretic perspective, our design enhances discriminative feature compactness and reduces structural entropy under occlusion and annotation scarcity. We conducted a systematic evaluation on the USSP dataset, and the results show that LAViTSPose outperforms existing methods on both sitting posture classification and face-orientation recognition while meeting real-time inference requirements. Full article
(This article belongs to the Special Issue Entropy in Machine Learning Applications, 2nd Edition)
Show Figures

Figure 1

16 pages, 1543 KB  
Article
Inferring Mental States via Linear and Non-Linear Body Movement Dynamics: A Pilot Study
by Tad T. Brunyé, Kana Okano, James McIntyre, Madelyn K. Sandone, Lisa N. Townsend, Marissa Marko Lee, Marisa Smith and Gregory I. Hughes
Sensors 2025, 25(22), 6990; https://doi.org/10.3390/s25226990 - 15 Nov 2025
Viewed by 559
Abstract
Stress, workload, and uncertainty characterize occupational tasks across sports, healthcare, military, and transportation domains. Emerging theory and empirical research suggest that coordinated whole-body movements may reflect these transient mental states. Wearable sensors and optical motion capture offer opportunities to quantify such movement dynamics [...] Read more.
Stress, workload, and uncertainty characterize occupational tasks across sports, healthcare, military, and transportation domains. Emerging theory and empirical research suggest that coordinated whole-body movements may reflect these transient mental states. Wearable sensors and optical motion capture offer opportunities to quantify such movement dynamics and classify mental states that influence occupational performance and human–machine interaction. We tested this possibility in a small pilot study (N = 10) designed to test feasibility and identify preliminary movement features linked to mental states. Participants performed a perceptual decision-making task involving facial emotion recognition (i.e., deciding whether depicted faces were happy versus angry) with variable levels of stress (via a risk of electric shock), workload (via time pressure), and uncertainty (via visual degradation of task stimuli). The time series of movement trajectories was analyzed both holistically (full trajectory) and by phase: lowered (early), raising (middle), aiming (late), and face-to-face (sequential). For each epoch, up to 3844 linear and non-linear features were extracted across temporal, spectral, probability, divergence, and fractal domains. Features were entered into a repeated 10-fold cross-validation procedure using 80/20 train/test splits. Feature selection was conducted with the T-Rex Selector, and selected features were used to train a scikit-learn pipeline with a Robust Scaler and a Logistic Regression classifier. Models achieved mean ROC AUC scores as high as 0.76 for stress classification, with the highest sensitivity during the full movement trajectory and middle (raise) phases. Classification of workload and uncertainty states was less successful. These findings demonstrate the potential of movement-based sensing to infer stress states in applied settings and inform future human–machine interface development. Full article
(This article belongs to the Special Issue Sensors and Data Analysis for Biomechanics and Physical Activity)
Show Figures

Figure 1

30 pages, 1077 KB  
Review
A Contemporary Multidimensional Insight into the Clinical and Pathological Presentation of Urological Conditions Associated with HIV: A Narrative Review
by Hannah Faherty, Jamshaid Nasir Shahid, Yousef Abu Osba, Maryam Jamshaid, Dushyant Mital and Mohamed H. Ahmed
Trop. Med. Infect. Dis. 2025, 10(11), 318; https://doi.org/10.3390/tropicalmed10110318 - 11 Nov 2025
Viewed by 674
Abstract
Human Immunodeficiency Virus (HIV) infection is associated with a wide spectrum of urological manifestations, reflecting both the direct effects of viral infection and the indirect consequences of immunosuppression, opportunistic infections, malignancies and long-term combined antiretroviral therapy (cART). This narrative review provides a contemporary, [...] Read more.
Human Immunodeficiency Virus (HIV) infection is associated with a wide spectrum of urological manifestations, reflecting both the direct effects of viral infection and the indirect consequences of immunosuppression, opportunistic infections, malignancies and long-term combined antiretroviral therapy (cART). This narrative review provides a contemporary, multifaceted overview of the clinical and pathological presentations of urological conditions in people living with HIV (PLWHIV), based on articles published between 1989 and 2025. Conditions discussed include HIV-associated nephropathy (HIVAN), opportunistic genitourinary infections, malignancies such as Kaposi sarcoma and lymphoma, as well as non-infectious complications such as HIV-associated nephropathy and erectile dysfunction (ED). The review highlights the evolving epidemiology of these conditions in the cART era, with a noted decline in opportunistic infections but a rising burden of chronic kidney disease and malignancies, largely due to improved survival and ageing of the HIV-positive population. Pathological insights are explored and discussed, including mechanisms of HIV-associated renal injury, such as direct viral infection of renal epithelial cells and genetic predispositions linked to Apolipoprotein L1 (APOL1) variants. In addition, psychosocial factors, including anxiety, stress, stigma, and alcohol use, are discussed, as they may contribute to late presentation to clinical urology services. The review also considers the challenges faced in low and middle-income countries, the impact of HIV on urological services, and the important role of palliative care in advanced disease. Ultimately, this review underscores the need for early recognition, comprehensive diagnostic and surgical evaluation, and integrated social, psychological, and palliative management strategies tailored to the unique needs of PLWHIV. A deeper understanding of the interplay between HIV, cART, psychosocial determinants, and urological health is essential for improving patient outcomes and guiding future research in this evolving field. Full article
(This article belongs to the Special Issue HIV Testing, Prevention and Care Interventions, 2nd Edition)
Show Figures

Figure 1

17 pages, 2093 KB  
Article
Plant Bioelectrical Signals for Environmental and Emotional State Classification
by Peter A. Gloor
Biosensors 2025, 15(11), 744; https://doi.org/10.3390/bios15110744 - 5 Nov 2025
Viewed by 1135
Abstract
In this study, we present a pilot investigation using a single Purple Heart plant (Tradescantia pallida) to explore whether bioelectrical signals for dual-purpose classification tasks: environmental state detection and human emotion recognition. Using an AD8232 ECG sensor at 400 Hz sampling rate, we [...] Read more.
In this study, we present a pilot investigation using a single Purple Heart plant (Tradescantia pallida) to explore whether bioelectrical signals for dual-purpose classification tasks: environmental state detection and human emotion recognition. Using an AD8232 ECG sensor at 400 Hz sampling rate, we recorded 3 s bioelectrical signal segments with 1 s overlap, converting them to mel-spectrograms for ResNet18 CNN (Convolutional Neural Network) classification. For lamp on/off detection, we achieved 85.4% accuracy with balanced precision (0.85–0.86) and recall (0.84–0.86) metrics across 2767 spectrogram samples. For human emotion classification, our system achieved optimal performance at 73% accuracy with 1 s lag, distinguishing between happy and sad emotional states across 1619 samples. These results should be viewed as preliminary and exploratory, demonstrating feasibility rather than definitive evidence of plant-based emotion sensing. Replication across plants, days, and experimental sites will be essential to establish robustness. The current study is limited by a single-plant setup, modest sample size, and reliance on human face-tracking labels, which together preclude strong claims about generalizability. Full article
(This article belongs to the Special Issue Biosensing Technology in Agriculture and Biological Products)
Show Figures

Figure 1

17 pages, 1645 KB  
Article
Cross-Dataset Emotion Valence Prediction Approach from 4-Channel EEG: CNN Model and Multi-Modal Evaluation
by Vladimir Romaniuk and Alexey Kashevnik
Big Data Cogn. Comput. 2025, 9(11), 280; https://doi.org/10.3390/bdcc9110280 - 5 Nov 2025
Viewed by 1011
Abstract
Emotion recognition based on electroencephalography (EEG) has gained significant attention due to its potential applications in human–computer interaction, affective computing, and mental health assessment. This study presents a convolutional neural network-based approach to emotion valence prediction model development using 4-channel headband EEG data [...] Read more.
Emotion recognition based on electroencephalography (EEG) has gained significant attention due to its potential applications in human–computer interaction, affective computing, and mental health assessment. This study presents a convolutional neural network-based approach to emotion valence prediction model development using 4-channel headband EEG data as well as its evaluation based on computer vision emotion valence recognition. We trained a model on the publicly available FACED and SEED datasets and tested it on a newly collected dataset recorded using a wearable BrainBit headband. The model’s performance is evaluated using both standard train–validation–test splitting and a leave-one-subject-out cross-validation strategy. Additionally, the model is evaluated on computer vision-based emotion recognition system to assess the reliability and consistency of EEG-based emotion prediction. Experimental results demonstrate that the CNN model achieves competitive accuracy in predicting emotion valence from EEG signals, despite the challenges posed by limited channel availability and individual variability. The findings show the usability of compact EEG devices for real-time emotion recognition and their potential integration into adaptive user interfaces and mental health applications. Full article
(This article belongs to the Special Issue Advances in Complex Networks)
Show Figures

Figure 1

20 pages, 3299 KB  
Article
WIS: A Technology of Wireless Non-Contact Incremental Training Sensing
by Guanjie Wang, Yu Wu, Hongyu Sun, Xinyue Zhang, Wanjia Li and Yanhua Dong
Electronics 2025, 14(21), 4326; https://doi.org/10.3390/electronics14214326 - 4 Nov 2025
Viewed by 362
Abstract
Wireless contactless human activity sensing is a new type of sensing method that uses the propagation characteristics of wireless signals to accurately perceive and understand human behavior. However, facing the huge amount of newly generated data and expanding action categories in the sensing [...] Read more.
Wireless contactless human activity sensing is a new type of sensing method that uses the propagation characteristics of wireless signals to accurately perceive and understand human behavior. However, facing the huge amount of newly generated data and expanding action categories in the sensing process, the traditional model needs to be retrained frequently. This model not only brings significant computational power overhead, but also seriously affects the real-time response speed of the system. To address this problem, this paper proposes a model, WIS (Wireless Incremental Sense), which is composed of two parts. The first part is the basic sensing module NFFCN (Nearest Neighbor Feature Fusion Classification). NFFCN is a fusion classification method based on Nearest Class Mean (NCM) classifier and Random Forest (RF). By combining the local feature extraction ability of NCM and the integrated learning advantage of RF, this method can efficiently extract human behavior features from wireless signals and achieve high-precision classification. The second part is the incremental learning module NFFCN-RTST, which uses the retraining subtree (RTST) incremental strategy to optimize the model. Unlike update leaf statistics (ULS) and the Incrementally Grow Tree (IGT) incremental strategy, RTST not only updates the statistical data of leaf nodes but also dynamically adjusts the previously learned splitting function, so as to better adapt to new data and categories. In the experimental validation on the rRuler and Stan WiFi datasets, the average recognition accuracy of NFFCN reaches 87.1% and 98.4%, respectively. In the class-incremental experimental validation, the recognition accuracy of WIS reaches 87% and 95%, respectively. Full article
Show Figures

Figure 1

9 pages, 229 KB  
Essay
Clash Actions: Indigenous Peoples’ Human Rights and Class Actions
by Cindy Blackstock and Pamela Palmater
Genealogy 2025, 9(4), 122; https://doi.org/10.3390/genealogy9040122 - 3 Nov 2025
Viewed by 1873
Abstract
As many face significant financial costs and legal barriers to accessing justice to remedy systemic human rights violations rooted in colonialism, they are increasingly turning to class action litigation for recognition of harms and to safeguard others. Drawing on Canadian examples, including a [...] Read more.
As many face significant financial costs and legal barriers to accessing justice to remedy systemic human rights violations rooted in colonialism, they are increasingly turning to class action litigation for recognition of harms and to safeguard others. Drawing on Canadian examples, including a class action involving First Nations children, this article examines the complex and sometimes conflicting relationship between class actions and human rights remedies. The paper highlights the risks of class actions displacing human rights awards, the ethical challenges in relationships between class counsel and Indigenous victims, and the limited effectiveness of settlements in preventing recurring injustices. The article concludes by calling for stronger regulation of class action lawyers and tethering such proceedings to the United Nations Declaration on the Rights of Indigenous Peoples and other human rights standards, including the United Nations Convention on the Rights of the Child. Full article
(This article belongs to the Special Issue Self Determination in First Peoples Child Protection)
25 pages, 3059 KB  
Article
A Lightweight Framework for Pilot Pose Estimation and Behavior Recognition with Integrated Safety Assessment
by Honglan Wu, Xin Lu, Youchao Sun and Hao Liu
Aerospace 2025, 12(11), 986; https://doi.org/10.3390/aerospace12110986 - 3 Nov 2025
Viewed by 643
Abstract
With the rapid advancement of aviation technology, modern aircraft cockpits are evolving toward high automation and intelligence, making pilot-cockpit interaction a critical factor influencing flight safety and efficiency. Pilot pose estimation and behavior recognition are critical for monitoring pilot state, preventing operational errors, [...] Read more.
With the rapid advancement of aviation technology, modern aircraft cockpits are evolving toward high automation and intelligence, making pilot-cockpit interaction a critical factor influencing flight safety and efficiency. Pilot pose estimation and behavior recognition are critical for monitoring pilot state, preventing operational errors, and enabling adaptive human–machine interaction, thus playing an essential role in aviation safety assurance and intelligent cockpit development. However, existing methods face challenges in real-time performance, reliability, and computational complexity in practical applications. Traditional approaches, such as wearable sensors and image-processing-based algorithms, demonstrate certain effectiveness but still exhibit limitations in aviation environments. To address these issues, this paper proposes a lightweight pilot pose estimation and behavior recognition framework, integrating Vision Transformer with depth-wise separable convolution to optimize the accuracy and efficiency of keypoint detection. Additionally, a novel multimodal data fusion technique is introduced, along with a scientifically designed evaluation system, to enhance the robustness and security of the system in complex environments. Experimental results on a pilot keypoint detection dataset captured in a simulated cockpit environment show that the proposed method achieves 81.9 AP, while substantially reducing model parameters and notably improving inference efficiency compared with HRNet. This study provides new insights and methodologies for the design and evaluation of aviation human-machine interaction systems. Full article
(This article belongs to the Section Air Traffic and Transportation)
Show Figures

Figure 1

18 pages, 1906 KB  
Article
Generalizable Interaction Recognition for Learning from Demonstration Using Wrist and Object Trajectories
by Jagannatha Charjee Pyaraka, Mats Isaksson, John McCormick, Sheila Sutjipto and Fouad Sukkar
Electronics 2025, 14(21), 4297; https://doi.org/10.3390/electronics14214297 - 31 Oct 2025
Viewed by 539
Abstract
Learning from Demonstration (LfD) enables robots to acquire manipulation skills by observing human actions. However, existing methods often face challenges such as high computational cost, limited generalizability, and a loss of key interaction details. This study presents a compact representation for interaction recognition [...] Read more.
Learning from Demonstration (LfD) enables robots to acquire manipulation skills by observing human actions. However, existing methods often face challenges such as high computational cost, limited generalizability, and a loss of key interaction details. This study presents a compact representation for interaction recognition in LfD that encodes human–object interactions using 2D wrist trajectories and 3D object poses. A lightweight extraction pipeline combines MediaPipe-based wrist tracking with FoundationPose-based 6-DoF object estimation to obtain these trajectories directly from RGB-D video without specialized sensors or heavy preprocessing. Experiments on the GRAB and FPHA datasets show that the representation effectively captures task-relevant interactions, achieving 94.6% accuracy on GRAB and 96.0% on FPHA with well-calibrated probability predictions. Both Bidirectional Long Short-Term Memory (Bi-LSTM) with attention and Transformer architectures deliver consistent performance, confirming robustness and generalizability. The method achieves sub-second inference, a memory footprint under 1 GB, and reliable operation on both GPU and CPU platforms, enabling deployment on edge devices such as NVIDIA Jetson. By bridging pose-based and object-centric paradigms, this approach offers a compact and efficient foundation for scalable robot learning while preserving essential spatiotemporal dynamics. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

Back to TopTop