Next Article in Journal
High-Performance Hydrogen Gas Sensor Based on Pd-Doped MoS2/Si Heterojunction
Previous Article in Journal
Distributed Collaborative Data Processing Framework for Unmanned Platforms Based on Federated Edge Intelligence
Previous Article in Special Issue
A Scalable Approach to IoT Interoperability: The Share Pattern
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Lightweight Vision Transformer for Frame-Level Ergonomic Posture Classification in Industrial Workflows

by
Luca Cruciata
1,†,
Salvatore Contino
1,†,
Marianna Ciccarelli
2,
Roberto Pirrone
1,
Leonardo Mostarda
3,
Alessandra Papetti
2 and
Marco Piangerelli
4,5,*
1
Department of Engineering, University of Palermo, 90128 Palermo, Italy
2
Department of Industrial Engineering and Mathematical Sciences, Poytechnic University of Marche, Via Brecce Bianche 12, 60131 Ancona, Italy
3
Department of Mathematics and Informatics, University of Perugia, Via Vanvitelli 1, 06123 Perugia, Italy
4
Computer Science Division, School of Science and Technology, University of Camerino, Via Madonna delle Carceri 7, 62032 Camerino, Italy
5
Vici & C. S.p.A., Via Gutenberg 5, 47822 Santarcangelo di Romagna, Italy
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2025, 25(15), 4750; https://doi.org/10.3390/s25154750 (registering DOI)
Submission received: 26 June 2025 / Revised: 25 July 2025 / Accepted: 30 July 2025 / Published: 1 August 2025
(This article belongs to the Special Issue Secure and Decentralised IoT Systems)

Abstract

Work-related musculoskeletal disorders (WMSDs) are a leading concern in industrial ergonomics, often stemming from sustained non-neutral postures and repetitive tasks. This paper presents a vision-based framework for real-time, frame-level ergonomic risk classification using a lightweight Vision Transformer (ViT). The proposed system operates directly on raw RGB images without requiring skeleton reconstruction, joint angle estimation, or image segmentation. A single ViT model simultaneously classifies eight anatomical regions, enabling efficient multi-label posture assessment. Training is supervised using a multimodal dataset acquired from synchronized RGB video and full-body inertial motion capture, with ergonomic risk labels derived from RULA scores computed on joint kinematics. The system is validated on realistic, simulated industrial tasks that include common challenges such as occlusion and posture variability. Experimental results show that the ViT model achieves state-of-the-art performance, with F1-scores exceeding 0.99 and AUC values above 0.996 across all regions. Compared to previous CNN-based system, the proposed model improves classification accuracy and generalizability while reducing complexity and enabling real-time inference on edge devices. These findings demonstrate the model’s potential for unobtrusive, scalable ergonomic risk monitoring in real-world manufacturing environments.

1. Introduction

Work-related musculoskeletal disorders (WMSDs) remain among the most pressing occupational health challenges worldwide, particularly in manual and industrial labor contexts. In the EU alone, WMSDs affect nearly 60% of workers, regardless of the sector or job role [1]. These disorders originate from a complex interplay of biomechanical, organizational, and psychosocial risk factors and commonly manifest in the back, shoulders, neck, and upper limbs. Among the key physical risk factors, manual material handling, awkward postures, repetitive motions, and exposure to vibrations have been consistently associated with the onset and progression of musculoskeletal disorders. Although limited exposure to these factors may not cause immediate harm, sustained or high-intensity demands often exceed physiological tolerance, increasing the risk of injury over time [2]. Large-scale surveys highlight the scale of this problem as follows: in Europe, 44% of employees are exposed to tiring or painful postures for at least a quarter of their working time, with 13% reporting such exposure for more than three-quarters of their shift. Likewise, repetitive movements of the hands or arms affect 61% of workers, with a significant share enduring these conditions for most of their workday [2]. Beyond pain and disability, WMSDs also involve high economic costs in terms of reduced productivity, absenteeism, and health care burdens [3]. The manufacturing sector, in particular, suffers from high rates of sick leave and absenteeism due to WMSDs, with the back and upper limbs among the areas most frequently affected [4]. For example, in Germany alone, WMSDs account for an estimated EUR 6.45 million in annual production losses and EUR 10.63 million in lost gross value added [5].
This scenario highlights the urgent need for effective risk assessment strategies capable of identifying hazardous conditions before they result in injury. To this end, traditional ergonomic assessment methods—such as OCRA (OCcupational Repetitive Action), REBA (Rapid Entire Body Assessment), and RULA (Rapid Upper Limb Assessment)—have been widely adopted due to their ease of use and standardization. These methods involve scoring postures and loads based on visual inspections or video analysis by trained evaluators. However, they suffer from limitations, including evaluator subjectivity, limited temporal resolution, and a strong reliance on manual interpretation, making them impractical for continuous, real-time, or large-scale monitoring [6,7]. Finally, while detailed biomechanical simulations offer insights into physical load distributions under postural strain, these approaches are computationally intensive and not directly applicable to scalable in-field ergonomic assessment [8].
These constraints have motivated the development of automated, objective methods based on sensor technologies and computer vision. Motion capture has become a cornerstone of biomechanical research, offering reliable quantitative insights into human movement dynamics [9]. A wide range of technologies are used for this purpose, including optical, wearable, and hybrid systems [10]. In laboratory settings, marker-based optical motion capture systems equipped with infrared cameras are still regarded as a gold standard due to their precision. However, their applicability is largely restricted to controlled environments, and they involve high setup costs and complex data processing workflows. Among more deployable alternatives, inertial measurement units (IMUs) have received widespread attention due to their portability and ability to capture full-body kinematics in non-laboratory environments. IMUs typically integrate accelerometers, gyroscopes, and magnetometers, enabling joint angle computation and motion segmentation. These systems have been applied to diverse ergonomic scenarios, such as warehouse operations [11], muscle load monitoring via EMG [12], and agricultural harvesting tasks [13]. A modular low-cost motion capture system capable of estimating full-body segment orientations using quaternion-based Kalman filtering, with additional gait recognition functionality, was proposed in [14]. More recently, an open source, sensor-based pipeline combining custom IMUs and OpenSim inverse kinematics, has been introduced and validated in a real automotive plant with <5% deviation in RULA scores compared to commercial references [15]. Despite their demonstrated potential, the use of IMUs in ergonomic risk assessment remains limited by several practical constraints [16]. These include sensor drift, the need for recalibration in dynamic conditions [17], discomfort from prolonged wear [18], short battery life [19], and the specialized expertise required for data elaboration and interpretation [20]. Moreover, some authors emphasized that the effective integration of wearable motion capture into ergonomic workflows often entails additional analytical effort, which can reduce operational efficiency and hinder widespread adoption [21]. Economic factors also play a role, with high-end systems remaining unaffordable for many small and medium enterprises [22]. Still, recent studies point to measurable improvements in ergonomic conditions through wearable technologies, with artificial intelligence emerging as a key enabler for automated risk detection [23]. Despite their growing technical maturity, the integration of wearable systems into routine industrial workflows for ergonomic risk monitoring remains limited, primarily due to usability constraints and the complexity of post-processing pipelines. While the development of tools that automate post-processing and support risk factor identification can help address many of the current limitations of sensor-based systems, their use in real-time applications—such as delivering immediate feedback to workers or managing human–machine collaboration—remains constrained. In such contexts, vision-based systems play a complementary role, enabling more responsive and comprehensive ergonomic monitoring.
As a result, interest has grown in vision-based markerless motion capture systems that leverage computer vision and machine learning (ML) or deep learning (DL) to estimate postures from RGB or RGB-D video [24,25]. These systems offer the advantage of being fully non-invasive, low-cost, and scalable. These approaches typically follow a two-stage pipeline, outlined as follows: first, body pose is estimated from RGB or RGB-D video using ML DL methods; then, joint angles or geometric features are extracted and mapped to ergonomic risk indices such as RULA, REBA, or OWAS. Technically, vision-based systems can be classified by spatial representation (2D vs. 3D), input modality (monocular or multi-view), and model architecture (single- or multi-person pipelines). Early implementations used 2D detection with OpenCV and Haarcascade [26], while more sophisticated models such as OpenPose [27] and VoxelPose [28] enabled 3D pose estimation via multi-camera or volumetric strategies. However, challenges such as occlusion, variability in lighting, camera calibration, and restricted field of view still compromise their performance in industrial settings [29,30,31,32].
Together, these points motivate the present study, which combines a single-view pose-estimation backbone with a lightweight temporal graph encoder. A wide range of studies have applied CNN-based methods for pose estimation and activity recognition. One approach employed a 3D CNN to classify writing postures [33], while other works proposed temporal CNNs to extract spatiotemporal features for human action recognition [34,35]. Several systems estimate joint angles from skeletons to compute ergonomic scores. For instance, RULA scores have been predicted using 2D joint data combined with Euclidean and angular metrics [36].
Other similar efforts rely on skeleton-based representations with varying levels of accuracy [24,37,38,39]. Despite promising results, these methods often lack robustness in uncontrolled industrial environments and remain sensitive to lighting, occlusion, and camera viewpoint. Furthermore, they usually depend on intermediate steps (e.g., joint angle estimation) that introduce error propagation and reduce real-time feasibility.
Recent advances have introduced annotated datasets and benchmarking efforts to improve vision-based ergonomic scoring. For instance, some authors extended the COCO dataset with ergonomic key points for REBA assessment and developed REBAPose, which improved pose inference and scoring accuracy [27]. However, this work remains confined to 2D and lacks validation in operational scenarios. PoseChecker [40] represents a validated markerless motion capture tool that aligns closely with marker-based optical motion capture and IMU references in terms of joint tracking and ergonomic scoring accuracy. Related to this, vision-language models have also been proposed to sanitize industrial datasets containing noisy or weak labels. For instance, some authors proposed VLSR, a CLIP-based framework that aligns image and text embeddings to cluster, refine, and consolidate labels in multi-label manufacturing datasets. While our study relies on high-fidelity motion capture for ground truth, such techniques are highly relevant for enabling scalable data annotation in real-world applications [41]. However, results may not generalize to real-world industrial scenarios characterized by occlusions and variable lighting. A 2D RGB MoCap system was evaluated across multiple real-world industrial environments, confirming its practical potential, although performance varied under conditions involving occlusions and nonstandard layouts [42]. Additional comparisons between Kinect V2, OpenCV, and IMU-based systems for torso posture estimation indicated that Kinect remains a viable option, despite its sensitivity to lighting conditions and limited tracking scope [43]. In addition, hybrid frameworks are emerging that combine posture detection with ergonomic reasoning or decision making. As an example, the integration of VideoPose3D with fuzzy logic for RULA scoring and task reassignment to collaborative robots has shown promising results in productivity and ergonomic load reduction, although it is limited to single-user offline setups [44].
An early work demonstrated that even a compact CNN trained on roughly one thousand images could already classify four industrial lifting postures with ≈94% accuracy, opening the door to camera-based ergonomic screening [45]. Building on that proof of concept, a Mask-R-CNN person segmentation step was added in [46]. This way posture and fall detection remained reliable in cluttered domestic rooms at real-time speed (≈28 fps on a GTX-1080). To cope with short-term occlusions common on factory floors, one study fused four consecutive frames within a multi-stream CNN, boosting the F-score to 0.88 even when a worker was momentarily hidden by a workpiece [47].
Although these vision-only systems rely on RGB or depth imagery, several groups began to exploit estimated 3-D skeletons for finer-grained reasoning. Çalışkan (2022) [48] tiled entire joint-coordinate sequences into “posture mosaics”, achieving perfect accuracy on the small KARD and MSR-Pairs datasets—an impressive result. Seeking a direct link to ergonomic scores, Hossain et al. (2023) [49] trained on 3.6 million synthetic poses to map 3-D key-points straight to REBA risk levels, achieving 89 % overall accuracy but still missing around 29 % of minority high-risk cases.
A complementary direction was explored by Kwon et al. (2022) [50], who showed that monocular RGB cameras could approximate expert RULA ratings ( ρ = 0.82 ) on their DyWHSE dataset—although heavy self-occlusion and personal protective equipment remained problematic. Finally, Jung et al. (2025) [51] reminded the community that accurate posture labels are only a means to an end; using 7270 compensation records, they predicted the work-relatedness of shoulder disorders with 84 % accuracy, demonstrating how machine learning can close the loop between exposure assessment and clinical outcomes.
The following three unresolved challenges emerge from this body of work:
  • Occlusion resilience—despite temporal fusion and person segmentation, long or permanent occlusions still degrade joint tracking and risk estimates.
  • Dataset realism and scale—many claims of near-perfect accuracy rely on small or proprietary datasets that do not capture the lighting, attire and workflow variability of real factories.
  • Outcome linkage—only a handful of studies quantify posture recognition through indices directly linked to musculoskeletal risk prediction.
In general, most vision-based systems rely on skeleton reconstruction and joint angle estimation to derive ergonomic scores, which may not be optimal under industrial constraints. In this context, the present study contributes to the field by introducing the use of a state-of-the-art neural network, the Vision Transformer (ViT). Compared to convolutional network approaches, this network manages more efficiently processing of the relevant information, without requiring extensive pre-processing of important image features. Furthermore, the ViT used in this work manages to provide the classification for each of the body parts through a single training, in contrast to the first version of this approach [52]. In this way, the overall computational cost is reduced considerably by speeding up the training and inference phases; the limited number of parameters could favour near real-time applications of the prediction, returning predictions in real time with a very high accuracy. The main contributions are summarized as follows:
  • A vision-based ergonomic risk classification framework is proposed, leveraging a ViT architecture for direct analysis of RGB images, without relying on skeleton reconstruction or joint angle estimation.
  • The system is designed to classify multiple body regions simultaneously within a single model, reducing computational complexity and simplifying training and inference procedures.
  • The approach is validated on a multimodal dataset of simulated industrial tasks, reflecting realistic working conditions and addressing common challenges such as posture variability and occlusion.
  • The classification output is directly linked to ergonomic risk indices, enabling frame-level posture risk assessment applicable to industrial monitoring scenarios.
This paper is organized as follows. Section 2 describes the data acquisition protocol, the dataset processing pipeline, and the architecture of the ViT-based classifier. Section 3 presents the experimental setup, evaluation metrics, and classification results. Section 4 discusses the findings, highlighting the strengths and limitations of the approach. Finally, Section 5 concludes the paper and outlines possible future research directions.

2. Materials and Methods

2.1. Dataset Creation

The dataset used in this study was originally developed and presented in our previous work, SPECTRE, where we proposed a preliminary vision-based method for ergonomic risk assessment in collaborative industrial environments [52]. The data acquisition protocol, fully described in that paper, involved the simulation of typical manual operations performed in an Italian kitchen manufacturer, such as handling, assembly, and quality control, reproduced in a laboratory setting. The acquisition setup was designed to replicate realistic industrial conditions, including poor lighting, obstacles, and dynamic operator movements, thereby introducing natural variability in postures, occlusions, and viewpoints.
Data acquisition was carried out using a dual system composed of a full-body inertial motion capture system (Xsens MVN) and a high-resolution RGB video camera, both operating at 60 frames per second. The motion capture system comprised 17 wireless MTw sensors positioned according to standard biomechanical placement guidelines, allowing the continuous recording of joint angles and body segment orientations. The video stream was used as input for the image-based posture classification, while the kinematic data served as ground truth for ergonomic labeling. Synchronization between the two data streams was ensured by displaying the motion capture frame counter within the camera’s field of view at the beginning of each recording. This enabled precise alignment of each video frame with the corresponding row in the Xsens output.
For each frame, the motion capture data were used to compute joint angles and segment orientations for the following key body parts: upper and lower arms, wrists, neck, and trunk, separately for left and right sides. A custom algorithm processed the motion capture output to derive these joint angles and compare them against threshold values derived from the RULA method [53]. Each body segment in each frame was then labeled as either “OK” (ergonomic posture) or “KO” (non-ergonomic posture), depending on whether the angles exceeded the critical values defined by RULA for that joint [52]. Dataset generation was performed through an automated pipeline with the following steps (Figure 1):
1.
Extraction of individual frames from the video recording.
2.
Removal of non-relevant frames (e.g., neutral postures, transitions).
3.
Matching of each frame to its corresponding motion data entry.
4.
Calculation of joint angles and ergonomic risk label for each body segment.
5.
Saving each frame as a labeled image, where the filename encodes the frame number and the ergonomic classification for the specific body part.
An illustrative example of the acquired data and the corresponding ergonomic risk labels is shown in Figure 2.

2.2. Data Processing

The dataset used consists of 47,822 samples with an active/inactive ratio of ∼6% for each of the classes present. The labels reported as active will describe the anatomical regions that have an incorrect anatomical conformation, while inactive will code the anatomical regions that have a correct and physiological position. In contrast to the previous work, the pre-processing of the images did not involve the use of MEDIAPIPE (https://github.com/google-ai-edge/mediapipe, accessed on 9 May 2025) to crop the subjects from the global image. This design choice has been made in order to build a computationally cheap pipeline, thus speeding up the inference process. In fact, leaving the image raw (as shown in Figure 3) allows inference to be performed using the frame acquired straight from the camera, which increases the amount of information available from it and complicates the classification process. To validate the classification performance in a more robust and reliable way, a 5-fold cross-validation has been applied, splitting the dataset into 5 different folds and evaluating the average performance on each generated validation set. This approach allows for more stress in the training phases, ensuring greater reliability on the results obtained, displayed as the average performance on each of the folds.

2.3. Network Architecture

The architecture used is a lightweight variant of ViT [54], designed to reduce the number of parameters while keeping classification performance high. This family of architectures, as is well known, are able to extract both local and global information, obtaining contextualized information that overcomes the limitations of CNNs. The latter, prior to the advent of Transformers, dominated the field of computer vision, finding applications in all domains of study, despite limitations related to the inability to abstract local information within complex contexts. For this reason, ViTs with their self-attention mechanisms have been able to improve global classification and segmentation performance by being able to correlate long-distance information within richly detailed images. One of the main characteristics of ViTs is related to the patching mechanism of the input image. In fact, an image whose dimensions are typically represented by H × W × C is divided into square paths of dimension P × P . Each patch is linearized into a vector of dimension P 2 × C and subsequently encoded in a latent space through a Multilayer Perceptron (MLP). The result obtained is a sequence of vectors { x 1 , x 2 , , x n } , where n = H W P 2 . This linearization process leads to a loss of spatial information, which is compensated for by the addition of positional embedding to keep it in memory. Each layer of the ViT typically consists of a Multi-Head Self-Attention (MHSA) module, a feed-forward (MLP) layer, a normalization layer (LayerNorm), and residual blocks to propagate lossless information forward. Classification is achieved through a token [CLS] that will be used as input for the classification block. The general schematization of the operation of a ViT is shown in Figure 4.
In summary, the main difference between CNNs and ViTs lies in the feature extraction mechanism;l the former use locally static convolutional filters, whereas ViTs adopt a global attention mechanism that can dynamically adapt to the relative importance of the various image regions. In the proposed work, a ViT has been implemented that will be described in detail in the next section, which aims to surpass in terms of efficiency and performance classical convolutional architectures such as those used for the same task in the previously published work [52].

Architecture Parameters

To evaluate the impact of hyperparameters on performance, a search grid was used, with the values shown in Table 1. The configuration chosen for the proposed architecture specifically includes the use of a patch_size parameter of 32. This choice allows us to subdivide the starting image into 49 patches of size 32 × 32 , thus providing the right trade-off between local and global information. In fact, given the nature of the image, it is essential to be able to extrapolate the features relating to the subject to be classified. In contrast to the previous approach, as discussed in Section 2.2, the proposed model does not use the cropped images but uses the entire frame. In order to improve the computational efficiency, different patch embedding values have been tested starting from 1024 ( 32 × 32 ), using all the information in the patches. The ideal size was identified as patch_embedding = 192, which allowed the efficient integration of the information present in each patch. To maximize feature extraction capabilities, six transformer blocks were inserted, each with 16 heads for Multi-Head Attention. Finally, self-attention bias has been added to extend the learning capabilities by allowing the architecture to better adapt to the data. Despite the depth, the neural architecture had a small number of parameters (around ∼ 5M) as shown in Figure 5.
In summary, the best architecture employed receives as input images of size 224 × 224 , which will be divided into patches of size 32. This will form the input representation for the 6 consecutive transformer blocks each of which will consist of a Multi-Head Attention with a number of heads equal to 16. A CLS token (Class token) used during the training phases will be added to each image initially to associate the analyzed patches with the reference class. This CLS token will be used in the output to the architecture to return the prediction. A visual representation of the final architecture is shown in Figure 6.

3. Results

The following section shows the results obtained with the architecture proposed in the present paper. The validation of our work has been carried out using the following metrics:
  • Accuracy, i.e., the percentage of correct predictions out of the total number of predictions
    Accuracy = TP + TN TP + TN + FP + FN
    where TP represents the True Positives, TN represents the True Negatives, FP represents the False Positives, and FN represents the False Negatives.
  • Area Under the Curve (AUC) reporting the model’s ability to discriminate between the two classes
    AUC = 0 1 TPR ( FPR 1 ( x ) ) d x
    where TPR is the True Positive Rate and FPR is the False Positive Rate.
  • Precision that considers among all positive predictions those that are truly positive
    Precision = TP TP + FP
  • Recall (Sensitivity) which measures the TP predictions that have been detected, giving a measure descriptive of the discriminative capacity of the model
    Recall = TP TP + FN
  • F1-Score, i.e., the harmonic mean between Precision and Recall, which provides an analytical compromise for highly unbalanced datasets such as the one under study
    F 1 = 2 · Precision · Recall Precision + Recall
In order to facilitate the structure of the results shown, the explicit name of the labelled classes has been coded as shown in Table 2.
Table 3 shows the average results obtained with the 5-fold cross-validation with reference to the metrics described above.
As can be seen, the results obtained are excellent for each classes. These results remain consistent considering the use of K-Fold cross-validation, further demonstrating the reliability of our architecture. Further proof of the outstanding performance is provided by the confusion matrices shown in Figure 7a–h. As can be observed, the confusion matrices show the very high performance of the model on all the classes analyzed. As can be observed, the amount of prediction of FP and FN is negligible or even absent for the classes TRUNK, LLA and NECK. Further evidence is provided by the Precision–Recall curves, as can be seen in Figure 8a–h. These curves show consistency and uniformity between the different classes, with a constant AUC value of 0.9990. This data support the confusion matrix, underlining a very favorable balance between precision and recall. The curves tend to maintain precision close to unity even for very high recall values, showing that the model is able to identify almost all positives without introducing significant false positives.
As can be seen for all classes, the error does not exceed 4 misclassifications, with a peak of 5 misclassifications for the Right Wrist class (RW). This occurrence could be caused by the pose and perspective of the acquired image. In fact, as can be seen in Figure 3, the right wrist is often covered by the equipment that the operator is handling. Despite this additional challenge, the proposed architecture succeeds in extracting the features to correctly discriminate the current pose that the anatomical region of reference is taking in that particular frame thanks to the lightweight ViT’s abilities to extract information. In fact, this architecture, thanks to the global/local mechanisms, manages to integrate the features for the correct class using contextual information, thus making up for the lack of information for hidden classes (e.g., RW).
In addition, a quantitative comparison in terms of Precision–Recall curves (shown in Table 4 has been conducted with the previous SPECTRE approach, highlighting numerically how the use of SPECTRE-ViT improves prediction accuracy while significantly reducing computation time.

4. Discussion

The results obtained in this study confirm the effectiveness of a ViT-based approach for frame-level ergonomic posture classification in an industrial context. The system demonstrated remarkably high performance across all anatomical regions evaluated, achieving an average AUC above 0.996 and F1-scores exceeding 0.99 for most classes, as summarized in Table 3. These findings underscore the robustness and reliability of the proposed architecture, particularly when benchmarked against existing state-of-the-art approaches in ergonomic risk estimation.
The proposed ViT-based architecture also offers clear advantages over existing vision-based posture recognition systems. Conventional approaches often rely on skeleton estimation or intermediate joint-angle computations, typically extracted via tools such as OpenPose or Kinect. These intermediate steps introduce inherent limitations, including error propagation, occlusion sensitivity, and viewpoint dependency. Moreover, many vision-based systems are validated on constrained or small-scale datasets, leading to inflated performance metrics that may not generalize to realistic industrial environments.
In contrast, our approach directly classifies ergonomic posture from raw RGB frames, avoiding the skeleton reconstruction pipeline entirely. This not only streamlines the processing workflow but also results in significant performance improvements. As shown in Table 3 and Table 4, our method outperforms the CNN-based SPECTRE framework across all evaluated classes, with AUC margins reaching up to 12.9% for challenging regions such as the Right Wrist (RW), which is typically subject to frequent occlusions.
Furthermore, while several existing systems require segmented or cropped images to improve performance, our ViT-based model processes full-frame images without any pre-processing. This design decision was motivated by the need to build a low-complexity and deployment-ready system that avoids the dependencies and fragility introduced by external segmentation or keypoint estimation libraries. Cropping and skeleton-based methods often suffer from failure modes in occluded or cluttered scenes, which are common in industrial environments. This design choice enhances deployment feasibility, allowing real-time inference directly on raw camera feeds, and this preserves important contextual cues from the environment. Such robustness is particularly relevant in industrial settings characterized by complex lighting, background clutter, and unpredictable occlusions.
Importantly, the SPECTRE-ViT system was explicitly designed for low-cost, scalable deployment in real-world industrial environments, including small- and medium-sized enterprises. Thanks to its compact architecture (∼5M parameters) and independence from external pre-processing (e.g., skeleton detection or image cropping), the model can perform real-time inference even on edge devices. Crucially, the model has been tested on raw RGB frames without relying on dedicated cameras or depth sensors. This makes it suitable for use with embedded cameras found in off-the-shelf smartphones or tablets. In combination with lightweight edge computing frameworks, SPECTRE-ViT can be deployed directly on mobile devices positioned in the workplace, enabling non-intrusive, cost-effective ergonomic monitoring without cloud connectivity or specialized hardware.
Collectively, these features underscore that our ViT-based system not only surpasses both IMU and prior CNN-based approaches but also advances the state of the art among vision-based ergonomic monitoring tools. It offers a compelling combination of classification accuracy, computational efficiency, and operational simplicity, making it suitable for scalable, real-time deployment in dynamic manufacturing environments. The ViT-based vision system offers a non-invasive, low-cost, and scalable alternative, suited to dynamic environments with varying lighting, perspectives, and occlusion patterns.
Compared to other posture monitoring systems, including our previous SPECTRE framework, the proposed approach yields superior results across all evaluated metrics, especially for anatomically challenging regions such as wrists. Table 5 provides a detailed comparison between the two architectures, highlighting improvements in accuracy, model simplicity, and hardware requirements. In addition, Table 6 positions our method relative to other recent literature in terms of input type, accuracy, and robustness. Most prior approaches rely on either joint estimation pipelines or handcrafted rule-based scoring, both of which introduce fragility under occlusion and impose computational burdens. Our method avoids these steps entirely and achieves higher generalization using a simpler end-to-end strategy.
Table 5 highlights the key improvements of the SPECTRE-ViT system over our previously published SPECTRE architecture. In particular, the ViT-based model achieves significantly better accuracy for difficult joints (e.g., Right Wrist), eliminates the need for image segmentation or skeleton reconstruction and supports real-time inference, even on embedded devices. These enhancements make it more suitable for scalable deployment in real-world industrial workflows. From a broader perspective, this work contributes to addressing the following three persistent challenges in the field: (1) occlusion resilience—mitigated through self-attention and full-frame processing; (2) dataset realism—enabled by using an industrially representative, multi-modal dataset; and (3) outcome linkage—by leveraging direct ergonomic labeling based on ground-truth motion capture data in alignment with the RULA methodology.
Despite the excellent results, some limitations should be acknowledged. First, the dataset is derived from a controlled reproduction of industrial tasks, which, although realistic, may not encompass the full range of variability in real production settings. Field validation in operational environments with real-time constraints and multiple operators is currently in progress and will be a critical next step to confirm the generalizability of the model under uncontrolled lighting, occlusion, and workflow conditions. Second, while the model operates in near real time, latency benchmarks under deployment scenarios are yet to be rigorously quantified. Lastly, expanding the classification taxonomy to cover additional ergonomic indices (e.g., REBA, OWAS) and complex postural interactions (e.g., multi-joint dynamics) could further enhance its applicability.

5. Conclusions

This study has presented a novel and efficient vision-based system for ergonomic risk assessment, built upon a lightweight ViT architecture. The proposed method leverages full-frame RGB images to predict posture-level ergonomic compliance for key anatomical regions without requiring pre-processing steps such as cropping, background removal, or skeleton tracking. The integration of this streamlined pipeline with a high-performing classification backbone resulted in consistently excellent metrics, with average F1-scores and AUC values exceeding 0.99 across all classes.
Compared to conventional CNN-based models and previously published approaches such as the SPECTRE framework, the ViT model proposed here offers notable advantages in terms of computational efficiency, classification accuracy, and adaptability to real-world conditions. This positions the model as a highly promising candidate for non-invasive, real-time ergonomic monitoring in industrial settings, especially in collaborative workspaces where wearable sensors may be intrusive or impractical.
Beyond technical performance, the methodological design prioritizes deployment feasibility. By removing the dependency on skeleton estimation or joint angle computation and reducing reliance on expensive hardware setups or expert calibration, the system becomes inherently more accessible to small and medium-sized enterprises (SMEs). This democratization of ergonomic assessment technology holds the potential to significantly impact workplace health and safety, particularly in sectors characterized by manual labor and high WMSD prevalence.
Moreover, the proposed architecture offers scalability and flexibility. The low parameter count (∼5M) ensures that it can be deployed on edge devices or embedded systems without requiring cloud-based processing. This opens the door to real-time ergonomic feedback loops, which could provide workers and supervisors with continuous, actionable insights during task execution—a key feature for proactive injury prevention and adaptive human–machine collaboration.
Nevertheless, the conclusions drawn must be contextualized within the limitations of the present study. While the dataset was acquired under realistic conditions and using dual-modality ground truth (RGB + IMU), it still represents a controlled laboratory environment. Future work should extend the validation phase to fully operational industrial floors, incorporating variations in lighting, background clutter, occlusion severity, and worker apparel. Furthermore, longitudinal studies assessing the long-term impact of deploying such systems on ergonomic outcomes and productivity are warranted.
From a technical standpoint, future research will explore the integration of temporal attention mechanisms to enable continuous pose tracking and risk accumulation over time. Additionally, expanding the posture taxonomy to include lower-limb assessments and dynamic task phases could further enhance coverage. Efforts to interface with robotic task allocation systems—as hinted by recent studies on ergonomic human–robot collaboration—are also under consideration.
In conclusion, the proposed ViT-based framework represents a significant step forward in the development of scalable, high-performance, and user-friendly ergonomic assessment tools. It successfully bridges the gap between laboratory-grade accuracy and field-level usability, paving the way for more intelligent, responsive, and inclusive safety management systems within Industry 5.0 paradigms. By enabling the early detection of risk factors and facilitating data-driven interventions, such systems contribute not only to occupational health but also to a more sustainable and human-centered model of industrial innovation.

Author Contributions

Conceptualization, M.P., S.C., L.C., and A.P.; methodology, M.P., S.C., L.C., M.C., and A.P.; software, M.P., S.C., and L.C.; validation, M.P., S.C., and L.C.; resources, S.C. and L.C.; data curation, M.C. and A.P.; writing—original draft preparation, M.P., S.C., M.C., A.P., and L.C.; writing—review and editing, M.P., S.C., M.C., A.P., and L.C.; supervision, L.M. and R.P.; project administration, L.M. and R.P.; funding acquisition, L.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work received funding from the project “rIAb: Intelligent system to support rehabilitation therapy”, financial incentives pursuant to the inter-ministerial decree of 6 December 2021 submitted pursuant to the Directorial Decree of 24 June 2022 and approved for funding with MIMIT provision protocol R.0001700.18-10-2024, to which the following code was associated: FTE0000451.

Institutional Review Board Statement

Informed consent was obtained from all subjects involved in the study.

Informed Consent Statement

The participants provided their written informed consent to participate in this study.

Data Availability Statement

Data are available upon reasonable request from the corresponding author.

Conflicts of Interest

Marco Piangerelli was employed by Vici & C S.p.A. The company had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results. The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. European Agency for Safety and Health at Work. Work-Related Musculoskeletal Disorders: Prevalence, Costs and Demographics in the EU; Publications Office of the European Union: Luxembourg, 2019. [Google Scholar]
  2. Parent-Thirion, A.; Biletta, I.; Cabrita, J.; Vargas, O.; Vermeylen, G.; Wilczynska, A.; Wilkens, M. Eurofound (2017), Sixth European Working Conditions Survey-Overview Report (2017 Update); Publications Office of the European Union: Luxembourg, 2017; pp. 1–164. [Google Scholar]
  3. Korhan, O.; Memon, A.A. Introductory chapter: Work-related musculoskeletal disorders. In Work-Related Musculoskeletal Disorders; IntechOpen: Rijeka, Croatia, 2019. [Google Scholar] [CrossRef]
  4. Copsey, S.; Irastorza, X.; Schneider, E. Work-Related Musculoskeletal Disorders in the EU—Facts and Figures; Publications Office of the European Union: Luxembourg, 2010; pp. 29–49. [Google Scholar]
  5. De Kok, J.; Vroonhof, P.; Snijders, J.; Roullis, G.; Clarke, M.; Peereboom, K.; van Dorst, P.; Isusi, I. Work-Related Musculoskeletal Disorders: Prevalence, Costs and Demographics in the EU; Publications Office of the European Union: Luxembourg, 2019; p. 215. [Google Scholar] [CrossRef]
  6. Malchaire, J.; Gauthy, R.; Piette, A.; Strambi, F. A Classification of Methods for Assessing and/or Preventing the Risks of Musculoskeletal Disorders; ETUI, European Trade Union Institute: Brussels, Belgium, 2011. [Google Scholar]
  7. Humadi, A.; Nazarahari, M.; Ahmad, R.; Rouhani, H. In-field instrumented ergonomic risk assessment: Inertial measurement units versus Kinect V2. Int. J. Ind. Ergon. 2021, 84, 103147. [Google Scholar] [CrossRef]
  8. Abbasi-Ghiri, A.; Ebrahimkhani, M.; Arjmand, N. Novel force–displacement control passive finite element models of the spine to simulate intact and pathological conditions; comparisons with traditional passive and detailed musculoskeletal models. J. Biomech. 2022, 141, 111173. [Google Scholar] [CrossRef]
  9. Bonakdar, A.; Houshmand, S.; Martinez, K.B.; Golabchi, A.; Tavakoli, M.; Rouhani, H. Fatigue assessment in multi-activity manual handling tasks through joint angle monitoring with wearable sensors. Biomed. Signal Process. Control 2025, 102, 107398. [Google Scholar] [CrossRef]
  10. Salisu, S.; Ruhaiyem, N.I.R.; Eisa, T.A.E.; Nasser, M.; Saeed, F.; Younis, H.A. Motion Capture Technologies for Ergonomics: A Systematic Literature Review. Diagnostics 2023, 13, 2593. [Google Scholar] [CrossRef] [PubMed]
  11. Battini, D.; Persona, A.; Sgarbossa, F. Innovative real-time system to integrate ergonomic evaluations into warehouse design and management. Comput. Ind. Eng. 2014, 77, 1–10. [Google Scholar] [CrossRef]
  12. Peppoloni, L.; Filippeschi, A.; Ruffaldi, E.; Avizzano, C. A novel wearable system for the online assessment of risk for biomechanical load in repetitive efforts. Int. J. Ind. Ergon. 2016, 52, 1–11. [Google Scholar] [CrossRef]
  13. Merino, G.; da Silva, L.; Mattos, D.; Guimarães, B.; Merino, E. Ergonomic evaluation of the musculoskeletal risks in a banana harvesting activity through qualitative and quantitative measures, with emphasis on motion capture (Xsens) and EMG. Int. J. Ind. Ergon. 2019, 69, 80–89. [Google Scholar] [CrossRef]
  14. Caputo, F.; Greco, A.; D’Amato, E.; Notaro, I.; Spada, S. On the use of Virtual Reality for a human-centered workplace design. Procedia Struct. Integr. 2018, 8, 297–308. [Google Scholar] [CrossRef]
  15. González-Alonso, J.; Simón-Martínez, C.; Antón-Rodríguez, M.; González-Ortega, D.; Díaz-Pernas, F.; Martínez-Zarzuela, M. Development of an end-to-end hardware and software pipeline for affordable and feasible ergonomics assessment in the automotive industry. Saf. Sci. 2024, 173, 106431. [Google Scholar] [CrossRef]
  16. Lim, S.; D’Souza, C. A narrative review on contemporary and emerging uses of inertial sensing in occupational ergonomics. Int. J. Ind. Ergon. 2020, 76, 102937. [Google Scholar] [CrossRef]
  17. Vlasic, D.; Adelsberger, R.; Vannucci, G.; Barnwell, J.; Gross, M.; Matusik, W.; Popović, J. Practical Motion Capture in Everyday Surroundings. ACM Trans. Graph. 2007, 26, 35-es. [Google Scholar] [CrossRef]
  18. Yadav, S.K.; Tiwari, K.; Pandey, H.M.; Akbar, S.A. A review of multimodal human activity recognition with special emphasis on classification, applications, challenges and future directions. Knowl.-Based Syst. 2021, 223, 106970. [Google Scholar] [CrossRef]
  19. Schall, M.C., Jr.; Sesek, R.F.; Cavuoto, L.A. Barriers to the Adoption of Wearable Sensors in the Workplace: A Survey of Occupational Safety and Health Professionals. Hum. Factors 2018, 60, 351–362. [Google Scholar] [CrossRef] [PubMed]
  20. Chen, K.Y.; Chou, L.W.; Lee, H.M.; Young, S.T.; Lin, C.H.; Zhou, Y.S.; Tang, S.T.; Lai, Y.H. Human Motion Tracking Using 3D Image Features with a Long Short-Term Memory Mechanism Model—An Example of Forward Reaching. Sensors 2022, 22, 292. [Google Scholar] [CrossRef]
  21. Lind, C.M.; Abtahi, F.; Forsman, M. Wearable Motion Capture Devices for the Prevention of Work-Related Musculoskeletal Disorders in Ergonomics—An Overview of Current Applications, Challenges, and Future Opportunities. Sensors 2023, 23, 4259. [Google Scholar] [CrossRef]
  22. Diego-Mas, J.A.; Poveda-Bautista, R.; Garzon-Leal, D. Using RGB-D sensors and evolutionary algorithms for the optimization of workstation layouts. Appl. Ergon. 2017, 65, 530–540. [Google Scholar] [CrossRef] [PubMed]
  23. Naranjo, J.E.; Mora, C.A.; Bustamante Villagómez, D.F.; Mancheno Falconi, M.G.; Garcia, M.V. Wearable Sensors in Industrial Ergonomics: Enhancing Safety and Productivity in Industry 4.0. Sensors 2025, 25, 1526. [Google Scholar] [CrossRef] [PubMed]
  24. Fernández, M.M.; Álvaro Fernández, J.; Bajo, J.M.; Delrieux, C.A. Ergonomic risk assessment based on computer vision and machine learning. Comput. Ind. Eng. 2020, 149, 106816. [Google Scholar] [CrossRef]
  25. Abobakr, A.; Nahavandi, D.; Hossny, M.; Iskander, J.; Attia, M.; Nahavandi, S.; Smets, M. RGB-D ergonomic assessment system of adopted working postures. Appl. Ergon. 2019, 80, 75–88. [Google Scholar] [CrossRef]
  26. Damle, R.; Gurjar, A.; Joshi, A.; Nagre, K. Human Body Skeleton Detection and Tracking. Int. J. Tech. Res. Appl. 2015, 3, 222–225. [Google Scholar]
  27. Slembrouck, M.; Luong, H.Q.; Gerlo, J.; Schütte, K.; Cauwelaert, D.V.; Clercq, D.D.; Vanwanseele, B.; Veelaert, P.; Philips, W. Multiview 3D Markerless Human Pose Estimation from OpenPose Skeletons. In Advanced Concepts for Intelligent Vision Systems; Springer International Publishing: Cham, Switzerland, 2020; pp. 166–178. [Google Scholar] [CrossRef]
  28. Tu, H.; Wang, C.; Zeng, W. End-to-End Estimation of Multi-Person 3D Poses from Multiple Cameras. arXiv 2020, arXiv:2004.06239. [Google Scholar] [CrossRef]
  29. Kim, W.; Sung, J.; Saakes, D.; Huang, C.; Xiong, S. Ergonomic postural assessment using a new open-source human pose estimation technology (OpenPose). Int. J. Ind. Ergon. 2021, 84, 103164. [Google Scholar] [CrossRef]
  30. Bibi, S.; Anjum, N.; Sher, M. Automated multi-feature human interaction recognition in complex environment. Comput. Ind. 2018, 99, 282–293. [Google Scholar] [CrossRef]
  31. Seo, J.; Lee, S. Automated postural ergonomic risk assessment using vision-based posture classification. Autom. Constr. 2021, 128, 103725. [Google Scholar] [CrossRef]
  32. Xiao, B.; Xiao, H.; Wang, J.; Chen, Y. Vision-based method for tracking workers by integrating deep learning instance segmentation in off-site construction. Autom. Constr. 2022, 136, 104148. [Google Scholar] [CrossRef]
  33. Liu, J.; Wang, Y.; Liu, Y.; Xiang, S.; Pan, C. 3D PostureNet: A unified framework for skeleton-based posture recognition. Pattern Recognit. Lett. 2020, 140, 143–149. [Google Scholar] [CrossRef]
  34. Andrade-Ambriz, Y.A.; Ledesma, S.; Ibarra-Manzano, M.A.; Oros-Flores, M.I.; Almanza-Ojeda, D.L. Human activity recognition using temporal convolutional neural network architecture. Expert Syst. Appl. 2022, 191, 116287. [Google Scholar] [CrossRef]
  35. Zhu, S.; Fang, Z.; Wang, Y.; Yu, J.; Du, J. Multimodal activity recognition with local block CNN and attention-based spatial weighted CNN. J. Vis. Commun. Image Represent. 2019, 60, 38–43. [Google Scholar] [CrossRef]
  36. Nayak, G.K.; Kim, E. Development of a fully automated RULA assessment system based on computer vision. Int. J. Ind. Ergon. 2021, 86, 103218. [Google Scholar] [CrossRef]
  37. Clark, R.A.; Mentiplay, B.F.; Hough, E.; Pua, Y.H. Three-dimensional cameras and skeleton pose tracking for physical function assessment: A review of uses, validity, current developments and Kinect alternatives. Gait Posture 2019, 68, 193–200. [Google Scholar] [CrossRef]
  38. Li, C.; Zhong, Q.; Xie, D.; Pu, S. Skeleton-based action recognition with convolutional neural networks. In Proceedings of the 2017 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Hong Kong, China, 10–14 July 2017; pp. 597–600. [Google Scholar] [CrossRef]
  39. Yoshikawa, Y.; Shishido, H.; Suita, M.; Kameda, Y.; Kitahara, I. Shot detection using skeleton position in badminton videos. In Proceedings of the International Workshop on Advanced Imaging Technology (IWAIT) 2021, Online, 5–6 January 2021. [Google Scholar]
  40. Bonakdar, A.; Riahi, N.; Shakourisalim, M.; Miller, L.; Tavakoli, M.; Rouhani, H.; Golabchi, A. Validation of markerless vision-based motion capture for ergonomics risk assessment. Int. J. Ind. Ergon. 2025, 107, 103734. [Google Scholar] [CrossRef]
  41. Mahjourian, N.; Nguyen, V. Sanitizing Manufacturing Dataset Labels Using Vision-Language Models. arXiv 2025, arXiv:2506.23465. [Google Scholar]
  42. Agostinelli, T.; Generosi, A.; Ceccacci, S.; Mengoni, M. Validation of computer vision-based ergonomic risk assessment tools for real manufacturing environments. Sci. Rep. 2024, 14, 27785. [Google Scholar] [CrossRef] [PubMed]
  43. Muntean, E.; Leba, M. Comparative Analysis of Wearable and Computer Vision-Based Methods for Human Torso Posture Estimation. In Proceedings of the 2024 9th International Conference on Mathematics and Computers in Sciences and Industry (MCSI), Rhodes Island, Greece, 22–24 August 2024; pp. 56–61. [Google Scholar] [CrossRef]
  44. Menanno, M.; Riccio, C.; Benedetto, V.; Gissi, F.; Savino, M.M.; Troiano, L. An Ergonomic Risk Assessment System Based on 3D Human Pose Estimation and Collaborative Robot. Appl. Sci. 2024, 14, 4823. [Google Scholar] [CrossRef]
  45. Hung, J.S.; Liu, P.L.; Chang, C.C. A deep learning-based approach for human posture classification. In Proceedings of the 2020 2nd International Conference on Management Science and Industrial Engineering, Osaka, Japan, 7–9 April 2020; pp. 171–175. [Google Scholar]
  46. Hasib, R.; Khan, K.N.; Yu, M.; Khan, M.S. Vision-based human posture classification and fall detection using convolutional neural network. In Proceedings of the 2021 International Conference on Artificial Intelligence (ICAI), Lucknow, India, 22–23 May 2021; pp. 74–79. [Google Scholar]
  47. Aghamohammadi, A.; Beheshti Shirazi, S.A.; Banihashem, S.Y.; Shishechi, S.; Ranjbarzadeh, R.; Jafarzadeh Ghoushchi, S.; Bendechache, M. A deep learning model for ergonomics risk assessment and sports and health monitoring in self-occluded images. Signal Image Video Process. 2024, 18, 1161–1173. [Google Scholar] [CrossRef]
  48. Çalışkan, A. Detecting human activity types from 3D posture data using deep learning models. Biomed. Signal Process. Control 2023, 81, 104479. [Google Scholar] [CrossRef]
  49. Hossain, M.S.; Azam, S.; Karim, A.; Montaha, S.; Quadir, R.; De Boer, F.; Altaf-Ul-Amin, M. Ergonomic Risk Prediction for Awkward Postures From 3D Keypoints Using Deep Learning. IEEE Access 2023, 11, 114497–114508. [Google Scholar] [CrossRef]
  50. Kwon, Y.J.; Kim, D.H.; Son, B.C.; Choi, K.H.; Kwak, S.; Kim, T. A work-related musculoskeletal disorders (WMSDs) risk-assessment system using a single-view pose estimation model. Int. J. Environ. Res. Public Health 2022, 19, 9803. [Google Scholar] [CrossRef]
  51. Jung, S.; Kim, B.; Kim, Y.J.; Lee, E.S.; Kang, D.; Kim, Y. Prediction of work-relatedness of shoulder musculoskeletal disorders as by using machine learning. Saf. Health Work 2025, 16, 113–121. [Google Scholar] [CrossRef]
  52. Ciccarelli, M.; Corradini, F.; Germani, M.; Menchi, G.; Mostarda, L.; Papetti, A.; Piangerelli, M. SPECTRE: A deep learning network for posture recognition in manufacturing. J. Intell. Manuf. 2022, 34, 3469–3481. [Google Scholar] [CrossRef]
  53. McAtamney, L.; Nigel Corlett, E. RULA: A survey method for the investigation of work-related upper limb disorders. Appl. Ergon. 1993, 24, 91–99. [Google Scholar] [CrossRef] [PubMed]
  54. Alrfou, K.; Kordijazi, A.; Zhao, T. Computer Vision Methods for the Microstructural Analysis of Materials: The State-of-the-art and Future Perspectives. arXiv 2022, arXiv:2208.04149. [Google Scholar]
Figure 1. Dataset generation pipeline.
Figure 1. Dataset generation pipeline.
Sensors 25 04750 g001
Figure 2. Example frames extracted from the acquired dataset. The labels indicate the body segments identified as risky in each posture.
Figure 2. Example frames extracted from the acquired dataset. The labels indicate the body segments identified as risky in each posture.
Sensors 25 04750 g002
Figure 3. Example image from the dataset used during training and validation phases (Adapted from [52]). The features extracted from this image are used for human pose classification in real-world factory environments.
Figure 3. Example image from the dataset used during training and validation phases (Adapted from [52]). The features extracted from this image are used for human pose classification in real-world factory environments.
Sensors 25 04750 g003
Figure 4. The ViT architecture employed in the pose classification pipeline. The input image is split into patches, which are embedded and passed through a series of transformation layers to produce a classification output.
Figure 4. The ViT architecture employed in the pose classification pipeline. The input image is split into patches, which are embedded and passed through a series of transformation layers to produce a classification output.
Sensors 25 04750 g004
Figure 5. Schematic summary of the architecture used. This representation shows all the layers and trainable parameters for each of them.
Figure 5. Schematic summary of the architecture used. This representation shows all the layers and trainable parameters for each of them.
Sensors 25 04750 g005
Figure 6. Visual representation of the ViT architecture proposed and employed in this work. The first part of the architecture shows the mechanism for generating patches and positional embeddings that will become inputs for the first transformer encoder. On the right side of the architecture, we see the CLS token that will lead to the prediction for each of the classes.
Figure 6. Visual representation of the ViT architecture proposed and employed in this work. The first part of the architecture shows the mechanism for generating patches and positional embeddings that will become inputs for the first transformer encoder. On the right side of the architecture, we see the CLS token that will lead to the prediction for each of the classes.
Sensors 25 04750 g006
Figure 7. (ah) Confusion Matrices for SPECTRE-ViT.
Figure 7. (ah) Confusion Matrices for SPECTRE-ViT.
Sensors 25 04750 g007
Figure 8. (ah) PR Curves for SPECTRE-ViT.
Figure 8. (ah) PR Curves for SPECTRE-ViT.
Sensors 25 04750 g008
Table 1. Hyperparameter grid used for training the Vision Transformer.
Table 1. Hyperparameter grid used for training the Vision Transformer.
HyperparameterTested Values
batch_size8, 16, 32, 64
optimizerAdam, AdamW, Adamax
learning rate (LR) 1 × 10 2 , 1 × 10 3 , 1 × 10 4
dropout0.1, 0.3, 0.5
Patch size8, 16, 32, 64
Embedding dim.192, 256, 512, 1024
Depth1, 2, 4, 6
qkv_biasTrue or False
Table 2. Encoding of labels in results.
Table 2. Encoding of labels in results.
LabelsCoded Labels
NeckNECK
TrunkTRUNK
Right Upper ArmRUA
Right Lower ArmRLA
Right WristRW
Left Upper ArmLUA
Left Lower ArmLLA
Left WristLW
Table 3. Average results for the 5 folds for each class.
Table 3. Average results for the 5 folds for each class.
LabelsAUCAccuracyF1PrecisionRecall
NECK0.99980.99610.99710.99710.9971
TRUNK0.99990.99650.99740.99700.9977
RUA0.99960.99280.99470.99620.9933
RLA0.99860.98940.99300.99510.9909
RW0.99620.98780.99270.98910.9963
LUA0.99950.99440.99420.99470.9938
LLA0.99990.99440.99640.99450.9983
LW0.99950.99090.99320.99610.9904
Table 4. Results obtained in terms of Precision–Recall (PR-AUC) between the SPECTRE-ViT architecture and the previous SPECTRE architecture.
Table 4. Results obtained in terms of Precision–Recall (PR-AUC) between the SPECTRE-ViT architecture and the previous SPECTRE architecture.
LabelsSPECTRE-ViTSPECTRE [52]
NECK1.00000.9870
TRUNK1.00000.9950
RUA0.99970.9600
RLA0.99900.9230
RW0.99970.8700
LUA0.99980.9980
LLA0.99980.9750
LW0.99860.9700
Best values are highlighted in bold.
Table 5. Comparison between the original SPECTRE framework (2022) and the proposed SPECTRE-ViT model (2025).
Table 5. Comparison between the original SPECTRE framework (2022) and the proposed SPECTRE-ViT model (2025).
SPECTRE [52]SPECTRE-ViT
ArchitectureParallel CNNs (one per body part)Unified Vision Transformer (ViT) with multi-label output
Input TypeSegmented RGB + skeleton overlayRaw RGB frames (full frame)
PreprocessingMediapipe segmentation + joint alignmentNone (no cropping or skeleton extraction)
Number of Models8 CNNs (independent classifiers)Single ViT model (shared encoder)
ExplainabilityLIME applied to each CNNNot applied (future work)
Lowest F1-score (Right Wrist)0.92600.9927
Highest F1-score (Trunk)0.94090.9974
Highest F1-score (Neck)0.93620.9971
Inference ModeOffline onlyNear real-time
Hardware RequirementsHigh (A100 GPU)Edge-capable (GTX-class or mobile)
Table 6. Comparison of SPECTRE-ViT with selected vision-based ergonomic monitoring systems.
Table 6. Comparison of SPECTRE-ViT with selected vision-based ergonomic monitoring systems.
StudyInput TypePipelineEvaluation MetricsLimitations
Nayak & Kim (2021) [36]2D Joint Data (from images)CNN-based joint localization → RULA scoringICC = 0.776–0.867 (reliability vs. experts)No F1 reported; requires side-view image pair; dependent on person detection
Zhu et al. (2019) [35]RGB videoSpatiotemporal CNN with spatial fusion95–97% (action recognition)Not task-agnostic; no ergonomic score estimation
Andrade-Ambriz et al. (2022) [34]RGB videoTemporal CNN trained on short clipsAccuracy = 96.4% (human activity)Not validated for ergonomic classification or real-time deployment
Aghamohammadi et al. (2024) [47]RGB video (4-frame sequence)Multi-route CNN with temporal fusionF1 = 0.88Temporal input required; occlusion robustness limited to short intervals
SPECTRE-ViT (2025)Raw RGB (single frame)Vision Transformer (ViT), direct multi-region classificationF1 > 0.99 (all regions)Latency not evaluated; field tests ongoing
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cruciata, L.; Contino, S.; Ciccarelli, M.; Pirrone, R.; Mostarda, L.; Papetti, A.; Piangerelli, M. Lightweight Vision Transformer for Frame-Level Ergonomic Posture Classification in Industrial Workflows. Sensors 2025, 25, 4750. https://doi.org/10.3390/s25154750

AMA Style

Cruciata L, Contino S, Ciccarelli M, Pirrone R, Mostarda L, Papetti A, Piangerelli M. Lightweight Vision Transformer for Frame-Level Ergonomic Posture Classification in Industrial Workflows. Sensors. 2025; 25(15):4750. https://doi.org/10.3390/s25154750

Chicago/Turabian Style

Cruciata, Luca, Salvatore Contino, Marianna Ciccarelli, Roberto Pirrone, Leonardo Mostarda, Alessandra Papetti, and Marco Piangerelli. 2025. "Lightweight Vision Transformer for Frame-Level Ergonomic Posture Classification in Industrial Workflows" Sensors 25, no. 15: 4750. https://doi.org/10.3390/s25154750

APA Style

Cruciata, L., Contino, S., Ciccarelli, M., Pirrone, R., Mostarda, L., Papetti, A., & Piangerelli, M. (2025). Lightweight Vision Transformer for Frame-Level Ergonomic Posture Classification in Industrial Workflows. Sensors, 25(15), 4750. https://doi.org/10.3390/s25154750

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop