1. Introduction
Back pain is a widespread and debilitating health condition that affects individuals throughout the world [
1]. According to the World Health Organization, back pain is one of the top five leading causes of disability worldwide [
2]. Around 70–85% of adults experience low back pain (LBP) at some point in their lifetime [
3]; among them, an average of 62% still experience pain one year after initial onset and 60% will suffer from a recurrent episode of pain [
4]. The total costs of LBP in the US exceed
$100 billion per year, including lost wages due to inability to work [
5].
Physical therapy is effective for preventing and treating LBP [
6,
7], but adherence to prescribed exercises is low [
8]. Barriers include the lack of personalization of exercises to a given patient’s needs, reduced engagement due to the gap the patients feel between supervised sessions in clinic and performing exercises at home without support or oversight [
9,
10], and the difficulty in determining whether patients are correctly adhering to the physical therapists’ recommendations at home—the clinician does not know what the patients are doing at home, but also LBP patients do not know whether they are performing exercises correctly due to impaired proprioception [
11,
12,
13].
Traditionally, determining the severity of the LBP and specific movements that exacerbate the LBP, prescribing the recommended exercises, and monitoring the progress throughout physical therapy are all done through clinical evaluations and interventions to enhance strength, stability, and mobility and ultimately mitigate pain and disability [
14,
15,
16]. These evaluations are typically conducted in person, and can be time consuming and difficult to conduct remotely. With the increasing need for remote healthcare solutions, accurately monitoring and assessing movement patterns at home has become essential for personalized evaluation, development of treatment plans, and prescription and monitoring of therapeutic interventions. Remote monitoring that quantifies the diverse activities patients engage in free-living environments can help identify movement patterns linked to LBP [
17] and monitor adherence to the prescribed home exercises.
Advances in sensor and deep learning technologies have provided promising alternatives for the remote monitoring and analysis of movements [
18,
19,
20]. Prior research has leveraged motion capture (MoCap) [
21], surface electromyography (EMG) [
22], and inertial measurement unit (IMU) sensors [
23] to train deep learning models to classify human movement. However, while these models have high-quality data streams, the sensors are expensive and not easily portable for home use and monitoring, making them not accessible to the general population.
1.1. Problem Setting and Challenges
There is a need for low-cost, portable sensors and Artificial Intelligence (AI) tools to capture low back movements in free-living environments. Motion Tape (MT) has emerged as a promising alternative to other sensors [
24]. MT is a flexible, wearable strain sensor fabricated by depositing a conductive nanomaterial network onto a commercial kinesiology tape (k-tape) substrate. The sensor adheres directly to the skin, allowing it to translate the skin strain induced by joint motion and muscle activation into measurable electrical signals. The sensing mechanism of MT is piezoresistive, where deformation of the nanomaterial network produces changes in electrical resistance that are proportional to the applied strain. During use, time-varying resistance signals are continuously recorded and interpreted as skin deformation associated with movement and muscle engagement [
24,
25]. Because MT is lightweight (on the order of grams), it minimally interferes with natural motion and is suited for prolonged wear outside laboratory settings. MT measurements are inherently influenced by factors such as sensor placement and skin adherence.
The task of classifying low back movements is challenging due to the multi-segmental nature of the lumbar spine, involving all three planes of motion. Furthermore, noise is inherent in the MT data; the sensors are limited in quantity, as they are custom fabricated in the research lab for this study; and human subject testing with multiple devices and under the supervision of a physical therapist has complex logistics. Thus, obtaining large datasets for AI is not feasible yet. However, training on small datasets leads to overfitting, while noise can cause high bias in model predictions [
26].
1.2. Our Approach
In this paper, we aim to achieve highly accurate movement classification using MT data. Given a dataset of MT data streams from multiple sensors placed on the low back, the primary purpose of this study is to train a model to distinguish between one of six low back movements, namely standing extension, forward flexion, left/right lateral bending, and seated left/right rotation. Our secondary purpose is to create a machine learning (ML) classification pipeline that addresses the problems that arise from a noisy and limited dataset. This is the first study of deep learning from MT data for low back movements.
We introduce the Motion-Tape Augmentation Inference Model (MT-AIM) a deep learning classification pipeline trained on MT data and leveraging both feature and data augmentation (see
Figure 1). To tackle the problem of noise, MT-AIM calculates the Discrete-Time Fourier Transform (DTFT) of the MT data, as well as leveraging a generative model to predict respective kinematic angles. In particular, we use marker-based MoCap data taken concurrently with MT data to train a model to generate kinematic angles based on the MT data. These features are then concatenated with the data before classification. Furthermore, an additional generative model is trained to learn the underlying distribution of the MT data and generate synthetic samples conditioned on movement type, thereby increasing the effective size of the training dataset.
2. Related Work
We consider related work from both ML and clinical perspectives. From a ML classification perspective, dealing with noisy datasets is a known challenge and several pipelines have been proposed to address it. Classical approaches often rely on preprocessing techniques such as data cleaning, feature selection, and dimensionality reduction to mitigate noise and enhance model performance [
27,
28]. In cases with many features, methods like Principal Component Analysis (PCA) and Independent Component Analysis (ICA) have been used to reduce the impact of noise by identifying and isolating the most informative features [
29].
In small-scale settings, where the available data may be insufficient for training complex models, techniques such as transfer learning and data augmentation have been applied to improve generalization by leveraging pretrained models or generating synthetic data to expand the training set [
30]. Recent advancements have focused on deep generative architectures that are specifically designed to handle noisy time-series data, including Conditional Variational Autoencoders (C-VAE) [
31], Fourier-based diffusion models (Diffusion-TS) [
32], and Generative Adversarial Networks (GAN) such as TimeGAN [
33]. Furthermore, Bayesian approaches, which model uncertainty in the data and the model parameters, have been shown to improve performance when data is limited or noisy [
34]. In addition, techniques like few-shot learning and active learning have gained attention for addressing small datasets by optimizing the use of available data and actively selecting the most informative samples for training [
35].
Many ML solutions exist for synthetic time-series generation, which we can apply to MT data. However, limited research exists related to conditioning the generation upon another time-series, which we apply to kinematic translation. The most similar work is TimeWeaver [
36]. In the paper, Narasimhan et al. use metadata to condition a diffusion model for data imputation. Our problem setting differs in that we are modeling physical relationships over time, rather than correlations with contextual metadata-based features, which motivates our different conditioning strategy.
From a clinical context, marker-based optical MoCap is the prevailing benchmark for movement research, but it demands significant costs and specialized knowledge [
37]. As a result, researchers have explored alternative methods leveraging various sensor data. Classical ML models like support vector machines and gradient-boosted decision trees have been applied to classify arm, head, and neck movements from video and marker-tracking sensors [
38,
39]. More sophisticated architectures, such as Convolutional Neural Network Long Short-Term Memory (CNN-LSTM) models and Transformers, have demonstrated success in classifying fidgety infant movements, squat variations, and foot movements [
40,
41,
42]. In domains more closely related to our focus on lower back kinematics, deep learning has enabled the mapping of markerless RGB-D camera data of flexion and extension to specific body segments for spine motion analysis [
43].
Previous work with MT has also been conducted. In the works of Loh et al. and Huang et al., strain sensor data collected with MT is used as input to convolutional autoencoders for classifying incorrect movements in warfighter marksmanship exercises and golf swing patterns [
44,
45]. Additionally, a three-layer feedforward neural network is constructed to identify subject-specific muscle parameters and predict motion for elbow flexion-extension motion [
46]. However, our low back use case presents new challenges for ML due to the multi-segmental nature of the lumbar spine, multi-planar movements that occur during activities, and substantial variability in skin stretch across different planes of motion. This is the first study that uses ML to classify low back movement based on MT data.
As our research shares similar facets to all aforementioned work, we draw upon the most similar and successful to compare and integrate for the MT-AIM pipeline. In particular, we implement C-VAE, Diffusion-TS, and TimeGAN generative models, and XGBoost, Transformer, and CNN-LSTM classification heads.
3. Materials
3.1. Dataset Description
A total of 10 healthy participants without LBP were recruited to serve as an initial control group to ensure the full range of sensors could be tested and for comparisons in future studies including patients with LBP. Inclusion criteria required participants to be between 18 and 65 years old, able to follow instructions in English, free of LBP in the preceding year, physically capable of walking and bending their backs, and willing to wear standardized testing attire [
24]. The mean age was
,
male and
female; the detailed demographics and anthropometric measurements are presented in [
24].
MT sensors were fabricated using a previously established procedure adapted from [
47]. Regions of k-tape (Rock Tape, Durham, NC, USA) were masked to define the 7.5 × 40 mm
2 sensing area. Graphene nanosheet (GNS) ink was prepared by dispersing graphene within an ethyl cellulose-ethanol solution (solvents from Sigma-Aldrich, St. Louis, MO, USA) deposited onto the k-tape by spray coating. Following the deposition of three spray-coated layers, the resulting GNS sensing element exhibited an initial unstrained resistance of 10 kOhm. Conductive silver paste (Conductor 3 Silver Ink from Voltera, Kitchener, ON, Canada) was applied to each end of the sensing region to form electrodes, then multistrand wires (from Digi-Key, Thief River Falls, MN, USA) were soldered to the electrodes.
Six MTs were arranged in a 3 × 2 matrix lateral to the spine, positioned to align with spinal junctions from T12 to S1 (see
Figure 2). Retroreflective marker-based 3D optical MoCap provided the reference kinematic data. An experienced physical therapist identified key lumbar spine and pelvic landmarks for marker and MT placement for each participant. First, the middle MTs 3 and 4 were placed to cross L2-to-L3 and L3-to-L4 junctions. Above them, the MTs 1 and 2 were placed to cross the T12-to-L1 and L1-to-L2 junctions. Lastly, the bottom MTs 5 and 6 were placed to cross the L4-to-L5 and L5-to-S1 junctions for most participants. MoCap reflective markers were placed on the spinous processes from T12 to L5 and also to the left and right of L1 and L4 to construct a modified version of a validated multi-segmental spine model, defining upper and lower lumbar segments. Additional markers were placed bilaterally on pelvic landmarks to define the pelvis segment [
24].
The movements tested were standing extension, forward flexion, lateral flexion (right and left), and seated rotation (right and left), as seen in
Figure 2. Each participant performed three repetitions per each of the six movements, totaling 180 time-series samples around 6 s each. Measurements across all sensors were recorded approximately every 20 ms.
Figure 3 illustrates an example of MT strains measured using all six sensors and three repetitions for each movement type by the same participant. Resistance
is normalized to the base MT resistance
in an unstretched position before the test:
, where
R is the electrical resistance measured at a point in time.
3.2. Data Processing
Motion capture data was processed using Qualisys Track Manager (version 2024.3, Qualisys North America, Inc., Buffalo Grove, IL, USA) to label marker trajectories and interpolate missing data, then transferred to Visual3D (C-Motion, Inc., Germantown, MD, USA). A multi-segmental spine model was used to calculate lumbar spine kinematic angles using Euler angles between upper lumbar, lower lumbar, and pelvis segments. The resulting angles were exported to MATLAB (version R2023a, MathWorks, Natick, MA, USA) for comparison with MT resistance data.
MT resistance data were read into MATLAB R2023a, converted to decimal, and each trial’s change in resistance was divided by its baseline. The resulting time series were filtered using a Hampel algorithm to discard outliers [
48]; the Hampel filtering used local windows within each sample. Trials with excessive noise, defined as resistance values greater than 10 standard deviations above the mean, were excluded, which resulted in removing less than 2% of the data.
Resistance and kinematic data were temporally aligned using synchronized start times, trimmed to match in duration, and normalized between
and 1 [
24]. The normalization was applied for each movement trial (a trial had three repetitions of the same movement) using that subject’s own min/max rather than global statistics. For the strain data,
corresponds to peak sensor compression and 1 corresponds to peak sensor tension. For kinematics data,
and 1 correspond to peak movement in each direction.
Figure 4 shows an example of processed MT and MoCap kinematics data for one flexion movement. MT sensors show tension during flexion, with the highest values for the lower MT sensors 5 and 6, then 3 and 4. This parallels a larger decrease in lower-to-pelvis and upper-to-lower Euler angles for the x axis, which corresponds to sagittal plane flexion.
4. Methods: MT-AIM
We introduce a deep learning framework designed to assess low back movement with MT sensor data. To address the challenges posed by limited sample size and signal noise within the MT dataset, we leverage conditional generative models to augment both the size and feature set of the data. This leads us to our solution, MT-AIM, which combines these data and feature augmentation techniques to optimally train our classification model. Specifically, the data augmentation model generates synthetic MT samples, which are then added to the original training set. Next, we apply DTFT transformations and generate kinematic joint-angle predictions conditioned on the MT samples, which are then concatenated with their corresponding original samples. The resulting augmented dataset is then used to train the classification models. See
Figure 1 for an overview of our approach. For all the models, hyperparameters were selected via an exhaustive grid search.
4.1. Generative Modeling for Data Augmentation
Classification models trained on datasets with limited sample sizes and noisy feature representations often suffer from overfitting and poor generalization to unseen data. A common and well-supported approach to mitigate this issue is the incorporation of synthetic data during training. To generate additional synthetic MT samples, we implemented and compared three generative models from related works: C-VAE [
49], Diffusion-TS [
32], and TimeGAN [
33]. Each model was conditioned on the type of movement to produce class-relevant MT time-series samples.
For the C-VAE model, the search space included the number of decoder layers and the dimensionality of the hidden state. The optimal configuration consisted of 4 decoder layers and a hidden state dimension of 12.
For the Diffusion-TS model, grid search was performed over the number of sampling steps, hidden state dimensionality, and the number of decoder layers. Given the model’s complexity, the best performance was achieved with 500 sampling steps, a hidden state size of 64, and 12 decoder layers.
For the TimeGAN model, we optimized the number of additional training iterations for the generator module, the latent space dimensionality, and the number of embedding layers. While model performance was relatively stable across parameter settings, we report the results using a 2-to-1 generator–discriminator iteration advantage, a latent dimension of 12, and 3 embedding layers.
We used the best performing movement-conditioned generative model to generate 20 synthetic samples per movement type, resulting in 120 synthetic MT samples in total, which were subsequently used to augment the training dataset for classification.
4.2. Generative Modeling for Feature Augmentation
To further improve the representational capacity of the model and enhance classification performance, we expanded the feature set associated with each sample. Richer feature representations are particularly important when working with noisy and low-dimensional physiological data, where subtle movement patterns may be difficult to capture using raw signals alone. Prior to augmentation, the original features consisted of six post-processed MT strain time series, one per placed MT (see
Figure 2), obtained by normalizing baseline adjusted resistance signals and filtering outliers. As an initial step, we then applied DTFT to each signal, producing 6 additional time-independent frequency-domain features, making it 12 features in total.
To further enrich the feature space with bio-mechanically meaningful information, we modified the previously used conditional generative models—C-VAE, Diffusion-TS, and TimeGAN—to predict joint-angle kinematics of the upper lumbar, lower lumbar, and pelvis segments, conditioned on MT data. Instead of conditioning on a scalar movement type, we need to condition the generative model on the entire time series of MT sensor data. Essentially, the model acts as a “translator” to map the limited MT data to the corresponding joint-angle kinematics.
To realize this, each conditional generative model was augmented with a transformer encoder layer followed by a fully connected projection layer, which encoded the conditional MT data into a representation compatible with the model’s internal embedding space (see
Figure 5). The resulting MT embeddings were summed into the internal embedding space, and the model was trained using a supervised loss computed against ground-truth joint-angle trajectories obtained from motion capture. This extension produced another six additional features, one for each predicted joint-angle, increasing the total feature dimensionality from 12 to 18. Hyperparameters between the MT and kinematics generators were shared.
4.3. Methods for Classification
Given our augmented dataset, now consisting of synthetic MT and expanded features, our goal is to distinguish among six low back movements using MT sensor data: standing extension, forward flexion, left/right lateral bending, and seated left/right rotation. No prior baselines exist for this specific sensor and task. We implemented and compared three standard architectures commonly used in related works for classifying time-series data: XGBoost [
50], Transformer [
51], and CNN-LSTM [
52]. Comparing to IMU- or EMG-based methods would be misleading, as the sensor modalities are fundamentally different.
For the XGBoost classifier, a tree-based model was employed, with grid search conducted over the learning rate and maximum tree depth. The optimal configuration was found to be a learning rate of 0.01 and a maximum depth of 3.
For the Transformer-based classifier, grid search was performed to tune the number of encoder layers and the dimensionality of the latent space. We considered encoder depths ranging from 2 to 5 layers and latent dimensions of 6, 12, or 24. The best-performing configuration included 4 encoder layers and a latent dimension of 12.
For the CNN-LSTM model, we applied grid search to optimize the hidden state dimensionality (32, 64, or 100) and the number of LSTM layers (1, 2, or 3). The optimal configuration was found to include a hidden state size of 100 and 2 LSTM layers.
We first evaluated these baseline models using both cross-sample test splits and leave-one-subject-out (LOSO) validation. As shown in
Table 1, we found no statistically significant difference in model performance between cross-sample and LOSO evaluations across five trials for all three base classifiers: CNN-LSTM, Transformer, XGBoost. If substantial subject-specific leakage existed, we would expect cross-sample validation to significantly outperform LOSO due to the model exploiting subject identity markers. Furthermore,
Table 2 shows the precision and recall for each of the 6 exercise classes for the CNN-LSTM classifier: there is no statistical difference between the two splits for all exercises. Classifying lateral bends and flexion results in the highest precision, whereas rotation and extension have lower precision, with left rotation as the exercise most difficult to classify.
To further investigate if there is subject-specific information in our dataset, we trained a dedicated subject classifier using 5-fold cross-validation, achieving only 11% ± 5.2 accuracy (the chance level is 10% for 10 subjects). This near-random performance confirms that subject-specific signatures are not strongly encoded in the features used for our primary classification task. PCA is also useful in capturing data variance—the PCA visualization from
Figure 6 demonstrates that subject clusters are not well-separated in this reduced feature space. Individual subjects’ samples are distributed throughout the feature space rather than forming distinct, isolated clusters. This indicates that subject variance is not a dominant source of variation in our features. Even sample outliers are not specific to one subject but rather appear scattered across multiple subjects. Our preprocessing pipeline was designed to extract exercise-related patterns while suppressing subject-specific confounds: Hampel filtering using local windows within each sample removes outlier artifacts and eliminates subject-specific noise patterns, and per movement min-max normalization removes subject-specific baseline offsets and amplitude identification markers.
As there is no significant difference between cross-sample and LOSO evaluations in our dataset, we adopted cross-sample splits for subsequent experiments, providing greater flexibility in train and test dataset composition and yielding clearer insights into the data requirements for model learning. We report mean ± std over five trials in each table to demonstrate consistency. Given our dataset of 180 samples and 10 subjects, a subject-split leaves 18 samples per subject, which limits flexibility for experimenting with different fold sizes. The use of stratified sampling allows to better understand how much data the model requires to reach optimal performance. LOSO evaluation will be important in subsequent studies on a larger population, including heterogeneous LBP characteristics.
5. Results and Discussion
To evaluate our pipeline, we must look differently at the conditional generative models and classification models used within. For generative models, we evaluate on typical time-series distance functions: Wasserstein Distance and Frechet Time-Series Distance (FTSD) [
53]. For classification models, we evaluate based on accuracy.
To summarize from the previous sections, our dataset consists of 10 participants, 6 movement types, and 3 repetitions per movement type, resulting in 180 original MT labeled samples. The labels are the specific movements supervised by a physical therapist in the lab. For data augmentation, we additionally generate 20 samples per movement type, resulting in 120 synthetic MT samples, thus bringing the total to 300 MT samples. The original features consist of 6 MT time series, corresponding to the 6 sensors on the low back. For feature augmentation, DTFT produces 6 additional features, and the generated kinematics produce another 6 features, bringing the total to 18 features.
5.1. Generated Motion Tape Evaluation
To evaluate our synthetic MT time-series data, we assess both distributional differences compared to real MT data and overall temporal alignment using Wasserstein and FTSD metrics. Evaluating on both metrics allows for richer insights into the structure and similarity of the sequences.
The Wasserstein Distance, also known as Earth Movers Distance, compares the shapes of two time-series distributions
A and
B, even if they do not align in time. It is commonly referred to as measuring the minimum “cost” of transforming one distribution into the other:
where
is the set of probability spaces whose marginal distributions correspond to
A and
B respectively.
To evaluate our synthetic data considering temporal structure, we choose FTSD. This metric measures the similarity between distributions by taking into account the location and ordering of points at certain steps in time:
where
d is the distance function between two points in the metric space of curves
A and
B, and
and
are all reparameterizations of
A and
B respectively.
To evaluate our generative models, we used all MT data for training, generating 120 new samples after convergence. This was repeated 5 times per model, with randomly initialized weights, to calculate average and standard deviation metrics for Wasserstein Distance and FTSD. As seen in
Table 3, the C-VAE and Diffusion-TS models significantly outperform the TimeGAN model for generating synthetic MT data.
5.2. Generated Kinematics Evaluation
We evaluate the generated kinematic time-series data using the same metrics as our generated MT data. Since the kinematic joint angles are a function of the MT data, it is expected that the distance to ground truth in generated kinematics samples would be smaller compared to our MT evaluations.
Table 3 and
Figure 7 confirm this expectation and also show the superiority of C-VAE and Diffusion-TS for the kinematic translation task. On the other hand, TimeGAN exhibited mode collapse, generating highly similar kinematic outputs despite the variation of input noise or movement.
5.3. Classification Evaluation
To assess the performance of our classification models, we evaluate them using accuracy. Given that our dataset is balanced and the primary objective is to effectively classify lower back movements, accuracy provides the most informative measure of model performance. When evaluating our models, we randomly select stratified MT samples making up 25% of all real data to be part of a test set. The rest of the real data along with synthetic data (depending on the experiment) is used for training. This procedure is conducted 5 times per model, with random weight initialization, where an average accuracy score and standard deviation is recorded.
In
Table 4 and
Figure 8 we compare the accuracy scores of different combinations of synthetic and classification models. We exclude TimeGAN as synthetic MT and kinematic translation modules due to poor performance recorded in
Table 3. As this is the first ML study for low back movement classification with MT data, no prior baselines exist for this specific task and sensor. We use true kinematics as an ablation study to assess model performance with ideal inputs, differentiating classification capability from estimation noise: feature augmentation in
Table 4 represents augmentation with true kinematics and DTFT features, whereas MT-AIM uses data augmentation and generated kinematics. The reason for using real kinematics for Feature Augment is to show the classification performance upper bound using only feature augmentation. It validates that kinematic information is valuable and supports our design choice to generate kinematics, showing the maximum possible benefit with just feature augmentation if generation were perfect.
From the visualizations, we can see that both augmenting the data and features significantly impacts the performance of our classification models. Furthermore, the models using the CNN-LSTM classification both reached over 99% average accuracy over the five trials. We also see from
Figure 8 that choosing Diffusion-TS over C-VAE, or vice versa, did not significantly impact performance; rather, the existence of augmentation is all that mattered. In general, more metrics beyond accuracy may provide a more comprehensive evaluation, as class-wise performance may vary. However, in this study we managed to train a model to near perfect accuracy; thus, our results highlight the validity of MT and augmentation approaches.
6. Limitations and Opportunities for Future Work
MT-AIM offers an effective pipeline for classifying lower back movements using MT. The standard approach for biometric monitoring technologies is to first conduct a study on a healthy population and then validate it in a clinical population. This study realizes the first step in a healthy population so that we can ensure full ranges of movement in all movement directions. These data were derived from a preliminary laboratory validation study under the supervision of a physical therapist; the movement types for this study were selected by a physical therapist as clinically relevant for LBP assessment. Future clinical validation studies are needed in larger samples and in people with and without LBP. Future work includes comparing the movements between people with and without LBP, identifying impairments and subtypes of LBP based on MT, and testing functional activities in free-living environments. Our work is promising in providing a framework for classification, with the goal to enable doctors to remotely assess and recommend physical therapy for LBP.
While many aspects need to be solved before this approach can be used in free-living environments for patient monitoring to inform decision making and improve outcomes, this classification pipeline is an important step in that direction. The clinical problem of managing LBP requires: evaluating movements to determine how movement is associated with the LBP problem, estimating the strain on the lower back throughout the day, assessing whether patients are performing exercises correctly, and assessing progression throughout treatment and risk of LBP development or recurrence. A new generation of MT is under development that transmits data wirelessly to a smartphone app, which sends it to the cloud for real-time monitoring. Furthermore, in our experiments, inference speed was quicker than 20 ms which is faster than the frequency of the sensor readings, so the computational cost is negligible for the task of classifying movements. However, experiments involving real-time monitoring and for longer duration will need new models to analyze compound movements with MT and assess progression throughout treatment, whose computational cost will need to be evaluated.
The small sample size limits our ability to conclude the model generalization to a population of people performing lower back movements. External datasets do not exist for MT, as it is a novel sensor. Small sample size is common at this stage of development for new sensors, as the sample size is purposely small because the data gathered is used to help improve sensors for future larger studies. Different people could have different MT readings based on height, weight, and other demographics, so more subjects are needed with a variety of body types. For example, individuals with obesity may have different skin strains, leading to difficulty in interpreting their MT readings and classifying the movements. Furthermore, individuals with LBP may have different movement patterns. Also, the amount of noise and size of dataset remained constant in our research, so we can’t make any conclusions about MT-AIM’s performance on other datasets of varying size and fidelity.
Variability in the GNS coverage on different MT sensors may cause inconsistent sensitivity, which can lead to data samples that are hard to classify. We are exploring new materials for fabricating MT using Multiwalled Carbon Nanotubes, which are promising in being less noisy than the GNS used in this study. However, there is a tradeoff between sensitivity and signal-to-noise ratio; thus, some noise is to be expected. The advantage of our approach is that even when the sensing technology is noisy, the framework still reaches very high accuracy.
Furthermore, MT baseline resistance values can drift over time. This can cause classification issues if the movement profiles drift. Since measurements were taken in a controlled setting in the lab, sensor drift was not a concern in our study. Each measurement session lasted about an hour per participant, including lab setup and interviews. We also normalized the sensors to baseline before each movement type. There is promising research to mitigate the sensor drift in the classification pipeline during prolonged use in free-living environments.
Because MT operates in response to local skin strain, variations in sensor placement can influence both signal magnitude and temporal characteristics. Changes in placement relative to underlying muscle groups may alter the local strain field experienced by the sensor, leading to differences in sensitivity to specific movements. This makes placement critical, as only the targeted muscle group or bony segment should be covered by the MT sensing region. In the event of misplacement, undesirable perturbations may arise from adjacent or unintended muscle groups within the same region. Similarly, orientation affects how skin deformation is transferred into the conductive network, which can impact repeatability across sessions or subjects.
It could be challenging for non-experts to reliably identify anatomical landmarks, but clinicians are expertly trained to do this. Therefore, in this study the sensors were placed on all participants by a physical therapist. Future studies that need non-expert application could create a template that is based on easily identified landmarks—like the PSIS (Posterior Superior Iliac Spine) visible as dimples on the back of the pelvis—and then the MTs are laid down according to the template.
The task of generating kinematic joint angles from MT data still has a lot of room for improvement. Physics-informed training seems like a promising avenue. This could lead to 3D modeling of human motion using MT. Another suggestion for future work would be to use MT for form analysis in athletic performance such as weight lifting. It would be interesting to see if the tape is precise enough to recognize small anomalies in human motion.
7. Conclusions
We developed MT-AIM, a deep learning pipeline to classify low back movement based on MT sensors. MT-AIM demonstrates the potential of using low-cost, portable wearable sensors, like MT, for classifying human motion, while also providing a viable solution for remote monitoring of lower back strain in people with LBP. By leveraging synthetic data generation and feature augmentation, MT-AIM addresses the challenges posed by the small scale and noise of MT data, achieving state-of-the-art performance despite these limitations.
The ability to generate synthetic MT data of a desired movement, alongside predictions of joint kinematics, significantly enhances the accuracy of our classification models. Moreover, this approach may not only aid in the remote monitoring of low back strain but also open the door to future applications with other wearable sensors.
However, the limited dataset suggests that further validation with larger, more diverse datasets is necessary. As this technology evolves, extending the model to incorporate more nuanced features, such as physics-based regularization terms, could lead to more complex insights on human motion modeling with MT. Future work should focus on validating the system with patients experiencing LBP and exploring real-world deployment scenarios to assess long-term reliability and clinical utility.
Author Contributions
Conceptualization, R.Y., E.F., S.P.G. and K.J.L.; Methodology, J.L., R.Y., E.F., and A.L.; Software, J.L., E.W. and S.P.G.; Validation, S.P.G.; Formal analysis, J.L., R.Y. and E.F.; Investigation, J.L., R.Y., E.F., A.L., S.P.G., E.W. and K.J.L.; Resources, E.F., S.P.G., K.J.L. and R.Y.; Data curation, E.W. and S.P.G.; Writing—original draft, J.L. and R.Y.; Writing—review & editing, J.L., E.F., A.L., E.W., S.P.G. and K.J.L.; Visualization, J.L. and E.W.; Project administration, E.F. and S.P.G.; Funding acquisition, E.F., S.P.G., K.J.L. and R.Y. All authors have read and agreed to the published version of the manuscript.
Funding
This research was supported by the U.S. National Science Foundation (NSF) under grant no. IIS-2205093 (PI: E.F.).
Institutional Review Board Statement
The study was conducted in accordance with the Declaration of Helsinki, and the protocol was approved by the San Diego State University Institutional Review Board (Protocol Number: HS-2022-0269; Approval Date: 25 January 2023).
Informed Consent Statement
Informed consent was obtained from all subjects involved in the study.
Data Availability Statement
The de-identified dataset used in this study is available upon appropriate request to the corresponding author. The data are not publicly available, owing to ethical concerns. The complete source code for MT-AIM is available in a public GitHub repository [
54].
Conflicts of Interest
Co-author Kenneth J. Loh is a co-founder of JAK Labs, Inc., a company that may potentially benefit from the research results. JAK Labs intends to commercialize Motion Tape for different physical exercise and sports markets, among others. The terms of this arrangement have been reviewed and approved by the University of California San Diego in accordance with its conflict-of-interest policies.
Abbreviations
The following abbreviations are used in this manuscript:
| AI | Artificial Intelligence |
| CNN-LSTM | Convolutional Neural Network Long Short-Term Memory |
| C-VAE | Conditional Variational Autoencoder |
| DTFT | Discrete-Time Fourier Transform |
| EMG | Electromyography |
| FTSD | Frechet Time-Series Distance |
| GAN | Generative Adversarial Network |
| GNS | Graphene nanosheet |
| K-tape | Kinesiology tape |
| ICA | Independent Component Analysis |
| IMU | Inertial Measurement Unit |
| LBP | Low back pain |
| LOSO | Leave-One-Subject-Out |
| ML | Machine Learning |
| MoCap | Motion Capture |
| MT | Motion Tape |
| MT-AIM | Motion-Tape Augmentation Inference Model |
| PCA | Principal Component Analysis |
References
- Chen, S.; Chen, M.; Wu, X.; Lin, S.; Tao, C.; Cao, H.; Shao, Z.; Xiao, G. Global, regional and national burden of low back pain 1990–2019: A systematic analysis of the Global Burden of Disease study 2019. J. Orthop. Transl. 2022, 32, 49–58. [Google Scholar] [CrossRef] [PubMed]
- Pan American Health Organization (PAHO). Leading Causes of Death and Disability. 2021. Available online: https://www.paho.org/en/enlace/leading-causes-death-and-disability (accessed on 3 February 2026).
- Andersson, G. Epidemiological features of chronic low-back pain. Lancet 1999, 354, 581–585. [Google Scholar] [CrossRef] [PubMed]
- Hestbaek, L.; Leboeuf-Yde, C.; Manniche, C. Low back pain: What is the long-term course? A review of studies of general patient populations. Eur. Spine J. 2003, 12, 149–165. [Google Scholar] [CrossRef] [PubMed]
- Katz, J.N. Lumbar disc disorders and low-back pain: Socioeconomic factors and consequences. J. Bone Jt. Surg.-Am. Vol. 2006, 88, 21–24. [Google Scholar] [CrossRef]
- Saragiotto, B.; Maher, C.; Yamato, T.; Costa, L.; Menezes Costa, L.; Ostelo, R.; Macedo, L. Motor control exercise for chronic non-specific low-back pain. Cochrane Database Syst. Rev. 2016, 2016, CD012004. [Google Scholar] [CrossRef]
- Steffens, D.; Maher, C.G.; Pereira, L.S.M.; Stevens, M.L.; Oliveira, V.C.; Chapple, M.; Teixeira-Salmela, L.F.; Hancock, M.J. Prevention of Low Back Pain: A Systematic Review and Meta-analysis. JAMA Intern. Med. 2016, 176, 199–208. [Google Scholar] [CrossRef]
- Anar, S. The effectiveness of home-based exercise programs for low back pain patients. J. Phys. Ther. Sci. 2016, 28, 2727–2730. [Google Scholar] [CrossRef]
- Palazzo, C.; Klinger, E.; Dorner, V.; Kadri, A.; Thierry, O.; Boumenir, Y.; Martin, W.; Poiraudeau, S.; Ville, I. Barriers to home-based exercise program adherence with chronic low back pain: Patient expectations regarding new technologies. Ann. Phys. Rehabil. Med. 2016, 59, 107–113. [Google Scholar] [CrossRef]
- Boutevillain, L.; Dupeyron, A.; Rouch, C.; Richard, E.; Coudeyre, E. Facilitators and barriers to physical activity in people with chronic low back pain: A qualitative study. PLoS ONE 2017, 12, e0179826. [Google Scholar] [CrossRef]
- Zheng, Z.; Wang, Y.; Wang, T.; Wu, Y.; Li, Y. A systematic review and meta-analysis on comparative kinematics in the lumbopelvic region in the patients suffering from spinal pain. J. Healthc. Eng. 2022, 2022, 7369242. [Google Scholar] [CrossRef]
- Ghamkhar, L.; Kahlaee, A.H. Pain and pain-related disability associated with proprioceptive impairment in chronic low back pain patients: A systematic review. J. Manip. Physiol. Ther. 2019, 42, 210–217. [Google Scholar] [CrossRef]
- Poesl, M.; Carvalho, G.F.; Adamczyk, W.M.; Schüßler, B.; Richter, M.; Luedtke, K.; Szikszay, T.M. Widespread proprioceptive acuity impairment in chronic back pain: A cross-sectional study. Arch. Phys. Med. Rehabil. 2023, 104, 1439–1446. [Google Scholar] [CrossRef] [PubMed]
- Henchoz, Y.; Kai-Lik So, A. Exercise and nonspecific low back pain: A literature review. Jt. Bone Spine 2008, 75, 533–539. [Google Scholar] [CrossRef] [PubMed]
- Nelson, B.W.; O’Reilly, E.; Miller, M.; Hogan, M.; Wegner, J.A.; Kelly, C. The clinical effects of intensive, specific exercise on chronic low back pain: A controlled study of 895 consecutive patients with 1-year follow up. Orthopedics 1995, 18, 971–981. [Google Scholar] [CrossRef] [PubMed]
- Marich, A.V.; Lanier, V.M.; Salsich, G.B.; Lang, C.E.; Van Dillen, L.R. Immediate effects of a single session of motor skill training on the lumbar movement pattern during a functional activity in people with low back pain: A repeated-measures study. Phys. Ther. 2018, 98, 605–615. [Google Scholar] [CrossRef]
- Silva, F.G.; da Silva, T.M.; Maher, C.; L Ferreira, M.; Costa, L.d.C.M.; Costa, L.O.P. What triggers an episode of acute low back pain? A protocol of a replication case-crossover study. BMJ Open 2021, 11, e040784. [Google Scholar] [CrossRef]
- Alschuler, K.N.; Hoodin, F.; Murphy, S.L.; Geisser, M.E. Ambulatory monitoring as a measure of disability in chronic low back pain populations. Clin. J. Pain 2011, 27, 707–715. [Google Scholar] [CrossRef]
- Schaller, A.; Rudolf, K.; Dejonghe, L.; Grieben, C.; Froboese, I. Influencing factors on the overestimation of self-reported physical activity: A cross-sectional analysis of low back pain patients and healthy controls. BioMed Res. Int. 2016, 2016, 1497213. [Google Scholar] [CrossRef]
- Dutta, A.; Ma, O.; Toledo, M.; Pregonero, A.F.; Ainsworth, B.E.; Buman, M.P.; Bliss, D.W. Identifying free-living physical activities using lab-based models with wearable accelerometers. Sensors 2018, 18, 3893. [Google Scholar] [CrossRef]
- Zhu, X.; Huang, P.Y.; Liang, J.; de Melo, C.M.; Hauptmann, A. STMT: A Spatial-Temporal Mesh Transformer for MoCap-Based Action Recognition. arXiv 2023, arXiv:2303.18177. [Google Scholar] [CrossRef]
- Aarotale, P.N.; Rattani, A. Machine Learning-based sEMG Signal Classification for Hand Gesture Recognition. In Proceedings of the 2024 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Lisbon, Portugal, 3–6 December 2024; pp. 6319–6326. [Google Scholar] [CrossRef]
- de Villa, S.G.; Casillas-Pérez, D.; Jiménez-Martín, A.; García-Domínguez, J.J. Inertial Sensors for Human Motion Analysis: A Comprehensive Review. IEEE Trans. Instrum. Meas. 2023, 72, 4006439. [Google Scholar] [CrossRef]
- Lee, A.; Wyckoff, E.; Farcas, E.; Godino, J.; Patrick, K.; Spiegel, S.; Yu, R.; Kumar, A.; Loh, K.J.; Gombatto, S. Preliminary Validity and Acceptability of Motion Tape for Measuring Low Back Movement: Mixed Methods Study. JMIR Rehabil. Assist. Technol. 2024, 11, e57953. [Google Scholar] [CrossRef] [PubMed]
- Spiegel, S.; Wyckoff, E.; Barolo, J.; Lee, A.; Farcas, E.; Godino, J.; Patrick, K.; Loh, K.J.; Gombatto, S.P. Motion Tape Strain During Trunk Muscle Engagement in Young, Healthy Participants. Sensors 2024, 24, 6933. [Google Scholar] [CrossRef] [PubMed]
- Varoquaux, G. Cross-validation failure: Small sample sizes lead to large error bars. arXiv 2017, arXiv:1706.07581. [Google Scholar] [CrossRef]
- Pau, S.; Perniciano, A.; Pes, B.; Rubattu, D. An Evaluation of Feature Selection Robustness on Class Noisy Data. Information 2023, 14, 438. [Google Scholar] [CrossRef]
- Goel, S.; Rajendran, S.; Gupta, N. CleanML: A Study for Evaluating the Impact of Data Cleaning on ML Classification Tasks. arXiv 2019, arXiv:1904.09483. [Google Scholar]
- Grünauer, A.; Vincze, M. Using Dimension Reduction to Improve the Classification of High-dimensional Data. arXiv 2015, arXiv:1505.06907. [Google Scholar] [CrossRef]
- Chatterjee, S.; Hazra, D.; Byun, Y.C.; Kim, Y.W. Enhancement of Image Classification Using Transfer Learning and GAN-Based Synthetic Data Augmentation. Mathematics 2022, 10, 1541. [Google Scholar] [CrossRef]
- He, H.; Zhang, Q.; Yi, K.; Shi, K.; Niu, Z.; Cao, L. Distributional Drift Adaptation with Temporal Conditional Variational Autoencoder for Multivariate Time Series Forecasting. IEEE Trans. Neural Netw. Learn. Syst. 2024, 36, 7287–7301. [Google Scholar] [CrossRef]
- Yuan, X.; Qiao, Y. Diffusion-TS: Interpretable Diffusion for General Time Series Generation. In Proceedings of the International Conference on Learning Representations (ICLR), Vienna, Austria, 7–11 May 2024. [Google Scholar]
- Yoon, J.; Jarrett, D.; van der Schaar, M. Time-series Generative Adversarial Networks. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Vancouver, BC, Canada, 8–14 December 2019; Volume 32. [Google Scholar]
- van der Heijden, M.; Velikova, M.; Lucas, P.J. Learning Bayesian networks for clinical time series analysis. J. Biomed. Inform. 2014, 48, 94–105. [Google Scholar] [CrossRef]
- Abdali, A.; Gripon, V.; Drumetz, L.; Boguslawski, B. Active Few-Shot Classification: A New Paradigm for Data-Scarce Learning Settings. arXiv 2022, arXiv:2209.11481. [Google Scholar]
- Narasimhan, S.S.; Agarwal, S.; Akcin, O.; Sanghavi, S.; Chinchali, S.P. Time Weaver: A Conditional Time Series Generation Model. In Proceedings of the 41st International Conference on Machine Learning (ICML), Vienna, Austria, 21–27 July 2024; pp. 37293–37320. [Google Scholar]
- Das, K.; de Paula Oliveira, T.; Newell, J. Comparison of markerless and marker-based motion capture systems using 95% functional limits of agreement in a linear mixed-effects modelling framework. Sci. Rep. 2023, 13, 22880. [Google Scholar] [CrossRef] [PubMed]
- Asghar, A.B.; Majeed, M.; Taseer, A.; Khan, M.B.; Naveed, K.; Jaffery, M.H.; Metwally, A.S.M.; Ejsmont, K.; Nejman, M. Classification and monitoring of arm exercises using machine learning and wrist-worn band. Egypt. Inform. J. 2024, 27, 100534. [Google Scholar] [CrossRef]
- Jiang, Y.; Sadeqi, A.; Miller, E.L.; Sonkusale, S. Head motion classification using thread-based sensor and machine learning algorithm. Sci. Rep. 2021, 11, 2646. [Google Scholar] [CrossRef]
- Kulvicius, T.; Zhang, D.; Poustka, L.; Bölte, S.; Jahn, L.; Flügge, S.; Kraft, M.; Zweckstetter, M.; Nielsen-Saines, K.; Wörgötter, F.; et al. Deep learning empowered sensor fusion boosts infant movement classification. Commun. Med. 2025, 5, 16. [Google Scholar] [CrossRef]
- Lee, J.; Joo, H.; Lee, J.; Chee, Y. Automatic Classification of Squat Posture Using Inertial Sensors: Deep Learning Approach. Sensors 2020, 20, 361. [Google Scholar] [CrossRef]
- Fandakli, S.; Okumus, H. Deep learning based ankle–foot movement classification for prosthetic foot. Neural Comput. Appl. 2024, 36, 11397–11407. [Google Scholar] [CrossRef]
- Wenghofer, J.; Beange, K.H.; Ramos, W.C.; Mavor, M.P.; Graham, R.B. Dynamic assessment of spine movement patterns using an RGB-D camera and deep learning. J. Biomech. 2024, 166, 112012. [Google Scholar] [CrossRef]
- Loh, K.; Huang, S.C.; Lin, Y.A. Enhancing Warfighter Marksmanship Performance using Motion Tape Elastic Fabric Sensors. In Proceedings of the Interservice/Industry Training, Simulation, and Education Conference (I/ITSEC), Orlando, FL, USA, 28 November–2 December 2022. [Google Scholar]
- Huang, S.C.; Lin, Y.A.; Pierce, T.; Wyckoff, E.; Loh, K. Measuring the Golf Swing Pattern Using Motion Tape for Feedback and Fault Detection. In Proceedings of the 14th International Workshop on Structural Health Monitoring, Stanford, CA, USA, 12–14 September 2023. [Google Scholar]
- Taneja, K.; He, X.; He, Q.; Zhao, X.; Lin, Y.A.; Loh, K.; Chen, J.S. A Feature-Encoded Physics-Informed Parameter Identification Neural Network for Musculoskeletal Systems. J. Biomech. Eng. 2022, 144, 121006. [Google Scholar] [CrossRef]
- Lin, Y.; Zhao, Y.; Wang, L.; Yeh, Y.P.Y.; Chiang, W.; Loh, K. Graphene K-Tape Meshes for Densely Distributed Human Motion Monitoring. Adv. Mater. Technol 2021, 6, 2000861. [Google Scholar] [CrossRef]
- Hampel, F.R. The influence curve and its role in robust estimation. J. Am. Stat. Assoc. 1974, 69, 383–393. [Google Scholar] [CrossRef]
- Pagnoni, A.; Liu, K.; Li, S. Conditional Variational Autoencoder for Neural Machine Translation. arXiv 2018, arXiv:1812.04405. [Google Scholar] [CrossRef]
- Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; Association for Computing Machinery: New York, NY, USA, 2016; pp. 785–794. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
- Sainath, T.N.; Vinyals, O.; Senior, A.W.; Sak, H. Convolutional, Long Short-Term Memory, Fully Connected Deep Neural Networks. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), South Brisbane, Australia, 19–24 April 2015; pp. 4580–4584. [Google Scholar] [CrossRef]
- Driemel, A.; Krivošija, A.; Sohler, C. Clustering time series under the Fréchet distance. arXiv 2015, arXiv:1512.04349. [Google Scholar] [CrossRef]
- Levy, J. Assessing Low Back Movement with Motion Tape Sensor Data Through Deep Learning. 2025. Available online: https://github.com/Jared-Levy/MS-ADAPT (accessed on 5 December 2025).
Figure 1.
Generative and classification data flow for MT-AIM. Processed MT and MoCap kinematics data are used to train synthetic MT and kinematics generators. The trained models are sampled to create synthetic MT and kinematics data. A DTFT is taken of all MT data. Finally, real, transformed, and synthetic data is all combined before being used to train a classifier.
Figure 1.
Generative and classification data flow for MT-AIM. Processed MT and MoCap kinematics data are used to train synthetic MT and kinematics generators. The trained models are sampled to create synthetic MT and kinematics data. A DTFT is taken of all MT data. Finally, real, transformed, and synthetic data is all combined before being used to train a classifier.
Figure 2.
(Left): The placement of six MTs (black stripes) and retroreflective markers used for 3D optical MoCap (gray circles). (Right): The six movements tested with human subjects: standing extension, standing forward flexion, standing lateral flexion (left and right), and seated rotation (left and right).
Figure 2.
(Left): The placement of six MTs (black stripes) and retroreflective markers used for 3D optical MoCap (gray circles). (Right): The six movements tested with human subjects: standing extension, standing forward flexion, standing lateral flexion (left and right), and seated rotation (left and right).
Figure 3.
Example MT normalized resistance for each movement type, repeated three times.
Figure 3.
Example MT normalized resistance for each movement type, repeated three times.
Figure 4.
(Left): Sample of processed MT data showing individual sensors. (Right): Sample of corresponding MoCap processed kinematics data showing Euler angles between the lower lumbar segment (L4–L5) and pelvis segments, and between the upper lumbar segment (L1–L3) and lower lumbar segment (L4–L5).
Figure 4.
(Left): Sample of processed MT data showing individual sensors. (Right): Sample of corresponding MoCap processed kinematics data showing Euler angles between the lower lumbar segment (L4–L5) and pelvis segments, and between the upper lumbar segment (L1–L3) and lower lumbar segment (L4–L5).
Figure 5.
General data flow and architecture for the kinematics generators. (a) An MT-conditioned Transformer Encoder produces an embedding that is summed with a Gaussian latent representation of the kinematics learned by a feedforward encoder. The combined latent is decoded into interpretable kinematic outputs, and the resulting loss is used for gradient-based optimization. (b) The learned kinematics latent is replaced with randomly sampled Gaussian noise , which is combined with the MT embedding and decoded to generate kinematics.
Figure 5.
General data flow and architecture for the kinematics generators. (a) An MT-conditioned Transformer Encoder produces an embedding that is summed with a Gaussian latent representation of the kinematics learned by a feedforward encoder. The combined latent is decoded into interpretable kinematic outputs, and the resulting loss is used for gradient-based optimization. (b) The learned kinematics latent is replaced with randomly sampled Gaussian noise , which is combined with the MT embedding and decoded to generate kinematics.
Figure 6.
PCA analysis of samples colored by subject.
Figure 6.
PCA analysis of samples colored by subject.
Figure 7.
Examples of generated kinematics compared with ground truth. Each row compares different low back movements and each column compares different model kinematic predictions for that row’s movement. The final column displays the mode collapse exhibited by TimeGAN.
Figure 7.
Examples of generated kinematics compared with ground truth. Each row compares different low back movements and each column compares different model kinematic predictions for that row’s movement. The final column displays the mode collapse exhibited by TimeGAN.
Figure 8.
Accuracy barplots evaluating the effect of different augmentations. Models with +A/+B use C-VAE/Diffusion-TS to generate synthetic MT data. Models with +D/+E use C-VAE/Diffusion-TS to generate synthetic MT data and the same model to generate conditional kinematics data. The dotted line on each plot represents that classifier’s base classification accuracy (no augmentations).
Figure 8.
Accuracy barplots evaluating the effect of different augmentations. Models with +A/+B use C-VAE/Diffusion-TS to generate synthetic MT data. Models with +D/+E use C-VAE/Diffusion-TS to generate synthetic MT data and the same model to generate conditional kinematics data. The dotted line on each plot represents that classifier’s base classification accuracy (no augmentations).
Table 1.
Base classifier accuracies with different types of split using 5-fold cross-validation.
Table 1.
Base classifier accuracies with different types of split using 5-fold cross-validation.
| | CNN-LSTM | Transformer | XGBoost |
|---|
| Subject Split | | | |
| Sample Split | | | |
Table 2.
Precision and recall per class for the base CNN-LSTM classifier with different types of split.
Table 2.
Precision and recall per class for the base CNN-LSTM classifier with different types of split.
| Exercise | Precision | Recall |
|---|
| Sample-Split | Subject-Split | Sample-Split | Subject-Split |
|---|
| Extension | | | | |
| Flexion | | | | |
| Left Lateral Bend | | | | |
| Right Lateral Bend | | | | |
| Left Rotation | | | | |
| Right Rotation | | | | |
| Average | | | | |
Table 3.
Comparison of MT and kinematics generative models. The FTSD and Wasserstein distance metrics measure the similarity between predicted and ground-truth time series (generated MT vs. real MT and generated kinematics vs. real MoCap kinematics). The bold values denote the best model for each scenario. C-VAE and Diffusion-TS models have similar performances, but both significantly outperform TimeGAN models.
Table 3.
Comparison of MT and kinematics generative models. The FTSD and Wasserstein distance metrics measure the similarity between predicted and ground-truth time series (generated MT vs. real MT and generated kinematics vs. real MoCap kinematics). The bold values denote the best model for each scenario. C-VAE and Diffusion-TS models have similar performances, but both significantly outperform TimeGAN models.
| Model | Motion Tape | Kinematics |
|---|
| FTSD | Wasserstein | FTSD | Wasserstein |
|---|
| C-VAE | | | | |
| Diffusion-TS | | | | |
| TimeGAN | | | | |
Table 4.
Classification accuracies of different models for low back movements. Data augmentation is the base classifier with the addition of synthetic MT data. Feature augmentation is the base classifier with the addition of real (rather than predicted) kinematics and DTFT feature augmentations. MT-AIM uses predicted MT and predicted kinematics from our best generators. Only the highest-performing combinations of data and feature augmentation models are shown (TimeGAN is left out). The bold values denote the best model for each scenario.
Table 4.
Classification accuracies of different models for low back movements. Data augmentation is the base classifier with the addition of synthetic MT data. Feature augmentation is the base classifier with the addition of real (rather than predicted) kinematics and DTFT feature augmentations. MT-AIM uses predicted MT and predicted kinematics from our best generators. Only the highest-performing combinations of data and feature augmentation models are shown (TimeGAN is left out). The bold values denote the best model for each scenario.
| Model | Base | Data Augment | Feature Augment | Both (MT-AIM) |
|---|
| XGBoost | | | | |
| Transformer | | | | |
| CNN-LSTM | | | | |
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |