VI-Net: View-Invariant Quality of Human Movement Assessment

We propose a view-invariant method towards the assessment of the quality of human movements which does not rely on skeleton data. Our end-to-end convolutional neural network consists of two stages, where at first a view-invariant trajectory descriptor for each body joint is generated from RGB images, and then the collection of trajectories for all joints are processed by an adapted, pre-trained 2D CNN (e.g. VGG-19 or ResNeXt-50) to learn the relationship amongst the different body parts and deliver a score for the movement quality. We release the only publicly-available, multi-view, non-skeleton, non-mocap, rehabilitation movement dataset (QMAR), and provide results for both cross-subject and cross-view scenarios on this dataset. We show that VI-Net achieves average rank correlation of 0.66 on cross-subject and 0.65 on unseen views when trained on only two views. We also evaluate the proposed method on the single-view rehabilitation dataset KIMORE and obtain 0.66 rank correlation against a baseline of 0.62.


Introduction
Beyond the realms of action detection and recognition, action analysis includes the automatic assessment of the quality of human action or movement, e.g. in sports action analysis [1,2,3,4], skill assessment [5,6], and patient rehabilitation movement analysis [7,8].For example, in the latter application, clinicians observe patients performing specific actions in the clinic, such as walking or sitting-to-standing, to establish an objective marker for their level of functional mobility.By automating such mobility disorder assessment using computer vision, health service authorities can decrease costs, reduce hospital visits, and diminish the variability in clinicians' subjective assessment of patients.
Recent RGB based action analysis methods, such as [2,3,6,4], are not able to deal with view-invariance when applied to viewpoints significantly different to their training data.To achieve some degree of invariance, some works such as [9,8,7,10,11,12,13], have made use of 3D human pose obtained from (i) Kinect, (ii) motion capture, or (iii) 3D pose estimation methods.Although the Kinect can provide 3D pose efficiently in optimal conditions, it is dependent on several parameters, including distance and viewing direction between the subject and the sensor.Motion capture systems (mocaps) tend to be highly accurate and view-invariant, but obtaining 3D pose by such means is expensive and time consuming, since it requires specialist hardware, software, and setups.These make mocaps unsuitable for use in unconstrained home or clinical or sports settings.Recently, many deep learning methods, e.g.[14,15,16,17,18], have been proposed to extract 3D human pose from RGB images.Such methods (a) either do not deal with view-invariance and are trained from specific views on their respective datasets (for example, [14,17] show that their methods fail  when they apply them on poses and view angles which are different from their training sets), (b) or if they handle view-invariance, such as [19,20], then they need multiple views for training.
To the best of our knowledge, there is no existing RGB-based, view-invariant method that assesses the quality of human movement.We argue here that using temporal pose information from RGB, can be repurposed, instead of skeleton points, for view-invariant movement quality assessment.In the proposed end-to-end View-Invariant Network (VI-Net in Fig. 1), we stack temporal heatmaps of each body joint (obtained from OpenPose [21]) and feed them into our view-invariant trajectory descriptor module (VTDM).This applies a 2D convolution layer that aggregates spatial poses over time to generate a trajectory descriptor map per body joint, which is then forged to be view-invariant by deploying the Spatial Transformer Network [22].Next, in our movement score module (MSM), these descriptor maps for all body joints are put through an adapted pre-trained 2D convolution model, such as VGG-19 [23] or ResNeXt-50 [24], to learn the relationship amongst the joint trajectories and estimate a score for the movement.Note, OpenPose has been trained on 2D pose datasets which means that our proposed method implicitly benefits from joint labelling.
Initially, we apply our method to a new dataset, called QMAR1 , that includes multiple camera views of subjects performing both normal movements and simulated Parkinsons and Stroke ailments for walking and sit-to-stand actions.We provide cross-subject and cross-view results on this new dataset.Recent works such as [25,26,27,28], provide cross-view results only when their network is trained on multiple views.As recently noted by Varol et al. [29], a highly challenging scenario in view-invariant action recognition would be to obtain cross-view results by training from only one viewpoint.While we present results using a prudent set of two viewpoints only within a multiview training scenario, we also rise to the challenge to provide cross-view results by training solely from a single viewpoint.We also present results on the single view rehabilitation dataset KIMORE [30] which provides 5 different types of lower back exercises in real patients suffering from Parkinsons, Stroke, and back pain.This work makes a number of contributions.We propose the first view-invariant method to assess quality of movement from RGB images and our approach does not require any knowledge about viewpoints or cameras during training or testing.Further, it is based on 2D convolutions only which is computationally cheaper than 3D temporal methods.We also present an RGB, multi-view, rehabilitation movement assessment dataset (QMAR) to both evaluate the performance of the proposed method and provide a benchmark dataset for future view-invariant methods.
The rest of the paper is organized as follows.We review the related works in Section 2 and our QMAR dataset in Section 3.Then, we present our proposed network in Section 4 and experimental results in Section 5. Finally, in Section 6, we conclude our work.

Related Work
Action analysis has picked up relative pace only in recent years with the majority of works covering one of either physical rehabilitation, sport scoring, or skill assessment [13].Here, we first consider example non-skeleton based methods (which are mainly on sport scoring), and then review physical rehabilitation methods as it is the main application focus of our work.Finally, given the lack of existing view-invariant movement analysis techniques, we briefly reflect on related view-invariant action recognition approaches.
Non-Skeleton Movement Analysis -A number of works have focused on scoring sports actions.Pirsiavash et al. [31] proposed an SVM based method, trained on spatio-temporal features of body poses, to assess the quality of diving and figure-skating actions.Although their method estimated action scores better than human non-experts, it was less accurate than human expert judgments.More recently, deep learning methods have been deployed to assess the quality of sport actions in RGB-only data, such as [1,2,32,3,4,33].For example, Li et al. [1] divided a video into several clips to extract their spatio-temporal features by differently weighted C3D networks and then concatenated the features for input to another C3D network to predict action scores.Parmar and Morris presented a new dataset and also used a C3D network to extract features for multi-task learning [3].
The authors in [4,33] propose I3D [34] based methods to analyse human movement.Pan et al. [4] combine I3D features with pose information by building join relation graphs to predict score movement.Tang et al. [33] proposed a novel loss function which addresses the intrinsic score distribution uncertainty of sport actions in the decisions by different judges.The use of 3D convolutions imposes a hefty memory and computational burden, even for a relatively shallow model, which we avoid in our proposed method.Furthermore, the performance of these methods are expected to drop significantly when they are applied on a different viewpoint since they are trained on appearance features which change drastically in varying viewpoints.
Rehabilitation Movement Assessment -Several works have focussed on such movement assessment, e.g.[9,35,36,10,7,8].For example, Crabbe et al. [9] proposed a CNN network to map a depth image to high-level pose in a manifold space made from skeleton data.Then, the high level poses were employed by a statistical model to assess quality of movement for walking on stairs.In [7], Sardari et al. extended the work in [9] by proposing a ResNetbased model to estimate view-invariant high-level pose from RGB images where the high-level pose representation was derived from 3D mocap data using manifold learning.The accuracy of their proposed method was good when training was performed from all views, but dropped significantly on unseen views.Liao et al. [8] proposed an LSTM based method for rehabilitation movement assessment from 3D mocap skeleton data and proposed a performance metric based on Gaussian mixture models to estimate their score.Elkholy et al. [36] extracted spatio-temporal descriptors from 3D Kinect skeleton data to assess the quality of movement for walking on stairs, sit-down, stand-up, and walking actions.They first classified each sequence into normal and abnormal by making a probabilistic model from descriptors derived from normal subjects, and then scored an action by modeling a linear regression on spatio-temporal descriptors of movements with different scores.Khokhlova et al. [10] proposed an LSTM-based method to classify pathological gaits from Kinect skeleton data.They trained several bi-directional LSTMs on different training/validation sets of data.For classification, they computed the weighted mean of the LSTM outputs.All the methods that rely on skeleton data are either unworkable or difficult to apply to in-the-wild scenarios for rehabilitation (or sports or skills) movement analysis.
View-Invariant Action Recognition -As stated in [37,26,29] amongst others, the performance of action recognition methods, such as [38,34,39,40,41] to name a few, drops drastically when they test their models from unseen views, since appearance features change significantly in different viewpoints.To overcome this, some works have dealt with viewpoint variations through skeleton data, e.g.[42,43,37,44].For example, Rahmani et al. [37] train an SVM on view-invariant feature vectors from dense trajectories of multiple views in mocap data via a fully connected neural network.Zhang et al. [44] developed a two-stream method, one LSTM and one convolutional model, where both streams include a view adaptation and a classification network.In each case, the former network was trained to estimate the transformation parameters of 3D skeleton data to a canonical view, and the latter classified the action.Finally, the output of the two streams were fused by weighted averaging of the two classifiers' outputs.
As providing skeleton data is difficult for in-the-wild scenarios, others such as [25,26,27,29,45] have focused on generating view-invariant features from RGB-D data.Li et al. [26] extract unsupervised view-invariant features by designing a recurrent encoder network which estimated 3D flows from RGB-D streams of two different views.In [29], the authors generated synthetic multi-view video sequence from one view, and then trained a 3D ResNet-50 [40] on both synthetic and real data to classify actions.Among these methods, Varol et al. [29] is the only work that provides cross-view evaluation through single view training, resulting in 49.4% accuracy on the UESTC dataset [46], which then was increased to 67.8% when they used additional synthetic multi-view data for training.Next, we first introduce our new RGB multi-view Quality of Movement Assessment for Rehabilitation dataset, QMAR.Then, we give the details of a recently released rehabilitation dataset KIMORE [30], a single-view dataset that includes RGB images and score annotations, making it suitable for single-view evaluation.

QMAR
QMAR was recorded using 6 Primesense cameras with 38 healthy subjects, 8 female and 30 male.Fig. 2 shows the position of the 6 cameras -3 different frontal views and 3 different side views.The subjects were trained by a physiotherapist to perform two different types of movements while simulating two ailments, resulting in four overall possibilities: a return walk to approximately the original position while simulating Parkinsons (W-P), and Stroke (W-S), and standing up and sitting down with Parkinson (SS-P) and Stroke (SS-S).The dataset includes RGB and depth (and skeleton) data, although in this work we only use RGB.As capturing depth data from the 6 Primesense cameras was not possible due to infrared interference, the depth and skeleton data were retained from only view 2 at ≈ 0 • and view 5 at ≈ 90 • .
The movements in QMAR were scored by the severity of the abnormality.The score ranges were 0 to 4 for W-P, 0 to 5 for W-S and SS-S, and 0 to 12 for SS-P.A score of 0 in all cases indicates a normally executed action.Sample frames from QMAR are shown in Fig. 3. Table 1 details the quality score or range and the number of frames and sequences for each action type.Table 2

KIMORE [30]
This is the only RGB single view rehabilitation movement dataset where the quality of movements have been annotated for quantitative scores.KIMORE [30] has 78 subjects (44 healthy, and 34 real patients suffering from Parkinson, Stroke, and back pain) performing five types of rehabilitation exercises (Ex #1 to Ex #5) for lower-back pain.All videos are frontal view -see sample frames in Fig. 4. KIMORE [30] provides two types of scores, P O S and CF S , with values in the range 0 to 50 for each exercise as defined by clinicians.P O S and CF S represent the motion of upper limbs and physical constraints during the exercise respectively.

Proposed Method
Although its appearance changes significantly when we observe an instance of human movement from different viewpoints, the 2D spatio-temporal trajectories generated by body joints in a sequence are affine transformations of each other.For example, see Fig. 5, where the trajectory maps of just the feet joints appear different in orientation, spatial location and scale.Thus, our hypothesis is that by extracting body joint trajectory maps that are translation, rotation, and scale invariant, we should be able to assess the quality of movement from arbitrary viewpoints one may encounter in-the-wild.
The proposed VI-Net network has a view-invariant trajectory descriptor module (VTDM) that feeds into a subsequent movement score module (MSM) as shown in Fig. 1.In VTDM, first a 2D convolution filter is applied on stacked heatmaps of each body joint over the video clip frames to generate a trajectory descriptor map.Then, the Spatial Transformer Network (STN) [22] is applied to the trajectory descriptor to make it view-invariant.The spatio-temporal descriptors from all body joints are then stacked as input into the MSM module, which can be implemented by an adapted, pre-trained CNN to learn the relationship amongst the joint trajectories and provide a score for the overall quality of movement.We illustrate the flexibility of MSM by implementing two different pre-trained networks, VGG-19 and ResNeXt-50, and compare their results.VI-Net is trained in an end-to-end manner.As the quality of movement scores in our QMAR dataset are discrete, we use classification to obtain our predicted score.Table 3 carries further details of our proposed VI-Net.
Generating a Joint Trajectory Descriptor -First, we extract human body joint heatmaps by estimating the probability of each body joint at each image pixel, per frame for a video clip with T frames, by applying OpenPose [21].Even though it may seem that our claim to be an RGB-only method may be undermined by the use of a method which was built by using joint labelling, the fact remains that OpenPose is used in this work as an existing tool, with no further joint labelling or recourse to non-RGB data.Other methods, e.g.[49], which estimate body joint heatmaps from RGB images can equally be used.
To reduce computational complexity, we retain the first 15 joint heatmaps of the BODY-25 version of OpenPose.This is further motivated by the fact, highlighted in [47], that the remaining joints only provide repetitive information.Then, for each body joint j ∈ {1, 2, ..., J = 15}, we stack its heatmaps over the T -frame video clip to get the 3D heatmap J j of size W ×H ×T which then becomes the input to our VTDM module.To obtain a body joint's trajectory descriptor Λ j , the processing in VTDM starts with the application of a convolution filter Φ on J j to aggregate its spatial poses over time, i.e.
where Λ j is of size W × H × 1.We experimented with both 2D and 3D convolutions, and found that a 3 × 3 2D convolution filter yields the best results.
Forging a View-Invariant Trajectory Descriptor -In the next step of the VTDM module, we experimented with STN [22], DCN [50,51], and ETN [52] networks, and found STN [22] the best performing option to forge a viewinvariant trajectory descriptor out of Λ j .
STN can be applied to feature maps of a CNN's layers to render the output translation, rotation, scale, and shear invariant.It is composed of three stages.At first, a CNN-regression network, referred to as the localisation network, is applied to our joint trajectory descriptor Λ j to estimate the parameters for a 2D affine transformation matrix, θ = f loc (Λ j ).Instead of the original CNN in [22], which applied 32 convolution filters followed by two fully connected (FC) layers, we formulate our own localisation network made up of 10 convolution filters followed by two FC layers.The rationale for this is that our trajectory descriptor maps are not as complex as RGB images, and hence fewer filters are sufficient to extract their features.The details of our localisation network's layers are provided in Table 3.Then, in the second stage, to estimate each pixel value of our view-invariant trajectory descriptor Λj , a sampling kernel is applied on specific regions of Λ j , where the centres of these regions are defined on a sampling grid.This sampling grid Γ θ (G) is generated from a general grid G = {(x g i , y g i )}, i ∈ {1, . . ., W ′ ×H ′ } and the predicted transformation parameters, such that where Γ θ (G) = {(x Λj i , y Λj i ), i ∈ {1, . . ., W ′ ×H ′ } are the centers of the regions of Λ j the sampling kernel is applied to, in order to generate the new pixel values of the output feature map Λj .Jaderberg et al. [22] recommend the use of different types of transformations to generate the sampling grid Γ θ (G) based on the problem domain.In VTDM, we use the 2D affine transformations shown in Eq. 2. Finally, the sampler takes both Λ j and Γ θ (G) to generate a view-invariant trajectory descriptor Λj from Λ j at the grid points by bilinear interpolation.Assessing the Quality of Human Movement -In the final part of VI-Net (see Fig. 1), the collection of view-invariant trajectory descriptors Λj for joints j ∈ {1, 2, ..., J = 15}, are stacked into a global descriptor Λ and passed through a pre-trained network in the MSM module to assess the quality of movement of the joints.VGG-19 and ResNeXt-50 were chosen for their state-of-the-art performances, popularity, and availability.For VGG-19, its first layer was replaced with a new 2D convolutional layer, including 3 × 3 convolution filters with channel size J (instead of 3 used for RGB input images), and for ResNeXt-50 its first layer was replaced with 7 × 7 convolution filters with channel size J.The last FC layer in each case was modified to allow movement quality scoring through classification where each score is considered as a class, i.e. for a movement with maximum score S, where S=4 for W-P, S=5 for W-S and SS-S, and S=12 for SS-P movements.The last FC layer of VI-Net has S + 1 output units.
Although VGG-19/ResNeXt-50 were trained on RGB images, we still benefit from their pretrained weights, since our new first layers were initialised with their original first layer weights.The output of this modified layer has the same size as the output of the layer it replaces (Table 3), so the new layer is compatible with the rest of network.In addition, we normalize the pixel values of the trajectory heatmaps to be between 0 and 255, i.e. the same as RGB images on which VGG and ResNeXt were trained on, and trajectory descriptor maps have shape and intensity variations -thus the features extracted from them would be as equally valid as for natural images on which VGG and ResNeXt operate.

Experiments and Results
We first report on two sets of experiments on QMAR to evaluate the performance of VI-Net to assess quality of movement, based around cross-subject and cross-view scenarios.Then, to show the efficiency of VI-Net on other datasets and movement types, we present its results also on the single view KIMORE dataset.We used Pytorch on two GeForce GTX 750 GPUs.All networks were trained for 20 epochs using stochastic gradient descent optimization with initial learning rate of 0.001, and batch size 5.To evaluate the performance of the proposed method, we used Spearman's rank correlation as used in [4,3,1].
Dataset Imbalance -It can be seen from Tables 1 and 2 that the number of sequences for score 0 (normal) is many more than the number of sequences for other individual scores, so we randomly selected 15 normal sequences for W-P, W-S, SS-S movements and 4 normal sequences for SS-P to mix with abnormal movements to perform all our experiments.To further address the imbalance, we applied offline temporal cropping to add new sequences.
Network Training and Testing -In both the training and testing phases, video sequences were divided into 16-frame clips (without overlaps).In training, the clips were selected randomly from amongst all video sequences of the training set, and passed to VI-Net.Then, the weights were updated following a cross-entropy loss, where f (.) is the S + 1 dimensional output of the last fully connected layer and s is the video clip's ground truth label/score.In testing, every 16-frame clip of a video sequence was passed to VI-Net.After averaging the outputs of the last fully connected layer across each class for all the clips, we then set the score for the whole video sequence as the maximum of the clip scores (see Fig. 6), i.e.
where k ∈ {0, 1, . . .S} and M is the number of clips for a video.
Comparative Evaluation -As we are not aware of any other RGB-based view-invariant method to assess quality of movement, we are unable to compare VI-Net's performance to other methods under a cross-view scenario.However,  for cross-subject and single-view scenarios, we evaluate against (a) a C3D baseline (fashioned after Parmar and Morris [3]) by combining the outputs of the C3D network to score a sequence in the test phase in the same fashion as in VI-Net, and (b) the pre-trained, fine-tuned I3D [34].We also provide an ablation study for all scenarios by removing STN from VI-Net to analyse the impact of this part of the proposed method.

Cross-Subject Quality of Movement Analysis
In this experiment, all available views were used in both training and testing, while the subjects performing the actions were distinct.We applied k-fold cross validation where k is the number of scores for each movement.Table 4 shows that VI-Net outperforms networks based on C3D (after [3]) and I3D [34] for all types of movements, regardless of whether VGG-19 or ResNeXt-50 are used in the MSM module.While I3D results are mostly competitive, C3D performs less well due to its shallower nature, and larger number of parameters, exacerbated by QMAR's relatively small size.We show in Section 5.3 that C3D performs significantly better on a larger dataset.
As ablation analysis, to test the effectiveness of STN, we present VI-Net's results with and without STN in Table 4.It can be observed that the improvements with STN are not necessarily consistent across the actions since when all viewpoints are used in training, the MSM module gets trained on all trajectory orientations such that the effect of STN is often overridden.

Cross-View Quality of Movement Analysis
We evaluate the generalization ability of VI-Net on unseen views by using cross-view scenarios, i.e. distinct training and testing views of the scene, while data from all subjects is utilised.Recent works such as [25,26,27,28], provide cross-view results only when their network is trained on multiple views.As recently noted by Varol et al. [29], a highly challenging scenario in view-invariant action recognition would be to obtain cross-view results by training from only one viewpoint.Therefore, we performed the training and testing for each movement type such that (i) we trained from one view only and tested on all other views (as reasoned in Section 1), and in the next experiment, (ii) we trained on a combination of one frontal view (views 1 to 3) and one side view (views 4 to 6) and tested on all other available views.Since for the latter case there are many combinations, we show results for only selected views: view 2 ≈ 0 • with all side views, and view 5 ≈ 90 • with all frontal views.Since in cross-view analysis all subjects are used in both training and testing, applying the C3D and I3D models would be redundant because they would simply learn the appearance and shape features of our participants in our study and their results would be unreliable.
In QMAR, when observing a movement from the frontal views, there is little or almost no occlusion of relevant body parts.However, when observing from side views, occlusions resulting from missing or noisy joint heatmaps from OpenPose, can occur for a few seconds or less (short-term), or for almost the whole sequence (long-term).Short term occlusions are more likely in walking movements W-P and W-S, while long-term occlusions occur more often in sit-to-stand movements (SS-P and SS-S).
The results of our view-invariancy experiments, using single views only in training, are shown in Table 5.It can be seen that for walking movements W-P and W-S, VI-Net is able to assess the movements from unseen views well, with the best results reaching 0.73 and 0.66 rank correlation respectively (yellow highlights), and only relatively affected by short term occlusions.However, for sit-to-stand movements SS-P and SS-S, the long-term occlusions during these movements affect the integrity of the trajectory descriptors and the performance of VI-Net is not as strong, with the best results reaching 0.52 and 0.56 respectively (orange highlights).Note, for all action types, when VI-Net has STN with adapted ResNeXt, it performs best on average.Table 6 shows the results for each movement type when one side view and one frontal view are combined for training.VI-Net's performance improves compared to the single view experiment above with the best results reaching 0.92 and 0.83 for W-P and W-S movements (green highlights) and 0.61 and 0.64 for SS-P and SS-S movements (purple highlights), because the network is effectively trained with both short-term and long-term occluded trajectory descriptors.These results also show that on average VI-Net performs better with adapted ResNeXt-50 for walking movements (W-P and W-S) and with adapted VGG-19 for sit-to-stand movements (SS-P and SS-S).This is potentially because ResNext-50's variety of filter sizes are better suited to the variation in 3D spatial changes of joint trajectories inherent in walking movements compared to VGG-19's 3 × 3 filters which can tune better to the more spatially restricted sit-to-stand movements.We also note that the fundamental purpose of STN in VI-Net is to ensure efficient cross-view performance is possible when the network is trained from a single view only.It would therefore be expected and plausible that STN's effect would diminish as more views are used since the MSM module gets trained on more trajectory orientations (which we verified experimentally by training with multiple views).

Single-View Quality of Movement Analysis
Next, we provide the results of VI-Net on the single-view KIMORE dataset, to illustrate that it can be applied to such data too.KIMORE provides two types of scores, P O S and CF S (see Subsection 3.2) which have a strong correspondence to each other, such that if one is low for a subject, so is the other.Hence, we trained the network based on a single, summed measure to predict a final score ranging between 0 and 100 for each action type.We include 70% of the subjects for training and retain the remaining 30% for testing ensuring each set contains a balanced variety of scores from low to high.
Table 7 shows the results of C3D baseline (after [3]), pre-trained, fine-tuned I3D [34] and VI-Net on KIMORE.It can be seen that VI-Net outperforms the other methods for all movement types except for Exercise #3.VI-Net with adapted VGG-19 performs better than with ResNeXt-50 for all movement types.This may be because, similar to sit-to-stand movements in QMAR, where VI-Net performs better with VGG-19, all movements types in KIMORE are also performed at the same location and distance from camera, and thus carry less variation in 3D trajectory space.This shows that our results are consistent in this sense across both datasets.
In addition, although all sequences in both training and testing sets have been captured from the same view, VI-Net's performance on average improves with STN.This can be attributed to STN improving the network generalization on different subjects.Also, unlike in QMAR's cross-subject results where C3D performed poorly, the results on KIMORE for C3D are promising because KIMORE has more data to help the network train more efficiently.

Conclusion
View-invariant human movement analysis from RGB is a significant challenge in action analysis applications, such as sports, skill assessment, and healthcare monitoring.In this paper, we proposed a novel RGB based view-invariant method to assess the quality of human movement which can be trained from a relatively small dataset and without any knowledge about viewpoints used for data capture.We also introduced QMAR, the only multi-view, non-skeleton, nonmocap, rehabilitation movement dataset to evaluate the performance of the proposed method, which may also serve well for comparative analysis for the community.We demonstrated that the proposed method is applicable to crosssubject, cross-view, and single-view movement analysis by achieving average rank correlation 0.66 on cross-subject and 0.65 on unseen views when trained from only two views, and 0.66 on single-view setting.

Figure 1 :
Figure 1: VI-Net has an view-invariant trajectory descriptor module (VTDM) and a movement score module (MSM) where the classifier output corresponds to a quality score.

Figure 4 :
Figure 4: Sample frames of KIMORE for five different exercises.

Figure 5 :
Figure 5: Walking example -all six views, and corresponding trajectory maps for feet.

Table 1 :
details the number of sequences for each score.Details of the movements in the QMAR dataset.

Table 2 :
Details of abnormality score ranges in the QMAR dataset

Table 4 :
Comparative cross-subject results on QMAR.

Table 5 :
Cross-view results for all actions with single-view training.

Table 6 :
Cross-view results for all actions with two-view training.