Next Article in Journal
Proposal for the Clustering of Characteristics to Identify Emotions in the Development of a Foreign Language Exam
Next Article in Special Issue
Extended Online DMD and Weighted Modifications for Streaming Data Analysis
Previous Article in Journal
Computational Methods in the Drug Delivery of Carbon Nanocarriers onto Several Compounds in Sarraceniaceae Medicinal Plant as Monkeypox Therapy
Previous Article in Special Issue
A Design Concept of an Intelligent Onboard Computer Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Modeling of Nonlinear Dynamic Processes of Human Movement in Virtual Reality Based on Digital Shadows

Laboratory of VR Simulators, Tambov State Technical University, 392000 Tambov, Russia
*
Author to whom correspondence should be addressed.
Computation 2023, 11(5), 85; https://doi.org/10.3390/computation11050085
Submission received: 24 March 2023 / Revised: 17 April 2023 / Accepted: 21 April 2023 / Published: 23 April 2023
(This article belongs to the Special Issue Mathematical Modeling and Study of Nonlinear Dynamic Processes)

Abstract

:
In virtual reality (VR) systems, a problem is the accurate reproduction of the user’s body in a virtual environment using inverse kinematics because existing motion capture systems have a number of drawbacks, and minimizing the number of key tracking points (KTPs) leads to a large error. To solve this problem, it is proposed to use the concept of a digital shadow and machine learning technologies to optimize the number of KTPs. A technique for movement process data collecting from a virtual avatar is implemented, modeling of nonlinear dynamic processes of human movement based on a digital shadow is carried out, the problem of optimizing the number of KTP is formulated, and an overview of the applied machine learning algorithms and metrics for their evaluation is given. An experiment on a dataset formed from virtual avatar movements shows the following results: three KTPs do not provide sufficient reconstruction accuracy, the choice of five or seven KTPs is optimal; among the algorithms, the most efficient in descending order are AdaBoostRegressor, LinearRegression, and SGDRegressor. During the reconstruction using AdaBoostRegressor, the maximum deviation is not more than 0.25 m, and the average is not more than 0.10 m.

Graphical Abstract

1. Introduction

Virtual reality (VR) systems are actively used to solve various problems, ranging from professional training to psychological rehabilitation and ending with the entertainment industry. At the current stage of VR development, one of the urgent problems is the accurate reproduction of the user’s body in a virtual environment because reliable visualization of the body directly affects the immersion effect [1]. The solution to this problem is inextricably linked with the exact positioning of a person in real space. The basic delivery package of most VR systems includes a head-mounted display (HMD, also called VR-headset) and two controllers for moving and interacting with objects. Actions can be tracked using headset-mounted cameras, using external sensors and base stations (Lighthouse technology) and, finally, using motion capture suits [2].
In the first case, it is possible to use computer vision technologies, but they provide high-quality reproduction only for hands, in particular, palms and fingers; the recognition of the rest of the body is difficult due to the specific camera’s location [3].
On the other hand, the use of Lighthouse allows you to accurately position the headset, controllers, and additional sensors (trackers). The disadvantage of this approach is the inconvenience of the simultaneous use of multiple sensors for positioning all the key points of a person [4].
The third approach is based on the use of expensive—and often sensitive to external electromagnetic interference—suits with many inertial sensors, which significantly limits their use in mass solutions [5]. In addition to complex motion capture suits, inertial measuring units can be used to track movement, but their accuracy due to the continuous accumulation of errors may be insufficient for a number of situations [6].
Thus, there is a significant problem that consists of the imperfection of existing technical means, which do not allow for building an inexpensive, efficient, and accurate system for capturing human movements for subsequent visualization in a virtual space. Within the framework of this study, the problem of modeling nonlinear dynamic processes of human movement in virtual reality based on a digital shadow is considered, taking into account the need to maximally simplify the motion capture system, that is, reduce the number of tracked points.
This article has the following structure: The first section discusses the subject area, the main scientific and technical problems, analyzes the existing works, studies the process of human movement and tracking, reviews possible solutions to the inverse kinematics problem using machine learning algorithms and metrics to evaluate their effectiveness, and formulates the purpose of this study. The second section presents the theoretical basis of the study, including an analysis of the data collection process on human movement in virtual reality, modeling nonlinear dynamic processes of human movement based on a digital shadow, formulating the problem of the tracking points number optimization, reviewing the algorithms used for machine learning and metrics to evaluate them. The “Results” section presents the characteristics of the generated dataset of user movements obtained from virtual avatars, a comparison of various machine learning models by objective and subjective metrics. The study ends with a discussion of the results obtained and conclusions.

1.1. Study of the Processes of Human Body Movement and Tracking

VR software developers solve this problem using two main methods [7]: direct and inverse kinematics using various motion capture and tracking systems. Let us consider the differences between these approaches, and also compare the technical means for obtaining initial data on human body kinematics when interacting with virtual reality.

1.1.1. Direct Kinematics

In the case of direct kinematics, a sufficient amount of information about all key tracking points (KTPs) of the user’s body is required because the movement of the skeleton is modeled hierarchically top-down, starting, for example, from the shoulder, further to the elbow, and ending with the hand, that is, the child segments move relative to the parent segments without affecting them. By KTP, we mean such a point in a three-dimensional space that uniquely defines a segment or node of the human body and is necessary for the avatar reconstruction in virtual space. This approach is used in motion capture suits; therefore, it is expensive and requires the collection of a large amount of data on the position of all parts of the human body.

1.1.2. Inverse Kinematics

In contrast to direct kinematics, inverse kinematics is based on the processing of a limited amount of data about KTP. In inverse kinematics, on the contrary, the movement of child segments leads to a change in the position of the parent segments, that is, the algorithm calculates the position and orientation of the shoulder and elbow based on the position and orientation of the hand. The first approach is used in motion capture suits; the second is in conditions of limited information about the KTP, for example, when the developer has access only to the position of the user’s hands, according to which they restore the coordinates of the points of the entire arm and even the torso or legs (taking into account the height of the HMD and the position of the hands) [8]. Since the process of human movement is nonlinear and dynamic [9], individual parts of the body move independently of each other, which makes the task of modeling inverse kinematics non-trivial.

1.1.3. Comparison of Approaches to Human Movements Tracking

Direct or inverse kinematics is based on obtaining information about the human body or its individual parts’ position. Analyzing various approaches to collecting data on human movement, the following areas can be distinguished:
-
Tracking the trajectories of human movements using a motion capture suit (for example, Perception Neuron—a system based on inertial sensors [10]), which allows for tracking a change in the position of 59 segments of the human body relative to the base (reference) point. The disadvantage of this approach is the lack of information about the absolute person’s position value in a three-dimensional space and the high probability of data distortion when located close to a source of electromagnetic interference;
-
High-precision tracking of a person’s key points using trackers and virtual reality controllers with low measurement error in the visibility zone of base stations (up to 9 × 9 m) [11]. In the case of leaving the visibility zone of base stations or their overlapping, signal loss from sensors is possible, which leads to incorrect data; the size and weight of trackers limit the possible number of sensors attached to a person;
-
The application of computer vision technologies based on a single camera, stereo cameras, or a system of several synchronized cameras to obtain corrected and more accurate data about a person’s position in a three-dimensional space by recognizing key points of the human body, including fingers and face [12]. When using this technology, there are problems with recognizing key fragments of a person’s silhouette when the object is moving quickly or in low-light conditions.
Thus, each of the considered approaches has a number of disadvantages caused by technical limitations. The solution may be a combination of the above approaches within a unified motion capture system, but this will inevitably lead to increasing its cost and the problem of using it for a wide audience [13]. Therefore, it is especially important to implement an approach that will allow for fast, cheap, and high-precision data collection on human movement processes in various scenarios. Without a sufficient amount of information with a high degree of reliability (the data do not contain interference, noise, and distortion), it is impossible to form a digital shadow. The use of high-precision motion capture suits makes it possible to form a digital representation of a person in virtual space; however, this approach is time-consuming and requires the involvement of a large number of participants with different physical parameters. The main disadvantage is the inability to simulate, in the real world, a number of hard-to-reach movement scenarios and unsafe movements.
The solution to this problem is to combine data from two fundamentally different sources: motion capture systems and a virtual avatar from the game scene [14]. Because the procedure for constructing a digital representation of the human body in virtual space is based on a set of points in the metric coordinate system, then for the final implementation of the virtual avatar, the data source will not have a fundamental difference. The proposed approach will significantly speed up the data collection process (by speeding up animation and parallelization in a virtual scene).
Given the above, to collect the training dataset, you can use data only from the virtual model. A necessary condition for using the data obtained in this way is the reliable values of all the KTP coordinates in the metric coordinate system. Since KTPs collected in this way will allow a complete reconstruction of a virtual avatar, this approach can completely replace data collection using a large number of trackers, as well as computer vision technology. Knowing all the coordinates of the person’s points, it is possible to determine the angles between the segments, which makes it possible to replace the motion capture suit with the proposed approach. The correctness of data from a virtual avatar is ensured by the fact that they are a combination or modification of information from existing high-precision motion capture systems. A comparison of approaches to data collection is presented in Table 1, which allows us to conclude that KTP sets obtained from a virtual avatar can be used as initial data on a person’s position in space. Such data are abstracted from the source, which will make it possible to implement more universal models and information processing tools on their basis. In addition, this method is more productive, has no restrictions on the number and location of KTP, and is also the most accurate. Because the KTP coordinates are read from a digital object, the accuracy depends on the resolution of the digital representation used (float or double data type). By default, virtual reality game development environments such as Unity3D use a vector of three float variables, the values of which are measured in meters, to store point coordinates. In the C# programming language, the float type provides accuracy up to 9 digits after the decimal point (this provides an error of up to 10−6 mm), when using the double type up to 17 digits after the decimal point (an error of up to 10−13 mm).

1.2. Approaches to Solving the Problem of Inverse Kinematics

Consider the existing research in the field of solving the problem of inverse kinematics using various human body motion capture systems.
In [15], five additional trackers fixed on the back, heels, and elbows of a person were used, which, together with two controllers and an HMD, make up eight key points, and based on their position, the developers solved the problem of inverse kinematics. The resulting system had a fairly low latency and high accuracy according to the subjective assessment of the participants in the control group.
The study in [16] is based on the analysis of human gait and the comparison of head movements in a VR-headset with the phases of the step, which made it possible to use only one point to restore leg movements using various approaches (Threshold, Pearson correlation method, SVM, and BLSTM). The BLSTM neural network provides the smallest error but cannot be used in real time, so the authors chose a method based on the Pearson correlation calculation between the acceleration of the head and legs. This approach is not universal and is limited to restoring points only with uniform walking.
A number of studies are focusing on a more accurate reconstruction of only certain parts of the body, such as the hands and fingers. For this, combinations of virtual reality gloves and additional inertial sensors for each finger [17] or a combination of cameras with a depth sensor and neural networks can be used to more accurately restore three-dimensional coordinates [18]. Neural networks are used to successfully reconstruct a three-dimensional model of the human body from a two-dimensional image, in which the main points of the skeleton are first recognized, and then, they are converted into a three-dimensional representation with the correction of the rotation angles of the segments [18].
Attempts to implement inverse kinematics using machine learning technologies have been carried out for a long time. Early works [19], rather pessimistically, assessed the possibility of solving this problem due to the insufficient performance and accuracy of neural networks at that time. However, in the future, significant progress can be observed in this direction; combinations of various neural networks are used to implement the inverse kinematics of robots [20], predicting body positions taking into account the environment based on a virtual skeleton [21] or a set of sensors [22]. The study in [23] considers the use of neural networks to restore the key points of a person’s silhouette when deleting from 10 to 40% of the data.
The analysis shows that for the approximation of the human movement process and the subsequent solution of the problem of inverse kinematics, one of the promising areas is the use of machine learning methods. This allows the use of a limited number of sensors or simplified motion capture systems, but requires the implementation of virtual avatar reconstruction algorithms. Finding the minimum number of KTPs will reduce the amount of information transmitted, the computational load, and the cost of motion-tracking system implementation. Thus, it is necessary to determine possible ways to solve this problem. The following are identified as the main directions for solving the problem of approximating the human movement process: a digital model, a digital twin, and a digital shadow [24]. Let us consider each of these options.

1.2.1. Digital Model

The most accurate and correct approach is to build a digital model that includes dependencies between all points of the human body, as well as kinematic and dynamic parameters in the body movement process. Thus, a digital model is a mathematical or digital representation of a physical object that functions independently of it but reliably reflects its characteristics and all processes occurring in it.
The development of such a mathematical model that fully describes all the patterns of human limb movement is a non-trivial task. In conditions when not all points of the body can be tracked or in the absence of a number of key points in the process of using inverse kinematics, this task may be unsolvable due to its high complexity. Therefore, the actual direction will be the implementation of numerical methods that provide acceptable accuracy in the reconstruction of the human body, considering a number of assumptions.

1.2.2. Digital Twin

A digital twin of an object is understood as a system consisting of a digital model (a system of mathematical and computer models that describes the structure and functionality of a real object) and bidirectional information links with the object. The founder of the concept of “digital twin”, M. Grieves, defines it as a “set of virtual information constructs that fully describes a potential or actual physical manufactured product from the micro atomic level to the macro geometrical level” [25].
A comprehensive database is needed to form a digital twin. In addition, for a number of areas, including objects with undetermined behavior, the creation of a digital twin is a task of high complexity and laboriousness with the need to verify, validate, and check the completeness and adequacy of the resulting model. For example, creating a digital twin of a person or even an animal remains a difficult task at the current level of technology. When solving the problem of precise positioning and reconstruction of the human body movements, the “digital twin” approach is also difficult to implement and redundant.

1.2.3. Digital Shadow

The concept of digital shadows was proposed as part of Industry 4.0 [26] as a kind of platform that combines information from various sources to enable real-time analysis of an object for decision making. A digital shadow is a mathematical or digital representation of an object with an automated data transfer channel from physical objects that influence the digital object.
Unlike the digital twin, the formation of a digital shadow does not require a comprehensive database, but one limited by the needs of a specific task. The digital shadow, built on the generalization and processing of large amounts of data on the nonlinear dynamic process of human movement, can be used to solve problems of direct and inverse kinematics and restore damaged or missing data. The concept of digital shadows corresponds to the principles of machine learning algorithms operation, which is also based on the collection and generalization of large amounts of information.
The considered digital shadow of the human movement processes should provide a connection between the initial data on the position of the person’s physical body and its reconstructed digital representation. Thus, the digital shadow realizes the complete restoration of the human body in the digital space, without using complex analytical calculations, due to the processing and generalization of data on a set of KTP.
Since for the successful implementation of the digital shadow, it is required to generalize the initial data on the position of the human physical body for subsequent reconstruction, this task is reduced to finding some regression dependence and, therefore, can be approximated using various machine learning algorithms.

1.3. Machine Learning Technologies for Solving the Task

In accordance with the previous analysis, it was found that the task of reconstructing the human body movements under the conditions of a limited number of KTPs can be successfully solved using machine learning methods. Consider the key approaches in this area, as well as the metrics, to assess the quality of their work.

1.3.1. Overview of Machine Learning Algorithms

Regression dependencies can be approximated using various machine learning algorithms that differ in their architecture, scope, and efficiency. Within the framework of this study, based on the analysis of existing works [27,28,29,30], the following algorithms were selected and implemented in common libraries for machine learning and big data processing:
LinearRegression—a linear model that optimizes the coefficients of the regression equation using the least squares method. The number of coefficients depends on the dimension of the input and output variables. It is distinguished by simple implementation and ease of interpretation of the results [31].
SGDRegressor—a linear regression model using stochastic gradient descent and various loss functions. The model shows high efficiency on large amounts of data [32].
Neural networks—nonlinear models with an arbitrary number of hidden layers, the ability to implement complex architectures (with multiple inputs and outputs), and use various nonlinear activation functions and loss functions. Due to the large number of parameters, they allow generalizing and identifying patterns with a sufficient volume of the training sample but require fine-tuning of the architecture and model parameters [33].
KNeighborsRegressor—k-nearest neighbor regression. The target is predicted by local interpolation of the targets associated with the nearest neighbors in the training set. The disadvantage of this method is that for its application, it is necessary to store the entire training set in memory, and performing regression can be computationally expensive because the algorithm analyzes all data points [32].
DecisionTreeRegressor—a model that predicts the value of a target variable by learning simple decision rules inferred from the data features. A tree can be seen as a piecewise constant approximation. The depth of the tree increases until a state of overfitting is observed [34]. A lack of depth will lead to the underfitting of the model.
RandomForestRegressor—a representative of averaging ensemble methods that allows you to combine the predictions of several basic models to improve the generalizing ability. A random forest is a meta estimator that fits a number of classifying decision trees on various sub-samples of the dataset and uses averaging to improve the predictive accuracy and control over-fitting [35].
AdaBoostRegressor—one of the ensemble boosting methods, which begins by fitting a regressor on the original dataset and then fits additional copies of the regressor on the same dataset but where the weights of instances are adjusted according to the error of the current prediction. As such, subsequent regressors focus more on difficult cases [36].
It should be noted that for a number of algorithms (SGDRegressor and AdaBoostRegressor) that do not support multiple outputs, the use of the MultiOutputRegressor class is required, which allows for multi-purpose regression.

1.3.2. Overview of Metrics for Assessing the Quality of Human Movements Reconstruction

To assess the quality of reconstruction of human body movements after applying machine learning methods, both existing common metrics and those adapted to the specifics of the problem being solved were used [37]. The main metrics include the following:
Mean square error (MSE):
M S E = 1 M j = 1 M i = 1 N v r ( v r i j r i j ) 2
Calculation time, defined as the average time of calling the “predict” function of the machine learning algorithm for one instance of the sample, in milliseconds.
T = i = 1 M t i M
We also introduced two additional metrics aimed at estimating the Euclidean distance. This is the average Euclidean distance between the reference and predicted points:
D m e a n = 1 M K j = 1 M i = 1 K ( v x i x i ) 2 + ( v y i y i ) 2 + ( v z i z i ) 2
The maximum Euclidean distance between the reference and predicted points within the entire avatar:
D max = 1 M j = 1 M max ( v x i x i ) 2 + ( v y i y i ) 2 + ( v z i z i ) 2
The total Euclidean distance between the reference and predicted points within the entire avatar:
D s u m = 1 M j = 1 M i = 1 K ( v x i x i ) 2 + ( v y i y i ) 2 + ( v z i z i ) 2
The chosen metrics, thus, allow us to evaluate both the speed of the compared algorithms and their accuracy from different points of view. It should be noted that although such metrics as MSE and D m e a n are common and generally accepted in solving problems of inverse kinematics, they do not consider the context and boundaries of body segments, which can lead to unrealistic solutions that do not correspond to human anatomy. D max makes it possible to give a pessimistic assessment of the methods, but does not reflect the quality of the reconstruction as a whole. Thus, these listed metrics can be used to analyze the solution quality, but do not provide an opportunity for an objective comparison of different approaches. The D s u m metric, on the contrary, reflects the total reconstruction error for all points of the human body, making it possible to identify both the general discrepancy and the deviation in individual KTP. Thus, D s u m will be used as the main criterion for evaluating machine learning algorithms. For its verification, an additional visual analysis of the results of solving the regression problem by a person is proposed.

1.4. Aim of This Study

The aim of this study is to model a nonlinear dynamic process of human movement in virtual reality under conditions of insufficient initial data on the position of key points of the body. The analysis of the subject area revealed that the achievement of the goal will be based on the following provisions:
-
To simplify the motion capture system, it is necessary to use an approach based on inverse kinematics;
-
The development of a mathematical model that fully describes all the patterns of human limb movement is a non-trivial task, as well as the implementation of a digital twin of the body movement process. In conditions when not all points of the body can be tracked or in the absence of a number of key points when using inverse kinematics, this task may be unsolvable due to its high complexity;
-
As the main method for solving the problem, the concept of digital shadows is proposed, which is based on the generalization of a large amount of data on various typical scenarios of human movements;
-
Machine learning algorithms will be used to approximate the regression dependency between a limited number of KTPs and a fully reconstructed human body model.

2. Materials and Methods

The research methodology includes the following steps:
-
Collecting data on KTP from one or more sources—motion capture systems (motion capture suits, VR trackers, IMU sensors, computer vision, and so on), as well as from a person’s virtual avatar;
-
Formalization of nonlinear dynamic motion processes in the form of a certain model adapted for subsequent software implementation of the digital shadow;
-
Optimization of the model to ensure effective operation in conditions of insufficient data on the person’s position;
-
Software implementation of a digital shadow, which allows modeling nonlinear dynamic motion processes in virtual reality under conditions of insufficient information, based on machine learning methods.

2.1. Modeling of Nonlinear Dynamic Processes of Human Movement Based on Digital Shadow

In the general case, the mathematical description of a digital shadow is a multidimensional vector of object/process characteristics at each moment of time. This vector includes both the values of input variables (internal and external environment parameters) and output parameters (internal and external). Then, in order to model nonlinear dynamic processes of human movement, at the first stage, we formalized the main components of the motion capture process and their characteristics.
Because, in the process of functioning, the digital shadow will interact with data from existing motion capture systems, it is necessary to analyze the specifics of their work. Three main directions for obtaining data on the movement process were identified above: a motion capture suit; virtual reality trackers; computer vision system.
When using a motion capture suit, a set is formed from one reference point and a number of segments (bones) located relative to it, the position of which is indicated using tilt angles along three axes. If necessary, the system allows recording, in addition to sensor rotation angle changes, its movement relative to the previous measurement [38]. We formalized the components of this process.
For a motion capture suit, the set of segments (bones) is defined as B = { b i | i = 1 N b } of size N b . For each segment b i , 3 or 6 values are given depending on the working mode:
  b i = b x i , b y i , b z i , b a x i , b a y i , b a z i   or   b i = b a x i , b a y i , b a z i
where b x i , b y i , b z i are the three-dimensional positions of the i -th sensor, relative to previous measurement;
b a x i , b a y i , b a z i are the three-dimensional rotations of the i -th sensor, relative to previous measurement;
Let us denote the reference point on the user’s back as b 0 = b x 0 , b y 0 , b z 0 , b a x 0 , b a y 0 , b a z 0 . A set of segments, B , is defined for each sample, so there is a correspondence between a set of discrete timestamps, T = { t j } , and a set of sensor values, B . Because links are given between the segments, we define the order of their sequence and connection as O B : B × B , which is a partially ordered set.
Next, we consider the formalization of the components of a motion capture system based on virtual reality trackers. Let N t r trackers T R = { t r i | i = 1 N t r } , each of which is described by a tuple:
t r i = t x i , t y i , t z i , t a x i , t a y i , t a z i ,
where t x i , t y i , t z i , t a x i , t a y i , t a z i are the absolute coordinates of i -th tracker along the X , Y , Z axes, as well as the angles of their rotation along these axes, respectively.
Let us assume that, by default, t r 0 is placed on the user’s back, t r 1 and t r 2 are placed on the hands (or are controllers that the user holds in their hands), t r 3 and t r 4 are on the legs. Further location of trackers can be carried out on the basis of an expert approach in key nodes of the human body.
The formalization of a motion capture system based on computer vision, C V , includes a number of additional steps since the initial data for this approach are a frame (or a set of frames) on which it is necessary to recognize a person and their key points [39]. Therefore, at the first stage, it is necessary to carry out the following transformation of frame, f , into a set of points, P = { p i } . Each point is represented as a tuple:
p i = p x i , p y i , p z i ,
where p x i , p y i , p z i are three-dimensional coordinates of point, p i , in the frame, f .
The layout of points along the human body depends on the selected body recognition method, but always includes the reference points of the arms, legs, torso, and head. Most neural network models position points by two axes ( X and Y ) due to the complexity of depth estimation when using a camera. A number of algorithms (for example, MediaPipe) emulate the determination of the p z i coordinate relative to some reference point, but this value is not accurate and does not allow positioning the object in space. To accurately obtain the depth value, computer stereo vision or a system of several cameras can be used [40].
After formalizing the initial data from the considered motion capture systems, let us move on to modeling the digital shadow of the human movement process, based on data collection from a virtual avatar. This model is based on the processing of virtual KTPs (VKTPs), which correspond to coordinates along three axes. In the case of a sufficient number of VKTPs, their position allows restoring the position and rotation angles of the virtual avatar segments in a similar way to the motion capture systems discussed above.
Let a set of VKTPs be given, corresponding to the key areas of the human body and allowing its subsequent reconstruction in the virtual space. Let us denote the total number of VKTPs as K .
Then, the total number of coordinate values can be specified through the V R vector:
V R = { v x 1 , v y 1 , v z 1 , , v x K , v y K , v z K }
The length of the V R vector is equal to N v r = 3 K . Then, the V R vector with the size of N v r uniquely determines the digital representation of a person with the required accuracy (i.e., it allows modeling all the necessary parts of the virtual body and their connections). Let us introduce an additional designation of the i -th element of the V R element of the v r i , where i = 1 N v r .
Along with the virtual representation of a person (their digital shadow), denoted by the vector, there is a real position of the user’s body in space. Let us introduce the following notation:
The vector of user body coordinates values in the real world, R Q :
R Q = { x 1 , y 1 , z 1 , , x Q , y Q , z Q }
where Q is the number of KTPs received from the real world (using any of the existing motion capture systems).
Length of the R Q vector is N r = 3 Q .
Coordinate vector, R = { r i | i = 1 N v r } , the dimensions of which correspond to the dimensions of V R , and the values of r i correspond to the real values of the user’s KTP. This vector is used as a reference for comparison with V R . Thus, the R V R relationship defines the correspondence between KTP and VKTP.
Element of vector, R Q , with i -th index: r i , where i = 1 N r . Values of r i elements can be obtained from any motion capture system that supports coordinates, for example, from tracker position values according to Formula (2).
Thus, when comparing R Q and V R vectors, one of three scenarios can occur:
-
The lengths are the same (when Q = K ), then if the condition of the KTP location relative to the VKTP placement scheme is met, all KTPs are used to build a virtual avatar;
-
If the number of real KTPs is less than the number of VKTPs (for Q < K ), including the situation when part of the KTP is discarded due to incorrect placement, there is a lack of initial data for constructing a virtual avatar, and it is required to implement algorithms for reconstructing the missing VKTPs based on incomplete input data;
-
If the number of real KTPs is greater than the number of VKTPs (when Q > K ), then the KTPs closest to the VKTPs are used to build a virtual avatar, the rest of the KTPs can be discarded.
In the three scenarios under consideration, when solving a specific problem, it is necessary to fix the location and number of KTPs (the set of KTPs is system-dependent and unchanged during operation). The accidental addition or removal of KTP during operation is not considered in this study, although it is a promising task from a practical point of view.
The second scenario is of greatest interest because the reconstruction of nonlinear dynamic processes of human movement based on insufficient data from real KTPs can reduce the complexity and cost of the motion capture system and increase the realism of displaying a user’s digital avatar in conditions of insufficient data. Consider the process of modeling nonlinear dynamic processes of human movement according to the second scenario using a digital shadow, which includes the following stages:
  • Let there be some motion capture system that does not allow for a complete reconstruction of the person’s digital representation (insufficient number of sensors, interference, poor conditions for body recognition), that is, Q < K .
  • The selected motion capture system transfers to the virtual environment a certain set of KTP values in the form of an R Q vector.
  • Each KTP is assigned the VKTP closest to it in position. Denote a subset of Q VKTPa as V R Q . Then, a relation is formed between two vectors: R Q V R Q , V R Q V R .
  • The machine learning algorithm, A , is selected, aimed at solving the regression problem of the following form:
    A : V R Q V R
Thus, a VKTP vector of length Q is input into a machine learning algorithm, and a reconstructed vector, V R , from all K VKTP is expected at the output. If the selected algorithm can reconstruct with sufficient accuracy, then it can be used to model nonlinear dynamic processes of human movement in conditions of insufficient initial data.
5.
After successful training of the algorithm, A , the vector R Q of real KTP is fed to its input, and the reconstruction of the virtual avatar is carried out: A ( R Q ) = V R .
Then, the formalized research problem has the form: it is necessary to find the type and parameters of the machine learning algorithm, A , for which the value of the total Euclidean distance between the predicted points ( V R vector) and real KTP ( R vector) will be minimal:
D s u m ( A ( R Q ) ) min .
where M is the size of the control sample;
v r i j , r i j are coordinate values of the virtual and real points for the j -th sample element.
An important component of solving problem (7) is minimizing the length of the vector R Q of the initial KTP:
Q min
To successfully solve problems (12) and (13), it is necessary to form a training sample of sufficient size and train a set of machine learning algorithms, the input of which will be given various combinations of points.

2.2. Formation of Sets of Key Human Tracking Points

An important component in solving the tasks is to determine the minimum possible length Q KTP, as well as their placement. Obviously, there are some boundary conditions: for small Q , the algorithms may not work, and in the case of Q K , the economic feasibility of problem-solving decreases.
The analysis of virtual human models showed that 18 VKTPs are enough to build them. This corresponds to many commercial solutions, for example, Perception Neuron based on 18 sensors [41]. Thus, when recording data, it is necessary to record N v r = 54 values (3 coordinates per point). The location of these points should correspond to the connection nodes of the main segments of the human body: hand, elbow, shoulder, extreme points of the foot, knee, hip joint, extreme and central points of the back, head. The layout and numerical identifiers of each KTP are shown in Figure 1.
Note that the selected KTPs are system-dependent to one degree or another; this is due to the fact that the human body can change the position of individual parts only within certain limits, and the lengths of segments between points are constant. Individual KTPs can be quite independent of the other points (e.g., 2 from 3, 4, or 5) and move in space without affecting the position of the other KTPs, but such scenarios are specific. In the collected dataset, it is necessary to take into account the typical movements of the human body, which may include the specific scenarios indicated above (when halves or sides of the human body move independently of each other). The presence of a large amount of data on various typical movements will allow building regression between KTP from the input set and the full model of the human body through approximation and generalization. However, a number of exceptional scenarios may not be processed correctly enough, which will require expanding the dataset.
Further, in the process of experimental studies, it will be necessary to determine the location and number of KTPs sufficient for the effective reconstruction of the human body in virtual reality.

3. Results

In accordance with the above methodology for modeling nonlinear dynamic processes of human movement based on a digital shadow, it is necessary to conduct experimental studies, including collecting a dataset of various human movements from a virtual avatar, analyzing it, creating and comparing machine learning models, and generalizing the obtained experimental results.

3.1. Experiment Scheme

In the Unity 2022.2.1.f1 development environment, a scene is created that hosts a virtual avatar of a 173 cm tall person. Many animation assets are imported into the scene and applied sequentially to the virtual avatar. A total of 18 VKTPs are set on the avatar according to the scheme shown in Figure 1. Next, all animations are launched in the scene, and the points are recorded at a frequency of 600 Hz in the CSV format.
The collected data are processed in Python software using data analysis libraries (Pandas and SciPy). Further, various machine learning models are implemented; they are trained on a training dataset and evaluated on a control data set (using objective quantitative metrics and subjective visual analysis).
The software was launched and tested on the following hardware configuration: AMD Ryzen 5900X, 64 GB RAM, Nvidia RTX 3060 Ti, NVMe SSD, and operating system—Windows 10.

3.2. Description of the Dataset of Human Movements

To form a training dataset, data were collected from a virtual scene with 256 different types of animation (source—CMU Graphics Lab Motion Capture Database [42]) at a frequency of 30 times per second. In total, this allowed us to collect 1,088,174 records from 18 VKTPs (54 values per record). A total of 870,539 records were used as a training sample, 217,635 records were used as a test sample. A control set of 14,000 records with unique animation was also formed, which does not participate in either the training or test samples.
Further, an analysis of the collected data for all 18 key tracking points (Table 2) was carried out, during which the average values of the coordinates of the points along the three axes and their standard deviation, as well as the range of minimum and maximum values of each VKTP, were determined.
Checking the data for normality using the Shapiro test shows that the data are normally distributed with probability, p = 0 . Similar results were shown by a normality test based on D’Agostino and Pearson’s criteria [43]. Thus, the use of parametric tests for data evaluation is inappropriate.
For the convenience of assessing the quality of machine learning algorithms, a software tool based on the Matplotlib library was implemented, which allows visualizing an arbitrary number of virtual avatars. Each avatar is built on the basis of 18 points, after which they are connected in the right order. Figure 2 shows examples from the collected dataset.
Thus, the collected data are very diverse, allowing us to conclude that the generated dataset is sufficient.

3.3. Determining the Optimal Layout of Tracking Points

To select the optimal location of KTPs and their combinations, a preliminary study was carried out, including the following steps:
A total of 131,089 variants from all possible combinations of KTP are formed, including 18 sets of 1 KTP and 131,071 sets from all possible combinations of KTP, given the limitation that they always use the KTP of the head (1). KTP (1) must be present in the sets due to the fact that a person is always wearing a helmet when using virtual reality. Thus, the indicated 131,071 sets were formed using a complete enumeration of all combinations of the remaining 17 KTPs.
Next, the LinearRegression algorithm was used for test reconstruction of the human body for all 131,089 KTP variants. For the training procedure, the previously designated training (more than 800,000 poses) and control (14,000 poses) datasets were used, which made it possible to cover most of the typical movements. The LinearRegression algorithm was chosen because of its high performance, ease of implementation, and repeatability of results.
For each Q from 1 to 18, the five best results were selected, as presented in Table 3, indicating the maximum deviation, D max , and its average deviation (SD) for the sample. Compact notation is used in the table if consecutive KTPs were used in the set. The results obtained allow us to determine the size limits of the KTP set. At Q < 3 , firstly, the maximum error is too high, and secondly, the absence of symmetrical points does not allow the correct process of all poses. At Q > 13 , an increase in the reconstruction error can be observed, since it is necessary to approximate individual small sets of KTPs with a fixed set of a large number of KTPs, which can be difficult in the case of the specific scenarios discussed earlier in Section 2.2 (in the case when the desired KTPs are sufficiently independent from the rest). When Q = 18 , all input data values are equal to the output data values, so there is no need to approximate such a dependence.
In addition to the size limits of the KTP sets, the conducted experiment makes it possible to form the priority of choosing the KTP in the process of expanding the sets from the smallest to the largest one. First, there is a clear superiority of certain points (2 and 3, 6 and 7, 16 and 17) due to their constant presence in almost all the best sets for various Q . Secondly, it is necessary to take into account the symmetry of pairs of some points (for example, 2 and 3, 6 and 7) when forming sets, because the use of only one point from a pair may reduce the reconstruction accuracy for one of the body halves. This leads to the need to use a two-point step when increasing the length of the KTP sets (i.e., Q = 3, 5, 7, and so on) so as not to break the symmetry. Finally, some KTPs, despite their frequent use in the best sets, are inconvenient for use in real life due to the difficulty of fixing sensors in these zones (for example, points 16 and 17) or can be replaced by the nearest points (for example, 18 by 13, 14 by 6, and 15 by 7).
Based on the assessment of the obtained preliminary results and the experience of experts (based on the greatest convenience of fixing trackers and their priority), the following order of adding points was formed: starting from the extreme points on the arms and legs and ending with intermediate positions (on the knees, elbows, back, and so on). As a result, 6 preferred sets were formulated, starting from the simplest with 3 KTPs and ending with the maximum with 13 KTPs:
  • 3 KTPs: head (1) and left (2) and right (3) hand;
  • 5 KTPs: 3 KTPs with the addition of the left (4) and right (5) foot (heel area);
  • 7 KTPs: 5 KTPs with the addition of the left (6) and right (7) elbows;
  • 9 KTPs: 7 KTPs with the addition of the left (8) and right (9) knee;
  • 11 KTPs: 9 KTPs with the addition of the left (10) and right (11) hip joint;
  • 13 KTPs: 11 KTPs with the addition of the lower (12) and upper (13) points of the back.
Thus, during the first stage of research, the optimal location and sets of KTPs were determined, on which it is further necessary to train various machine learning algorithms.

3.4. Description of Used Machine Learning Models

To select the optimal solution, several machine learning algorithms were analyzed, the main parameters of which are presented in Table 4. Model development was carried out in the PyCharm environment using Python 3.9 interpreter.
At the preliminary stage, several architectures of multilayer neural networks were considered, some of which, due to an insufficient or excessive number of neurons and layers, did not achieve high accuracy values. As a result, two different architectures of neural networks were considered: the first, simpler one transforms the KTP vector (from 3 to 13) into a set of 18 VKTPs. The second one, which has multiple inputs, in addition to the current KTP data, uses, as the second input, a vector of 54 values from 18 VKTPs reconstructed in the previous step. On the one hand, this makes it possible to take into account previous body positions more correctly; on the other hand, it can lead to error accumulation. All neural networks were trained for 10 epochs, and the accuracy on the test set varied from 94 to 95 percent. The neural network models were implemented using the Keras framework based on Tensorflow.
Among the machine learning algorithms for solving the regression problem, the following were chosen: KNeighborsRegressor (KNN), DecisionTreeRegressor (DTR), LinearRegression (LR), RandomForestRegressor (RFR), SGDRegressor (SGDR), and AdaBoostRegressor (ADA). For their implementation, scikit-learn library (version 0.24.2) was used. The description and parameters of the models are presented in Table 4.

3.5. Analysis of Experimental Data

The selected machine learning algorithms were tested on 14,000 unique records not involved in the learning process. A comparison of algorithms by D m e a n , D max , D s u m , and MSE metrics, as well as regression time, T , for one set of input data is presented in Table 5. The algorithms are grouped depending on the number of initial VKTPs. The three best results in the group are selected in bold. The measurement of the calculation time of one reconstruction was performed on the CPU in one thread to achieve equivalent conditions.
Next, a visual comparison of the results obtained was carried out to determine the best algorithm because even high scores on the selected metrics may not provide a realistic reconstruction of the virtual avatar. To do this, every hundredth set of restored VKTPs was saved from the control group. Figure 3 and Figure 4 show comparisons of all algorithms for a random sample from the control group. The blue color denotes the reference position of the skeleton (vector R ), and the red color indicates the reconstructed VKTP (vector V R ).
As a result, there is a correlation between the visual realism of the reconstructed human body models and the metrics used: for lower values of D m e a n , D max , and MSE, there is an improvement in the quality of the reconstruction relative to the high values. The main criterion— D s u m —with the equality of other metrics, allows you to unambiguously determine the best solution, which, according to the observer, has an advantage from a visual point of view. Thus, this criterion is applicable for an objective assessment of the human body reconstruction quality.

4. Discussion

An analysis of the machine learning algorithm performances shows that the average reconstruction time does not exceed 2 milliseconds, which provides a sufficient calculation speed and will not affect the performance of virtual reality systems. The exception is the KNR algorithm, the VKTP calculation time for which is up to 37 milliseconds (at Q = 13 ), which makes it difficult to use it in real time.
Next, we analyze the minimum number of KTPs required for the reconstruction of a virtual avatar. The preliminary experiments show that the use of less than 3 and more than 13 KTPs is inappropriate; in addition, the need for symmetry of the used KTPs was revealed, which led to the formation of 6 sets. These sets are made up of KTP combinations that showed the best results in the preliminary tests to evaluate the number and location of KTPs (Section 3.3).
The conducted experimental studies and visual analysis show that the use of three points is not sufficient for all algorithms. The maximum deviation D max exceeds 0.4 m, and the average D m e a n does not decrease below 0.15 m. These results are not acceptable. The reconstructions of virtual avatars shown in Figure 3 and Figure 4 have significant deviations in the region of the lower extremities; the position of the legs does not correspond to the real one. Thus, it is impossible to reliably reconstruct the legs of a virtual avatar, especially in complex poses, without having information about the KTP in the region of the lower extremities. With Q = 5 , machine learning algorithms are much better at positioning the lower extremities. The use of seven KTPs has a positive effect on difficult hand positions, especially in the elbow area. Using nine or more KTPs has a positive impact, but it is not significant or visually sensitive. Let us summarize the recommendations for choosing the number of KTPs:
-
Q = 3 : an option for simplified motion capture systems that do not require foot tracking and precise hand positioning;
-
Q = 5 : the optimal solution for most virtual reality systems due to the ability to accurately position all the user’s limbs;
-
Q = 7 : a solution for systems that require high accuracy in the positioning of the upper limbs, providing a reconstruction error comparable to more complex models.
Given the need to minimize the number of KTPs for a wide range of virtual reality systems, the optimal accuracy of virtual avatar reconstruction is achieved with five and seven input KTPs.
A comparison of various machine learning algorithms made it possible to evaluate them by objective and subjective metrics. The best results in quantitative metrics at 3 Q 7 were shown by LinearRegression (LR), SGDRegressor (SGDR), AdaBoostRegressor (ADA), and neural networks (with single and multiple inputs). Visual analysis shows that, in the most important range 3 Q 7 , these algorithms are generally comparable and show a corresponding quality similar to quantitative metrics. With an increase in the number of KTPs, neural networks begin to show worse results due to generalization because they introduce some deviations for all points. The remaining algorithms showed their inefficiency: DecisionTreeRegressor (DTR) and RandomForestRegressor (RFR) form an averaged universal body position that is too far from the real one, and KNeighborsRegressor (KNR) shows closer results but has the same drawback. Additionally, the KNR model has the largest size among all.
Taking into account the final accuracy by the criterion D s u m , model size and performance, the optimal machine learning algorithms for 5 Q 7 are (in descending order): AdaBoostRegressor, LinearRegression, and SGDRegressor. The other algorithms are not recommended for use.
The results obtained during the experiments are consistent in their accuracy with the studies carried out in [44], where, using six KTPs (HMD and five trackers), an average deviation of 0.02 m (for body segments, not specific KTPs) and a calculation time of 1.65 ms were obtained. In study [45], various algorithms produced a deviation from 0.07 to 0.11 m when using six KTPs.
Thus, the novelty of this study lies in the use of the digital shadow concept for modeling nonlinear dynamic processes of human movement in virtual reality and the development of algorithms for collecting and processing information to determine the minimum number of tracking points, as well as conducting experimental studies to find the optimal machine learning algorithm that allows for implementing reconstruction of the user’s virtual avatar under conditions of insufficient data.
Further research will be related to the integration of optimal machine learning algorithms into virtual reality systems. This will require the addition of new stages in the data processing pipeline in VR systems [46]. The functioning of such systems will not be based on the direct use of data from KTP (for example, virtual reality trackers), but on sending them to digital shadow software, receiving a restored VKTP set, forming a virtual avatar, and changing the final person’s model position in the scene. An avatar formed in this way will be as close as possible to the real position of the body, taking into account the minimum number of tracked points.

5. Conclusions

In this study, the problem of modeling nonlinear dynamic processes of human movement in virtual reality was considered to simplify motion capture systems and reduce the number of KTPs. To solve it, the concept of a digital shadow was proposed, including the formalization of the motion process based on the analysis of the functioning of various motion capture systems, the development of a procedure for generating a dataset from a virtual avatar, and the formulation and solution of the problem of minimizing the number of key points for tracking a person based on machine learning algorithms.
In the course of this study, the existing approaches to solving the problem of inverse kinematics in the conditions of a limited number of tracking points were analyzed using various machine learning algorithms. Further, a comparison of motion capture systems was carried out, and shortcomings of the data obtained from them were identified, which led to the development of a methodology for collecting information about human movements from a virtual avatar. This allows you to increase the speed of formation, size, and diversity of the dataset by simulating various scenarios in a virtual scene.
Next, the simulation of the human movement process in virtual reality was carried out, and the main components of the digital shadow were formalized, after which the task of optimizing the number of KTPs was formulated with the selection of the best machine learning algorithm.
A total of six main combinations of KTP were identified, for which various machine learning algorithms were compared for the reconstruction of the entire virtual avatar. It was found that the use of three KTPs is not sufficient for all algorithms, the maximum deviation is more than 0.4 m, and the average is more than 0.15 m; in addition, it is impossible to reliably reconstruct the legs of a virtual avatar, especially in complex poses. The use of five KTPs eliminates this problem, and seven KTPs allow complex hand positions to be reconstructed in the elbow area. The use of nine or more KTPs has a positive effect, but it is not significant and visually sensitive, so five to seven KTPs are recommended as the optimal number of points for various use cases. When comparing algorithms under these conditions for solving the regression problem, the best results were shown (in descending order): AdaBoostRegressor, LinearRegression/SGDRegressor, neural network (with multiple inputs), and neural network (with one input). When using AdaBoostRegressor, the maximum deviation is not more than 0.25 m, and the average is not more than 0.10 m.
The novelty of this study is the use of the digital shadow concept and the collection of data from a virtual avatar to simulate nonlinear dynamic processes of human movement in virtual reality. The proposed approaches make it possible to reduce the cost and complexity of motion capture systems, speed up the process of collecting information for training machine learning algorithms, and increase the amount of data by simulating various motion scenarios in virtual space.

Author Contributions

Conceptualization, A.O. and D.D.; methodology, A.O.; software, D.T.; validation, A.O. and A.V.; formal analysis, A.O. and A.V.; investigation, A.O., D.D. and A.V.; resources, D.T. and A.V.; data curation, D.T.; writing—original draft preparation, A.O.; writing—review and editing, A.O. and A.V.; visualization, D.T.; supervision, A.O. and D.D.; project administration, D.D.; funding acquisition, D.D. All authors have read and agreed to the published version of the manuscript.

Funding

The research was funded by a grant of the Russian Science Foundation (grant number: 22-71-10057; https://rscf.ru/en/project/22-71-10057/ (accessed on 20 April 2023)).

Data Availability Statement

Data available in a publicly accessible repository that does not issue DOIs. Publicly available datasets were analyzed in this study. This data can be found here: mocap.cs.cmu.edu. The database was created with funding from NSF EIA-0196217.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Al-Jundi, H.A.; Tanbour, E.Y. A framework for fidelity evaluation of immersive virtual reality systems. Virtual Real. 2022, 26, 1103–1122. [Google Scholar] [CrossRef]
  2. Obukhov, A.D.; Volkov, A.A.; Vekhteva, N.A.; Teselkin, D.V.; Arkhipov, A.E. Human motion capture algorithm for creating digital shadows of the movement process. J. Phys. Conf. Ser. 2022, 2388, 012033. [Google Scholar] [CrossRef]
  3. Oudah, M.; Al-Naji, A.; Chahl, J. Hand Gesture Recognition Based on Computer Vision: A Review of Techniques. J. Imaging 2020, 6, 73. [Google Scholar] [CrossRef] [PubMed]
  4. Nie, J.Z.; Nie, J.W.; Hung, N.T.; Cotton, R.J.; Slutzky, M.W. Portable, open-source solutions for estimating wrist position during reaching in people with stroke. Sci. Rep. 2021, 11, 22491. [Google Scholar] [CrossRef] [PubMed]
  5. Hindle, B.R.; Keogh, J.W.L.; Lorimer, A.V. Inertial-Based Human Motion Capture: A Technical Summary of Current Processing Methodologies for Spatiotemporal and Kinematic Measures. Appl. Bionics Biomech. 2021, 2021, 6628320. [Google Scholar] [CrossRef]
  6. Filippeschi, A.; Schmitz, N.; Miezal, M.; Bleser, G.; Ruffaldi, E.; Stricker, D. Survey of Motion Tracking Methods Based on Inertial Sensors: A Focus on Upper Limb Human Motion. Sensors 2017, 17, 1257. [Google Scholar] [CrossRef]
  7. Gonzalez-Islas, J.-C.; Dominguez-Ramirez, O.-A.; Lopez-Ortega, O.; Peña-Ramirez, J.; Ordaz-Oliver, J.-P.; Marroquin-Gutierrez, F. Crouch Gait Analysis and Visualization Based on Gait Forward and Inverse Kinematics. Appl. Sci. 2022, 12, 10197. [Google Scholar] [CrossRef]
  8. Parger, M.; Mueller, J.H.; Schmalstieg, D.; Steinberger, M. Human upper-body inverse kinematics for increased embodiment in consumer-grade virtual reality. In Proceedings of the 24th ACM Symposium on Virtual Reality Software and Technology, Tokyo, Japan, 28 November 2018; pp. 1–10. [Google Scholar] [CrossRef]
  9. Van Emmerik, R.E.; Rosenstein, M.T.; McDermott, W.J.; Hamill, J. A Nonlinear Dynamics Approach to Human Movement. J. Appl. Biomech. 2004, 20, 396–420. [Google Scholar] [CrossRef]
  10. Wu, Y.; Tao, K.; Chen, Q.; Tian, Y.; Sun, L. A Comprehensive Analysis of the Validity and Reliability of the Perception Neuron Studio for Upper-Body Motion Capture. Sensors 2022, 22, 6954. [Google Scholar] [CrossRef]
  11. Ikbal, M.S.; Ramadoss, V.; Zoppi, M. Dynamic Pose Tracking Performance Evaluation of HTC Vive Virtual Reality System. IEEE Access 2020, 9, 3798–3815. [Google Scholar] [CrossRef]
  12. Hellsten, T.; Karlsson, J.; Shamsuzzaman, M.; Pulkkis, G. The Potential of Computer Vision-Based Marker-Less Human Motion Analysis for Rehabilitation. Rehabil. Process Outcome 2021, 10, 11795727211022330. [Google Scholar] [CrossRef]
  13. Chen, W.; Yu, C.; Tu, C.; Lyu, Z.; Tang, J.; Ou, S.; Fu, Y.; Xue, Z. A Survey on Hand Pose Estimation with Wearable Sensors and Computer-Vision-Based Methods. Sensors 2020, 20, 1074. [Google Scholar] [CrossRef]
  14. Degen, R.; Tauber, A.; Nüßgen, A.; Irmer, M.; Klein, F.; Schyr, C.; Leijon, M.; Ruschitzka, M. Methodical Approach to Integrate Human Movement Diversity in Real-Time into a Virtual Test Field for Highly Automated Vehicle Systems. J. Transp. Technol. 2022, 12, 296–309. [Google Scholar] [CrossRef]
  15. Caserman, P.; Garcia-Agundez, A.; Konrad, R.; Göbel, S.; Steinmetz, R. Real-time body tracking in virtual reality using a Vive tracker. Virtual Real. 2019, 23, 155–168. [Google Scholar] [CrossRef]
  16. Feigl, T.; Gruner, L.; Mutschler, C.; Roth, D. Real-time gait reconstruction for virtual reality using a single sensor. In Proceedings of the 2020 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), Recife, Brazil, 9–13 November 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 84–89. [Google Scholar]
  17. Liu, H.; Zhang, Z.; Xie, X.; Zhu, Y.; Liu, Y.; Wang, Y.; Zhu, S.-C. High-Fidelity Grasping in Virtual Reality using a Glove-based System. In Proceedings of the International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 5180–5186. [Google Scholar]
  18. Li, J.; Xu, C.; Chen, Z.; Bian, S.; Yang, L.; Lu, C. HybrIK: A Hybrid Analytical-Neural Inverse Kinematics Solution for 3D Human Pose and Shape Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 3383–3393. [Google Scholar]
  19. Oyama, E.; Agah, A.; MacDorman, K.F.; Maeda, T.; Tachi, S. A modular neural network architecture for inverse kinematics model learning. Neurocomputing 2001, 38, 797–805. [Google Scholar] [CrossRef]
  20. Bai, Y.; Luo, M.; Pang, F. An Algorithm for Solving Robot Inverse Kinematics Based on FOA Optimized BP Neural Network. Appl. Sci. 2021, 11, 7129. [Google Scholar] [CrossRef]
  21. Kratzer, P.; Toussaint, M.; Mainprice, J. Prediction of Human Full-Body Movements with Motion Optimization and Recurrent Neural Networks. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1792–1798. [Google Scholar] [CrossRef]
  22. Bataineh, M.; Marler, T.; Abdel-Malek, K.; Arora, J. Neural network for dynamic human motion prediction. Expert Syst. Appl. 2016, 48, 26–34. [Google Scholar] [CrossRef]
  23. Cui, Q.; Sun, H.; Li, Y.; Kong, Y. Efficient human motion recovery using bidirectional attention network. Neural Comput. Appl. 2020, 32, 10127–10142. [Google Scholar] [CrossRef]
  24. Kritzinger, W.; Karner, M.; Traar, G.; Henjes, J.; Sihn, W. Digital Twin in manufacturing: A categorical literature review and classification. IFAC-PapersOnLine 2018, 51, 1016–1022. [Google Scholar] [CrossRef]
  25. Grieves, M. Origins of the Digital Twin Concept. 2016. Available online: https://www.researchgate.net/publication/307509727 (accessed on 11 April 2023).
  26. Krüger, J.; Wang, L.; Verl, A.; Bauernhansl, T.; Carpanzano, E.; Makris, S.; Fleischer, J.; Reinhart, G.; Franke, J.; Pellegrinelli, S. Innovative control of assembly systems and lines. CIRP Ann. 2017, 66, 707–730. [Google Scholar] [CrossRef]
  27. Ribeiro, P.M.S.; Matos, A.C.; Santos, P.H.; Cardoso, J.S. Machine Learning Improvements to Human Motion Tracking with IMUs. Sensors 2020, 20, 6383. [Google Scholar] [CrossRef]
  28. Stančić, I.; Musić, J.; Grujić, T.; Vasić, M.K.; Bonković, M. Comparison and Evaluation of Machine Learning-Based Classification of Hand Gestures Captured by Inertial Sensors. Computation 2022, 10, 159. [Google Scholar] [CrossRef]
  29. Yacchirema, D.; de Puga, J.S.; Palau, C.; Esteve, M. Fall detection system for elderly people using IoT and ensemble machine learning algorithm. Pers. Ubiquitous Comput. 2019, 23, 801–817. [Google Scholar] [CrossRef]
  30. Halilaj, E.; Rajagopal, A.; Fiterau, M.; Hicks, J.L.; Hastie, T.J.; Delp, S.L. Machine learning in human movement biomechanics: Best practices, common pitfalls, and new opportunities. J. Biomech. 2018, 81, 1–11. [Google Scholar] [CrossRef]
  31. Bartol, K.; Bojanić, D.; Petković, T.; Peharec, S.; Pribanić, T. Linear regression vs. deep learning: A simple yet effective baseline for human body measurement. Sensors 2022, 22, 1885. [Google Scholar] [CrossRef] [PubMed]
  32. Turgeon, S.; Lanovaz, M.J. Tutorial: Applying Machine Learning in Behavioral Research. Perspect. Behav. Sci. 2020, 43, 697–723. [Google Scholar] [CrossRef]
  33. Pavllo, D.; Feichtenhofer, C.; Auli, M.; Grangier, D. Modeling Human Motion with Quaternion-Based Neural Networks. Int. J. Comput. Vis. 2020, 128, 855–872. [Google Scholar] [CrossRef]
  34. Almeida, R.O.; Munis, R.A.; Camargo, D.A.; da Silva, T.; Sasso Júnior, V.A.; Simões, D. Prediction of Road Transport of Wood in Uruguay: Approach with Machine Learning. Forests 2022, 13, 1737. [Google Scholar] [CrossRef]
  35. Li, B.; Bai, B.; Han, C. Upper body motion recognition based on key frame and random forest regression. Multimedia Tools Appl. 2020, 79, 5197–5212. [Google Scholar] [CrossRef]
  36. Sipper, M.; Moore, J.H. AddGBoost: A gradient boosting-style algorithm based on strong learners. Mach. Learn. Appl. 2022, 7, 100243. [Google Scholar] [CrossRef]
  37. Kanko, R.M.; Laende, E.K.; Davis, E.M.; Selbie, W.S.; Deluzio, K.J. Concurrent assessment of gait kinematics using marker-based and markerless motion capture. J. Biomech. 2021, 127, 110665. [Google Scholar] [CrossRef]
  38. Choo, C.Z.Y.; Chow, J.Y.; Komar, J. Validation of the Perception Neuron system for full-body motion capture. PLoS ONE 2022, 17, e0262730. [Google Scholar] [CrossRef]
  39. Al-Faris, M.; Chiverton, J.; Ndzi, D.; Ahmed, A.I. A Review on Computer Vision-Based Methods for Human Action Recognition. J. Imaging 2020, 6, 46. [Google Scholar] [CrossRef]
  40. Zheng, B.; Sun, G.; Meng, Z.; Nan, R. Vegetable Size Measurement Based on Stereo Camera and Keypoints Detection. Sensors 2022, 22, 1617. [Google Scholar] [CrossRef] [PubMed]
  41. Sers, R.; Forrester, S.; Moss, E.; Ward, S.; Ma, J.; Zecca, M. Validity of the Perception Neuron inertial motion capture system for upper body motion analysis. Meas. J. Int. Meas. Confed. 2020, 149, 107024. [Google Scholar] [CrossRef]
  42. The Daz-Friendly Bvh Release of Cmu Motion Capture Database. Available online: https://www.sites.google.com/a/cgspeed.com/cgspeed/motion-capture/the-daz-friendly-bvh-release-of-cmus-motion-capture-database (accessed on 22 March 2023).
  43. Demir, S. Comparison of Normality Tests in Terms of Sample Sizes under Different Skewness and Kurtosis Coefficients. Int. J. Assess. Tools Educ. 2022, 9, 397–409. [Google Scholar] [CrossRef]
  44. Zeng, Q.; Zheng, G.; Liu, Q. PE-DLS: A novel method for performing real-time full-body motion reconstruction in VR based on Vive trackers. Virtual Real. 2022, 26, 1391–1407. [Google Scholar] [CrossRef]
  45. Yi, X.; Zhou, Y.; Xu, F. Transpose: Real-time 3D human translation and pose estimation with six inertial sensors. ACM Trans. Graph. 2021, 40, 1–13. [Google Scholar] [CrossRef]
  46. Obukhov, A.; Volkov, A.A.; Nazarova, A.O. Microservice Architecture of Virtual Training Complexes. Inform. Autom. 2022, 21, 1265–1289. [Google Scholar] [CrossRef]
Figure 1. Scheme and order of KTP placement, where 1–18—numerical identifiers of each KTP.
Figure 1. Scheme and order of KTP placement, where 1–18—numerical identifiers of each KTP.
Computation 11 00085 g001
Figure 2. Visualization of random samples from the dataset.
Figure 2. Visualization of random samples from the dataset.
Computation 11 00085 g002
Figure 3. Visual comparison of virtual avatar reconstruction quality using different machine learning algorithms: (a) LR; (b) SGDR; (c) NN (S); (d) NN (M).
Figure 3. Visual comparison of virtual avatar reconstruction quality using different machine learning algorithms: (a) LR; (b) SGDR; (c) NN (S); (d) NN (M).
Computation 11 00085 g003
Figure 4. Visual comparison of virtual avatar reconstruction quality using different machine learning algorithms: (a) KNR; (b) DTR; (c) RFR; (d) ADA.
Figure 4. Visual comparison of virtual avatar reconstruction quality using different machine learning algorithms: (a) KNR; (b) DTR; (c) RFR; (d) ADA.
Computation 11 00085 g004
Table 1. Comparison of different approaches to collecting human movement data.
Table 1. Comparison of different approaches to collecting human movement data.
MethodNumber of KTPAccuracyMeasurement Sample RateTracking Area
Motion capture suit12 … 32±1 mmUp to 250 HzRadius up to 150 m
VR trackers1 … 68±0.7 mm60 … 120 HzFrom 1.5 × 2 m
to 9 × 9 m
Computer vision17 … 33From ±5 mm
to ±40 mm
15 … 120 HzLimited by camera field of view
Our methodUnlimitedUp to ±10−6 mm (using “float”),
up to ±10−13 mm (using “double”)
Up to 600 HzUnlimited
Table 2. Statistical description of the dataset.
Table 2. Statistical description of the dataset.
PointsMean ± SD[Min; Max]
#XYZXYZ
10.01 ± 0.11.425 ± 0.2260.007 ± 0.089[−0.526; 0.544][−1.002; 1.622][−0.52; 0.724]
2−0.004 ± 0.0520.976 ± 0.192−0.003 ± 0.049[−0.25; 0.248][−1.072; 1.26][−0.28; 0.497]
30.002 ± 0.0311.116 ± 0.1970.001 ± 0.028[−0.197; 0.274][−0.92; 1.28][−0.201; 0.594]
40.005 ± 0.071.336 ± 0.2140.004 ± 0.062[−0.422; 0.443][−0.937; 1.516][−0.426; 0.677]
50.012 ± 0.1051.29 ± 0.21−0.007 ± 0.102[−0.41; 0.436][−0.939; 1.471][−0.421; 0.566]
60.02 ± 0.1711.109 ± 0.215−0.014 ± 0.173[−0.654; 0.63][−1.046; 1.64][−0.593; 0.624]
70.028 ± 0.2241.045 ± 0.277−0.005 ± 0.222[−0.877; 0.827][−1.155; 1.858][−0.774; 0.806]
8−0.003 ± 0.1061.292 ± 0.2110.014 ± 0.107[−0.42; 0.415][−0.879; 1.476][−0.412; 0.751]
9−0.01 ± 0.1691.104 ± 0.2120.023 ± 0.173[−0.666; 0.645][−1.037; 1.65][−0.598; 0.774]
100.001 ± 0.221.035 ± 0.270.025 ± 0.216[−0.871; 0.84][−1.133; 1.864][−0.778; 0.944]
11−0.003 ± 0.0880.849 ± 0.19−0.012 ± 0.086[−0.322; 0.343][−1.224; 1.31][−0.332; 0.394]
120.006 ± 0.130.502 ± 0.199−0.008 ± 0.13[−0.615; 0.605][−1.46; 1.616][−0.614; 0.607]
130.001 ± 0.1870.129 ± 0.225−0.02 ± 0.186[−0.999; 1.017][−1.809; 2.032][−1.015; 0.98]
140.014 ± 0.2110.049 ± 0.227−0.016 ± 0.211[−1.125; 1.154][−1.885; 2.177][−1.148; 1.057]
15−0.012 ± 0.090.846 ± 0.190.0 ± 0.086[−0.331; 0.331][−1.215; 1.31][−0.346; 0.513]
16−0.009 ± 0.1280.498 ± 0.1990.012 ± 0.126[−0.591; 0.605][−1.479; 1.624][−0.598; 0.789]
17−0.022 ± 0.1860.123 ± 0.2230.006 ± 0.179[−1.003; 1.001][−1.85; 2.046][−0.968; 1.013]
18−0.011 ± 0.2050.041 ± 0.2260.015 ± 0.2[−1.144; 1.048][−1.936; 2.156][−1.117; 1.122]
SD—standard deviation.
Table 3. Comparison results of different KTP sets.
Table 3. Comparison results of different KTP sets.
QUsed KTPs D max QUsed KTPs D max ± SD
120.444 ± 0.14121,150.431 ± 0.147
30.448 ± 0.133 1,70.437 ± 0.143
60.449 ± 0.130 1,30.438 ± 0.148
70.452 ± 0.133 1,20.441 ± 0.154
140.469 ± 0.150 1,60.442 ± 0.141
31,3,170.381 ± 0.14241,2,3,170.332 ± 0.149
1,3,50.388 ± 0.142 1,3,16,170.333 ± 0.119
1,2,170.389 ± 0.144 1,2,16,170.335 ± 0.125
1,7,170.389 ± 0.145 1,3,6,170.336 ± 0.146
1,3,90.393 ± 0.154 1,3,5,160.338 ± 0.122
51–3,16,170.248 ± 0.0961–3,9,16,170.243 ± 0.109
1–3,5,160.254 ± 0.100 1–3,8,16,170.245 ± 0.110
1,3,6,16,170.256 ± 0.099 1–3,5,9,160.246 ± 0.111
1–3,4,170.256 ± 0.101 1–4,8,170.248 ± 0.098
1,2,7,16,170.258 ± 0.099 1–3,7,16,170.249 ± 0.098
71–4,9,16,170.241 ± 0.11181–3,7,9,10,16,170.233 ± 0.104
1–3,6,9,16,170.241 ± 0.109 1–4,9,16–180.236 ± 0.109
1–3,5,9,16,170.242 ± 0.109 1–4,6,9,16,170.239 ± 0.112
1–5,9,160.243 ± 0.111 1–3,5,6,9,16,170.240 ± 0.113
1–3,5,8,16,170.244 ± 0.114 1–3,6–8,16,170.241 ± 0.111
91–5,7–9,130.220 ± 0.122101–3,7–9,11,14,16,170.195 ± 0.06
1–4,6,9,16–180.229 ± 0.114 1–8,9,130.196 ± 0.08
1–4,7,9,16–180.230 ± 0.120 1–3,5,7–9,11,14,160.197 ± 0.07
1–3,6,7,9,10,16,170.231 ± 0.108 1–4,7–9,11,14,170.198 ± 0.07
1–4,8,9,13,15,170.231 ± 0.115 1–5,7–9,11,140.200 ± 0.08
111–9,13,170.182 ± 0.06121–11,180.163 ± 0.049
1–9,13,160.184 ± 0.066 1–10,15,180.169 ± 0.031
1–4,6–9,13,16,170.191 ± 0.077 1–9,11,14,180.170 ± 0.037
1–3,6–11,16,170.193 ± 0.085 1–9,13,16,170.179 ± 0.064
1–5,7–9,11,14,180.193 ± 0.062 1–3,5–11,16,180.180 ± 0.101
131–11,17,180.162 ± 0.048141–11,16–180.161 ± 0.038
1–11,16,180.163 ± 0.049 1–11,13,17,180.165 ± 0.039
1–11,13,180.166 ± 0.049 1–11,13,16,180.166 ± 0.041
1–9,11,14,17,180.169 ± 0.036 1–9,11,14,16–180.168 ± 0.026
1–9,11,14,16,180.169 ± 0.037 1–10,15–180.170 ± 0.025
151–11,13,16–180.164 ± 0.048161–11,13,14,16–180.176 ± 0.053
1–9,11,13,14,16–180.173 ± 0.043 1–11,13,15–180.178 ± 0.053
1–10,13,15,16–180.174 ± 0.042 1–11,13–170.182 ± 0.011
1–11,14,16–180.175 ± 0.052 1–13,16–180.184 ± 0.017
1–11,15–180.176 ± 0.051 1–11,14–180.191 ± 0.012
171–11,13–180.192 ± 0.013181–181 × 10−5 ± 0.0
1–9,11–180.279 ± 0.094
1–14,16–180.317 ± 0.021
1–13,15–180.317 ± 0.021
1–7,9–180.743 ± 0.318
SD—standard deviation.
Table 4. Parameters of selected machine learning algorithms.
Table 4. Parameters of selected machine learning algorithms.
ModelDescription
LinearRegression (LR)Ordinary least squares linear regression with default parameters.
SGDRegressor (SGDR)Standard linear model fitted with stochastic gradient descent with loss = “squared_loss” passed to MultiOutputRegressor.
Simple neural network (NN (S))Multilayer neural network with input of 9–39 neurons, 3 hidden dense layers of 100 neurons with ReLU activation function, 2 dropout layers (20% dropout rate), output—54 neurons.
Multiple neural network (NN (M))Multilayer neural network with two inputs (9–39 neurons and 54), 3 hidden dense layers of 100 neurons, 2 dropout layers (20% dropout rate), output—54 neurons.
KNeighborsRegressor (KNN)Standard regression based on k-nearest neighbors algorithm = “ball_tree”.
DecisionTreeRegressor (DTR)Standard decision tree regressor with max_depth = 5.
RandomForestRegressor (RFR)Standard random forest with n_estimators = 10, max_depth = 5.
AdaBoostRegressor (ADA)Standard AdaBoost regressor based on DecisionTreeRegressor with n_estimators = 100, max_depth = 10.
Table 5. Results of machine learning algorithm comparison.
Table 5. Results of machine learning algorithm comparison.
QMethod D m e a n ± SD D max ± SD D s u m ± SDMSET, ms
3LR0.153 ± 0.0520.4 ± 0.172.757 ± 0.9450.014<0.001
SGDR0.153 ± 0.0520.4 ± 0.1692.756 ± 0.9440.0140.112
NN (S)0.173 ± 0.0640.423 ± 0.1983.114 ± 1.1590.0171.273
NN (M)0.17 ± 0.0630.409 ± 0.193.051 ± 1.1330.0161.548
KNR0.183 ± 0.0710.437 ± 0.2183.288 ± 1.2780.0191.228
DTR0.221 ± 0.0690.463 ± 0.1443.977 ± 1.2440.0220.001
RFR0.216 ± 0.0660.443 ± 0.1393.894 ± 1.180.0210.001
ADA0.173 ± 0.0480.451 ± 0.1723.119 ± 0.8580.0180.300
5LR0.102 ± 0.0350.26 ± 0.1021.835 ± 0.6380.007<0.001
SGDR0.102 ± 0.0350.26 ± 0.11.841 ± 0.6380.0070.008
NN (S)0.141 ± 0.0380.26 ± 0.0732.544 ± 0.6840.0081.214
NN (M)0.128 ± 0.0440.255 ± 0.0662.311 ± 0.7890.0081.542
KNR0.147 ± 0.0520.282 ± 0.0832.65 ± 0.940.01012.463
DTR0.243 ± 0.070.499 ± 0.1444.381 ± 1.2530.027<0.001
RFR0.236 ± 0.0670.484 ± 0.1414.252 ± 1.2040.0250.001
ADA0.095± 0.0280.246 ± 0.0761.718 ± 0.5090.0060.316
7LR0.089 ± 0.0320.256 ± 0.1011.6 ± 0.5780.006<0.001
SGDR0.09 ± 0.0320.255 ± 0.0991.617 ± 0.5760.0060.011
NN (S)0.142 ± 0.0410.257 ± 0.0842.559 ± 0.7420.0091.221
NN (M)0.134 ± 0.0460.258 ± 0.0732.411 ± 0.8230.0081.553
KNR0.145 ± 0.0520.276 ± 0.0862.604 ± 0.940.00915.877
DTR0.23 ± 0.0710.477 ± 0.1454.141 ± 1.2740.024<0.001
RFR0.23 ± 0.0710.477 ± 0.1454.139 ± 1.2720.0240.001
ADA0.08 ± 0.0270.24 ± 0.081.441 ± 0.4790.0050.335
9LR0.072 ± 0.0360.251 ± 0.1271.297 ± 0.6440.005<0.001
SGDR0.075 ± 0.0360.247 ± 0.1231.351 ± 0.650.0050.013
NN (S)0.125 ± 0.0340.232 ± 0.0882.253 ± 0.6110.0071.211
NN (M)0.131 ± 0.0390.234 ± 0.0652.354 ± 0.6970.0071.563
KNR0.139 ± 0.0480.254 ± 0.0722.501 ± 0.870.00920.011
DTR0.23 ± 0.0710.477 ± 0.1454.141 ± 1.2760.024<0.001
RFR0.23 ± 0.0710.477 ± 0.1454.142 ± 1.2750.0240.001
ADA0.058 ± 0.0220.207 ± 0.0761.041 ± 0.390.0030.353
11LR0.039 ± 0.0230.192 ± 0.0730.708 ± 0.4140.002<0.001
SGDR0.059 ± 0.0340.195 ± 0.0621.055 ± 0.6090.0030.012
NN (S)0.133 ± 0.0360.242 ± 0.0622.4 ± 0.650.0071.22
NN (M)0.133 ± 0.0430.252 ± 0.0652.398 ± 0.7680.0081.56
KNR0.138 ± 0.0480.252 ± 0.0752.476 ± 0.8560.00829.341
DTR0.23 ± 0.0710.477 ± 0.1454.141 ± 1.2780.024<0.001
RFR0.23 ± 0.0710.477 ± 0.1454.14 ± 1.2770.0240.002
ADA0.034 ± 0.010.18 ± 0.0430.621 ± 0.1850.0010.427
13LR0.028 ± 0.0050.186 ± 0.0180.507 ± 0.0820.001<0.001
SGDR0.056 ± 0.0320.148 ± 0.0621.014 ± 0.580.0020.054
NN (S)0.128 ± 0.0370.239 ± 0.0642.307 ± 0.6660.0071.212
NN (M)0.129 ± 0.0430.229 ± 0.0582.33 ± 0.7740.0071.572
KNR0.137 ± 0.0470.251 ± 0.0752.474 ± 0.8460.00837.614
DTR0.23 ± 0.0710.477 ± 0.1454.141 ± 1.2780.024<0.001
RFR0.23 ± 0.0710.477 ± 0.1454.14 ± 1.2780.0240.001
ADA0.03 ± 0.0050.184 ± 0.0180.537 ± 0.090.0010.418
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Obukhov, A.; Dedov, D.; Volkov, A.; Teselkin, D. Modeling of Nonlinear Dynamic Processes of Human Movement in Virtual Reality Based on Digital Shadows. Computation 2023, 11, 85. https://doi.org/10.3390/computation11050085

AMA Style

Obukhov A, Dedov D, Volkov A, Teselkin D. Modeling of Nonlinear Dynamic Processes of Human Movement in Virtual Reality Based on Digital Shadows. Computation. 2023; 11(5):85. https://doi.org/10.3390/computation11050085

Chicago/Turabian Style

Obukhov, Artem, Denis Dedov, Andrey Volkov, and Daniil Teselkin. 2023. "Modeling of Nonlinear Dynamic Processes of Human Movement in Virtual Reality Based on Digital Shadows" Computation 11, no. 5: 85. https://doi.org/10.3390/computation11050085

APA Style

Obukhov, A., Dedov, D., Volkov, A., & Teselkin, D. (2023). Modeling of Nonlinear Dynamic Processes of Human Movement in Virtual Reality Based on Digital Shadows. Computation, 11(5), 85. https://doi.org/10.3390/computation11050085

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop