Next Article in Journal
Experimental and Computational Study of Archery Arrows Fletched with Straight Vanes
Previous Article in Journal
Evaluation of 3-Axial Knee Joint Torques Produced by Compression Sports Tights in Running Motion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Proceeding Paper

Multi-Technology Correction Based 3D Human Pose Estimation for Jump Analysis in Figure Skating †

1
School of Information, Production and Systems, Waseda University, Kitakyushu 808-0315, Japan
2
School of Artificial Intelligence, Xidian University, Xi’an 710071, China
3
School of Sport Sciences, Waseda University, Tokyo 169-8050, Japan
*
Author to whom correspondence should be addressed.
Presented at the 13th conference of the International Sports Engineering Association, Online, 22–26 June 2020.
Proceedings 2020, 49(1), 95; https://doi.org/10.3390/proceedings2020049095
Published: 15 June 2020

Abstract

:
Jump analysis in figure skating is important. Recovering the 3D pose of a figure skater has become increasingly important. However, issues such as restrictions from an athlete’s clothing, self-occlusion, abnormal pose and so on will result in poor results. This paper proposes a multi-technology correction framework to obtain a 3D human pose. The framework consists of three key components: temporal information-based mutational point correction, multi-perspective-based reconstructed point selection and trajectory smoothness-based inaccurate point correction. Firstly, temporal information is used to correct the mutational points at the 2D level. Secondly, a multi-perspective is used to select the correct spatial points at the 3D level. Thirdly, trajectory smoothness is used to correct inaccuracies at the 3D level. This work will serve the purpose of displaying the 3D animated pose of a figure skater. The quality grade of the result rate on the test sequences is 87.25%.

1. Introduction

Recently, estimating 3D human poses in sports has attracted academic interests for its vast potential. Analyzing the 3D jump pose in figure skating is an active research area, as it plays a significant role for a figure skater’s behavior understanding. It can not only objectively evaluate the performance of a figure skater’s jump, but also enhance the audience’s entertainment experience by displaying the jump height, spatial trajectory, action details and other information obtained from a figure skating video. Unfortunately, the diverse variations in background, costume, abnormal pose, self-occlusion, illumination and camera parameters make it a challenging problem. Recent advanced technologies in estimating 3D pose have not covered these variations appropriately. Motivated by these problems, this work has developed a transformation system to generate a 3D pose conditioned on the corresponding 2D pose.
The current related work cannot meet the requirements for estimating the 3D pose of figure skaters. It is well known that the goal of 3D human pose estimation is to localize key points of single or multiple human bodies in a 3D space. However, most of the previous 3D human pose estimation methods utilize a convolutional neural network and performed well on large-scale publicly datasets [1,2,3,4]. However, many methods estimate the relative 3D pose to a reference point in the body. Then, according to the prior information such as the length of the bone, to calculate the final 3D pose is done by adding the 3D coordinates of the reference point to the estimated relative coordinates [5,6]. The 3D coordinates obtained by these methods are not the real positions of key points in the space. This is not a suitable method for the analysis of jumps in figure skating because it lacks a 3D annotated dataset. Currently, a 3D dataset is much more difficult to obtain because accurate 3D pose annotation requires using motion capture in indoor artificial settings. However, these are not possible for figure skating which requires a large venue.
Different from the previous 3D pose estimation method, this work obtains the athlete’s 3D pose with fully spatial realism by considering the particularity of figure skating as shown in Figure 1. The whole system takes as the input a sequence of images capturing the motion of a figure skater from a synchronized multi-perspective and outputs the 3D joints of the target person in the form of a 3D human model video. The three proposed methods are as follows: temporal information-based mutational point correction, multi-perspective-based reconstructed point selection and trajectory smoothness-based inaccurate point correction. The details of the proposals will be introduced in Section 3.
The rest of this paper is organized as follows. The whole system is introduced in Section 2. Details of the proposed methods are explained in Section 3. Finally, the experiment and conclusion are in Section 4 and Section 5, respectively.

2. Framework

2.1. 2D Pose Estimation
The work first obtains the 2D pixel values of the human joints through the multi-person 2D pose estimation method OpenPose [7] as shown in Figure 2. This approach uses a nonparametric representation, which is referred to as part affinity fields (PAFs), to associate body parts with individuals in the image. The architecture encodes the global context, allowing a greedy bottom-up parsing step to achieve a real-time performance. The architecture is designed to jointly learn part locations and their association.

2.2. Camera Calibration and Binocular Stereo Reconstruction

In order to determine the relationship between the 3D geometric position of the figure skater’s joint spatial point and its corresponding 2D point in the image, it is necessary to establish a geometric model of camera imaging. These geometric model parameters are camera parameters, and the process of solving the parameters is called camera calibration. The accuracy of the calibration result and the stability of the algorithm directly affect the quality of the results. The solution process of the camera calibration matrix is as follows: z is an unknown scale factor which corresponds to the depth; u and v are the pixel coordinate values; X, Y and Z are the spatial coordinate values; and   M 3 X 4 is the camera calibration matrix.
z [ u v 1 ] = M 3 X 4 [ X Y Z 1 ]
Binocular stereo vision mimics human eyes to obtain 3D information and consists of two cameras. The two cameras form a triangular relationship with the measured object in space. As shown in Figure 3b, the spatial coordinate can be obtained according to the calibration matrix and the human joints’ pixel value of the two camera planes.

2.3. Dataset

As shown in Figure 3a, which is a standard figure skating venue, with six cameras placed every 60 degrees within the auditorium, the visual fields of these cameras cover the red area simultaneously. The resolution of sequences is 1920 times 1080, and the frame rate is 60 frames per second (fps). Synchronized images which contain the target from six perspectives are available at any moment.

3. Proposed Methods

3.1. Mutational Point Correction

Due to the particularity of figure skating, such as the restrictions of figure skater’s clothing, the self- occlusion, abnormal pose, large venues and other factors, accurate 2D detection results cannot be obtained even after an advanced 2D human pose estimation method [7]. Proposal 1 will mainly solve the problems at the 2D level as shown in Figure 4. The errors in 2D information can be roughly divided into three types of mutations: the left and right are reversed, the wrong identification and the sudden appearance and disappearance of the joints.
Except for the key points in the face and neck, all the key points have left and right sides. All three kinds of mutation points are first judged with the threshold value of the previous frame. Here, the threshold is defined as a circle with the key point of the previous frame as the center and a certain pixel value as the radius. If the threshold is exceeded, the values of the left and right sides will be exchanged. Thus, the mutation caused by a left–right inversion can be corrected. Then, perform the second threshold judgment where wrong identification can be detected. As shown in Figure 5b, the identification of these errors can be judged as a very serious error, which is difficult to correct. If an inappropriate correction method is used, it will lead to error superposition and introduce a larger error, so discard the serious error. Due to the figure skater’s self-occlusion, the key points will appear and disappear suddenly, which is difficult to correct at the 2D level, so correct it at the 3D level.

3.2. Reconstructed Point Selection

Six cameras are used to capture the dataset. In theory, according to the stereo binocular reconstruction principle, C 6 2 spatial points will be reconstructed because the principle of reconstruction requires a clear triangular relationship between two different cameras. Therefore, it is necessary to discard the camera combinations which cannot form a distinct triangle with the target athlete such as 0-1, 2-5 and 3-4. For reconstructed points which can be used, if any two are not equal, the weighted average will be calculated. Otherwise, use majority rules to get the accurate one. The weighted average (w) is calculated as follows: S is the reconstruction point, η is the weight of S and the value of η is determined according to the relative displacement from the previous frame.
w = η 1 S 1 + η 2 S 2 + + η n S n
η = 30 ( x t - x t - 1 ) 2 + ( y t - y t - 1 ) 2 + ( z t - z t - 1 ) 2

3.3. Inaccurate Point Correction

In order to correct inaccurate reconstructed points, it is necessary to utilize the smoothness of the spatial trajectory. Figure 6a shows the spatial trajectory of the skater’s left elbow. As shown in Figure 6b, the red dot is the key point of the first two frames, and the gray is the current reconstructed point. A straight line can be determined according to the key points of the previous two frames. If the reconstructed point of the current frame is far away from this line, the position of the current frame’s key point needs to be predicted on the line.

4. Experiment

The experiment is run on the test videos which contain 51 groups of jumps in total. The test videos are six corner perspectives of the figure skating venue. The test dataset contains 23 groups of 1 turn jump, 10 groups of 2 turn jumps, 15 groups of 3 turn jumps and 3 groups of a falling jump. For the software environment, the proposed method is implemented on C++ and OpenCV 3.4.1.
The definition of a successful frame is that if comparing the experimental results with the reference results, the 14 joint points coincide within a certain error range, then it is considered a successful frame, as shown in Figure 7a. The allowable range of error is defined by the fact that the experimental results do not exceed the sphere with a radius of 30 mm centered on the reference point, as shown in Figure 7b.
Figure 7. The quality grade of results is determined by comparing them with the reference spatial points which are reconstructed from the manually labeled 2D pixel values. (a) Evaluation method; (b) the allowable range of error; (c) experimental result. The quality grade of results is defined as
Figure 7. The quality grade of results is determined by comparing them with the reference spatial points which are reconstructed from the manually labeled 2D pixel values. (a) Evaluation method; (b) the allowable range of error; (c) experimental result. The quality grade of results is defined as
Proceedings 49 00095 g007
quality   grade = successful   frames total   frames

5. Conclusions

This work has developed a system to obtain the 3D jump pose of a figure skater. At the core of the approach, this method corrects inaccurate or even erroneous reconstruction results by combining spatial-temporal information and a multi-perspective during the process of 2D-to-3D pose transformation. The proposed system outperforms previous 3D pose estimation in terms of spatial the coordinates’ authenticity. The quality grade of the experimental result is 87.25%, based on the test sequences which have different types of jumps in figure skating. For future work, we plan to change the network of pose estimation as needed, adding key points such as hands and feet, at the same time as improving the algorithm to realize real-time. After these modifications, the system can be applied to the objective performance evaluation of figure skaters and the real-time display of figure skating TV broadcasts to enhance the audience entertainment experience.

Acknowledgments

This work was supported by Waseda University Grant for Special Research Projects (2019Q-055).

References

  1. Ionescu, C.; Papava, D.; Olaru, V.; Sminchisescu, C. Human3.6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. TPAMI 2014, 36, 1325–1339. [Google Scholar] [CrossRef] [PubMed]
  2. Mehta, D.; Rhodin, H.; Casas, D.; Fua, P.; Sotnychenko, O.; Xu, W.P.; Theobalt, C. Monocular 3d human pose estimation in the wild using improved cnn supervision. In Proceedings of the International Conference on 3DVision, Qingdao, China, 10–12 October 2017. [Google Scholar] [CrossRef]
  3. Shotton, J.; Fitzgibbon, A.; Cook, M.; Sharp, T.; Finocchio, M.; Moore, R.; Kipman, A.; Blake, A. Realtime human pose recognition in parts from single depth images. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 20–25 June 2011; pp. 1297–1304. [Google Scholar]
  4. Pavllo, D.; Feichtenhofer, C.; Grangier, D.; Auli, M. 3D human pose estimation in video with temporal convolutions and semi-supervised training. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 7753–7762. [Google Scholar]
  5. Pavlakos, G.; Zhou, X.W.; Derpanis, K.G.; Daniilidis, K. Coarse-to-fine volumetric prediction for single-image 3d human pose. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 7025–7034. [Google Scholar]
  6. Yang, W.; Ouyang, W.L.; Wang, X.L.; Ren, J.; Li, H.S.; Wang, X.G. 3d human pose estimation in the wild by adversarial learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 5255–5264. [Google Scholar]
  7. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime multi-person 2d pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 7291–7299. [Google Scholar]
Figure 1. A frame belonging to the jump and the corresponding 3D pose.
Figure 1. A frame belonging to the jump and the corresponding 3D pose.
Proceedings 49 00095 g001
Figure 2. 2D information from OpenPose.
Figure 2. 2D information from OpenPose.
Proceedings 49 00095 g002
Figure 3. Six cameras are placed in the audience. The camera’s field of view is centered on the red dotted frame. Any two synchronized cameras which can form a distinct triangle with the athlete can reconstruct a spatial point. (a) The size of the venue and the location of camera; (b) binocular stereo reconstruction.
Figure 3. Six cameras are placed in the audience. The camera’s field of view is centered on the red dotted frame. Any two synchronized cameras which can form a distinct triangle with the athlete can reconstruct a spatial point. (a) The size of the venue and the location of camera; (b) binocular stereo reconstruction.
Proceedings 49 00095 g003
Figure 4. Proposal 1 introduced in 3.1 is applied at the 2D level. Proposal 2 in 3.2 and proposal 3 in 3.3 are applied at the 3D level.
Figure 4. Proposal 1 introduced in 3.1 is applied at the 2D level. Proposal 2 in 3.2 and proposal 3 in 3.3 are applied at the 3D level.
Proceedings 49 00095 g004
Figure 5. In order to get good reconstruction results, the system should first correct the errors as much as possible at the 2D level. (a) Left–right inversion; (b) wrong identification; (c) sudden appearance and disappearance; (d) the flowchart of proposal 1; (e) threshold.
Figure 5. In order to get good reconstruction results, the system should first correct the errors as much as possible at the 2D level. (a) Left–right inversion; (b) wrong identification; (c) sudden appearance and disappearance; (d) the flowchart of proposal 1; (e) threshold.
Proceedings 49 00095 g005
Figure 6. For the 3D result of the current frame, it is necessary to judge whether it is accurate. An inaccurate spatial point requires a suitable predictive value. (a) The spatial trajectory of the left elbow; (b) determination and prediction of key points.
Figure 6. For the 3D result of the current frame, it is necessary to judge whether it is accurate. An inaccurate spatial point requires a suitable predictive value. (a) The spatial trajectory of the left elbow; (b) determination and prediction of key points.
Proceedings 49 00095 g006
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tian, L.; Cheng, X.; Honda, M.; Ikenaga, T. Multi-Technology Correction Based 3D Human Pose Estimation for Jump Analysis in Figure Skating. Proceedings 2020, 49, 95. https://doi.org/10.3390/proceedings2020049095

AMA Style

Tian L, Cheng X, Honda M, Ikenaga T. Multi-Technology Correction Based 3D Human Pose Estimation for Jump Analysis in Figure Skating. Proceedings. 2020; 49(1):95. https://doi.org/10.3390/proceedings2020049095

Chicago/Turabian Style

Tian, Limao, Xina Cheng, Masaaki Honda, and Takeshi Ikenaga. 2020. "Multi-Technology Correction Based 3D Human Pose Estimation for Jump Analysis in Figure Skating" Proceedings 49, no. 1: 95. https://doi.org/10.3390/proceedings2020049095

APA Style

Tian, L., Cheng, X., Honda, M., & Ikenaga, T. (2020). Multi-Technology Correction Based 3D Human Pose Estimation for Jump Analysis in Figure Skating. Proceedings, 49(1), 95. https://doi.org/10.3390/proceedings2020049095

Article Metrics

Back to TopTop