Next Article in Journal
Assessing the Design of Interactive Radial Data Visualizations for Mobile Devices
Next Article in Special Issue
Lithium Metal Battery Quality Control via Transformer–CNN Segmentation
Previous Article in Journal
Multispectral Satellite Image Analysis for Computing Vegetation Indices by R in the Khartoum Region of Sudan, Northeast Africa
Previous Article in Special Issue
Novel Light Convolutional Neural Network for COVID Detection with Watershed Based Region Growing Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Future Prediction of Shuttlecock Trajectory in Badminton Using Player’s Information

Graduate School of Science and Technology, Keio University, Yokohama 223-8852, Japan
*
Author to whom correspondence should be addressed.
J. Imaging 2023, 9(5), 99; https://doi.org/10.3390/jimaging9050099
Submission received: 27 March 2023 / Revised: 30 April 2023 / Accepted: 8 May 2023 / Published: 11 May 2023
(This article belongs to the Special Issue Computer Vision and Deep Learning: Trends and Applications)

Abstract

:
Video analysis has become an essential aspect of net sports, such as badminton. Accurately predicting the future trajectory of balls and shuttlecocks can significantly benefit players by enhancing their performance and enabling them to devise effective game strategies. This paper aims to analyze data to provide players with an advantage in the fast-paced rallies of badminton matches. The paper delves into the innovative task of predicting future shuttlecock trajectories in badminton match videos and presents a method that takes into account both the shuttlecock position and the positions and postures of the players. In the experiments, players were extracted from the match video, their postures were analyzed, and a time-series model was trained. The results indicate that the proposed method improved accuracy by 13% compared to methods that solely used shuttlecock position information as input, and by 8.4% compared to methods that employed both shuttlecock and player position information as input.

1. Introduction

Recently, computer vision technologies have been employed to automate the analysis of video clips from net sports matches, such as tennis, volleyball, and badminton. These technologies enable player pose detection [1] and ball detection/tracking [2], allowing for the extraction of crucial information from match videos. This information can then be used to determine the high-level context, such as the players’ actions, during the match.
Predicting the future movement of the ball, shuttlecock, and opponent is a critical aspect of sports video analysis. The ability to forecast these movements during a rally can give players a significant advantage over their competitors. Many players rely on their experience to make these predictions, and in fast-paced sports such as badminton, being able to anticipate the movement of the shuttlecock even a fraction of a second ahead can mean the difference between winning and losing the match. Performance analysis of players can also be considered so that players can understand strategies in matches. If a player’s performance aligns with the prediction made by the proposed model, it implies that the shuttlecock’s trajectory may be easily predicted by the competing player in the match. Players aim to play unpredictably to win the match.
Most research focused on predicting future movements in net sports has centered around predicting the landing point of the ball or shuttlecock [3,4,5,6,7,8]. However, in badminton, the shuttlecock must be hit without bouncing and from a higher, faster forward position, making the prediction of the landing point insufficient for gaining an advantage in the game. To achieve this, it is necessary to predict the shuttlecock’s trajectory. Currently, trajectory prediction studies are limited to short-term events, such as the serve in table tennis or the toss in volleyball, and have not yet been applied to rallies in badminton.
This paper presents a method for predicting the future trajectory of the badminton shuttlecock during a match. One of the simple methods utilizes sequential models, such as recurrent neural networks (RNNs) [9], to model the motion of the shuttlecock by inputting previous shuttlecock trajectories and outputting future shuttlecock trajectories. To better reflect a player’s decision-making process during a match, the proposed method also considers the player’s position and posture information in addition to the shuttlecock’s position information.
The proposed method was evaluated using the shuttlecock trajectory dataset [10] and demonstrated improved accuracy compared to methods that utilized only shuttlecock position information or methods that employed both shuttlecock and player position information. The results of the study can be summarized as follows:
  • This is a pioneering study on predicting the trajectory of the badminton shuttlecock during a match.
  • The proposed method predicts the shuttlecock’s trajectory by considering the player’s position and posture information, in addition to the shuttlecock’s position information.
  • The results of the experiments show that the proposed method outperforms previous methods that use only shuttlecock position information as input or methods that use both shuttlecock and player position information as input.

2. Materials and Methods

2.1. Related Work

2.1.1. Future Predictions in Net Sports

Recently, research on future prediction in various net sports, such as tennis, volleyball, table tennis, and badminton, has been growing in popularity. This section introduces previous research in the field and compares them with the current research. As shown in Table 1, previous research on future prediction in net sports has focused on predicting shot direction, landing point, stroke, and trajectory. A more detailed discussion of previous research in each sport is presented below.
In tennis, Shimizu et al. [11] were the first to predict the future shot direction in three categories: right cross, left cross, and straight. They did so by using the player’s continuous position and posture information up until the moment the ball was hit. They also developed a new dataset with shot directions for their study. However, in badminton, predicting only the direction is insufficient, as players have different movements for low and high trajectories.
In table tennis, Wu et al. [4,5] predicted the landing point of the service by using the player’s motion information up until the moment just before hitting the ping-pong ball. This information was obtained through pose estimation. In volleyball, Sato et al. [6] predicted the landing point of the ball by using its velocity and 3D position, with an average error of about 0.3 m and about 1.5 s before the ball hit the floor. In badminton, Waghmare et al. [3] calculated the speed and direction of the shuttlecock to predict its landing point, using a two-dimensional laser scanner. While these methods help the player reach the landing point more quickly, they are not enough to give the player an advantage in the game by allowing them to hit the shuttlecock back faster and higher.
Stroke prediction involves predicting both the shot type and landing point. In tennis, Fernando et al. [7] predicted strokes by using a semi-supervised generative adversarial network (SGAN) [14], which combined a memory model with the automatic feature learning capabilities of a deep neural network. In badminton, Wang et al. [8] utilized a network called shuttleNet to predict the next stroke based on the current stroke. This was the first study to address stroke prediction in sports. While stroke prediction provides a better prediction of the shuttlecock’s trajectory than landing point prediction, it still falls short compared to trajectory prediction.
In volleyball, Suda et al. [12] predicted the trajectory of the toss 0.3 s before it actually occurred, using the setter player’s 3D joint positions. In table tennis, Lin et al. [13] predicted the trajectory of a subsequent serve by using a dual-network method. The method involved learning two separate parabolic trajectories: one from the service point to the landing point on the table (parabola 1) and one from the landing point to the hitting point (parabola 2). While trajectory prediction has been studied in both volleyball and table tennis, it has not yet been adequately studied in badminton.
All of the previous research has been conducted in recent years, and research on future predictions in net sports is still in its developmental stage. While shot direction prediction in tennis by Shimizu et al. [11] and stroke prediction in badminton by Wang et al. [8] have been explored, trajectory prediction has not been thoroughly studied, particularly in the case of badminton. Therefore, the aim of the current study is to predict the shuttlecock’s trajectory in badminton.

2.1.2. Object Detection

Object detection is the task of identifying objects with specific attributes within an image or video and determining their location and extent by surrounding them with a bounding box.
There are two types of deep learning-based object detection methods: one-stage methods [15,16,17] that directly detect the target object from the input image, and two-stage methods [18,19,20] that first identify candidate regions in the input image and then perform detailed detection for each region. One-stage methods, such as YOLOv4 [17], prioritize processing speed and are suitable for real-time applications. Two-stage methods, such as Region-CNN (R-CNN) [18], Fast R-CNN [19], and Faster R-CNN [20], may have a slower processing speed compared to one-stage methods, but they offer higher detection accuracy. For this study, Faster R-CNN is used, as it is one of the highest-performing two-stage methods and provides more accurate information about the player’s position and posture.

2.1.3. Pose Estimation

Pose estimation is the task of identifying information about a person’s posture (eyes, nose, limbs, etc.) from an image or video showing a person. The joint points of a person are obtained as keypoints.
There are two types of deep learning-based pose estimation methods: the top-down method [1,21,22,23,24,25,26,27], which performs pose estimation for each object identified after object detection in the image, and the bottom-up method [28,29,30,31,32,33], which first performs pose estimation for all objects in the image by connecting each keypoint to other objects of the same type. The former method tends to be more accurate because it estimates the posture of each object one by one, while the latter method is less accurate because it is challenging to learn to connect keypoints between the same objects. In this method, object detection and pose estimation are independent of each other in the top-down method, and HRNet [1] is applied to pose estimation after object detection.

2.2. Method

2.2.1. Overview

This paper proposes a method for predicting the trajectory of a shuttlecock over the next (future) n f frames, based on the shuttlecock position, the positions, and postures of the two competing players detected during the previous (past) n p frames. The proposed method consists of two modules: a detection and pose estimation module for past frames, and a time series model module for future prediction, as illustrated in Figure 1.
The first module takes a sequence of past frames from badminton match videos as input and performs shuttlecock detection, player detection, and player pose estimation for each image.
In the second module, the dimensions of the shuttlecock position, player position, and player posture information are each aligned in two dimensions. The six-dimensional information is then combined and used as input for trajectory prediction.

2.2.2. Detection and Pose Estimation from Past Frames

The shuttlecock position information is obtained from a shuttlecock detector, such as TrackNet [34]. These are two-dimensional coordinates x s and y s on the image.
For player detection and pose estimation, the MMPose framework [35] is utilized, which includes several pose estimation models and pre-trained models. An object detection model detects players and records their positions as bounding boxes, while a pose estimation model estimates the posture of the players and records their keypoints.
To detect humans, an object detector is utilized; we employ Faster R-CNN [20] trained on the Microsoft Common Objects in Context (MS COCO) dataset [36]. The detector provides bounding boxes and confidence scores as detection results. Bounding boxes are represented by four two-dimensional coordinate points on the image when a human is enclosed by a rectangle. The confidence score ranges from 0 to 1 and indicates the likelihood that the object within the detected bounding box is a human. By using the confidence score, only players, not referees or spectators, are detected. Referees and spectators that are detected alongside players have a lower confidence score, as they may be sitting, have only their faces in the image, be facing sideways, or appear small. Thus, the person with the highest confidence score is identified as the player, and the player’s bounding box is obtained, as shown in Figure 2.
The player position information is obtained by calculating the normalized center coordinates ( x p l a y e r , y p l a y e r ) of the player with each coordinate in Figure 3 as follows:
( x p l a y e r , y p l a y e r ) = ( b b o x l e f t + b b o x r i g h t ) / 2 i m g w i d t h , ( b b o x t o p + b b o x b o t t o m ) / 2 i m g h e i g h t .
The players are distinguished by assigning a number to each of them in the image, starting with the player on the lower side. The player position information is represented as four-dimensional coordinates ( x p 1 , y p 1 , x p 2 , y p 2 ) , where x p 1 represents the x coordinate of the player who is shown on the lower side of the image.
After the player detection, HRNet [1], which has been trained by MS COCO, is used as the pose estimator. As shown in Figure 4, 17 joints (eyes, ears, nose, shoulders, elbows, wrists, hips, knees, and ankles) are detected as keypoints in the bounding box obtained from player detection.
The coordinate values are expressed as absolute coordinates on the image. For example, if the coordinates of a keypoint are ( x k , y k ) on the image, the normalized absolute coordinates ( a x k , a y k ) of the keypoint are calculated using the values in Figure 3 as follows:
( a x k , a y k ) = x k w , y k h ,
where w , and h are the width and height of the input image, respectively. The posture information of the two players is represented as a 68-dimensional feature vector ( x p 1 k 1 , y p 1 k 1 , , x p 2 k 17 , y p 2 k 17 ) , where x p i k j represents the x coordinate of the first keypoint of the player who is shown on the lower side of the image.

2.2.3. Time Series Model for Future Prediction

The proposed method predicts the trajectory by inputting three types of information: the shuttlecock position information obtained by the shuttlecock detector, the player position information obtained by the object detector, and the player posture information obtained by the pose estimation.
At this point, the shuttlecock position information is 2-dimensional, the player position information is 4-dimensional, and the player posture information is 68-dimensional. In order to make three types of information into a feature vector of the same space, each of them is embedded into two dimensions by a fully connected layer as shown in Table 2. The shuttlecock position information and the player position information and the player posture information are combined to form a 6-dimensional feature vector.
The combined information is fed into a long short-term memory (LSTM) network [37], which is the second module used for predicting the shuttlecock trajectory. The LSTM network uses multiple past inputs stacked together as its input, and its output is then forwarded to the fully connected layer for further processing.

2.3. Experiment

2.3.1. Dataset

The shuttlecock trajectory dataset [10] was utilized, which was created for training and testing the TrackNet [34] and TrackNetV2 [38] models for badminton applications. This dataset comprises 26 match videos for training and 3 match videos for testing. The match videos have a resolution of 1280 × 720 and a frame rate of 30 fps, and are separated by rallies, which refer to recordings that start with a serve and end with a score.
Each frame in the dataset provides information about the shuttlecock’s position and the moment it hits the racket. Twenty-three matches were used for training, excluding amateur matches, while three matches were used for testing. The professional matches used in this study were singles matches held in international tournaments between 2018 and 2021. The 23 match rally videos for training were randomly split into an 80% training set and a 20% validation set. The three-match rally videos for testing were used as the test set.
Data cleansing was carried out on the dataset to enhance prediction accuracy. The position coordinates of the shuttlecock in the dataset were set to ( 0 , 0 ) whenever it was hidden by a person or not visible, which could negatively affect learning since the shuttlecock would appear to move unnaturally in the frames before and after it. Frames with such issues were removed from the dataset to ensure that consecutive frames with unnatural shuttlecock movements were not used for learning.
Data augmentation was also carried out to increase the diversity of the training data. Since flipping a badminton match video upside down would no longer be appropriate for the sport, only left–right flipping and translations were performed. The original image was flipped left and right with a probability of 50% and then translated to the right or bottom in a range of 0 to 50 pixels relative to the width and height of the image. Figure 5 shows the results of the data augmentation.

2.3.2. Evaluation Metrics

Two types of displacement errors, namely the average displacement error (ADE) and final displacement error (FDE), were used as evaluation metrics in this experiment. ADE is the average of errors across all output frames, while FDE is the error at the final point of the output trajectory. The Euclidean distance was calculated using the two-dimensional coordinates in a 1280 × 720 pixel image, and the unit of measurement is pixels. ADE is more important in this task since it predicts the trajectory for multiple frames rather than just the landing point.

2.3.3. Network Training

We implemented the proposed method in PyTorch [39] (1.12.1+cu102, with Python 3.7.13) and ran it on the NVIDIA TITAN RTX processing unit using CUDA 11.4. For LSTM, the number of layers was set to 3, the hidden layer to 128 dimensions, and the network was optimized using Adam [40], with a weight decay of 10 4 , the momentums β 1 = 0.5 and β 2 = 0.999 , and a learning rate of 0.02. The model was trained for 400 epochs with four input frames and 12 output frames for all cases.
The mean squared error (MSE) was employed as the loss function:
L o s s = 1 D i = 1 D ( F ( x ) i g t i ) 2 ,
where F is the time-series model LSTM of the proposed method, x is its input data and g t is the ground-truth data of the output shuttlecock position. F ( x ) and g t are D-dimensional vectors and F ( x ) i is the value of F ( x ) on the i dimension. The output results were compared with the ground-truth data to calculate the error. Then, the parameters were updated to reduce the error using backpropagation.

2.3.4. Other Models

Baseline Models

To verify the effectiveness of the proposed method, two baseline models were set up. One is to input only the shuttlecock position information without any player information, and the other is to input the shuttlecock and player position information without any player posture information.

Other Time-Series Models

LSTM was used as the model for trajectory prediction, but four other time-series models were also examined: an RNN [9], gated recurrent unit (GRU) [41], Transformer [42] and sequence-to-sequence learning with neural networks (Seq2Seq) [43]. The network parameters were set the same as in the proposed method model for all time-series models, except for the learning rate. Only the learning rate was set as shown in Table 3.

Other Representations of Posture Information

In addition to the absolute coordinate values on the image used in this method, we also examined other ways of representing the posture of the players: the relative joint positions in the bounding box and the heat map generated for the pose estimation.
If the coordinates of a keypoint are ( x k , y k ) on the image, the normalized relative coordinates ( r x k , r y k ) to the bounding box are calculated using the values in Figure 3 as follows:
( r x k , r y k ) = x k b b o x l e f t b b o x r i g h t b b o x l e f t , y k b b o x t o p b b o x b o t t o m b b o x t o p .
The accuracy of three inputs to the time-series model was compared: absolute coordinate values on the image, relative coordinate values to the bounding box, and heatmap generated by pose estimation, respectively.
When using absolute coordinates on the image, the spatial coordinates are the same as the position information of the shuttlecock and players. When using relative coordinates to the bounding box, the spatial coordinates are different from the position information of the shuttlecock and players. Relative coordinates have a more significant influence on the posture information than absolute coordinates because the change in posture is greater with relative coordinates. Figure 6a shows the overview in the case of using the joint positions to represent the postures of the players.
When using the heatmap, a heatmap was first generated for each of the two players in the pose estimation part and superimposed, as shown in Figure 7. Then, the heatmap including the posture information for the two players was sent to ResNet-18 [44] to extract a 512-dimensional feature vector, which was input to the time-series model together with the shuttlecock position and the players’ positions. Figure 6b shows the overview in the case of using the heatmap to represent the postures of the players.

2.3.5. Data Augmentation Values

Experiments were conducted for data augmentation with different probabilities of left–right flipping and ranges of parallel shifts. The probability of left–right reversal was set to 0%, 25%, 50%, or 75%, where 0% means that the image is not flipped. The range of translation was set to 0 pixels, 50 pixels, or 100 pixels, where 0 means no translation.

2.3.6. The Number of Frames of Past/Future

Experiments were conducted with the number of past and future frames fixed at 4 and 12, respectively. For the results shown in Section 3.3, five different combinations of the number of past and future frames were tested by increasing the number of past frames from 4 to 12, while keeping the total number of past and future frames at 16.

3. Results

3.1. Comparison with Other Models

As for the quantitative evaluation, the results of comparing this method with the baseline methods and the methods using other time-series models are shown in Table 4 and Table 5. As a qualitative evaluation, the results of comparing this method with the two baseline methods are shown in Figure 8, and the results of comparing this method with methods using other time-series models are shown in Figure 9.
The proposed method shows the best results for ADE and FDE. It also qualitatively predicts the trajectories closest to the ground truth.
Compared to the method with the baseline methods, the proposed method improves ADE by about 13% compared to the method using only the shuttlecock position information as input, and by about 8.4% compared to the method using the shuttlecock and player position information as input. This shows that the proposed method effectively uses all the information on the shuttlecock position, player position, and player posture. Compared to the method with the highest accuracy for each time-series model, the proposed method using LSTM improves ADE by about 9.8%, 5.0%, 20%, and 12% compared to the methods using RNN, GRU, Transformer, and Seq2Seq, respectively. Therefore, it is shown that the proposed method using LSTM is the best model among the time-series models considered.

3.2. Comparison by Data Augmentation Values

The results of the experiments with different probabilities of left–right flipping and different ranges of translation are shown in Table 6 and Table 7. The proposed method shows the best results for ADE and FDE. This indicates that a 50% probability of left–right flipping and a 50-pixel range of translation are appropriate among those considered in this study, and that the data augmentation is effective.

3.3. Comparison of the Number of Frames of Past/Future

The accuracy improves as the number of input frames increases and the number of frames to be predicted decreases as shown in Figure 10, Table 8 and Table 9.

4. Discussion and Conclusions

4.1. Limitation

This study proposed a method for predicting future trajectories in badminton match videos. However, several limitations exist in the framework.
The first limitation is the position of the cameras. The dataset used in this study includes high-level tournaments, such as the World Championships, all of which were recorded from a distance so that the entire court could be seen. This makes it easy to obtain information on shuttlecocks and players. However, recording from a distance requires equipment, such as a tripod and a photographer, making it difficult to set up a camera when attending a match alone or when there is no access to a suitable recording location. In addition, since the camera is recording from a distance, it is likely that people may pass in front of the camera. If there are many frames where the shuttlecock is not in the angle of view and cannot be tracked, it would be difficult to predict.
Second is the adaptation to sudden changes in trajectory. As shown in Figure 11, if the trajectory changes at the moment of hitting back or the moment the shuttlecock touches the floor, it is quite difficult to predict. Even if it could be predicted that the trajectory would change after bouncing back, the direction of the change could be wrong. In this method, the position and posture information of the player was input in addition to the shuttlecock’s position information, which enabled some frames to adapt to changes in trajectory, but there were still many frames that were difficult to predict.

4.2. Future Work

There are two things that need to be worked on in the future to make this research work with in-the-wild data.
The first is to improve the generalization performance of the model. For the practical scenario, it is necessary to confirm that the model is general enough by testing the proposed model on multiple data sets in badminton (e.g., those with different camera locations and those containing matches of amateur players) and other net sports, such as tennis and table tennis. For this purpose, it would also be useful to be able to automate the rally segmentation and shuttlecock position detection in match videos.
The second is to perform trajectory prediction in 3D space. Since this research is based on the novel task of performing trajectory prediction in badminton, we performed the prediction in 2D space as a first step. If we can further develop this task to predict in 3D space and project the predicted trajectory into real space, it can be used for training to predict the trajectory of a shuttlecock hit by an opponent or to prevent an opponent from predicting the trajectory of a shuttlecock hit by oneself. This can further help players improve their skills.

4.3. Conclusions

This paper approached the novel task of predicting the trajectory of the shuttlecock in a badminton match video and proposed a trajectory prediction method that uses information about the shuttlecock’s position and the players’ positions and the players’ postures. Experiments comparing the proposed method with the baseline method were conducted to confirm the effectiveness of the proposed method. Furthermore, experiments with different time-series models show that the LSTM used in this method achieves the highest accuracy.

Author Contributions

Conceptualization, Y.N., R.H. (Ryo Hachiuma) and H.S.; methodology, Y.N., R.H. (Ryo Hachiuma), R.H. (Ryosuke Hori) and H.S.; software, Y.N., R.H. (Ryo Hachiuma) and R.H. (Ryosuke Hori); validation, Y.N.; formal analysis, Y.N.; investigation, Y.N.; resources, H.S.; data curation, Y.N.; writing—original draft preparation, Y.N.; writing—review and editing, Y.N., R.H. (Ryo Hachiuma), R.H. (Ryosuke Hori) and H.S.; visualization, Y.N.; supervision, H.S.; project administration, R.H. (Ryo Hachiuma) and H.S.; funding acquisition, H.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep High-Resolution Representation Learning for Human Pose Estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 5686–5696. [Google Scholar] [CrossRef]
  2. Cao, Z.; Liao, T.; Song, W.; Chen, Z.; Li, C. Detecting the shuttlecock for a badminton robot: A YOLO based approach. Expert Syst. Appl. 2021, 164, 113833. [Google Scholar] [CrossRef]
  3. Waghmare, G.; Borkar, S.; Saley, V.; Chinchore, H.; Wabale, S. Badminton shuttlecock detection and prediction of trajectory using multiple 2 dimensional scanners. In Proceedings of the IEEE First International Conference on Control, Measurement and Instrumentation (CMI), Kolkata Section, India, 8–10 January 2016; pp. 234–238. [Google Scholar] [CrossRef]
  4. Wu, E.; Perteneder, F.; Koike, H. Real-time Table Tennis Forecasting System based on Long Short-term Pose Prediction Network. In Proceedings of the SIGGRAPH Asia Posters (SA), Brisbane, Australia, 17–20 November 2019; pp. 23:1–23:2. [Google Scholar] [CrossRef]
  5. Wu, E.; Koike, H. FuturePong: Real-time table tennis trajectory forecasting using pose prediction network. In Proceedings of the Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems (CHI EA), Honolulu, HI, USA, 25–30 April 2020; pp. 1–8. [Google Scholar] [CrossRef]
  6. Sato, K.; Sano, Y.; Otsuki, M.; Oka, M.; Kato, K. Augmented Recreational Volleyball Court: Supporting the Beginners’ Landing Position Prediction Skill by Providing Peripheral Visual Feedback. In Proceedings of the 10th Augmented Human International Conference (AH), Reims, France, 11–12 March 2019; Wolf, K., Zhang, H., Taïar, R., Seigneur, J.M., Eds.; ACM: New York, NY, USA, 2019; pp. 15:1–15:9. [Google Scholar]
  7. Fernando, T.; Denman, S.; Sridharan, S.; Fookes, C. Memory Augmented Deep Generative models for Forecasting the Next Shot Location in Tennis. IEEE Trans. Knowl. Data Eng. 2019, 32, 1785–1797. [Google Scholar] [CrossRef]
  8. Wang, W.Y.; Shuai, H.H.; Chang, K.S.; Peng, W.C. ShuttleNet: Position-Aware Fusion of Rally Progress and Player Styles for Stroke Forecasting in Badminton. In Proceedings of the AAAI Conference on Artificial Intelligence (AAAI), Virtual, 22 February–1 March 2022; Volume 36, pp. 4219–4227. [Google Scholar] [CrossRef]
  9. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  10. Ik, T.U. Shuttlecock Trajectory Dataset. 2020. Available online: https://hackmd.io/Nf8Rh1NrSrqNUzmO0sQKZw (accessed on 1 January 2020).
  11. Shimizu, T.; Hachiuma, R.; Saito, H.; Yoshikawa, T.; Lee, C. Prediction of Future Shot Direction using Pose and Position of Tennis Player. In Proceedings of the 2nd International Workshop on Multimedia Content Analysis in Sports (MMSports), Nice, France, 25 October 2019; pp. 59–66. [Google Scholar] [CrossRef]
  12. Suda, S.; Makino, Y.; Shinoda, H. Prediction of Volleyball Trajectory Using Skeletal Motions of Setter Player. In Proceedings of the 10th Augmented Human International Conference (AH), Reims, France, 11–12 March 2019; pp. 16:1–16:8. [Google Scholar] [CrossRef]
  13. Lin, H.I.; Yu, Z.; Huang, Y.C. Ball Tracking and Trajectory Prediction for Table-Tennis Robots. Sensors 2020, 20, 333. [Google Scholar] [CrossRef] [PubMed]
  14. Denton, E.L.; Gross, S.; Fergus, R. Semi-Supervised Learning with Context-Conditional Generative Adversarial Networks. arXiv 2016, arXiv:1611.06430. [Google Scholar]
  15. Redmon, J.; Divvala, S.K.; Girshick, R.B.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE Computer Society: New York, NY, USA, 2016; pp. 779–788. [Google Scholar]
  16. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  17. Bochkovskiy, A.; Wang, C.; Liao, H.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  18. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar] [CrossRef]
  19. Girshick, R. Fast R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  20. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 7–12 December 2015; Volume 28, pp. 91–99. [Google Scholar]
  21. Toshev, A.; Szegedy, C. DeepPose: Human Pose Estimation via Deep Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; IEEE Computer Society: New York, NY, USA, 2014; pp. 1653–1660. [Google Scholar]
  22. Tompson, J.; Jain, A.; LeCun, Y.; Bregler, C. Joint Training of a Convolutional Network and a Graphical Model for Human Pose Estimation. In Proceedings of the Conference on Neural Information Processing System (NIPS), Montreal, QC, Canada, 8–13 December 2014; Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N.D., Weinberger, K.Q., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2014; pp. 1799–1807. [Google Scholar]
  23. Wei, S.E.; Ramakrishna, V.; Kanade, T.; Sheikh, Y. Convolutional Pose Machines. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE Computer Society: New York, NY, USA, 2016; pp. 4724–4732. [Google Scholar]
  24. Newell, A.; Yang, K.; Deng, J. Stacked Hourglass Networks for Human Pose Estimation. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 483–499. [Google Scholar]
  25. Chen, Y.; Wang, Z.; Peng, Y.; Zhang, Z.; Yu, G.; Sun, J. Cascaded Pyramid Network for Multi-Person Pose Estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; IEEE Computer Society: New York, NY, USA, 2018; pp. 7103–7112. [Google Scholar]
  26. Xiao, B.; Wu, H.; Wei, Y. Simple Baselines for Human Pose Estimation and Tracking. In Lecture Notes in Computer Science, Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer: Berlin/Heidelberg, Germany, 2018; Volume 11210, pp. 472–487. [Google Scholar]
  27. Duan, H.; Lin, K.Y.; Jin, S.; Liu, W.; Qian, C.; Ouyang, W. TRB: A Novel Triplet Representation for Understanding 2D Human Body. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9478–9487. [Google Scholar]
  28. Pishchulin, L.; Insafutdinov, E.; Tang, S.; Andres, B.; Andriluka, M.; Gehler, P.V.; Schiele, B. DeepCut: Joint Subset Partition and Labeling for Multi Person Pose Estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE Computer Society: New York, NY, USA, 2016; pp. 4929–4937. [Google Scholar]
  29. Insafutdinov, E.; Pishchulin, L.; Andres, B.; Andriluka, M.; Schiele, B. DeeperCut: A Deeper, Stronger, and Faster Multi-person Pose Estimation Model. In Lecture Notes in Computer Science, Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer: Berlin/Heidelberg, Germany, 2016; Volume 9910, pp. 34–50. [Google Scholar]
  30. Cao, Z.; Simon, T.; Wei, S.E.; Sheikh, Y. Realtime Multi-person 2D Pose Estimation Using Part Affinity Fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–16 July 2017; IEEE Computer Society: New York, NY, USA, 2017; pp. 1302–1310. [Google Scholar]
  31. Newell, A.; Huang, Z.; Deng, J. Associative Embedding: End-to-End Learning for Joint Detection and Grouping. In Proceedings of the Conference on Neural Information Processing System (NIPS), Long Beach, CA, USA, 4–9 December 2017; Guyon, I., von Luxburg, U., Bengio, S., Wallach, H.M., Fergus, R., Vishwanathan, S.V.N., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; pp. 2277–2287. [Google Scholar]
  32. Papandreou, G.; Zhu, T.; Chen, L.C.; Gidaris, S.; Tompson, J.; Murphy, K. PersonLab: Person Pose Estimation and Instance Segmentation with a Bottom-Up, Part-Based, Geometric Embedding Model. In Lecture Notes in Computer Science, Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer: Berlin/Heidelberg, Germany, 2018; Volume 11218, pp. 282–299. [Google Scholar]
  33. Martinez, G.H.; Raaj, Y.; Idrees, H.; Xiang, D.; Joo, H.; Simon, T.; Sheikh, Y. Single-Network Whole-Body Pose Estimation. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6981–6990. [Google Scholar]
  34. Huang, Y.C.; Liao, I.N.; Chen, C.H.; İk, T.U.; Peng, W.C. TrackNet: A Deep Learning Network for Tracking High-speed and Tiny Objects in Sports Applications. In Proceedings of the 16th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Taipei, Taiwan, 18–21 September 2019; pp. 1–8. [Google Scholar] [CrossRef]
  35. Sengupta, A.; Jin, F.; Zhang, R.; Cao, S. mm-Pose: Real-Time Human Skeletal Posture Estimation Using mmWave Radars and CNNs. IEEE Sens. J. 2020, 20, 10032–10044. [Google Scholar] [CrossRef]
  36. Lin, T.Y.; Maire, M.; Belongie, S.J.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014; Volume 8693, pp. 740–755. [Google Scholar]
  37. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  38. Sun, N.E.; Lin, Y.C.; Chuang, S.P.; Hsu, T.H.; Yu, D.R.; Chung, H.Y.; İk, T.U. TrackNetV2: Efficient Shuttlecock Tracking Network. In Proceedings of the International Conference on Pervasive Artificial Intelligence (ICPAI), Taipei, Taiwan, 3–5 December 2020; pp. 86–91. [Google Scholar] [CrossRef]
  39. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic differentiation in Pytorch. In Proceedings of the Conference on Neural Information Processing System (NIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  40. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  41. Cho, K.; van Merrienboer, B.; Gülçehre, Ç.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  42. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.u.; Polosukhin, I. Attention is All you Need. In Proceedings of the Conference on Neural Information Processing System (NIPS), Long Beach, CA, USA, 4–9 December 2017; Volume 30, pp. 5998–6008. [Google Scholar]
  43. Sutskever, I.; Vinyals, O.; Le, Q.V. Sequence to Sequence Learning with Neural Networks. In Proceedings of the Conference on Neural Information Processing System (NIPS), Montreal, QC, Canada, 8–13 December 2014; Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N.D., Weinberger, K.Q., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2014; pp. 3104–3112. [Google Scholar]
  44. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
Figure 1. Overview of the proposed method.
Figure 1. Overview of the proposed method.
Jimaging 09 00099 g001
Figure 2. Player detection.
Figure 2. Player detection.
Jimaging 09 00099 g002
Figure 3. Coordinates of image and bounding box.
Figure 3. Coordinates of image and bounding box.
Jimaging 09 00099 g003
Figure 4. Pose estimation.
Figure 4. Pose estimation.
Jimaging 09 00099 g004
Figure 5. Data augmentation. (a) Original image, (b) image after left–right flipping, and (c) image with 100 translations to the right and 100 translations to the bottom.
Figure 5. Data augmentation. (a) Original image, (b) image after left–right flipping, and (c) image with 100 translations to the right and 100 translations to the bottom.
Jimaging 09 00099 g005
Figure 6. Other time-series models and other way of representing the players’ posture.
Figure 6. Other time-series models and other way of representing the players’ posture.
Jimaging 09 00099 g006
Figure 7. Heatmap. (c) is a heat map for two players superimposed on (a,b).
Figure 7. Heatmap. (c) is a heat map for two players superimposed on (a,b).
Jimaging 09 00099 g007
Figure 8. Future predictions of shuttlecock trajectories compared to baseline 1 and baseline 2.
Figure 8. Future predictions of shuttlecock trajectories compared to baseline 1 and baseline 2.
Jimaging 09 00099 g008
Figure 9. Future predictions of shuttlecock trajectories compared to other models.
Figure 9. Future predictions of shuttlecock trajectories compared to other models.
Jimaging 09 00099 g009
Figure 10. Comparison according to the number of frames of input/output.
Figure 10. Comparison according to the number of frames of input/output.
Jimaging 09 00099 g010
Figure 11. The adaptation to sudden changes in trajectory.
Figure 11. The adaptation to sudden changes in trajectory.
Jimaging 09 00099 g011
Table 1. Previous research of future predictions in net sports.
Table 1. Previous research of future predictions in net sports.
YearAuthorSportsPrediction Target
2019Shimizu et al. [11]TennisShot direction
2016Waghmare et al. [3]BadmintonShuttlecock landing point
2019, 2020Wu et al. [4,5]Table tennisServe landing point
2019Sato et al. [6]VolleyballBall landing point
2019Fernando et al. [7]TennisStroke
2022Wang et al. [8]BadmintonStroke
2019Suda et al. [12]VolleyballToss trajectory
2020Lin et al. [13]Table tennisServe trajectory
2022ProposedBadmintonShuttlecock trajectory
Table 2. Number of dimensions.
Table 2. Number of dimensions.
TargetOriginalAfter Embedding
Shuttlecock position22
Player position42
Player posture682
All746
Table 3. Learning rate for each time-series model.
Table 3. Learning rate for each time-series model.
ModelsLearning Rate
RNN0.005
GRU0.01
Transformer0.001
Seq2Seq0.02
Table 4. ADE results. The last row shows the result of the proposed method. The alphabet of the player posture represents the input method to the time-series model: A is the absolute coordinates on the image, R is the relative coordinates in the bounding box, and H is the heat map. Bold font indicates the best (the minimum) ADE.
Table 4. ADE results. The last row shows the result of the proposed method. The alphabet of the player posture represents the input method to the time-series model: A is the absolute coordinates on the image, R is the relative coordinates in the bounding box, and H is the heat map. Bold font indicates the best (the minimum) ADE.
Input DataModelsADE (Pixel)
Shuttlecock PositionPlayer PositionPlayer PostureMatch AMatch BMatch CAverage
--LSTM54.140.745.847.6
-LSTM51.636.543.945.2
RRNN52.038.944.945.9
HRNN63.446.060.258.4
ARNN55.438.049.848.8
RGRU48.936.142.943.6
HGRU59.842.852.951.1
AGRU53.437.043.545.4
RTransformer59.344.549.051.5
HTransformer81.065.773.474.3
ATransformer60.344.247.451.6
RSeq2Seq40.453.845.847.3
HSeq2Seq53.840.745.947.7
ASeq2Seq56.746.643.048.9
RLSTM46.534.441.041.4
HLSTM56.444.451.951.3
ALSTM48.333.039.141.4
Table 5. FDE results. The last row shows the result of the proposed method. The alphabet of the player posture represents the input method to the time-series model: A is the absolute coordinates on the image, R is the relative coordinates in the bounding box, and H is the heat map. Bold font indicates the best (the minimum) FDE.
Table 5. FDE results. The last row shows the result of the proposed method. The alphabet of the player posture represents the input method to the time-series model: A is the absolute coordinates on the image, R is the relative coordinates in the bounding box, and H is the heat map. Bold font indicates the best (the minimum) FDE.
Input DataModelsFDE (Pixel)
Shuttlecock PositionPlayer PositionPlayer PostureMatch AMatch BMatch CAverage
--LSTM94.174.175.081.5
-LSTM88.063.470.976.1
RRNN90.166.774.578.0
HRNN108.777.2101.198.7
ARNN94.064.174.579.1
RGRU87.467.669.876.2
HGRU109.376.796.092.9
AGRU95.870.073.580.3
RTransformer97.174.277.184.0
HTransformer128.695.7109.0113.8
ATransformer99.071.975.283.7
RSeq2Seq64.391.373.577.9
HSeq2Seq93.772.378.683.1
ASeq2Seq92.073.264.978.1
RLSTM81.561.668.071.7
HLSTM97.579.186.788.4
ALSTM83.159.063.370.4
Table 6. ADE results from changing the probability of left–right flipping and the range of translation in augmentation.
Table 6. ADE results from changing the probability of left–right flipping and the range of translation in augmentation.
AugmentationADE (Pixel)
Shift Range (Pixel)Probability of Flip (%)Match AMatch BMatch CAverage
0050.036.840.243.2
02548.834.539.242.0
05051.437.941.244.3
07548.834.541.343.2
50051.738.040.744.3
502549.035.139.742.6
505048.333.039.141.4
507548.036.441.742.5
100050.639.441.844.6
1002549.136.041.742.8
1005051.838.343.646.6
1007551.838.243.044.9
Table 7. FDE results from changing the probability of left–right flipping and the range of translation in augmentation.
Table 7. FDE results from changing the probability of left–right flipping and the range of translation in augmentation.
AugmentationFDE (Pixel)
Shift Range (Pixel)Probability of Flip (%)Match AMatch BMatch CAverage
0085.766.768.174.7
02585.262.565.972.8
05086.867.269.074.7
07583.761.168.373.4
50086.766.868.075.0
502584.164.565.973.3
505083.159.063.370.4
507581.763.666.571.3
100087.665.668.175.0
1002583.263.368.672.2
1005085.762.868.575.3
1007589.068.270.176.3
Table 8. ADE results when input/output frames are changed.
Table 8. ADE results when input/output frames are changed.
Input FramesOutput FramesADE (Pixel)
Match AMatch BMatch CAverage
41248.333.039.141.4
61042.731.535.137.2
8831.925.929.029.5
10622.422.425.323.5
12419.019.218.718.9
Table 9. FDE results when input/output frames are changed.
Table 9. FDE results when input/output frames are changed.
Input FramesOutput FramesFDE (Pixel)
Match AMatch BMatch CAverage
41283.159.063.370.4
61070.851.556.761.0
8852.043.947.248.0
10636.135.837.836.7
12428.529.826.928.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nokihara, Y.; Hachiuma, R.; Hori, R.; Saito, H. Future Prediction of Shuttlecock Trajectory in Badminton Using Player’s Information. J. Imaging 2023, 9, 99. https://doi.org/10.3390/jimaging9050099

AMA Style

Nokihara Y, Hachiuma R, Hori R, Saito H. Future Prediction of Shuttlecock Trajectory in Badminton Using Player’s Information. Journal of Imaging. 2023; 9(5):99. https://doi.org/10.3390/jimaging9050099

Chicago/Turabian Style

Nokihara, Yuka, Ryo Hachiuma, Ryosuke Hori, and Hideo Saito. 2023. "Future Prediction of Shuttlecock Trajectory in Badminton Using Player’s Information" Journal of Imaging 9, no. 5: 99. https://doi.org/10.3390/jimaging9050099

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop