Next Article in Journal
Security at the Edge for Resource-Limited IoT Devices
Next Article in Special Issue
Building Flexible, Scalable, and Machine Learning-Ready Multimodal Oncology Datasets
Previous Article in Journal
IS-CAT: Intensity–Spatial Cross-Attention Transformer for LiDAR-Based Place Recognition
Previous Article in Special Issue
Hierarchical Fusion Network with Enhanced Knowledge and Contrastive Learning for Multimodal Aspect-Based Sentiment Analysis on Social Media
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Music-Driven Dance Generation Method Based on a Spatial-Temporal Refinement Model to Optimize Abnormal Frames

1
State Key Laboratory of Media Convergence and Communication, Communication University of China, Beijing 100024, China
2
Key Laboratory of Acoustic Visual Technology and Intelligent Control System, Ministry of Culture and Tourism, Communication University of China, Beijing 100024, China
3
Beijing Key Laboratory of Modern Entertainment Technology, Communication University of China, Beijing 100024, China
4
School of Information and Communication Engineering, Communication University of China, Beijing 100024, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(2), 588; https://doi.org/10.3390/s24020588
Submission received: 22 November 2023 / Revised: 14 December 2023 / Accepted: 10 January 2024 / Published: 17 January 2024

Abstract

:
Since existing music-driven dance generation methods have abnormal motion when generating dance sequences which leads to unnatural overall dance movements, a music-driven dance generation method based on a spatial-temporal refinement model is proposed to optimize the abnormal frames. Firstly, the cross-modal alignment model is used to learn the correspondence between the two modalities of audio and dance video and based on the learned correspondence, the corresponding dance segments are matched with the input music segments. Secondly, an abnormal frame optimization algorithm is proposed to carry out the optimization of the abnormal frames in the dance sequence. Finally, a temporal refinement model is used to constrain the music beats and dance rhythms in the temporal perspective to further strengthen the consistency between the music and the dance movements. The experimental results show that the proposed method can generate realistic and natural dance video sequences, with the FID index reduced by 1.2 and the diversity index improved by 1.7.

1. Introduction

Hearing and vision are important human sensory systems, and audio and visual signals are common modalities in life. The naturally occurring connection between the two motivates humans to learn and make inferences about audio and video and explore the correlation between these two modalities to better understand human cognition. The music-driven dance generation task implemented in this paper is a study of the multi-modal fusion of audio and video. This task focuses on mapping between music features and motion features, which can be used for dance creation in the fields of art or sports or action generation for characters in games and has great research and application potential.
Traditional music-driven dance generation methods analyze factors such as the loudness, rhythm, emotion, style, and beat of music and dance [1,2], propose similarity matrices [3], and construct motion graphs [4,5]. In addition, refs. [6,7] proposed a probabilistic framework for generating dance from music. Hidden Markov models [8] are also widely used in dance movement choreography [9,10,11,12]. Due to the excessive consideration of professional factors, the traditional methods cannot be flexibly applied to random music and dance scenes.
Recently, based on the rapid development of deep learning technologies, some deep learning-based approaches have emerged for music-driven dance generation tasks. For example, Lee et al. [13] proposed a generative method of analysis and synthesis. In the analysis stage, dance is decomposed into a series of basic dance units, and the model learns how to move through these units; in the synthesis stage, the model generates a complete dance sequence by organizing multiple basic dance movements based on the input music. Luka et al. [14] designed a Chor-RNN framework where, after training on original data, an LSTM network was utilized to achieve the prediction of dance movements. Refs. [15,16,17,18] all designed corresponding generation frameworks based on LSTM networks for generating 3D dance movements, but the generated dance sequences had the problem of abnormal movements. In order to improve the naturalness of the generated dance movements, Ren et al. [19] proposed a new pose perceptual loss in the generation stage, but the large number of motion key points still limits the quality of the synthesized dances. Yalta et al. [20] proposed to solve this kind of problem using weakly supervised learning, but the quality of the generated dances was still unsatisfactory due to a lack of human choreography experience. Sun et al. [21] proposed a cross-modal correlation framework based on GAN that aims to effectively correlate music and dance through probability. Although these approaches note the importance of specific neural network structure design for the music-driven dance generation tasks, they ignore the problem of the unnatural movements of the generated dance sequences.
Aiming at the problem of unnatural dance movements generated using existing deep learning methods in music-driven dance generation tasks, this paper proposes a method based on the spatial-temporal refinement model to optimize the abnormal frames, The overall flowchart is shown in Figure 1. Firstly, a cross-modal alignment model [22] is used to learn the correspondence between the audio and the dance video using pose features as an intermediary to match the best pose segments with any music based on the learned correspondence and synthesize the pose segments into a complete dance sequence. Since the pose segments are discrete, there is an abnormal motion problem between the segments. Secondly, an abnormal frame optimization algorithm is designed to solve the problem, the pixel offsets of the key points corresponding to the two adjacent frames are first calculated to locate the position of the abnormal frame and the corresponding two keyframes are determined, the keyframes are input into the spatial refinement model, and consecutive natural action frames are generated in the middle of the two keyframes and the abnormal frames are replaced. Finally, the temporal refinement model is used to enforce consistency constraints between the music and the dance.
The main contributions of this paper are summarized as follows: (1) learning the correspondence between music and dance sequences through a cross-modal alignment model to achieve the matching of optimal pose segments for any input music; (2) proposing an abnormal frame optimization algorithm, which detects the abnormal frames and carries out the optimization of the abnormal frames using the spatial refinement model; (3) aligning the music beats and the dance rhythms using the temporal refinement model to achieve the consistency constraints on the overall dance movements and music.

2. Related Works

In Section 2, we introduce the research works related to cross-modality generation and human movement prediction. In Section 2.1, we introduce the cross-modality generation works between text, image, audio, video, and other modalities. In Section 2.2, we introduce related works on human movement prediction, which help to understand the way dance movements are generated in music-driven dance generation tasks.

2.1. Cross-Modality Generation

The cross-modality generation task aims to explore the correlation between different modalities. Currently, there are extensive research works on text and images, such as image captioning [23,24] and text-to-image synthesis [25,26]. There are also works on audio and video cross-modal generation [27,28] that uses visual cues to generate background sounds for target objects in videos. Refs. [29,30] predict the motion of mouth signs from audio, focusing on speech and lip synchronization. Ref. [31] builds a hierarchical system for predicting lip movements and synthesizing speech videos. Ref. [32] trained an LSTM network on the input audio of a violin or piano performance to predict the player’s hand movements.

2.2. Human Movement Prediction

Movement prediction is full of randomness and spatial-temporal complexity, so dance movement generation is a challenging task. Graph convolutional networks (GCNs) can capture the spatial-temporal relationships between bodies well and extract high-level features of the human skeleton, thus in Lebailly et al. [33] the relationship between joints is learned using graph neural networks (GCNs) and the GCN architecture is combined with the initial layer of time to achieve motion prediction. The time initial layer learns long short-term information by processing the inputs of different sequence lengths. AS-GCN [34] captures richer action-specific correlations through an inference model with an encoder–decoder structure. Refs. [35,36] proposed a two-stream adaptive graph convolutional network model (2s-AGCN) based on ST-GCN and AS-GCN. This model utilizes the similarity between graph joints and adaptively learns graph structures in addition to the natural connections in the human skeleton, thus improving the accuracy of human movement recognition.

3. Methodology

In this paper, we propose a music-driven dance generation method based on the optimization of anomalous frames using a spatial-temporal refinement model, as shown in Figure 2. The method consists of three main parts: (1) learning the cross-modal correspondence relationship, (2) proposing the abnormal frame optimization algorithm (AFOA), and (3) placing constraints on the consistency of the music and the dance movements. In Section 3.1, the cross-modal alignment model is utilized to learn the correspondence between audio and dance video; based on the learned correspondence, the best dance clip is matched with the input music and synthesized into a complete dance sequence. In Section 3.2, the abnormal frame optimization algorithm is proposed to solve the problem of abnormal movements in the dance sequence. In Section 3.3, the temporal refinement model is used to align the AFOA-optimized dance sequences and music to strengthen the consistency constraints between music and dance.

3.1. Learning the Cross-Modal Correspondence Relationship

Learning the correspondence relationship between music and dance is the key to achieving consistent audio-visual integration and is also an important step in the dance generation process. Therefore, this section uses the cross-modal alignment model [22] to learn the correspondence relationship between audio features and pose features by using the pose features as an intermediary to connect audio and dance videos. The learning process of this correspondence relationship is as follows:
Firstly, as shown in the first part of Figure 2, given an original domain paired audio A and dance video V , MFCC is used to extract the audio features of audio A and convert the frequency domain signal into low-dimensional vectors. The audio encoder consists of Bi-LSTM and FC to extract the forward and backward information of the audio to enrich the audio features. The OpenPose algorithm [37] is used to extract the pose features P of the dance video V . The pose encoder consists of an ST-GCN [38] that learns the spatial and temporal modalities of the data to extract the depth features of the pose sequence. After the feature extraction and the coding process of the audio and dance video, the audio A , dance video V , and pose feature P of the original domain can be represented as a ternary group { x i o = ( a i o , v i o , p i o ) | x O } i = 1 N , where N is the total number of frames.
Secondly, the data from the original domain are input into the cross-modal alignment model C m ( · ) , and the correspondence between the two modalities is learned by measuring the distance between the audio features a i o and the pose features p i o . The learning process of the correspondence can be expressed as Equation (1), where θ is the network parameter of the cross-modal alignment model C m ( · ) , and f ( · ) is the feature function.
θ C = a r g   min θ C m a i o , p i o θ f i = j C m a i o , p i o θ f i j
Finally, based on the learned correspondence θ c between the audio and pose features, the best pose segments p ¯ can be matched from the original domain sequence database D (described in Section 4.1) for any target music a t . This process can be represented as:
p ¯ = a r g   min p D C m a t , p θ c

3.2. The Abnormal Frame Optimization Algorithm

The best pose segments p ¯ obtained above are synthesized into a complete pose sequence. Since the pose segments in database D are discrete, there will be abnormal motion problems in adjacent frames at the connection between the pose segments. Therefore, in this section, the abnormal frame optimization algorithm (AFOA) is designed to solve the problem of abnormal movements and ensure the natural continuity of the dance movements. As shown in Figure 3, the AFOA includes two steps: firstly, detecting abnormal frame n by calculating the pixel offsets of the key points corresponding to two adjacent frames; secondly, two keyframes k 1 , k 2 are determined based on the detected abnormal frame and input into the spatial refinement module (SRM), a continuous and natural pose sequence is generated between the two keyframes, and the pose segment where the corresponding abnormal frame is located is replaced. The abnormal frame optimization algorithm is introduced in two parts: detecting the abnormal frame in Section 3.2.1 and the spatial refinement model network structure in Section 3.2.2.

3.2.1. Detecting the Abnormal Frame

In order to remove the abnormal frames in the dance sequence and make the dance sequence continuous and natural, the abnormal frames must be detected first. The discontinuous frames in the pose sequence are detected by calculating the pixel offsets Δ P of the key points corresponding to the two adjacent frames. The calculation of Δ P is shown in Equation (3), where f 1 and f 2 are the two adjacent frames, i are the key points labeled 0–22, and ( x , y ) are the pixel coordinates of the key points.
Δ P = i = 0 22 ( | x f 1 i x f 2 i | + | y f 1 i y f 2 i | )
We compared the discontinuous frames detected when Δ P = 8 , Δ P = 10 , Δ P = 12 , as shown in Figure 4. The discontinuous frames detected when Δ P = 8 were actually back-to-back frames by observation. The discontinuous frames detected when Δ P = 12 have a larger mutation compared with the two frames before and after observation, so anomalous frames may be missed when setting a larger Δ P . In contrast, setting Δ P = 10 can detect the discontinuous frames between the front and back frames well, so the frame with a key point movement larger than 10 pixels is set as a discontinuous frame [22].
In order to generate a continuous natural pose sequence through the spatial refinement model (SRM), two keyframes need to be set. Assuming that the detected abnormal frame is the nth frame, if k 1 = n m and k 1 = n m are confirmed as non-abnormal frames, then k 1 and k 2 can be defined as the keyframes. As shown in Figure 3, k 1 and k 2 input into the spatial refinement model (SRM) as keyframes generate a natural transitional dance movement sequence between k 1 and k 2 , and replace the segment where the abnormal frame is located to make the movements in the whole dance sequence continuous and natural. In this experiment, we set m = 3 , which generates a new 5-frame movement sequence. It achieves the purpose of replacing the segment where the abnormal frame is located and controls the complexity of the network.

3.2.2. Spatial Refinement Model Network Structure

In order to generate a continuous pose sequence between two keyframes to replace the pose sequence where the corresponding abnormal frame is located, this paper utilizes the spatial refinement model (SRM) [39] for movement prediction between keyframes k 1 and k 2 . As shown in Figure 5, the spatial refinement model (SRM) mainly includes encoders, decoders, and controllers. The spatial refinement model uses a root trajectory controller and a speed controller based on a transformer, which can better learn inter-frame context information and achieve fine-grained motion control. The SRM uses the first LSTM to embed the pose information of the historical frame into the potential space and uses the second LSTM to predict the posture of the next frame. The following is a detailed introduction to the spatial refinement model.
The spatial refinement model (SRM) mainly consists of encoders, decoders, and controllers. Encoders: the state encoder aims to receive pose information X t = p t , o t r , c t , the velocity encoder aims to receive velocity information v t in order to perceive the dynamics of the movement, and the position encoder aims to receive the root position information p t + 1 for the next frame. Decoders: the root decoder aims to predict the positional information of the root joint, and the state decoder aims to predict the relative position and velocity information of the other joints. Controllers: the target controller receives the pose information of the keyframes and enables the network to perceive the distance between the predicted frame and the target keyframe. The root trajectory controller and velocity controller are based on the transformer construct. The transformer structure [40,41,42] can model the dependency of sequences by using the correlation between keys and different tokens. This allows the network to capture the temporal context information of dance movement trajectories p t + 1 , u and velocity constraints f t + 1 , u , and thus learn smooth context representation.
In order to enable the spatial refinement model (SRM) to generate naturally continuous dance sequences between keyframes k 1 and k 2 , this paper uses four loss functions—reconstruction loss, root trajectory smoothing loss, keyframe consistency loss, and velocity consistency loss [39], to supervise the model in an end-to-end form. These four loss functions will be introduced separately below.
(1) Reconstruction Loss: The reconstruction loss is constructed by the mean square error (MSE) loss to measure the similarity between the predicted movement and the real movement and motivates the network to generate a motion sequence that satisfies the constraints. The reconstruction loss is shown in Equation (4), where n is the length of the sequence, X t = p t , o t r , c t represents the true value information at time t , including the root position information p t , the rotation angle o t r of the root, and the velocity factor v t .
L r e c = 1 n t = k 1 + 1 k 2 X ^ t X t 2
(2) Root Trajectory Smoothing Loss: It enhances temporal consistency by minimizing the differences in the spatial position and rotation angle of the root joint between frame t and frame t 1 . The root trajectory smoothing loss is shown in Equation (5):
L r o o t = 1 n ( t = k 1 + 1 k 2 p t p t 1 2 + t = k 1 + 1 k 2 o t r o t 1 r 2 )
(3) Keyframe consistency loss: The purpose of the spatial refinement model is to generate natural transitional dance movements between two keyframes, which requires ensuring the continuity of the predicted movements near the keyframes. The keyframe consistency loss is shown in Equation (6), where s is the number of frames affected by the keyframe, refer to [39] to set s = 3 .
L k e y = 1 2 s ( t = k 1 + 1 k 1 + s p ^ t p k 1 2 + t = k 2 s + 1 k 2 p ^ t p k 2 2 )
(4) Velocity consistency loss: It ensures that the velocity of the synthesized dance sequence is consistent with the given control conditions, as shown in Equation (7), where f t is the velocity constraint given according to the overall dance sequence, and f ~ t is the velocity of the calculated predicted dance movements.
L v f a c = 1 n t = k 1 + 1 k 2 f ~ t f t 2
The complete loss function is shown in Equation (8), where ω r e c , ω r o o t , ω k e y and ω v f a c are the corresponding loss weights, set to 0.3, 0.15, 0.2, and 0.15, respectively.
L = ω r e c L r e c + ω r o o t L r o o t + ω k e y L k e y + ω v f a c L v f a c

3.3. Music and Dance Consistency Constraints

After processing using the spatial refinement model, the abnormal movements in the dance sequences are well smoothed from the spatial perspective. But overall, there will be a phenomenon where the music beat is inconsistent with the dance rhythm. Therefore, the temporal refinement model [22] is used to further constrain the music and dance from a temporal perspective. The constraint is to align the music beat with the dance rhythm.
Firstly, the beats of the target audio a t are extracted using the Librosa library [43], and the dance rhythms are defined as postural movements with large variations, as shown in Equation (9):
d μ = a r g   max j ( | p j t p j 1 t | ) ,   j i , i + ω c
Secondly, a sliding window of ω c frame size is set to search the local maximum to align. After the detected music beat a μ , the ω c frames can be reorganized as i , μ and μ , i + ω c . For the preceding and the following frames of the current beat point, the cubic fitting interpolation method is adopted to obtain the aligned pose p ^ t . The alignment function is denoted as:
a * = a r g   min a x = i μ ( m = 0 3 a m x m p x t ) 2
p ^ t = a 0 * + m = 1 3 a m * x m
After the optimization of the temporal refinement model, as shown in Figure 6, the music beats and dance rhythms can be aligned well, and the music and dance movements can be better integrated as one.

4. Experiments

In this section, we first describe the dataset used in the experiment and the implementation details. Then, the proposed method is compared with other methods, and ablation studies using different optimization strategies are conducted to prove the effectiveness of our proposed method. Finally, the dance sequences generated by different methods are visualized for subjective evaluation and comparison.

4.1. Dataset and Implementation Details

The experiments in this paper are performed on the dataset of Guo et al. [22], which collects dance videos from real scenarios and can obtain the dancer’s movements and audio at the same time. The dataset contains 122 female dance videos and 32 male dance videos and each video length ranges from 3 to 5 min, totaling 9 h. The resolution of the videos is 1920 × 1080 and is processed at a standard frame rate of 24 fps, and the size of the dance image extracted from the frame is adjusted to a uniform size. The OpenPose algorithm [37] is used to extract human key points from the dance images. Each pose consists of 23 joint points. The frames whose key points are not fully detected and abnormal frames with sudden motion changes are deleted. After processing the abnormal data, the pose sequences are intercepted into different pose segments and finally constructed as pose segment database D .
To evaluate the quality of the generated pose sequences, 80% of the dataset was randomly selected for training and the remaining 20% for testing. The model in this paper was implemented using PyTorch on an RTX-3090 GPU and trained for 500 epochs. The spatial refinement model was trained using the Adam optimizer with a learning rate of 0.001 and batch size of 128.

4.2. Evaluation Metrics

(1) FID: Fréchet inception distance (FID) is a commonly used evaluation index in image processing, which can effectively calculate the feature distance between the real samples and the generated samples. The smaller the FID value, the closer the generated data is to the real data, and the better the effect of the model. The calculation of FID is shown in Equation (12), where T r a c e is the trace of the matrix, m and C are the mean and covariance of the probability distribution of the generated data, respectively, and m T and C T are the mean and covariance of the probability distribution of the real data, respectively.
d 2 m , C , m T , C T = m m T 2 2 + T r a c e ( C + C T 2 ( C C T ) 1 / 2 )
(2) Beat Hit Rate: R h i t is the ratio of the number of motion beats aligned with music to the total number of motion beats, R h i t calculated as shown in Equation (13). The total number B K of motion beats and the number B A of motion beats aligned with the music beats is counted. If the input music is synchronized with the generated dance movements, the indicator R h i t will be higher.
R h i t = B A B K
(3) Diversity and multimodality: Diversity refers to the differences between the generated dance movements, the greater the differences, the greater the diversity. The diversity is calculated by randomly sampling the 40 dances generated and calculating the average FID value between them. The larger the value, the greater the differentiation between all the movements, and the greater the diversity of the dances. Multimodality refers to the ability to generate different dances given the same musical conditions, demonstrated through visualization.

4.3. Comparison of the Results to Other Methods

This paper compares the methods of Lee et al. [13] and Guo et al. [22] using four metrics: FID, beat hit rate, diversity, and multimodality.
The results of comparing this paper’s method with the methods of Lee et al. [13] and Guo et al. [22] are shown in Table 1. The dance sequence generated using this paper’s method has the smallest FID value, and the generated dance sequence is closer to the real values. The increase in beat hit rate indicates that the consistency between the generated dance rhythms and the music beats is improved, indicating that the temporal refinement model imposes an effective consistency constraint on the music and dance. The diversity metric is improved by 1.7, indicating that compared with the traditional TSD algorithm used by Guo et al. [22], the spatial refinement model based on LSTM and the transformer proposed in this paper can generate rich dance movements. Correspondingly, the multimodal metric is also improved, further proving that the method in this paper can generate rich dance movements.

4.4. Ablation Study

In this section, the ablation experiments of the three processing methods for abnormal frames are compared: no abnormal frame optimization processing, processing based on the traditional TSD algorithm, and processing based on the AFOA algorithm proposed in this paper. As shown in Table 2, it can be seen through the results that the experimental results of no abnormal frame optimization processing are the worst; after Guo et al. [22] optimized the abnormal frames using the traditional TSD algorithm, the FID value and diversity are improved. The experimental results of the AFOA algorithm proposed in this paper are the best, indicating that the optimization strategy of generating continuous natural dance movements through the spatial refinement model between keyframes k 1 and k 2 is effective. It ensures the natural continuity of dance movements and improves the diversity of dance movements.

4.5. Visualization of the Generated Dance Sequence

The generated dance sequences are visualized for subjective evaluation and comparison. As shown in Figure 7, two different dance sequences can be generated given the same music, reflecting the multimodality of the method in this paper. As shown in Figure 8, the comparison before and after the spatial refinement model optimizes the abnormal frames. Through the visualization results, it can be seen that after the spatial refinement model optimizes the abnormal frames, the dance movements can be made natural and continuous.
The dance sequences generated using the method of this paper can be visually compared with the methods of Lee et al. [13] and Guo et al. [22]. As shown in Figure 9, the dance movements generated by Lee et al.’s method have fewer changes and the dance diversity is poor. In contrast, Guo et al.’s method has significant improvements, but the generated movements also have abnormal jitter. The abnormal frame optimization algorithm (AFOA) proposed in this paper generates natural and continuous pose segments between two keyframes through the spatial refinement model (SRM). It achieves the purpose of optimizing abnormal frames, making dance movements continuous and natural, and improving the diversity of dance movements.

5. Conclusions and Discussion

Aimed at the problem of dance movement mutation in existing music-driven dance generation methods when generating dance sequences, this paper proposes a method to optimize abnormal frames based on a spatial-temporal refinement model. The method consists of three parts: (1) The cross-modal alignment model is used to learn the correspondence between audio and dance video modalities, match the best dance segments for any input music based on the learned correspondence, and synthesize it into a complete dance sequence. (2) Since the dance segments are discrete, there is a mutation problem in the dance movements between segments. In order to solve this problem, the abnormal frame optimization algorithm is proposed. First, by calculating the pixel offset of the corresponding key points of two adjacent frames, the positions of the abnormal frame and the corresponding two keyframes are determined. Then, the spatial refinement model generates a natural transition pose sequence between the two keyframes and replaces the pose sequence where the corresponding abnormal frame is located to optimize the abnormal frame. (3) The temporal refinement model is used to align the music beats and dance rhythms to enhance the consistency of the music and dance movements from the temporal perspective. Experiments have shown that the method proposed in this paper can generate dance sequences with natural and continuous movements.
Based on the music-driven dance generation method proposed in this paper, a virtual reality-based real-life dance performance system can be designed. The stage combined with virtual reality technology can bring novel and intuitive artistic audio-visual experience to the audience. The system can assist professionals in choreographing dance movements and can also provide dance movement instruction for amateurs who love to dance.
Packaging the demo system as a software tool can make the system easier to deploy in more application scenarios. In future work, we plan to collect and integrate music genres, dance styles and other factors to make the generated dance more delicate and relevant to the theme.

Author Contributions

Conceptualization, H.W. and Y.S.; methodology, H.W. and Y.S.; software, H.W.; validation, H.W., Y.S. and T.W.; formal analysis, H.W.; investigation, H.W.; resources, W.J.; data curation, T.W.; writing—original draft preparation, H.W.; writing—review and editing, H.W.; visualization, H.W.; supervision, W.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Funds for the National Key R&D Program of China, Ministry of science and technology of China (2021YFF0901705) and the Fundamental Research Funds for the Central Universities.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Datasets during the current study are available from https://drive.google.com/file/d/1vPyOqaIT-nmB5Yb8HQ0FZk8Usg2RD8Vp/view (accessed on 21 November 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xia, G.; Tay, J.; Dannenberg, R.; Veloso, M. Autonomous robot dancing driven by beats and emotions of music. In Proceedings of the 11th International Conference on Autonomous Agents and Multiagent Systems—Volume 1, Valencia, Spain, 4–8 June 2012; pp. 205–212. [Google Scholar]
  2. LaViers, A.; Egerstedt, M. Style based robotic motion. In Proceedings of the 2012 American Control Conference (ACC), Montreal, QC, Canada, 27–29 June 2012; pp. 4327–4332. [Google Scholar]
  3. Kim, J.W.; Fouad, H.; Sibert, J.L.; Hahn, J.K. Perceptually motivated automatic dance motion generation for music. Comput. Animat. Virtual Worlds 2009, 20, 375–384. [Google Scholar] [CrossRef]
  4. Xu, J.; Takagi, K.; Sakazawa, S. Motion synthesis for synchronizing with streaming music by segment-based search on metadata motion graphs. In Proceedings of the 2011 IEEE International Conference on Multimedia and Expo, Barcelona, Spain, 11–15 July 2011; pp. 1–6. [Google Scholar]
  5. Fan, R.; Xu, S.; Geng, W. Example-based automatic music-driven conventional dance motion synthesis. IEEE Trans. Vis. Comput. Graph. 2012, 18, 501–515. [Google Scholar] [PubMed]
  6. Ofli, F.; Erzin, E.; Yemez, Y.; Tekalp, A.M. Learn2dance: Learning statistical music-to-dance mappings for choreography synthesis. IEEE Trans. Multimed. 2012, 14, 747–759. [Google Scholar] [CrossRef]
  7. Fukayama, S.; Goto, M. Music content driven automated choreography with beat-wise motion connectivity constraints. In Proc. Sound Music Comput. Proc. SMC 2015, 3, 177–183. [Google Scholar]
  8. Sonnhammer, E.L.; Von Heijne, G.; Krogh, A. A hidden markov model for predicting transmembrane helices in protein sequences. Proc. Int. Conf. Intell. Syst. Mol. Biol. 1998, 6, 175–182. [Google Scholar] [PubMed]
  9. Manfrè, A.; Infantino, I.; Vella, F.; Gaglio, S. An automatic system for humanoid dance creation. Biol. Inspired Cogn. Archit. 2016, 15, 1–9. [Google Scholar] [CrossRef]
  10. Augello, A.; Infantino, I.; Manfré, A.; Pilato, G.; Vella, F.; Chella, A. Creation and cognition for humanoid live dancing. Robot. Auton. Syst. 2016, 86, 128–137. [Google Scholar] [CrossRef]
  11. Manfré, A.; Augello, A.; Pilato, G.; Vella, F.; Infantino, I. Exploiting interactive genetic algorithms for creative humanoid dancing. Biol. Inspired Cogn. Archit. 2016, 17, 12–21. [Google Scholar] [CrossRef]
  12. Qin, R.; Zhou, C.; Zhu, H.; Shi, M.; Chao, F.; Li, N. A music-driven dance system of humanoid robots. Int. J. Humanoid Robot. 2018, 15, 1850023. [Google Scholar] [CrossRef]
  13. Lee, H.; Yang, X.; Liu, M.; Wang, T.; Lu, Y.; Yang, M.; Kautz, J. Dancing to Music. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 3586–3596. [Google Scholar]
  14. Crnkovic-Friis, L.; Crnkovic-Friis, L. Generative Choreography using Deep Learning. In Proceedings of the International Conference on Innovative Computing and Cloud Computing, Paris, France; 2016; pp. 272–277. [Google Scholar] [CrossRef]
  15. Martinez, J.; Black, M.J.; Romero, J. On Human Motion Prediction Using Recurrent Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4674–4683. [Google Scholar]
  16. Komura, T.; Habibie, I.; Holden, D.; Schwarz, J.; Yearsley, J. A recurrent variational autoencoder for human motion synthesis. In Proceedings of the 28th British Machine Vision Conference, London, UK, 4–7 September 2017; pp. 1–12. [Google Scholar]
  17. Holden, D.; Saito, J.; Komura, T. A deep learning framework for character motion synthesis and editing. ACM Trans. Graph. 2016, 35, 138–145. [Google Scholar] [CrossRef]
  18. Li, Z.; Zhou, Y.; Xiao, S.; He, C.; Huang, Z.; Li, H. Auto-Conditioned Recurrent Networks for Extended Complex Human Motion Synthesis. arXiv 2018. [Google Scholar] [CrossRef]
  19. Ren, X.; Li, H.; Huang, Z.; Chen, Q. Music-oriented Dance Video Synthesis with Pose Perceptual Loss. arXiv 2019, arXiv:1912.06606v1. [Google Scholar]
  20. Yalta, N.; Watanabe, S.; Nakadai, K.; Ogata, T. Weakly-Supervised Deep Recurrent Neural Networks for Basic Dance Step Generation. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–8. [Google Scholar]
  21. Sun, G.; Wong, Y.; Cheng, Z.; Kankanhalli, M.; Geng, W.; Li, X. DeepDance: Music-to-Dance Motion Choreography with Adversarial Learning. IEEE Trans. Multimed. 2020, 23, 497–509. [Google Scholar] [CrossRef]
  22. Guo, X.; Zhao, Y.; Li, J. DanceIt: Music-inspired dancing video synthesis. IEEE Trans. Image Process. 2021, 30, 5559–5572. [Google Scholar] [CrossRef] [PubMed]
  23. Karpathy, A.; Li, F.F. Deep visual-semantic alignments for generating image descriptions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  24. Vinyals, O.; Toshev, A.; Bengio, S.; Erhan, D. Show and tell: A neural image caption generator. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  25. Reed, S.; Akata, Z.; Yan, X.; Logeswaran, L.; Schiele, B.; Lee, H. Generative adversarial text to image synthesis. In Proceedings of the 33rd International Conference on Machine Learning, New York, NY, USA, 20–22 June 2016. [Google Scholar]
  26. Zhang, H.; Xu, T.; Li, H.; Zhang, S.; Huang, X.; Wang, X.; Metaxas, D. StackGAN: Text to photo-realistic image synthesis with stacked generative adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  27. Owens, A.; Isola, P.; McDermott, J.; Torralba, A.; Adelson, E.; Freeman, W. Visually indicated sounds. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  28. Davis, A.; Rubinstein, M.; Wadhwa, N.; Mysore, G.; Durand, F.; Freeman, W. The visual microphone: Passive recovery of sound from video. In ACM Transactions on Graphics; Association for Computing Machinery: New York, NY, USA, 2014. [Google Scholar]
  29. Karras, T.; Aila, T.; Laine, S.; Herva, A.; Lehtinen, J. Audio-driven facial animation by joint end-to-end learning of pose and emotion. In ACM Transactions on Graphics; Association for Computing Machinery: New York, NY, USA, 2017. [Google Scholar]
  30. Suwajanakorn, S.; Seitz, S.M.; Kemelmacher-Shlizerman, I. Synthesizing Obama: Learning lip sync from audio. In ACM Transactions on Graphics; Association for Computing Machinery: New York, NY, USA, 2017. [Google Scholar]
  31. Chen, L.; Li, Z.; Maddox, R.K.; Duan, Z.; Xu, C. Lip Movements Generation at a Glance. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 16–28. [Google Scholar]
  32. Zhou, Y.; Wang, Z.; Fang, C.; Bui, T.; Berg, T. Visual to sound: Generating natural sound for videos in the wild. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  33. Lebailly, T.; Kiciroglu, S.; Salzmann, M.; Fua, P.; Wang, W. Motion Prediction Using Temporal Inception Module. In Proceedings of the Asian Conference on Computer Vision, Kyoto, Japan, 30 November–4 December 2020. [Google Scholar]
  34. Li, M.; Chen, S.; Chen, X.; Zhang, Y.; Wang, Y.; Tian, Q. Actional-structural graph convolutional networks for skeleton-based action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 3595–3603. [Google Scholar]
  35. Wang, H.; Ho, E.S.L.; Shum, H.P.H.; Zhu, Z. Spatio-temporal manifold learning for human motions via long-horizon modeling. IEEE Trans. Vis. Comput. Graph. 2021, 27, 216–227. [Google Scholar] [CrossRef] [PubMed]
  36. Shi, L.; Zhang, Y.; Cheng, J.; Lu, H. Two-stream adaptive graph convolutional networks for skeleton-based action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 12026–12035. [Google Scholar]
  37. Cao, Z.; Hidalgo, G.; Simon, T.; Wei, S.-E.; Sheikh, Y. OpenPose: Realtime multi-person 2D pose estimation using part affinity fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  38. Yan, S.; Xiong, Y.; Lin, D. Spatial temporal graph convolutional networks for skeleton-based action recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 7444–7452. [Google Scholar]
  39. Pan, J.; Wang, S.; Bai, J.; Dai, J. Diverse Dance Synthesis via Keyframes with Transformer Controllers. Comput. Graph. Forum 2021, 40, 71–83. [Google Scholar] [CrossRef]
  40. Banar, N.; Daelemans, W.; Kestemont, M. Character-level transformer-based neural machine translation. In Proceedings of the International Conference on Natural Language Processing and Information Retrieval, Seoul, Republic of Korea, 18–20 December 2020; pp. 149–156. [Google Scholar]
  41. Zhang, D.; Zhang, H.; Tang, J.; Wang, M.; Hua, X.; Sun, Q. Feature pyramid transformer. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 323–339. [Google Scholar]
  42. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kir-Illov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  43. McFee, B.; Raffel, C.; Liang, D.; Ellis, D.P.; McVicar, M.; Battenberg, E.; Nieto, O. Librosa: Audio and music signal analysis in Python. In Proceedings of the 14th Python in Science Conference, Austin, TX, USA, 6–12 July 2015; pp. 18–25. [Google Scholar]
Figure 1. The overall flowchart of this paper is used to align the music beats and dance rhythms to further enhance the consistency between the music and the dance movements. Through spatial and temporal constraints, natural and continuous dance movements are generated.
Figure 1. The overall flowchart of this paper is used to align the music beats and dance rhythms to further enhance the consistency between the music and the dance movements. Through spatial and temporal constraints, natural and continuous dance movements are generated.
Sensors 24 00588 g001
Figure 2. Framework diagram of the music-driven dance generation method based on the spatial-temporal refinement model to optimize abnormal frames. 1, 2, 3 in the figure denote steps 1, 2 and 3, while corresponding to the contents of Section 3.1, Section 3.2 and Section 3.3, respectively.
Figure 2. Framework diagram of the music-driven dance generation method based on the spatial-temporal refinement model to optimize abnormal frames. 1, 2, 3 in the figure denote steps 1, 2 and 3, while corresponding to the contents of Section 3.1, Section 3.2 and Section 3.3, respectively.
Sensors 24 00588 g002
Figure 3. Abnormal frame optimization algorithm (AFOA) framework.
Figure 3. Abnormal frame optimization algorithm (AFOA) framework.
Sensors 24 00588 g003
Figure 4. Abnormal frames in a dance sequence.
Figure 4. Abnormal frames in a dance sequence.
Sensors 24 00588 g004
Figure 5. Network framework for the spatial refinement model (SRM).
Figure 5. Network framework for the spatial refinement model (SRM).
Sensors 24 00588 g005
Figure 6. Temporal refinement model aligning music beats and dance rhythms. The first row of red dashed boxes indicates the detected music beats and the second row of red dashed boxes indicates the dance rhythms.
Figure 6. Temporal refinement model aligning music beats and dance rhythms. The first row of red dashed boxes indicates the detected music beats and the second row of red dashed boxes indicates the dance rhythms.
Sensors 24 00588 g006
Figure 7. Multimodality: given the same musical conditions two different dance sequences can be generated.
Figure 7. Multimodality: given the same musical conditions two different dance sequences can be generated.
Sensors 24 00588 g007
Figure 8. The spatial refinement model optimizes abnormal dance movements.
Figure 8. The spatial refinement model optimizes abnormal dance movements.
Sensors 24 00588 g008
Figure 9. Visual comparison between the method of this paper and the methods of Lee et al. [13] and Guo et al. [22].
Figure 9. Visual comparison between the method of this paper and the methods of Lee et al. [13] and Guo et al. [22].
Sensors 24 00588 g009
Table 1. Comparison of objective indicators of dance generation between the method proposed in this paper and other methods. indicates that larger values are better, and indicates that smaller values are better.
Table 1. Comparison of objective indicators of dance generation between the method proposed in this paper and other methods. indicates that larger values are better, and indicates that smaller values are better.
MethodFID ↓Beat Hit Rate ↑Diversity ↑Multimodality ↑
Real dances5.951.6%53.5-
Lee et al. [13]12.865.1%53.247.8
Guo et al. [22]8.670.3%58.657.8
Ours7.472.6%60.358.3
Table 2. Ablation studies with different optimization strategies. indicates that larger values are better, and indicates that smaller values are better.
Table 2. Ablation studies with different optimization strategies. indicates that larger values are better, and indicates that smaller values are better.
MethodFID ↓Diversity ↑
Real dances5.953.5
No abnormal frame optimization process11.256.2
Guo-TSD8.658.6
Ours-AFOA algorithm7.460.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, H.; Song, Y.; Jiang, W.; Wang, T. A Music-Driven Dance Generation Method Based on a Spatial-Temporal Refinement Model to Optimize Abnormal Frames. Sensors 2024, 24, 588. https://doi.org/10.3390/s24020588

AMA Style

Wang H, Song Y, Jiang W, Wang T. A Music-Driven Dance Generation Method Based on a Spatial-Temporal Refinement Model to Optimize Abnormal Frames. Sensors. 2024; 24(2):588. https://doi.org/10.3390/s24020588

Chicago/Turabian Style

Wang, Huaxin, Yang Song, Wei Jiang, and Tianhao Wang. 2024. "A Music-Driven Dance Generation Method Based on a Spatial-Temporal Refinement Model to Optimize Abnormal Frames" Sensors 24, no. 2: 588. https://doi.org/10.3390/s24020588

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop