Next Article in Journal
Application of the E-Nose as a Non-Destructive Technique in the Early Detection of Monilinia laxa on Plum (Prunus domestica L.)
Previous Article in Journal
Targetless Radar–Camera Calibration via Trajectory Alignment
Previous Article in Special Issue
Change in Acoustic Parameters of Electric Guitar Strings Under Dynamic Loading
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Joint Learning of Emotion and Singing Style for Enhanced Music Style Understanding

1
School of Humanities and Arts, Hunan Institute of Traffic Engineering, Hengyang 421219, China
2
Department of Crop and Soil Sciences, College of Agriculture and Environmental Sciences, University of Georgia, Tifton, GA 31793, USA
*
Authors to whom correspondence should be addressed.
Sensors 2025, 25(24), 7575; https://doi.org/10.3390/s25247575 (registering DOI)
Submission received: 5 November 2025 / Revised: 30 November 2025 / Accepted: 10 December 2025 / Published: 13 December 2025

Abstract

Understanding music styles is essential for music information retrieval, personalized recommendation, and AI-assisted content creation. However, existing work typically addresses tasks such as emotion classification and singing style classification independently, thereby neglecting the intrinsic relationships between them. In this study, we introduce a multi-task learning framework that jointly models these two tasks to enable explicit knowledge sharing and mutual enhancement. Our results indicate that joint optimization consistently outperforms single-task counterparts, demonstrating the value of leveraging inter-task correlations for more robust singing style analysis. To assess the generality and adaptability of the proposed framework, we evaluate it across various backbone architectures, including Transformer, TextCNN, and BERT, and observe stable performance improvements in all cases. Experiments on a benchmark dataset, which were self-constructed and collected through professional recording devices, further show that the framework not only achieves the best accuracy on both tasks on our dataset under a singer-wise split, but also yields interpretable insights into the interplay between emotional expression and stylistic characteristics in vocal performance.
Keywords: music singing style; music emotion; AI4Music; multi-task learning music singing style; music emotion; AI4Music; multi-task learning

Share and Cite

MDPI and ACS Style

Chen, Y.; Mao, J.; Wang, R.-F. Joint Learning of Emotion and Singing Style for Enhanced Music Style Understanding. Sensors 2025, 25, 7575. https://doi.org/10.3390/s25247575

AMA Style

Chen Y, Mao J, Wang R-F. Joint Learning of Emotion and Singing Style for Enhanced Music Style Understanding. Sensors. 2025; 25(24):7575. https://doi.org/10.3390/s25247575

Chicago/Turabian Style

Chen, Yuwen, Jing Mao, and Rui-Feng Wang. 2025. "Joint Learning of Emotion and Singing Style for Enhanced Music Style Understanding" Sensors 25, no. 24: 7575. https://doi.org/10.3390/s25247575

APA Style

Chen, Y., Mao, J., & Wang, R.-F. (2025). Joint Learning of Emotion and Singing Style for Enhanced Music Style Understanding. Sensors, 25(24), 7575. https://doi.org/10.3390/s25247575

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop