You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

14 January 2023

Using Deep Learning to Recognize Therapeutic Effects of Music Based on Emotions

,
,
,
and
1
Faculty of Electrical Engineering and Computer Science, Transilvania University of Brasov, 500036 Brasov, Romania
2
Romanian Academy of Scientists, 050044 Bucharest, Romania
3
Romanian Academy of Technical Sciences, 010413 Bucharest, Romania
4
Electrical Machines and Drives Department, Technical University of Cluj Napoca, 400027 Cluj-Napoca, Romania
This article belongs to the Special Issue Machine Learning and Signal Processing Based Acoustic Sensors

Abstract

Music is important in everyday life, and music therapy can help treat a variety of health issues. Music listening is a technique used by music therapists in various clinical treatments. As a result, music therapists must have an intelligent system at their disposal to assist and support them in selecting the most appropriate music for each patient. Previous research has not thoroughly addressed the relationship between music features and their effects on patients. The current paper focuses on identifying and predicting whether music has therapeutic benefits. A machine learning model is developed, using a multi-class neural network to classify emotions into four categories and then predict the output. The neural network developed has three layers: (i) an input layer with multiple features; (ii) a deep connected hidden layer; (iii) an output layer. K-Fold Cross Validation was used to assess the estimator. The experiment aims to create a machine-learning model that can predict whether a specific song has therapeutic effects on a specific person. The model considers a person’s musical and emotional characteristics but is also trained to consider solfeggio frequencies. During the training phase, a subset of the Million Dataset is used. The user selects their favorite type of music and their current mood to allow the model to make a prediction. If the selected song is inappropriate, the application, using Machine Learning, recommends another type of music that may be useful for that specific user. An ongoing study is underway to validate the Machine Learning model. The developed system has been tested on many individuals. Because it achieved very good performance indicators, the proposed solution can be used by music therapists or even patients to select the appropriate song for their treatment.

1. Introduction

Counseling, music therapy, physical exercise, and other practices can help an individual’s mental health. Music, on the other hand, is essential in everyday life. Music could both cause and stabilize mood swings [1].
Studies have shown that listening to music has the benefits of lowering heart rate and stress hormone levels [2]; thus, it has been used to reduce stress and anxiety for many years. Although music has long been used in therapy, few intelligent/smart applications can assist and support the medical therapist in selecting the most appropriate songs for his patients. A. Raglio et al. [3] used machine learning methods to identify the main predictors of the relaxation effects of listening to music. The authors assigned approximately 300 participants to listen to random music for 9 min. However, the process was not automated because relaxation levels were recorded before and after the listening experience using a Visual Analog Scale (VAS). The effect music listening had on relaxation was therefore predicted using a decision tree.
Listening to music is a technique used by music therapists in a variety of clinical treatments [4,5]. Research conducted on the importance of listening to music shows that many health problems can be treated using music therapy. An experimental study in the field of listening to music, chosen by the participants after stress exposure, has discovered that it improves mental balance [6]. According to recent surveys conducted by music therapists from various countries, a percentage of 42.7% of practitioners in this field use music in the therapy they provide to their patients [7].
Recent advances in Artificial Intelligence and Machine Learning, particularly since 2010, have enriched the signal processing landscape by providing new tools for signal estimation, classification, prediction, and manipulation. This not only resulted in significant performance gains in various long-standing problem domains (e.g., speech, computer vision, medical diagnosis), but it also enabled the development of new classes of nonlinear functions. Open questions in fundamental signal processing include robustness, adaptivity, and performance analysis. According to B. Sandler [8], incorporating new techniques into emerging architectures will most likely result in new system-level solutions for a variety of applications, leveraging their strengths while overcoming inherent weaknesses.
Music has been shown to have physical and mental health benefits, including improved cardiovascular health, is strongly connected to reducing cases of dementia in older populations, and improves markers of general mental well-being, such as stress reduction. A UK-based research team used a spectrum scale of anxious/negative to calm/positive to assess the effects of various musical pieces on listeners [9]. They gathered the initial data from listener reports, proven to be a reliable predictor of emotional response. They later fed this information into a supervised machine learning algorithm, which predicted additional types of music with strong affective properties. Testing the new music led to a feedback loop that showed that the ML system is an efficient way of identifying songs with desired effects. They analyzed specific data from MIDI files—such as pitch, melody, timing, and dynamics—and correlated each characteristic with the scores on the positivity scale.
Previous research has not thoroughly addressed the relationship between music features and their therapeutic effects on patients using the appropriate audio signal processing. In the current experiment, the dominant emotion conveyed by a specific musical sequence was chosen using an Artificial Intelligence model. The basic emotion wheel [10] describes the types of emotions into which songs are classified. A categorical approach was used, with the music divided into groups, and each group was described with an adjective (e.g., -sad, happy, boring, etc.). The experiment aimed to create a Machine Learning model that could predict whether a specific song has therapeutic effects on a specific person.
Aside from the introduction, this paper is divided into four sections. The second presents related studies, the third describes the material and methods used in the current study, the fourth focuses on presenting the results, and the final one describes the conclusions.

3. Materials and Methods

This section describes the materials and methods used in the current paper’s experiment. The goal was to create a Machine Learning model that can predict whether a specific song has therapeutic effects on a specific person. The model will consider a person’s musical and emotional preferences, as well as the previously mentioned aspects in terms of frequencies.
The full pipeline of the experiment is the following (Figure 1):
Figure 1. Experiment Pipeline.
  • Extract audio features;
  • Exploratory Data Analysis;
  • Clean Dataset;
  • Train Initial Machine Learning Model;
  • Evaluate metrics of the Model;
  • Design & Develop Machine Learning Classifier.
Because everything has a vibration, music therapy and sound healing focus on specific frequencies. Six of these frequencies, known as solfeggio frequencies, are specific tones known since Antiquity to have a beneficial effect on the mind and body. They were used in various rituals and ceremonies from ancient India to medieval Europe. Researchers have only recently begun to solve the mystery surrounding these frequencies, including 528 Hz, a tone that has gained attention for its power of healing and emotional release [24].
Solfeggio frequencies became popular again in the 1970s. The six important frequencies thought to raise vibrations generating therapeutical effects and helping heal are the following [25,26]:
  • 96 Hz—helps people eliminate feelings such as fear, guilt, and grief;
  • 432 Hz—clears negativity and triggers a positive change. [26] shows that music tuned to 432 Hz slows down the heart rate compared to 440 Hz;
  • 528 Hz—one of the most important (also known as “love frequency”) is the frequency of transformation and DNA repair, and also helps to increase awareness;
  • 639 Hz—helps reinforce relationships and connections and increases empathy and harmony;
  • 741 Hz—known as a detoxifying frequency, it also helps solve problems;
  • 852 Hz—beneficial for spiritual self-fulfillment.
K. Akimoto et al. [24] discovered that music tuned to the frequency of 528 Hz significantly reduced stress after only a few minutes of listening. Another study [27] found that 528 Hz reduced the toxic effects of ethanol, which is the main ingredient in alcoholic beverages. Furthermore, the authors observed that this frequency increased cell life by about 20%.

3.1. Audio Feature Extraction Data Exploration and Cleaning

A typical audio processing process involves acquiring data and extracting acoustic features relevant to the problem, followed by decision-making schemes involving detection, classification, and knowledge fusion.
The musical features used in this experiment are the following:
  • Spectral characteristics:
    • Spectral centroid—the mean frequency of the signal weighted by the magnitude;
    • Spectral roll-off—how many frequencies are concentrated below a certain threshold;
    • Spectral flux—how much the frequency varies over time;
    • Mel Frequency Cepstral Coefficients (MFCCs).
  • Temporal characteristics:
    • Zero-crossing rate—the number of time domain crossings in a frame;
    • Temporal centroid;
    • Log attack time—the time required to reach the maximum amplitude of a signal from a minimum time threshold.
  • Melodic/harmonic characteristics:
    • Tone class profile;
    • The clarity of the key;
    • Harmonic change;
    • Musical module.
  • Rhythmic characteristics:
    • Beat histogram (measured in beats per minute);
    • Medium tempo.
In this experiment, the Python library Librosa was used [28]. This library includes several methods for signal processing and extraction, such as spectral and rhythmic features.
The Million Song dataset was used for this experiment. This collection is a free collection of audio features and metadata for one million pieces of contemporary popular music [29].
The extracted audio features can be seen on a spectrogram, which depicts the frequency spectrum of a signal as it changes over time. Figure 2 depicts a song’s spectrogram as a heat map, with the intensity shown by varying color gradients.
Figure 2. Spectrum of frequencies of signal (Spectrogram).
The Python script saves all computed features to a CSV file. The CSV contains one column for each of the following characteristics: tempo (in beats per minute), root-mean-square (RMS), chronogram, mel-spectrogram, spectral centroid, spectral contrast, spectral roll-off, zero-crossing rate, harmonizing, and Mel-frequency cepstral coefficients (MFCCs).
The exploratory data analysis step was performed after extracting the audio features from the dataset. The main goal of this step was to determine which characteristics could be used as strong indicators to make an accurate prediction. The distribution of each variable was graphically plotted and then analyzed (as shown in Figure 3 for average beats).
Figure 3. Audio features file.
The most relevant 10 audio features were selected after analyzing each feature, and the others were removed from the dataset.
Because labeled data is used in classification problems, these labels were applied to each audio file. The songs were divided into four categories, each represented by a different color: “energetic”, “calm”, “happy”, and “sad”. These categories were chosen based on M. Nuzzolo’s [30] article, which explains the best way to categorize music by mood.

3.2. Training the Machine Learning Model

A MinMaxScaler was used in the feature normalization process to ensure that all values between 0 and 1 were preserved, while also preserving the original shape of the data [31]. Finally, the dataset was divided, with 80% designated for training and 20% for testing.
Figure 4 depicts the entire pipeline and steps for developing the Machine Learning model.
Figure 4. ML Model Development Pipeline.
The model was built using the Keras library, designed to allow the rapid development of deep neural networks [32]. Because the main goal is to categorize songs into four mood categories (happy, sad, energetic, or calm), the ML model is a multi-class neural network. In this experiment, a KerasClassifier with a ReLU (Rectified Linear Unit) activation function was used.
Several possible values were tested and compared when selecting the input and output layers, as well as the activation functions. The layers of the developed neural network are as follows (Figure 5):
Figure 5. Neural Network Architecture.
  • An input layer with 10 audio features as input;
  • A deeply connected hidden layer with multiple nodes with a Rectified Linear Unit (ReLU) activation function;
  • An output layer containing four outputs (one for each category) with a Softwax activation function. Therefore, a classifier with an estimator role was also needed.
The estimator was evaluated using K-Fold Cross Validation. After experimenting with various possible values, the number of splits was set to K = 10. The model’s overall accuracy was 91.49%. The model was trained on 8000 samples in the current experiment.

3.3. Evaluating the Model

A confusion matrix was plotted using the Seaborn Library and Matplotlib to examine the model’s performance in detail (Figure 6). The accuracy of the model was also calculated.
Figure 6. Confusion Matrix.
With a final accuracy score of 94% and an examination of the Confusion Matrix, the model classified calm, happy, and energetic songs very well, but the accuracy for sad songs was slightly lower at 85%. It also attempted to improve the model’s accuracy by modifying some parameters, such as batch size, number of epochs, and the aggregation or deletion of some features used to train the model.
Deep Learning Algorithms can be used to implement ideas or projects involving the automation of tasks that require a significant amount of time to interpret. They can also assist in learning more about the world of data science and music taste trends.
Finally, a dedicated function was developed to predict the state conveyed by a song passed as a parameter using the previously created neural network.

3.4. Web Application for Classification

The classification is based on a dataset whose features have already been extracted by the algorithm presented in Section 3.3, using a web application developed in HTML5 with JavaScript. To perform the classification, the files containing the extracted features generated by the Python application must be provided in either CSV or JSON format. The data is also validated and, if necessary, normalized to ensure that the algorithm performs optimally.
The classification algorithm’s default parameters are defined in the JavaScript file, but they can be changed by the user via the web application’s user interface (UI). The following parameters had default values: number of epochs, learning rate, test data set size, number of hidden units, and activation function for both hidden and output layers.
These parameters are set to the values shown in Table 1 by default.
Table 1. Default parameter values.
The user interface used to configure the parameters is illustrated in Figure 7.
Figure 7. Model Training Web Application UI.
The songs are categorized when one presses the Classify button, and the results are displayed in the browser’s console. When the classifier is run, the performance parameters for each step are displayed. The accuracy and loss improved with each run, reaching a final accuracy value of more than 90% after 120 epochs.
Another web application was developed, allowing the music therapists to select (i) some characteristics for the patient, (ii) a song, and, based on the trained Machine Learning Model described above, the app will be able to indicate whether that specific song will have a therapeutical effect for the patient.
The application asks the patient about his favorite type of music and his current mood before making a prediction. If the selected song is inappropriate, the application, using Machine Learning, will recommend another type of music that may be useful for that user. Figure 8 depicts the User Interface of the Application.
Figure 8. ClassificationWeb Application UI.

4. Results and Discussions

This section summarizes the specific findings of this article and suggests opportunities and recommendations for further research. The research was carried out with the assistance of the Competence Valorization and Transfer Center (CVTC) from the Transylvania University of Brasov-Romania, in partnership with the Faculty of Electrical Engineering and the Faculty of Music at Transilvania University.
Although the developed Machine Learning model performed well in both the training and evaluation phases, it is critical to test and validate with people who are willing to use therapy as a solution to various problems. An ongoing study is being conducted to validate the Machine Learning solution, and the developed system has already been tested on a large number of people. The participants were either CVCT members or Music Therapy Master Program students, who all had different moods and musical tastes. They signed an ERB agreement to comply with the General Data Protection Regulation (GDPR). These subjects used the application on various days and in various moods, and the Machine Learning model correctly predicted and chosen in approximately 91.6% of cases.
Because previous studies have not thoroughly addressed the relationship between music features and their therapeutical effects via audio signal processing, this paper proposed a Machine Learning solution for recognizing the therapeutic effect conveyed by music. As a classifier, the algorithm described in this paper employs a multi-class neural network. It comprises an input layer with ten features, a deeply connected hidden layer with multiple hidden units, and an output layer. A web application for customizing the hyperparameters for the machine learning model, as well as another application for predicting whether a song is suitable for a specific person, was also developed.
Figure 9 depicts the train and validation loss values over all epochs. Furthermore, the model had 89% precision, 91% recall, and an F1-score of 0.899.
Figure 9. Performance indicators.
This current experiment, however, can be improved. The proposed solution has the limitation that users must select their current mood, which is somewhat subjective. If users are unsure about which mood to choose, one solution to improving the current application could be to allow them to select multiple options. In addition, future enhancements will use edge AI methods for data processing, allowing the Machine Learning Model to be deployed directly on the PSoC6 microcontroller device. The ModusToolbox environment allows one to create and run pre-trained machine-learning models on the PSoC6 directly. The model is compatible with this microcontroller and can be, technically, deployed on any IoT device because it was created using the Keras library in Python. By doing so, the entire system becomes portable and can be controlled via Bluetooth or Wi-Fi if an internet connection is available.

5. Conclusions

Music listening has long been used in clinical treatments by music therapists. As studies on the importance of listening to music have been conducted, and many health problems can be remedied with the help of music therapy, it is critical for practitioners or even patients to use an intelligent system for selecting the right music for each patient.
Because previous studies have not addressed the relationship between music features and their therapeutic effects on patients using audio signal processing, this paper attempts to address this issue. The experiment aimed to create a Machine-Learning model that could predict whether a specific song has therapeutic effects on a specific person. The model was trained to consider the solfeggio frequencies as well as the characteristics of a specific person in terms of music and emotions. A section of the freely available Million Dataset was used to train the machine learning model.
According to Section 3, the model achieved very good performance indicators and an overall accuracy of more than 94%. There is also an ongoing validation process for people who want to use music therapy to treat their problems. So far, it has been tested on several people with positive results. As a result, the proposed solution can be used by therapists and others who want to benefit from the therapeutic effects of music.

Author Contributions

Conceptualization, H.A.M., T.C. and D.U.; methodology, C.S. and H.H.; software, H.A.M.; validation, D.U., C.S. and H.H.; formal analysis, H.A.M. and D.U.; investigation, H.A.M.; resources, H.A.M., T.C. and D.U.; data curation, H.A.M.; writing—original draft preparation, H.A.M. and T.C.; writing—review and editing, D.U. and C.S.; visualization, D.U., C.S. and H.H.; supervision, D.U.; project administration, C.S.; funding acquisition, D.U. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

We would like to express our deep appreciation to the Cypress/Infineon company for providing us with free PSoC6 kits, facilitating this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jensen, K.; Ystad, S.; Kronland-Martinet, R. Computer Music Modeling and Retrieval. In Proceedings of Sense of Sounds: 4th International Symposium, CMMR, Copenhagen, Denmark, August 2007; Lecture Notes in Computer Science. Springer: Warsaw, Poland, 2007. [Google Scholar]
  2. Bardekar, A.; Gurjar, A.A. Study of Indian Classical Ragas Structure and its Influence on Hu-man Body for Music Therapy. In Proceedings of the 2nd International Conference on Applied and Theoretical Computing and Communication Technology (iCATccT), Bangalore, India, 21–23 July 2016; pp. 119–123. [Google Scholar] [CrossRef]
  3. Raglio, A.; Imbriani, M.; Imbriani, C.; Baiardi, P.; Manzoni, S.; Gianotti, M.; Castelli, M.; Vanneschi, L.; Vico, F.; Manzoni, L. Machine learning techniques to predict the effectiveness of music therapy: A randomized controlled trial. Comput. Methods Programs Biomed. 2020, 185, 105160. [Google Scholar] [CrossRef] [PubMed]
  4. Pérez-Ros, P.; Cubero-Plazas, L.; Mejías-Serrano, T.; Cunha, C.; Martínez-Arnau, F.M. Preferred Music Listening Intervention in Nursing Home Residents with Cognitive Impairment: A Randomized Intervention Study. J. Alzheimers Dis. 2019, 70, 433–442. [Google Scholar] [CrossRef] [PubMed]
  5. de Witte, M.; Spruit, A.; van Hooren, S.; Moonen, X.; Stams, G.J. Effects of Music Interventions on Stress-Related Outcomes: A Systematic Review and Two Meta-Analyses. Health Psychol Rev. 2019, 14, 294–324. [Google Scholar] [CrossRef] [PubMed]
  6. Hohmann, L.; Bradt, J.; Stegemann, T.; Koelsch, S. Effects of music therapy and music-based interventions in the treatment of substance use disorders: A systematic review. PLOS One 2017, 12, e0187363. [Google Scholar] [CrossRef] [PubMed]
  7. Kern, P.; Tague, D.B. Music Therapy Practice Status and Trends Worldwide: An International Survey Study. J. Music. Ther. 2017, 54, 255–286. [Google Scholar] [CrossRef] [PubMed]
  8. Sandler, B. AI: A Signal Processing Perspective. Available online: https://sigport.org/documents/ai-signal-processing-perspective (accessed on 15 August 2022).
  9. Williams, D.; Hodge, V.; Wu, C. On the use of AI for Generation of Functional Music to Improve Mental Health. Front. Artif. Intell. 2020, 3, 497864. [Google Scholar] [CrossRef] [PubMed]
  10. Emotion Wheel. Available online: https://positivepsychology.com/emotion-wheel/ (accessed on 1 December 2022).
  11. Manaris, B.; Krehbiel, D.; Roos, P.; Zalonis, T. Armonique: Experiments In Content-based Similarity Retrieval Using Power-law Melodic And Timbre Metrics, ISMIR 2008—Session 3a—Content-Based Retrieval. Categ. Similarity 2018, 343–348. [Google Scholar]
  12. Hung, H.T.; Ching, J.; Doh, S.; Kim, N.; Nam, J.; Yang, Y.H. EMOPIA: A multi-modal pop piano dataset for emotion recognition and emotion-based music generation, 2021. In Proceedings of the International Society for Music Information Retrieval Conference 2021, Online, 7–12 November 2021. [Google Scholar] [CrossRef]
  13. Lee, H.; Hoeger, F.; Schoenwiesner, M.; Park, M.; Jacoby, N. Cross-cultural mood perception in pop songs and its alignment with mood detection algorithms, 2021. In Proceedings of the International Society for Music Information Retrieval Conference 2021, Online, 7–12 November 2021. [Google Scholar] [CrossRef]
  14. Chaki, S.; Doshi, P.; Bhattacharya, S.; Patnaik, P. Explaining Perceived Emotion Predictions in Music: An Attentive Approach, 2020. In Proceedings of the 21st International Society for Music Information Retrieval Conference 2020, Virtual, 11–16 October 2020. [Google Scholar] [CrossRef]
  15. de Berardinis, J.; Cangelosi, A.; Coutinho, E. The multiple voices of musical emotions: Source separation for improving music emotion recognition models and their interpretability, 2020. In Proceedings of the 21st International Society for Music Information Retrieval Conference 2020, Virtual, 11–16 October 2020. [Google Scholar] [CrossRef]
  16. Korzeniowski, F.; Nieto, O.; McCallum, M.; Won, M.; Oramas, S.; Schmidt, E. Mood classification using listening data, 2020. In Proceeding of the International Society for Music Information Retrieval Conference 2020, Virtual, 11–16 October 2020. [Google Scholar] [CrossRef]
  17. Greer, T.; Ma, B.; Sachs, M.; Habibi, A.; Narayanan, S. A Multimodal View into Music’s Effect on Human Neural, Physiological, and Emotional Experience. In Proceedings of the 27th ACM International Conference on Multimedia (MM’19). Association for Computing Machinery, New York, NY, USA, 21–25 October 2019; pp. 167–175. [Google Scholar] [CrossRef]
  18. Kim, Y.E.; Schmidt, E.M.; Migneco, R.; Morton, B.G.; Richardson, P.; Scott, J.; Speck, J.A.; Turnbull, D. Music emotion recognition: A state of the art review. Proc. Ismir. 2010, 86, 937–952. [Google Scholar]
  19. Delbouys, R.; Hennequin, R.; Piccoli, F.; Royo-Letelier, J.; Moussallam, M. Music mood detection based on audio and lyrics with deep neural net. arXiv 2018, arXiv:1809.07276. [Google Scholar]
  20. Xu, L.; Sun, Z.; Wen, X.; Huang, Z.; Chao, C.; Xu, L. Using machine learning analysis to interpret the relationship between music emotion and lyric features. PeerJ Comput. Sci. 2021, 7, e785. [Google Scholar] [CrossRef] [PubMed]
  21. Cowie, R.; Douglas-Cowie, E.; Tsapatsoulis, N.; Votsis, G.; Kollias, S.; Fellenz, W.; Taylor, J. Emotion recognition in human-computer interaction, 2001. IEEE Signal Process. Mag. 2001, 18, 32–80. [Google Scholar] [CrossRef]
  22. Hoang, M.-H.; Kim, S.-O.; Yang, H.-J.; Lee, G.-S. Context-Aware Emotion Recognition Based on Visual Relationship Detection. IEEE Access 2021, 9, 90465–90474. [Google Scholar] [CrossRef]
  23. Ramet, G.; Garner, P.N.; Baeriswyl, M.; Lazaridis, A. Context-Aware Attention Mechanism for Speech Emotion Recognition. In Proceedings of the 2018 IEEE Spoken Language Technology Workshop (SLT), Athens, Greece, 18–21 December 2018; pp. 126–131. [Google Scholar] [CrossRef]
  24. Akimoto, K.; Hu, A.; Yamaguchi, T.; Kobayashi, H. Effect of 528 Hz Music on the Endocrine System and Autonomic Nervous System. Health 2018, 10, 1159–1170. [Google Scholar] [CrossRef]
  25. Nakajima, Y.; Tanaka, N.; Mima, T.; Izumi, S.I. Stress Recovery Effects of High- and Low-Frequency Amplified Music on Heart Rate Variability. Behav. Neurol. 2016, 2016, 1–8. [Google Scholar] [CrossRef] [PubMed]
  26. Calamassi, D.; Pomponi, G.P. Music Tuned to 440 Hz Versus 432 Hz and the Health Effects: A Double-blind Cross-over Pilot Study; Elsevier(Science Direct): Amsterdam, The Netherlands, 2019; Volume 15, pp. 283–290. ISSN 1550-8307. [Google Scholar] [CrossRef]
  27. Babayi, T.; Riazi, G.H. The Effects of 528 Hz Sound Wave to Reduce Cell Death in Human Astrocyte Primary Cell Culture Treated with Ethanol. J. Addict. Res. Ther. 2017, 8, 1–5. [Google Scholar] [CrossRef]
  28. Librosa Library Documentation. Available online: https://librosa.org/doc/latest/index.html (accessed on 25 November 2022).
  29. Bertin-Mahieux, T.; Ellis, D.; Whitman, B.; Lamere, P. The Million Song Dataset. In Proceedings of the 12th International Society for Music Information Retrieval Conference, Miami, Florida, USA, 24–28 October 2011. [Google Scholar] [CrossRef]
  30. Nuzzolo, M. Music Mood Classification, Electrical and Computer Engineering Design Handbook; 2015; Chapter 3. Available online: https://sites.tufts.edu/eeseniordesignhandbook/2015/music-mood-classification/ (accessed on 15 December 2022).
  31. Raymaekers, J.; Zamar, R. Pooled variable scaling for cluster analysis. Bioinformatics 2020, 36, 849–3855. [Google Scholar] [CrossRef] [PubMed]
  32. Keras API Documentation. Available online: https://keras.io/ (accessed on 13 December 2022).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.