You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

14 June 2023

HapticSOUND: An Interactive Learning Experience with a Digital Musical Instrument

,
,
,
,
and
1
Department of Applied Informatics, University of Macedonia, 54636 Thessaloniki, Greece
2
Department of Industrial Engineering and Management, International Hellenic University, 57001 Thessaloniki, Greece
3
Medical Informatics Laboratory, Department of Medicine, Democritus University of Thrace, 68100 Alexandroupolis, Greece
*
Author to whom correspondence should be addressed.

Abstract

In this paper, an interactive learning experience is proposed, aiming to involve museum visitors in a personalized experience of the transmittal of cultural knowledge in an active and creative way. The proposed system, called HapticSOUND, consists of three subsystems: (a) the Information, where visitors are informed about the traditional musical instruments; (b) the Entertainment, where visitors are entertained by playing serious games to virtually assemble traditional musical instruments by a set of 3D objects; and (c) the Interaction, where visitors interact with a digital musical instrument which is an exact 3D-printed replica of a traditional musical instrument, where cameras have been placed to capture user gestures and machine learning algorithms have been implemented for gesture recognition. The museum visitor can interact with the lifelike replica to tactilely and aurally explore the instrument’s abilities, producing sounds guided by the system and receiving real-time visual and audio feedback. Emphasis is given to the Interaction Subsystem, where a pilot study was conducted to evaluate the usability of the subsystem. Preliminary results were promising since the usability was satisfactory, indicating that it is an innovative approach that utilizes sensorimotor learning and machine learning techniques in the context of playing sounds based on real-time gesture and fingering recognition.

1. Introduction

Music has evolved tremendously in recent years, with new technologies and instruments being developed to create new and innovative sounds. One of the most significant developments in the music industry has been the introduction of Digital Musical Instruments (DMIs) [1,2]. DMIs are devices that produce sounds through digital processing rather than through mechanical or acoustic means. The evolution of DMIs can be traced back to the 1920s when the first digital synthesizer, called Thereminvox, was developed by Leon Theremin, who used hand movements to control the pitch and volume of the sound produced. Since then, the development of DMIs has been rapid and dynamic, with each generation of instruments introducing new sounds and possibilities for musicians and producers. The introduction of MIDI in the 1980s revolutionized the music industry by enabling musicians to connect different instruments and equipment and to control various parameters of DMIs through physical movements, leading to the development of more sophisticated DMIs. Today, DMIs come in different forms, including keyboards, guitars, drums, percussion, wind, brass, and stringed instruments. DMIs are designed to mimic the sound of their acoustic counterparts, but they offer more flexibility and versatility in terms of sound production. One of the significant advantages of DMIs is that they can be easily customized and reprogrammed to suit the specific needs of the user [3]. This customization allows for greater control over the sound production process and encourages experimentation with sound design. In addition, gesture-controlled DMIs allow for greater expressivity and creativity in music performance [4]. Musicians can use their physical movements to shape and manipulate the sound produced by the instrument, creating unique and dynamic performances. Gesture also allows for a more intuitive and natural way of interacting with DMIs [5]. Therefore, gesture-controlled DMIs have changed the way humans create, perform, and interact with music.
However, learning to play DMIs comes with some challenges, including the need for musicians to develop new motor skills and new perceptual abilities [6]. Additionally, because DMIs can produce a wide range of sounds, learning to control and manipulate these sounds with their gestures can be overwhelming for beginners [7], since they need to integrate multiple sensory modalities. Apart from that, DMIs allow a high degree of flexibility in mapping gestures to sound. This flexibility can lead to complex mappings where the relationship between a specific gesture and its corresponding sound output may not be immediately intuitive or straightforward [8,9]. Finally, the gesture-based control of DMIs also involves the use of technology, which can be challenging for individuals.
In this paper, an innovative approach is proposed to alleviate the aforementioned limitations, called HapticSOUND, which provides a unique interactive experience to museum visitors about traditional musical instruments. The approach utilizes sensorimotor learning, which refers to the process by which individuals learn to use their sensory and motor systems to interact with the environment [10]. Specifically, through the Interaction Subsystem, users perform gestures (motor skills) on a 3D-printed DMI to produce sounds. In addition, the subsystem has integrated sensory feedback, such as visual, auditory, and annotated information, related to the performed gestures, and provides levels with an increasing degree of difficulty for acquiring and refining motor skills based on sensory feedback. Apart from this subsystem, HapticSOUND provides the Information Subsystem where users learn about traditional musical instruments, as well as the Entertainment Subsystem where they play a 3D gesture-based puzzle game. The rest of the paper is organized as follows: In Section 2, the related work is provided. Next, in Section 3, the description of the HapticSOUND system is introduced, and three subsystems are described, placing more emphasis on the Interaction Subsystem (Section 4). In Section 5, the pilot study to evaluate the usability of the Interaction Subsystem is reported, and the results are discussed in Section 6. Finally, Section 7 concludes the paper and presents limitations and future work.

3. HapticSOUND Architecture

The HapticSOUND project aims at creating an analysis, documentation, and demonstration system for traditional musical instruments in museums. HapticSOUND will be established in the Cretan Ethnology Museum (CEM) in Greece, which involves visitors in a personalized experience of the transmittal of cultural knowledge in an active, pleasant, and creative way. The proposed system aims to act as a bridge between the visitors of the CEM museum and the cultural heritage of Crete (Greece). To achieve this, the proposed system involves the creation of a demonstrative system for informing CEM visitors about the musical instruments exhibited in the museum, promoting their familiarization with the sounds they produce, the manner they create sounds, and how they are constructed. The visitor’s participation is supported in experiential learning by providing opportunities for visitors to learn how to play the 3D-printed DMI, evolving from simple note creation to musical patterns. Moreover, visitors have the opportunity to study the structural model of the musical instrument in the context of a 3D serious game, in which they have to assemble it.
The next sections provide an in-depth explanation of the HapticSOUND system and its subsystems. The architecture is shown in Figure 1 and consists of three subsystems: (1) the Information Subsystem (provides audio tracks and informative text), (2) the Entertainment Subsystem (provides a 3D gesture-based serious puzzle game and gamification elements), and (3) the Interaction Subsystem (provides the user interaction with an exact 3D-printed replica of a traditional musical instrument).
Figure 1. HapticSOUND Architecture.
The user interacts with HapticSOUND and its subsystems through the user interface, which has been implemented in the Unity game engine (https:/unity.com, accessed on 10 July 2022), and the use of the Leap Motion Controller (https://www.ultraleap.com, accessed on 22 January 2023), with which the system captures and interprets the user’s gestures (hands and fingers). For the user’s convenience, virtual hands are also placed in the 3D space (Figure 2), following the user’s gestures in real time and providing visual feedback to him/her. In the next subsections, the three subsystems are briefly described.
Figure 2. (a) HapticSOUND system and museum installation; (b) the user interface with the three subsystems and the visualization of the user’s hand.

3.1. Information Subsystem

The purpose of the Information Subsystem is to inform museum visitors about the traditional Cretan musical instruments. Specifically, it allows the user to display text, videos, and 3D representations of traditional musical instruments on the screen (i.e., lute and oud), navigate collections and digital exhibitions, and listen to samples of traditional musical instruments. The architecture of the Information Subsystem is presented in Figure 3.
Figure 3. Information Subsystem.
The basic unit of the Information Subsystem is the Content Management System (CMS), located in the central system of HapticSOUND along with the Information Subsystem database. The CMS was created using the OMEKA software v3.1.1 (https://omeka.org/, accessed on 5 November 2022), which is an open-source platform specifically designed for creating, managing, searching, and publishing artifacts, digital exhibitions, and collections in museums, libraries, and exhibitions. OMEKA provides the ability to create and present exhibitions and collections consisting of digital artifacts from various digital collections and offers a fully customizable viewing environment. OMEKA allows data and metadata import and export for all required file types and provides the ability to extend its capabilities through the installation of plugins. Finally, using the OMEKA platform ensures interoperability with other systems through the use of established standards (Figure 4).
Figure 4. OMEKA software is embedded in the user interface of the Information Subsystem.
The user interface unit is the public part of the Information Subsystem through which the user navigates to the museum’s digital collections and exhibitions, searches for information, views photos and texts, and plays videos and music. The Information Subsystem operates in a web browser environment; therefore, the WebView is responsible for loading and displaying the CMS’s web content in the HapticSOUND system environment, which has been implemented in the Unity game engine.
The Administrator Interface unit is the private part of the Information Subsystem, where users with privileged rights (museum staff, project members, etc.) execute management actions, such as importing objects, collections, and plugins, as well as creating and managing user accounts.

3.2. Entertainment Subsystem

The purpose of the Entertainment Subsystem is to engage visitors in creative activities while they are entertained. To achieve this, the Entertainment Subsystem includes engaging game mechanics (such as puzzles) and gamification elements (such as leaderboards). Specifically, the Entertainment Subsystem is a digital 3D gesture-based serious puzzle game, which employs the Leap Motion controller to receive gestural input from the user. The visitor virtually assembles the traditional Cretan lute that is disassembled into its component parts. Therefore, to solve the puzzle, the visitor has to identify the lute’s parts, reflect on the lute’s structural model, and interact with the virtual representations of the lute’s parts to gain an in-depth understanding of its structure and functionality. Nevertheless, the Entertainment Subsystem aims to foster positive feelings in the users by engaging them in an amusing and productive activity, performed in a realistic and attractive 3D environment. The architecture of the Entertainment Subsystem is presented in Figure 5.
Figure 5. Entertainment Subsystem [20].
Apart from the virtual representations of the lute’s parts, the Entertainment Subsystem contains the following sub-units: (a) the Configuration sub-unit, which sets the difficulty level of the game, (b) the User Feedback sub-unit, which allows the user to evaluate his/her experience, (c) the Help sub-unit, which provides help to the user by explaining the purpose of the game and the manner it is played, (d) the Leaderboards sub-unit, which manages and displays the results board showing the players with the best performance, and (e) the Game Manager sub-unit, which coordinates the game’s operations. The Game Manager generates the session data that include statistics, measurements, and scores, while the User Feedback sub-unit creates the rating data that are stored in a back-end storage facility (i.e., a database) of the HapticSOUND Central System. The Entertainment Subsystem was implemented in the Unity 3D game development engine with C# as the programming language, as was the Information Subsystem (Figure 6).
Figure 6. The User Interface of Entertainment Subsystem along with the 3D gesture-based serious puzzle game.

3.3. Interaction Subsystem

The purpose of the Interaction Subsystem is to enable visitors to interact with the DMI and generate sounds using a sensorimotor approach. Specifically, it includes a DMI that is an accurate replica of the Traditional Musical Instrument (TMI replica) in physical dimensions, where two Time-of-Flight (ToF) cameras have been placed on the TMI replica, as shown in Figure 7, to capture, process, and recognize the user’s gestures.
Figure 7. TMI replica with 2 ToF cameras.
The Interaction Subsystem offers two scenarios of use: Case 1—the user performs single musical notes and receives auditory feedback in real time; and Case 2—the user performs a music excerpt, imitating the technical performance of an expert music teacher, and receives visual and auditory feedback in real time. The architecture of this Subsystem is presented in Figure 8.
Figure 8. Interaction Subsystem.
The user interacts with the 3D-printed DMI, which is the TMI replica (Figure 7), while cameras detect his/her gestures through algorithms, which are processed by the Raspberry Pi Subsystem. Then, these data are sent to the central system of HapticSOUND via Wi-Fi in order for the performed gesture to be recognized. To achieve this, the system was trained beforehand with expert gestures by applying machine learning algorithms and then these gestures were mapped to sounds that have been stored in the HapticSOUND database. As a result, the user executes a gesture and the system recognizes the performed gesture and simultaneously produces sounds. Apart from the audio feedback, the system also provides visual feedback to the user.

4. Implementation of the Interaction Subsystem

In this section, the implementation of the Interaction Subsystem is described in more detail. As already mentioned in Section 3.3, two ToF cameras are used to capture the user’s gestures. Camera 1 records and monitors the distance of the instrumentalist’s left hand touching the keyboard (we consider the performer to be right-handed), while camera 2 records the tempo given by the instrumentalist with his/her right hand (Figure 9).
Figure 9. ToF Camera Placement on the Music Instrument’s Replica.
These ToF cameras are IFM3D O3X101, which provide point clouds, depth maps, and grayscale images. The resolution of the camera is 224 × 172 pixels with corresponding viewing angles of 60° × 45°. The IFM3D cameras capture the user’s finger gestures. A Region of Interest (ROI) in the view field of each camera is specified (Figure 10) and the captured data corresponding to the ROI are processed by the Raspberry Pi 4 microcontroller. For finger gesture recognition, an algorithm was developed in Python, utilizing the OpenCV library. The algorithm runs on the microcontroller in order to extract features such as position coordinates (including depth), angles, velocity, etc., which are the input data to the system.
Figure 10. (a) ROI image from ToF camera in 2D, (b) ROI image from ToF camera in 3D (capturing the depth).
In addition, the user interface was implemented in the Max/MSP programming environment, which is one of the most popular software environments for creating interactive and real-time DMIs [5,13,15] that can be used to recognize a wide range of hand gestures. The user interface was then embedded into Unity, through the use of mira.frame (https://docs.cycling74.com/max7/refpages/mira.frame, accessed on 7 May 2023), where the Interaction System runs (Figure 11).
Figure 11. (a) User interface of the interaction subsystem; (b) visual feedback (green annotation and percentage for the correct gestures).
For Case 1 (performing single musical notes and receiving auditory feedback in real time), initially, all single notes are recorded while the expert music teacher performs with the real traditional musical instrument and stored in the HapticSOUND Database. Next, the user performs fingerings on the TMI replica, where his/her gestures are captured. These gestural data, which correspond to predefined values of variables A, B, and C (Figure 9), are mapped to single notes each time the user’s hands take a specific position. To achieve this, the mapping that has been implemented is a well-known technique called direct (explicit) mapping [21] and one-to-one, where an input parameter is mapped to a composition parameter. As a result, single notes are produced (auditory feedback) according to the user’s fingerings. For example, when the distance of the left hand is A1 with a specific finger arrangement B1 (output by camera 1) and the rhythm given by the right hand is C1 (output by camera 2), the sound of note D1, which was recorded from the corresponding position and is in the database, is triggered. Summarizing, in this case, the intuitive and straightforward nature of direct and one-to-one mapping simplifies the learning process for users using the DMI because it allows them to quickly establish a connection between their gestures and the sound they want to produce, facilitating a more natural and expressive interaction with the instrument.
For Case 2 (performing a music excerpt and receiving visual and auditory feedback in real time), both the gestural data and the corresponding audio excerpt are captured while the expert music teacher performs with the real traditional musical instrument. These data feed the machine learning algorithm (training phase) and, as a result, gestural data are mapped to audio excerpts through the technique of implicit mapping [22]. Then, the user (novice) tries to perform the same gestures on the TMI replica as the expert music teacher, in order to reproduce in real time the same music sound. To achieve this, both the real video of the expert music teacher as well as an annotated fretboard with yellow annotations are displayed to the user in order to place his/her fingers in the correct positions (Figure 11a). Therefore, the user imitates a gesture, and cameras record his/her gesture and send the corresponding data to the system via Wi-Fi. Then, the system recognizes the performed gesture and predicts the music sound (recognition phase) that corresponds to the specific gesture by comparing the performed gesture to that of the expert. The music sound is generated by the system in real time and depends on how much the user’s gestural data converge or diverge from that of the expert (audio feedback), which helps users to detect and correct errors in their performance. As they explore and experiment with different gestures, they receive feedback that indicates whether their gestures produce the desired sound. By comparing the intended outcome with the actual result, they can identify discrepancies and adjust their gestures accordingly. Over time, this feedback loop helps refine motor skills and improve the accuracy of the mapping between gestures and sound. The system also provides visual feedback, in addition to audio feedback, by displaying green annotations for the correct gesture as well as the total percentage of the similarity of the user’s performed gesture in relation to the gesture of the expert (Figure 11b). Case 2 provides three levels with an increasing degree of difficulty and, by repeatedly performing gestures and observing the corresponding auditory or visual feedback, users learn to associate specific gestures with desired musical outcomes. Finally, the machine learning algorithms that were implemented are Hidden Markov Models (HMM) [23] and Dynamic Time Warping (DTW) [24]. The advantage of DTW is that it allows the time alignment between the gesture of the user and of the expert, depending on the speed of the gesture that is performed by the user. Specifically, Gesture Follower was used, developed by IRCAM [25,26], which implements both HMM and DTW. Additional advantages are that Gesture Follower adopts a template-based method and applies training with a single sample (one-shot learning), minimizing the training process. Summarizing, in this case, a sensorimotor approach is implemented to contribute to the learning process through the integration of feedback, error detection, and correction. It involves a dynamic process of acquiring, integrating, and refining motor skills based on sensory feedback since it helps users develop an embodied understanding of the DMI, enhancing their ability to control and express themselves through the mapping between gestures and sound.

5. Pilot Study

This section presents the small-scale preliminary study that was conducted to evaluate the usability of the Interaction Subsystem. To achieve this, the System Usability Scale (SUS) questionnaire was selected [27], which is one of the most well-known instruments for usability assessment, and it provides reliable results even with a small sample size [28]. The questionnaire consists of 10 statements, with a five-point Likert scale ranging from “Strongly Agree” to “Strongly Disagree”. Scores are assigned based on the response, with some statements receiving reverse scoring. The scores for all 10 statements are then summed to obtain a total score between 0 and 100, with higher scores indicating higher perceived usability. The SUS score can be interpreted in the context of the study, with scores above 68 considered average [27].

5.1. Participants

A one-day workshop was organized at the University of Macedonia (Greece) to evaluate the Interaction Subsystem, where, in total, fourteen (n = 14) participants took part in the study, who were students, researchers, and academics with varying levels of musical experience from the Department of Music Science & Art and the Department of Applied Informatics. The sample comprised nine (n = 9) men and five (n = 5) women with an average age of 35 years. In addition, 36% (n = 5) of the subjects had a senior level of musical experience, 21% (n = 3) of them had mid-level experience, 29% (n = 4) had an intermediate level, while the rest 14% (n = 2) had entry-level musical experience.

5.2. Procedure

The pilot study included two phases corresponding to the two cases that the Interaction Subsystem provides. More specifically, in Phase 1 (Case 1), each user experimented with the TMI replica in order to execute single musical notes and explore the possibilities of the TMI replica. The duration of Phase 1 was approximately 2 min per user.
Then, the user moved on to Phase 2 (Case 2), where s/he performed on the TMI replica three music excerpts of a famous Cretan traditional song named “Erotokritos”. The three music excerpts corresponded to the three levels of sensorimotor learning that were of increasing difficulty. The user moved to the next excerpt only if s/he passed the score of 60% in the previous level. An annotated fretboard was displayed to him/her, which s/he had to follow by imitating the gesture of the expert music teacher in order to produce the proper music sound. The duration of Phase 2 ranged from approximately 2 to 5 min depending on the pace and the success level of each user.

6. Results and Discussion

The overall SUS score for the usability assessment of the Interaction Subsystem is 70, which is considered acceptable and can be interpreted, as presented in Figure 12. It is worth noting that in Figure 12, the positive meaning of each question was taken into account, and a number close to 5 indicates user satisfaction. Through the analysis of users’ responses to each question, the most important usability weaknesses are related to “Background Knowledge”, “User Confidence”, and “Need for Tech Support”.
Figure 12. Analysis of SUS Score per SUS Question.
The above findings can be further supported by the SUS score per user, which is depicted in Figure 13.
Figure 13. SUS Score per Participant (U1–U14).
Specifically, the two users who had entry-level musical experience (U13 and U14) were not very confident in holding the TMI replica, placing their hands in the correct positions, and looking at the annotated fretboard at the same time (“User Confidence”). As a result, they preferred to have the support of a technical person in order to be able to successfully perform the musical gestures (“Need for Tech Support”). In addition, some of the users either with high or low musical experience (U1, U2, U5, U10, U13, and U14) agreed that there were many things that needed to be learned for the first time (“Background Knowledge”). The expert musicians claimed that although they know how to play a real traditional musical instrument, it was not the same as the 3D-printed TMI replica. It is worth mentioning that despite encountering usability issues, all participants were able to successfully complete the three music excerpts (levels) of “Erotokritos”. Although errors occurred during the execution of the gestures and the produced sound, with the aid of audio and visual feedback, all participants were able to make corrections and align their performed gestures with the expert gestures.

7. Conclusions and Future Work

Leveraging the aforementioned, this paper presents an approach called HapticSOUND that provides a unique interactive experience to museum visitors with traditional musical instruments. HapticSOUND comprises three subsystems: (a) the Information Subsystem, where users can learn information about traditional musical instruments, (b) the Entertainment Subsystem, where users learn to virtually assemble traditional musical instruments through a 3D serious game, and (c) the Interaction Subsystem, where users interact with a DMI which is a 3D TMI replica. Emphasis is given to the design and implementation of the Interaction Subsystem which utilizes sensorimotor learning and machine learning techniques to enhance the user experience. Users can explore the abilities of the 3D-printed TMI replica in a more engaging and intuitive way. In addition, a preliminary pilot study was conducted on 14 participants in order to evaluate the usability of the Interaction Subsystem. The findings were promising since the usability was satisfactory, revealing that the proposed approach can provide added value to the implementation of similar future endeavors.
However, there are certain limitations of the present study that should be addressed and improved upon in future research. Firstly, although the analysis is robust in small samples [28], a larger-scale experiment could be necessary to further investigate the usability of the system and extract more valuable conclusions. Moreover, in order to enhance the generalizability of the validation process and to mitigate the potential impact of the limited sample size in the present study, it would be beneficial to include a more diverse and representative sample of the population of interest. Another important aspect that has to be taken into consideration is the investigation of whether the learning outcomes and motor skills of the user will be enhanced with the use of the Interaction Subsystem. Finally, the HapticSOUND system is planned to be established in the Cretan Ethnology Museum (CEM), where all the aforementioned improvements in the evaluation of the subsystem will be accomplished in the near future as well as some modifications might also be required to ensure its effective application in real conditions.

Author Contributions

Conceptualization, C.V., A.T., D.T., T.I., M.N.K. and I.M.; methodology, C.V., A.T., D.T., T.I., M.N.K. and I.M.; software, C.V., D.T. and M.N.K.; validation, C.V.; writing—original draft preparation, C.V., D.T. and T.I.; writing—review and editing, C.V., A.T., D.T., M.N.K. and I.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been co-financed by the Operational Programme Competitiveness Entrepreneurship Innovation (EPANEK) 2014–2020—(European Regional Development Fund of the European Union and by national resources), grant number: MIS 5066859—and it was implemented through the General Secretariat for Research and Innovation (GSRI) of the Greek Ministry of Development and Investments.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Miranda, E.; Wanderley, M. New Digital Musical Instruments: Control and Interaction beyond the Keyboard; AR Editions: Madison, WI, USA, 2006. [Google Scholar]
  2. O’Modhrain, S. A framework for the evaluation of digital musical instruments. Comput. Music. J. 2011, 35, 28–42. [Google Scholar] [CrossRef]
  3. Malloch, J.; Sinclair, S.; Wanderley, M.M. A Network-Based Framework for Collaborative Development and Performance of Digital Musical Instruments. In Computer Music Modeling and Retrieval. Sense of Sounds. CMMR 2007; Kronland-Martinet, R., Ystad, S., Jensen, K., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2008; Volume 4969. [Google Scholar] [CrossRef]
  4. Malloch, J.; Wanderley, M.M. Embodied Cognition and Digital Musical Instruments. In The Routledge Companion to Embodied Music Interaction; Lesaffre, M., Maes, P.-J., Leman, M., Eds.; Routledge Handbooks Online: New York, NY, USA, 2017. [Google Scholar]
  5. Volioti, C.; Manitsaris, S.; Hemery, E.; Hadjidimitriou, S.; Charisis, V.; Hadjileontiadis, L.; Katsouli, E.; Moutarde, F.; Manitsaris, A. A Natural User Interface for Gestural Expression and Emotional Elicitation to Access the Musical Intangible Cultural Heritage. J. Comput. Cult. Herit. 2018, 11, 1–20. [Google Scholar] [CrossRef]
  6. Malloch, J.; Birnbaum, D.; Sinyor, E.; Wanderley, M.M. Towards a New Conceptual Framework for Digital Musical Instruments. In Proceedings of the 9th International Conference on Digital Audio Effects, Montreal, QC, Canada, 18–20 September 2006; pp. 49–52. [Google Scholar]
  7. Alber, N.; Alonso, M.; Geiger, G.; Jordà, S. The Reactable: A collaborative musical instrument for playing and understanding music. Herit. Mus. 2010, 4, 36–43. [Google Scholar]
  8. Fels, F.; Gadd, A.; Mulder, A. Mapping Transparency Through Metaphor: Towards More Expressive Musical Instruments. Organ. Sound 2002, 7, 109–126. [Google Scholar] [CrossRef]
  9. Schloss, W.A. Using Contemporary Technology in Live Performance: The Dilemma of the Performer. J. New Music. Res. 2003, 32, 239–242. [Google Scholar] [CrossRef]
  10. Wolpert, D.M.; Diedrichsen, J.; Flanagan, J.R. Principles of sensorimotor learning. Nat. Rev. Neurosci. 2011, 12, 739–751. [Google Scholar] [CrossRef] [PubMed]
  11. Turchet, L.; McPherson, A.P.; Barthet, M. Real-Time Hit Classification in a Smart Cajón. Front. ICT 2018, 5, 16. [Google Scholar] [CrossRef]
  12. Erdem, Ç.; Lan, Q.; Jensenius, A.R. Exploring relationships between effort, motion, and sound in new musical instruments. Hum. Technol. 2020, 16, 310–347. [Google Scholar]
  13. Rasamimanana, N.; Bevilacqua, F.; Schnell, N.; Guedy, F.; Flety, E.; Maestracci, C.; Zamborlin, B.; Frechin, J.; Petrevski, U. Modular musical objects towards embodied control of digital music. In Proceedings of the Fifth International Conference on Tangible, Embedded, and Embodied Interaction. Association for Computing Machinery, New York, NY, USA, 23–26 January 2011; pp. 9–12. [Google Scholar] [CrossRef]
  14. Gillian, N.E.; Paradiso, J.A. Digito: A Fine-Grain Gesturally Controlled Virtual Musical Instrument. In Proceedings of the International Conference on New Interfaces for Musical Expression, Ann Arbor, MI, USA, 21–23 May 2012. [Google Scholar] [CrossRef]
  15. Dalmazzo, D.; Ramirez, R. Air violin: A machine learning approach to fingering gesture recognition. In Proceedings of the 1st ACM SIGCHI International Workshop on Multimodal Interaction for Education, New York, NY, USA, 13 November 2017; pp. 63–66. [Google Scholar] [CrossRef]
  16. Hofmann, A. Study for virtual keyboard instrument and hand tracking in a VR environment. In Proceedings of the International Conference on New Interfaces Musical Expression, Art installation for NIME 2021, Shanghai, China, 14–18 June 2021. [Google Scholar] [CrossRef]
  17. Lee, M. Entangled: A Multi-Modal, Multi-User Interactive Instrument in Virtual 3D Space Using the Smartphone for Gesture Control. In Proceedings of the International Conference on New Interfaces Musical Expression, Shanghai, China, 14–18 June 2021. [Google Scholar] [CrossRef]
  18. Mittal, A.; Gupta, R. MuTable (Music Table): Turn any surface into musical instrument. arXiv 2022, arXiv:2212.14125. [Google Scholar]
  19. Bresin, R.; Elblaus, L.; Frid, E.; Favero, F.; Annersten, L.; Berner, D.W.; Morreale, F. Sound forest/ljudskogen: A large-scale string-based interactive musical instrument. In Proceedings of the Sound and Music Computing, Hamburg, Germany, 31 August–3 September 2016. [Google Scholar]
  20. Katsantonis, M.N.; Manikas, A.; Mavridis, I. Design of a Cultural Heritage Gesture-Based Puzzle Game and Evaluation of User Experience. Appl. Sci. 2023, 13, 5493. [Google Scholar] [CrossRef]
  21. Rabiner, L.R. A tutorial on hidden Markov models and selected applications in speech recognition. Proc. IEEE 1989, 77, 257–285. [Google Scholar] [CrossRef]
  22. Lee, M.A.; Wessel, D. Connectionist models for real-time control of synthesis and compositional algorithms. In Proceedings of the International Computer Music Conference (ICMC), San Jose, CA, USA, 14–18 October 1992; pp. 277–280. [Google Scholar]
  23. Van Nort, D.; Wanderley, M.M.; Depalle, P. On the Choice of Mappings Based on Geometric Properties. In Proceedings of the 2004 Conference on New Interfaces for Musical Expression (NIME ’04), Hamamatsu, Japan, 3–5 June 2004; National University of Singapore: Singapore, 2004; pp. 87–91. [Google Scholar]
  24. Fang, C. From Dynamic Time Warping (DTW) to Hidden Markov Model (HMM) Final Project Report for ECE 742 Stochastic Decision. March 2009. Available online: https://www.semanticscholar.org/paper/From-Dynamic-Time-Warping-(-DTW-)-to-Hidden-Markov-Fang/048cd300b39c49eab62a4f91457ea3a0aa6b9bb7 (accessed on 16 May 2023).
  25. Bevilacqua, F.; Guédy, F.; Schnell, N.; Fléty, E.; Leroy, N. Wireless sensor interface and gesture-follower for music pedagogy. In Proceedings of the International Conference of New Interfaces for Musical Expression, New York, NY, USA, 6–10 June 2007; pp. 124–129. [Google Scholar]
  26. Bevilacqua, F.; Zamborlin, B.; Sypniewski, A.; Schnell, N.; Guédy, F.; Rasamimanana, N. Continuous realtime gesture following and recognition. LNAI 2010, 5934, 73–84. [Google Scholar]
  27. Brooke, J. SUS: A “quick and dirty” usability scale. In Usability Evaluation in Industry; Jordan, P.W., Thomas, B., Weerdmeester, B.A., McClelland, I.L., Eds.; Taylor & Francis: London, UK, 1996; pp. 189–194. [Google Scholar]
  28. Tullis, T.S.; Stetson, J.N. A Comparison of Questionnaires for Assessing Website Usability. In Proceedings of the Usability Professionals Association (UPA), Minneapolis, MN, USA, 7–11 June 2004. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.