Special Issue "Embodied and Spatial Interaction"
A special issue of Multimodal Technologies and Interaction (ISSN 2414-4088).
Deadline for manuscript submissions: closed (30 April 2019).
Interests: multimodal human-computer interaction; gestural interaction; user experience; accessibility; users with special needs
Embodied and spatial interaction has gained a lot of attention lately, but there is little research carried out from multimodal interaction perspective. In this special issue, we welcome submissions related to embodied and spatial interaction, which cover any combination of multimodal interaction means. Articles could relate to design, implementation and evaluation of multimodal solutions. In particular, we welcome articles which focus on fusion and fission of different modalities, including multi-dimensional analysis of multimodal interaction (e.g., combinations of subjective and objective metrics) in this setting. Other possible topics include case studies (e.g., system descriptions and evaluation results) of multimodal embodied and spatial systems, multimodal interaction techniques for embodied and spatial interaction (e.g., combination of gestural and spoken interaction), and domain-specific case studies (e.g., embodied and spatial interaction in industrial settings and healthcare).
In this context, embodied and spatial interaction could be interpreted rather freely, and interaction can take place both in physical environments (e.g., build environments) and virtual environments (e.g., virtual reality), including their combinations. Different interaction means could include gestures (e.g., mid-air gestures), spoken interaction, haptic feedback, gaze tracking, motion capture, wearable computing, and human interaction with IoT-data (among others). From evaluation viewpoint, research on multi-dimensional analysis (e.g., subjective metrics such as UX questionnaires, and objective metrics such as log-data and biometric information) are particularly welcome. Also, embodied and spatial interaction studies with special user groups (including assistive technology, people with special needs, ICT4D) are within the scope of this Special Issue.
This Special Issue aims to provide a collection of high quality research articles that address challenges in multimodal embodied and spatial interaction. Both theoretical and applied research studies are welcome, and practical case studies in different domains are particularly welcome.
Prof. Markku Turunen
Manuscript Submission Information
Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All papers will be peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.
Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Multimodal Technologies and Interaction is an international peer-reviewed open access quarterly journal published by MDPI.
Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 1000 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.
- Spatial interaction
- Embodied interaction
- Interaction analysis
- Interaction design
- User Experience
- Spoken interaction
- Haptic feedback