A Survey on Multi-User Conversational Interfaces
Abstract
1. Introduction
- Development of a survey paper following the PRISMA flow method [1]. Following this approach, we filter 181 articles, analysing in detail 59 final candidates. From these candidates, we provide an analysis of the venues and journals in which they were published. Additionally, we outline possible existing connections between authors and countries of the analysed collection. Although dispersed, we observe a growing community and indicate several networks of potential collaborators.
- Proposal of a definition of the MUCIs term based on related works and renowned dictionaries, given the lack of consensus in the literature.
- Identification of the most important aspects of contemporary MUCIs by highlighting patterns and common design approaches. To create a broad understanding, the relevant publications of the sample are categorised into four sections: Application Domains and Examples, User and Dialogue Modelling, Multi-Modality, and Evaluation Methods.
- Provision of guidelines for future implementations of MUCIs and their current challenges to address.
2. Scope and Methodology
3. Quantitative Paper Analysis
3.1. Relevance of Conferences and Journals
3.2. Networks of Research Collaborations
4. Capturing Multi-User Conversational Interfaces
4.1. Application Domains and Examples
4.1.1. Health Care
4.1.2. Social Conversation
4.1.3. Service Robots
4.1.4. Tutoring
4.1.5. Negotiation and Conflict Resolution
- MU collaborative interactions vs. MU competitive interactions: For example, in social conversation contexts, tasks are often collaborative in nature [11,13], with users working together toward a shared goal. In contrast, competitive scenarios consist of a different dynamic, where users may interact with the system in parallel, without being explicitly aware of each other’s intentions and ambitions. This can lead to implicit or explicit competition, where users compete for the system’s attention or resources, either unintentionally, as observed in [10], or deliberately, as demonstrated in [21,22]. Such interactions between users influence both the task and behaviour of the agent and thus require a different optimisation of the dialogue policy.
- Task-oriented designs vs. open-domain designs: Another meaningful way to categorise the reviewed works is by distinguishing between task-oriented and open-domain systems. Task-oriented systems are specifically designed and customised to achieve a specific objective, such as Tangy [12], which assists users in scheduling and playing bingo games. In contrast, open-domain systems aim to engage in open conversations without a narrowly defined goal, as shown in [15,16]. These differences in system goals also have significant implications for dialogue policy design and the handling of MU dialogues. While user interactions are often structured and goal-driven in task-oriented dialogues, reducing the need for extensive individual customisation, open-domain systems may require more nuanced adaptation to individual users and in-group dynamics to maintain engaging and contextually appropriate interactions. Moreover, the application of user models may facilitate the decision-making in domains such as tutoring [25,26] or conflict resolution [33].
4.2. User and Dialogue Modelling
4.3. Multi-Modality in Multi-User Conversational Interfaces
4.3.1. Virtual Agents
4.3.2. Robotic Agents
4.4. Evaluation Methods
5. Discussion
- Role differentiation: Most system developers assign distinct roles to users and system (e.g., patients and health coach [14], customers and assistant [3], learners and tutor [5]), which guide the flow of dialogues and facilitate to manage turn-taking. One reason for this may be that systems which support clear role distributions tend to manage MU input more effectively and ensure balanced participation.
- Limited dialogue management: Several publications, particularly in tutoring and conflict resolution, describe conceptual frameworks or systems that only partially implement dialogue logic [25,32]. As a result, dynamic conversational capabilities remain underdeveloped in contemporary approaches. Instead of employing data-driven approaches, systems often rely on scripted or semi-structured dialogues with minimal adaptability to unexpected user input or spontaneous conversational shifts. However, suitable training data for MUCIs remains scarce across many applications and use cases.
- Rigid interaction behaviour: A significant number of interfaces still rely on pre-defined and fixed conversation structures, which implies that the system is processing user input rather sequentially and separately from the group context. In contrast, some systems (e.g., [11,22]) demonstrate capabilities to handle truly simultaneous input from multiple users with dynamic turn-taking. Nevertheless, this seems necessary to enable natural interactions in group settings, where users may speak out of turn, overlap, or engage in side conversations. Without support for simultaneous input and context-aware coordination, systems risk selecting inappropriate or intrusive responses.
- Prototypical systems: Many systems are developed as research prototypes or simulations. Although the findings of their corresponding user studies offer valuable insights, fully functional and deployable real-world MUCIs have yet to be realised. Exceptions are seen in service robotics and personal assistants that operate in constrained environments, such as [64,65]. Bridging the gap between experimental prototypes and robust real-world MUCIs remains a critical challenge, requiring advances in dialogue management, multi-user conversation analysis, and multi-modal integration.
- Multi-Modality: The use of embodiment and multi-modality through robots or virtual avatars has shown great potential for maintaining engagement and managing attention in groups. Systems that integrate multi-modal input and output (e.g., gaze, speech, and gestures) tend to be more robust in turn-taking or user state tracking [24,49]. However, implementing and synchronising multi-modal channels remains technically demanding and resource-intensive, which may limit scalability. In addition, such systems may consist of hardware setups that are difficult to replicate and transfer to further use-cases.
- Application-orientated design: MUCIs should be adapted to the social and physical context in which they operate. This includes considerations such as the number of users, their spatial arrangement, and the interaction setting (e.g., public spaces, homes, classrooms). A good contextual fit improves relevance, usability, and user engagement.
- Adaptive dialogue management: Future systems should be able to handle interruptions, overlapping speech, and changing in-group dynamics. Effective dialogue policies require flexibility and, where possible, adaptive mechanisms that can respond to dynamic user behaviours, leaving behind fixed interaction patterns. While data-driven methods like reinforcement learning and transformer-based dialogue models have shown notable success in single-user CUIs, their potential in multi-user scenarios needs further exploration.
- Natural turn-taking strategies: As a characteristic of natural group interaction, users often communicate through a combination of speech, gaze, gestures, and facial expressions. MUCIs should integrate multi-modal channels in a coordinated way to manage turn-taking effectively, taking into account both individual and group-level cues.
- Standardised evaluation metrics: The evaluation of MUCIs remains fragmented, with a lack of consistent metrics and methodologies. To support meaningful comparison and progress in the field, it is necessary to apply shared evaluation standards that assess subjective user feedback, task performance, and group-level interaction quality. While short-term, lab-based studies provide valuable initial insights, future research should prioritise long-term, real-world deployments to fully capture user behaviour and social interaction over time.
- Ethical considerations: Although significant efforts on system development and dialogue management were presented in the reviewed articles, few explicitly address ethical aspects such as user privacy, fairness in interaction, and accessibility. These considerations are particularly critical in shared or public environments in which multiple users engage simultaneously and the system has to balance competing goals, complex social dynamics, and differing user expectations. Future work should include these principles in the design and evaluation of MUCIs.
6. Conclusions and Outlook
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
AI | Artificial Intelligence |
DS | Dialogue Systems |
CUI | Conversational User Interface |
MU | Multi-User |
MUCI | Multi-User Conversational Interface |
Appendix A. Most Relevant Articles
References
- Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. Int. J. Surg. 2021, 88, 105906. [Google Scholar] [CrossRef] [PubMed]
- Cole, V.; Boutet, M. ResearchRabbit (product review). J. Can. Health Libr. Assoc./J. L’Assoc. BibliothèQues Santé Can. 2023, 44, 43–47. [Google Scholar] [CrossRef]
- Johansson, M.; Skantze, G. Opportunities and Obligations to Take Turns in Collaborative Multi-Party Human-Robot Interaction. In Proceedings of the 16th Annual Meeting of the Special Interest Group on Discourse and Dialogue, Prague, Czech Republic, 2–4 September 2015; pp. 305–314. [Google Scholar] [CrossRef]
- Skantze, G. Turn-taking in Conversational Systems and Human-Robot Interaction: A Review. Comput. Speech Lang. 2021, 67, 101178. [Google Scholar] [CrossRef]
- Moubayed, S.A.; Skantze, G.; Beskow, J. The furhat back-projected humanoid head–lip reading, gaze and multi-party interaction. Int. J. Humanoid Robot. 2013, 10, 1350005. [Google Scholar] [CrossRef]
- Suzuki, S.; Anuardi, M.N.A.M.; Sripian, P.; Matsuhira, N.; Sugaya, M. Multi-user Robot Impression with a Virtual Agent and Features Modification According to Real-time Emotion from Physiological Signals. In Proceedings of the 2020 29th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), Naples, Italy, 31 August–4 September 2020; pp. 1006–1012. [Google Scholar] [CrossRef]
- Panhwar, M.A.; Saddar, S.; Aftab, A.; Memon, K.A.; Raza, M.O.; ZhongLiang, D.; Noor-ul-Ain. Adaptive Interface for Multi-users and Its Evaluation. Int. J. Comput. Sci. Netw. Secur. 2019, 19, 198–201. [Google Scholar]
- Gu, J.C.; Tao, C.; Ling, Z.H. Who Says What to Whom: A Survey of Multi-Party Conversations. In Proceedings of the IJCAI, Vienna, Austria, 23–29 July 2022; pp. 5486–5493. [Google Scholar]
- Traum, D. Issues in Multiparty Dialogues. In International Workshop on Agent Communication Languages; Dignum, F., Ed.; Springer: Berlin/Heidelberg, Germany, 2004; pp. 201–211. [Google Scholar]
- Lee, G.; An, K.; Yun, S.S.; Choi, J. A simultaneous robot service scheme for Multi-users. In Proceedings of the 2015 12th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI), Goyangi, Republic of Korea, 28–30 October 2015; pp. 373–374. [Google Scholar]
- Skantze, G.; Johansson, M.; Beskow, J. Exploring turn-taking cues in multi-party human–robot discussions about objects. In Proceedings of the 2015 ACM on International Conference on Multimodal Interaction, Seattle, WA, USA, 9–13 November 2015; pp. 67–74. [Google Scholar]
- Louie, W.Y.G.; Vaquero, T.; Nejat, G.; Beck, J.C. An autonomous assistive robot for planning, scheduling and facilitating multi-user activities. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; pp. 5292–5298. [Google Scholar]
- Dohsaka, K.; Asai, R.; Higashinaka, R.; Minami, Y.; Maeda, E. Effects of conversational agents on activation of communication in thought-evoking multi-party dialogues. IEICE Trans. Inf. Syst. 2014, 97, 2147–2156. [Google Scholar] [CrossRef]
- Snaith, M.; De Franco, D.; Beinema, T.; Den Akker, H.O.; Pease, A. A dialogue game for multi-party goal-setting in health coaching. In Proceedings of the 7th International Conference on Computational Models of Argument, COMMA 2018, Warsaw, Poland, 12–14 September 2018; pp. 337–344. [Google Scholar]
- Todo, Y.; Nishimura, R.; Yamamoto, K.; Nakagawa, S. Development and evaluation of spoken dialog systems with one or two agents through two domains. In Proceedings of the International Conference on Text, Speech and Dialogue, Pilsen, Czech Republic, 1–5 September 2013; pp. 185–192. [Google Scholar]
- Shibahara, Y.; Yamamoto, K.; Nakagawa, S. Effect of sympathetic relation and unsympathetic relation in multi-agent spoken dialogue system. In Proceedings of the 2016 International Conference on Advanced Informatics: Concepts, Theory and Application (ICAICTA), Penang, Malaysia, 16–19 August 2016; pp. 1–6. [Google Scholar]
- Clavel, C.; Cafaro, A.; Campano, S.; Pelachaud, C. Fostering user engagement in face-to-face human-agent interactions: A survey. In Toward Robotic Socially Believable Behaving Systems-Volume II; Esposito, A., Jain, L., Eds.; Springer: Cham, Switzerland, 2016; pp. 93–120. [Google Scholar]
- Liu, C.; Liu, K.; He, S.; Nie, Z.; Zhao, J. Incorporating Interlocutor-Aware Context into Response Generation on Multi-Party Chatbots. In Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL), Hong Kong, China, 3–4 November 2019; Bansal, M., Villavicencio, A., Eds.; Association for Computational Linguistics (ACL): Hong Kong, China, 2019; pp. 718–727. [Google Scholar] [CrossRef]
- Ouchi, H.; Tsuboi, Y. Addressee and response selection for multi-party conversation. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, Austin, TX, USA, 1–5 November 2016; pp. 2133–2143. [Google Scholar] [CrossRef]
- Ju, D.; Feng, S.; Lv, P.; Wang, D.; Zhang, Y. Learning to Improve Persona Consistency in Multi-party Dialogue Generation via Text Knowledge Enhancement. In Proceedings of the 29th International Conference on Computational Linguistics, Gyeongju, Republic of Korea, 12–17 October 2022; pp. 298–309. [Google Scholar]
- Keizer, S.; Kastoris, P.; Foster, M.E.; Deshmukh, A.; Lemon, O. Evaluating a social multi-user interaction model using a Nao robot. In Proceedings of the 23rd IEEE International Symposium on Robot and Human Interactive Communication, Edinburgh, UK, 25–29 August 2014; pp. 318–322. [Google Scholar] [CrossRef]
- Keizer, S.; Foster, M.E.; Gaschler, A.; Giuliani, M.; Isard, A.; Lemon, O. Handling uncertain input in multi-user human–robot interaction. In Proceedings of the The 23rd IEEE International Symposium on Robot and Human Interactive Communication, Edinburgh, UK, 25–29 August 2014; pp. 312–317. [Google Scholar] [CrossRef]
- Ben-Hanania, A.; Goldberg, J.; Cauchard, J.R. Multi-User Control for Domestic Robots with Natural Interfaces. In Proceedings of the 17th International Conference on Mobile and Ubiquitous Multimedia, Cairo, Egypt, 25–28 November 2018; pp. 433–439. [Google Scholar] [CrossRef]
- Divekar, R.R.; Drozdal, J.; Chabot, S.; Zhou, Y.; Su, H.; Chen, Y.; Zhu, H.; Hendler, J.A.; Braasch, J. Foreign language acquisition via artificial intelligence and extended reality: Design and evaluation. Comput. Assist. Lang. Learn. 2021, 35, 1–29. [Google Scholar] [CrossRef]
- Koutsombogera, M.; Deligiannis, M.; Giagkou, M.; Papageorgiou, H. Towards modelling multimodal and multiparty interaction in educational settings. In Toward Robotic Socially Believable Behaving Systems-Volume II; Springer International Publishing: Cham, Switzerland, 2016; pp. 165–184. [Google Scholar]
- Laine, T.H.; Lee, W.; Moon, J.; Kim, E. Building confidence in the metaverse: Implementation and evaluation of a multi-user virtual reality application for overcoming fear of public speaking. Int. J.-Hum.-Comput. Stud. 2025, 199, 103487. [Google Scholar] [CrossRef]
- Helvert, J.; Rosmalen, P.; Börner, D.; Petukhova, V.; Alexandersson, J. Observing, coaching and reflecting: A multi-modal natural language-based dialogue system in a learning context. In Proceedings of the 11th International Conference on Intelligent Environments, Prague, Czech Republic, 5–17 July 2015; Volume 19, pp. 220–227. [Google Scholar]
- Al Moubayed, S.; Lehman, J. Toward better understanding of engagement in multiparty spoken interaction with children. In Proceedings of the 2015 ACM on International Conference on Multimodal Interaction, Seattle, WA, USA, 9–13 November 2015; pp. 211–218. [Google Scholar]
- Hiraoka, T.; Georgila, K.; Nouri, E.; Traum, D.; Nakamura, S. Reinforcement learning in multi-party trading dialog. In Proceedings of the 16th Annual Meeting of the Special Interest Group on Discourse and Dialogue, Prague, Czech Republic, 2–4 September 2015; pp. 32–41. [Google Scholar] [CrossRef]
- Cuayáhuitl, H.; Keizer, S.; Lemon, O. Learning to trade in strategic board games. In CGW 2015, GIGA 2015. Communications in Computer and Information Science; Springer: Cham, Switzerland, 2015; Volume 614, pp. 83–95. [Google Scholar] [CrossRef]
- Xiao, G.; Georgila, K. A comparison of reinforcement learning methodologies in two-party and three-party negotiation dialogue. In Proceedings of the The Thirty-First International Flairs Conference, Melbourne, FL, USA, 21–23 May 2018. [Google Scholar]
- Yuan, J.; Yao, L.; Hao, Z.; Wang, Y. Multi-party Dialogue Games for Dialectical Argumentation. J. Comput. 2012, 7, 2564–2571. [Google Scholar] [CrossRef]
- Nurgaliyev, K.; Di Mauro, D.; Khan, N.; Augusto, J.C. Improved multi-user interaction in a smart environment through a preference-based conflict resolution virtual assistant. In Proceedings of the 2017 International Conference on Intelligent Environments (IE), Seoul, Republic of Korea, 21–25 August 2017; pp. 100–107. [Google Scholar] [CrossRef]
- Li, J.; Louie, W.Y.G.; Mohamed, S.; Despond, F.; Nejat, G. A user-study with tangy the bingo facilitating robot and long-term care residents. In Proceedings of the 2016 IEEE international symposium on robotics and intelligent sensors (IRIS), Tokyo, Japan, 17–20 December 2016; pp. 109–115. [Google Scholar] [CrossRef]
- Tahir, Y.; Dauwels, J.; Thalmann, D.; Magnenat Thalmann, N. A user study of a humanoid robot as a social mediator for two-person conversations. Int. J. Soc. Robot. 2020, 12, 1031–1044. [Google Scholar] [CrossRef]
- Kim, S.; Eun, J.; Oh, C.; Suh, B.; Lee, J. Bot in the Bunch: Facilitating Group Chat Discussion by Improving Efficiency and Participation with a Chatbot. In Proceedings of the CHI ’20: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, New York, NY, USA, 25–30 April 2020; pp. 1–13. [Google Scholar] [CrossRef]
- Anbro, S.J.; Houmanfar, R.A.; Thomas, J.; Baxter, K.; Harris, F.C., Jr.; Crosswell, L.H. Behavioral Assessment in Virtual Reality: An Evaluation of Multi-User Simulations in Healthcare Education. J. Organ. Behav. Manag. 2022, 43, 92–136. [Google Scholar] [CrossRef]
- Fan, J.; Beuscher, L.; Newhouse, P.A.; Mion, L.C.; Sarkar, N. A robotic coach architecture for multi-user human–robot interaction (RAMU) with the elderly and cognitively impaired. In Proceedings of the 2016 25th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), New York, NY, USA, 26–31 August 2016; pp. 445–450. [Google Scholar] [CrossRef]
- Zhu, L.; Zhang, Z.; Wang, J.; Wang, H.; Wu, H.; Yang, Z. Multi-Party Empathetic Dialogue Generation: A New Task for Dialog Systems. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Dublin, Ireland, 22–27 May 2022; pp. 298–307. [Google Scholar] [CrossRef]
- Adikari, A.; De Silva, D.; Moraliyage, H.; Alahakoon, D.; Wong, J.; Gancarz, M.; Chackochan, S.; Park, B.; Heo, R.; Leung, Y. Empathic conversational agents for real-time monitoring and co-facilitation of patient-centered healthcare. Future Gener. Comput. Syst. 2022, 126, 318–329. [Google Scholar] [CrossRef]
- Kumar, R.; Chauhan, D.S.; Dias, G.; Ekbal, A. Modelling Personalized Dialogue Generation in Multi-Party Settings. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; pp. 1–6. [Google Scholar] [CrossRef]
- Zhang, H.; Chan, Z.; Song, Y.; Zhao, D.; Yan, R. When Less Is More: Using Less Context Information to Generate Better Utterances in Group Conversations. In Natural Language Processing and Chinese Computing: 7th CCF International Conference, NLPCC 2018, Hohhot, China, August 26–30, 2018, Proceedings, Part I 7; Zhang, M., Ng, V., Zhao, D., Li, S., Zan, H., Eds.; Springer: Cham, Switzerland, 2018; pp. 76–84. [Google Scholar] [CrossRef]
- Gropp, M.; Schmidt, A.; Kleinbauer, T.; Klakow, D. Platon: Dialog Management and Rapid Prototyping for Multilingual Multi-user Dialog Systems. In Proceedings of the International Conference on Text, Speech, and Dialogue, Brno, Czech Republic, 12–16 September 2016; pp. 478–485. [Google Scholar] [CrossRef]
- Wang, D.; Hakkani-Tür, D.; Tur, G. Understanding computer-directed utterances in multi-user dialog systems. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 8377–8381. [Google Scholar] [CrossRef]
- Toxtli, C.; Monroy-Hernández, A.; Cranshaw, J. Understanding chatbot-mediated task management. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, Montreal, QC, Canada, 21–26 April 2018; pp. 1–6. [Google Scholar] [CrossRef]
- Aylett, M.P.; Romeo, M. You Don’t Need to Speak, You Need to Listen: Robot Interaction and Human-Like Turn-Taking. In Proceedings of the 5th International Conference on Conversational User Interfaces, Eindhoven, The Netherlands, 19–21 July 2023. [Google Scholar] [CrossRef]
- Makri, E.; Spiliotopoulos, D.; Vassilakis, C.; Margaris, D. Human behaviour in multimodal interaction: Main effects of civic action and interpersonal and problem-solving skills. J. Ambient. Intell. Humaniz. Comput. 2020, 11, 5991–6006. [Google Scholar] [CrossRef]
- Makri, E.; Koryzis, D.; Spiliotopoulos, D.; Svolopoulos, V. Metalogue’s Virtual Agent for Negotiation: Its Effects on Learning Experience, Metacognitive and Individual-and-Community-Level Attitudes Pre-and-Post Interaction. In Proceedings of the EDULEARN19, 11th International Conference on Education and New Learning Technologies, Palma, Spain, 1–3 July 2019; IATED Academy: Valencia, Spain, 2019; pp. 1542–1551. [Google Scholar] [CrossRef]
- Addlesee, A.; Cherakara, N.; Nelson, N.; Hernandez Garcia, D.; Gunson, N.; Sieińska, W.; Dondrup, C.; Lemon, O. Multi-party Multimodal Conversations Between Patients, Their Companions, and a Social Robot in a Hospital Memory Clinic. In Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics: System Demonstrations, St. Julians, Malta, 17– 22 March 2024; Aletras, N., De Clercq, O., Eds.; ACL: Kerrville, TX, USA, 2024; pp. 62–70. [Google Scholar]
- Kondo, Y.; Takemura, K.; Takamatsu, J.; Ogasawara, T. A gesture-centric android system for multi-party human–robot interaction. J. Hum.-Robot. Interact. 2013, 2, 133–151. [Google Scholar] [CrossRef]
- Fan, J.; Bian, D.; Zheng, Z.; Beuscher, L.; Newhouse, P.A.; Mion, L.C.; Sarkar, N. A robotic coach architecture for elder care (ROCARE) based on multi-user engagement models. IEEE Trans. Neural Syst. Rehabil. Eng. 2016, 25, 1153–1163. [Google Scholar] [CrossRef] [PubMed]
- Louie, W.Y.G.; Nejat, G. A social robot learning to facilitate an assistive group-based activity from non-expert caregivers. Int. J. Soc. Robot. 2020, 12, 1159–1176. [Google Scholar] [CrossRef]
- Tsamis, G.; Chantziaras, G.; Giakoumis, D.; Kostavelis, I.; Kargakos, A.; Tsakiris, A.; Tzovaras, D. Intuitive and Safe Interaction in Multi-User Human Robot Collaboration Environments through Augmented Reality Displays. In Proceedings of the 2021 30th IEEE International Conference on Robot & Human Interactive Communication (RO-MAN), Vancouver, BC, Canada, 8–12 August 2021; pp. 520–526. [Google Scholar] [CrossRef]
- Oertel, C.; Jonell, P.; Kontogiorgos, D.; Mora, K.F.; Odobez, J.M.; Gustafson, J. Towards an engagement-aware attentive artificial listener for multi-party interactions. Front. Robot. AI 2021, 8, 189. [Google Scholar] [CrossRef] [PubMed]
- Kraus, M.; Klein, S.; Wagner, N.; Minker, W.; André, E. A Pilot Study on Multi-Party Conversation Strategies for Group Recommendations. In Proceedings of the 6th ACM Conference on Conversational User Interfaces, Luxembourg, 8–10 July 2024. [Google Scholar] [CrossRef]
- Straßmann, C.; Eudenbach, C.; Arntz, A.; Eimler, S.C. “Don’t Judge a Book by its Cover”: Exploring Discriminatory Behavior in Multi-User-Robot Interaction. In Proceedings of the Companion of the 2024 ACM/IEEE International Conference on Human-Robot Interaction, Boulder, CO, USA, 11–15 March 2024; pp. 1023–1027. [Google Scholar] [CrossRef]
- Wagner, N.; Kraus, M.; Tonn, T.; Minker, W. Comparing Moderation Strategies in Group Chats with Multi-User Chatbots. In Proceedings of the CUI 2022: 4th Conference on Conversational User Interfaces, Glasgow, UK, 26–28 July 2022. [Google Scholar] [CrossRef]
- Li, B.; Lou, R.; Segonds, F.; Merienne, F. Multi-user interface for co-located real-time work with digital mock-up: A way to foster collaboration? Int. J. Interact. Des. Manuf. (IJIDeM) 2017, 11, 609–621. [Google Scholar] [CrossRef]
- Paetzel-Prüsmann, M.; Lehman, J.F.; Gomez, C.J.; Kennedy, J. An Automatic Evaluation Framework for Social Conversations with Robots. In Proceedings of the 2024 International Symposium on Technological Advances in Human-Robot Interaction, Boulder, CO, USA, 9–10 March 2024; pp. 56–64. [Google Scholar] [CrossRef]
- Richter, V.; Carlmeyer, B.; Lier, F.; Meyer zu Borgsen, S.; Schlangen, D.; Kummert, F.; Wachsmuth, S.; Wrede, B. Are you talking to me? Improving the Robustness of Dialogue Systems in a Multi Party HRI Scenario by Incorporating Gaze Direction and Lip Movement of Attendees. In Proceedings of the Fourth International Conference on Human Agent Interaction, 4–7 October 2016; Biopolis: Singapore; pp. 43–50. [Google Scholar] [CrossRef]
- Alofs, T.; Theune, M.; Swartjes, I. A tabletop interactive storytelling system: Designing for social interaction. Int. J. Arts Technol. 2015, 8, 188–211. [Google Scholar] [CrossRef]
- Canévet, O.; He, W.; Motlicek, P.; Odobez, J.M. The MuMMER Dataset for Robot Perception in Multi-party HRI Scenarios. In Proceedings of the 2020 29th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), Naples, Italy, 31 August–4 September 2020; pp. 1294–1300. [Google Scholar] [CrossRef]
- Faria, M.; Silva, R.; Alves-Oliveira, P.; Melo, F.S.; Paiva, A. “Me and you together” movement impact in multi-user collaboration tasks. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 2793–2798. [Google Scholar] [CrossRef]
- Müller, A.; Richert, A. No One is an Island—Investigating the Need for Social Robots (and Researchers) to Handle Multi-Party Interactions in Public Spaces. In Proceedings of the 2023 32nd IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), Busan, Republic of Korea, 28–31 August 2023; pp. 1772–1777. [Google Scholar] [CrossRef]
- Ostrowski, A.K.; Fu, J.; Zygouras, V.; Park, H.W.; Breazeal, C. Speed Dating with Voice User Interfaces: Understanding How Families Interact and Perceive Voice User Interfaces in a Group Setting. Front. Robot. AI 2022, 8, 730992. [Google Scholar] [CrossRef] [PubMed]
- Skantze, G.; Al Moubayed, S. IrisTK: A statechart-based toolkit for multi-party face-to-face interaction. In Proceedings of the 14th ACM International Conference on Multimodal Interaction, Santa Monica, CA, USA, 22–26 October 2012; pp. 69–76. [Google Scholar] [CrossRef]
(TS = (Multi-user dialog* system*) OR TS = (Multi-party dialog* system*) |
OR |
TS = (Multi-user NEAR/15 (“conversation* agent” OR “user interface” |
OR |
chatbot* OR robot* OR multi-modal* OR assistan*)) |
OR |
TS = (Multi-user NEAR/2 evaluation)) NOT TS=(network OR MIMO) |
Conference | Number of Papers |
---|---|
IEEE International Symposium on Robot and Human Interactive Communication (IEEE RO-MAN) | 6 |
Conversational User Interfaces (CUI) | 4 |
Annual Meeting of the Association for Computational Linguistics (ACL) | 4 |
International Conference on Robotics and Automation (ICRA) | 2 |
International Conference on Intelligent Environments (IE) | 2 |
Text, Speech and Dialogue (TSD) | 2 |
International Conference on Intelligent Robots and Systems (IROS) | 2 |
ACM International Conference on Multi-Modal Interaction (ICMI) | 2 |
Augmented Reality, Virtual Reality, and Computer Graphics (AVR) | 2 |
Publication | User Model | Decision Making for Turn-Taking | Implemented System Prototype | Considering In-Group Dynamics |
---|---|---|---|---|
[21,35,36] | ✓ | ✓ | ✓ | ✓ |
[3] | ✓ | ✓ | ✓ | |
[25] | ✓ | ✓ | ✓ | |
[14] | ✓ | ✓ | ||
[11,12,33,37,38,39] | ✓ | ✓ | ✓ | |
[40,41] | ✓ | ✓ | ||
[42] | ✓ | |||
[22,28] | ✓ | ✓ | ✓ | |
[15,16,34] | ✓ | ✓ | ||
[43,44] | ✓ | |||
[45] | ✓ |
Publication | User Study | Performance-Based Metrics | Group or User-Centred Metrics | Data Collection |
---|---|---|---|---|
[11,21,28] | ✓ | ✓ | user-centred | ✓ |
[36] | ✓ | ✓ | group-centred | ✓ |
[12,15,22,34,37,45,55,56] | ✓ | ✓ | user-centred | |
[25,57] | ✓ | user-centred | ✓ | |
[35,58] | ✓ | user-centred | ||
[16,33,59] | ✓ | group-centred | ||
[3,39,60] | ✓ | group-centred | ✓ | |
[42] | ✓ | group-centred | ||
[40,41,44] | ✓ | user-centred |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wagner, N.; Kraus, M.; Minker, W.; Griol, D.; Callejas, Z. A Survey on Multi-User Conversational Interfaces. Appl. Sci. 2025, 15, 7267. https://doi.org/10.3390/app15137267
Wagner N, Kraus M, Minker W, Griol D, Callejas Z. A Survey on Multi-User Conversational Interfaces. Applied Sciences. 2025; 15(13):7267. https://doi.org/10.3390/app15137267
Chicago/Turabian StyleWagner, Nicolas, Matthias Kraus, Wolfgang Minker, David Griol, and Zoraida Callejas. 2025. "A Survey on Multi-User Conversational Interfaces" Applied Sciences 15, no. 13: 7267. https://doi.org/10.3390/app15137267
APA StyleWagner, N., Kraus, M., Minker, W., Griol, D., & Callejas, Z. (2025). A Survey on Multi-User Conversational Interfaces. Applied Sciences, 15(13), 7267. https://doi.org/10.3390/app15137267