You are currently viewing a new version of our website. To view the old version click .
  • Article
  • Open Access

3 September 2024

Detailed Analysis of Responses from Older Adults through Natural Speech: Comparison of Questions by AI Agents and Humans

,
,
and
1
Simulation of Complex Systems Lab, Department of Human and Engineered Environmental Studies, Graduate School of Frontier Sciences, The University of Tokyo, Tokyo 277-8563, Japan
2
AI-UX Design Research Institution, Advanced Institute of Industrial Technology, 10-40 Higashi-Oi 1-Chome, Shinagawa, Tokyo 140-0011, Japan
3
Institute of Gerontology (IOG), The University of Tokyo, Tokyo 113-8656, Japan
4
Institute for Future Initiatives (IFI), The University of Tokyo, Tokyo 113-0033, Japan
This article belongs to the Section Behavioral and Mental Health

Abstract

In recent years, an increasing number of studies have begun to use conversational data in spontaneous speech to estimate cognitive function in older people. The providers of spontaneous speech with older people used to be physicians and licensed psychologists, but it is now possible to have conversations with fully automatic AI agents. However, it has not yet been clarified what differences exist in conversational communication with older people when the examiner is either a human or an AI agent. In this study, elderly people living in the community and attending a silver human resource center and a day service center were the subjects. Dialogues were conducted using generic interview items for estimating cognitive function through daily conversation, which were developed through research on estimation methods for cognitive function. From the data obtained from the dialogues, we compared the effects of human–AI interaction on the number of utterances, speaking time, and silence time. This study was conducted at a facility in Japan and included 32 subjects (12 males and 20 females). The results showed significant differences between human and AI dialogue in the number of utterances and silent time. This study suggests the effectiveness of AI in communication with older people and explores the possibility of using AI in social welfare.

1. Introduction

In modern societies, including Japan, the rapid aging of the population poses a significant challenge. As of 2020, elderly individuals aged 65 and above comprise about 28% of the total population in Japan, and this proportion is expected to exceed 30% by 2030 []. The demographic shift towards an older population significantly impacts healthcare, welfare, and economic activities. Social isolation, which worsens as society ages, is a critical issue. Many elderly individuals experience reduced contact with family and friends, leading to increased feelings of loneliness, which studies have shown to adversely affect psychological and physical health []. This is partly due to a lack of social support.
Moreover, cognitive decline in the elderly is a significant concern. Aging is commonly associated with reductions in memory and cognitive abilities [], which can impair independence and degrade the quality of life for the elderly. However, recent research indicates that appropriate social and cognitive stimuli can maintain or even improve cognitive functions [], particularly suggesting that dialogue and social activities could reduce the risk of dementia.
To address social isolation and cognitive decline, the use of artificial intelligence (AI) is gaining attention. With advancements in AI technology, new possibilities for elderly support are being explored. AI can complement and sometimes substitute traditional caregiving and support methods. For instance, AI-driven communication support systems can potentially reduce social isolation among the elderly. These systems utilize speech recognition and natural language processing technologies to facilitate conversations with the elderly []. Additionally, communications facilitated through AI-equipped robots or virtual agents can provide the elderly with opportunities for regular dialogue and information exchange [].
Furthermore, AI could also aid in maintaining and enhancing the cognitive functions of the elderly. By integrating AI into cognitive training programs, customized interventions suitable for individual cognitive states can be developed, which are potentially effective in preventing dementia []. AI is also used for the early detection of cognitive impairments through automated speech analysis and digital voice evaluation [,]. Moreover, AI excels in analyzing large data sets and recognizing patterns, contributing to elderly health management and medical predictions [].
As research and development progress, the role of AI in supporting the elderly is expected to grow significantly. Utilizing AI’s capabilities and advancing human-centered design could improve the quality of life for the elderly. Silver human resource centers and day service centers play crucial roles in enabling elderly social participation and maintaining physical and mental health through interactions with others. However, the impact of the quality and frequency of dialogues in these settings on the psychological health of the elderly is not yet fully understood.
This study investigates the differences in the number of utterances, speech duration, and silence times in dialogues conducted with elderly individuals from silver human resource centers and day service centers, comparing interactions with humans and AI. Cognitive functions were also measured using the MMSE. These validations examined the impact of AI-facilitated dialogues on the elderly, suggesting that AI-mediated communication could effectively ameliorate social isolation among the elderly.

3. Methods

3.1. Experimental Design

Participants underwent three assessments: a cognitive function test using the Mini-Mental State Examination (MMSE), routine conversations with humans, and conversations mediated by AI. The MMSE used can be administered in 6–10 min and consists of 11 items that assess orientation, memory, calculation ability, language abilities, and visuospatial skills, with a total score of 30 points. A score below 23 suggests possible dementia, and a score below 27 suggests mild cognitive impairment (MCI) []. The Japanese version of the MMSE (MMSE-J) was utilized in this study []. For the conversational tests, generic hearing items developed by Igarashi et al. (2023) for estimating cognitive function through routine conversation were adopted []. These items were designed to assess family history, physical condition, interests, daily activities, and cognitive abilities related to memory and orientation. They were validated by five certified psychologists working in hospitals. These items were further refined for this study, categorizing them into six areas: temporal orientation, spatial orientation, family history, daily activities, physical condition, and interests, with a total of 30 questions designed for both AI and human conversations (Table 1). Conversations based on these questions were conducted over a month apart. This study was conducted with the approval of the ethical review by the University of Tokyo and Tokyo Metropolitan Institute of Technology (code: 23009, 19 October 2023).
Table 1. Daily conversation used in interpersonal and agent conversations, consisting of 30 questions across 5 areas (copyright Igarashi et al., 2023 []).

3.2. Participants

According to the definition by the World Health Organization (WHO), elderly people are individuals aged 65 years and older. In Japan, the Ministry of Health, Labor and Welfare also references the WHO definition in its resources. Additionally, the Statistics Bureau of Japan classifies individuals aged 65 and above as elderly. Based on this, the present study also defined elderly people as those aged 65 and older. This study was conducted at a senior talent center and a day service center in Tokyo, involving 32 elderly participants (12 males and 20 females). The senior talent center refers to a resource center for individuals aged 65 and over who have retired from work but still possess a willingness to work, offering opportunities for temporary jobs through mediation. The day service center is a facility providing day services specifically for dementia patients. It serves elderly individuals whose cognitive functions are declining but do not require long-term institutionalization, offering a place for them to prevent further cognitive decline and engage in communication with others. The participants were elderly people living in the community who were not long-term residents of institutions or hospitals. In addition, to avoid the influence of dialect, participants were residents of Tokyo. Furthermore, this study was limited to those who had no physical ailments such as hearing loss and had no difficulty in listening to and speaking in general conversation. Participants were informed of the purpose and content of this study and their consent was obtained. For elderly people with cognitive decline, this study was explained to their family members and their consent was obtained.

3.3. Data Collection

The number of utterances, speech duration, and silence periods were recorded for both AI-mediated and human conversations, allowing for quantitative analysis of the interaction quality and the elderly’s responses. The content of the conversations was transcribed by a professional service.
Several modules were used to implement the AI agent, an avatar with AI dialogue capabilities. First, the author modeled the AI agent using VRoid Studio [] to design it as if it were having an interpersonal conversation. For displaying the 3D character in a browser, @pixiv/three-vrm [] was used. @pixiv/three-vrm is a library for loading and displaying VRM, a format for handling humanoid 3D avatar model data using three.js, on a browser, and is open-sourced by Pixiv. The background of the character incorporates the background of the room in the facility where the dialogue is taking place, creating an environment in which the user seems to be conversing with a real person.
In this study, the author executed a conversation with a human being and designed the AI avatar to resemble the appearance of a person (Figure 1). The aim was to minimize the influence of the avatar’s appearance on the test results, even though there are individual differences in impressions of appearance, because using an avatar whose appearance is significantly different from the person the participants assumes would be speaking may affect the test results. The Web Speech API (Speech Recognition) [] was used for user speech recognition. Since speech recognition can be difficult when the timing of the agent’s speech and the participant’s speech overlap, we implemented a system in which the participants wear a pin microphone and speech is recognized only while a button is pressed. Responses from the system were generated using the Koeiro API [], and the AI agent’s mouth was designed to move in sync with the speech using lip-sync functionality. The quantitative data obtained was analyzed using Microsoft® Excel® 2021 MSO.
Figure 1. (a) The appearance of the author who conducted the interpersonal conversation and (b) the appearance of the AI agent modeled based on the author’s appearance.

3.4. Protocols in Conversation Design

The conversation design protocol was created using a semi-structured interview technique with fixed question content. The conversational protocol was designed to ensure that the completion of a conversation about a particular question is clear, with the AI agent always responding to the user’s answer with a response related to its content before moving on to the next question. Only the portion of the response corresponding to the user’s answer was implemented using the ChatGPT API [] (Figure 2 and Figure 3).
Figure 2. Protocol of daily conversation system for cognitive function estimation by AI agents.
Figure 3. Configuration diagram showing the entire system.

4. Results

The study participants comprised 32 elderly individuals (12 males and 20 females) living in the community. Participants from the senior talent center included nine males and six females, with an average age of 74.73 years (SD = 5.48) and an average MMSE score of 28.67. The MMSE scores for participants from the senior talent center ranged from a maximum of 30 to a minimum of 26.
Participants from the day service center included 3 males and 14 females, with an average age of 82.59 years (SD = 4.96) and an average MMSE score of 13.47. The MMSE scores for participants from the day service center ranged from a maximum of 25 to a minimum of 7. Participants were categorized into high and low cognitive function groups based on an MMSE score cutoff of 25.

4.1. Conversations with Humans or AI

For conversations with a human interlocutor, the overall average number of utterances was 113.91, the average speech duration was 32.88 s, and the average silence duration was 4.19 s (Table 2). Specifically, for participants from the senior talent center, the average number of utterances was 123.96, the average speech duration was 42.56 s, and the average silence duration was 4.16 s. For participants from the day service center, the average number of utterances was 104.87, the average speech duration was 24.24 s, and the average silence duration was 4.21 s.
Table 2. Mean and SD for participants’ age, MMSE, number of answer words, speech time, and silent time.
In conversations with the AI, the overall average number of utterances was 59.55, the average speech duration was 17.25 s, and the average silence duration was 5.01 s. Specifically, for participants from the senior talent center, the average number of utterances was 21.86, the average speech duration was 10.22 s, and the average silence duration was 3.76 s. For the participants from the day service center, the average number of utterances was 92.58, the average speech duration was 23.42 s, and the average silence duration was 5.34 s.

4.2. Comparison between Senior Talent Center and Day Service Center

For participants from the senior talent center, the overall average number of utterances was 73.76, the average speech duration was 26.65 s, and the average silence duration was 4.04 s. In human conversations, the average number of utterances was 123.96, the average speech duration was 42.56 s, and the average silence duration was 4.16 s. In AI conversations, the average number of utterances was 21.86, the average speech duration was 10.22 s, and the average silence duration was 3.76 s.
For participants from the day service center, the overall average number of utterances was 98.58, the average speech duration was 23.83 s, and the average silence duration was 4.86 s. In human conversations, the average number of utterances was 104.87, the average speech duration was 24.24 s, and the average silence duration was 4.21 s. In AI conversations, the average number of utterances was 92.58, the average speech duration was 23.42 s, and the average silence duration was 5.34 s (Table 2).

5. Analysis

5.1. Comparison of Number of Answer Words per Question

The number of utterances was analyzed for its correlation with age and cognitive function using a two-tailed t-test for unequal variances at a significance level of 5%. No significant difference was observed across the entire dataset (p = 4.97, t = 5.04). Comparing conversations with AI and humans among participants from the senior talent center showed no significant differences (p = 3.09, t = 14.7). However, among the day service center participants, the average number of utterances in human conversations was significantly different from the AI conversations, being lower for humans (92.58) than for AI (104.88) (p = 0.05, t = −1.96). In conversations between AI and humans, a significant difference in the number of utterances was found between senior human resources and day service participants only in the AI conversations (p = 0.02, t = 2.28 for human; p = 1.75, t = −16.67 for AI). Figure 4 presents these results in a box plot format. Sections marked with a single asterisk indicate a significant difference at the 5% level, while sections marked with two asterisks indicate a significant difference at the 1% level.
Figure 4. Comparison of high- and low-cognitive-function groups in terms of number of utterances. *: Significant differences were found at the 5% level.

5.2. Comparison of Speech Duration per Answer

The impact of age and cognitive function on speech duration was analyzed using a two-tailed t-test for unequal variances at a 5% level. No significant differences were found (p = 0.13, t = 1.53). Similarly, no significant differences were observed in speech duration between AI and human conversations among participants from either the senior talent center (p = 5.50, t = 10.07) or the day service center (p = 0.57, t = 0.57).
There was also no significant difference between the speech time of silver and day service participants when comparing the speech time of silver and day service participants when interacting with humans (p = 3.83, t = 5.56), and no significant difference was found between the speech time of silver and day service participants when interacting with AI (p = 9.80, t = −10.74). Figure 5 presents these results in a box plot format.
Figure 5. Comparison of high- and low-cognitive-function groups in terms of seconds of speech.

5.3. Comparison of Silence Duration per Answer

Silence duration was analyzed for its correlation with age and cognitive function using a two-tailed t-test for unequal variances. A significant difference was found (p = 0.00, t = −2.80). In the senior talent center, no significant difference in terms of silence duration was observed between AI and human conversations (p = 0.44, t = 0.77). However, for the day service center participants, silence duration was significantly longer in AI conversations compared to human conversations, indicating possible discomfort or confusion among some elderly individuals when interacting with AI (p = 0.00, t = −3.05). Figure 6 presents these results in a box plot format. Within the group with cognitive decline, there was a significant difference in silence duration between the conversations with AI and the conversations with humans. Additionally, in the conversations with AI, there was a significant difference in silence duration between the high and low cognitive function groups. Sections marked with two asterisks indicate a significant difference at the 1% level.
Figure 6. Comparison of high- and low-cognitive-function groups in terms of silent seconds. **: Significant differences were found at the 1% level.

6. Discussion

6.1. Effectiveness of AI Agents

This study confirmed that AI agents could be effective in engaging elderly individuals in conversation, particularly those with diminished cognitive functions. AI conversations resulted in a higher number of utterances compared to human conversations among participants from the day service center (p = 0.05, t = −1.96), suggesting that AI could facilitate dialogue under certain conditions equally well or better than humans. However, no significant differences in the number of utterances were observed among participants from the senior talent center (p = 3.09, t = 14.7), indicating that the effectiveness of AI may vary depending on the cognitive state of the participants. The increase in silence duration during AI conversations also deserves attention; it may indicate that AI conversations could impose cognitive or emotional burdens on some elderly individuals, necessitating further refinement in AI agent design to ensure psychological comfort. Additionally, research by Wang et al. (2022) has explored the potential of silence duration as a diagnostic biomarker for cognitive impairment, suggesting that if the differences in silence duration between human and AI conversations are mitigated, AI agents could contribute to the diagnosis of cognitive disorders [].

6.2. Differences in Responses by Question Item

This study compared the seconds of speech and silence for each question item in conversations with AI and humans among participants from the day service center. A paired two-tailed t-test at a 5% level showed significant differences in speech seconds for questions about siblings (p = 0.00, t = 3.30), post-middle school activities (p = 0.00, t = 3.20), meal preparation (p = 0.04, t = −2.28), laundry (p = 0.02, t = −2.58), and respected individuals (p = 0.00, t = 3.50). The average speech duration was longer in human interactions for questions about siblings (40.65 s for humans vs. 13.00 s for AI), post-middle school activities (36.56 s for humans vs. 15.07 s for AI), and respected individuals (45.94 s for humans vs. 21.67 s for AI). These questions typically elicited responses related to memories or emotions. In contrast, questions about current daily activities such as meal preparation and laundry prompted longer responses in AI conversations, indicating the effectiveness of AI in facilitating dialogue on routine topics.
Cognitive fluctuations in dementia patients may affect the activities of daily living (ADL) and the quality of life [], and the various challenges faced in the daily lives of dementia patients can create difficulties for both the patients themselves and their families []. As such, the assessment of ADL is considered highly important for rehabilitation and treatment [].
A study by Yamada et al. (2021) found that assessing speech responses to questions about daily life using a tablet-based system could detect early signs of Alzheimer’s disease with accuracy, comparable to neuropsychological tests []. In our study, the trend indicating the effectiveness of AI in facilitating conversations about daily life suggests the potential use of AI in ADL assessment, providing new insights into the possibilities for utilizing AI in this area.

6.3. Effects of Age and Cognitive Function on Question Responses

Participants from the senior talent center, who generally had better cognitive function, showed no significant differences in speech duration or number of utterances between the AI and human conversations (p = 5.50, t = 10.07 for speech duration; p = 3.09, t = 14.7 for utterances), suggesting that AI agents can be as effective as humans in engaging cognitively intact elderly individuals. Similarly, no significant differences in silence duration were observed (p = 0.44, t = 0.77). On the other hand, participants from the day service center, who had lower cognitive functions, showed no significant differences in speech duration or number of utterances between AI and human conversations (p = 0.57, t = 0.57 for speech duration; p = 1.75, t = −16.67 for utterances). However, the increased silence duration in AI conversations suggests that AI interactions might be causing discomfort or confusion for these individuals, highlighting the need to customize AI dialogue systems based on individual cognitive abilities.

6.4. Future Work

Given that AI can elicit speech and speech duration at levels comparable to or better than human interactions, it stands as an effective dialogue method. However, the increased silence duration observed in AI conversations suggests room for improvement in AI’s questioning and response strategies. While age and cognitive function do not appear to significantly influence the number of words spoken or the duration of speech, they do affect silence duration, indicating the need for further exploration of factors that influence this aspect. Diederich et al. (2022) reviewed multiple studies on conversational agents. They highlighted that, due to advances in natural language processing, conversational agents are gaining increasing attention in both academia and practice. The authors emphasized the importance of designing conversational agents to meet their intended purposes and better understanding how humans interact with them. They also identified the need to accommodate diverse user groups, such as individuals with disabilities and the elderly, as a future challenge []. The fact that we were able to obtain quantitative data by examining the cognitive function of each interaction with a human or with an AI by dividing the cognitive function into groups will contribute to the development of AI agents for a variety of users in the future.

7. Limitations

This study conducted dialogue experiments based on a set of questions designed for routine conversation. Future research should explore the long-term effects of sustained interactions with AI to evaluate the sustainability and effectiveness of AI-supported communication aids.
In this study, the subjects were limited to elderly persons living in the community, and the conversation style was semi-structured interviews. Different results could be obtained with the elderly residing in nursing homes or hospitals or with a different conversation style. In Japan, it is difficult to select elderly people with cognitive decline as research subjects, and the sample size was small for this experiment. However, verification by prototype is necessary under such circumstances, and therein lies the contribution of this study. In the future, a follow-up study with a larger sample size is needed, using this study as a starting point.
In this study, the subjects were limited to elderly persons living in the community, and the conversation style was semi-structured interviews. Different results could be obtained if the elderly were residing in nursing homes or hospitals or if different conversation styles were used. In addition, although the MMSE was used for the cognitive function test in this study, different results may be indicated when different tests are used. In Japan, it is difficult to select elderly people with cognitive decline as research subjects, and the sample size in this experiment was small. However, even under such circumstances, verification by prototype is necessary, and this is where the contribution of this study lies. In the future, a follow-up study with a larger sample size is needed, using this study as a starting point.

8. Conclusions

This study examined the interaction dynamics of elderly participants during conversations mediated by AI versus those with humans. Key metrics, such as the number of utterances, speech duration, and silence duration, were recorded and analyzed to understand the impact of cognitive function and age on communication patterns.
The results indicated that while there were no significant differences in the overall number of utterances and speech duration between AI and human conversations, noticeable differences emerged when segmented by participant groups. Specifically, participants from the day service center exhibited significantly longer silence durations during AI-mediated conversations, suggesting a potential discomfort or difficulty in engaging with AI compared to human interlocutors. For participants from the senior talent center, the difference in interaction quality between AI and human conversations was minimal, indicating a potential adaptation or familiarity with technology. The absence of significant differences in utterance count and speech duration between AI and human interactions in this group supports the feasibility of using AI for routine conversational engagement with cognitively healthier elderly individuals.
The findings underscore the importance of tailoring AI-mediated interactions to accommodate varying levels of cognitive function and comfort with technology among elderly populations. Future research should focus on refining AI interaction protocols to enhance engagement and reduce discomfort, particularly for those with lower cognitive function or less familiarity with AI.

Author Contributions

Conceptualization, T.I.; methodology, T.I.; software, T.I.; validation, T.I.; data curation, T.I.; writing—original draft preparation, T.I.; writing—review and editing, T.I.; visualization, T.I.; supervision, K.I., K.N. and Y.C.; project administration, K.I., K.N. and Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study was conducted with the approval of the ethical review boards of the University of Tokyo and the Tokyo Metropolitan Institute of Technology (code: 23009, 19 October 2023).

Data Availability Statement

Based on the requirements for the ethical review and the protocols outlined by our University for storing and sharing data, our data, which include information on individuals with dementia, will be disclosed upon reasonable request.

Acknowledgments

This study was conducted with the cooperation of the Silver Human Resources Center. We would like to thank all the study participants. In this study, the agent used Open AI’s ChatGPT API to generate appropriate responses that match the elderly person’s responses. All individuals listed in this section agree to this statement.

Conflicts of Interest

The authors declare no conflicts of interest. Tsukushikai Medical Corporation had no role in the design of this study; in the collection, analysis, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Cabinet Office Japan. Annual Report on the Aging Society: 2020 (Summary); Cabinet Office: Tokyo, Japan, 2020.
  2. Cornwell, E.Y.; Waite, L.J. Social disconnectedness, perceived isolation, and health among older adults. J. Health Soc. Behav. 2009, 50, 31–48. [Google Scholar] [CrossRef] [PubMed]
  3. Salthouse, T.A. Consequences of age-related cognitive declines. Annu. Rev. Psychol. 2012, 63, 201–226. [Google Scholar] [CrossRef] [PubMed]
  4. Hertzog, C.; Kramer, A.F.; Wilson, R.S.; Lindenberger, U. Enrichment effects on adult cognitive development: Can the functional capacity of older adults be preserved and enhanced? Psychol. Sci. Public Interest 2008, 9, 1–65. [Google Scholar] [CrossRef] [PubMed]
  5. Pu, L.; Moyle, W.; Jones, C.; Todorovic, M. The effectiveness of social robots for older adults: A systematic review and meta-analysis of randomized controlled studies. Gerontologist 2018, 58, e37–e51. [Google Scholar] [CrossRef] [PubMed]
  6. Broadbent, E.; Garrett, J.; Jepsen, N.; Li Ogilvie, V.; Ahn, H.S.; Robinson, H.; Peri, K. Using robots at home to support patients with chronic obstructive pulmonary disease: Pilot randomized controlled trial. J. Med. Internet Res. 2018, 20, e382. [Google Scholar] [CrossRef] [PubMed]
  7. Chan, M.Y.; Haber, S.; Drew, L.M.; Park, D.C.; Campbell, K.L. Longitudinal evidence for functional specialization of the neural circuit supporting working memory training. Cereb. Cortex 2020, 30, 5497–5510. [Google Scholar] [CrossRef] [PubMed]
  8. Ambrosini, E.; Cid, M.; de Isla, C.G.; Salamanca, P.; Borghese, N.A.; Ferrante, S.; Caielli, M.; Milis, M.; Loizou, C.; Azzolino, D.; et al. Automatic speech analysis to early detect functional cognitive decline in elderly population. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019. [Google Scholar] [CrossRef] [PubMed]
  9. Robin, J.; Xu, M.; Kaufman, L.D.; Simpson, W. Using digital speech assessments to detect early signs of cognitive impairment. Front. Digit. Health 2021, 3, 749758. [Google Scholar] [CrossRef] [PubMed]
  10. Miotto, R.; Wang, F.; Wang, S.; Jiang, X.; Dudley, J.T. Deep learning for healthcare: Review, opportunities and challenges. Brief. Bioinform. 2018, 19, 1236–1246. [Google Scholar] [CrossRef] [PubMed]
  11. Liu, L.; Stroulia, E.; Nikolaidis, I.; Miguel-Cruz, A.; Rios Rincon, A. Smart homes and home health monitoring technologies for older adults: A systematic review. Int. J. Med. Inform. 2016, 91, 44–59. [Google Scholar] [CrossRef] [PubMed]
  12. Hasan, S.A.; Sapci, H.A. Innovative assisted living tools, remote monitoring technologies, artificial intelligence-driven solutions,. and robotic systems for aging societies: Systematic review. JMIR Aging 2019, 2, e15429. [Google Scholar] [CrossRef] [PubMed]
  13. Asgharian, P.; Panchea, A.M.; Ferland, F. A review on the use of mobile service robots in elderly care. Robotics 2022, 11, 127. [Google Scholar] [CrossRef]
  14. Igarashi, T.; Nihei, M.; Inoue, T.; Sugawara, I.; Kamata, M. Eliciting a User’s Preferences by the Self-Disclosure of Socially Assistive Robots in Local Households of Older Adults to Facilitate Verbal Human–Robot Interaction. Int. J. Environ. Res. Public Health 2022, 19, 11319. [Google Scholar] [CrossRef] [PubMed]
  15. NEC Platforms Papero I. Available online: https://www.necplatforms.co.jp/solution/papero_i/index.html (accessed on 26 April 2024).
  16. Kobayashi, T.; Miyazaki, T.; Arai, K. Study on a system to detect early signs of dementia through a social media intermediary robot. IEICE Trans. Inf. Syst. 2022, 105, 533–545. [Google Scholar] [CrossRef]
  17. Ma, Q.; Chen, K.; Chan AH, S.; Teh, P. A review of recent studies. In Proceedings of the Human Aspects of IT for the Aged Population: Design for Aging: First International Conference, ITAP 2015, Los Angeles, CA, USA, 2–7 August 2015; Proceedings, Part I 1. Springer International Publishing: Berlin/Heidelberg, Germany, 2015. [Google Scholar] [CrossRef]
  18. Lee, C.; Coughlin, J.F. PERSPECTIVE: Older adults’ adoption of technology: An integrated approach to identifying determinants and barriers. J. Prod. Innov. Manag. 2015, 32, 747–759. [Google Scholar] [CrossRef]
  19. Blanka, K.; Poulova, P. Older people and technology acceptance. In Proceedings of the Human Aspects of IT for the Aged Population: Acceptance, Communication and Participation: 4th International Conference, ITAP 2018, Held as Part of HCI International 2018, Las Vegas, NV, USA, 15–20 July 2018; Proceedings, Part I 4. Springer International Publishing: Berlin/Heidelberg, Germany, 2018. [Google Scholar] [CrossRef]
  20. Luxton, D.D. (Ed.) Artificial Intelligence in Behavioral and Mental Health Care; Academic Press: Cambridge, MA, USA, 2015. [Google Scholar] [CrossRef]
  21. Walker, G.; Morris, L.A.; Christensen, H.; Mirheidari, B.; Reuber, M.; Blackburn, D.J. Characterising spoken responses to an intelligent virtual agent by persons with mild cognitive impairment. Clin. Linguist. Phon. 2021, 35, 237–252. [Google Scholar] [CrossRef] [PubMed]
  22. NTT Communications. Free Trial of “Brain Health Check Toll-Free Dial” Begins. Available online: https://www.ntt.com/about-us/press-releases/news/article/2022/0921.html (accessed on 26 April 2024).
  23. Hikino, J.; Nakano, Y.; Yasuda, K. Communication support for dementia patients using a conversational agent. In Proceedings of the 73rd National Convention of the Information Processing Society of Japan, Tokyo, Japan, 2–4 March 2011; Volume 1, pp. 195–196. [Google Scholar]
  24. Li, H.; Zhang, R.; Lee, Y.C.; Kraut, R.E.; Mohr, D.C. Systematic review and meta-analysis of AI-based conversational agents for promoting mental health and well-being. NPJ Digit. Med. 2023, 6, 236. [Google Scholar] [CrossRef] [PubMed]
  25. Oh, Y.H.; Chung, K.; Ju, D.Y. Differences in interactions with a conversational agent. Int. J. Environ. Res. Public Health 2020, 17, 3189. [Google Scholar] [CrossRef] [PubMed]
  26. Oliveira, J.D.; Engelmann, D.C.; Kniest, D.; Vieira, R.; Bordini, R.H. Multi-agent interaction to assist visually-impaired and elderly people. Int. J. Environ. Res. Public Health 2022, 19, 8945. [Google Scholar] [CrossRef] [PubMed]
  27. Rodríguez-Martínez, A.; Amezcua-Aguilar, T.; Cortés-Moreno, J.; Jiménez-Delgado, J.J. Qualitative Analysis of Conversational Chatbots to Alleviate Loneliness in Older Adults as a Strategy for Emotional Health. Healthcare 2023, 12, 62. [Google Scholar] [CrossRef] [PubMed]
  28. Schöbel, S.; Schmitt, A.; Benner, D.; Saqr, M.; Janson, A.; Leimeister, J.M. Charting the evolution and future of conversational agents: A research agenda along five waves and new frontiers. Inf. Syst. Front. 2024, 26, 729–754. [Google Scholar] [CrossRef] [PubMed]
  29. Fukaya, Y.; Kitamura, T.; Koyama, S.; Yamakuma, K.; Sato, S. Analysis of utterances by older persons in ‘life-worldly’ communication with caregivers in Japan. J. Nurs. Care 2016, 5, 1000367. [Google Scholar] [CrossRef] [PubMed]
  30. Schröder, J.; Knebel, M. P2-061: Dementia and communication. Alzheimer’s Dement. 2010, 6, S331. [Google Scholar] [CrossRef]
  31. Lee, H.; Gayraud, F.; Hirsch, F.; Barkat-Defradas, M. Speech Dysfluencies in Normal and Pathological Aging: A Comparison between Alzheimer Patients and Healthy Elderly Subjects. ICPhS 2011, 17, 153–156. [Google Scholar]
  32. Igarashi, T.; Umeda-Kameyama, Y.; Kojima, T.; Akishita, M.; Nihei, M. Questionnaires for the Assessment of Cognitive Function Secondary to Intake Interviews in In-Hospital Work and Development and Evaluation of a Classification Model Using Acoustic Features. Sensors 2023, 23, 5346. [Google Scholar] [CrossRef] [PubMed]
  33. Chen, D.; Alsadoon, A.; Prasad, P.W.C.; Elchouemi, A. Early diagnosis of Alzheimer using mini mental state examination method: MMSE. In Proceedings of the 2017 8th International Conference on Information and Communication Systems (ICICS), Irbid, Jordan, 4–6 April 2017. [Google Scholar] [CrossRef] [PubMed]
  34. Sugishita, M.; Koshizuka, Y.; Sudou, S.; Sugishita, K.; Hemmi, I.; Karasawa, H.; Ihara, M.; Asada, T.; Mihara, B. Validity and reliability of the Japanese version of the Mini-Mental State Examination (MMSE-J). J. Cogn. Neurosci. 2018, 20, 91–110. [Google Scholar] [PubMed]
  35. Igarashi, T.; Umeda-Kameyama, Y.; Kojima, T.; Akishita, M.; Nihei, M. Assessment of adjunct cognitive functioning through intake interviews integrated with natural language processing models. Front. Med. 2023, 10, 1145314. [Google Scholar] [CrossRef] [PubMed]
  36. 3D Character Creation Software VRoid Studio. Available online: https://vroid.com/studio (accessed on 26 April 2024).
  37. GitHub Pixiv Three-Vrm. Available online: https://github.com/pixiv/three-vrm/releases (accessed on 26 April 2024).
  38. Web Speech API—MDN Web Docs. Available online: https://developer.mozilla.org/ja/docs/Web/API/Web_Speech_API (accessed on 26 April 2024).
  39. AI Service for Generating Voice and Facial Motion Koemotion. Available online: https://rinna.co.jp/news/2023/06/20230612.html (accessed on 26 April 2024).
  40. Introducing ChatGPT and Whisper APIs. Available online: https://openai.com/index/introducing-chatgpt-and-whisper-apis/ (accessed on 26 April 2024).
  41. Wang, H.L.; Tang, R.; Ren, R.J.; Dammer, E.B.; Guo, Q.H.; Peng, G.P.; Cui, H.L.; Zhang, Y.M.; Wang, J.T.; Xie, X.Y.; et al. Speech silence character as a diagnostic biomarker of early cognitive decline and its functional mechanism: A multicenter cross-sectional cohort study. BMC Med. 2022, 20, 380. [Google Scholar] [CrossRef] [PubMed]
  42. Sun, M.; Mainland, B.J.; Ornstein, T.J.; Mallya, S.; Fiocco, A.J.; Sin, G.L.; Shulman, K.I.; Herrmann, N. The association between cognitive fluctuations and activities of daily living and quality of life among institution-alized patients with dementia. Int. J. Geriatr. Psychiatry 2018, 33, e280–e285. [Google Scholar] [CrossRef] [PubMed]
  43. Terada, S.; Nakashima, M.; Wakutani, Y.; Nakata, K.; Kutoku, Y.; Sunada, Y.; Kondo, K.; Ishizu, H.; Yokota, O.; Maki, Y.; et al. Social problems in daily life of patients with dementia. Geriatr. Gerontol. Int. 2019, 19, 113–118. [Google Scholar] [CrossRef] [PubMed]
  44. Aiko, O.; Maejima, S.-I.; Arai, H. Assessment of Activities of Daily Living (ADL) in the Elderly and Persons with Dementia. Jpn. J. Rehabil. Med. 2021, 58, 991–997. [Google Scholar] [CrossRef]
  45. Yamada, Y.; Shinkawa, K.; Kobayashi, M.; Nishimura, M.; Nemoto, M.; Tsukada, E.; Ota, M.; Nemoto, K.; Arai, T. Tablet-based automatic assessment for early detection of Alzheimer’s disease using speech responses to daily life questions. Front. Digit. Health 2021, 3, 653904. [Google Scholar] [CrossRef] [PubMed]
  46. Stephan, D.; Benedikt, B.A.; Stefan, M.; Lutz, K. On the design of and interaction with conversational agents: An organizing and assessing review of human-computer interaction research. J. Assoc. Inf. Syst. 2022, 23, 96–138. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.