1. Introduction
Information and communication technologies (ICTs) became facilitators of a large portfolio of services, devices, and applications aimed at improving people’s lives. The use of ICT for neuropsychological assessment brings with it some relevant advantages [
1], such as the standardization of administration procedures, greater accuracy in the presentation of response times and latencies, simpler and less error-prone data collection, and the possibility of capturing and monitoring dynamic indicators like cognitive markers such as writing time latency [
2] or other non-cognitive indicators such as gait patterns [
3]. By facilitating the administration of tests digitally or via online applications, ICT can also support better randomization of stimuli presentation across repeated administrations with the generation of varying stimuli.
Presently, we have at our disposal advanced data analysis models that allow us to take advantage of the information collected during the interaction with users to evaluate their cognitive capacity. One of the most prominent approaches in this field is machine learning, which is considered a branch of artificial intelligence that focuses on creating computing solutions that do not require task-specific programming [
4], but are able to learn by training test cases or data, allowing the program to acquire knowledge and improve its performance based on accumulated experience. Currently, these machine learning-based analytics techniques are being increasingly used [
5,
6]. In the field of cognitive health, machine learning is complemented with advanced statistical techniques (e.g., multi-variate linear regression, logistic regression, classification and regression trees (CART), etc.) to infer users’ cognitive abilities from their interaction with different types of technological tools, such as tablets or smart speakers. These techniques are also combined with other advanced analytical and prediction systems, such as neural networks [
7] and item response theory [
8].
Additionally, the use of intelligent conversational agents hosted by tablet computers or smart speakers is an emerging field of research, with the few published studies mostly quasi-experimental, and both their validity and reliability are rarely estimated. Despite this, recent literature points in favor of the use of conversational agents in support of patients suffering from mild cognitive impairment (MCI) [
9,
10]. In other words, intelligent agents become an element that, at least in principle, supports the introduction of tablets and smart speakers as the interaction devices of choice to provide assistance and support to senior adults, exploiting their intuitive and simple interaction capabilities.
In this context, the digital divide between senior adults and the general population represents a significant disparity in access to and utilization of digital technologies, particularly the internet and modern devices like smartphones, tablets, and smart speakers. This divide is primarily driven by factors related to age, as older individuals may have grown up during a time when such technologies were less prevalent or were not a part of their education [
11]. Additionally, socioeconomic status can play a role, as seniors from lower-income backgrounds may face barriers in obtaining the necessary equipment and internet access [
12]. Apart from limiting the opportunities for seniors to access critical services, connect with loved ones, and participate in the digital aspects of modern life, the digital divide may generate negative perceptions about new technological services and devices, which in turn may compromise some of the benefits expected from the introduction of technology in healthcare settings [
13,
14].
We can find in the literature some developments in oral and touch interactions for senior adults in cognitive assessment technologies. For example, [
15] focuses on the use of cognitive interaction systems (CISs) to enhance the quality of life for older adults, while [
16] aimed to develop performance evaluation measures for touch-based target selection difficulties in senior adults. Additionally, the use of virtual environments, such as non-immersive exergames and immersive virtual reality environments, has shown potential for assessing cognitive capabilities in elderly individuals [
17]. Furthermore, a robotic platform was developed to conduct cognitive orientation assessments for the elderly, using natural language processing to evaluate user responses [
18]. Finally, an analysis of older users’ experience with portable electronic devices highlighted difficulties with finding and configuring settings, as well as issues with keyboard, font size, and touchscreen interactions [
19].
Recent studies have also explored technology acceptance models applicable to senior users. Among them, Ref. [
20] proposed a new conceptual framework for technology acceptance targeted at older Korean adults living in rural areas, finding significant positive paths from external controls, attitudinal beliefs, and cognitive health to internal abilities. Ref. [
21] discussed an online mapping technology (OMT) acceptance model for older adults, finding that facilitating conditions, compatibility, and self-satisfaction positively influenced the perceived ease of use and perceived usefulness of OMT. Ref. [
22] compared frail, prefrail, and robust groups to find that gerontechnology self-efficacy, health contexts and abilities, and attitudinal factors significantly affect intentions to use daily living assistive technologies. Ref. [
23] collected a systematic literature analysis that served to identify factors influencing older adults’ use of sports technology, including intention, perceived ease of use, perceived usefulness, and attitudes. These studies provide insights into technology acceptance among senior adults and can inform the development of targeted interventions and education programs.
The first field studies carried out by the atlanTTic Research Center at the University of Vigo in collaboration with the main socio-health entities in the region provided initial evidence on the psychometric validity of this approach to cognitive evaluation and its pertinence to assess different relevant aspects related to cognitive decline. These studies were carried out by means of pilots involving more than 150 participants, in which two instruments were used. The first one consists of three serious games running on a tablet computer and is aimed at assessing episodic memory, executive functions, and semantic memory, which are the main cognitive markers of MCI and dementia [
24]. These games are based on psychometrically validated pen-and-paper tests, but provide a more interactive and motivating environment than their classical counterparts (i.e., the California Verbal Learning Test [
25], the Rotor Pursuit Test [
26], and the Pyramids and Palm Trees Test [
27], respectively). The second tool is a cognitive impairment detection agent called DigiMoca [
28]. This tool consists of an intelligent conversational agent running on a smart speaker that behaves as an adaptation of the telephone version of the classic MoCa test (T-MoCa [
29]), which evaluates attention, concentration, memory, language, abstraction, calculation, and orientation.
No matter the advantages that may be brought by the incorporation of state-of-the art technological devices in cognitive screening for dementias, it is essential to investigate the perception of users when using them, as this aspect is instrumental to investigating the acceptance and, ultimately, the feasibility of the introduction of these technologies in actual clinical settings. Knowing the experience and opinions of users can also provide invaluable information to improve and adapt technological tools to their specific needs, allowing us to identify possible barriers or challenges that hinder the successful implementation of these tools, and thus be able to take appropriate measures to overcome them. This is particularly important in the case of senior users. A person-centered approach helps to ensure that cognitive screening technologies are more accessible and user-friendly for the older population, enabling accurate early detection, providing more comprehensive monitoring of their cognitive status, and providing appropriate technological support to their daily activities and to maintain their cognitive reserve.
This study relies on the technology acceptance model (TAM) [
30] to assess user acceptability and perception. This instrument is a theoretical model developed in the field of research on technology adoption and acceptance and is based on the premise that the acceptance and use of technology are mainly influenced by individuals’ perceptions of its usefulness and ease of use, variables that eventually determine an individual’s attitude towards technological devices.
TAM is administered by means of a five-point Likert scale questionnaire and aims to capture information on the perception and application of technologies in daily life. The Likert scale is a widely used tool for measuring attitudes and perceptions, as it allows respondents to express their opinions on an intuitive scale, offering a structured way to gauge perceived utility, ease of use, and enjoyment. The five-point Likert scale provides a range of responses sufficient for capturing subtle differences in senior users’ perceptions.
In this study, the target technologies are digital cognitive tests performed on a tablet and a smart speaker. Participating subjects complete a TAM questionnaire before and after the use of the two instruments under study. The information collected is used to evaluate three main aspects, namely perceived usefulness (PU), which refers to the degree to which a person believes that the use of a particular technology will improve their performance when doing a specific activity; perceived ease of use (PEOU), which indicates to what extent a person believes that, by using a technological system, they will perform their tasks with less effort; and finally, perceived enjoyment (PE), which refers to the degree to which a person finds an activity to be pleasant or enjoyable when carried out with the support of a certain technology. These perceptions influence the individual’s attitude toward technological tools, their intention to use them, and ultimately, their actual usage patterns.
Thus, the main objective of this research is to conduct a comparative study on the acceptability of the two tools mentioned above, a tablet and a smart speaker hosting Panoramix and DigiMoca, respectively, by regular attendees of a day care center for seniors. The study will focus on assessing how users perceive, accept, and enjoy each of these tools in a day center context and how they perceive them in relation to their classical counterparts, namely pen-and-paper tests in the case of Panoramix and person-to-person interviews in the case of DigiMoca. As a secondary objective, information on the perceived acceptability of the day center professionals involved will also be gathered.
2. Materials and Methods
This study involved a total of 25 volunteers, 16 women and 9 men, aged between 60 and 93 years, living in the Vigo area, Spain, with a mean age of 79.80 and a standard deviation of 7.842 years. In addition, the scores of each of the participants on the Barthel scale [
31] were collected. This scale measures autonomy for activities of daily living, with a maximum score of 100 indicating the maximum degree of independence. Thirteen participants scored 100 (i.e., completely independent), one person scored 95 (i.e., low dependence), ten people scored between 61 and 90 (i.e., moderate dependence), and the remaining participant scored 60 (i.e., severe dependence).
Participants were recruited at the Parque Castrelos day care center; all of them signed an informed consent form approved by the Research Ethics Committee of Galicia, which also approved this study (protocol 2023/15 [
32]), thus ensuring compliance with the applicable ethical regulations and the protection of the privacy of the participants according to the EU’s General Data Protection Regulation.
In order to define the characteristics of the sample, the initial inclusion and exclusion criteria applied are enumerated below:
Inclusion factors:
Persons without cognitive impairment (GDS 1), very mild cognitive impairment (GDS 2), and mild cognitive impairment (GDS 3).
Persons over 65 years of age.
Users of Parque Castrelos day center.
Exclusion factors:
Persons unable to consent to the study.
Persons with moderate or advanced cognitive impairment (GDS ≥ 4).
Individuals with an active psychiatric condition.
Individuals with severe hearing or vocal impairments.
Persons declaring technological aversion or phobia.
Illiterate individuals.
All participants were carefully selected according to the previously mentioned criteria. GDS refers to the Global Deterioration Scale [
33], a popular instrument for the assessment of primary degenerative dementia and the delineation of its stages. It was considered convenient, upon recommendation of the day center, to include two additional subjects who had a GDS level within the inclusion criteria but whose ages were lower than the 65 years initially proposed (i.e., 60 and 62 years old). Eventually, 16 people classified as GDS 1 or GDS 2, which indicates normal cognitive function or with very slight impairment, participated in the study, together with 9 people classified as GDS 3, which corresponds to mild cognitive impairment (cf.
Table 1).
The selection and classification of participants based on the GDS scale was carried out by staff external to the study, specifically the multidisciplinary team of Parque Castrelos day care center, which is composed of a psychologist, a social educator, a physiotherapist, and the coordinator of the team of geroculturists.
All participants had already undergone at least one classical cognitive assessment process based on pencil-and-paper tests and, more specifically, CLVT’s Spanish version and the card version of Pyramids and Palm Trees. These instruments were the ones that would be administered by means of a tablet computer. Consequently, during the pilot, information would also be gathered to compare perceptions about classical and digital tests. However, this was not the case for T-MoCa’s smart speaker version, and no participant had previous experience with only-verbal cognitive assessments. To guarantee a similar baseline for both devices, participants were also tested with T-MoCa before interacting with tablets and smart speakers. This also served as confirmation of participants’ GDS level.
Each subject participated in four individual weekly sessions, carried out over a period of four weeks. For logistic reasons (i.e., only a small number of smart speakers were available), the subject sample was randomly divided into two groups (cf.
Table 2):
- Group 1
composed of 13 people, 4 of whom were men and 9 were women. Within this group, 9 participants were classified as GDS 1 or 2, while 4 of them presented mild cognitive impairment (GDS 3). These participants started their pilot sessions by interacting with the smart speaker.
- Group 2
composed of 12 people: 5 men and 7 women. In this group, 7 users were assessed as GDS 1 or 2, and 5 as GDS 3. Participants in this group started the pilot sessions with MoCa’s telephone version (T-MoCa).
Table 3 summarizes the relations between group and gender variables. During the first session, the perception and acceptability of popular technological devices were evaluated by means of the TAM questionnaire. Once the survey was completed, T-MoCa or its smart speaker version was administered, depending on the subject’s assigned group (i.e., one group was evaluated using T-MoCa and the second group using DigiMoca). All participants (i.e., both groups) were eventually evaluated with both T-MoCa and its digital counterpart, and T-MoCa outcomes served as a golden standard. As T-MoCa and its smart speaker version are conceptually equivalent, participants were administered T-MoCa and DigiMoca in separate sessions to minimize possible learning effects.
During the second session, each participant was administered the cognitive assessment solution not experienced in the first session, that is, group 1, which was previously administered T-MoCa, was assessed with DigiMoca, while group 2, which did not interact with T-MoCa during the first session, was assessed by means of a smart speaker. This second session took approximately 15 min.
During the third session, all participants interacted with a tablet computer to play the Panoramix serious games.
Finally, in the fourth and last session, the instrument hosted by the smart speaker was administered for a second time to all participants, and the session was completed with a post-pilot acceptability analysis using the TAM model discussed above.
IBM’s statistical package SPSS v. 27 [
34] was used to perform statistical analysis. Descriptive statistics and frequencies were obtained for demographic data (i.e., gender, GDS level, group, and age variables) and questionnaire responses. The descriptive statistics utilized helped to summarize and describe the perceptions gathered, as they provide a clear understanding of the central tendency, variability, and distribution of responses, enabling us to identify patterns and trends in the perceptions of senior adults regarding both devices. For the analysis of TAM responses, a reliability analysis was also performed using the Cronbach’s alpha coefficient (
). This coefficient provides a measure of reliability that indicates the internal consistency of a set of items in a questionnaire [
35]. It makes it possible to evaluate the reliability of the responses and to ensure that items in the questionnaire consistently measure the variable intended to be measured.
In addition, TAM items were grouped into three dimensions: (1) perceived usefulness (PU), (2) perceived ease of use (PEOU), and (3) perceived enjoyment (PE). Next, the mean of each dimension was calculated, and a visual grouping was carried out with the aim of facilitating data visualization and subsequent analysis. Finally, as one of the aims of this study was to detect different perceptions about technology before and after interacting with the devices mentioned, a study of the normality distribution was carried out using the Wilcoxon rank test for related samples of less than 50 participants [
36]. In this context, this test aids in verifying whether the responses collected from senior adults adhere to a normal distribution, which is essential for conducting our statistical analyses accurately. The choice of using the Wilcoxon rank test for smaller samples ensures the assumption of normality for subsequent statistical analyses. It is most relevant to validate this assumption, especially with smaller sample sizes, to avoid biased conclusions or inaccurate interpretations of the data. This made it possible to decide between parametric and nonparametric tests to analyze the existence of significant differences before and after administration of the technological instruments.
The combination of the Likert scale for perceptions, descriptive statistics for summarizing data, Cronbach’s alpha coefficient for reliability analysis, and Wilcoxon rank test for verifying data normality facilitated a comprehensive analysis of perceptions and acceptance before and after exposure to tablets and smart speakers among senior adults. They provided valuable insights into how their opinions change following interaction with these technologies, aiding in understanding preferences and potential barriers to adoption.
For the analysis of acceptability and usability from the perspective of health practitioners at the day center, the Post-Study System Usability Questionnaire (PSSUQ) was utilized. This questionnaire includes a total of 19 items that collect information on the general perceptions of usability and usefulness of technology. Responses correspond to a Likert scale with 7 options, reflecting a range of perceptions from 1 (strongly disagree) to 7 (strongly agree). PSSUQ collects data on three dimensions of interest, namely system usability (items 1–8), information quality (items 9–15), and interface quality (items 16–18) [
37].
This questionnaire was applied to three professionals from different fields who were present during the administration of the smart speaker-based cognitive evaluations. As pointed out above, these professionals are part of the multidisciplinary team of the Parque Castrelos day center, and include a therapist, a social educator, and a psychologist, aged between 28 and 50 years. In this study, due to the limited size of the sample and the fact that all three participants already had extensive experience with tablet computers, only information on their perceptions of smart speakers was gathered.
3. Results
As discussed above, a reliability analysis of TAM responses was performed using the Cronbach’s alpha coefficient. The result obtained, , confirms the high reliability of the TAM survey used.
3.1. Frequency of Use and General Perception of Technological Devices
First, an analysis of the frequency of use of technological instruments and the general perception of technology was carried out. As depicted in
Figure 1, television is the most used technological tool (
, cf.
Table 4); 52% of respondents (13 subjects) reported frequent use of the television, while 36% (9 individuals) used this device intensively.
In contrast, tablet computers (
) along with e-mail (
) and social networks (
) were the tools least used by participants in this pilot study, with a high percentage of people claiming that they never utilized them. More specifically, 88% never used a tablet or e-mail, and 84% never used social networks. In relation to computers (
) and the internet (
), 80% and 72%, respectively, claimed that they never used them. On the other hand, as shown in
Figure 1, cell phone usage (
) presents greater variability due to the fact that the sample is dispersed along all frequency of use ranges.
Despite the fact that participants reported a low frequency of use of most technological devices with the exception of television (
), 52% of them perceived technology as very useful, while 40% considered it to be quite useful (cf.
Figure 2). In addition, 88% of respondents admitted to enjoying using technological devices very much (60%) or quite a lot (28%). In terms of ease of use, 44% of subjects perceived technology as very easy to use in general, while 32% found it quite easy to use.
On the other hand, when comparing the perception of technologies before and after interacting with tablets and smart speakers (cf.
Table 5), it can be observed that the mean values are very close, and the Wilcoxon test indicates that there are no significant differences in any of the dimensions (PU, PEOU, PE) before and after interacting with the devices (cf.
Table 6). Despite this, as can be seen in
Figure 2, perceived ease of use is the only dimension with a clear trend towards improvement. Specifically, it can be observed that 32% of participants considered the technological devices to be quite useful instruments before the intervention, and after interacting with them, this percentage increased to 44%. In addition, an increase in the maximum score of the scale in this dimension (perceived ease of use) is evident, rising from 44% before the study to 52% after the study.
3.2. Users’ Perceptions of Smart Speakers
In relation to the interaction with a smart speaker, significant differences (
) were observed in the three dimensions of the TAM perception survey before and after the study (cf.
Table 7).
Although it is true that a large majority, specifically 80%, of the participants were unaware of this type of technology before the pilot, once they interacted with it, a clearly positive trend can be observed, with most participants tending to rate it with high scores of 4 or 5 on the Likert scale (cf.
Figure 3). This can be observed in the post-test averages, which present high scores in relation to the three dimensions, indicating a positive perception in terms of its usefulness (
), ease of use (
), and enjoyment (
) (cf.
Table 8).
With respect to the perception of usefulness, before interacting with the smart speaker, 44% of subjects knew about this type of device but admitted to never having used it. Among them, four people (16%) perceived smart speakers as not useful at all, but this dropped to 0% after the experience. Five users (20%) already considered them to be quite useful, with scores of 4 and 5 on the Likert scale. These values increased considerably to 40% (10 of 25 participants) who rated smart speakers with a score of 4 (i.e., quite useful), while 48% (12 individuals) rated them with a score of 5. On the other hand, two participants (8%) opted for an intermediate score of 3 before and after administration, and only one person considered them not very useful, with a score of 2 (cf.
Figure 3).
In relation to the perceived ease of use, 80% of the sample had never used a smart speaker and therefore could not rate its ease of use before the study. Only five people had this experience; three of them (12%) rated smart speakers with a score of 5, considering them very easy to use, and two people (8%) considered them moderately easy to use, with a score of 3. After interacting with the speakers, these values clearly improved, since most users (56%) rated them as very easy to use, and 32% considered them quite easy to use. On the other hand, 8% of users kept scoring speakers with a 3 on the Likert scale, and only one participant did not consider smart speakers as particularly easy tools to use.
Finally, when questioned about perceived enjoyment, the five people who previously used smart speakers rated them as moderately fun (two individuals), while the other three perceived them as quite or very entertaining. After the experience, 52% of participants enjoyed them quite a lot (score of 4), and 36% provided a score of 5 out of 5. On the other hand, the two subjects who rated smart speakers with a 3 confirmed this score post-test, while only one person admitted not having enjoyed the device at all.
3.3. Users’ Perceptions of Tablet Computers
As in the case of smart speakers, non-parametric analyses of related samples indicate significant differences in the perception of participants in the three dimensions analyzed before and after interacting with tablet computers to play serious cognitive games (
, cf.
Table 9).
In addition, post-test averages present fairly high scores, indicating a favorable perception in all three domains of perceived usefulness (
), perceived ease of use (
), and perceived enjoyment (
) (cf.
Table 10).
When performing the comparative analysis before and after the interaction with a tablet, it was observed that the majority of users (92%) never used similar tools and were unaware of their operation, and 16% were aware of the existence of serious games on tablets but had never played them (cf.
Figure 4). The fact that 16% of users were aware of the existence of serious games on tablets but had never played with them stems from the difference between 92% of users not reporting their perceived ease of use or perceived enjoyment and 76% of users not reporting their perceived usefulness. The difference, 16%, who perceive a degree of usefulness but have not utilized those devices corresponds to people aware of the technological instrument, in most cases from other family members or peers attending the daycare center, but who have never interacted with it. This was confirmed explicitly by participants in this group.
Regarding the perception of usefulness after the study, 40% of participants rated this device with a score of 4, indicating that they consider it a fairly useful technological resource, and 32% perceived it as very useful (i.e., a score of 5). However, 28% rated tablets’ usefulness with a more conservative score of 3. In any case, these results reflect a positive perception of the usefulness of serious games. It should be noted that, before the application of the test, 12% of participants who were aware of this type of instrument rated them with a score lower than 3, but this perception changed after the test was administered, since no scores lower than 3 were obtained.
In relation to the perceived ease of use dimension, results show that, after the study, 48% of subjects considered this tool to be very easy to use, while 28% rated it as fairly easy to interact with (score of 4), and 24% as moderately easy (score of 3). On the other hand, in the analysis of perceived enjoyment, it is observed that the majority of participants, specifically 44%, perceived the activity as most entertaining. Likewise, 32% provided a score of 4, and 16% rated it with a 3.
3.4. Tablets vs. Smart Speakers: Comparative Analysis
In relation to the preference between smart speakers and classical interaction by means of a face-to-face interview or telephone conversation, participants provide responses that do not let us extract clear conclusions due to their limited statistical significance (
), which in turn reflect a broad range of perceptions for a small population and subtle differences. In any case, the average score is 2.84 (cf.
Table 11), which corresponds to a score that indicates that there is no clear preference between a smart speaker and traditional ways of interaction. Specifically, 28% (seven people) indicated that they did not prefer a smart speaker when compared to person-to-person communication (score of 2 out of 5). On the other hand, six of the participants (24%) expressed a more moderate stance, assigning a score of 3. In contrast, eight participants expressed a clear preference for the smart speaker option, five of them (20%) assigned a score of 4, and three respondents (12%) rated it with the maximum score of 5 points, indicating that they consider smart speakers as a highly preferable tool compared to a traditional interaction when performing cognitive assessment (cf.
Figure 5).
Regarding the preference between tablets hosting videogames and traditional pen-and-paper tests, it can be observed that there is a slightly more defined trend to prefer the administration of games by means of a tablet than in the case of smart speakers compared to traditional tests. The mean score obtained in this case is 3.52 (cf.
Table 11), which suggests that participants consider that the introduction of tablets offers certain advantages compared to traditional tests. Specifically, 28% (seven individuals) provided a score of 5, reporting a clear preference for the tablet option, and six people (24%) rated tablets with a score of 4. Only 8% favored pen-and-paper tests, and 28% did not express a definite preference, with a score of 3 out of 5.
Finally, when comparing the preference between smart speakers (
) and tablets (
), differences were found with a value of
(cf.
Table 12). These data suggest that there is some tendency to prefer the tablet hosting serious games over the smart speaker option, although this preference is barely statistically significant.
In the case of tablets, seven individuals (28%) show a clear preference for this tool, with a score of 4 out of 5, and another 28% do so with the highest score of 5 (cf.
Figure 6). This contrasts with the 24% and 4%, respectively, of users preferring the smart speaker, with most of them (52%) leaning towards a lower score of 2.
The average values of the dimensions assessed by the TAM survey (cf.
Table 13) show that the values of perceived usefulness (
), perceived ease of use (
), and perceived enjoyment (
) are slightly higher in the case of smart speakers than in the case of tablets (i.e.,
,
, and
), although again they cannot be considered statistically significant differences (i.e.,
,
,
, cf.
Table 14). Thus, we can conclude that there are no significant differences in the perception of usefulness, ease of use, or enjoyment for both approaches, although participants show a certain tendency to prefer the smart speaker option.
3.5. Practitioners’ Perceptions
The mean total score obtained was 4.41, with a standard deviation of 1.41 (cf.
Table 15), indicating an overall positive score but with some variability in the responses. In relation to the dimensions, the usefulness of the system obtained an average score of 4.81, which suggests that the evaluators considered that smart speakers were useful overall as a tool to administer cognitive assessment. On the other hand, in relation to the quality of the information, the values obtained (
) indicate that the quality provided by the smart speakers was considered moderate. Finally, with respect to the quality of the interface, a mean score of
was obtained, which suggests that the surveyed professionals perceived the interface as relatively good.
Practitioners also provided their perceptions in relation to the positive and negative aspects of smart speakers as cognitive assessment instruments, which confirmed the values provided when answering the questionnaire. All the information collected, especially the negative aspects, is valuable for further developing and optimizing the instrument since it allows the identification of areas where improvements and possible adjustments are necessary for a more effective introduction.
Among the positive aspects, it was noted that smart speakers facilitate the introduction of new technologies and dynamic learning, being considered useful and easy to use. In addition, it was emphasized that they provide clear and understandable information for users without sensory deficits. The novelty of the instrument was also perceived as attractive and interesting. Finally, its accessibility as a tool and its usefulness in the evaluation and detection of cognitive impairment were highlighted.
On the other hand, professionals identified some negative aspects that should be taken into account for the improvement of the instrument. All three contributors agree that response times should be extended since some users with cognitive impairments require more time to react to an automated voiced agent.
It was also pointed out that smart speakers are not adapted for users with hearing deficits. In addition, it was suggested to add the possibility of increasing the volume further, even adaptively, to facilitate access for people with hearing deficits. In addition, the need for a pre-assessment phase was pointed out to let subjects experience the interaction with the smart agent and thus familiarize them with the device prior to the actual cognitive evaluation session.
4. Discussion
The multifaceted perspectives of this work are examined below, encompassing senior adults’ perceptions of the technologies utilized in this study, insights into health practitioners’ and caretakers’ viewpoints, as well as an analysis of the study’s limitations.
4.1. Senior Adults’ Perceptions
As might be expected, the results obtained in relation to frequency of use indicate that television is the most-used technological device among participants (). This phenomenon can be attributed to a number of factors, such as elders’ poorer technological skills, lower purchasing power that impedes access to more advanced and expensive technologies, a lack of instructions and guidance to acquire technological skills needed for newer devices, a lack of knowledge and confidence in handling new technologies, as well as the persistence of more traditional habits. Television is often perceived as simpler and less overwhelming compared to newer technologies, which can have a steeper learning curve in a context where older adults have limited exposure to newer technologies during their formative years, so they may not have had the same opportunities to learn how to use smartphones or computers.
In addition, health-related barriers may hinder access to and learning of new technologies in this age group, while television has been a part of daily life for several decades, and many older adults are more comfortable with the technology with which they grew up. They may find it easier to use and navigate compared to newer devices, as television is typically more accessible for seniors who may have physical limitations or impairments that make using newer technologies, such as computers or tablets, more challenging.
On the other hand, participants’ general perception about popular technologies before and after their participation in this study is quite positive, and no significant ad hoc or post hoc differences were found, suggesting that their participation in this study did not have a significant impact on their perception. Furthermore, the data gathered indicate that the lack of use of new technologies is not due to a lack of interest or negative perception, since they value them as very useful (, ), easy to use (, ), and enjoyable (, ). It can also be inferred that the low use of technologies in general by this sample of people over 65 is not due to a lack of interest or perceived difficulty; instead, it can be hypothesized that it is due to existing barriers to accessing them.
In the case of the smart speakers, the results obtained reflect significantly positive perceptions of participants towards them. Statistically significant differences were found before and after their administration, since participants had no prior knowledge of them. Despite the lack of initial information, the average perceived usefulness (), perceived ease of use (), and perceived enjoyment () were quite high, as the majority expressed a favorable perception in all three domains. These results indicate a positive perception of these devices among participants, despite the lack of previous familiarity with them.
As for tablet computers, statistically significant differences were also found in all three perception dimensions before and after interacting with them. As in the previous case, this is explained by the absence of previous experience due to the lack of knowledge of this type of device. However, once the interaction sessions were completed, the perception was very favorable, with averages similar to those obtained for smart speakers (i.e., , , and ). In conclusion, despite the fact that participants lacked previous experience and knowledge about tablet computers, the experience was positive for them, concluding that they are quite useful, easy to use, and entertaining.
When analyzing the differences between the two devices (i.e., smart speakers and tablets), the results show a preference for tablets. However, as can be inferred from average values, there are no statistically significant differences in perception, as evaluated by the TAM survey. Both are similarly rated as very useful with low significance values (, , ). However, when exploring the preference between both tools for cognitive assessment, a slightly significant trend was found in favor of playing games with a tablet () compared to talking to a smart assistant hosted by a smart speaker (), with a significance value of . These results may be due to the more dynamic and interactive approach of game playing along with the favorable visual context it offers compared to the purely auditory approach of the smart speaker, which would result in a slight tendency to prefer videogames.
As for the comparison between interacting with a smart agent and the traditional face-to-face or telephone interview, no clear preference was found for one approach or the other (). However, in the case of the tablet-hosted videogames vs. pen-and-paper tests, there is a slight tendency in favor of the former (, ). This suggests that participating users consider that the introduction of tablets offers certain advantages compared to traditional tests, which may be due to aspects such as the higher level of interactivity or the active role users play with this tool.
The favorable perceptions of both smart speakers and tablets suggest that these tools are well-received and considered suitable for use among senior adults. This positive reception implies that these technologies can potentially serve as effective tools for cognitive assessment and engagement within this age group. In addition, the study’s findings, indicating perceived usefulness and added value compared to traditional approaches, highlight the potential benefits that smart speakers and tablets offer in cognitive assessment. This suggests that these modern technologies could provide advantages over conventional methods, possibly leading to more engaging and effective assessment techniques.
Moreover, the positive ratings for ease of use are especially relevant, particularly for older adults. The ease of interaction with these devices indicates that user-friendly interfaces are instrumental in ensuring successful engagement and positive experiences among senior adults. This aspect is especially relevant for technology adoption in this demographic, as ease of use directly relates to their satisfaction and overall user experience.
These positive perceptions also indicate a shift in cognitive assessment practices. Integrating user-friendly technology like smart speakers and tablets into assessment methods for senior adults may enhance engagement, accuracy, and the overall acceptability of such assessments. It also suggests the potential for these tools to be utilized in various cognitive training and intervention programs for older adults. In addition, the study’s positive outcomes have broader implications for the acceptance of technology among aging populations, as it highlights that, with appropriate design considerations and user-friendly interfaces, modern technology can be embraced and utilized effectively by senior adults, challenging stereotypes about their reluctance to adopt new technologies. However, the findings also stress the importance of developing tailored technological solutions that cater to the unique needs and preferences of our elders. This requires designing intuitive interfaces, clear instructions, and functionalities that align with the capabilities and interests of older users.
4.2. Practitioners’ Perceptions
Due to the limited number of practitioners participating in this study, although the results obtained are clearly positive, they cannot be considered to be statistically significant. However, a wider multicenter study performed by our team and that involved 24 professional caregivers [
38] indicated that the overall usability perception from practitioners is generally positive, slightly under 6 on a 7-point scale, and never drops below 5 for any of the dimensions in
Table 15, even if considering specific demographic groups based on gender, career field, and years of experience. In the mentioned extended study, with respect to practitioners’ overall usability, we obtained an average value of
with an internal consistency of (
) for all participants and all items, which did not significantly change depending on gender or career experience. When inquired about the perceived usefulness, the average rating was
(
), which was again not considerably affected by gender or career experience; although, for male practitioners, the consistency value dropped to
. The values obtained for information quality were similar to the previous dimensions, with an average value of
overall (
), with higher consistency for females (
) than males (
). Finally, with respect to interface quality, the overall mean rating was
, with the lowest internal consistency value overall (
) and considerable differences between demographic groups, with a higher
for females than males (
). This is the only dimension where internal consistency dropped below an
acceptable level, and it is probably due to the small number of PSSUQ items that it considers (i.e., only three).
4.3. Limitations
This study’s findings should be interpreted within the context of certain limitations that warrant consideration and potential implications for future research. The sample might include participants who were already comfortable with technology or had positive attitudes toward it. Future research could aim for a broader and more diverse sample that includes individuals with varying degrees of technological familiarity or those who might be more resistant to technology adoption.
Due to the implemented pilot’s logistics and the duration of the interactive sessions, perceptions were assessed after a brief interaction with the devices. Longer-term studies assessing sustained engagement and usability over time would provide a more comprehensive understanding of technology acceptance among senior adults. Along the same line, participants’ initial positive perceptions might change with prolonged use or exposure to more complex tasks. Future studies could track participants’ experiences over time to assess potential challenges or changes in perceptions as they become more familiar with the devices.
Future research could also further explore contextual factors, such as socio-economic status or cultural background, to provide a more nuanced understanding of perceptions across diverse groups, and could incorporate additional measures, such as cognitive performance, task completion rates, and qualitative feedback, to gain a more holistic view of technology acceptance and its impact on cognitive assessment.