Literacy in Artificial Intelligence as a Challenge for Teaching in Higher Education: A Case Study at Portalegre Polytechnic University

: The growing impact of artificial intelligence (AI) on Humanity is unavoidable, and therefore, “AI literacy” is extremely important. In the field of education—AI in education (AIED)—this technology is having a huge impact on the educational community and on the education system itself. The present study seeks to assess the level of AI literacy and knowledge among teachers at Portalegre Polytechnic University (PPU), aiming to identify gaps, find the main opportunities for innovation and development, and seek the degree of relationship between the dimensions of an AI questionnaire, as well as identifying the predictive variables in this matter. As a measuring instrument, a validated questionnaire based on three dimensions (AI Literacy, AI Self-Efficacy, and AI Self-Management) was applied to a sample of 75 teachers in the various schools of PPU. This revealed an average level of AI literacy (3.28), highlighting that 62.4% of responses are at levels 3 and 4 (based on a Likert scale from 1 to 5). The results also demonstrate that the first dimension is highly significant for the total dimensions, i.e., for AI Literacy, and no factor characterizing the sample is a predictor, but finding a below-average result in the learning factor indicates a pressing need to focus on developing these skills.


Introduction and Literature Review
The recent dissemination of artificial intelligence (AI) to the general public has promoted studies on its application in everyday life.The growing impact of AI on Humanity is unavoidable, and therefore, it is extremely important to understand what it is and what it can do.The set of skills that include the use, application, and interaction with AI is currently called "AI literacy".
The importance of this topic arises, from the outset, in the field of education-AI in education (AIED)-where this technology is having a huge impact on the educational community and on the education system itself.Studying the use of AIED is a search for solutions that can add value to the teaching-learning process, supporting teachers and students, highlighting the human factor, their thinking skills, teamwork and flexibility, management of knowledge, ethics, and responsibility [1].
The scientific term "artificial intelligence", as a science of intelligent machines, according to [2,3], dates back to 1956.This was followed during the 1980s by a great development of intellectual skills in machines, as well as the first attempts to replicate the teaching process using AI [1].Ref. [3] states that the entire education system should be reviewed, not only to make it more practical, but also more open to the world of work and to anticipate the transformations in knowledge.AIED began as a field of recreation and research for computer scientists, with a great impact on education [4], and today fuels the controversy referred to by [5] regarding the use of AIED and the fear that the machine will replace the teacher [6,7].
The acquisition and development of digital skills are seen as essential tools to facilitate lifelong learning, and are therefore one of the main economic concerns in most developed countries."Literacy", the ability to read and write and to perceive and interpret what is read, undergoes an important development when it becomes clear that, despite having the ability to write and read, some people are unable to understand the meaning of what they read.In terms of information and communication technologies (ICTs), digital literacy has been studied in depth, but there is still no consensual definition, because the ability to use a computer is currently an insufficient measure to define digital literacy [8].
When AI is introduced into the concept of digital literacy, the scenario becomes even more complex.According to [9], AI literacy is more than knowing how to use AI-driven tools, as it involves lower-and higher-level thinking skills to understand the knowledge and capabilities behind AI technologies and make work easier.For this author, it will not be possible to adequately understand this technology as long as we insist on considering it only as knowledge and skills, as AI involves attitudes and moral decision making for the development of AI literacy and its responsible use.According to [10,11], AI literacy is composed of different competencies, enabling individuals to critically evaluate the use of those kinds of technologies, to communicate and collaborate with AI, and to use it in different contexts, its objective being to describe the skills necessary for a basic understanding of AI.
The widely reported and recognized need for AI regulation leads to new steps towards this.On 26 October 2023, the Secretary-General of the United Nations (UN), António Guterres, launched a high-level multisectoral advisory body on AI, to identify risks, challenges, and main opportunities, while more recently the Spanish presidency of the EU Council announced that the EU co-legislators, the Council and the European Parliament, had reached a provisional agreement on the world's first rules for AI, advancing the preparation of a regulation aiming to ensure that AI in use in the EU should be safe and respect European rights and values [12].
On 26 January 2024, the Council of the European Union approved the Proposal for a Regulation of the European Parliament and of the Council, establishing harmonized rules in the field of artificial intelligence (Artificial Intelligence Act) and amending certain legislative acts.Aiming to ensure a high level of protection of health, safety, and fundamental rights, including democracy, the rule of law, and environmental protection, the possible first AI Act includes sanctions for non-compliance, impact assessment on fundamental rights, provisions for testing high-risk AI systems, and rules and obligations for all general-purpose AI models, regulating the development, deployment, and use of artificial intelligence systems [13].
In Portugal, the most recent document with official recommendations for the use of AI is a guide for ethical, transparent, and responsible Artificial Intelligence in Public Administration, published in July 2022 by the Agency for Administrative Modernization [14].This document advances a structuring conceptualization for ethical, responsible, and transparent AI, identifies barriers, challenges, and dangers, and presents recommendations and a tool for risk assessment.Despite very complete content, the level of dissemination and the respective scope for an effective contribution to AI literacy in Portuguese public administration are unknown.
Although research into AIED has at its heart the desire to support student learning, experience from other areas of AI suggests that this ethical intention is not, in itself, sufficient [15][16][17][18].Complementing this, education research shows that factors such as teacher-student interaction, educational programs, teachers' attitudes, and their decisions made in the classroom are related to the ethical dimension [18].
So, for the ethical dimension of AI, there is a need to consider issues such as equity, responsibility, transparency, partiality, autonomy, and inclusion, and also distinguish between "doing ethical things" and "doing things ethically", in order to understand and make pedagogical choices that are ethical and take into account the ever-present possibility of unintended consequences [18][19][20][21].In academia, where the citizens of the future are prepared, people will tend to use these AI tools constantly, irresponsibly, and without ethical principles in their studies [22].As an example of this, it was recently detected that around 200 scientific studies had been written with ChatGPT and accepted by scientific reviews (https://www.dailymail.co.uk/sciencetech/article-13211523/ChatGPT-scandal-AI-generated-scientific-papers.html,accessed on 28 March 2024).
According to [23], the generalized use of AIED can potentially harm teacher-student interaction and compromise the independent and capable student's development.This threat may be amplified by marketing efforts to make the public believe in neutral and objective AI algorithms.This reveals two dimensions of the real ethical problem of AI: the ethical user-system relation in AI systems and ethical use of AI systems by users.Much work needs to be done in this area and, in this context, it is recognized by [4] that most AIED researchers do not have the training to deal with emerging ethical issues.
Indeed, Ref. [24] suggests some principles for ethical and reliable AIED that should be considered, namely (i) Governance and management principle: AIED governance and management must take into account interdisciplinary and multi-stakeholder perspectives, as well as all ethical considerations from relevant domains, including, among others, data ethics, learning analytics ethics, computational ethics, human rights and inclusion; (ii) Principle of transparency of data and algorithms: The process of collecting, analyzing, and communicating data must be transparent, with informed consent and clarity about data ownership, accessibility, and the objectives of its use; (iii) Accountability principle: AIED regulation must explicitly address recognition and responsibility for the actions of each stakeholder involved in the design and use of systems, including the possibility of auditing, the minimization and communication of negative side effects, trade-offs, and compensation; (iv) Principle of sustainability and proportionality: AIED must be designed, developed, and used in a way that does not disrupt the environment, the global economy, and society, namely the labor market, culture, and politics; (v) Privacy principle: AIED must guarantee the user's informed consent and maintain the confidentiality of user information, both when they provide information and when the system collects information about them; (vi) Security principle: AIED must be designed and implemented to ensure that the solution is robust enough to effectively safeguard and protect data against cybercrime, data breaches, and corruption threats, ensuring the privacy and security of sensitive information; (vii) Safety principle: AIED systems must be designed, developed, and implemented according to a risk management approach, in order to protect users from unintentional and unexpected harm and reduce the number of serious situations; (viii) Principle of inclusion in accessibility: The design, development, and implementation of AIED must take into account infrastructure, equipment, skills, and social acceptance, allowing equitable access and use of AIED; (ix) Human-centered AIED principle: The aim of AIED should be to complement and enhance human cognitive, social, and cultural capabilities, while preserving meaningful opportunities for freedom of choice and ensuring human control over AI-based work processes.
In turn, ref. [25] states that definitions of AI literacy differ in terms of the exact number and configuration of skills it entails, and referring to [26], indicates that an analysis of conceptualizations of AI literacy in education can be organized into four concepts: (1) knowing and understanding AI, (2) using and applying AI, (3) evaluating and creating AI, and (4) AI ethics.For that author, the vast majority of conceptualizations of AI literacy are parallel to Bloom's taxonomy in terms of its general configuration of skills.Considering that this taxonomy constitutes the basis of countless formulations of competences in schools and universities, this is of enormous importance and correlation with AIED.
It is still difficult to measure AI literacy.Four published scales are currently used to carry out this measurement, three of which are not school-focused, but can be used for more general measurement purposes.As they are not based on established theoretical models of competences, it makes the interpretation of the latent factors of these scales seem arbitrary [25].In fact, Carolus et al. [25] developed a new measuring instrument based on the existing literature on AI literacy, which is modular, meets psychometric requirements, and includes other psychological skills in addition to the classical ones of AI literacy.
Although it is not objectively clear how the development of AI can be applied to education systems, enthusiasm is growing, with excessive optimism regarding the potential to transform current education systems [27].Ref. [4] sought to identify potential aspects of threat, excitement, and promise of AIED and highlighted the importance of traditional pedagogical values, such as skepticism, continuing to argue that the ultimate goal of education should be to promote responsible citizens and healthy educated minds.Therefore, the adoption of ethical frameworks for the use and development of AIED is extremely important, ensuring that it will be continually discussed and updated in light of the rapid development of AI techniques and their potential for widespread application [28].
At the same time, a set of questions must be carefully considered and comprehensively addressed as soon as possible: "What will be the future role of the teacher, and other school personnel, in education with AI systems?And how does this align with our beliefs or pedagogical theories?Do educational leaders and teachers have enough knowledge in the field of AI to distinguish a poorly developed system from a good one?Or how to apply them appropriately in the education context?Furthermore, how can we protect student and teacher data when the skills and knowledge to develop AIED systems are in the hands of for-profit organizations and not in the education sector?".In particular, the issue of aligning AI with pedagogical theory must remain on the table, as any new technology integrated into education must be designed to fill a pedagogical need [4].
Although the use of questionnaires to assess literacy in AI is still limited, mainly because there are not many validated questionnaires yet, it is possible to find some evidence on studies in AI literacy in higher education in the literature.For example, ref. [29] concludes that the use of tools associated with artificial intelligence, in an exploratory learning environment context, can benefit teaching itself.Ref. [30] relates the knowledge of AI by teachers with data literacy, proposing an approach to reflect those data literacy competencies to use AI.Based on a literature review and through a survey applied in Serbia, in this case among students, ref. [31] concludes that AI, together with machine learning, has the potential to improve the learning levels of the student population.In a recent study, ref. [32] analyzes the adoption of artificial intelligence in higher education practices, relating it to literacy levels and their opinion on the conditions under which the use of AI tools is defensible, identifying clear concerns with justice and responsibility, as well as the lack of knowledge about the phenomenon of AI.In fact, despite the increase in knowledge about AI applied to education, it is still a challenge, taking into account the current context [33].
Considering the relevance of understanding the use of artificial intelligence in education, in particular in the higher education context, the present study seeks to assess the level of AI literacy and knowledge among lecturers at Portalegre Polytechnic University (PPU), aiming to identify gaps and find the main opportunities for innovation and development so that the education system can adopt AIED as an ally in promoting higher-quality education better prepared for the challenges of the future.As specific objectives, we seek to assess the degree of relationship between the dimensions of AI literacy and identify the predictive factors.
The remainder of this paper is organized as follows: Section 2 presents the materials and methods, in particular, the questionnaire; Section 3 presents the results; Section 4 discusses the results and concludes the analysis.

Materials and Methods
Despite the high number of studies produced on AI literacy to date, its measurement is still complex.The difficulties of conceptualization and the fact that many articles on the subject originate in an educational context limit the development of measurement scales and therefore their adoption in different contexts.
Ref. [25] developed a measuring instrument that builds on the existing literature on AI literacy.The questionnaire presented is modular (including distinct facets that can be used independently), is easily applicable to professional life, meets psychometric requirements, and includes other psychological skills besides the classic facets of AI literacy, having been tested for its factorial structure.Therefore, the questionnaire by [25] was applied in this study, adapted to the Portuguese language.It consists of 29 questions, based on three dimensions-AI Literacy, AI Self-Efficacy, and AI Self-Management-measured using a 5-point Likert scale (from 1 = "totally disagree" to 5 = "completely agree").A brief note about the questionnaire used in this paper is the following: we decided to keep the original names of the dimensions proposed by [25], although the first dimension (AI Literacy) may be confused with the final result, which, in fact, is named by the authors as a scale.
In the first dimension (AI Literacy), using and applying AI, according to [34], means applying knowledge, concepts, and applications of AI in different scenarios and implies understanding the applications of AI and how it can affect one's life.In turn, the knowing and understanding AI factor means knowing the basic functions of AI and knowing how to use its applications, covering the acquisition of fundamental concepts, skills, knowledge, and attitudes that do not require prior knowledge, as well as understanding the technologies underlying techniques and basic concepts underlying AI in different products and services.The AI ethics factor, advanced by the same author, means human-centered considerations (for example, equity, responsibility, transparency, ethics, and security), therefore incorporating knowledge of ethical issues relating to AI technologies.Still in this dimension, but by the authors of [10,35], it means distinguishing between technological equipment that uses and does not use AI.
The second dimension (AI Self-Efficacy) integrates the Problem Resolution factor.According to [35], this means voluntary behavior aimed at solving problems, based on belief in the advantages of behavioral success, external approval, and the level of control of internal and external factors.The learning factor, according to [36], means understanding how AI learns and can be affected by data, that is, having a basic understanding of how AI and machine learning work, as well as knowledge of the implications of data quality, feedback, and one's own data of interaction.Still on this factor, ref. [37] integrates skills that allow the development of adaptive knowledge to make self-learning and technological evolution profitable, with [38] including the level of availability for AI.
The third and final dimension (AI Self-Management) integrates the AI persuasion literacy factor, which, according to [36], means understanding how the human-like characteristics of AI systems can unconsciously manipulate users' perceptions and behaviors and thus thwart attempts to influence them.According to the same author, the Emotion Regulation factor means the constructive management of negative emotions (such as frustration and anxiety) when interacting with AI systems.
The present sample was made up of 75 teachers from the various schools of PPU, from a total of 225, corresponding to one third of the population.The whole set of teachers includes a relevant number of invited professionals, who are normally less likely to answer this kind of questionnaire, resulting in the 75 completed answers.The questionnaire was presented online, and all teachers were asked to answer it, through the institutional email.Participation was completely voluntary and anonymous.The questionnaire was preceded by an explanation of the objectives and respondents' consent was explicitly obtained, following PPU procedures.
The respondents are aged from 25 to over 50, with 70.7% being 45 or over.A total of 41 participants were female (54.7%), 33 participants were male (44.0%), and 1 chose not to specify (1.3%).A major proportion of participants came from the School of Technology, Management and Design (40.0%), followed by the School of Health (29.3%), the School of Education and Sciences (22.7%), and the School of Biosciences of Elvas (8.0%).It is noteworthy that most participants teach in more than one study cycle, with 25.3% teaching in higher technical courses and bachelor's degrees, 22.7% teaching bachelor's and master's degrees, and 21.3% teaching only bachelor's degrees, while 20.0% teach in the three study cycles (higher technical courses and bachelor's and master's degree).The main areas of basic training for participants are health (29.3%), social and behavioral sciences (13.3%), and business sciences (12.0%).
The instrument used in the present study designed by [25] covers the three dimensions based on the existing literature (AI Literacy, AI Self-Efficacy, and AI Self-Management), each containing more than one descriptor factor.The whole questionnaire can be consulted in Table A1.

Results
The results of the questionnaire, presented in Table 1, reveal an average level of AI literacy (3.28), highlighting that 62.4% of responses are at levels 3 and 4. The AI Literacy dimension recorded the highest average response (3.56), highlighting that the factor of using and applying AI had the highest average response (3.85), followed by the ethics of AI factor with an average of 3.73.Still in this dimension, the knowing and understanding AI factor had an average response of 3.46 and detecting AI was at 3.20 (see Figure 1).In fact, most respondents reveal the capacity to use and apply AI as well as saying they are able to act in an ethical way, regarding AI.
In turn, the AI Self-Efficacy dimension obtained the lowest average response (2.86), highlighting that learning was the factor with the lowest average response (2.49).Specifically in this factor, 65.4% of participants responded with levels 2 and 3, reflecting that they have more difficulty in handling problems and challenges related to AI (see Figure 2).
Finally, in the AI Self-Management dimension, which had an average response of 3.41, the AI persuasion literacy factor had an average response of 3.27 and Emotion Regulation had an average of 3.55, revealing respondents' greater perception of the possibility of controlling their emotions regarding the use of AI than in considering the influence of AI in their daily life (see Figure 3).In turn, the AI Self-Efficacy dimension obtained the lowest average response (2.86), highlighting that learning was the factor with the lowest average response (2.49).Specifically in this factor, 65.4% of participants responded with levels 2 and 3, reflecting that they have more difficulty in handling problems and challenges related to AI (see Figure 2).Finally, in the AI Self-Management dimension, which had an average response of 3.41, the AI persuasion literacy factor had an average response of 3.27 and Emotion Regulation had an average of 3.55, revealing respondents' greater perception of the possibility of controlling their emotions regarding the use of AI than in considering the influence of AI in their daily life (see Figure 3).In turn, the AI Self-Efficacy dimension obtained the lowest average response (2.86), highlighting that learning was the factor with the lowest average response (2.49).Specifically in this factor, 65.4% of participants responded with levels 2 and 3, reflecting that they have more difficulty in handling problems and challenges related to AI (see Figure 2).Finally, in the AI Self-Management dimension, which had an average response of 3.41, the AI persuasion literacy factor had an average response of 3.27 and Emotion Regulation had an average of 3.55, revealing respondents' greater perception of the possibility of controlling their emotions regarding the use of AI than in considering the influence of AI in their daily life (see Figure 3).A more detailed analysis of the means and standard deviation of each question can be seen in Appendix B (Table A2), with the respective mean values, ranging from 2.40 ("Despite the rapid changes in the field of artificial intelligence, I can always keep up to date") to 4.25 ("I can operate AI applications in everyday life.").
We continue our analysis by calculating the internal consistency of the instrument, obtaining a value of 0.930 for Cronbach's Alpha.The correlation coefficients between each question and the total suggest good internal validity indices that exceed the critical index (<0.20),including the items with the lowest value, highlighting that the vast majority of items have a correlation greater than 0.35, with some items reaching an index above 0.5 (see Table 2).It is noteworthy that in the field of AI Self-Management, two factors related to AI persuasion literacy have a correlation value lower than 0.3.Even so, it was decided to keep them as their elimination would not make significant improvements to the result of the instrument and because, above all, the aim is to maintain the theoretical framework chosen for the objective of this study, that is, to assess the sample's level of AI literacy.A more detailed analysis of the means and standard deviation of each question can be seen in Appendix B (Table A2), with the respective mean values, ranging from 2.40 ("Despite the rapid changes in the field of artificial intelligence, I can always keep up to date") to 4.25 ("I can operate AI applications in everyday life.").
We continue our analysis by calculating the internal consistency of the instrument, obtaining a value of 0.930 for Cronbach's Alpha.The correlation coefficients between each question and the total suggest good internal validity indices that exceed the critical index (<0.20),including the items with the lowest value, highlighting that the vast majority of items have a correlation greater than 0.35, with some items reaching an index above 0.5 (see Table 2).It is noteworthy that in the field of AI Self-Management, two factors related to AI persuasion literacy have a correlation value lower than 0.3.Even so, it was decided to keep them as their elimination would not make significant improvements to the result of the instrument and because, above all, the aim is to maintain the theoretical framework chosen for the objective of this study, that is, to assess the sample's level of AI literacy.To study the construct validity, the principal component factor analysis (PCFA) method with varimax rotation was chosen.Following the procedure, a Keyser-Meyer-Olkin measure of 0.812 was obtained after rotation, which reflects a reasonable variance of the factors [39].Bartlett's test of sphericity is associated with a chi-square of 1783.012.Factor extraction followed the method advocated by [40], which consists of reading the scree plot graph.Results of the component factor analysis are presented in Table 3.After the analysis, reading of the scree plot graph suggested the existence of three factors.So, considering three factors, the results were relatively aligned with the reference, with those factors explaining 58.6% of the variance found (the first factor explained 36.09%, the second 16.20%, and the third 6.31%).
In terms of correlations between dimensions, as presented in Table 4, there was a very high level of correlation resulting from the three-factor factor analysis, with correlation values from 0.501 to 0.766 between dimensions and values from 0.636 to 0.949 between the total dimensions and each of them.The results demonstrate that the first dimension is highly significant for the total dimensions, i.e., for AI Literacy in the sample.

Discussion and Conclusions
The objective of this study was to assess the AI literacy of PPU teachers, in order to identify gaps and find the main opportunities for innovation and development, specifically seeking to assess the degree of relationship between the dimensions of AI literacy and identify what could be the predictive variables in this matter.
The results of the questionnaire revealed an overall average level of AI literacy in the sample, meaning it will be desirable to implement strategies to develop faculty skills in AI matters given the growing impact of AIED on the education community and the education system itself.A higher level of AI literacy will allow us to find and implement better solutions to add value to the teaching-learning process through AI technologies and simultaneously support teachers and students, consistent, for example, with the results of [30].
The correlation between the three dimensions studied allows us to conclude that the AI Literacy dimension is the biggest predictor of the level of AI literacy score in the sample, which, integrating the factors of use and application, knowledge and understanding, detection, and ethics, suggests specific development of these skills to increase the participants' overall level of literacy.However, the below-average result of the learning factor, incorporating understanding of the functioning of AI technologies, the development of adaptive knowledge, and the level of availability for AI, indicates the pressing need to focus on developing these skills through awareness-raising policies and targeted training actions.
From the study of correlations, it was concluded that no factor characterizing the sample (age, gender, study cycle taught, or area of training) is a predictor.Therefore, these factors do not explain the sample's level of literacy, nor do they determine or limit it.
Highlighted is the higher level of the application of knowledge, concepts, and applications of AI in different scenarios and awareness of ethical issues relating to AI technologies such as equity, responsibility, transparency, ethics, and safety of teachers, given the recent emergence of public and widespread use of AI applications by the education community (like, for example, [32]).
The results presented in this work are relevant to the reality studied, allowing specific measures to be taken to increase the level of AI literacy in its various components, not only in terms of knowledge of the different possible tools for teaching and their possible integration into the teaching-learning process, but also in terms of how to deal with challenges related to ethical concerns.A more in-depth analysis of the tools that may be used by both students and teachers may also be relevant, promoting the possible involvement of the academic community in the joint analysis of these issues.Also associated with the level of AI literacy, creating initiatives that can demystify the use of AI in higher education, in particular, by demonstrating ways in which it can be useful in the daily lives of the academic community, could help to encourage the fair use of this type of tool.
The sample size preventing generalizations and the fact that it is not possible to make comparisons of the application of the same measuring instrument are the main limitations identified.This leads to suggesting future work applying the instrument to the student body at PPU and expanding similar studies to Portuguese polytechnic higher education, looking for possible predictors in a broader educational community and identifying intervention priorities to increase AI literacy in academia in Portugal.The fact of applying this study to a single higher education institution is another limitation.In the future, these results could be complemented with other assessments not only in other institutions, but also in other professional fields, including technical and non-technical professions, and in other areas where AI could be used.A final note is that this is one of the first applications of the questionnaire to higher education, making it difficult to make comparisons.This should also be considered in future work, even comparing different kinds of higher education institutions.

Figure 1 .
Figure 1.Results of the AI Literacy dimension.

Figure 2 .
Figure 2. Results of the AI Self-Efficiency dimension.

Figure 1 .
Figure 1.Results of the AI Literacy dimension.

Figure 1 .
Figure 1.Results of the AI Literacy dimension.

Figure 2 .
Figure 2. Results of the AI Self-Efficiency dimension.

Figure 2 .
Figure 2. Results of the AI Self-Efficiency dimension.

Information 2024 , 14 Figure 3 .
Figure 3. Results of the AI Self-Management dimension.

Figure 3 .
Figure 3. Results of the AI Self-Management dimension.

Table 1 .
Results of the AI Literacy questionnaire.

Table 3 .
Component factor analysis with varimax rotation method.

Table A1 .
23 everyday life, I can interact with AI in a way that makes my tasks easier.5Ineverydaylife,Icanworktogethergainfully with artificial intelligence.6Icancommunicategainfullywithartificial intelligence in everyday life.ifIamdealingwithanapplication based on artificial intelligence.13Icandistinguishdevicesthatuse AI from devices that do not.14Icandistinguishif I interact with AI or a "real human".upwith the latest innovations in AI applications.22Despitetherapidchanges in the field of artificial intelligence, I can always keep up to date.23Althoughthere are often new AI applications, I manage to always be "up to date".Cont.

Table A2 .
Mean and standard deviation for each question.