1. Introduction
AI has spread widely, causing a deep transformation in different fields, including industry, healthcare, energy, environment, education, finance, transportation, judicial settings, and logistics (
Brynjolfsson & McAfee, 2014;
Chui et al., 2018;
Leal et al., 2023). For example, in the healthcare field, AI diagnoses diseases more accurately and prescribes personalized treatments, improving the quality of healthcare (
Haenlein & Kaplan, 2021;
Johnson et al., 2021). In the business sector, it optimizes strategic business decision-making by analyzing immense amounts of data and extracting precise patterns, improving planning and management (
Dwivedi et al., 2021). It is also effective in education, where it developspersonalized learning systems that takeinto account students’ needs, thereby reducing educational gaps and improving teaching quality (
Zhai et al., 2022). In the legal field, it can contribute to improving judicial processes and decision-making, and in industry, it facilitates the intelligent automation of processes, especially in the field of mechanics (
Valavanidis, 2023). There have even been widespread advances in revolutionary AI applications related to people’s daily lives, such as personal assistants (e.g., Alexa, Siri;
Brill et al., 2019), self-driving cars (
Hong et al., 2021;
Manoharan, 2019), or social robots (
Glas et al., 2016;
Hentout et al., 2019).
1.1. Attitudes Towards AI
Despite being hailed as one of the world’s most important innovations (
Valavanidis, 2023), artificial intelligence (AI) often elicits ambivalent public attitudes, particularly regarding issues like privacy, unemployment, and social justice (
Zhu et al., 2025). In Spain, for example, the 2024 National Survey on the Social Perception of Science and Technology (
Fundación Española para la Ciencia y la Tecnología, 2025) revealed widespread enthusiasm for AI. The survey indicated high usage rates (over 80%) and a clear recognition of AI’s benefits, with improved process efficiency being the most cited advantage (49.8%). However, this enthusiasm is coupled with significant public distrust. Many citizens admit to not understanding how AI works and express concerns about personal data security (42.4%), a false sense of security in AI systems (32.6%), and increased control by governments (28.9%) and companies (16.7%).
These concerns are echoed across various sectors. In the public sphere, AI raises fears of increased government surveillance and reduced personal freedoms (
Sobrino-García, 2021). In the workplace, it may drive structural changes that create new opportunities while threatening traditional jobs (
Méndez et al., 2008). In education, AI can enhance personalized learning but may also diminish human interaction (
Galindo-Domínguez et al., 2023). Similarly, in the justice system, it could speed up legal decision-making but also risks undermining human legal analysis and the integrity of judicial outcomes (
Lozano et al., 2021). Furthermore, AI is not immune to perpetuating gender biases, making it essential to identify and address these disparities (
Grassini & Ree, 2023;
Arees et al., 2024).
Public attitudes toward AI will not only affect its immediate acceptance and implementation but will also play a crucial role in shaping its future development (
Zhu et al., 2025). Understanding these attitudes helps identify fears and uncertainties, enabling organizations to address these concerns and foster greater technology adoption (
Kim & Lee, 2024;
World Economic Forum, 2024). Moreover, studying public attitudes is key to guiding AI’s development responsibly and informing public training and digital literacy initiatives in response to AI’s rapid advancement (
Montag & Ali, 2024). Ultimately, this process reveals critical ethical concerns, which areessential for ensuring that AI development remains aligned with human values (
Mittelstadt, 2019;
United Nations System Chief Executives Board for Coordination, 2022;
Uzougbo et al., 2024).
1.2. AI Attitude Measurement
Models of acceptance and adoption of new technologies have been fundamental to understanding how individuals interact with emerging tools in different contexts, including artificial intelligence (e.g.,
Ali et al., 2024). Among the most influential are the Technology Acceptance Model (TAM;
Davis, 1989;
Davis et al., 1989), which emphasizes perceived usefulness and perceived ease of use, and the Unified Theory of Acceptance and Use of Technology (UTAUT;
Venkatesh et al., 2003), which extends TAM by incorporating factors such as social influence and facilitating conditions.
These frameworks make it possible to identify the key determinants of the intention to use and attitude toward technologies, serving as the foundation for developing specific instruments that measure attitudes and perceptions of AI, as well as for interpreting how psychological and contextual variables influence its acceptance and effective use.
Several instruments have been developed to assess public attitudes toward AI. Among the most widely cited is the General Attitudes towards Artificial Intelligence Scale (GAAIS;
Schepman & Rodway, 2020), which has been validated across diverse cultural contexts (
Şahin & Yıldırım, 2024). While shorter scales have emerged recently—such as the AI Attitude Scale (AIAS-4;
Grassini, 2023), Attitudes Towards AI (ATAI;
Sindermann et al., 2021), and the Threats of Artificial Intelligence (TAI) scale (
Kieslich et al., 2021)—these often focus on specific dimensions like fear. In contrast, the GAAIS is well-suited for capturing the full complexity of individual attitudes, addressing the limitations of shorter scales that may not fully represent these views or may overemphasize negative perceptions (
Zhu et al., 2025).
In their initial validation study,
Schepman and Rodway (
2020) used Exploratory Factor Analysis (EFA) to identify two core dimensions of attitudes toward AI: a positive dimension reflecting its social and personal utility, and a negative dimension capturing common concerns. The scale demonstrated strong psychometric properties, including good convergent and discriminant validity against existing measures. The authors also found that comfort with technology was a strong predictor of favorable AI attitudes, while perceived capability was a weaker predictor. Additionally, they noted that people viewed AI more positively for data-intensive applications (e.g., astronomy, cybersecurity) but more negatively in domains requiring human judgment (e.g., psychological counseling, arts) (
Schepman & Rodway, 2020).
In a subsequent study, the same authors performed a Confirmatory Factor Analysis (CFA), re-confirming the two-factor structure of the GAAIS (
Schepman & Rodway, 2023). They also explored its relationship with the Big Five personality traits, finding that higher levels of introversion, conscientiousness, agreeableness, and general trust were all associated with more positive or tolerant attitudes toward AI.
1.3. The Present Study
Despite the exponential growth of artificial intelligence and the cultural factors that shape its social perception, there are no instruments in the Spanish context capable of measuring public attitudes toward AI while accounting for their ambivalent nature. Therefore, this study aims to adapt and validate the Spanish version of the GAAIS, analyzing its psychometric properties to offer a reliable tool that supports empirical research and future investigations.
2. Materials and Methods
2.1. Participants
The selection of participants was carried out using a convenience approach, and the sample was expanded using the snowball method. The questionnaire was administered in digital format via an online web form. The invitation to participate was distributed through social networks, and participants were also asked to share the link with others, following the logic of chain sampling. Data were collected between February and December 2024. Participation was voluntary, and no compensation was provided. After listwise deletion of cases with missing values, the final sample consisted of 644 individuals. The sample was 50.8% male (n = 327) and 49.1% female (n = 316), with one participant (0.2%) selecting “Other.” Ages ranged from 18 to 78 years (M = 33.06, SD = 14.91). Regarding educational attainment, 5.6% (n = 36) had completed primary education, 29.2% (n = 188) had completed secondary education, and 65.1% (n = 419) had a university-level education.
Furthermore, participants were asked whether they worked or studied in a field related to artificial intelligence, new technologies, computing, or big data. The results showed that 7.3% of individuals (n = 47) had a direct link to the field of artificial intelligence, 16.6% (n = 107) indicated that their current occupation was related to new technologies, 21.1% (n = 136) indicated a link to the field of computing, and 6.7% (n = 43) stated that their work activity was related to big data.
2.2. Procedure and Data Analysis
The 20 items of the General Attitudes Towards Artificial Intelligence Scale (GAAIS;
Schepman & Rodway, 2023) were translated from English to Spanish. A committee of experts then reviewed the translated items to ensure semantic and cultural equivalence, followed by a back-translation to verify fidelity to the original scale.
All statistical analyses were performed in R (RStudio IDE, Version 2024.12.1). First, an exploratory data analysis was performed on all study variables. This included calculating basic descriptive statistics (see
Appendix B) and applying relevant normality tests to examine the distributions and identify potential violations of assumptions. Items on the GAAIS and the Attitude towards Artificial Intelligence (ATAI) scale—selected to assess convergent and discriminant validity—were reverse-coded where necessary. Subsequently, Spearman’s rho (ρ) correlations were calculated for all item pairings between the two scales.
To validate the two-factor structure proposed by
Schepman and Rodway (
2023) through Confirmatory Factor Analysis (CFA), different structural equation models were tested using the lavaan package (
Rosseel, 2012). This was carried out based on a polychoric correlation matrix (
Bandalos & Finney, 2018), using the Weighted Least Squares Mean and Variance Adjusted (WLSMV) estimator, as it is a robust estimator, especially suitable for ordinal data that does not necessarily fit a multivariate normal distribution. Model fit was assessed using the chi-square and degree of freedom ratio (χ
2/df), the comparative fit indices Comparative Fit Index (CFI;
Bentler, 1990) and the Tucker–Lewis Index (TLI;
Tucker & Lewis, 1973), and the absolute fit indices, taking into account the root mean square error of approximation (RMSEA;
Steiger, 1990) and the standardized root mean square residual (SRMR;
Bentler, 1995). For the interpretation of results, values equal to or less than three were considered adequate for the chi-square ratio per degree of freedom (χ
2/df ≤ 3). Regarding the comparative fit indices (CFI and TLI), values greater than 0.90 were considered indicative of a good fit (
Hu & Bentler, 1999). For the absolute fit indices, RMSEA values equal to or less than 0.05 were used as a reference to describe a good model fit, and values between 0.05 and 0.08 were used to indicate an adequate fit (
Browne & Cudeck, 1993). For the SRMR, a threshold of 0.08 was established to distinguish models characterized by a good fit to the data from models with a poorer fit, as this index is sensitive to model misspecification (
Hu & Bentler, 1999).
Scale reliability was evaluated using Cronbach’s α (
Cronbach, 1951), ordinal α (
Zumbo et al., 2007), and McDonald’s ω (
McDonald, 1999) coefficients. Convergent and discriminant validity were assessed by calculating the Average Variance Extracted (AVE), applying the Fornell–Larcker criterion (1981), and computing the Heterotrait–Monotrait Ratio (HTMT;
Henseler et al., 2015). Additionally, Spearman’s rho correlations between the GAAIS and ATAI dimensions were analyzed using a permutation test to determine if correlations between theoretically related dimensions were significantly stronger than those between less related dimensions.
Finally, a cross-validation procedure was implemented to assess the stability of the factor solution across different subsamples. A factorial invariance analysis was also performed between groups by sex to examine whether the factor structure remained constant across the four levels of invariance: configural, metric, scalar, and strict. Models were estimated using multigroup confirmatory factor analysis with the WLSMV estimator and theta parameterization, which allows explicit estimation of residual variances required for testing strict invariance. When comparing the different invariance models, changes in the fit indices were taken as evaluation criteria. Specifically, differences in CFI (ΔCFI ≤ 0.01), RMSEA (ΔRMSEA ≤ 0.015), and SRMR (ΔSRMR ≤ 0.01 for metric invariance and ΔSRMR ≤ 0.03 for scalar invariance) were considered, following the recommendations of
Chen (
2007).
2.3. Measures
The GAAIS contains 20 items divided into two dimensions: Positive GAAIS (12 items) and Negative GAAIS (8 items). The first dimension reflects perceived opportunities, benefits, and positive emotions related to AI, while the negative dimension focuses on concerns and negative emotions. Additionally, the authors include an attention control element to verify whether the questions had been read, to exclude participants who may have responded randomly. Each item is rated on a five-point Likert scale ranging from 1 (“strongly disagree”) to 5 (“strongly agree”). Items belonging to the Negative GAAIS are reverse-scored so that, in both subscales, higher scores represent more positive attitudes toward artificial intelligence. The mean score of each subscale can be used as an indicator of the respective attitudinal dimension, and the authors do not recommend computing a single total score for the entire scale. Thus, higher values in the Positive GAAIS indicate greater enthusiasm and perceived utility of AI (e.g., “Artificial intelligence can provide new economic opportunities for this country”), whereas lower scores in the Negative GAAIS reflect stronger concerns and distrust toward it (e.g., “Artificial intelligence is used to spy on people”).
Similarly, the Attitude towards Artificial Intelligence (ATAI) scale (
Sindermann et al., 2021) was used to assess convergent validity. It consists of five items across two dimensions: Acceptance and Fear. For both the GAAIS and ATAI, responses were collected on a five-point Likert-type scale ranging from 1 (“Strongly Disagree”) to 5 (“Strongly Agree”).
3. Results
3.1. Descriptive Results
Spearman’s rho correlation matrix among the items showed values ranging from −0.119 to 0.598, with several non-significant associations. These results suggest considerable heterogeneity among some items, but do not indicate significant multicollinearity. Items 3 and 5 had the highest number of non-significant correlations. Items within the Positive GAAIS (PG) dimension tend to be more closely related to each other, with higher correlations within the dimension. This pattern also held true for the Negative GAAIS (NG) dimension. In contrast, the correlations obtained between items from the Positive GAAIS (PG) and the Negative GAAIS (NG) dimensions were notably lower.
3.2. Validity
A two-factor correlated model based on the original scale structure was evaluated, which showed unsatisfactory fit indices: χ2 = 848.234, df = 169, p < 0.001, CFI = 0.851, TLI = 0.832, RMSEA = 0.082, and SRMR = 0.069. Based on the analysis of the item factor loadings and modification indices, an adjusted model was proposed. In this second version, four items (i3, i5, i6, and i13) were eliminated due to their low factor loadings (λ < 0.40) and their limited contribution to the construct. Similarly, the high residual variances associated with these items were a factor in their removal, reflecting a significant degree of error not explained by the model and compromising its psychometric adequacy. Furthermore, acorrelation between the errors of item pairs i4-i11, i7-i18, and i9-i10 was allowed due to their conceptual similarities.
In this second scenario, the results of the chi-square test (χ
2 = 388.180, df = 100,
p < 0.001) indicated a better model fit compared to the initial model and the base model, which assumes independence between variables. Although the result was significant (
p < 0.001), it is important to note that this statistic is especially sensitive to sample size and can exaggerate significance in large samples, as is the case here, so it should be interpreted with caution. However, the lower value obtained in the factorial model compared to the base model suggests a considerably better fit. The χ
2/dfratio (388.180/100) was 3.88, slightly above the conventional cutoff for a good fit (χ
2/df ≤ 3), but within the range often regarded as acceptable when other fit indices are satisfactory. Regarding the absolute fit indices (see
Table 1), the Root Mean Square Error of Approximation (RMSEA) showed a value of 0.069, with a 95% confidence interval between 0.059 and 0.078, and the standardized root mean square residual (SRMR) was 0.051. The RMSEA value can be considered adequate, while the SRMR indicates a good model fit. Similarly, regarding the comparative fit indices, the CFI obtained a value of 0.926, and the TLI a value of 0.911. Both values confirm the good fit of the model. As part of the analysis, a unifactorial model was also specified to assess its fit compared to the previous model. As shown in
Table 1, this model had the poorest fit of the three (χ
2 = 2250.817, df = 170,
p < 0.001, CFI = 0.630, TLI = 0.586, RMSEA = 0.138, SRMR = 0.117).
The final model (see
Figure 1) consisted of 16 items, 10 for the Positive GAAIS subscale and 6 for the Negative GAAIS subscale. All factor loadings were statistically significant, with values ranging from 0.485 to 0.786 (all λs > 0.4).
3.3. Reliability
Regarding the scale’s reliability, Cronbach’s α was 0.854, supporting an adequate internal consistency. As for the subscales, Cronbach’s α was 0.840 for the Positive GAAIS and 0.808 for the Negative GAAIS. The ordinal alpha coefficient was 0.879 for the total scale, 0.873 for the positive subscale, and 0.838 for the negative subscale. Finally, the McDonald’s ω was 0.873 for the overall scale, 0.836 for the Positive GAAIS, and 0.804 for the Negative GAAIS. Overall, these indices indicate that both the scale and the subscales exhibit adequate and consistent reliability, supporting their use for assessing the construct (see
Table 2).
3.4. Convergent and Discriminant Validity
The model showed limited internal convergent validity (see
Table 2). The AVE values were slightly below the recommended threshold of 0.50 (Positive GAAIS = 0.413; Negative GAAIS = 0.472; Total GAAIS = 0.435). The average item coefficient of determination (R
2) was consistent with these results (0.42, 0.49, and 0.45, respectively), indicating that the latent factors explained 42% and 49% of the variance in their corresponding items.
In addition, internal discriminant validity was adequate. According to the
Fornell and Larcker (
1981) criterion, the square root of the AVE for each construct was greater than the correlations between constructs (Positive GAAIS = 0.643 > 0.480; Negative GAAIS = 0.687 > 0.480). The HTMT index (
Henseler et al., 2015) was 0.472, below the critical threshold of 0.85, indicating adequate discrimination between the two factors.
Furthermore, Spearman’s rho correlation (ρ) was calculated between the GAAIS subscales (Positive and Negative) and the dimensions of the Attitude Towards Artificial Intelligence (ATAI;
Sindermann et al., 2021), Acceptance, and Fear. Overall, there was a strong correlation between the two instruments (ρ = 0.766,
p < 0.001). This was also true for the theoretically related sub-dimensions. Thus, Positive GAAIS and Acceptance (ATAI) showed a correlation of 0.657 (
p < 0.001), and Negative GAAIS and Fear (ATAI) had a correlation of 0.800 (
p < 0.001). Regarding discriminant validity, the lowest correlations in the matrix correspond to the relationship between the Positive GAAIS subscale and Fear (ATAI), as well as between the Negative GAAIS subscale and Acceptance (ATAI). In the first case, a value of ρ = 0.368,
p < 0.001, was obtained, and in the second, a correlation of ρ = 0.460 (
p < 0.001) was obtained. Furthermore, statistically significant differences were observed (see
Table 3) between the correlations of the theoretically related dimensions (PG ↔ Acceptance, NG ↔ Fear) and the correlations of the less related dimensions (PG ↔ Fear, NG ↔ Acceptance).
3.5. Evaluation of Factorial Solution Stability
To assess the consistency of the scale’s factor structure, the total sample (
N = 644) was randomly divided into two subsamples (using an odd-even split for cross-validation;
n1 = 322,
n2 = 322), and a Confirmatory Factor Analysis (CFA) was performed on each half. The structure identified above was replicated in both subsamples, yielding fit indices that were consistent with the full sample and the subsamples, albeit with slight variations (see
Table 4). With the reduced sample size, the χ
2/dfratio values decreased considerably compared to the initial analysis. The absolute and comparative fit indices remained within acceptable ranges in all cases. Although there were slight variations, the results do not seriously compromise the model’s stability. Factor loadings were similar between the two analyses, with standardized loadings ranging from 0.38 to 0.80. These results suggest that the scale’s factor structure is consistent and generalizable to different subsamples, reinforcing its structural validity.
Finally, the model’s invariance across sexes was tested. Adequate fit indices were obtained for the initial measurement models for men (χ
2 = 295.215, df = 100,
p < 0.001, CFI = 0.900, RMSEA = 0.077, SRMR = 0.071) and women (χ
2 = 227.463, df = 100,
p < 0.001, CFI = 0.927, RMSEA = 0.072, SRMR = 0.056), prior to testing measurement invariance. As shown in
Table 5, the results support full factorial invariance. The fit indices are adequate at all levels of analysis, and the differences observed between them do not exceed the cutoff points typically used to compare the different types of invariance (
Chen, 2007). Thus, the instrument measures the construct equally for men and women in terms of factor structure, loadings, intercepts, and error variances, allowing for valid comparisons between the two groups.
4. Discussion
It is essential to have valid and reliable instruments to assess public attitudes across different sociocultural contexts. To this end, this study adapted the General Attitudes towards Artificial Intelligence Scale (GAAIS;
Schepman & Rodway, 2020) into Spanish (see
Appendix A). The results from the different analyses supported the original two-factor structure (Positive GAAIS and Negative GAAIS). The final model had acceptable fit indices and adequate internal consistency. To obtain this level of fit, four items were eliminated from the original scale due to their poor psychometric performance and limited theoretical contribution. From a conceptual standpoint, the removed items addressed aspects closely related to those already represented in the retained items, suggesting that their exclusion contributes to a more streamlined model without compromising construct coverage. The final scale proposed for use in Spanish consists of 16 items divided into two subscales: Positive GAAIS (10 items) and Negative GAAIS (6 items).
The results showed limited convergent validity (AVE < 0.50). However, this could be due to the nature of the construct being measured, since attitudes involve cognitive, affective, and behavioral elements (
Rosenberg & Hovland, 1960) that are not necessarily perfectly aligned but are part of the same general construct. Thus, different facets of one’s attitude toward AI can be relatively independent or even conflicting. A good example would be individuals who believe that AI is useful but who nevertheless fear it and do not want to use it in their daily lives. Furthermore, the items that make up the scale cover a wide range of ideas applied to different contexts (social benefits, work efficiency, personal comfort with using AI), an aspect that could also contribute to the scale’s low internal convergent validity. In contrast, adequate discriminant validity was found, supporting the differentiation between both dimensions (Positive GAAIS and Negative GAAIS). Moreover, the correlations between the GAAIS and ATAI subscales showed statistically significant differences between theoretically related and less related dimensions. Overall, these results support the instrument’s validity in distinguishing between the different positive and negative aspects of artificial intelligence. The factor solution was found to be consistent across different subsamples, and full factorial invariance was confirmed across gender, indicating that the scale measures the construct equally in men and women, allowing for valid comparisons between groups without measurement bias.
The moderate correlation between the two factors is consistent with that reported in other adaptations (
Sacco et al., 2025;
Schepman & Rodway, 2020,
2023;
Zhu et al., 2025). Regarding reliability, the coefficients obtained are in line with those reported by
Şahin and Yıldırım (
2024) in their meta-analytic review, which integrated data from 19 independent studies conducted across different regions of Asia, Europe, and the Americas. The authors reported Cronbach’s alpha values of 0.881, 0.828, and 0.863 for the total, negative, and positive subscales, respectively, figures that are like those obtained in the present study (see
Table 2).
In comparison with other instruments designed to assess attitudes toward artificial intelligence, such as the ATAI (
Sindermann et al., 2021), the bifactorial structure of the GAAIS corresponds to the dimensions of acceptance and fear proposed in that questionnaire. Thus, both instruments consistently capture the coexistence of positive and negative components in attitudes toward AI. However, the GAAIS provides a broader and more balanced assessment by including both cognitive beliefs and emotional responses without overemphasizing the negative affective component, as occurs in the ATAI, where three of the five items focus on fear. Meanwhile, the AIAS (
Grassini, 2023) seeks a brief and unidimensional measure of general attitude toward artificial intelligence, which facilitates its use in large samples but limits its ability to capture the attitudinal ambivalence that characterizes this construct. Overall, the Spanish version of the GAAIS preserves the theoretical complexity and sensitivity of the original scale, offering a balance between conceptual breadth and psychometric robustness that distinguishes it from other available instruments.
Regarding the scoring and interpretation of the scale, it is recommended to follow the original instructions provided by the authors (
Schepman & Rodway, 2020). Items belonging to the Negative GAAIS are reverse-scored so that higher values in both subscales represent more positive attitudes toward artificial intelligence. The mean scores of each subscale are used as indicators of positive and negative attitudes, where higher scores on the Positive GAAIS reflect greater enthusiasm and perceived utility, whereas lower scores on the Negative GAAIS indicate stronger concerns and distrust.
4.1. Implications
The rapid incorporation of artificial intelligence into various areas of social and professional life poses new challenges for psychological research. In this context, having valid and reliable instruments is essential to understand how people perceive, evaluate, and accept these technologies, as well as to guide their responsible and equitable implementation.
Having psychometrically sound instruments is crucial not only to better understand social perceptions surrounding artificial intelligence but also to inform the design of public policies and intervention strategies that promote the ethical, equitable, and socially responsible use of these emerging technologies (
Lozano et al., 2021;
Stein et al., 2024;
Valavanidis, 2023;
World Economic Forum, 2024). From a practical perspective, the validation of the GAAIS provides a reliable tool to assess attitudes toward AI, with potential for application and comparison across different professional contexts in future research. One of its most relevant contributions is its ability to capture the ambivalence that characterizes many of these attitudes, combining positive perceptions related to AI’s innovative potential with ethical, labor, and social concerns. This feature enables a more nuanced understanding of the factors influencing the acceptance or rejection of these technologies and can inform the design of training programs, institutional strategies, and awareness campaigns aimed at fostering informed and responsible adoption. Likewise, the results obtained with the GAAIS may contribute to generating empirical evidence relevanttoinforming future public policies and regulatory frameworks that support the ethical and safe use of AI, as well as the protection of individuals’ fundamental rights (
Sobrino-García, 2021).
From a theoretical perspective, this validation contributes to the development of the field of attitudes toward artificial intelligence by providing empirical evidence of its multidimensional structure and the coexistence of positive and negative evaluations toward the same technology. Furthermore, it opens the possibility of conducting cross-cultural comparisons and examining how variables such as education or technological experience shape the perception and acceptance of AI. Overall, the GAAIS contributes to strengthening a more comprehensive conceptual framework that helps to understand the conditions under which artificial intelligence can be integrated in a sustainable, ethical, and socially beneficial manner.
4.2. Limitations and Future Directions
It is important to consider several limitations that may affect the interpretation and generalization of the results. First, no prior pilot test was conducted, which could have helped identify potential difficulties in item comprehension, administration issues, or technical aspects that might affect data quality. Additionally, the sample used for this study is not necessarily representative of the entire population of Spain, since it was obtained through convenience sampling, which implies that the results should be interpreted with caution. Future studies should consider using probabilistic or stratified sampling methods to enhance external validity and representativeness. Likewise, the temporal stability of the instrument was not assessed through a test–retest procedure; this lack of evidence limits the conclusions regarding the instrument’s consistency over time.
Finally, another potential limitation is the method effect associated with reverse-worded items; that is, items worded negatively or whose direction is contrary to the general construct. This effect, which we could not control for in this study, can interfere with the instrument’s structure. Although these items are usually recoded to facilitate interpretation, they can add unwanted variance related to cognitive load, item comprehension, or systematic response patterns that reflect item format rather than construct content (
Dueber et al., 2021;
Tomás et al., 2012,
2013;
Weijters et al., 2013). In this case, since the two dimensions are composed entirely of positively and negatively worded items, an artificial factor could have been generated by shared variance among the negative items (the Negative GAAIS subscale), which is not due to the construct itself. Therefore, future research should determine the extent to which this effect exists in the GAAIS and, if necessary, propose alternative models that are capable of controlling for it.