Validation Using Structural Equations of the “Cursa-T” Scale to Measure Research and Digital Competencies in Undergraduate Students

: Research competencies are considered essential in fields such as science, academia


Introduction 1.Research Skills in Undergraduate Students
Research competencies encompass the set of skills and knowledge essential for effectively conducting research endeavors [1].Essentially, they denote the proficiency in utilizing scientific knowledge, formulating pertinent questions, and deriving conclusions based on evidence, thereby contributing to a comprehensive understanding of the world or the transformations occurring within it.Ramírez-Armenta et al. [2] underscore the significance of these competencies in facilitating the research activities of individuals, emphasizing the need to acknowledge variations in mastery and to delineate the thresholds between levels of acquisition, such as at the bachelor's, undergraduate, or postgraduate levels.
These competencies hold paramount importance across various domains, including but not limited to science, academia, and technology, where the generation of new knowledge or the resolution of intricate problems is imperative.Fostering research skills in undergraduate students serves a dual purpose: not only does it prepare them for advanced studies at the postgraduate level, but it also provides them with valuable research experiences that can lead to groundbreaking discoveries, innovations, or projects.
Moreover, recognizing the pivotal role of recent and high-quality research results in informed decision-making within professional settings; it becomes imperative to instill the methodology of utilizing scientific evidence throughout the entire span of university education [1,[3][4][5][6].The ability to critically evaluate and apply research findings is an asset that enhances professional competence and contributes to advancements in diverse fields.
In the realm of health, specific competencies in research are identified as foundational for the training of healthcare professionals, as highlighted by various studies [4,7,8].These competencies equip health practitioners with the necessary tools to navigate the complex landscape of healthcare research, ensuring that they can contribute meaningfully to advancements in medical knowledge and healthcare practices.As the healthcare field continually evolves, the cultivation of research competencies becomes indispensable for professionals striving to stay abreast of the latest developments and contribute to the improvement of health outcomes.

The Importance of Technology for Health Research
Technology is already part of the daily routine of teachers and students, but it requires skill, because if used properly it supports university work, thus facilitating academic work [9].
In the context of research competencies, are located technological skills considered as relevant, such as the following: (a) information search and management based on the use of specialized databases, online search engines, and other tools for accessing relevant scientific literature and efficiently managing the information obtained; (b) use of software for analysis, visualization, interpretation, and presentation of health research results; (c) use of online communication and collaboration technologies, such as videoconferencing, email, and platforms that allow students to interact with other researchers, share results, and collaborate on research projects; (d) use of technologies such as mobile devices, sensors, and specialized applications to collect information accurately and efficiently; (e) privacy considerations when using technology in health research in compliance with established ethical standards [10][11][12].
All of these are fostered in both graduate and undergraduate students, ranging in skill level and proficiency.It is important to note that the technological skills required may vary depending on the specific field of health.

Measuring Competencies
In this respect, in the review of the literature it was observed that there are few works in which valid and reliable instruments are shared that measure investigative competencies with respect to undergraduate training; even fewer are specific to the areas of health, and something that is missing as a fundamental dimension is the inclusion of technology as a basis for carrying out research nowadays [13,14].Digital competencies have been measured and studied as a construct separate from the ability to carry out research or as an activity far removed from undergraduate students, with instruments with psychometric properties dedicated to assessing skills in postgraduate students [15][16][17][18][19].
To address this gap, it is imperative to explore models of competition from various international contexts.This section delves into models employed in Europe, Australia, and the United States, offering historical context and scrutinizing the reasons behind the design of these competition frameworks.Understanding the evolution of these models is vital for assessing their relevance and effectiveness in measuring research skills at the undergraduate level.
European countries have been at the forefront of developing frameworks to evaluate research skills among undergraduate students.The European Higher Education Area (EHEA) has witnessed a proliferation of competency-based assessment models, such as the European Qualifications Framework (EQF) and the Tuning Educational Structures in Europe project.These initiatives were born out of the Bologna Process, aimed at harmonizing higher education systems across Europe.
The EQF, for instance, was designed to facilitate comparability and transparency of qualifications across European countries, fostering a competitive environment that encourages the development of research skills.The Bologna Process sought to create a standardized and competitive landscape to enhance the quality and relevance of education, aligning it with the demands of the knowledge-based economy [16].
In Australia, the Quality Indicators for Learning and Teaching (QILT) framework has been instrumental in measuring research skills at the undergraduate level.Originating from a need to enhance the accountability and transparency of higher education institutions, the QILT framework integrates student feedback, graduate outcomes, and employer satisfaction to assess the overall quality of education.
The competitive nature of this model stems from Australia's commitment to providing students with a world-class education that equips them with research skills applicable in a global context.The QILT framework reflects a historical shift towards outcomeoriented education, emphasizing the practical relevance of research skills for students in the Australian higher education landscape [16,17].
In the United States, the National Survey of Student Engagement (NSSE) and the Council for Accreditation of Educator Preparation (CAEP) are examples of models assessing undergraduate research skills.The historical context of these models lies in the U.S. commitment to promoting innovation, critical thinking, and research prowess among its students.
The competitive aspect of these models is rooted in the U.S. emphasis on global leadership in research and development.The models aim to ensure that undergraduate education aligns with the evolving needs of a knowledge-driven economy, fostering a culture of competitiveness and excellence [18].
While these international models of competition have undoubtedly contributed to shaping undergraduate research skills assessment, a critical examination is necessary.The design of these models often reflects regional priorities, and their applicability to diverse contexts may be limited.Additionally, a potential criticism lies in the inherent competition these models foster, potentially overshadowing collaborative aspects of research, which are equally crucial in today's interconnected world.
However, a nuanced critique is essential to ensure that these models effectively address the multifaceted dimensions of research competencies, including the integration of technology and collaborative approaches.
Based on the above lines, the main objective of this research is linked to the design and validation process of a scale to measure the inquiry and digital competencies of undergraduate students.

Sample
The sample consists of 1104 health science students from a Mexican university, specifically, 837 women and 267 men studying at university.The average age is 26 years old.Together, all of them agreed to participate in the research after being informed of its main objectives (informed consent).With 1104 participants in a population of 16,000 people, a maximum margin of sampling error of ±2.85% is obtained with a confidence level of 95%.

Data Collection Instrument
To collect the data, a questionnaire called "Cursa-T" was developed.It was developed on the basis of similar studies [20][21][22][23][24].Each item is measured on a Likert scale of 5 intervals, where 0 represents the minimum value and 4 the maximum.
The questionnaire has a short heading in which the main objective of the research is presented.It then proceeds to collect the socio-demographic characteristics of the participants.Next, 51 items are presented to assess the students' self-perception of research competencies (Table 1).The dimensions of the instrument are intricately linked to the Research Competencies Development Program, a framework grounded in the four foundational pillars of education-knowing, knowing how, doing, and being-within the specific institutional context where the evaluation took place.These pillars serve as the guiding principles for shaping a comprehensive approach to research competencies, acknowledging the multifaceted nature of education and skill development.
Simultaneously, the theoretical framework advanced by Juárez and Torres [9], emphasizing investigative competencies at varying levels-basic, intermediate, and advanced-was thoughtfully incorporated into the evaluative process.This theoretical perspective provides a nuanced understanding of research competencies, recognizing the evolution and progression of these skills across different proficiency levels.By considering these hierarchical levels, the evaluation endeavors to capture the diverse spectrum of competencies individuals may possess, ensuring a comprehensive and nuanced assessment.
This dual consideration of the institution's educational pillars and the tiered framework proposed by Juárez and Torres not only enriches the evaluation process but also aligns it with a broader educational philosophy.It acknowledges the dynamic interplay between foundational knowledge, practical application, and the evolving nature of research As such, the evaluation instrument aims to reflect a holistic perspective that encompasses both the institution-specific educational principles and the overarching theoretical framework, fostering a more nuanced and contextually relevant understanding of research competencies.

Data Collection and Analysis Procedure
The questionnaire was carried out in digital format, through the "Microsoft Forms" platform.Data collection was carried out during the months of October 2022 and January 2023.The purposes of the study were explained and the collaboration of the students was requested.The anonymity of the participants was ensured at all times.
The data matrix has been modified for operational reasons.Thus, new "dimensions" variables have been created (Dim.A, Dim.B, Dim.C, Dim.D, and Dim.E) calculated from the average of the items that compose them, with a minimum value of 0 and maximum 4. The reliability, discriminant validity, and convergent validity of the questionnaire have been calculated using the following coefficients: Cronbach's Alpha, Composite Reliability (CR), Extracted Average Variance (AVE), and Maximum Shared Variance (MSV).Additionally, and to compare these results, the inferential analysis method between items and dimensions is used.To do this, the bivariate correlational analysis technique is used through Spearman's correlation coefficient ρ.For its part, the construct validity of the test has been obtained through an exploratory factor analysis (EFA).The method used to select the factors is the principal components method.The obtained factors are rotated orthogonally using the Varimax method with Kaiser normalization.Once the number of factors has been determined, a confirmatory factor analysis (CFA) is performed.Confirmatory factor analysis is used to check whether the theoretical measures of the model are consistent through modeling diagrams and use of structural equations [25].That is, it is checked to determine whether the data fit the hypothetical measurement model revealed by the exploratory factor analysis.The method used to contrast the theoretical model has been weighted least squares (WLS), which provides consistent estimates in samples that do not fit normality criteria [25].For this last procedure, the AMOS computer program was used, capable of revealing complex hypothetical relationships between variables, using structural equation modeling (SEM).At the same time, it has been proven that the data are not normally distributed through a descriptive study in which asymmetry and kurtosis have been taken into account.The Kolmogorov-Smirnov goodness-of-fit test has confirmed this verification, with significance (p-value) equal to 0.000 for all items (non-normal distribution).

Results
The assessment of the reliability of the "Cursa-T" questionnaire involved the calculation of Cronbach's alpha coefficient, both in its entirety and for each of its distinct dimensions.The comprehensive analysis yielded an impressive Cronbach's alpha index of 0.983, signifying an exceptionally high level of reliability for the questionnaire [26].This index, surpassing the threshold of 0.9, unequivocally underscores the robustness and dependability of the instrument.
Furthermore, a granular examination of the individual dimensions within the questionnaire reveals equally commendable reliability coefficients for each.Specifically, the reliability coefficients for the dimensions are as follows: A. Knowledge of research methodology (0.960); B. Use of statistics for research (0.960); C. Scientific report (0.967); D. Oral support of projects (0.959); and E. Technological applications in research (0.954).Notably, all these dimensions exceed the recommended threshold of 0.7, thereby affirming their high reliability and indicating their efficacy in capturing the intended constructs [27].
To provide a more comprehensive evaluation, additional statistical metrics were considered, including the Composite Reliability (CR), Average Extracted Variance (AVE), and Maximum Shared Variance (MSV).The outcomes of these calculations, along with reference values for model adjustment, are detailed in Table 2.These supplementary analyses contribute to a more nuanced understanding of the questionnaire's reliability by examining factors such as convergent validity, shared variance, and overall model fit [28].In summary, the rigorous examination of the "Cursa-T" questionnaire, encompassing both global and dimensional perspectives, underscores its robust reliability, positioning it as a sound and trustworthy tool for assessing research competencies.The combination of Cronbach's alpha coefficients and additional statistical parameters enhances the overall confidence in the questionnaire's ability to yield reliable and valid results in the evaluation of research-related skills and knowledge.
All figures obtained adjust to the reference values.Therefore, the reliability of the model (CR) as well as its convergent (AVE) and discriminant validity (MSV) are demonstrated.Next, we proceed to analyze the simple correlations of each item with the theoretical dimension or construct.The results are shown in Table 3.Values that do not correspond to the corresponding item-dimension relationship have been eliminated to facilitate reading the table.
Any value surpassing the threshold of 0.707 is deemed acceptable for an item to be included as a constituent of its respective dimension [29].Consequently, all items have met this criterion and have been seamlessly incorporated into their designated dimensions.
The assessment of construct validity for the test was conducted using an exploratory factor analysis, the results of which are detailed in Table 4. Prior to this analysis, the suitability of factor analysis was verified through the Kaiser-Meyer-Olkin (KMO) test, yielding a highly significant coefficient of 0.983.Additionally, Bartlett's sphericity test confirmed the applicability of factor analysis, with a significance level (p-value) of 0.000, affirming the appropriateness of conducting the factor analysis.
These preliminary tests not only justified the utilization of factor analysis but also underscored the robustness of the dataset for such an analytical approach.The high KMO coefficient signifies the adequacy of the sample for factor analysis, while the significant Bartlett's sphericity test implies that the variables within the dataset are interrelated, justifying their inclusion in the factor analysis.
The subsequent exploratory factor analysis, as elaborated in Table 4, delves into the underlying structure of the test, unraveling the latent factors that contribute to the observed patterns of responses.This analysis not only facilitates a deeper understanding of the construct validity of the test but also sheds light on the relationships between variables and the overarching dimensions they collectively represent.Values that do not correspond to the corresponding item-dimension relationship have been eliminated to facilitate reading the table.The outcomes, elucidating 73.78% of the variance, establish the existence of the five proposed theoretical factors: Knowledge of research methodology (Dim.A), Use of statistics for research (Dim.B), Scientific report (Dim.C), Oral support of projects (Dim.D), and Technological applications in research (Dim.E).These factors encapsulate the fundamental dimensions underpinning the observed patterns of responses.To validate and refine the theoretical model derived from exploratory factor analysis (EFA), a confirmatory factor analysis (CFA) is employed.Figure 1   This iterative process, from EFA to CFA, contributes to the refinement and validation of the theoretical model, enhancing its accuracy in capturing the complex interplay between variables and dimensions.Through this analytical journey, the study aims to establish a robust and reliable framework that aligns with the underlying constructs of interest, ultimately enhancing the credibility of the assessment tool.The CFA serves as a critical step in corroborating the theoretical underpinnings identified through EFA.By examining the structural relationships between the observed variables and their respective latent factors, CFA evaluates the goodness-of-fit of the proposed model.The graphical representation in Figure 1 allows for a clear and concise depiction of how each item aligns with its designated dimension and the interconnections between different dimensions.
This iterative process, from EFA to CFA, contributes to the refinement and validation of the theoretical model, enhancing its accuracy in capturing the complex interplay between variables and dimensions.Through this analytical journey, the study aims to establish a robust and reliable framework that aligns with the underlying constructs of interest, ultimately enhancing the credibility of the assessment tool.
The factor loadings within the dimensions range from 0.747 to 0.886, indicating robust levels of correlation between the observed items and their respective latent factors.These values signify a high degree of association, underscoring the strength of the relationship between the measured variables and the underlying dimensions.Additionally, the interrelationships among different factors (dimensions) also exhibit predominantly high levels, with an average correlation coefficient of 0.694.This collective pattern of strong associations reinforces the coherence and internal consistency of the proposed model.
To further assess the model's fit, Table 5 presents both the obtained values and reference benchmarks based on Lévy Mangin and collaborators [27].Key indicators, including Chi-Square (CMIN), goodness-of-fit index (GFI), parsimonious goodness-of-fit index (PGFI), normalized fit index (NFI), and normalized parsimonious fit index (PNFI), are scrutinized to evaluate the model's overall fit to the observed data.This comprehensive evaluation helps ascertain the adequacy of the proposed model in capturing the underlying structure of the dimensions and their interrelationships.
These statistical metrics serve as valuable tools for gauging the quality of the model, providing insights into its goodness of fit, parsimony, and overall suitability for representing the intricate relationships within the data.The examination of both factor loadings and model fit parameters collectively contributes to a comprehensive understanding of the model's reliability and effectiveness in capturing the intended constructs.
Evidently, all the obtained results align harmoniously with the reference values.Consequently, these findings serve to validate and affirm the proposed theoretical model.

Discussion and Conclusions
Since 2007, Llanos and Dáger [30] underscored the significance of cultivating research competencies among undergraduate students, particularly in the context of addressing challenges within Latin America, because of the limited scientific production at universities in this higher education level.Aliaga-Pacora et al. [31] underlines the necessity to identify and provide a clear process of how knowledge and abilities, which together form the investigative competency, align with the purpose of allowing the mastery of practical actions that generate knowledge for professionals.The presented instrument serves as a pivotal step, allowing for diagnostic assessments to inform decisions about training programs, particularly within the high school framework that shapes the university's educational model [4,11].
Contrary to the misconception that undergraduate education does not aim to produce researchers, studies by Ramos et al. [3] and Chávez-Ayala et al. [12] assert that it fosters positive attitudes toward research at this level, with some students adopting scientific research practices in their professional endeavors.Consequently, the necessity arises for valid and reliable instruments that can gauge the evolution of these competencies across different training levels.
While all investigative competencies integrated into the dimensions of the instrument are desirable in the context of health professionals' education, it is crucial to recognize the potential for specific approaches based on the distinct needs or objectives of various health areas, such as medicine, nursing, pharmacy, among others [32][33][34][35].Thus, a thorough analysis is essential, considering the unique profiles of each training program to determine advances or modifications in study plans [36,37].
At the same time, it is necessary to develop explanatory models of the use of technologies when carrying out research.The framework and results of this study underscore the pivotal role of technology, encompassing devices, software, and platforms.However, there is a recognized need to explore the influence of social networks on research skills.Acknowledging social networks as valuable reflections, given their capacity to facilitate participation in academic communities, becomes imperative.Teaching these skills to health-focused undergraduate students becomes crucial, preparing them for an increasingly digitized environment and bolstering their ability to contribute to research advancements in the health domain [38].
In conclusion, this research contributes an instrument that enables periodic evaluations of the effectiveness of teaching research competencies.It serves as a catalyst for decisionmaking based on evidence, the promotion of innovation, and, most importantly, the preparation of students for the academic and professional challenges awaiting them in the dynamic field of health.
portrays the structural diagram outlining the item-dimension relationships and dimension-dimension correlation indices, providing a visual representation of the hypothesized model.
between different dimensions.

Table 1 .
Description of the questionnaire items.

Table 2 .
Convergent and discriminant validity of the model.

Table 3 .
Correlations of the items with the associated dimensions.