1. Introduction
The assessment of learning in the classroom is one of the most relevant processes to lead students towards the achievement of the competencies, skills, contents or abilities included in a curriculum or program [
1,
2,
3,
4]. This is because evaluation, if it is formative, should go hand in hand with two relevant components. The first of them is associated with the quality of the evidence of learning that is generated through the review of assignments, oral participations and exams that are applied throughout a semester or academic cycle [
5]. This quality of evidence is directly associated with the meaningfulness, realism and contextual validity of the assessment conditions set for students [
2,
3,
6]. For example, an essay exam will more closely correspond to the intellectual demands of a given professional field as opposed to one where students must select the correct answer purposely from a list of alternatives [
1,
2,
7].
The second component or deployment is related to the possibility that university professors use this evidence to provide feedback on their teaching and learning processes [
3,
4,
7]. In this way, we could say that we would be facing an evaluative process of formative nature; otherwise, we would be measuring or contemplating how students are progressing in terms of their learning [
8]. For this reason, it could be affirmed that formative assessment is a process oriented towards students learning what they should learn. In other words, assessment understood in this way is aimed at learning; this is in total opposition to the practices that understand assessment as a simple verification of what has been learned. In this sense, the assessment of learning is above all a permanent disposition to know in a reflexive way the students’ learning processes in order to provide feedback [
2,
3]. This, as mentioned above, requires university professors to be attentive to the progress of their students and the possible problems they may encounter when learning. In the same way, is very important that evidence (in the form of tests or exams) enables as reliable a construction as possible, as it is on the basis of this and of the reflective processes that optimizing decisions will be taken in the form of feedback on both learning and teaching themselves [
2].
In this light, assessment is not the same as grading, which measures or verifies the quantity of what has been learnt. Nor can we equate assessment with a mere value judgement, as this would lead us to the contemplation of a reality in which there is no intentionality in the interest of improving learning [
3,
5,
7]
In view of the above, it would seem that the more formative approaches to assessment are more deeply rooted in the thinking and practice of basic education teachers (pre-school, primary, and secondary) than in university lecturers. It is possible that the reasoning that justifies not practicing truly formative assessment still persists, alluding to the fact that university students are adults and do not deserve help from their professors; when in fact—like any other student—they need help in a framework of warmth in terms of treatment [
9]. All this points towards the personal conditions that are added to the pedagogical and professional competences that every university professor must possess; the same ones that are projected in the assessment displays in the classroom [
10].
In short, assessment is a profoundly human act committed to the students whom we wish at all times to see succeed through their university studies. Such a well-meaning disposition elevates formative assessment to a practice that will lead to the greatest number of students (hopefully all of them) being able to achieve what the degree program expects of them. Thus, the success of the university education and the quality of the graduates’ professional displays would depend to a large extent on those timely aids, pedagogical adjustments, and meaningful feedback that we offer to the learners [
3].
On the other hand, it is worth mentioning that Peruvian universities in the framework of the current reform generated by the current University Law 30220 of 2014 have seen the need to deploy greater efforts for better teaching. Under these influences, assessment had to be improved significantly thanks to the development of courses, workshops and specializations in university teaching. A series of very determined actions had thus been initiated to optimize the teaching skills of professors and to investigate the impact of the measures undertaken; all this in a context of greater competitiveness of Peruvian universities to conquer better positions in terms of scientific production, respectability, and expectant location within higher education [
11].
However, these dynamics underwent abrupt changes and modifications as a result of the COVID-19 pandemic [
2]. In this case, it is not that improvements in teaching had been stopped, but rather, more urgent issues had to be addressed at that time, such as the acquisition of platforms for online teaching, professor training in the use of these platforms, and the transfer of teaching materials to a virtual space little known by the vast majority of professors [
7]. Likewise, programs had to be designed to establish teaching skills through digital platforms. Some of these training spaces would have been developed under a face-to-face education mode and others from a perspective closer to engineering than to pedagogy. In other cases, it was reported that some university owners ordered the merger of several sections or classrooms into one, where a professor who previously used to teach 25 or 35 students went from one moment to another to having more than 100 students in a single virtual classroom [
2,
3]. This would have occurred to a greater extent in private for-profit universities. In addition, these facts would have occurred in other latitudes with respect to the excessive increase in the number of students per classroom [
12].
Thus, the above-mentioned situation would have had a significant impact on the quality of teaching and assessment itself, and where, since it is practically impossible to assess 100 or 125 students through essay exams, an excessive use of multiple-choice exams would have been resorted to; situation that would have had a negative impact on the emotional state and motivation of professors and cause burnout in the context of the pandemic [
13,
14,
15]. At the same time, it is to be expected that—in other cases—professors have transferred the strengths and weaknesses of their face-to-face teaching (and their ways of assessing) to the online modality. However, this is also a problem for students because a significant percentage of them graduate from secondary education without having achieved the basic competencies needed to enter the university world [
16]. This translates into problems in oral expression, in the search for reliable information, in writing and at the level of thinking in general, among others. This is largely explained by the results of the 2018 PISA (Peru), which revealed the serious deficiencies of our young people who are close to finishing high school [
16,
17].
In addition, recent studies report the existence of problems that have become a pandemic, such as plagiarism, lack of academic integrity, the increase in artificial grading, the difficulty of grading group work, and the absence of face-to-face relationships between professors and students [
12,
18]. However, the evidence also points to favorable perceptions of inclusion, well-being, and satisfaction with online assessment in the context of pandemic-related evidence [
12,
19].
On the other hand, other studies report that in pandemics, scores would have risen considerably on the tests [
20], and that student recognized that the online modality demanded a greater effort, but with important advantages [
21]. It should be noted that most of these studies compare scores before and during the pandemic, but by means of multiple-choice exams [
22]. The number of research studies (such as this one) that analyze the types of exams used in online classes through a formative assessment perspective, and where the type of exam matters, especially given the need to provide feedback to students [
23], is small. Other researchers have reported the need to transcend the traditional multiple-choice exam that was used during this health crisis with insistence [
12,
20,
23,
24,
25]. Specifically, perceptual studies of students indicate that during the pandemic, multiple-choice tests have been favored over oral and essay tests [
26,
27,
28]. However, we have not recorded studies that tell us about students’ perceptions of assessment according to university degree or program, and, for this reason, research has concentrated on reporting global results [
5,
23]. Given this, we anticipate that each degree program has a particular culture that would be reflected in the assessment practiced in the classroom [
29,
30].
Purpose of Study
By virtue of the aforementioned situation, the general objective of the present research is to study the perceptions and preferences of university students with respect to the assessment of their learning in the classroom in the context of the emergency online teaching provided during the COVID-19 pandemic by. We thus have as specific objectives: (1) To characterize perceptions regarding the tests given by professors; (2) To describe the perceived emphases regarding the multiple-choice tests applied by their professors; (3) To identify preferences regarding the types of tests; and (4) To identify the perceptions of students of which tests are the most relevant for their professional education.
In this way, it is intended to provide inputs for debate, reflection, and decision making in this thematic field, where technological criteria (applications, programs, and automated systems to provide multiple-choice tests) seem to have taken precedence [
31]. For this reason, researchers in this field believe that, in the context of the COVID-19 pandemic, traditional tests focused more on memorization and on aspects of little relevance to students were applied [
32,
33]. Thus, this study aims to contribute, also, to this knowledge problem.
2. Materials and Method
2.1. Research Design
A quantitative approach study was developed using the survey method and the questionnaire was constructed as an instrument [
34]. In terms of its characteristics, this research is observational (non-experimental), transversal, descriptive, comparative, and correlational.
2.2. Selection of Participants
The sample consisted of 240 undergraduate university students from a private for-profit university that serves vulnerable socioeconomic sectors of Metropolitan Lima (Peru). These students had entered the university in the month of March 2021, during the COVID-19 pandemic. They were surveyed between the months of October and November (2021) of that same year while completing their second semester of their course. It should be noted that in that second semester, the total population of students was 303; therefore, the sample represents 79.21% of the total population. Likewise, these students (from this on-site university) were immersed from the beginning of their training in emergency virtual education due to the health situation. It is important to mention that the study was census-based, and therefore, the link with the instrument was sent to 100% of the students enrolled in the second semester.
Table 1 shows the distribution of the sample by degree program.
Ages ranged from 16 to 54 years, with a mean of 25.4 and a standard deviation of 8.105. It should be noted that 72% of the subjects in the sample mentioned that they study and work in different private and public companies. Likewise, 90% of them were born in a province or region other than Lima and came from the highlands (Andean and Altiplano zones) that have an altitude of more than 1500 m above sea level. Fundamentally, these students came from Huancayo, Ayacucho, Abancay, Arequipa, and Puno. Additionally, according to the internal census of this university, 89% (of those enrolled in this institution) will be the first generation of university graduates in their entire family. Finally, in terms of socioeconomic status, the university has 70% of students living in poverty, and has one of the lowest tuition pensions in the country.
2.3. Questionnaire
A questionnaire (
Appendix A) was created, where identification data, such as sex, age and degree program, are present in the first part. The second part comprises 11 multiple-choice questions that explore students’ perceptions regarding the most common types of tests (and their emphases) that were given by their university professors. The third part of the questionnaire explores preferences as to how to be assessed (essay tests or multiple-choice tests, for example). The fourth part of this instrument asks about evaluative perceptions regarding what type of test they consider most relevant to their preparation as future professionals.
It should be specified that this self-administered instrument was given thanks to Google Forms during the second COVID-19 wave. It is also important to point out that, at all times, we asked students about the perceptions regarding the assessment practiced (especially about the tests that their professors provided through the virtual classrooms) by their teachers in virtual emergency education in Peru. This research questionnaire had content validity, determined through the participation of 3 expert judges. All these judges were academics with doctoral degrees in educational sciences (one of them was a doctor in educational psychology), and with at least 10 years of experience in creating this type of test and in the thematic field of assessment. The questionnaire (anonymous) was answered under informed consent and voluntarily. It should be noted that this instrument was developed on the basis of a series of exploratory group interviews conducted by the main author of this study with students at the university.
2.4. Procedures
Access was granted to the university students thanks to the coordination established with the university’s highest authorities. These authorities were informed in writing about the nature of this research and the commitment to keep the name of the institution confidential was made explicit. In this way, a list was received with the e-mail addresses of all the students who were in their second semester. Then, the link to the questionnaire was sent by email and there were two more reminders so that they could participate in this study. Likewise, a general report of the results was given to the president of the university with recommendations regarding the classroom evaluation.
2.5. Data Analysis
Statistical analyses were performed in the SPSS program (Version 27 for Windows). Specifically, we performed descriptive and comparative analyses. The chi-square test (χ2 test) was also applied to study whether there were significant associations of the phenomena under study with the sex and age of the students.
3. Results
This research focuses on characterizing, from a quantitative approach, the perceptions that students at a university in Lima (Peru) have regarding the tests used by their professors in the context of emergency virtual education due to the health crisis caused by COVID 19. This, by virtue of the fact that each professional training program would present its own habits, idiosyncrasies, and evaluative practices that would be associated with its own pedagogical traditions. In this way, these degree-specific emphases (and those that are common) would be presented to the teaching staff, who have the power to decide which tests they will use (and which ones they exclude) in their classrooms.
3.1. Tests and Examinations Used according to Degree Program
By virtue of the above, it can be seen in
Table 2 that the students indicated that they would have had to take a greater number of response selection tests (or the so-called “objective tests” in the Peruvian educational system). It is necessary to point out that law students are the ones who believed that they sat these tests less frequently (37.5%); the rest of the students of all the programs considered that it was these tests that were mostly used in their assessment. In this context, psychology (91.7%) and nutrition (89.9%) students considered these multiple-choice tests (MC) to be used the most by their professors.
It should be added that in this study we have not considered student perceptions of the feedback received from university professors. Therefore, it is especially recommended that this aspect be considered in future scientific research in this area.
3.2. Types of Multiple-Choice Tests
Additionally, the purpose was to understand the student´s appreciation of the multiple-choice tests (MC) given by their professors by degree program. In this item, which is expressed as a result in
Table 3, In addition, we sought to know the students’ appreciation of the multiple-choice (MC) test applied by their teachers by degree. In this item, which is expressed as a result in
Table 2, students were asked about the way in which these tests (MC) were applied by their teachers (in general terms). It was found that the most frequently administered tests combined memoristic items (because they required memorising irrelevant data) and thinking or reasoning oriented items. It was found that the most frequently given tests combined memorizing items (because they demanded irrelevant data to be memorized) and those oriented towards thinking or reasoning. This is followed by tests with items for understanding and thinking (reasoning), and, to a lesser extent, exclusively memorizing test, where much lower percentages (from 0.0% to 9.8%) are evidenced for all the degree programs. It is important to mention, as something positive, that no student of psychology, nutrition, and law believed that there were exclusively memorizing answer selection tests in their university courses. As for the MC more exclusively oriented towards thinking or reasoning, it was found that the students’ perception was that these tests were given to a greater extent in nutrition (44.4%), administration (39.4%) and law (37.5).
However, it should be noted that we must consider these results (especially those presented in
Table 3) with caution since the students’ perceptions of the multiple-choice tests depend on their categories, knowledge of the subject matter, and thoughts about them. For this reason, it is likely that at the time of answering the questionnaire, many of the students may have had the opportunity to reflect on these issues for the first time. In any case, and in spite of these epistemological tensions, these perceptions refer us to the way in which they consider the tests they had to take.
3.3. Preferences on How to Be Assessed
We were also interested in finding out the preferences of students in all degree programs for two types of instruments to evaluate their learning throughout their professional training (
Table 4): essay tests (ET) and multiple-choice test (MC). Thus, we found that the preferences of university students pointed more to the MC than to the ET. It is worth mentioning that this preference for MC was present in all degree programs except accounting (26.1%) and marketing (43.5%). Likewise, we identified that nursing (68.3%) and pharmacy (65.5%) students were the ones who preferred MC to a greater extent. Additionally, we found that the proportion of law students preferring multiple choice and essay tests was the same (50%), which is striking as they are in a career that demands a lot of reading, discussion, debate, and handling of sources, which made us think (by way of conjecture) that these students would prefer essay tests to a greater extent. We also had the same reasoning with respect to psychology students.
Given this, we have no further information to contextualize these results in terms of the actual tests and exams that had to be taken in these emergency virtual classes.
3.4. Perceptions about the Most Important Tests
However, when students were asked about the most important tests for their university education, we found that the essay tests stood out in all the degree programs (
Table 5). In other words, MC are preferred in almost all degree programs, knowing that they would not be as important or would be the least relevant for their preparation for the professional field. In this case, it should be noted that 85% of the students at this university come (according to data from its academic office) from the public schools with the lowest learning results in Metropolitan Lima (Peru), as measured by the tests applied by the Ministry of Education. Thus, as is well known, the students who graduated from these basic education schools evidenced poor study habits, less ability to understand what they read, and possess fewer resources to express themselves adequately in writing. Thus, this paradox finding reveals a clear understanding that the essay tests (ET) are the most important; however, they are not preferred because the students’ anticipated problems in answering them successfully, given their educational deficiencies due to the high school or middle school they attended.
In this way, and in a more specific way, we found that law (87.5%), administration (84.8%), and marketing (82.6%) students are those who valued these essay tests to a greater extent as the most important or relevant for professional training. The other degree programs whose students perceived essay tests favorably were pharmacy (61.8%) and accounting (69.6%).
We also wanted to know whether age (ordinal variable) was associated with the preference for response selection tests or essay tests. In this case, there was no association between these variables (χ2 = 0.430, df = 1, p > 0.512). Similarly, we did not find a significant association between the age of the students and their perception of which tests are the most relevant for professional training (χ2 = 1.566, df = 1, p > 0.211). In this sense, it was determined that the marked preferences for multiple-choice tests and the belief that essay tests are the most relevant for vocational training were not associated with age. In other words, in all age groups, the perceptions are similar.
Along the same lines, we found that the sex of the students was not associated with preferences for either multiple-choice test or essay tests (χ2 = 3.175 df = 2, p > 0.204). Along the same lines, we found that there was no association between the sex of the students and their perception of developmental or essay (EE) tests as the most important for their professional training (χ2 = 1500, df = 2, p > 0.472). These last two analyses confirm that these preferences (multiple-choice tests) and the considerations about the most important tests (essay tests) are associated with previous school experiences (unsatisfactory in terms of communicative competencies) and the search for maximum ease. Or perhaps it is likely that these results refer to students anticipating (out of habit and developed abilities) that they would do worse (or require a different approach to study) with essay tests.
4. Discussion
This quantitative research aimed to explore the perceptions and preferences regarding classroom assessment in students of a private for-profit university, which mostly serves vulnerable populations. In this context, according to the figures of this university’s report, 70% of students are in poverty and 85% of them have completed their high school studies in public schools in the peripheral areas of the city of Lima. These data are very important because we are probably dealing with a student population (represented in the sample) with largely unsatisfactory school experiences and poor cultural stimuli (books, reflective dialogues, and student models). This fact has most probably generated the identified paradox that reveals that multiple choice exams (MC) are preferred, and, at the same time, essay test (ET) is recognized as the most appropriate for university education. Given this, it is possible that for these students, it is more complicated to have to write an answer because, most likely, in basic education they had few opportunities to develop this competence. On the other hand, answering a multiple-choice exam—even one of simple memorization—would be more within reach, given the low cognitive demand of these tests; even more so when there are still professors in much of the Peruvian educational system who are more focused on having their students retain large amounts of data, dates, principles, and formulas in their memory [
2,
7,
23,
29].
In addition to this, it was clearly identified that the tests that were given by university professors in this study were mostly response selection (MC) tests, and, within these, there was emphasis on the tests that combine memorized items and explore thinking or reasoning [
23]. However, as noted in the results, the degree program variable was very revealing, in the sense that it allowed us to identify differentiated perceptions. This is most likely related to the existence of different professional cultures within each university, which would explain the ways of approaching teaching, assessment, and the tasks entrusted to students [
35,
36].
In any case, it will be necessary to develop qualitative studies aimed at knowing the perceptions, attributions, and reasons that professors have for using some types of tests and not others. To know the reasons certain tests are chosen, we would have to consider the specialty of the professors in these courses since, in the current study, it has been found that not only would the students find the assessment process different, but so would the professors [
37]. Additionally, other variables or characteristics of the students should be considered in future studies in order to discriminate more clearly this phenomenon (even more so when we only found the degree program as the only relevant variable). For example, it would be advisable to consider in the future the semester and the type of subject (general education and specialization) that would make it possible to more precisely identify and understand the perceptions on the exams given in the classroom [
37]. However, also, it would be pertinent to research the university professors themselves to learn about the tests they give and the reasons they have for considering them for their work in the classroom.
On the other hand, we must understand that during the pandemic some universities in Peru (especially for-profit ones) increased the number of students in virtual classrooms. This fact, which has not been investigated in depth, was probably due to the need to compensate for the economic crisis that many Peruvian universities went through in the face of the large student desertion that occurred mainly during the first year of the health crisis due to COVID-19.
As a consequence of the above (the fact of working with classrooms with 80 or 100 students), it is very likely that a significant number of professors opted to use multiple-choice exams instead of essay tests. However, it should also be noted that Peru was the country (according to the World Health Organization) with the highest number of deaths due to the COVID-19 pandemic in the world. This undoubtedly had an equal impact on students and their university professors; the emotional impact was compounded by the economic impact due to the shortage of medical oxygen, medical supplies, and beds in public hospitals. Therefore, it is likely that this situation had some impact or influence on the perceptions of the students who responded to the questionnaire, and also, on the emphasis on the tests given by their professors in the framework of emergency virtual education.