Next Article in Journal
Tech-Enhanced Vocabulary Acquisition: Exploring the Use of Student-Created Video Learning Materials in the Tertiary-Level EFL (English as a Foreign Language) Flipped Classroom
Next Article in Special Issue
Exploratory Study on Geometric Learning of Students with Blindness in Mainstream Classrooms: Teachers’ Perspectives Using the Van Hiele Theory
Previous Article in Journal
The Impact of Integrated Project-Based Learning and Flipped Classroom on Students’ Computational Thinking Skills: Embedded Mixed Methods
Previous Article in Special Issue
The Use of Guided Reflections in Learning Proof Writing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Impact of the COVID-19 Pandemic upon Mathematics Assessment in Higher Education

by
Eabhnat Ní Fhloinn
1,* and
Olivia Fitzmaurice
2
1
School of Mathematical Sciences, Dublin City University, D09 V209 Dublin, Ireland
2
School of Education, University of Limerick, V94 T9PX Limerick, Ireland
*
Author to whom correspondence should be addressed.
Educ. Sci. 2025, 15(4), 449; https://doi.org/10.3390/educsci15040449
Submission received: 18 December 2024 / Revised: 27 March 2025 / Accepted: 1 April 2025 / Published: 3 April 2025

Abstract

:
Historically, the assessment of mathematics in higher education comprised closed-book, summative, proctored examinations. Related disciplines and subjects like statistics, mathematics education, and the history of mathematics lend themselves to a broader range of assessment techniques that have been reported to provide a more balanced picture of students’ abilities. In 2020, an online environment for the teaching and learning of mathematics was imposed on the academic world globally as a result of the COVID-19 pandemic. In an effort to teach and assess remotely while maintaining institutional academic standards, the majority of lecturers were in a situation where closed-book, proctored assessments were not an option. As a result, other methods were adopted. This paper reports on an investigation into how mathematics lecturers worldwide assessed mathematics before the pandemic, during the initial lockdown restrictions, and in the immediate aftermath, while some restrictions were still in place, to see if any changes were sustained. There was a statistically significant difference in the proportion of respondents who used many of the assessment types investigated across the three time periods, including open-book timed, open-book untimed, closed-book, multiple-choice questions, online proctored, in-person proctored, presentations, projects, and assignments. The majority of those who favoured closed-book proctored examinations prior to the pandemic moved to timed open-book assessments. Differences between the weightings of final examinations versus continuous assessments were also statistically significant, with greater weight given to continuous assessment once the pandemic began. Respondents’ satisfaction levels with their assessments were significantly different also, with the highest satisfaction levels prior to the pandemic and the lowest during the initial lockdown restrictions. Academic integrity was a key concern of the majority of respondents when assessing the learning outcomes of their modules and played a role in the vehicle of assessment they chose.

1. Introduction

The assessment of mathematics in higher education has long depended on closed-book, summative, proctored examinations (Iannone & Simpson, 2011; Davies et al., 2024). Arguments have been made to support the assertion that mathematics is in some way “different” as a discipline, in that it lends itself more to traditional forms of teaching and assessment, more so than might be the case in other subjects—but equally, more recently, such arguments have been refuted (Becher, 1994; Ní Fhloinn & Carr, 2017). Particularly since the early years of this millennium, research has been conducted into the possible modes of assessment into which mathematics could naturally expand, citing the fact that the traditional timed, closed-book assessment does not effectively assess skills such as problem-solving or the use of IT (Challis et al., 2003). It seems clear at least that mathematics assessment of this kind may fall short of the mark, as it may not give a fully comprehensive picture of a student’s learning achievement (Burton & Haines, 1997). Although assessment of a broader range of skills can be achieved through approaches such as report writing, projects, and oral examinations (Niss, 1998), it appears that alternative modes of assessment are only more prevalent in modules of statistics, history of mathematics, mathematics education, and final-year projects (Iannone & Simpson, 2011).
During the initial university closures due to COVID-19 in 2020, mainly out of necessity, a broader range of formative and summative assessment methods were embraced by the mathematics teaching community (Fitzmaurice & Ní Fhloinn, 2021). This was a positive outcome of the move to remote teaching, as alternative assessment strategies assess a broader range of learning outcomes (Pegg, 2003). However, while this change happened during the first months of the pandemic, when lecturers had to pivot to online assessment with little or no time to plan, it is of interest to determine what happened the following year, when lecturers still often had to assess mathematics online but had significantly more notice of the fact.
Two decades ago, online assessment in mathematics started becoming more prevalent as the internet became more conducive to mathematics (Engelbrecht & Harding, 2004). The use of technology in education is seen as an inherent component of a teaching and learning environment that seeks to fulfil the diverse needs of students in the 21st century (Valdez & Maderal, 2021). Digital technologies offer compelling tools to conduct formative assessments effectively in mathematics (Barana et al., 2021). Online assessment, or E-assessment, comprises an extensive range of assessment types, including but not limited to online essays and computer-marked online examinations (James, 2016). Online examinations are an efficient means of conducting diagnostic, formative, and summative assessments and providing students with the opportunity to perform to the best of their ability (Valdez & Maderal, 2021).
Recent years have seen exponential growth in the different modes of online assessment that are available. Research by Davies et al. (2024), conducted in 2024, however, indicates that Computer-Aided Assessment (CAA) in tertiary mathematics remains underutilised despite these reported advancements in assessment methods. The continued over-reliance by university-level mathematics lecturers on closed-book written exams referred to above prompts questions about whether CAA can provide a more effective alternative, particularly in formative assessment. Some studies have documented that CAA has been shown to improve examination performance (Greenhow, 2015), although Greenhow (Greenhow, 2015, 2019) recommends that it should complement rather than replace traditional mathematics assessments. When used effectively, online formative assessment has the potential to nurture a learner- and assessment-centred focus using formative feedback and enhanced learner engagement with worthwhile learning experiences (Gikandi et al., 2011). However, it should be noted that distance assessment is not always experienced as a positive for students, as they can struggle with access to technology and resources or simply with feelings of isolation (Kerka & Wonacott, 2000). Multiple-choice questions have also been shown to have the potential for bias in relation to students with varying learning styles or in relation to confidence levels (Sangwin, 2013). Specific training or knowledge of question creation is also needed, as questions assessed via computer-aided assessment are different from those graded by hand (Greenhow, 2015).
Iannone and Simpson’s more recent investigation into summative assessments in the UK (Iannone & Simpson, 2022) found that closed-book examinations are still tremendously popular; however, E-assessment has increased significantly in the decade since their last study on mathematics assessment. Systems like STACK and NUMBAS are widespread in many universities, mainly as part of some coursework components of first-year modules. These systems are attractive because of the time-saving aspects of electronic marking and the provision of rapid feedback (Iannone & Simpson, 2022). STACK is an open-source tool that can seamlessly integrate with learning management systems like Moodle, enabling the generation of randomised, automatically graded mathematics questions. Davies et al.’s (2024) work ratifies STACK as an assessment vehicle that delivers instant feedback, with adaptability in assessing complex mathematical concepts. NUMBAS is another open-source, web-based assessment tool designed specifically for mathematics. Similar to STACK, it provides interactive, automatically graded assignments and instantaneous feedback to participating students, which they tend to put great value on (Lishchynska et al., 2021). While the questions are customised and randomised, this can be quite a time-consuming process for lecturers who choose it (Lishchynska et al., 2021).
Valdez and Maderal (2021) state that online assessments in mathematics are increasing in use and popularity over the traditional paper-and-pen type as they evaluate student learning without the need for everyone to be physically present in the same room. The decrease in cost and increase in the availability of powerful technology have altered how many mathematics lecturers assess their modules (Stacey & Wiliam, 2012). Yet Iannone and Simpson (2022) found that there remains a relatively low level of variety in what they call ‘the assessment diet’ in mathematics in HE in the UK. There is a question over whether reasoning can be assessed online as efficiently as in a traditional examination setting; however, Sangwin (2019) demonstrated that typical closed-book exam questions in linear algebra could be replicated in an e-assessment system.
Academic integrity is a primary concern when selecting modes of assessment. Universities must uphold the academic veracity and exit standards of their degrees to preserve their reputation, and a move to online assessment can coincide with grade inflation if students are given increased time to complete assessments (Henley et al., 2022). Instances of academic misconduct have been shown to increase when assessment is fully in online format. Contract cheating refers to instances where students hire someone else to complete their work or provide answers on their behalf (Liyanagamage et al., 2025). Lancaster and Cotarlan (2021) found a 196% increase in contract cheating requests across five STEM subjects when comparing the periods from April to August 2019 and April to August 2020. This rise coincided with the shift to online assessments due to the COVID-19 pandemic. Trenholm (2007) contends that proctoring truly is the only method to eliminate cheating in online exams. However, Eaton and Turner’s (2020) research on E-proctoring, the systematic remote visual monitoring of students as they complete assessments, may have a detrimental impact on students’ mental health and well-being. Sarmiento and Prudente (2019) demonstrated that ways around this are achievable. Their work illustrates that it is possible to assess online and limit opportunities for copying. They used MyOpenMath to generate individualised homework assignments for students. They found it not only limited copying but also had a significant positive impact on students’ homework that was submitted and on their summative assessment performance (Sarmiento et al., 2018).
While there is an abundance of studies that examine student perceptions about online learning, there is a dearth in the literature on online assessment of mathematics (Valdez & Maderal, 2021), specifically which areas of mathematics lend themselves more to being assessed online. The university closures due to the COVID-19 pandemic forced lecturers worldwide out of their comfort zones and normal practices when it came to assessing their students. This research investigates the extent to which lecturers migrated from their conventional assessments when pressurised to do so, the lessons learnt during this time, the assessment changes that were preserved when they had a little more time to think and plan, and those that were discarded on their second attempt at online assessment. The research questions we explore in this paper are as follows:
  • What assessment types were used by mathematics lecturers before the pandemic, during the initial university closures, and during the academic year 2020/2021?
  • Were there changes to the weightings given to final examinations versus continuous assessment during these time periods?
  • Did mathematics lecturers observe any changes in grade distribution within their modules during these time periods?
  • Were mathematics lecturers satisfied with their assessment approaches during these time periods?
  • What do mathematics lecturers believe are the easiest and most difficult aspects of mathematics to assess online?

2. Materials and Methods

2.1. Sample

The profile of the respondents in the survey can be seen in detail in Table 1, with gender, age, years of experience in teaching mathematics in higher education, and employment status given. There was a total of 190 respondents to the survey. The gender breakdown of the respondents was 52% female, which does not reflect the population of mathematics lecturers in higher education, as this is predominantly male. The survey was sent to a mailing list of female mathematicians in Europe, which likely accounts for the high response rate from female mathematicians. The age profile was fairly evenly spread, with 85% of respondents between 30 and 59 years of age and a similar percentage in permanent employment. Their teaching experience in higher education reflected the age profile, with 40% of respondents having more than 20 years of experience.
The respondents were based primarily in Europe, with only 12% based outside of the continent, most of these in the United States. The highest proportion by far was based in Ireland, as are the two researchers in this study. In total, respondents from 27 different countries answered the survey. Further details can be found in (Ní Fhloinn & Fitzmaurice, 2022).
Respondents were also asked about their mathematics teaching in the academic year 2020/2021 to provide context for their responses. Half of the respondents lectured students taking non-specialist (service) mathematics, while 60% taught students undertaking a mathematics major. We defined “small” classes as those having less than 30 students, “medium” as being those between 30 and 100 students, and “large” as more than 100 students. From our sample, 58% had small classes, 57% had medium classes, and 37% had large classes. Almost three-quarters of respondents did all their teaching online that year, with a further 17% doing it almost all online. Prior to the pandemic, 75% of respondents had done no online teaching of any kind, with a further 13% having done only a little.

2.2. Survey Instrument

This study falls under the umbrella of a larger investigation into the remote lecturing of tertiary-level students of mathematics during the COVID-19 pandemic. Our research design comprised the creation and dissemination of a purposely designed survey to investigate how lecturers and students were coping with remote mathematics education during the initial stages of the COVID-19 pandemic around the world. In 2020, we created an initial survey that would shed light on how mathematics lecturers were responding to the challenge of a spontaneous and forced move to remote teaching. The survey questions were original questions devised by the authors. The outcome of this research is reported in (Ní Fhloinn & Fitzmaurice, 2022).
To get a comprehensive insight into this experience and period of time where there were varying levels of restrictions enforced in 2021, we disseminated a follow-up survey in order to make comparisons between remote teaching in 2020 with that 12 months later. It is this follow-up survey on which this study focuses. It was largely based on the previous survey, with questions adapted to account for the fact that lecturers now had a year of remote teaching and assessment experience. It was an anonymous survey created in Google Forms, with a consent to participate checkbox on the landing page. We piloted the survey with a panel of experts within two university mathematics departments to increase the reliability and validity of the instrument. The lecturers were asked to review the questions to check for clarity of phrasing and if the questions were unbiased and aligned with the intended constructs. Suggestions for items we had not included were also welcomed. We asked the panel of experts to identify ambiguities, inconsistencies, or potential misinterpretations in a bid to improve content validity. Their feedback helped us refine question wording and format, thus increasing reliability. This process was conducted so that the survey would accurately address our research questions and produce consistent results across different respondents. The relevant survey questions are shown in Appendix A.
The survey began with a number of profiling questions, the results of which are shown in Table 1. There were then sections on their mathematics teaching allocation, the types of technology they used and the purpose of this, student experience, remote teaching experience, and personal circumstances, as well as the section which is the focus of this paper—online assessment in mathematics. Within this section, there were nine questions, four of which were open-ended.

2.3. Data Collection

The survey was made available exclusively online, using Google Forms, and was advertised via mailing lists for mathematics lecturers, as well as at mathematics education conferences relevant to those working in higher education.

2.4. Data Analyses

Respondents were asked in many places to give their approach at three specific time points: prior to the pandemic, during the initial university closures from March–June 2020, and during the academic year 2020/2021. Throughout the paper, we will refer to these as Period 1, Period 2, and Period 3 for simplicity.
Quantitative analysis was done using SPSS (version 29). For investigating binary changes over the three time periods, we used Cochran’s Q test, as this is a non-parametric test for comparing binary outcomes across three or more time periods for the same group of subjects. In our results reported below, a statistically significant result is one for which the p-value is less than 0.05. Where the results of Cochran’s Q tests were significant, we conducted post-hoc McNemar tests to ascertain where the significant differences lay. In this case, we needed to utilise a Bonferroni adjustment on the results because of making multiple comparisons, and so a statistically significant result is one for which the p-value is less than 0.017. To investigate changes in Likert-scale data over the three time periods, a Friedman test was used, as this is a non-parametric test for comparing ordinal data across three or more time periods for the same group of subjects. Where the results of the Friedman test were significant, we conducted post-hoc Wilcoxon signed-rank tests as above for the McNemar tests, with a similar Bonferroni adjustment.
For the qualitative responses, grounded theory was utilised, specifically general inductive analysis. Both researchers independently coded the responses initially to provide greater reliability in the results. Inter-coder reliability was 91%. We used the open discussion to categorise the disputed data points. This revealed minor differences in the interpretation of some data that, when clarified, led to almost perfect agreement.

3. Results

3.1. Assessment Types

Respondents were asked about the types of assessments that they conducted at three specific time points: Period 1, Period 2, and Period 3. The results are shown in Figure 1.
A Cochran’s Q test was conducted on each assessment type, and the outcomes are shown in Table 2 below. From this, we found that there was a statistically significant difference in the proportion of respondents using most of these assessment types over time: namely, open-book timed, open-book untimed, closed-book, MCQ, online proctored, in-person proctored, presentations, projects and assignments.
Further post-hoc McNemar tests were conducted for the assessment types that were statistically significant, and the results are shown in Table 3 below. From this, it can be seen that for open-book timed and online proctored assessments, there was a statistically significant difference between each of the three time periods when considered in pairs—so there was a significant difference between Periods 1 and 2, Periods 1 and 3, and Periods 2 and 3. For open-book untimed, closed-book, MCQ and in-person proctored assessments, there was a statistically significant difference between Periods 1 and 2 and Periods 1 and 3, but not between Periods 2 and 3. For presentations, there was no significant difference between Periods 1 and 3, although the other two pairings were significant. For projects and assignments, the only significant difference was between Periods 2 and 3.

3.2. Assessment Weightings

To measure the difference in weightings between different assessment types, respondents were asked how much their final examination was worth in comparison with continuous assessment in their modules. They had to select a percentage band, which was 20% wide (e.g., 100% final exam, 80–99% final exam, etc.). The results are shown in Figure 2. As can be seen in the chart, for Period 1, the most common approach was to have between 60–79% of the marks for the module assigned to the final examination (n = 68), which dropped during Periods 2 and 3 (n = 46) but remained the most common weighting throughout, although 40–59% became a very close second (n = 42) in Period 3. When a Friedman test was conducted on this data, there was a statistically significant difference between the weightings of final examinations and continuous assessments, measured across the three time periods (χ2(2) = 27.752, p < 0.001). Post-hoc analysis was conducted using Wilcoxon signed-rank tests with a Bonferroni correction applied, and this found statistically significant results between Period 1 and Period 2 (Z = −4.112, p < 0.001) and Period 1 and Period 3 (Z = −3.879, p < 0.001), but not between Period 2 and Period 3 (Z = −0.866, p = 0.387).
In fact, if we consider the changes for any individual respondent, shown in Figure 3, we find that 112 respondents did not alter how much their final exam was worth in relation to the continuous assessment at any point during the three time periods under investigation. Very few put more weight on final examinations between Periods 1 and 2, with most who made changes moving towards a greater weighting on continuous assessment (either a 1–19% increase or 60–79% increase, suggesting they either made minor weighting adjustments or substantial ones). Between Periods 2 and 3, the biggest adjustments were plus or minus 1–19% change in weightings, with respondents almost equally adjusting towards more final examination weighting or more continuous assessment weighting. Overall, looking at the change between Periods 1 and 3, there was a noticeable shift back towards increased weighting for continuous assessment, although again the most common response was a 1–19% increase in weighting.
There were 73 additional comments on this question, many describing specifics about their approach, but some elaborating on reasons for the choices made in terms of assessment at the various stages. The most common themes are highlighted in Table 4 below, along with a sample comment made under each theme.

3.3. Grade Distribution

Respondents were then asked if they observed any differences in the distribution of grades within their modules during Period 3. Only just over 3% felt that there was a big difference in grade distribution that year, with 22% stating that there was no difference at all and the remaining responses falling somewhere between these poles, as shown in Table 5.
When asked for their interpretation of any differences they observed, there were 89 further comments made. The most common themes to emerge are shown in Table 6.

3.4. Satisfaction with Assessments

Respondents then rated their satisfaction levels with their assessments during the three time periods in question. The results are shown in Figure 4. It can be seen that three-quarters of respondents to this question were either satisfied or very satisfied with their assessments during Period 1. This dropped as low as two-fifths of respondents for Period 2 and recovered to 55% for Period 3. When a Friedman test was conducted, there was a statistically significant difference between satisfaction levels with assessments over time (χ2(2) = 84.874, p < 0.001). Post-hoc analysis was conducted using Wilcoxon signed-rank tests with a Bonferroni correction applied, and this found statistically significant results between each of the three time periods when compared with each other in terms of respondent satisfaction with their assessments (Period 1 vs. Period 2: Z = −7.333, p < 0.001; Period 2 vs. Period 3: Z = −4.564, p < 0.001; Period 1 vs. Period 3: Z = −5.286, p < 0.001).

3.5. Easiest/Hardest Aspects of Mathematics to Assess Online

Finally, respondents were asked two more general questions about assessment in mathematics: namely, what aspects of mathematics they considered to be most difficult and easiest to assess online. The most common themes to emerge from the 123 responses to each question are shown in Table 7.
Table 8 and Table 9 show the most common themes under “hardest to assess” and “easiest to assess”, respectively, along with the frequency of the theme and some sample comments made by respondents.

4. Discussion

In this study, we explored the assessment approaches taken by mathematics lecturers at three distinct periods to determine whether the forced inclusion of online assessment as a result of the COVID-19 campus closures might alter their opinions towards diversifying their assessment longer-term. Assessment in mathematics typically over-emphasises replication of content (facts) and skills (techniques) (Pegg, 2003). Assessment design should be heavily influenced by the mathematics that is considered most important for students to learn (Stacey & Wiliam, 2012). The respondents in this sample conducted almost no online teaching in Period 1, exclusively online teaching in Period 2, and almost all online teaching in Period 3. As such, the latter two periods did not reflect the “normal” situation for many of them but instead acted as a unique snapshot of time whereby they had to use online assessment.

4.1. What Assessment Types Were Used by Mathematics Lecturers Before the Pandemic, During the Initial University Closures, and During the Academic Year 2020/2021?

Our first research question explored the assessment types used by mathematics lecturers during the three periods under investigation. We found statistically significant changes in assessment types used by our sample in relation to 9 of the 12 assessment types investigated. We established that our sample was typical of the mathematics lecturer population during Period 1, in that 70% of respondents used closed-book examinations and 57% used assignments, both of which would be commonly used mathematics assessment approaches (Iannone & Simpson, 2022). By Period 3, open-book timed examinations had replaced the more popular closed-book examinations, even more so than in Period 2, when many respondents did not have sufficient time to implement such changes in their assessment. The findings clearly show a broader range of assessment modes in use in Period 3. After open-book timed assessments, projects, assignments and MCQs are the most popular modes of assessment. This is of interest as, according to Niss (1998), relaxing timing restrictions for students allows one to assess a broader range of concepts and skills. Suurtamm et al. (2016) say that implementing a range of assessment approaches allows students multiple opportunities to utilise feedback and demonstrate their learning. When a lecturer relies on one type of assessment, students often become experts at predicting likely assessment areas and choose their areas of revision accordingly (Burton & Haines, 1997). Seeley (2005) recommends designing assessments by “incorporating problem-solving, open-ended items, and problems that assess understanding as well as skills”, making assessment an integral part of teaching and learning. Of the three assessment types where no significant difference was observed, oral assessments were the most commonly used (by just over 20% of respondents). Although this was an under-utilised assessment type, it does appear to have been one that was able to be used across all three time periods, regardless of whether the assessment was in-person or online. This makes it an assessment type worthy of further consideration in relation to mathematics (Iannone et al., 2020).

4.2. Were There Changes to the Weightings Given to Final Examinations Versus Continuous Assessment During These Time Periods?

The proportion of lecturers who gave 80–100% final exams remained steady between Periods 1 and 3. There was a shift from the proportion who offered a 60–79% final exam towards a 40–59% exam. We do observe a statistically significant increase in lecturers favouring 100% continuous assessment in Period 3 when compared with Period 1. The number of those in favour of a 100% continuous assessment is in line with those in favour of a 100% final exam.

4.3. Did Mathematics Lecturers Observe Any Changes in Grade Distribution Within Their Modules During These Time Periods?

Overall, lecturers did not observe a vast difference in grade distribution between the three periods in question, although they gave varying reasons for this. The most prevalent theme was that of the type of examination that the students undertook. Some comments related this to a higher weighting on continuous assessment (“Grades are possibly higher due to the large amount of coursework/continuous assessment”), while others referenced the “binary right/wrong marking” of some online assessment tools, particularly in relation to multiple-choice questions. A number of respondents spoke of the impact of using open-book examinations where previously they would have had closed-book (“I don’t think I would have the same number of A’s if it was the usual proctored exam”). However, opinions were mixed on whether this was more or less difficult for the students, with one respondent stating, “open book exams plus timed online format skewed harder than in previous years as no book answers” while another felt that “open book made it easy for students to look at worked examples”. The next most common theme was that of copying—either another student’s work or using an online tool to cheat. Respondents observed that “cheating online is very easy” and that “students adapt to online exams and are more likely to cheat”. Conversely, one respondent felt that “it doesn’t appear that there was more cheating than there usually is”; however, this respondent stated that their module was fully assessed by continuous assessment both before the pandemic and during the academic year 2020/2021. Additionally, several respondents remarked on how, although they suspected there were students cheating, there was not as big a difference in the grade distribution as might have been expected because “there was probably cheating, but at the same time some students were less motivated”, and others echoed approaches similar to that in (Sarmiento & Prudente, 2019). Another strong theme to emerge was that of an increased divide in student performance, with many respondents pointing out how difficult it was for some of their students (More bimodal—perhaps reflecting the difficulties a subgroup of students had with online learning”). Several respondents felt that, although it was more difficult for some students, others coped well (“I feel that some students were not able to concentrate online. The high achievers were able to perform well in both situations”), or in fact that some students excelled in this situation (Struggling students withdrew but focused students had more time to study”). Several respondents also mentioned student engagement as a reason for greater disparity in grades, stating that “some students didn’t/couldn’t engage with the online environment” and “engagement in sessions was not as high during online learning as students were not as accountable as they would be in-person”. This could partly be attributed to the pandemic conditions, in which students from lower socio-economic backgrounds, or those with caring responsibilities, were impacted more in terms of engaging with online learning (Ní Fhloinn & Fitzmaurice, 2021). Finally, a number of respondents also alluded to their own lack of experience in setting different types of assessments as a reason for a differing grade profile, with some stating that they felt they made the assessment too easy as a result (“it was hard to pitch the level of difficulty on Webwork…I think that was too generous and the resultant marks were slightly higher”), while others felt the opposite had occurred (“Students scored higher in some classes because my questions were a bit too ‘traditional’”). Others simply observed that “there is a learning curve in terms of making a good online multiple-choice exam when all study aids are allowed. It is possible but requires some clever thinking”. This echoes the warning of Greenhow (2015) in relation to the creation of online assessments.

4.4. Were Mathematics Lecturers Satisfied with Their Assessment Approaches During These Time Periods?

Although the academics had increased the number of assessment methods in use, fewer respondents were satisfied with their assessments in Period 3 than before the pandemic in Period 1. The differences in satisfaction levels between all three time periods were significant, with the highest satisfaction ratings given pre-pandemic, and the next highest in Period 3. Overall satisfaction was significantly lower in Period 2, which was to be expected, as this was when lecturers had to pivot to new assessment types with very little time or resources to implement or plan these new assessments.

4.5. What Do Mathematics Lecturers Believe Are the Easiest and Most Difficult Aspects of Mathematics to Assess Online?

Our final research question was in relation to what aspects of mathematics lecturers believed were easiest or most difficult to assess online. The end of Period 3 represented a unique snapshot in time to obtain this information, whereby the majority of lecturing staff had been obliged to conduct online assessments in mathematics, regardless of their previous experience with online teaching (Ní Fhloinn & Fitzmaurice, 2021). We particularly wanted to investigate if mathematics lecturers perceive an online environment to be more conducive to the assessment of some mathematical areas than others. Respondents mostly spoke about the topic being assessed. The most common theme among “hardest to assess” was that of “proof”. Many respondents pointed out the difficulty of assessing theorems online in the same manner as they would have in a closed, proctored examination (“I believe there is much merit in asking students to learn to state theorems and to prove them. But this is tricky to assess online”). Other respondents noted that it was difficult to assess students’ ability to “reproduce a long technical proof”, although some respondents did question whether there was evidence of understanding previously in this approach (“Understanding (or, perhaps, if we are very honest, memory of) proofs.”), and observed that “it does make you think about key aspects of a proof not just memory and regurgitation”. The small number of instances of “proof” being mentioned in terms of being easiest to assess online leaned towards “standard questions that are not (easily) solved on webpages or computer software like Mathematica. Such as ‘show this function is continuous’, ‘show this subgroup is normal’” or “applications of theorems”.
Notably, “computation” appeared as the second-most common theme in terms of “most difficult to assess” and as the most common theme in terms of “easiest to assess”. Those who found it easiest to assess tended to make comments around the ease of using software to develop “questions that have parameters are relatively easy to vary without changing the difficulty of the question”. In contrast, many of those who found it hardest to assess linked this difficulty to the issue of students cheating, either by getting the solution from their classmates or online software (“I think any procedural questions for service maths modules…are difficult to assess online—the fact that there are so many websites where fully worked solutions to questions can be easily generated makes it very difficult to stand over the integrity of online assessments”). Similarly, some mentioned the difficulty of assessing basic computational techniques in open-book examinations (“As we only used open-book methods when assessments were online, it is hard to gauge if students have gained mathematical skills or if they are very dependent on copying procedures from worked examples”).
Yet more mentioned the drawbacks of “MCQs [multiple-choice questions] that involve lots of arithmetic/algebraic manipulation—it is difficult to give half marks/quarter marks to students who have done it partly correct but reached an incorrect result”.
Plagiarism, copying or “cheating” was the third most common theme in the aspects that are hardest to assess. This was frequently linked to either the “proof” or “computation” themes, but also often cited on its own, with one respondent stating that the hardest thing to assess was “anything where you want to ensure the work is the student’s own” and another that it was “anything you can find the answer to online”. The difficulty of detecting such activity was summed up by one respondent who said that “catching collusion can be tricky, as sometimes there is basically one obvious way to get to the right answer”. Another observed that “the fact that there are so many websites where fully worked solutions to questions can be easily generated makes it very difficult to stand over the integrity of online assessments”. This theme emerged a few times also among the comments in relation to the easiest aspects to assess; usually as an addendum to other comments, such as “However, there is no way to control or prevent students to contact each other during a timed exam they sit at home. Therefore, it is not clear if the exams assess if a particular student has understood the material”.
The theme of “understanding” occurred almost equally frequently in both the hardest and easiest aspects to assess online. Among those who deemed it hardest to assess online, many simply mentioned “conceptual understanding” without elaboration; however, a few mentioned that it was difficult to know “whether students understand definitions (because open book)” and “as ‘bookwork’ questions (e.g., proofs of theorems) are not possible…it can be difficult to assess a student’s understanding of the theory of any particular topic”. For the respondents who deemed “understanding” easiest to assess, again several simply mentioned “conceptual understanding”. However, a couple explained that “more abstract concepts are possibly a little easier to assess online as the understanding of the students can be tested in different ways” and that misconceptions could be exposed “using basic mcqs or true false. It doesn’t require any written mathematics and can expose key misunderstanding. Although why those exist might require alternative approaches”.
The theme “online solving” made specific reference to students using apps, websites or software to solve the questions they were asked in their assessments and only appeared in relation to the hardest aspects to assess. One respondent linked this to a need to consider what it is that is assessed and the purpose of such assessment, even outside the need to assess online (“The availability of online computational software and problem-solving sites makes it very difficult to assess students online. I don’t see any way around this without radically altering what we assess for. To some extent, this might be worth discussing even in the absence of online education”).
The three other most common themes of “theory”, “skills” and “reasoning” again appeared only in the hardest aspects to assess online. Respondents did not elaborate much on any of these themes, mostly just mentioning “theoretical content”, “basic skills” or “mathematical reasoning”. However, one respondent did elaborate in terms of reasoning to explain that “the problem I had was assessing their ability to think with insight into unfamiliar problems”.
In terms of the easiest aspects to assess online, the second-most common was “none” with no further comments made on this. A number of respondents mentioned “multiple-choice questions (MCQs)”, stating “multiple choice questions and quizzes can be easily developed and completed which does help with assessment between sessions”. However, another respondent cautioned that “aspects that can be validly assessed by multichoice questions lend themselves to being tested in hands-on mode … but the design of good MCQs is very challenging, and not all knowledge and skills can be covered”.
Finally, both “programming” and “statistics” were mentioned a number of times as being the easiest to assess online. In relation to the former, one respondent observed that “programming…, by design, has been assessed online for a very long time—anything else wouldn’t really make sense”. For statistics, another respondent suggested “giving students real data to work with and setting them online questions based on their individual data and/or project work. Students are very engaged too as they feel it is useful”.

5. Conclusions

To conclude, only one significant change in assessment approach was observed across the three periods of study; a move from closed- to open-book timed examinations. Academic integrity was clearly a strong consideration of the lecturers who participated in this study. For them, this refers to cheating and significantly influences the assessment strategy they adopt. This echoes the findings of Henley et al. (2022), whose large-scale survey of mathematics departments across the UK and Ireland found academic misconduct on the part of students to be of significant concern over the same time period as this study. The authors state that a community-wide approach will be necessary moving forward if open-book online assessments are to be continued. They found that increasing the time available for students to complete the assessments remotely led to increases in instances of plagiarism and cheating (Henley et al., 2022). Assigning randomised and personalised assessments is one method suggested in recent studies for reducing this misconduct (Eaton & Turner, 2020; Sarmiento & Prudente, 2019). While this may sound very laborious and time-consuming, Henley et al. (2022) state that this is not necessarily true. Respondents to their survey listed several methods of achieving this, including, for example, LaTeX Macros, which can be used to create personalised mathematical problems and generate random values for assessments. Online proctoring may reduce incidents of cheating (Lancaster & Cotarlan, 2021) but then result in negative mental health outcomes for others (Eaton & Turner, 2020).
Moving away from closed-book, proctored assessments in any capacity is likely to lead to grade inflation for a variety of reasons. Our findings echo those of Iannone and Simpson (2022), who in 2021 found that there is little variety in the assessment strategies of mathematics departments in Higher Education institutions in the UK. They too found that the weighting of final assessments was shifting towards the inclusion of other methods, but progress is slow. Despite the unprecedented visit of a pandemic and a drastic sudden move to remote teaching and assessment, assessment in mathematics remained largely unchanged. Research conducted since the pandemic provides evidence that remote assessment of mathematics posed a momentous challenge for many lecturers during this unprecedented period (Cusi et al., 2023). More worryingly, a study conducted with almost 470 university mathematics lecturers in Kuwait and the UK during the same time period as reported in this study found that the lecturers were very much unconvinced of the merits of online assessment in mathematics and would likely revert to their tried and trusted methods when restrictions were lifted (Hammad et al., 2025). Continued professional development programmes that support the effective implementation of practical and fair assessment practices and provide practical strategies and hands-on training are necessary if progress forward instead of backwards is to be sustained (Cusi et al., 2023).
The future of mathematics education holds great promise for technological innovations and the integration of Generative Artificial Intelligence (GenAI). Personalised learning is expected to become more widespread, with adaptive learning platforms and intelligent tutoring systems analysing students’ needs and providing tailored instruction and assessment. This study was undertaken when the capacities of GenAI were considerably weaker than they are at the time of writing, and given the concerns expressed by lecturing staff in relation to academic integrity at this earlier time, we feel that this is likely to be at the forefront of any decisions taken in relation to new assessment approaches to mathematics in the future. A follow-up mixed-methods study is planned to investigate the current status of assessment in mathematics globally, with the emergence of GenAI and its availability to both lecturers and students.
There are a number of limitations which should be taken into account in relation to this study. Firstly, 36% of respondents were based in Ireland, with a total of 88% of respondents based in Europe, meaning that the results can only be considered reflective of practice in European countries and may not be generalisable further afield. In addition, the survey was only available in English; it was distributed and conducted online via mailing lists and advertisements at relevant mathematics education conferences. As a result, it is unknown how representative it is of the general population of mathematics lecturers in higher education.

Author Contributions

Conceptualization, E.N.F. and O.F.; methodology, E.N.F. and O.F.; validation, E.N.F. and O.F.; formal analysis, E.N.F. and O.F.; investigation, E.N.F. and O.F.; writing—original draft preparation, O.F.; writing—review and editing, E.N.F.; visualization, E.N.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board of Dublin City University (protocol code: DCUREC/2020/079; date of approval: 1 May 2020).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Survey Questions on Assessment

  • How much of your assessment was conducted online versus in-person this academic year (2020–2021)? (5-point Likert scale)
  • What types of assessment did you conduct? (Please tick all that apply)
Assessment TypeBefore the PandemicDuring Initial University Closures (March–June 2020)Academic Year 2020–2021
Open-book timed
Open-book untimed
Closed book
Multiple-choice questions (MCQ)
Online proctored
In-person proctored
Presentations
Oral assessments
Essays
Screencasts
Projects
Assignments
  • How much was your final exam worth (versus continuous assessment) in your modules?
% WeightingBefore the PandemicDuring Initial University Closures (March–June 2020)Academic Year 2020-2021
100% Final Exam
80–99% Final Exam
60–79% Final Exam
40–59% Final Exam
20–39% Final Exam
1–19% Final Exam
100% Continuous Assessment
  • Any comments on changes to your approach to assessment this academic year?
  • Was there a difference in the distribution of grades within your modules this academic year? (5-point Likert scale)
  • If you saw a difference, why do you think this was?
  • How satisfied were you with your assessments? (5-point Likert scale)
  • What aspects of maths do you think are most difficult to assess online?
  • What aspects of maths do you think are easiest to assess online?

References

  1. Barana, A., Marchisio, M., & Sacchet, M. (2021). Interactive feedback for learning mathematics in a digital learning environment. Education Sciences, 11(6), 279. [Google Scholar]
  2. Becher, T. (1994). The significance of disciplinary differences. Studies in Higher Education, 19(2), 151–161. [Google Scholar] [CrossRef]
  3. Burton, L., & Haines, C. (1997). Innovation in teaching and assessing mathematics at university level. Teaching in Higher Education, 2(3), 273–293. [Google Scholar]
  4. Challis, N., Houston, K., & Stirling, D. (2003). Supporting good practice in assessment in mathematics, statistics and operational research. Available online: https://www.advance-he.ac.uk/knowledge-hub/supporting-good-practice-assessment-mathematics-statistics-and-operational-research (accessed on 17 December 2024).
  5. Cusi, A., Schacht, F., Aldon, G., & Swidan, O. (2023). Assessment in mathematics: A study on teachers’ practices in times of pandemic. ZDM–Mathematics Education, 55(1), 221–233. [Google Scholar] [CrossRef]
  6. Davies, B., Crisan, C., Geraniou, E., & Smart, T. (2024). A department-wide transition to a new mode of computer-aided assessment using STACK. International Journal of Research in Undergraduate Mathematics Education, 10, 850–870. [Google Scholar] [CrossRef]
  7. Eaton, S. E., & Turner, K. L. (2020). Exploring academic integrity and mental health during COVID-19: Rapid review. Journal of Contemporary Education Theory & Research (JCETR), 4(2), 35–41. [Google Scholar] [CrossRef]
  8. Engelbrecht, J., & Harding, A. (2004). Combing online and paper assessment in a web-based course in undergraduate mathematics. Journal of Computers in Mathematics and Science Teaching, 23(3), 217–231. [Google Scholar]
  9. Fitzmaurice, O., & Ní Fhloinn, E. (2021). Alternative mathematics assessment during university closures due to COVID-19. Irish Educational Studies, 40(2), 187–195. [Google Scholar] [CrossRef]
  10. Gikandi, J. W., Morrow, D., & Davis, N. E. (2011). Online formative assessment in higher education: A review of the literature. Computers & Education, 57(4), 2333–2351. [Google Scholar]
  11. Greenhow, M. (2015). Effective computer-aided assessment of mathematics; Principles, practice and results. Teaching Mathematics and Its Applications, 34(3), 117–137. [Google Scholar] [CrossRef]
  12. Greenhow, M. (2019). Embedding e-assessment effectively. MSOR Connections, 17(3), 63. [Google Scholar]
  13. Hammad, S., Wood, R., Hunter, G., & Dimitriadis, C. (2025). University mathematics lecturers’ experiences of teaching and learning during COVID-19 Pandemic: A comparative study between Kuwait and United Kingdom. Eurasia Journal of Mathematics, Science and Technology Education, 21(3), em2603. [Google Scholar] [CrossRef]
  14. Henley, M., Grove, M., & Hilliam, R. (2022). University mathematics assessment practices during the COVID-19 pandemic. MSOR Connections, 20(3), 5–17. [Google Scholar] [CrossRef]
  15. Iannone, P., Czichowsky, C., & Ruf, J. (2020). The impact of high-stakes oral performance assessment on students’ approaches to learning: A case study. Educational Studies in Mathematics, 103, 313–337. [Google Scholar] [CrossRef]
  16. Iannone, P., & Simpson, A. (2011). The summative assessment diet: How we assess in mathematics degrees. Teaching Mathematics and Its Applications, 30(4), 186–196. [Google Scholar] [CrossRef]
  17. Iannone, P., & Simpson, A. (2022). How we assess mathematics degrees: The summative assessment diet a decade on. Teaching Mathematics and Its Applications: An International Journal of the IMA, 41(1), 22–31. [Google Scholar] [CrossRef]
  18. James, R. (2016). Tertiary student attitudes to invigilated, online summative examinations. International Journal of Educational Technology in Higher Education, 13, 1–13. [Google Scholar]
  19. Kerka, S., & Wonacott, M. E. (2000). ERIC clearinghouse on adult, career and vocational education: Practitioner file. ERIC. [Google Scholar]
  20. Lancaster, T., & Cotarlan, C. (2021). Contract cheating by STEM students through a file sharing website: A COVID-19 pandemic perspective. International Journal for Educational Integrity, 17(1), 3. [Google Scholar] [CrossRef]
  21. Lishchynska, M., Palmer, C., & Cregan, V. (2021). Online mathematics supports in lieu of university tutorials during COVID-19: Student and lecturer perspectives. Teaching Mathematics and Its Applications, 40(4), 417–435. [Google Scholar] [CrossRef]
  22. Liyanagamage, N., Asylguzhina, E., Vel, P., Khan, Z. R., Fernando, M., & Lancaster, T. (2025). Methodological and ethical approaches to studying contract cheating: A systematic literature review. Journal of Academic Ethics, 1–27. [Google Scholar] [CrossRef]
  23. Niss, M. A. (1998). Dimensions of geometry and assessment. In Perspective on the teaching of geometry for the 21st century: An ICMI study (pp. 263–274). Kluwer Academic Publishers. [Google Scholar]
  24. Ní Fhloinn, E., & Carr, M. (2017). Formative assessment in mathematics for engineering students. European Journal of Engineering Education, 42(4), 458–470. [Google Scholar] [CrossRef]
  25. Ní Fhloinn, E., & Fitzmaurice, O. (2021). Challenges and opportunities: Experience of mathematics lecturers engaged in emergency remote teaching during the COVID-19 pandemic. Mathematics, 9(18), 2303. [Google Scholar] [CrossRef]
  26. Ní Fhloinn, E., & Fitzmaurice, O. (2022, June 24–25). Getting to grips with online mathematics education during the COVID-19 pandemic [Conference session]. Science and Mathematics Education Conference (SMEC) 2022, Dublin, Ireland. [Google Scholar] [CrossRef]
  27. Pegg, J. (2003). Assessment in mathematics. In J. M. Royer (Ed.), Mathematical Cognition (pp. 227–259). Information Age Publishing Inc. [Google Scholar]
  28. Sangwin, C. (2013). Computer aided assessment of mathematics. OUP Oxford. [Google Scholar]
  29. Sangwin, C. (2019, February 6–10). Developing and evaluating an online linear algebra examination for university mathematics [Conference session]. Eleventh Congress of the European Society for Research in Mathematics Education, Utrecht, The Netherlands. [Google Scholar]
  30. Sarmiento, C. P., Lapinid, M. R. C., & Prudente, M. S. (2018). The effect of online homework in the performance of mathematics of accounting and finance students. Advanced Science Letters, 24(11), 7956–7960. [Google Scholar]
  31. Sarmiento, C. P., & Prudente, M. S. (2019). Preventing Homework copying through online homework in a math class. The Normal Lights, 13(1), 27–55. [Google Scholar] [CrossRef]
  32. Seeley, C. L. (2005, September). Assessing to learn and learning to assess. NCTM News. [Google Scholar]
  33. Stacey, K., & Wiliam, D. (2012). Technology and assessment in mathematics. In Third international handbook of mathematics education (pp. 721–751). Springer. [Google Scholar]
  34. Suurtamm, C., Thompson, D. R., Kim, R. Y., Moreno, L. D., Sayac, N., Schukajlow, S., Silver, E., Ufer, S., & Vos, P. (2016). Assessment in mathematics education: Large-scale assessment and classroom assessment. Springer Nature. [Google Scholar]
  35. Trenholm, S. (2007). A review of cheating in fully asynchronous online courses: A math or fact-based course perspective. Journal of Educational Technology Systems, 35(3), 281–300. [Google Scholar] [CrossRef]
  36. Valdez, M. T. C. C., & Maderal, L. D. (2021). An analysis of students’ perception of online assessments and its relation to motivation towards mathematics learning. Electronic Journal of E-Learning, 19(5), 416–431. [Google Scholar]
Figure 1. Responses to the question “What types of assessment did you conduct?” (n = 190). Respondents were permitted to tick more than one response, as appropriate.
Figure 1. Responses to the question “What types of assessment did you conduct?” (n = 190). Respondents were permitted to tick more than one response, as appropriate.
Education 15 00449 g001
Figure 2. Responses to the question “How much was your final exam worth (versus continuous assessment) in your modules?” (n = 178).
Figure 2. Responses to the question “How much was your final exam worth (versus continuous assessment) in your modules?” (n = 178).
Education 15 00449 g002
Figure 3. Changes in individuals’ module weightings between the Final Exam (FE) and Continuous Assessment (CA) between (i) pre-pandemic and March–June 2020 (ii) March–June 2020 and the academic year 2020–2021 and (iii) pre-pandemic and the academic year 2020–2021.
Figure 3. Changes in individuals’ module weightings between the Final Exam (FE) and Continuous Assessment (CA) between (i) pre-pandemic and March–June 2020 (ii) March–June 2020 and the academic year 2020–2021 and (iii) pre-pandemic and the academic year 2020–2021.
Education 15 00449 g003
Figure 4. Responses to the question: “How satisfied were you with your assessments?” where respondents could select one response from a five-point Likert scale as shown in the chart. Respondents were asked the question in relation to the three time periods: before the pandemic; from March to June 2020; and during the academic year 2020–2021.
Figure 4. Responses to the question: “How satisfied were you with your assessments?” where respondents could select one response from a five-point Likert scale as shown in the chart. Respondents were asked the question in relation to the three time periods: before the pandemic; from March to June 2020; and during the academic year 2020–2021.
Education 15 00449 g004
Table 1. Profile statistics of survey respondents (n= 190), showing their gender, age, years of experience teaching mathematics in higher education, and current employment status.
Table 1. Profile statistics of survey respondents (n= 190), showing their gender, age, years of experience teaching mathematics in higher education, and current employment status.
Number%
Gender
Male8746%
Female9952%
(Blank)42%
Age
20–29 years84%
30–39 years5227%
40–49 years6032%
50–59 years5026%
60+ years1910%
(Blank)11%
Experience teaching maths in higher education
0–1 year21%
2–3 years116%
3–5 years1810%
5–10 years2815%
10–15 years3921%
15–20 years179%
20+ years7540%
Employment status
Ph.D./Postdoc21%
Short-term contract (≤1 yr)42%
Long-term contract (>1 yr)2212%
Permanent16285%
Table 2. Results of Cochran’s Q tests conducted on each assessment type, where a p-value of p < 0.05 represents a statistically significant difference in the proportion of respondents using each assessment type.
Table 2. Results of Cochran’s Q tests conducted on each assessment type, where a p-value of p < 0.05 represents a statistically significant difference in the proportion of respondents using each assessment type.
Assessment TypeCochran’s Q Valuep-ValueOutcome
Open-book timed138.078<0.001Significant
Open-book untimed19.818<0.001Significant
Closed-book163.538<0.001Significant
MCQ27.167<0.001Significant
Online proctored44.4<0.001Significant
In-person proctored107.079<0.001Significant
Presentations15.826<0.001Significant
Oral Assessments2.5830.275Not significant
Essays4.5450.103Not significant
Screencasts4.7500.093Not significant
Projects22.067<0.001Significant
Assignments8.6150.013Significant
Table 3. Results of McNemar tests conducted on the assessment types that showed as statistically significant in Table Y. A Bonferroni adjustment was applied so that p-values below 0.017 were considered significant. P1 = Period 1, P2 = Period 2 and P3 = Period 3.
Table 3. Results of McNemar tests conducted on the assessment types that showed as statistically significant in Table Y. A Bonferroni adjustment was applied so that p-values below 0.017 were considered significant. P1 = Period 1, P2 = Period 2 and P3 = Period 3.
Assessment TypePeriods Comparedp-ValueOutcome
Open-book timedP1vP2<0.001Significant
P1vP3<0.001Significant
P2vP3<0.001Significant
Open-book untimedP1vP2<0.001Significant
P1vP3<0.001Significant
P2vP30.832Not significant
Closed-bookP1vP2<0.001Significant
P1vP3<0.001Significant
P2vP30.064Not significant
MCQP1vP2<0.001Significant
P1vP3<0.001Significant
P2vP30.307Not significant
Online proctoredP1vP2<0.001Significant
P1vP3<0.001Significant
P2vP3<0.001Significant
In-person proctoredP1vP2<0.001Significant
P1vP3<0.001Significant
P2vP30.092Not significant
PresentationsP1vP20.002Significant
P1vP30.864Not significant
P2vP3<0.001Significant
ProjectsP1vP20.027Not significant
P1vP30.021Not significant
P2vP3<0.001Significant
AssignmentsP1vP20.481Not significant
P1vP30.096Not significant
P2vP30.004Significant
Table 4. Responses given to the question “Any comments on changes to your approach to assessment this academic year?” (n = 73).
Table 4. Responses given to the question “Any comments on changes to your approach to assessment this academic year?” (n = 73).
ThemesNumber of ResponsesSample Comment
Cheating10Open book online assessments led to lots of cheating (e.g., Chegg)
No choice5The university rules did not allow changing the balance of exam vs. continuous assessment
Previous experience4Multiple choice exam last June in Linear Algebra was awful to create and very bad at assessing students—average mark was 90%. This year’s Stack based questions for quizzes and exams on Moodle was much more discriminating
Time-consuming4Assessment became much more difficult. In particular, the process of creating an exam and finding appropriate problems (which could not be solved trivially using online software or Math Stack Exchange) was extremely time-consuming
Table 5. Responses given to the question “Was there a difference in the distribution of grades within your modules this academic year?”, where 1 = no difference and 5 = a big difference (n = 181).
Table 5. Responses given to the question “Was there a difference in the distribution of grades within your modules this academic year?”, where 1 = no difference and 5 = a big difference (n = 181).
RankingNumber of Responses
140
252
354
429
56
Table 6. The five most common themes mentioned by respondents when asked, “If you saw a difference (in grades), why do you think this was?”.
Table 6. The five most common themes mentioned by respondents when asked, “If you saw a difference (in grades), why do you think this was?”.
ThemeNumber of OccurrencesSample Comments
Exam type40Grades are possibly higher due to the large amount of coursework/continuous assessment.
I don’t think I would have the same number of A’s if it was the usual proctored exam.
Copying19Students adapt to online exams and are more likely to cheat.
It doesn’t appear that there was more cheating than there usually is.
Student divide19More bimodal—perhaps reflecting the difficulties a subgroup of students had with online learning.
Struggling students withdrew but focused students had more time to study.
Engagement9Engagement in sessions was not as high during online learning as students were not as accountable as they would be in-person.
Some students didn’t/couldn’t engage with the online environment.
Lecturer experience9It was hard to pitch the level of difficulty on Webwork…I think that was too generous and the resultant marks were slightly higher.
Students scored higher in some classes because my questions were a bit too ‘traditional’.
Table 7. The most common themes that emerged when respondents were asked “What aspects of maths do you think are most difficult to assess online?” and “What aspects of maths do you think are easiest to assess online?” (n = 123).
Table 7. The most common themes that emerged when respondents were asked “What aspects of maths do you think are most difficult to assess online?” and “What aspects of maths do you think are easiest to assess online?” (n = 123).
Themes: Hardest to AssessFrequencyThemes: Easiest to AssessFrequency
Proof29Computation38
Computation26None13
Cheating24Understanding11
Understanding16MCQs8
Online solving13Cheating5
Theory12Proof5
Skills9Programming4
Reasoning7Statistics4
Table 8. The most common themes emerged when respondents were asked, “What aspects of maths do you think are most difficult to assess online?” along with sample comments under each theme (n = 123).
Table 8. The most common themes emerged when respondents were asked, “What aspects of maths do you think are most difficult to assess online?” along with sample comments under each theme (n = 123).
ThemeNumber of OccurrencesSample Comments
Proof29I believe there is much merit in asking students to learn to state theorems and to prove them. But this is tricky to assess online.
Computation26I think any procedural questions for service maths modules…are difficult to assess online—the fact that there are so many websites where fully worked solutions to questions can be easily generated makes it very difficult to stand over the integrity of online assessments.
Cheating24The fact that there are so many websites where fully worked solutions to questions can be easily generated makes it very difficult to stand over the integrity of online assessments.
Understanding16As ‘bookwork’ questions (e.g., proofs of theorems) are not possible…it can be difficult to assess a student’s understanding of the theory of any particular topic.
Online solving13The availability of online computational software and problem-solving sites makes it very difficult to assess students online. I don’t see any way around this without radically altering what we assess for. To some extent, this might be worth discussing even in the absence of online education.
Theory12Theoretical content
Skills9Basic skills
Reasoning7The problem I had was assessing their ability to think with insight into unfamiliar problems.
Table 9. Most common themes that emerged when respondents were asked “What aspects of maths do you think are easiest to assess online?” along with sample comments under each theme (n = 123).
Table 9. Most common themes that emerged when respondents were asked “What aspects of maths do you think are easiest to assess online?” along with sample comments under each theme (n = 123).
ThemeNumber of OccurrencesSample Comments
Computation38Questions that have parameters are relatively easy to vary without changing the difficulty of the question.
None13There is really no aspect of Maths that is easy to assess online.
Understanding11More abstract concepts are possibly a little easier to assess online as the understanding of the students can be tested in different ways.
MCQs8Multiple choice questions and quizzes can be easily developed and completed which does help with assessment between sessions.
Cheating5However, there is no way to control or prevent students to contact each other during a timed exam they sit at home. Therefore, it is not clear if the exams assess if a particular student has understood the material.
Proof5Standard questions that are not (easily) solved on webpages or computer software like Mathematica. Such as ‘show this function is continuous’, ‘show this subgroup is normal’.
Programming4Programming…, by design, has been assessed online for a very long time—anything else wouldn’t really make sense.
Statistics4Giving students real data to work with and setting them online questions based on their individual data and/or project work. Students are very engaged too as they feel it is useful.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ní Fhloinn, E.; Fitzmaurice, O. The Impact of the COVID-19 Pandemic upon Mathematics Assessment in Higher Education. Educ. Sci. 2025, 15, 449. https://doi.org/10.3390/educsci15040449

AMA Style

Ní Fhloinn E, Fitzmaurice O. The Impact of the COVID-19 Pandemic upon Mathematics Assessment in Higher Education. Education Sciences. 2025; 15(4):449. https://doi.org/10.3390/educsci15040449

Chicago/Turabian Style

Ní Fhloinn, Eabhnat, and Olivia Fitzmaurice. 2025. "The Impact of the COVID-19 Pandemic upon Mathematics Assessment in Higher Education" Education Sciences 15, no. 4: 449. https://doi.org/10.3390/educsci15040449

APA Style

Ní Fhloinn, E., & Fitzmaurice, O. (2025). The Impact of the COVID-19 Pandemic upon Mathematics Assessment in Higher Education. Education Sciences, 15(4), 449. https://doi.org/10.3390/educsci15040449

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop