1. Introduction
In the US, the transition from early university mathematics courses that are calculational in nature to proof-based courses—where students are expected to read, write, and assess proofs—has been a significant modern pedagogical challenge [
1]. Even students who were successful in prior calculational courses, such as calculus and differential equations, may struggle when confronted with a proof-based course [
2]. Historically and inter-institutionally, an alarmingly high number of students earn non-passing grades in their first proof-based course. At many institutions of higher education, a first course in proof writing aligns with one of two categories: (1) a course with a primary focus and learning objectives that address specific mathematical content, which we describe as a Content-Based Introduction to Proof (CBIP); or (2) a course for which the primary focus is on proof structure and techniques, which we call a Fundamental Introduction to Proof (FIP) [
3]. In CBIPs, proofs (and the associated linguistic and logical content) are taught through the lens of other mathematical content, for example, linear algebra, abstract algebra, or real analysis. Alternatively, FIPs teach the fundamentals of proof writing through symbolic logic, sets, relations, and elementary number theory.
In the early 1980s, there was a concerted national movement to address students’ difficulty in transitioning to proof-based courses by creating FIP courses [
4]. These courses aimed to teach students how to effectively communicate in the language of mathematics and, in particular, how to write formal proofs such as those required in upper-level courses. Today, most collegiate mathematics departments in the U.S. have incorporated some form of an FIP course as a requirement for baccalaureate math programs, although the syllabus and learning objectives of FIPs vary widely [
3,
4,
5]. Marty [
4] addressed the effectiveness of CBIP and FIP courses through a 10-year study at their institution. Longitudinal data, based on grade outcomes in future courses, were compared between student populations who took CBIP-type courses and student populations who took FIP-type courses. The findings, based on a population of approximately 300 students, led the author to suggest that it is more effective to focus on developing students’ approaches to mathematical content (as emphasized in FIPs) than to focus on the mathematical content itself. The study concluded that FIPs also increase students’ confidence and ability to take ownership in their mathematical maturation.
Developing best practices for teaching proof writing in the mathematical sciences has been the focus of much scrutiny and has proven to be a formidable task [
6,
7,
8,
9,
10,
11,
12,
13,
14]. Alcock’s study [
15], based on interviews with mathematicians teaching introductory proof material, addressed the complexity of the thought processes involved in proof writing. The article concludes with the suggestion that the transition to proof courses should address four independent thinking modes: instantiation, structural, creative, and critical modes. The study of Moore [
2] looked specifically at the cognitive difficulties for students transitioning to proofs at a large university, concluding that the three main difficulties are (a) foundational concept understanding, (b) mathematical language and notation, and (c) starting a proof. The FIP in our study focuses on skills in (b) and (c) through novel self-reflective writing. Consequently, we identified a previously unaddressed opportunity to directly analyze the effectiveness of transition-to-proof reflective writing strategies.
The present study focused on an FIP course we recently developed to bolster the success rate of students in the subsequent Introduction to Real Analysis class. The latter course is considered difficult by many students and instructors. The modified course sequence begins with (1) a prerequisite introductory linear algebra course, followed by (2) the optional FIP course, predicted to prime student success in (3) the subsequent introductory analysis course. This sequencing allows for a direct comparison of students who choose the path of linear algebra to real analysis and those who choose to take the FIP between these core courses. The newly developed FIP course includes active learning techniques that have been shown to increase the efficacy of teaching mathematical proofs [
4]. While building from the thinking modality work of Alcock [
15], our approach of using reflective writing as a framework for students to develop self-awareness of their understanding of their proof-writing skills represents a novel interventionist [
13] pedagogical approach compared to the traditional pedagogical approaches to transition-to-proof courses.
The goal of the present work was to assess the innovative use of student self-reflective writing-to-learn (WTL) exercises, scaffolded by a modality-based prompt and assessment rubric. Specifically, we developed a modality rubric that consists of a Likert-scale evaluation and field(s) for open-ended response self-evaluations for each of four modalities: mechanical (the mechanical modality corresponds to remembering, which is different from [
15] the instantiative modality, which corresponds to a deeper understanding of the definitions), structural, creative, and critical thinking modalities of proof writing. Students were also asked to evaluate their use of the type-setting program LaTeX. However, here, we focus on the conceptual elements of proof writing, that is, the mechanical, structural, creative, and critical modes of thinking (see
Appendix B for example modality rubrics and
Section 2 for detailed descriptions of the modalities). The framework of these modalities and the modality-based rubric are inspired by prior, related work [
8,
16,
17]. In weekly homework assignments, students were asked to reflect on their performance specifically in the context of their comfort with the modalities. With this structure, we posed three research questions:
Does having students write reflections in the interventional (FIP) course support their ability to be metacognitive about their own proof-writing processes?
Does having students write reflections in the interventional course impact their performance and success in that course?
Does having students write reflections in the interventional course impact their performance in the subsequent Introduction to Real Analysis course?
We addressed these questions over three semesters of the FIP course through analyzing students’ weekly self-reflections (both with Likert ratings and open-ended responses) and correlating with students’ grades in the prerequisite, interventional, and subsequent introductory analysis courses. As such, we addressed a gap in the research in mathematics education regarding the cognitive processes involved in proof writing [
1]. Specifically, we investigated the effectiveness of a novel approach that has students engage in reflective writing in a transition-to-proof course. While there is extensive research on reflective writing and student success, our novel contribution is combining the modality rubric, one of the best practices for teaching proof writing, with reflective writing.
While the focus of this work is on the impact of reflective writing on student achievement in an introductory proof-writing course, there is a related question of the impact of the new course as an intervention relative to a student’s overall success. We analyze the impact of the new course in a subsequent paper [
18] using data from a pre–post-assessment and analyzing students’ grades in the prerequisite course, the interventional course, and the subsequent course for students who took and did not take the interventional course. Preliminary results show a positive effect of the interventional course on student learning and success [
18].
The next section lays out the theoretical framework for our study design and methodologies. The Materials and Methods section describes the interventional course, the learning modalities, the instruments used, and the qualitative and quantitative methods used to analyze the student reflections. The Results section includes a discussion of student performance in the progression of courses as a function of the quality of their reflections, an analysis of their growth in metacognition, and an analysis of student reflections relative to the grader’s evaluation of student performance. Finally, the Discussion section contains a summary of our results, conclusions, and other considerations.
4. Results
4.1. Analyses of Reflections for Quality, Metacognition, and Variability across Cohorts
The three raters independently rated the body of each student’s written reflections over the 12 homework assignments to provide a single, holistic rating for each student’s work. In determining these ratings, we agreed that typical “exceptional” reflections explicitly addressed the prompt by explicating multiple modalities and consistently provided examples and/or reasoning. Reflections rated “acceptable” often explicitly responded to the prompt, sometimes provided specific examples of areas of struggle or success from the student’s proofs, and/or provided some reasoning for the self-rating. A holistic rating of “developing” was assigned if the student neglected to submit some responses and/or the reflections responded to the prompt in nonspecific ways. A rating of “incomplete” was assigned where a student did not complete the reflections or the writing was extremely vague with respect to the prompt. See
Table 1. Example responses characteristic of each quality level are shown in
Table 2. One participant completed over 50% of reflection opportunities, yet their reflections were rated as “incomplete” (instead of “developing”, as suggested by the completion parameter) because the majority of their reflections were a single word, which was insufficient for addressing the prompt. Nearly all participants whose reflections were rated as exceptional completed at least 11 of 12 assignments. However, the reflections of one participant who completed only nine assignments were rated as exceptional due to their overall quality.
As shown in
Table 2, students whose writing was rated as “developing” or “incomplete” skipped writing the reflections or wrote them using very vague language. They generally did not describe their struggles or successes and did not refer to the modalities or any specific elements of the proofs they wrote that week. Thus, their reflections did not provide good evidence that they understood the four modalities and could appreciate how to operationalize them in order to write their proofs.
4.2. Analysis of Student Growth in Metacognition
In this section, we analyze student growth in metacognition based on a longitudinal analysis of student reflective writing over the semester, as well as a longitudinal comparison of student ratings with the ratings of the grader, whom we consider an expert in this context.
4.2.1. Longitudinal Analysis of Student Reflective Writing
To frame our analysis of metacognition in the students’ reflective writing over the semester, we looked for evidence of their awareness of and ability to control their own thinking and learning processes, specifically their use of the four thinking modalities to self-regulate their learning of the proof-writing process, as well as evidence of metacognitive growth in their reflective writing over the semester. Students who were able to articulate their thinking processes through accurate use of the modalities to explain their approaches to the homework in their reflections and/or were able to clearly express the extent of their understanding of how the modalities supported their ability to write the homework proofs were judged as showing evidence of metacognition. All 15 students whose sets of reflections were rated exceptional or acceptable showed evidence of metacognition in one or more of the reflections they wrote.
As an example of metacognition in reflection, we highlight a few of Student 13’s reflections, which were selected for their typicality as well as their richness in detail. This student’s set of reflections was rated exceptional overall and showed early and consistent evidence of metacognition. In Homework 2, Student 13 wrote:
Mechanical: I think I did a better job at using symbols than homework 1, but I’m still not sure whether it’s greater than, as good, and hopefully not worse than homework 1…
(Student 13, Homework 2)
In the first clause of this statement (“I think I did a better job at using symbols than homework 1”), Student 13 shows awareness of improvement in their use of symbols compared to in the previous week’s homework. In the second clause (“but I’m still not sure whether it’s greater than, as good, and hopefully not worse than homework 1”), they also demonstrate awareness of the limits of their current understanding, mentioning what they struggled with or were still unsure about. Like all students whose reflections were rated exceptional , Student 13 cited specific elements of that week’s proof exercise that showed they understood how the modalities relate to the thinking involved in writing those proofs. See, for example, the rest of Student 13’s Homework 2 reflection:
…[Structural]: I think I was able to show off parts of the statement in the truth table in order to fully determine if two statements were equivalent, but there could be something that I could be missing from the tables. Creative: I was able to show some connection between statements in order to determine if statements were logically equivalent, but I feel like I wasn’t able to fully explain some of the statements as to why they were either true or false. Critical: I think I was able [to] show how statements were either true or false, and was able to show my thought process with the truth tables.
(Student 13, Homework 2)
In response to the creative modality, Student 13 was able to articulate an awareness of strengths (“I was able to show some connection between statements”) but also identify areas of improvement (“I feel like I wasn’t able to fully explain some of the statements”). Thus, Student 13 is demonstrating signs of metacognition in their reflections by being able to identify their own strengths and weaknesses in writing the proofs in Homework 2.
While Student 13’s set of reflections are representative of writing that was rated exceptional overall, it is noteworthy that only five students out of 36 across the entire data pool wrote reflections that were rated this highly. In total, 9 of the 10 students whose sets of reflections we rated as acceptable wrote some entries that were highly metacognitive and self-critical (39 out of 108 instances of reflection across 12 homeworks for 9 students), but they did not consistently, throughout the entire semester, ground their references to the modalities in actual performance or the particulars of that week’s proofs. In total, 8 of 10 students whose set of reflections were rated acceptable showed inconsistent evidence of progress in metacognition over the semester. However, two students whose reflections were rated acceptable showed consistent development in their ability to reflect metacognitively about proof writing over the semester. Student 9’s reflections, for example, are representative of such growth, and were selected to highlight here because the contrast between their early and later writings is vivid. Compare Student 9’s reflection on Homework 3 with what they wrote for their final reflection on Homework 12:
I believe that the proofs I provided used adequate detail and included logical connections between statements in this homework.
(Student 9, Homework 3)
I believe that I have made a lot of improvement on how to approach a proof problem. I think I really built the mechanical and [structural] modality being that I can see a proof and immediately be able to break it down into the multiple subsections in order to solve it in [its] entirety. However, going from one point to the next is where I think I struggle. I have a tendency to set up each part of the proof or create a form of a shell of what needs to be filled in and then I have a hard time filling in some of the gaps. I think the repetition of proofs in class and on the homework have allowed me to improve on doing so, however I think I need a bit more practice on my own over the winter break…
(Student 9, Homework 12)
In the earlier reflection on Homework 3, Student 9 wrote in generalities with no explicit reference to the modalities or to their process, whereas by Homework 12, Student 9 articulated strengths and weaknesses in their proof-writing process with explicit reference to the modalities. Thus, while Student 13, classified as writing exceptional reflections, showed signs of metacognition in their responses as early as Homework 2, Student 9, classified as writing acceptable reflections, showed development of metacognition over the semester.
4.2.2. Longitudinal Analysis of Student–Grader Differences
We examined the question of whether grader–student differences in rating decrease over time using Pearson’s Correlation test and found that as the semester progressed, student and grader ratings became more similar in the structural and creative modalities (
,
Table 3). A similar trend is evident for both the mechanical and critical modalities, but not significantly so. The negative Pearson Correlation values show that the linear relationship for all four modalities is negative: As time passed, the difference in ratings between graders and students decreased.
The symbol indicates that the p-value of significance is less than . Through analyzing student reflections as well as comparing student–grader differences over time, there is evidence of student growth in metacognition over the semester. Below, we will show that students who wrote high-quality reflections demonstrated strong performances in both the interventional course and the advanced course.
4.3. Comparison across Semesters
The proportion of the students classified as writing either high- or low-quality reflections was fairly consistent over the three semesters (see
Table 4). Notably, the relative proportions of reflection quality observed in the second iteration of the course, which coincided with an abrupt transition to emergency remote instruction in spring 2020 due to the COVID-19 pandemic, stand out in comparison to the semesters prior and following.
In summary, students were grouped by the quality of their reflection responses over the course of the semester. Student reflections showed both evidence of metacognition as well as growth in reflective writing and development of metacognition. Despite changes to the prompt in an effort to elicit better self-reflections and an abrupt transition to online instruction, the proportions of reflection quality remained fairly consistent over the three semesters.
4.4. Analyses of Quantitative Trends
We compared the achievement of participants in the prerequisite course (linear algebra), the interventional course (FIP), and the subsequent course (Introduction to Real Analysis) based on participants’ quality of self-reflections in the interventional course. Specifically, we analyzed math course grade data on the basis of reflection quality. We proceeded with two statistical perspectives: (1) comparisons of students’ course grades between multiple groups of written reflection quality and (2) beta linear regression analyses to evaluate the possible role(s) of linear algebra grades, FIP grades, and written reflection quality as predictors of course grades in Introduction to Real Analysis.
Course performance diverged between those students who had high-quality (exceptional or acceptable quality) reflections in the interventional course compared to those who had low-quality (developing or incomplete quality) reflections as shown in
Figure 2. This difference in performance between the reflection quality groups is more prominent in both the interventional and introductory analysis course grades, as compared to prerequisite core course grades.
Figure 2 illustrates that the performance differential between the students who wrote high-quality reflections and the students who wrote low-quality reflections increases with course progression. A non-parametric, individual samples test (Mann–Whitney U) was applied to the data, with reflection quality (high vs. low) as the group qualifier. Statistical results suggest that differences between course grades among those students (
) who wrote high- vs. low-quality reflections were significant in the interventional (
) and introductory analysis courses (
), and not significant in the prerequisite core course (
).
4.5. Reflection Quality and Prior Student Course Achievement as Predictors of Future Success
A beta linear regression test shows that reflection quality, treated as two category levels, is a significant predictor of the Real Analysis grade (, ). A separate beta linear regression test shows that the Linear Algebra grade (scale of 0–4) is not a significant predictor of the Real Analysis grade (). A third beta linear regression test shows that the FIP grade is a significant predictor of the Real Analysis grade (, p = 0.022), where the beta variable represents whether students passed the interventional course with a C grade or higher. In summary, whether we apply regression analysis or a group comparison approach to the quantitative data, we see similar results. Specifically, reflection quality during the FIP is a significant predictor of future success in the Real Analysis course, while grades in the prerequisite Linear Algebra course are not. Likewise, group comparison results suggest that differences between course grades among those students who wrote high- vs. low-quality reflections were significant in the interventional (FIP) and introductory analysis courses, and not significant in Linear Algebra. Together, these aligned findings reinforce our confidence that having students write reflections throughout the interventional FIP had a positive impact on students’ success in both that course and the subsequent introductory analysis course.
6. Conclusions
Our analysis of students’ written reflections structured around thinking modalities of proof writing showed that students who wrote high-quality reflections performed better in both the interventional course and the subsequent introduction to analysis course. Furthermore, student performance in the prerequisite course did not predict student performance in the interventional and introduction to analysis courses. It was not just that diligent students performed well both with reflective writing and in the math content of their courses. If that were the case, we would expect their performance in the prerequisite course to correlate with their categorization as writing high-quality reflections, which it did not. Thus, we conclude that repeated exposure to guided self-reflection using the lens of the modalities supports growth in the students’ awareness of their own abilities pertaining to proof writing. In particular, the modalities provided students with a framework for discerning what they understood and what they needed help with to understand. Our results support the potential power of repeated, modality-based self-reflection as a strategy to improve students’ ability to write better proofs, and thus impact outcomes in future proof-based courses.
While the current work highlights the impact of reflective writing to support students’ metacognition around proof writing, future work could focus on the best way to guide students toward more effective reflective writing, thus improving both their self-awareness and outcomes in future courses. For example, future work could investigate the role of the frequency and intensity of reflective writing activities on developing proof-writing abilities. Additionally, our work did not directly measure how the students used the modalities while writing their proofs. Further research into students’ metacognitive processes—specifically think-aloud-type methods that reveal students’ thought processes as they write proofs—could yield additional insights into common misconceptions and misconstruals that would help instructors tailor their lectures, activities, and assignments accordingly. Lastly, supporting instructors and TAs in understanding theory and practice around reflective writing could enable them to better help students develop reflective writing skills.