MD Student Perceptions of ChatGPT for Reflective Writing Feedback in Undergraduate Medical Education
Abstract
1. Introduction
2. Materials and Methods
2.1. Study Design
2.2. Study Participants
2.3. Data Collection
2.4. Analysis
3. Results
3.1. Quantitative Comparison of Feedback Modalities
3.2. Qualitative Comparison of Feedback Modalities
3.3. Theme 1: Promoting Deeper Reflection Through Highlighting Areas of Improvement and Providing Guiding Questions
“The feedback from ChatGPT was thorough and insightful, offering a detailed analysis of my reflection. It effectively highlighted key strengths… while also identifying areas for improvement. I appreciated the specific questions posed by ChatGPT, as they prompted deeper reflection and encouraged me to consider how my experiences have shaped my professional identity.”
“I was pleasantly surprised by how detailed the feedback provided by ChatGPT was, and ChatGPT’s feedback provided more information regarding areas and steps for improvement. I especially liked how ChatGPT’s feedback provided guiding questions to help me engage in deeper reflection.”
“The ChatGPT feedback also was superior in affirming the strengths in my feedback, and demonstrated a profound appreciation and understanding of the significance of my reflection process… The ChatGPT feedback was also excellent in providing clear, meaningful critiques on how I can further challenge myself. It posed specific questions I could ask myself, or areas in my reflection that I could dig deeper in. It is truly as if the AI had a mind of its own…”
3.4. Theme 2: ChatGPT’s Misuse and Paraphrasing of Quotations
“The ChatGPT feedback was generally more detailed, however it was very formulaic and had a lot of “fluff.” Many quotations from my reflections were utilized, but the comments on these quotes mainly just paraphrased.”
“I also really disliked the direct quotes that ChatGPT pulled from my reflection—some of them are completely meaningless in the grand scheme of what I wrote in my reflection.”
3.5. Theme 3: Easily Identifiable AI Tone or “Voice”
“One critique I’d give is that the feedback still reads like stereotypical AI writing… there was a tendency for the AI to be unnecessarily verbose while simultaneously not sounding any more professional or intelligent. Certain everyday conversational words like “emphasize” or “underline” were replaced by less commonly used synonyms like “underscore”, and almost every verb was accompanied by an adverb.”
“I did think it still felt surprisingly very personalized and didn’t feel cold or robotic…if I did not know it was AI that wrote it, I would have believed it was comments from a tutor.”
3.6. Theme 4: Valuing LF’s Unique Perspective and Life Experiences
“Even if the feedback it presented sounds good on paper, one cannot discount the rich life experiences that my LF has, and the profound and personal ways in which my LF understands me from our many sessions together. This is something that ChatGPT will never be able to have, and as a result, I cannot confidently say that I would certainly prefer ChatGPT’s feedback over my LF’s-rather, I am uncertain about how I feel.”
“The most valuable thing about the feedback from my LFs is that it was written by actual people… whom I have grown to respect in their wealth of experience and often very wise perspective on the world…. It was an important experience for me to have my personal reflections read by actual people… And their feedback… actually pointed out things I didn’t see at the time but personally value highly.”
“I have known my LF since Fall 2023 and respect him greatly. His feedback, even just the few sentences, meant more to me because I knew it was coming from him and was reflective of his expertise in the field and observation of me as a student.”
4. Discussion
5. Conclusions
Supplementary Materials
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Gilson, A.; Safranek, C.W.; Huang, T.; Socrates, V.; Chi, L.; Taylor, R.A.; Chartash, D. How does ChatGPT perform on the United States Medical Licensing Examination (USMLE)? The implications of large language models for medical education and knowledge assessment. JMIR Med. Educ. 2023, 9, 45312. [Google Scholar] [CrossRef] [PubMed]
- Scherr, R.; Halaseh, F.F.; Spina, A.; Andalib, S.; Rivera, R. ChatGPT interactive medical simulations for early clinical education: Case study. JMIR Med. Educ. 2023, 9, 49877. [Google Scholar] [CrossRef] [PubMed]
- Liu, S.; Wright, A.P.; Patterson, B.L.; Wanderer, J.P.; Turer, R.W.; Nelson, S.D.; McCoy, A.B.; Sittig, D.F.; Wright, A. Using AI-generated suggestions from ChatGPT to optimize clinical decision support. JAMIA 2023, 30, 1237–1245. [Google Scholar] [CrossRef] [PubMed]
- Sandars, J. The use of reflection in medical education: AMEE Guide No. 44. Med. Teach. 2009, 31, 685–695. [Google Scholar] [CrossRef] [PubMed]
- Holden, M.D.; Buck, E.; Luk, J.; Ambriz, F.; Boisaubin, E.V.; Clark, M.A.; Mihalic, A.P.; Sadler, J.Z.; Sapire, K.J.; Spike, J.P.; et al. Professional identity formation: Creating a longitudinal framework through TIME (Transformation in Medical Education). Acad. Med. 2015, 90, 761–767. [Google Scholar] [CrossRef] [PubMed]
- Wald, H.S.; Davis, S.W.; Reis, S.P.; Monroe, A.D.; Borkan, J.M. Reflecting on reflections: Enhancement of medical education curriculum with structured field notes and guided feedback. Acad. Med. 2009, 84, 830–837. [Google Scholar] [CrossRef] [PubMed]
- Uygur, J.; Stuart, E.; De Paor, M.; Wallace, E.; Duffy, S.; O’Shea, M.; Smith, S.; Pawlikowska, T. A Best Evidence in Medical Education systematic review to determine the most effective teaching methods that develop reflection in medical students: BEME Guide No. 51. Med. Teach. 2019, 41, 3–16. [Google Scholar] [CrossRef] [PubMed]
- De la Croix, A.; Veen, M. The reflective zombie: Problematizing the conceptual framework of reflection in medical education. Perspect. Med. Educ. 2018, 7, 394–400. [Google Scholar] [CrossRef] [PubMed]
- Lee, G.B.; Chiu, A.M. Assessment and feedback methods in competency-based medical education. Ann. Allergy Asthma Immunol. 2022, 128, 256–262. [Google Scholar] [CrossRef] [PubMed]
- Dyment, J.E.; O’Connell, T.S. Assessing the quality of reflection in student journals: A review of the research. Teach. High. Edu. 2011, 16, 81–97. [Google Scholar] [CrossRef]
- Lim, J.Y.; Ong, S.Y.K.; Ng, C.Y.H.; Chan, K.L.E.; Wu, S.Y.E.A.; So, W.Z.; Tey, G.J.C.; Lam, Y.X.; Gao, N.L.X.; Lim, Y.X.; et al. A Systematic Scoping Review of Reflective Writing in Medical Education. BMC Med. Educ. 2023, 23, 12. [Google Scholar] [CrossRef] [PubMed]
- Shum, S.B.; Sándor, Á.; Goldsmith, R.; Bass, R.; McWilliams, M. Towards Reflective Writing Analytics: Rationale, Methodology and Preliminary Results. J. Learn. Anal. 2017, 4, 58–84. [Google Scholar] [CrossRef]
- Hanlon, C.D.; Frosch, E.M.; Shochet, R.B.; Buckingham Shum, S.J.; Gibson, A.; Goldberg, H.R. Recognizing Reflection: Computer-Assisted Analysis of First Year Medical Students’ Reflective Writing. Med. Sci. Educ. 2021, 31, 109–116. [Google Scholar] [CrossRef] [PubMed]
- Park, J. Medical students’ patterns of using ChatGPT as a feedback tool and perceptions of ChatGPT in a Leadership and Communication course in Korea: A cross-sectional study. J. Educ. Eval. Health Prof. 2023, 20, 29. [Google Scholar] [CrossRef] [PubMed]
- Tangadulrat, P.; Sono, S.; Tangtrakulwanich, B. Using ChatGPT for Clinical Practice and Medical Education: Cross-Sectional Survey of Medical Students’ and Physicians’ Perceptions. JMIR Med. Educ. 2023, 9, 50658. [Google Scholar] [CrossRef] [PubMed]
- Spooner, M.; Larkin, J.; Liew, S.C.; Jaafar, M.H.; McConkey, S.; Pawlikowska, T. “Tell me what is ‘better’!” How medical students experience feedback, through the lens of self-regulatory learning. BMC Med. Educ. 2023, 23, 895. [Google Scholar] [CrossRef] [PubMed]
- Eidt, L.B. Feedback in medical education: Beyond the traditional evaluation. Rev. Assoc. Medica Bras. 2023, 69, 9–12. [Google Scholar] [CrossRef] [PubMed]
- Hardavella, G.; Aamli-Gaagnat, A.; Saad, N.; Rousalova, I.; Sreter, K.B. How to give and receive feedback effectively. Breathe Sheff. Engl. 2017, 13, 327–333. [Google Scholar] [CrossRef] [PubMed]
- Hsieh, H.F.; Shannon, S.E. Three approaches to qualitative content analysis. Qual. Health Res. 2005, 15, 1277–1288. [Google Scholar] [CrossRef] [PubMed]
- Noble, C.; Young, J.; Brazil, V.; Krogh, K.; Molloy, E. Developing residents’ feedback literacy in emergency medicine: Lessons from design-based research. AEM Educ. Train. 2023, 7, 10897. [Google Scholar] [CrossRef] [PubMed]
Section | Content |
---|---|
Role Definition | Play the role of a medical school educator in a first-year reflection-based course on professional identity development. I will provide a student’s reflection and ask you to provide feedback. |
Instructions for Feedback | 1. Draw upon specific quotations from the reflection that represent good examples of the criteria listed below. |
2. Draw upon specific quotations that represent superficial reflection or criteria that were not thoroughly expressed. | |
3. Explain why these quotations were chosen. | |
4. Provide guiding questions to help the student probe deeper in their reflection. | |
5. Conclude with encouraging, validating words to motivate future reflection. | |
Note: Please omit section headings so the feedback reads as a personalized letter. | |
Criteria for Evaluating Reflection (along with associated textual examples) | Reference to past perspectives: Previously, prior to, before, I used to, earlier, until now, at first, initially, formerly. |
Challenge/discomfort: Suddenly, however, I was surprised, I felt, I didn’t know, I wasn’t sure, I was scared, I was overwhelmed, it was hard. | |
Pausing to reflect: I stopped, I paused, I thought, this made me think, I reflected, I noticed. | |
Learning something specific: I learned, I recognized, I realized, I understood, I became. | |
Future implementation: Since, now, going forward, I will, I am beginning to, I can see how. |
Statement No. | Feedback Dimension | Survey Statement |
---|---|---|
1 | Personalization | The feedback I received was personalized and showed a genuine concern for my learning. |
2 | Identifying Strengths and Areas for Improvement | The feedback helped me identify my strengths and areas that need improvement. |
3 | Motivates Openness to Future Feedback | The feedback has motivated me to be more open to future feedback. |
4 | Highlights Specific Observations | The feedback was specific and clear, while avoiding vagueness. |
5 | Balances Tone | The feedback maintained a good balance of tone, neither being too harsh nor overly congratulatory. |
6 | Provides Steps for Improvement | The feedback provided clear steps for how I can improve. |
7 | Considers Emotional Implications of Critical Feedback | The feedback considered the emotional impact of receiving critical comments. |
8 | Promotes Self-Regulation | The feedback has helped me in self-monitoring and evaluating my own learning independently. |
LF Mean (SD) | ChatGPT Mean (SD) | Mean Difference | Cohen’s d | One-Sided p | Two-Sided p | |
---|---|---|---|---|---|---|
Personalization | 4.00 (1.00) | 3.60 (1.06) | 0.40 | 0.39 | 0.186 | 0.373 |
Identifying Strengths and Areas for Improvement | 3.20 (1.21) | 4.13 (0.99) | −0.93 | −0.84 | 0.024 * | 0.048 * |
Motivates Openness to Future Feedback | 3.73 (0.80) | 3.27 (1.03) | 0.47 | 0.50 | 0.093 | 0.187 |
Highlights Specific Observations | 3.53 (0.83) | 3.53 (0.92) | 0.00 | 0.00 | 0.500 | 1.000 |
Balances Tone | 4.13 (0.92) | 3.93 (1.16) | 0.20 | 0.19 | 0.328 | 0.655 |
Provides Steps for Improvement | 2.67 (1.05) | 4.27 (0.88) | −1.60 | −1.65 | 0.001 * | 0.002 * |
Considers Emotional Implications of Critical Feedback | 3.87 (1.13) | 3.53 (0.99) | 0.33 | 0.32 | 0.194 | 0.388 |
Promotes Self-Regulation | 3.40 (1.12) | 3.53 (0.99) | 0.13 | −0.12 | 0.390 | 0.779 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Published by MDPI on behalf of the Academic Society for International Medical Education. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Haider, N.; Morjaria, L.; Sheth, U.; Al-Jabouri, N.; Sibbald, M. MD Student Perceptions of ChatGPT for Reflective Writing Feedback in Undergraduate Medical Education. Int. Med. Educ. 2025, 4, 27. https://doi.org/10.3390/ime4030027
Haider N, Morjaria L, Sheth U, Al-Jabouri N, Sibbald M. MD Student Perceptions of ChatGPT for Reflective Writing Feedback in Undergraduate Medical Education. International Medical Education. 2025; 4(3):27. https://doi.org/10.3390/ime4030027
Chicago/Turabian StyleHaider, Nabil, Leo Morjaria, Urmi Sheth, Nujud Al-Jabouri, and Matthew Sibbald. 2025. "MD Student Perceptions of ChatGPT for Reflective Writing Feedback in Undergraduate Medical Education" International Medical Education 4, no. 3: 27. https://doi.org/10.3390/ime4030027
APA StyleHaider, N., Morjaria, L., Sheth, U., Al-Jabouri, N., & Sibbald, M. (2025). MD Student Perceptions of ChatGPT for Reflective Writing Feedback in Undergraduate Medical Education. International Medical Education, 4(3), 27. https://doi.org/10.3390/ime4030027