Abstract
In this discussion paper based on preliminary data, the safety and other quality criteria of ChatGPT-4o-generated exercise plans for patients with type 2 diabetes mellitus (T2DM) are evaluated. The study team created three fictional patient profiles varying in sex, age, body mass index, secondary diseases/complications, medication, self-rated physical fitness, weekly exercise routine and personal exercise preferences. Three distinct prompts were used to generate three exercise plans for each fictional patient. While Prompt 1 was very simple, Prompt 2 and Prompt 3 included more detailed requests. Prompt 3 was optimized by ChatGPT itself. Three coaching experts reviewed the exercise plans for safety and other quality criteria and discussed their evaluations. Some of the exercise plans showed serious safety issues, especially for patients with secondary diseases/complications. While most exercise plans incorporated key training principles, they showed some deficits, e.g., insufficient feasibility. The use of more detailed prompts (Prompt 2 and Prompt 3) tended to result in more elaborate exercise plans with better ratings. ChatGPT-4o-generated exercise plans may have safety issues for patients with T2DM, indicating the need to consult a professional coach for feedback before starting a training program.
1. Introduction
Tailored exercise interventions and physical activity are cornerstones in managing type 2 diabetes mellitus (T2DM), providing benefits such as improved insulin sensitivity and reduced cardiovascular disease risk [1,2]. Designing exercise plans for patients with T2DM presents particular challenges. Factors such as secondary complications and individual preferences, among other things, must be considered to ensure safe training and encourage adherence to the exercise program [3]. Therefore, developing exercise plans for patients with T2DM requires expert knowledge and is time-consuming. Tailored exercise plans are not available for all patients with T2DM, especially not for those lacking access to experienced professional coaches. To overcome these problems, types of artificial intelligence (AI), such as Large Language Models (LLMs) might be useful.
LLMs such as ChatGPT have the potential to address health-related issues [4]. Launched in November 2022 by OpenAI, ChatGPT is an AI-powered LLM trained on extensive datasets, allowing it to generate human-like text [5]. According to OpenAI’s technical report, GPT-4 exhibits capabilities comparable to those of human experts in certain complex professional and academic tasks [6]. However, ChatGPT has received mixed reactions within academia and the broader scientific community, reflecting an ongoing debate about the benefits and risks associated with AI technology [7]. While ChatGPT offers significant potential, it also has certain limitations, such as the generation of fictitious references and artificial hallucinations [8,9]. Moreover, Li et al. [10] identified additional limitations, including the tendency to present inaccurate information in a convincingly authoritative manner.
Customer opinion on AI in healthcare and sports is shaped by (a lack of) trust, perceived benefits and ethical considerations [11,12]. According to the Technology Acceptance Model (TAM), people are more likely to use AI technology if they perceive it as useful and easy to use [13]. In healthcare, the potential of AI to improve diagnosis and treatment is widely recognized, but concerns remain, e.g., regarding data privacy, decision-making autonomy or the erosion of the therapist–patient relationship [12]. These concerns highlight the need to increase public trust by creating a framework that balances innovation with ethical considerations [14].
Evidence on the quality of ChatGPT-generated exercise plans remains limited, which is not surprising given the novelty of research in this area. Düking et al. [15] evaluated the quality of ChatGPT-generated exercise plans for healthy athletes. Their findings showed that the quality of exercise plans generated by ChatGPT for runners was highly dependent on the information provided to the LLM, and that the best rated exercise plan was not optimal.
Due to the increasing interest in AI and ChatGPT in the healthcare sector and particular challenges in creating exercise plans for T2DM patients, this article contributes to the following research questions:
- Are exercise plans generated by ChatGPT-4o safe for patients with T2DM?
- Do ChatGPT-4o-generated exercise plans align with the American Diabetes Association (ADA) and the American College of Sports Medicine (ACSM) recommendations for physical activity in T2DM patients [1,16]?
- How do the safety and overall quality of ChatGPT-4o-generated exercise plans vary based on the quality of the prompts used?
2. Materials and Methods
This study was conducted in accordance with the Declaration of Helsinki. To develop and evaluate ChatGPT-generated exercise plans, we adopted the methodology used in previous studies [15,17]. Specifically, we (i) defined three fictional patients and specific prompts entered in ChatGPT-4o which differed in terms of the information provided, (ii) defined the quality criteria for the assessment of ChatGPT-4o-generated exercise plans and (iii) involved experienced raters in assessing the generated exercise plans based on the defined criteria.
2.1. Generating Exercise Plans Using ChatGPT-4o
The free-of-charge ChatGPT-4o software (OpenAI, L.L.C., San Francisco, CA, USA) was used to create exercise plans tailored to individuals with T2DM.
To reflect a variety of users with T2DM, the study team created three fictional patient profiles that differed in terms of sex, age, body mass index, secondary complications, medication, self-rated physical fitness, weekly exercise routines and personal exercise preferences (Table 1).
Table 1.
Details of the fictional patients.
Using this information, the following prompts were used to generate exercise plans (see Table 2). Prompt 1 was very basic, while Prompt 2 was more specific. The third prompt was refined by ChatGPT itself in line with Giray’s prompt engineering guidelines [18] using Prompt 2 as the starting prompt.
Table 2.
Prompts used for the creation of exercise plans.
Using these prompts and each patient’s specific information, a total of nine exercise plans were generated on 16 September 2024.
The ChatGPT input for prompt engineering as well as all generated exercise plans are provided in Supplementary Material S1.
2.2. Assessment of the Quality of ChatGPT-4o-Generated Exercise Plans
Three authors (SA, LS, CB) participated in rating the exercise plans. All three authors are experienced professional training experts with academic backgrounds in sports and therapy sciences (holding at least a bachelor’s degree in sports science) and have acquired a qualification in diabetes management during their studies. All raters are members of the “Diabetes, sports and exercise” research group at the German Sport University Cologne and have contributed to multiple exercise research projects involving T2DM patients.
Quality criteria were defined in an initial roundtable discussion between the raters based on the American Diabetes Association (ADA) and the American College of Sports Medicine (ACSM)’s recommendations for physical activity in T2DM patients [1,16]. Where appropriate, detailed indicators were defined for each quality criterion (Table 3 and Table 4, left column). Quality criteria were classified as either ‘L: low’, ‘M: moderate’, ‘S: strong’ or ‘N/A: not applicable’. A rating of ‘low’ implies that the quality criterion was not sufficiently met or that something planned could potentially be dangerous for patients, ‘moderate’ that the quality criterion was partially met and ‘high’ that it was (almost) fully met. After the criteria had been individually and independently assessed by the raters (at home), they met for a joint discussion. In the case of divergent ratings, the decision was made by a majority (2:1) or, if all three raters had arrived at different ratings, the middle rating, i.e., ‘moderate’ was assigned.
Table 3.
Assessment of the ChatGPT-4o-generated exercise plans: safety issues.
Table 4.
Assessment of the ChatGPT-4o-generated exercise plans: training-related quality criteria.
The intraclass correlation coefficient (ICC) was calculated to estimate interrater reliability using IBM SPSS Statistics for Windows, version 30 (IBM Corp., Armonk, NY, USA). The ICC was 0.763 (95% CI: 0.669, 0.834), indicating ‘good’ agreement between the raters [19].
3. Results
3.1. Safety Issues
The safety ratings for the exercise plans are presented in Table 3. Several plans exhibited considerable safety gaps:
- The most common safety concern across all exercise plans was that users were not made aware of the need for a medical check-up prior to starting the training program, especially in cases of secondary conditions necessitating such a check-up (Patients 2 and 3) or when high-intensity workouts were recommended for sedentary patients [1].
Potential contraindications were generally taken into consideration, with only a few exceptions:
- High-intensity training was recommended for Patient 3 with proliferative retinopathy. High-intensity training is not recommended for patients with proliferative retinopathy due to the risk of triggering vitreous hemorrhage or retinal detachment [1].
All training plans included additional safety guidelines. However, some information was missing, e.g.:
- It was not always mentioned that patients with (insulin-treated) diabetes should monitor their glucose levels before, during and after exercise to avoid hypoglycemia [16].
- It was not always mentioned that patients with high blood pressure should avoid holding their breath during strength exercises to prevent exorbitantly high and dangerous blood pressure peaks [16].
Using more detailed prompts (Prompts 2 and 3) generally resulted in minimal quality improvement (Figure 1, Prompt 1: 4× L, 3× M, 2× S, Prompt 2: 2× L, 2× M, 5× S, Prompt 3: 3× L, 2× M, 4× S), primarily in terms of tailoring the plans for individuals with secondary complications by avoiding conflicts between training and potential contraindications and providing additional safety instructions.
Figure 1.
Frequency distribution of the ratings ‘low’, ‘moderate’ and ‘strong’ across all fictional patients depending on the prompts used.
3.2. Other Training-Related Quality Criteria
Training-relevant quality criteria beyond safety are presented in Table 4. Most exercise plans included a clear training goal, although this goal was not always very specific. The training content usually matched the broad goals provided, and most training plans adhered to the ADA’s recommendations for physical activity. All training plans included guidance on progressive training load increases. In many cases, the programs’ feasibility was questioned. Personal preferences were always taken into account. The most common shortcomings were:
- Training goals were not always closely aligned with SMART principles (specific, measurable, achievable, relevant, time-bound).
- Contrary to the ADA’s recommendations, no balance exercises were suggested for the older fictional patient (Patient 3).
- Missing details on the level of intensity, e.g., target training heart rate or whether to train to muscle failure during strength exercises or not.
- Many plans only included sample exercises, and it was sometimes unclear whether the user could or should add further exercises.
- The total training time did not match the proposed exercise volume.
- Many exercises seemed overwhelming, i.e., some complex bodyweight exercises seemed inappropriate for beginners.
Using more detailed prompts (Prompts 2 and 3) sometimes led to improvements (e.g., in terms of a defined training goal and/or an appropriate progressive increase in training volume and/or intensity). However, in some cases, the quality was obviously not changed or declined minimally (in terms of the feasibility of the program or its alignment with personal preferences) (Figure 1, Prompt 1: 5× L, 2× N/A, 10× M, 7× S, Prompt 2: 1× L, 15× M, 8× S, Prompt 3: 2× L, 13× M, 9× S).
4. Discussion
The aim of this article was to evaluate the safety and other quality criteria of ChatGPT-generated exercise plans for patients with T2DM by coaching experts.
In this exploratory study based on preliminary data, two distinct prompts (Prompt 1 and Prompt 2) were used, reflecting inputs that most users would likely enter in a similar manner. Additionally, a third prompt was refined through prompt engineering by ChatGPT to generate the highest achievable output quality.
Some exercise plans revealed significant safety concerns. There was no guidance on obtaining a medical check-up before starting the training program, where this would be appropriate. Many, but not all possible safety instructions were mentioned by ChatGPT. However, this could pose serious risks. The training plans were usually unsafe for the study’s fictional patients with diabetes, particularly for those with secondary diseases. Nearly all training plans adhered to important training principles. However, there were some deficiencies in accurately controlling training intensity and ensuring overall feasibility.
These findings are partially consistent with previous studies. Düking et al. [15] rated ChatGPT training plans for healthy runners, Washif et al. [17] for healthy strength athletes, while Dergaa et al. [20] assessed training plans for patients with various conditions (but only one of them with T2DM). All research groups found that ChatGPT-generated plans do not (yet) meet the standards of coaching experts, primarily due to limitations in complexity and the level of detail.
Individuals with T2DM should aim for at least 150 min of moderate-intensity physical activity or 75 min of vigorous-intensity physical activity per week, in addition to engaging in strength training at least twice a week. Older adults with diabetes should also consider flexibility and balance exercises [1,16]. While most exercise plans in the present study adhered to general practice guidelines, they often lacked instructions and only provided sample exercises. Guidance on how to perform exercises, especially complex resistance exercises, would be useful for patients and could help prevent potential injuries [20,21].
One positive aspect is that all exercise plans were tailored to the user’s personal preferences, which is crucial since enjoyment plays an important role in adherence to the training program [22].
On the downside, there are some serious safety risks, some of which have already been mentioned. For example, ChatGPT recommended a vigorous-intensity exercise program for patients with proliferative retinopathy. This could have serious consequences, such as triggering vitreous hemorrhage or retinal detachment [16]. In many cases, important safety instructions were absent, e.g., guidance on monitoring glucose levels over an adequate post-exercise period, which is crucial for insulin-treated patients to reduce the risk of exercise-induced hypoglycemic events [23]. Similarly, Dergaa et al. [20] noted a lack of additional safety instructions, e.g., regarding glucose monitoring or foot care. However, they described only one case scenario with a T2DM patient who was only taking metformin and did not have other diseases/any secondary complications.
The safety deficiencies observed among the different cases in the present study raise the question of the ethical responsibility of AI developers (e.g., OpenAI) in ensuring that their tools provide safe and reliable recommendations for vulnerable groups like T2DM patients. It may be an ethical obligation to inform users of the limitations and risks of relying on AI-generated plans without professional oversight, including risks of unsafe recommendations. There is a need for built-in safeguards in AI systems to reduce the risks when used in medical or fitness contexts. Ultimately, this raises questions about liability in case of adverse events. Another pressing issue is data privacy when patient data are entered and processed. Clear rules and laws are needed to protect users [24]. Patients must be informed and consent to how their data are used, ensuring full transparency [25].
In the present study, the overall quality of exercise plans improved with more detailed prompts (Prompt 2 and Prompt 3). Düking et al. [15] used different prompts and also reported an increase in overall quality when providing more inputs. It can be assumed in this context that the majority of users are more likely to use simple prompts and may not use a complex prompt or prompt engineering to optimize their queries. As a result, many users may receive an exercise plan of lower quality. On average, prompt engineering did not lead to a further improvement in quality when comparing the exercise plans for Prompt 2 and Prompt 3. This could be explained by the reason that Prompt 2 already contained many details for the query. However, this approach could be helpful, especially for inexperienced users, to achieve a better result.
Future research should investigate the potential of creating tailored models for the target group. Making a “custom GPT” (customgpt.ai, 07.01.2025) could be distributed to the target group through links or QR codes, enhancing accessibility and usability. Future AI applications could be optimized by integrating user feedback loops or collaborative AI–human decision-making systems to increase the perceived level of optimal support and improve their ability to generate safer and more tailored exercise plans [26].
A strength of the present study is the inclusion of different fictional patients and of different prompts for each of these fictional patients which differed in the information provided to ChatGPT-4o. While this allowed a thorough analysis of ChatGPT-4o’s capabilities in generating exercise plans for patients with T2DM, the results of our study might differ for other use cases and other versions of ChatGPT or other LLMs. Future research needs to critically evaluate other LLMs and use cases for LLMs. Additionally, as we showed that results differ depending on the input information provided, the users of LLMs need to be educated on how to use LLMs safely and effectively in practice.
However, some limitations remain. The study is based on only three fictional patients, limiting the generalizability of the findings. Using fictional profiles instead of real patient data simplifies the creation of different patient scenarios (with no to several secondary diseases) but does not necessarily make it possible to capture the complexity of real-world cases. Since all raters are part of the research team, inherent biases may exist. Future studies should therefore be conducted using real patient data, involving a larger and more diverse sample and external raters (blinded to the prompts used).
5. Conclusions
Our results imply that ChatGPT-4o-generated exercise plans are not (yet) fully safe for patients with T2DM, especially not for those with secondary diseases/complications. While some plans effectively address various training-related aspects, others clearly fall short. Using more detailed prompts generally yields more elaborate exercise plans. However, based on our analysis, the results of this study imply that exercise plans created with ChatGPT-4o for individuals with T2DM are not able to provide fully individualized guidance tailored to the specific needs of the patients. When T2DM patients use ChatGPT-4o to generate exercise plans, it is recommended to involve a professional exercise coach to provide feedback before the training. It is also conceivable that ChatGPT can assist exercise coaches in designing exercise plans. In both scenarios, there is a high potential for AI–human collaboration to enhance exercise planning outcomes. Coaches should pay particular attention to whether a prior medical check-up is advisable and whether any precautions due to secondary diseases need to be considered. In addition, the training goal should be formulated in detail, and the selection of exercises should be critically reviewed. To improve the quality of the prompts used, patients or coaches should provide as much detail as possible and request that the current guidelines of the relevant medical societies should be followed.
Supplementary Materials
The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/sports13040092/s1, File S1: ChatGPT input for prompt engineering; ChatGPT-4o-generated exercise plans.
Author Contributions
Conceptualization, P.D. and C.B.; methodology, P.D. and C.B.; validation, C.B.; formal analysis, S.A., L.S. and C.B.; investigation, S.A., L.S. and C.B.; data curation, S.A. and L.S.; writing—original draft preparation, S.A. and C.B.; writing—review and editing, S.A., L.S., P.D. and C.B.; visualization, L.S. and C.B.; supervision, T.K. and C.B.; project administration, T.K. and C.B. All authors have read and agreed to the published version of the manuscript.
Funding
IST University of Applied Sciences Düsseldorf provided funding for the open access publication.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The original contributions presented in the study are included in the article/Supplementary Materials, further inquiries can be directed to the corresponding author.
Acknowledgments
Free-of-charge ChatGPT-4o (OpenAI, L.L.C., San Francisco, CA, USA) was used to create exercise plans tailored for individuals with T2DM.
Conflicts of Interest
Christian Brinkmann reports relationships with Abbott, Novo Nordisk and Lilly. The companies had no role in the design of the study, in the collection, analyses, or interpretation of the data, in the writing of the manuscript or in the decision to publish the results.
Abbreviations
The following abbreviations are used in this manuscript:
| ACSM | American College of Sports Medicine |
| ADA | American Diabetes Association |
| AI | Artificial intelligence |
| ICC | Intraclass correlation coefficient |
| LLM | Large Language Models |
| T2DM | Type 2 diabetes mellitus |
| TAM | Technology Acceptance Model |
References
- Kanaley, J.A.; Colberg, S.R.; Corcoran, M.H.; Malin, S.K.; Rodriguez, N.R.; Crespo, C.J.; Kirwan, J.P.; Zierath, J.R. Exercise/Physical Activity in Individuals with Type 2 Diabetes: A Consensus Statement from the American College of Sports Medicine. Med. Sci. Sports Exerc. 2022, 54, 353–368. [Google Scholar] [CrossRef]
- Shah, S.Z.A.; Karam, J.A.; Zeb, A.; Ullah, R.; Shah, A.; Haq, I.U.; Ali, I.; Darain, H.; Chen, H. Movement is Improvement: The Therapeutic Effects of Exercise and General Physical Activity on Glycemic Control in Patients with Type 2 Diabetes Mellitus: A Systematic Review and Meta-Analysis of Randomized Controlled Trials. Diabetes Ther. 2021, 12, 707–732. [Google Scholar] [CrossRef]
- Brinkmann, C. Road map for personalized exercise medicine in T2DM. Trends Endocrinol. Metab. 2023, 34, 789–798. [Google Scholar] [CrossRef] [PubMed]
- Biswas, S.S. Potential Use of Chat GPT in Global Warming. Ann. Biomed. Eng. 2023, 51, 1126–1127. [Google Scholar] [CrossRef] [PubMed]
- OpenAI. OpenAI: Models GPT-4. Available online: https://platform.openai.com/docs/models (accessed on 31 July 2024).
- OpenAI; Achiam, J.; Adler, S.; Agarwal, S.; Ahmad, L.; Akkaya, I.; Aleman, F.L.; Almeida, D.; Altenschmidt, J.; Altman, S.; et al. GPT-4 Technical Report. arXiv 2023, arXiv:2303.08774. [Google Scholar]
- Sallam, M. ChatGPT Utility in Healthcare Education, Research, and Practice: Systematic Review on the Promising Perspectives and Valid Concerns. Healthcare 2023, 11, 887. [Google Scholar] [CrossRef] [PubMed]
- Chen, T.-J. ChatGPT and other artificial intelligence applications speed up scientific writing. J. Chin. Med. Assoc. 2023, 86, 351–353. [Google Scholar] [CrossRef]
- Goddard, J. Hallucinations in ChatGPT: A Cautionary Tale for Biomedical Researchers. Am. J. Med. 2023, 136, 1059–1060. [Google Scholar] [CrossRef]
- Li, J.; Dada, A.; Puladi, B.; Kleesiek, J.; Egger, J. ChatGPT in healthcare: A taxonomy and systematic review. Comput. Methods Programs Biomed. 2024, 245, 108013. [Google Scholar] [CrossRef]
- Jud, M.; Thalmann, S. AI in digital sports coaching—A systematic review. Manag. Sport. Leis. 2025, 8, 1–17. [Google Scholar] [CrossRef]
- Witkowski, K.; Okhai, R.; Neely, S.R. Public perceptions of artificial intelligence in healthcare: Ethical concerns and opportunities for patient-centered care. BMC Med. Ethics 2024, 25, 74. [Google Scholar] [CrossRef] [PubMed]
- Ma, J.; Wang, P.; Li, B.; Wang, T.; Pang, X.S.; Wang, D. Exploring User Adoption of ChatGPT: A Technology Acceptance Model Perspective. Int. J. Hum. Comput. Interact. 2025, 41, 1431–1445. [Google Scholar] [CrossRef]
- Al-kfairy, M.; Mustafa, D.; Kshetri, N.; Insiew, M.; Alfandi, O. Ethical Challenges and Solutions of Generative AI: An Interdisciplinary Perspective. Informatics 2024, 11, 58. [Google Scholar] [CrossRef]
- Düking, P.; Sperlich, B.; Voigt, L.; van Hooren, B.; Zanini, M.; Zinner, C. ChatGPT Generated Training Plans for Runners are not Rated Optimal by Coaching Experts, but Increase in Quality with Additional Input Information. J. Sports Sci. Med. 2024, 23, 56–72. [Google Scholar] [CrossRef]
- ElSayed, N.A.; Aleppo, G.; Aroda, V.R.; Bannuru, R.R.; Brown, F.M.; Bruemmer, D.; Collins, B.S.; Hilliard, M.E.; Isaacs, D.; Johnson, E.L.; et al. 5. Facilitating Positive Health Behaviors and Well-being to Improve Health Outcomes: Standards of Care in Diabetes—2023. Diabetes Care 2023, 46, S68–S96. [Google Scholar] [CrossRef]
- Washif, J.A.; Pagaduan, J.; James, C.; Dergaa, I.; Beaven, C.M. Artificial intelligence in sport: Exploring the potential of using ChatGPT in resistance training prescription. Biol. Sport 2024, 41, 209–220. [Google Scholar] [CrossRef]
- Giray, L. Prompt engineering with ChatGPT: A guide for academic writers. Ann. Biomed. Eng. 2023, 51, 2629–2633. [Google Scholar] [CrossRef]
- Koo, T.K.; Li, M.Y. A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research. J. Chiropr. Med. 2016, 15, 155–163. [Google Scholar] [CrossRef]
- Dergaa, I.; Saad, H.B.; El Omri, A.; Glenn, J.M.; Clark, C.C.T.; Washif, J.A.; Guelmami, N.; Hammouda, O.; Al-Horani, R.; Reynoso-Sánchez, L.; et al. Using artificial intelligence for exercise prescription in personalised health promotion: A critical evaluation of OpenAI’s GPT-4 model. Biol. Sport 2024, 41, 221–241. [Google Scholar] [CrossRef]
- Takeuchi, K.; Nakamura, M. The optimal duration of high-intensity static stretching in hamstrings. PLoS ONE 2020, 15, e0240181. [Google Scholar] [CrossRef]
- Vilafranca Cartagena, M.; Tort-Nasarre, G.; Arnaldo, R. Barriers and facilitators for physical activity in adults with type 2 diabetes mellitus: A scoping review. Int. J. Environ. Res. Public. Health 2021, 18, 5359. [Google Scholar] [CrossRef] [PubMed]
- Heller, S.R.; Peyrot, M.; Oates, S.K.; Taylor, A.D. Hypoglycemia in patient with type 2 diabetes treated with insulin: It can happen. BMJ Open Diabetes Res. Care 2020, 8, e001194. [Google Scholar] [CrossRef] [PubMed]
- Wang, C.; Liu, S.; Yang, H.; Guo, J.; Wu, Y.; Liu, J. Ethical Considerations of Using ChatGPT in Health Care. J. Med. Internet Res. 2023, 25, e48009. [Google Scholar] [CrossRef] [PubMed]
- Tangsrivimol, J.A.; Darzidehkalani, E.; Virk, H.U.H.; Wang, Z.; Egger, J.; Wang, M.; Hacking, S.; Glicksberg, B.S.; Strauss, M.; Krittanawong, C. Benefits, limits, and risks of ChatGPT in medicine. Front. Artif. Intell. 2025, 8, 1518049. [Google Scholar] [CrossRef]
- Ravšelj, D.; Keržič, D.; Tomaževič, N.; Umek, L.; Brezovar, N.; AIahad, N.; Abdulla, A.A.; Akopyan, A.; Segura, M.W.A.; AlHumaid, J.; et al. Higher education students’ perceptions of ChatGPT: A global study of early reactions. PLoS ONE 2025, 20, e0315011. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).