Retention Factors in STEM Education Identified Using Learning Analytics: A Systematic Review
Abstract
:1. Introduction
1.1. The Scope of This Study
- A summary of factors, sub-factors, and features contributing to retention learning analytics in STEM education is presented. It provides researchers with guidance regarding which factors and features have been explored using learning analytics in the context of the digital transformation in the higher education sector.
- The features measured for each of the retention factors are evaluated and discussed. In the review of LA studies, it demonstrates how each factor can be quantified based on available datasets from systems (e.g., LMS and SIS) to predict student likelihood of persisting.
- This review also highlights the features that significantly contributed to STEM retention. This will facilitate the feature selection process and improve exploratory modelling efficiency in LA studies.
1.1.1. Research Questions
2. Methodology
2.1. Search Strategy
2.2. Selection Criteria
2.2.1. Inclusion Criteria
- The articles were published in a peer-reviewed journal or conference;
- The study employed learning analytics techniques with empirical evidence;
- The study explored an online learning environment and examined data collected from a LMS or a similar online learning platform;
- The study subjects’ major should be within STEM field;
- The sample population should be from higher education institutions.
2.2.2. Exclusion Criteria
- The articles did not fulfil the eligibility criteria;
- The study did not contribute to the research questions.
2.3. Selection of Publications
3. Results
3.1. Type of Publication
3.2. Characteristics of Selected Studies
3.3. STEM Retention Factors, Sub-Factors and Features
3.3.1. Student Personal Characteristics
3.3.2. Student Enrolment Properties
3.3.3. Student Prior Academic Performance
3.3.4. Student Current Academic Performance
3.3.5. Student Behavioural Engagement
3.3.6. Student Academic Engagement
3.3.7. Course Design
4. Findings and Discussion
4.1. Retention Factors and Quantified Features Explored by LA in STEM Education
4.1.1. Student Personal Characteristics
4.1.2. Student Enrolment Properties
4.1.3. Student Prior Academic Performance
4.1.4. Student Current Academic Performance
4.1.5. Student Behavioural Engagement
4.1.6. Student Academic Engagement
4.1.7. Course Design
4.2. Recommendations for Future Research
- Seven retention factors and their sub-factors and features have been summarised. The power of each feature associated with each factor should be critically examined in future learning analytics studies. In addition, relationships between factors should be further evaluated.
- Features from static factors (student personal characteristics, student enrolment properties, and prior academic performance) demonstrated promising results in early detection of at-risk students. Learning analytics currently tends to explore dynamic data on the learning process leading to a snapshot of student learning patterns and predictions of learning outcomes for the purpose of providing intervention. However, these static factors and features are under explored in retention learning analytics studies; therefore, there is a need to further explore the potential of this data in LA for early prediction and/or even assist institutions in the phase of applicant selection to address the STEM retention challenge.
- The limited consideration towards the course design factor has been identified after reviewing the key publications. Most studies are student-focused, and the results emphasise that more attention and further research could be aimed at the pedagogical design of assessment, such as problem-based and cooperative learning to improve students’ interactions and maintain interest in STEM courses.
- Factor and feature selection and the data form denoted should direct studies in retention towards learning analytics. Combining features from different factors may yield better prediction results, though different data selection can result in varying predictive performance in identifying at-risk students [33]. Predictive model development should carefully select features and should avoid choosing purely one kind of retention factor. The results from one factor in learning analytics are far less accurate. [34]. A better performance model might be developed by combining other, even individually less effective, factors or features.
- The analysis may not be generalisable for all subjects in STEM disciplines, as the assessment and marking of each subject varies. Additionally, prediction models can also be affected by the pedagogical circumstances in the subject [15]. Future researchers could employ data from subjects in the same discipline that share similar pedagogy design and assessment format.
4.3. Limitations
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Almatrafi, O.; Johri, A.; Rangwala, H.; Lester, J. Retention and Persistence among STEM Students: A Comparison of Direct Admit and Transfer Students across Engineering and Science. In Proceedings of the 124th ASEE American Society for Engineering Education, Columbus, OH, USA, 25–28 June 2017; p. 18453. [Google Scholar]
- Sithole, A.; Chiyaka, E.T.; McCarthy, P.; Mupinga, D.M.; Bucklein, B.K.; Kibirige, J. Student Attraction, Persistence and Retention in STEM Programs: Successes and Continuing Challenges. High. Educ. Stud. 2017, 7, 46–59. [Google Scholar] [CrossRef]
- Sultana, S.; Khan, S.; Abbas, M.A. Predicting performance of electrical engineering students using cognitive and non-cognitive features for identification of potential dropouts. Int. J. Electr. Eng. Educ. 2017, 54, 105–118. [Google Scholar] [CrossRef]
- Iivari, N.; Sharma, S.; Ventä-Olkkonen, L. Digital transformation of everyday life—How COVID-19 pandemic transformed the basic education of the young generation and why information management research should care? Int. J. Inf. Manag. 2020, 55, 102183. [Google Scholar] [CrossRef] [PubMed]
- Anh, T.V.; Nguyen, H.T.T.; Linh, N.T.M. Digital Transformation: A Digital Learning Case Study. In Proceedings of the the 2019 The World Symposium on Software Engineering (WSSE), Wuhan, China, 20–23 September 2019; pp. 119–124. [Google Scholar]
- Martin, L. Foundations for Good Practice: The Student Experience of Online Learning in Australian Higher Education during the COVID-19 Pandemic; Tertiary Education Quality and Standards Agency: Melbourne, Australia, 2020; pp. 1–25. [Google Scholar]
- Francis, B.K.; Babu, S.S. Predicting Academic Performance of Students Using a Hybrid Data Mining Approach. J. Med. Syst. 2019, 43, 162. [Google Scholar] [CrossRef] [PubMed]
- Prieto, L.P.; Rodríguez-Triana, M.J.; Martínez-Maldonado, R.; Dimitriadis, Y.; Gašević, D. Orchestrating learning analytics (OrLA): Supporting inter-stakeholder communication about adoption of learning analytics at the classroom level. Australas. J. Educ. Technol. 2019, 35, 14–33. [Google Scholar] [CrossRef]
- Ferguson, R. Learning analytics: Drivers, developments and challenges. Int. J. Technol. Enhanc. Learn. 2012, 4, 304–317. [Google Scholar] [CrossRef]
- Gibson, D.C.; Ambrose, M.; Gardner, M. Self-organising maps and student retention: Understanding multi-faceted drivers. In Proceedings of the 32nd Annual Conference of the Australasian Society for Computers in Learning and Tertiary Education (ASCILITE 2015), Perth, Australia, 30 November–3 December 2019; pp. 112–120. [Google Scholar]
- Nieuwoudt, J.E.; Pedler, M.L. Student Retention in Higher Education: Why Students Choose to Remain at University. J. Coll. Stud. Retent. Res. Theory Pract. 2021, 1–24. [Google Scholar] [CrossRef]
- Muljana, P.S.; Luo, T. Factors contributing to student retention in online learning and recommended strategies for improvement: A systematic literature review. J. Inf. Technol. Educ. Res. 2019, 18, 19–57. [Google Scholar] [CrossRef] [Green Version]
- Spikol, D.; Ruffaldi, E.; Landolfi, L.; Cukurova, M. Estimation of Success in Collaborative Learning Based on Multimodal Learning Analytics Features. In Proceedings of the 17th International Conference on Advanced Learning Technologies (ICALT 2017), Timisoara, Romania, 3–7 July 2017; pp. 269–273. [Google Scholar]
- Finnegan, C.; Morris, L.V.; Lee, K. Differences by Course Discipline on Student Behavior, Persistence, and Achievement in Online Courses of Undergraduate General Education. J. Coll. Stud. Retent. Res. Theory Amp Pract. 2008, 10, 39–54. [Google Scholar] [CrossRef]
- Gašević, D.; Dawson, S.; Rogers, T.; Gasevic, D. Learning analytics should not promote one size fits all: The effects of instructional conditions in predicting academic success. Internet High. Educ. 2016, 28, 68–84. [Google Scholar] [CrossRef]
- Shafiq, D.A.; Marjani, M.; Habeeb, R.A.A.; Asirvatham, D. Student Retention using Educational Data Mining and Predictive Analytics: A Systematic Literature Review. IEEE Access 2022, 10, 72480–72503. [Google Scholar] [CrossRef]
- Liliana, L.; Santosa, P.I.; Kusumawardani, S.S. Completion factor in massive open online course in developing countries: A literature review in 2015–2021. World J. Educ. Technol. Curr. Issues 2022, 14, 456–472. [Google Scholar] [CrossRef]
- Shaikh, U.U.; Asif, Z. Persistence and Dropout in Higher Online Education: Review and Categorization of Factors. Front. Psychol. 2022, 13, 902070. [Google Scholar] [CrossRef] [PubMed]
- Moher, D.; Liberati, A.; Tetzlaff, J.; Altman, D.G. Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement. Ann. Intern. Med. 2009, 151, 264–269. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Umer, R.; Susnjak, T.; Mathrani, A.; Suriadi, S. A learning analytics approach: Using online weekly student engagement data to make predictions on student performance. In Proceedings of the 2018 International Conference on Computing, Electronic and Electrical Engineering (ICE Cube 2018), Quetta, Pakistan, 12–13 November 2019. [Google Scholar]
- Zacharis, N.Z. A multivariate approach to predicting student outcomes in web-enabled blended learning courses. Internet High. Educ. 2015, 27, 44–53. [Google Scholar] [CrossRef]
- Gkontzis, A.F.; Kotsiantis, S.; Panagiotakopoulos, C.T.; Verykios, V.S. A predictive analytics framework as a countermeasure for attrition of students. Interact. Learn. Environ. 2019, 3, 1028–1043. [Google Scholar] [CrossRef]
- Iatrellis, O.; Savvas, I.Κ.; Fitsilis, P.; Gerogiannis, V.C. A two-phase machine learning approach for predicting student outcomes. Educ. Inf. Technol. 2021, 26, 69–88. [Google Scholar] [CrossRef]
- Broadbent, J. Academic success is about self-efficacy rather than frequency of use of the learning management system. Australas. J. Educ. Technol. 2016, 32, 38–49. [Google Scholar] [CrossRef] [Green Version]
- Cohen, A. Analysis of student activity in web-supported courses as a tool for predicting dropout. Educ. Technol. Res. Dev. 2017, 65, 1285–1304. [Google Scholar] [CrossRef]
- Al-Shabandar, R.; Hussain, A.J.; Liatsis, P.; Keight, R. Analyzing Learners Behavior in MOOCs: An Examination of Performance and Motivation Using a Data-Driven Approach. IEEE Access 2018, 6, 73669–73685. [Google Scholar] [CrossRef]
- Sunday, K.; Ocheja, P.; Hussain, S.; Oyelere, S.S.; Balogun, O.S.; Agbo, F.J. Analyzing student performance in programming education using classification techniques. Int. J. Emerg. Technol. Learn. 2020, 15, 127–144. [Google Scholar] [CrossRef]
- Celis, S.; López, D.; Silva, J. Analyzing the influence of online behaviors and learning approaches on academic performance in first year engineering. In Proceedings of the 2nd Latin American Conference on Learning Analytics, Valdivia, Chile, 18–19 March 2019. [Google Scholar]
- Carter, A.S.; Hundhausen, C.D.; Adesope, O. Blending measures of programming and social behavior into predictive models of student achievement in early computing courses. ACM J. Trans. Comput. Educ. 2017, 17, 12. [Google Scholar] [CrossRef]
- Cirigliano, M.M.; Guthrie, C.D.; Pusic, M.V. Click-level Learning Analytics in an Online Medical Education Learning Platform. Teach. Learn. Med. 2020, 32, 410–421. [Google Scholar] [CrossRef] [PubMed]
- Howard, E.; Meehan, M.; Parnell, A. Contrasting prediction methods for early warning systems at undergraduate level. Internet High. Educ. 2018, 37, 66–75. [Google Scholar] [CrossRef] [Green Version]
- Nam, S.; Lonn, S.; Brown, T.; Davis, C.S.; Koch, D. Customized course advising: Investigating engineering student success with incoming profiles and patterns of concurrent course enrollment. In Proceedings of the 4th International Conference on Learning Analytics and Knowledge (LAK 2014), Indianapolis, IN, USA, 24–28 March 2014; pp. 16–25. [Google Scholar]
- Tsao, N.L.; Kuo, C.H.; Guo, T.L.; Sun, T.J. Data Consideration for At-Risk Students Early Alert. In Proceedings of the 6th IIAI International Congress on Advanced Applied Informatics, Hamamatsu, Japan, 9–13 July 2017; pp. 208–211. [Google Scholar]
- Bognar, L.; Fauszt, T. Different learning predictors and their effects for moodle machine learning models. In Proceedings of the 11th IEEE International Conference on Cognitive Infocommunications, Mariehamn, Finland, 23–25 September 2020; pp. 405–410. [Google Scholar]
- Pereira, F.D.; Oliveira, E.; Cristea, A.; Fernandes, D.; Silva, L.; Aguiar, G.; Alamri, A.; Alshehri, M. Early dropout prediction for programming courses supported by online judges. In Proceedings of the 20th International Conference on Artificial Intelligence in Education (AIED 2019), Chicago, IL, USA, 25–29 June 2019; pp. 67–72. [Google Scholar]
- Arakawa, K.; Hao, Q.; Deneke, W.; Cowan, I.; Wolfman, S.; Peterson, A. Early Identification of Student Struggles at the Topic Level Using Context-Agnostic Features. In Proceedings of the 53rd Annual ACM Technical Symposium on Computer Science Education (SIGCSE 2022), Virtual, 3–5 March 2022; pp. 147–153. [Google Scholar]
- Bravo-Agapito, J.; Romero, S.J.; Pamplona, S. Early prediction of undergraduate Student’s academic performance in completely online learning: A five-year study. Comput. Hum. Behav. 2021, 115, 106595. [Google Scholar] [CrossRef]
- Cannistrà, M.; Masci, C.; Ieva, F.; Agasisti, T.; Paganoni, A.M. Early-predicting dropout of university students: An application of innovative multilevel machine learning and statistical techniques. Stud. High. Educ. 2021, 47, 1935–1956. [Google Scholar] [CrossRef]
- Qiu, F.; Zhu, L.; Zhang, G.; Sheng, X.; Ye, M.; Xiang, Q.; Chen, P.K. E-Learning Performance Prediction: Mining the Feature Space of Effective Learning Behavior. Entropy 2022, 24, 722. [Google Scholar] [CrossRef]
- Russell, J.E.; Smith, A.; Larsen, R. Elements of Success: Supporting at-risk student resilience through learning analytics. Comput. Educ. 2020, 152, 103890. [Google Scholar] [CrossRef]
- De Almeida Neto, F.A.; Castro, A. Elicited and mined rules for dropout prevention in online courses. In Proceedings of the 2015 IEEE Frontiers in Education Conference (FIE 2015), El Paso, TX, USA, 21–24 October 2015; p. 7344048. [Google Scholar]
- Aguiar, E.; Chawla, N.V.; Brockman, J.; Ambrose, G.A.; Goodrich, V. Engagement vs. performance: Using electronic portfolios to predict first semester engineering student retention. In Proceedings of the 4th International Conference on Learning Analytics and Knowledge (LAK 2014), Indianapolis, IN, USA, 24–28 March 2014; pp. 103–112. [Google Scholar]
- Pereira, F.D.; Fonseca, S.C.; Oliveira, E.H.T.; Cristea, A.I.; Bellhauser, H.; Rodrigues, L.; Oliveira, D.B.F.; Isotani, S.; Carvalho, L.S.G. Explaining individual and collective programming students’ behaviour by interpreting a black-box predictive model. IEEE Access 2021, 9, 117097–117119. [Google Scholar] [CrossRef]
- DeMonbrun, R.M.; Brown, M.G. Exploring the relationship between the use of learning technologies and student success in the engineering classroom. In Proceedings of the 124th ASEE Annual Conference and Exposition, Columbus, OH, USA, 25–28 June 2017; p. 19716. [Google Scholar]
- Williams, L.; Titus, K.J.; Pittman, J.M. How Early is Early Enough: Correlating Student Performance with Final Grades. In Proceedings of the 5th Conference on Computing Education Practice (CEP 2021), Virtual, 7 January 2021; pp. 13–16. [Google Scholar]
- Martínez-Carrascal, J.A.; Márquez Cebrián, D.; Sancho-Vinuesa, T.; Valderrama, E. Impact of early activity on flipped classroom performance prediction: A case study for a first-year Engineering course. Comput. Appl. Eng. Educ. 2020, 28, 590–605. [Google Scholar] [CrossRef]
- Rafique, A.; Khan, M.S.; Jamal, M.H.; Tasadduq, M.; Rustam, F.; Lee, E.; Washington, P.B.; Ashraf, I. Integrating Learning Analytics and Collaborative Learning for Improving Student’s Academic Performance. IEEE Access 2021, 9, 167812–167826. [Google Scholar] [CrossRef]
- Nespereira, C.G.; Dai, K.; Vilas, A.F.; Redondo, R.P.D. Is the LMS access frequency a sign of students’ success in face-to-face higher education? In Proceedings of the 2nd International Conference on Technological Ecosystems for Enhancing Multiculturality (TEEM 2014), Salamanca, Spain, 1–3 October 2014; pp. 283–290. [Google Scholar]
- Jiang, X.; Liu, W.; Liu, J. Learning Analytics in a Blended Computer Education Course. In Proceedings of the International Workshop on Artificial Intelligence and Education (WAIE 2019), Singapore, 25–27 November 2019; pp. 6–12. [Google Scholar]
- Lacave, C.; Molina, A.I.; Cruz-Lemus, J.A. Learning Analytics to identify dropout factors of Computer Science studies through Bayesian networks. Behav. Inf. Technol. 2018, 37, 993–1007. [Google Scholar] [CrossRef]
- Mai, T.T.; Bezbradica, M.; Crane, M. Learning behaviours data in programming education: Community analysis and outcome prediction with cleaned data. Future Gener. Comput. Syst. 2022, 127, 42–55. [Google Scholar] [CrossRef]
- Feldman-Maggor, Y.; Blonder, R.; Tuvi-Arad, I. Let them choose: Optional assignments and online learning patterns as predictors of success in online general chemistry courses. Internet High. Educ. 2022, 55, 100867. [Google Scholar] [CrossRef]
- Koutcheme, C.; Sarsa, S.; Hellas, A.; Haaranen, L.; Leinonen, J. Methodological Considerations for Predicting At-risk Students. In Proceedings of the 24th Australasian Computing Education Conference (ACE 2022), Virtual, 14–17 February 2022; pp. 105–113. [Google Scholar]
- Vinker, E.; Rubinstein, A. Mining Code Submissions to Elucidate Disengagement in a Computer Science MOOC. In Proceedings of the 12th International Conference on Learning Analytics and Knowledge (LAK 2022), Virtual, 21–25 March 2022; pp. 142–151. [Google Scholar]
- Martínez-Navarro, Á.; Verdú, E.; Moreno-Ger, P. Mining Pre-Grade Academic and Demographic Data to Predict University Dropout. In Radical Solutions for Digital Transformation in Latin American Universities; Burgos, D., Branch, J.W., Eds.; Springer: Singapore, 2021; pp. 197–215. ISBN 978-981-16-3941-8. [Google Scholar]
- Sisovic, S.; Matetic, M.; Bakaric, M.B. Mining student data to assess the impact of moodle activities and prior knowledge on programming course success. In Proceedings of the 16th International Conference on Computer Systems and Technologies (CompSysTech ’15), Dublin, Ireland, 25–26 June 2015; pp. 366–373. [Google Scholar]
- Silva, J.; Romero, L.; Solano, D.; Fernandez, C.; Lezama, O.B.P.; Rojas, K. Model for predicting academic performance through artificial intelligence. In Proceedings of the International Conference on Computational Methods and Data Engineering, Sonipat, India, 30–31 January 2021. [Google Scholar]
- Marbouti, F.; Diefes-Dux, H.A.; Madhavan, K. Models for early prediction of at-risk students in a course using standards-based grading. Comput. Educ. 2016, 103, 1–15. [Google Scholar] [CrossRef] [Green Version]
- Chango, W.; Cerezo, R.; Romero, C. Multi-source and multimodal data fusion for predicting academic performance in blended learning university courses. Comput. Electr. Eng. 2021, 89, 106908. [Google Scholar] [CrossRef]
- Waddington, R.J.; Nam, S. Practice exams make perfect: Incorporating course resource use into an early warning system. In Proceedings of the 4th International Conference on Learning Analytics and Knowledge (LAK 2014), Indianapolis, IN, USA, 24–28 March 2014; pp. 188–192. [Google Scholar]
- Ulfa, S.; Fatawi, I. Predicting Factors That Influence Students’ Learning Outcomes Using Learning Analytics in Online Learning Environment. Int. J. Emerg. Technol. Learn. 2020, 16, 4–17. [Google Scholar] [CrossRef]
- Nouri, J.; Saqr, M.; Fors, U. Predicting performance of students in a flipped classroom using machine learning: Towards automated data-driven formative feedback. In Proceedings of the 10th International Conference on Society and Information Technologies (ICSIT 2019), Orlando, FL, USA, 12–15 March 2019; pp. 79–82. [Google Scholar]
- Cagliero, L.; Canale, L.; Farinetti, L.; Baralis, E.; Venuto, E. Predicting student academic performance by means of associative classification. Appl. Sci. 2021, 11, 1420. [Google Scholar] [CrossRef]
- Coussement, K.; Phan, M.; De Caigny, A.; Benoit, D.F.; Raes, A. Predicting student dropout in subscription-based online learning environments: The beneficial impact of the logit leaf model. Decis. Support Syst. 2020, 135, 113325. [Google Scholar] [CrossRef]
- Wang, C.; Chang, L.; Liu, T. Predicting Student Performance in Online Learning Using a Highly Efficient Gradient Boosting Decision Tree. In Proceedings of the 12th IFIP TC 12 International Conference on Intelligent Information Processing (IIP 2022), Qingdao, China, 27–30 May 2022; pp. 508–521. [Google Scholar]
- Kennedy, G.; Coffrin, C.; Barba, P.d.; Corrin, L. Predicting success: How learners’ prior knowledge, skills and activities predict MOOC performance. In Proceedings of the Fifth International Conference on Learning Analytics And Knowledge (LAK 2015), Poughkeepsie, NY, USA, 16–20 March 2015; pp. 136–140. [Google Scholar]
- Ademi, N.; Loshkovska, S.; Kalajdziski, S. Prediction of Student Success Through Analysis of Moodle Logs: Case Study. In Proceedings of the 11th International ICT Innovations Conference (ICT Innovations 2019), Ohrid, North Macedonia, 17–19 October 2019; pp. 27–40. [Google Scholar]
- Cerezo, R.; Sánchez-Santillán, M.; Paule-Ruiz, M.P.; Núñez, J.C. Students’ LMS interaction patterns and their relationship with achievement: A case study in higher education. Comput. Educ. 2016, 96, 42–54. [Google Scholar] [CrossRef]
- Jovanović, J.; Saqr, M.; Joksimović, S.; Gašević, D. Students matter the most in learning analytics: The effects of internal and instructional conditions in predicting academic success. Comput. Educ. 2021, 172, 104251. [Google Scholar] [CrossRef]
- Xing, W.; Chen, X.; Stein, J.; Marcinkowski, M. Temporal predication of dropouts in MOOCs: Reaching the low hanging fruit through stacking generalization. Comput. Hum. Behav. 2016, 58, 119–129. [Google Scholar] [CrossRef]
- Bertolini, R.; Finch, S.J.; Nehm, R.H. Testing the Impact of Novel Assessment Sources and Machine Learning Methods on Predictive Outcome Modeling in Undergraduate Biology. J. Sci. Educ. Technol. 2021, 30, 193–209. [Google Scholar] [CrossRef]
- Itani, A.; Brisson, L.; Garlatti, S. Understanding Learner’s Drop-Out in MOOCs. In Proceedings of the 19th International Conference on Intelligent Data Engineering and Automated Learning (IDEAL 2018), Madrid, Spain, 21–23 November 2018; pp. 233–244. [Google Scholar]
- Akçapınar, G.; Altun, A.; Aşkar, P. Using learning analytics to develop early-warning system for at-risk students. Int. J. Educ. Technol. High. Educ. 2019, 16, 40. [Google Scholar] [CrossRef]
- Suhonen, S.J.; Kinnari-Korpela, H. Using Moodle data for early warning of dropping out. In Proceedings of the 47th SEFI Annual Conference 2019—Varietas Delectat: Complexity is the New Normality, Budapest, Hungary, 16–19 September 2020; pp. 1113–1122. [Google Scholar]
- Venant, R.; Sharma, K.; Vidal, P.; Dillenbourg, P.; Broisin, J. Using sequential pattern mining to explore learners’ behaviors and evaluate their correlation with performance in inquiry-based learning. In Proceedings of the 12th European Conference on Technology Enhanced Learning (EC-TEL 2017), Tallinn, Estonia, 12–15 September 2017; pp. 286–299. [Google Scholar]
- Zhuhadar, L.; Daday, J.; Marklin, S.; Kessler, B.; Helbig, T. Using survival analysis to discovering pathways to success in mathematics. Comput. Hum. Behav. 2019, 92, 487–495. [Google Scholar] [CrossRef]
- Casey, K.; Azcona, D. Utilizing student activity patterns to predict performance. Int. J. Educ. Technol. High. Educ. 2017, 14, 4. [Google Scholar] [CrossRef] [Green Version]
- Massing, T.; Reckmann, N.; Blasberg, A.; Otto, B.; Hanck, C.; Goedicke, M. When is the Best Time to Learn?—Evidence from an Introductory Statistics Course. Open Educ. Stud. 2021, 3, 84–95. [Google Scholar] [CrossRef]
- Parkes, S.; Benkwitz, A.; Bardy, H.; Myler, K.; Peters, J. Being more human: Rooting learning analytics through resistance and reconnection with the values of higher education. High. Educ. Res. Dev. 2020, 39, 113–126. [Google Scholar] [CrossRef]
- Cochran, J.D.; Campbell, S.M.; Baker, H.M.; Leeds, E.M. The Role of Student Characteristics in Predicting Retention in Online Courses. Res. High. Educ. 2014, 55, 27–48. [Google Scholar] [CrossRef]
- Nicoletti, M.d.C. Revisiting the Tinto’s Theoretical Dropout Model. High. Educ. Stud. 2019, 9, 52–64. [Google Scholar] [CrossRef]
- Wladis, C.; Conway, K.M.; Hachey, A.C. The Online STEM Classroom—Who Succeeds? An Exploration of the Impact of Ethnicity, Gender, and Non-traditional Student Characteristics in the Community College Context. Community Coll. Rev. 2015, 43, 142–164. [Google Scholar] [CrossRef] [Green Version]
- Rosson, M.B.; Carroll, J.M.; Sinha, H. Orientation of Undergraduates Toward Careers in the Computer and Information Sciences. ACM J. Trans. Comput. Educ. 2011, 11, 1–23. [Google Scholar] [CrossRef]
- Kosasi, S.; Vedyanto; Kasma, U.; Yuliani, I.D.A.E. The Mediating Role of Learning Analytics to Improve Student Academic Performance. In Proceedings of the 2nd International Conference on Cybernetics and Intelligent System (ICORIS 2020), Manado, Indonesia, 27–28 October 2020; pp. 1–6. [Google Scholar]
- Derr, K.; Hübl, R.; Ahmed, M.Z. Using Test Data for Successive Refinement of an Online Pre-Course in Mathematics. In Proceedings of the 14th European Conference on e-Learning (ECEL 2015), Hatfield, UK, 29–30 October 2015; pp. 173–180. [Google Scholar]
- Derr, K. Identifying Consistent Variables in a Heterogeneous Data Set: Evaluation of a Web-Based Pre-Course in Mathematics. Electron. J. E-Learn. 2017, 15, 82–93. [Google Scholar]
- Yarygina, O. Learning analytics of CS0 students programming errors: The case of data science minor. In Proceedings of the 23rd International Academic Mindtrek Conference, AcademicMindtrek 2020, Tampere, Finland, 29–30 January 2020; pp. 149–152. [Google Scholar]
- Caspari-Sadeghi, S. Applying Learning Analytics in Online Environments: Measuring Learners’ Engagement Unobtrusively. Front. Educ. 2022, 7, 840947. [Google Scholar] [CrossRef]
- Balasooriya, I.; Mor, E.; Rodríguez, M.E. Engagement analytics: A microlevel approach to measure and visualize student engagement. In Software Data Engineering for Network eLearning Environments; Caballé, S., Conesa, J., Eds.; Lecture Notes on Data Engineering and Communications Technologies; Springer: Cham, Switzerland, 2018; Volume 11, pp. 47–66. ISBN 978-3-319-68318-8. [Google Scholar]
- Lagman, A.C.; Adao, R.T. Embedding machine learning algorithm models in decision support system in predicting student academic performance using enrollment and admission data. In Proceedings of the 8th International Workshop on Computer Science and Engineering (WCSE 2018), Bangkok, Thailand, 28–30 June 2018; pp. 298–302. [Google Scholar]
- Muñoz-Merino, P.J.; Valiente, J.A.R.; Kloos, C.D. Inferring higher level learning information from low level data for the Khan Academy platform. In Proceedings of the the Third International Conference on Learning Analytics and Knowledge (LAK 2013), Leuven, Belgium, 8–12 April 2013; pp. 112–116. [Google Scholar]
- Tinto, V. Reflections on Student Persistence. Stud. Success 2017, 8, 1–8. [Google Scholar] [CrossRef] [Green Version]
- Zimmerman, B.J.; Bandura, A.; Martinez-Pons, M. Self-Motivation for Academic Attainment: The Role of Self-Efficacy Beliefs and Personal Goal Setting. Am. Educ. Res. J. 1992, 29, 663–676. [Google Scholar] [CrossRef]
- Morris, L.V.; Finnegan, C.; Wu, S.-S. Tracking student behavior, persistence, and achievement in online courses. Internet High. Educ. 2005, 8, 221–231. [Google Scholar] [CrossRef]
- Holmes, N. Student perceptions of their learning and engagement in response to the use of a continuous e-assessment in an undergraduate module. Assess. Eval. High. Educ. 2015, 40, 1–14. [Google Scholar] [CrossRef]
- Ferguson, R.; Clow, D.; Beale, R.; Cooper, A.J.; Morris, N.; Bayne, S.; Woodgate, A. Moving through MOOCS: Pedagogy, learning design and Patterns of Engagement. In Proceedings of the 10th European Conference on Technology Enhanced Learning (EC-TEL 2015), Toledo, Spain, 15–18 September 2015; pp. 70–84. [Google Scholar]
Type | Search Terms | Search Query Example |
---|---|---|
Purpose | retention OR attrition OR completion OR dropout OR success | Scopus: TITLE-ABS-KEY (“learning analytics” OR “academic analytics” OR “educational analytics”) AND TITLE-ABS-KEY (retention OR attrition OR completion OR dropout OR success) |
Technique | “learning analytics” OR “academic analytics” OR “educational analytics” |
Reference | Key Purpose | Data Source | Data Collection Methods * | Sample (N) | Subjects ** | Commencing Students |
---|---|---|---|---|---|---|
[20] | Identifying at-risk student | For-credit institution | AD | 99 | M | N |
[21] | Predicting performance | For-credit institution | AD | 134 | T | Y |
[22] | Predicting dropout | For-credit institution | AD | 492 | T | Y |
[23] | Predicting performance | For-credit institution | AD | 1100 | T | N |
[24] | Predicting performance | For-credit institution | AD, Q | 310 | S | N |
[25] | Predicting dropout | For-credit institution | AD | 362 | M | Y |
[26] | Predicting performance | MOOC | AD | 597,692 | T | N |
[27] | Predicting performance | For-credit institution | AD | 239 | T | Y |
[28] | Predicting performance | For-credit institution | AD, Q | 784 | E | Y |
[29] | Predicting performance | For-credit institution | AD | 140 | T | N |
[30] | Predicting performance | MOOC | AD | 2794 | S | N |
[31] | Predicting performance | For-credit institution | AD | 136 | M | Y |
[32] | Predicting performance | For-credit institution | AD | 12,836 | E | N |
[33] | Identifying at-risk student | For-credit institution | AD | 224 | T, E, M | N |
[34] | Identifying at-risk student | For-credit institution | AD | 57 | M | N |
[35] | Predicting dropout | For-credit institution | AD | N/A | T | N |
[36] | Identifying struggle student | For-credit institution | AD | 312 | T | N |
[37] | Predicting performance | For-credit institution | AD | 802 | S,T | N |
[38] | Predicting dropout | For-credit institution | AD | 31,071 | E | N |
[39] | Predicting performance | MOOC | AD | 6272 | S, T, E, M | N |
[40] | Predicting performance | For-credit institution | AD, Q | 2864 | S | Y |
[41] | Predicting dropout | For-credit institution | AD | 1421 | T | N |
[42] | Predicting dropout | For-credit institution | AD | 429 | E | Y |
[43] | Predicting performance | For-credit institution | AD | 2056 | T | Y |
[44] | Predicting performance | For-credit institution | AD | 695 | E | Y |
[45] | Predicting performance | For-credit institution | AD | 197 | T | Y |
[46] | Predicting performance | For-credit institution | AD | 400 | E | Y |
[47] | Identifying at-risk student | For-credit institution | Q | 164 | T | Y |
[48] | Predicting performance | For-credit institution | AD | 419 | E | N |
[49] | Identifying at-risk student | For-credit institution | AD | 130,170 | T | N |
[50] | Predicting dropout | For-credit institution | AD | 383 | T | N |
[51] | Identifying at-risk student | For-credit institution | AD | 400 | T | Y |
[52] | Predicting performance | For-credit institution | AD, Q | 954 | S | N |
[53] | Predicting dropout | For-credit institution | AD | 2713 | T | N |
[54] | Identifying at-risk student | MOOC | AD | 4064 | T | N |
[55] | Predicting dropout | For-credit institution | AD | 274 | T | N |
[56] | Identifying at-risk student | For-credit institution | AD | 163 | T | Y |
[57] | Predicting performance | For-credit institution | AD | 1232 | E | N |
[58] | Identifying at-risk student | For-credit institution | AD | 3063 | E | Y |
[59] | Predicting performance | For-credit institution | AD | 57 | T | Y |
[60] | Identifying at-risk student | For-credit institution | AD | 8762 | S | Y |
[61] | Predicting performance | For-credit institution | AD | 53 | T | N |
[62] | Predicting performance | For-credit institution | AD | 251 | T | N |
[63] | Identifying at-risk student | For-credit institution | AD | 5000 | S, T, M | Y |
[64] | Predicting dropout | MOOC | AD | 10,554 | T | N |
[65] | Predicting performance | MOOC | AD | 6455 | S, T, E, M | N |
[66] | Identifying at-risk student | MOOC | AD | 7409 | M | N |
[67] | Identifying at-risk student | For-credit institution | AD | 260 | T | N |
[68] | Predicting performance | MOOC | AD | 32,621 | T | N |
[69] | Predicting performance | For-credit institution | AD | 2472 | S | N |
[70] | Predicting dropout | MOOC | AD | 3617 | T | N |
[71] | Predicting performance | For-credit institution | AD | 3225 | S | N |
[72] | Predicting dropout | MOOC | AD | 20,142 | T | N |
[73] | Identifying at-risk student | For-credit institution | AD | 76 | T | N |
[74] | Predicting dropout | For-credit institution | AD | 83 | S, M | N |
[75] | Predicting performance | For-credit institution | AD | 85 | T | Y |
[76] | Predicting dropout | For-credit institution | AD | 728 | M | N |
[77] | Identifying at-risk student | For-credit institution | AD | 111 | T | N |
[78] | Predicting performance | For-credit institution | AD | 753 | M | Y |
Retention Factors | References | Number of Studies | Percentage of Studies |
---|---|---|---|
Student personal characteristics | [21,22,26,28,31,37,38,40,42,44,47,50,52,55,62,64,68,71,76] | 19 | 32% |
Student enrolment properties | [23,26,28,31,40,42,46,47,50,55,71,73] | 12 | 20% |
Prior academic performance | [21,23,26,28,32,38,40,42,44,47,50,52,55,56,63,68,71,76] | 18 | 31% |
Current academic performance | [20,22,23,24,25,27,28,29,31,32,33,34,37,38,40,42,45,47,49,52,54,58,59,60,62,66,67,70,71,72,73,76,78] | 33 | 56% |
Behavioural engagement | [20,21,22,24,26,27,28,30,35,37,39,41,42,43,46,48,49,52,53,56,57,59,60,61,63,65,67,68,69,70,71,73,74,77] | 34 | 58% |
Academic engagement | [20,21,22,24,26,27,29,30,31,33,34,35,36,37,39,41,42,43,44,46,48,49,51,52,53,56,57,59,60,61,62,63,64,65,66,67,68,69,70,72,73,75,77,78] | 46 | 78% |
Course design | [25,34,37,41,49,54] | 6 | 10% |
Factor | Sub-Factors | Features | ||
---|---|---|---|---|
Student Personal Characteristics | Basic demographic features | Gender (18) | Age/DOB (10) | Disability (1) |
Race/Ethnicity (7) | Citizenship (7) | Working status (2) | ||
City of residence (4) | First generation university student (4) | |||
Socioeconomic features | Family structure (1) | Family income (5) | Parental education (4) | |
# siblings (1) | Parental occupation (1) | Socioeconomic level (1) |
Factor | Sub-Factors | Features | ||
---|---|---|---|---|
Student Enrolment Properties | Administrative | Adm. Type (3) | Study load (3) | International student (1) |
Enrolment date (1) | Intended degree (4) | Distance from campus (1) | ||
Internship (1) | Dormitory (1) | # course enrolled (3) | ||
Financial | Financial aid (1) | Scholarship (2) | Award (1) |
Factor | Sub-Factors | Features | ||
---|---|---|---|---|
Prior Academic Performance | Educational background | Educational level (6) | High school (4) | GPA high school (9) |
No. of credits (2) | GPA prior Uni (2) | |||
Admission test | Score of admission test (SAT, ACT, PSU, GTA) (11) | |||
Maths score (8) | English score (2) | Science score (3) |
Factor | Sub-Factors | Features | ||
---|---|---|---|---|
Current Academic Performance | Overall academic performance | GPA (8) | GPA ranking (1) | GPA semester(s) (9) |
Score course(s) (7) | Credit earned semester(s) (1) | |||
Assessment performance | Assignment score (11) | Assignment pass/fail (1) | Final exam score (3) | |
Quiz/Test score (5) | Lab score (1) |
Factor | Sub-Factors | Features | ||
---|---|---|---|---|
Behavioural Engagement | Clickstream | # login (20) | # notification view (2) | # course page view (11) |
# days login (3) | # announcement view (3) | # glossaries view (2) | ||
# clicks (2) | Online time (10) | First day login (2) | ||
Class participation | Attendance (4) |
Factor | Sub-Factors | Features | ||
---|---|---|---|---|
Academic Engagement | Learning material | # resources view (25) | # unique days visit (3) | # weekly content click (1) |
# back jump (1) | # lecture viewed (6) | # resources downloaded (3) | ||
Hyperlink click (5) | Time spent per module (2) | Specific course activity (3) | ||
Content complete percentage (1) | Video content (8) | Video solution (1) | ||
Assessment participation | # instruction view (2) | # task finished (16) | Quiz review (2) | |
# quiz (8) | Quiz on time (1) | Exam reflection (2) | ||
# submission file (2) | Extracurricular quiz (2) | Exam preparation (2) | ||
# attempt (8) | Days assignment submitted in advance (1) | |||
Grade view (2) | Time spent on assessments (4) | |||
Interactivity | Discussion/Forum viewed (17) | Discussion/Forum posted/commented (21) | Post about learning concept (1) | |
Collaborative communication (3) | Messages send to instructor (2) | Messages received from instructor (1) | ||
Peer assessment (1) | Consultation attend (4) |
Factor | Features | ||
---|---|---|---|
Course Design | # assignment/quiz (3) | Content pieces (1) | Pass rate (1) |
# distinct assignments (1) | Assessments completion rate (1) | ||
# clickstream (1) | Average activity days (2) |
Factor | Sub-Factors | Significant Features | ||
---|---|---|---|---|
Student Personal Characteristics | Basic demographic features | Age/DOB | City of residence | Gender |
Student Enrolment Properties | Administrative | Intended degree | ||
Financial | Financial aid | Scholarship | ||
Prior Academic Performance | Educational background | Educational level | GPA prior Uni | GPA high school |
Admission test | Score of admission test (SAT, ACT, PSU, GTA) | Maths score | Science score | |
Current Academic Performance | Overall academic performance | GPA | GPA semester(s) | Credit earned semester(s) |
Assessment performance | Assignment score | Quiz/Test score | Lab score | |
Behavioural Engagement | Clickstream | # login | # days login | # course page view |
# notification view | ||||
Class participation | Attendance | |||
Academic Engagement | Learning material | # resources view | # unique days visit | # lecture viewed |
# resources downloaded | Video content | |||
Assessment participation | # task finished | Time spent on assessments | Days assignment submitted in advance | |
Interactivity | Discussion/Forum viewed | Discussion/Forum posted/commented | ||
Course Design | # assignment/quiz |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Li, C.; Herbert, N.; Yeom, S.; Montgomery, J. Retention Factors in STEM Education Identified Using Learning Analytics: A Systematic Review. Educ. Sci. 2022, 12, 781. https://doi.org/10.3390/educsci12110781
Li C, Herbert N, Yeom S, Montgomery J. Retention Factors in STEM Education Identified Using Learning Analytics: A Systematic Review. Education Sciences. 2022; 12(11):781. https://doi.org/10.3390/educsci12110781
Chicago/Turabian StyleLi, Chunping, Nicole Herbert, Soonja Yeom, and James Montgomery. 2022. "Retention Factors in STEM Education Identified Using Learning Analytics: A Systematic Review" Education Sciences 12, no. 11: 781. https://doi.org/10.3390/educsci12110781
APA StyleLi, C., Herbert, N., Yeom, S., & Montgomery, J. (2022). Retention Factors in STEM Education Identified Using Learning Analytics: A Systematic Review. Education Sciences, 12(11), 781. https://doi.org/10.3390/educsci12110781