Assessment Analytics in Digital Assessments
Definition
1. Introduction
2. Data in Digital Assessments
2.1. Data Types
2.2. Interpretive Cautions
3. Analytic Approaches
3.1. Descriptive and Diagnostic Analytics
3.2. Measurement-Oriented Modeling
3.3. Sequence and Process-Oriented Analyses
3.4. Predictive Models
3.5. Recent Developments
4. Validity, Fairness, and Responsible Use
4.1. Validity Arguments
4.2. Fairness and Accessibility
4.3. Ethics, Privacy, and Governance
5. Implementation Considerations
5.1. Assessment Design and Improvement
5.2. Operational Monitoring and Quality Assurance
6. Two Case Studies of Assessment Analytics
6.1. A Hypothetical University Assessment
6.1.1. Data Sources
6.1.2. Analytical Approach
6.1.3. Implications for Practice
6.2. National Assessment of Educational Progress (NAEP)
6.2.1. Data Sources
6.2.2. Analytical Approach
6.2.3. Implications for Practice
7. Conclusions
Directions for Future Research
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Long, P.; Siemens, G. Penetrating the fog: Analytics in learning and education. EDUCAUSE Rev. 2011, 46, 30–40. [Google Scholar]
- Yildirim-Erbasli, S.N.; Bulut, O. Designing predictive models for early prediction of students’ test-taking engagement in computerized formative assessments. J. Appl. Test. Technol. 2023, 24, 34–47. [Google Scholar]
- Bridgeman, B.; Lennon, M.L.; Jackenthal, A. Effects of Screen Size, Screen Resolution, And Display Rate on Computer-Based Test Performance. ETS Res. Rep. Ser. 2001, 2001, i-23. [Google Scholar] [CrossRef]
- van Bakel, W.T.H. Navigating Exams: Identifying Test-Taking Navigation Behaviour. Master’s Thesis, Utrecht University, Utrecht, The Netherlands, 2021. [Google Scholar]
- Larmuseau, C.; Cornelis, J.; Lancieri, L.; Desmet, P.; Depaepe, F. Multimodal learning analytics to investigate cognitive load during online problem solving. Br. J. Educ. Technol. 2020, 51, 1548–1562. [Google Scholar] [CrossRef]
- Ifenthaler, D. Learning analytics. In The SAGE Encyclopedia of Educational Technology; Spector, J.M., Ed.; Sage Publications: Thousand Oaks, CA, USA, 2015; Volume 2, pp. 447–451. [Google Scholar]
- Ellis, C. Broadening the scope and increasing the usefulness of learning analytics: The case for assessment analytics. Br. J. Educ. Technol. 2013, 44, 662–664. [Google Scholar] [CrossRef]
- He, S.; Cui, Y. A systematic review of the use of log-based process data in computer-based assessments. Comput. Educ. 2025, 228, 105245. [Google Scholar] [CrossRef]
- Bulut, O.; Yildirim-Erbasli, S.N.; Gorgun, G. Assessment analytics for digital assessments: Identifying, modeling, and interpreting behavioral engagement. In Assessment Analytics in Education—Designs, Methods and Solutions; Sahin, M., Ifenthaler, D., Eds.; Springer: Cham, Switzerland, 2024; pp. 35–60. [Google Scholar] [CrossRef]
- Yildirim-Erbasli, S.N.; Bulut, O. The impact of students’ test-taking effort on growth estimates in low-stakes educational assessments. Educ. Res. Eval. 2021, 26, 368–386. [Google Scholar] [CrossRef]
- Jerez, D.; Mazzullo, E.; Bulut, O. Exploring slow responses in international large-scale assessments using sequential process analysis. Computers 2026, 15, 64. [Google Scholar] [CrossRef]
- Lehman, B.; Graesser, A. To resolve or not to resolve? That is the big question about confusion. In International Conference on Artificial Intelligence in Education; Springer International Publishing: Cham, Switzerland, 2015; pp. 216–225. [Google Scholar]
- Ghafournia, N.; Afghari, A. The Interaction between reading comprehension cognitive test-taking strategies, test performance, and cognitive language learning strategies. Procedia-Soc. Behav. Sci. 2013, 70, 80–84. [Google Scholar] [CrossRef]
- Lovett, B.J.; Lewandowski, L.J.; Potts, H.E. Test-taking speed: Predictors and implications. J. Psychoeduc. Assess. 2017, 35, 351–360. [Google Scholar] [CrossRef]
- Leighton, J.P.; Gokiert, R.J. The cognitive effects of test item features: Informing item generation by identifying construct irrelevant variance. In Proceedings of the Annual Meeting of the National Council on Measurement in Education, Montreal, QC, Canada, 12–14 April 2005. [Google Scholar]
- Bayrak, F.; Aydın, F.; Yurdugül, H. Navigational behavior patterns of learners on dashboards based on assessment analytics. In Visualizations and Dashboards for Learning Analytics; Springer International Publishing: Cham, Switzerland, 2021; pp. 251–268. [Google Scholar]
- Kuhfeld, M.; Soland, J. Using assessment metadata to quantify the impact of test disengagement on estimates of educational effectiveness. J. Res. Educ. Eff. 2020, 13, 147–175. [Google Scholar] [CrossRef]
- Camara, W.J.; Harris, D.J. Impact of technology, digital devices, and test timing on score comparability. In Integrating Timing Considerations to Improve Testing Practices; Routledge: London, UK, 2020; pp. 104–121. [Google Scholar]
- Lang, S.C.; Elliott, S.N.; Bolt, D.M.; Kratochwill, T.R. The effects of testing accommodations on students’ performances and reactions to testing. Sch. Psychol. Q. 2008, 23, 107. [Google Scholar] [CrossRef]
- Bishop, S.; Egan, K. Detecting erasures and unusual gain scores: Understanding the status quo. In Handbook of Quantitative Methods for Detecting Cheating on Tests; Routledge: London, UK, 2016; pp. 193–213. [Google Scholar]
- Sinharay, S.; Johnson, M.S. Three new methods for analysis of answer changes. Educ. Psychol. Meas. 2017, 77, 54–81. [Google Scholar] [CrossRef] [PubMed]
- Roscoe, R.D.; Snow, E.L.; Allen, L.K.; McNamara, D.S. Automated detection of essay revising patterns: Applications for intelligent feedback in a writing tutor. Grantee Submiss. 2015, 10, 59–79. [Google Scholar]
- Bridgeman, B. A simple answer to a simple question on changing answers. J. Educ. Meas. 2012, 49, 467–468. [Google Scholar] [CrossRef]
- Deane, P.; Zhang, M.; Hao, J.; Li, C. Using keystroke dynamics to detect nonoriginal text. J. Educ. Meas. 2026, 63, e12431. [Google Scholar] [CrossRef]
- Ahn, B.T.; Harley, J.M. Facial expressions when learning with a queer history app: Application of the control value theory of achievement emotions. Br. J. Educ. Technol. 2020, 51, 1563–1576. [Google Scholar] [CrossRef]
- Zhou, C.; Yuan, F.; Huang, T.; Zhang, Y.; Kaner, J. The impact of interface design element features on task performance in older adults: Evidence from eye-tracking and EEG signals. Int. J. Environ. Res. Public Health 2022, 19, 9251. [Google Scholar] [CrossRef]
- Mohammadyari, S.; Singh, H. Understanding the effect of e-learning on individual performance: The role of digital literacy. Comput. Educ. 2015, 82, 11–25. [Google Scholar] [CrossRef]
- Holmes, S. Time Limits and Speed of Working in Assessments: When, and to What Extent, Should Speed of Working Be Part of What Is Assessed; Ofqual research report 25/7267/2; Ofqual: Coventry, UK, 2025. [Google Scholar]
- Zumbo, B.D.; Hubley, A.M. (Eds.) Understanding and Investigating Response Processes in Validation Research; Springer International Publishing/Springer Nature: Cham, Switzerland, 2017. [Google Scholar] [CrossRef]
- Emerson, A.; Cloude, E.B.; Azevedo, R.; Lester, J. Multimodal learning analytics for game-based learning. Br. J. Educ. Technol. 2020, 51, 1505–1526. [Google Scholar] [CrossRef]
- Sharma, K.; Giannakos, M.N. Multimodal data capabilities for learning: What can multimodal data tell us about learning? Br. J. Educ. Technol. 2020, 51, 1450–1484. [Google Scholar] [CrossRef]
- Yan, L.; Zhao, L.; Gašević, D.; Martinez-Maldonado, R. Scalability, sustainability, and ethicality of multimodal learning analytics. In Proceedings of the 12th International Learning Analytics and Knowledge Conference, New York, NY, USA, 21–25 March 2022. [Google Scholar] [CrossRef]
- Yildirim-Erbasli, S.N.; Gorgun, G. Disentangling the relationship between ability and test-taking effort: To what extent the ability levels can be predicted from response behavior? Technol. Knowl. Learn. 2025, 30, 1475–1497. [Google Scholar] [CrossRef]
- Ulitzsch, E.; Yildirim-Erbasli, S.N.; Gorgun, G.; Bulut, O. An explanatory mixture IRT model for careless and insufficient effort responding in self-report measures. Br. J. Math. Stat. Psychol. 2022, 75, 668–698. [Google Scholar] [CrossRef]
- Nagy, G.; Ulitzsch, E. A multilevel mixture IRT framework for modeling response times as predictors or indicators of response engagement in IRT models. Educ. Psychol. Meas. 2022, 82, 845–879. [Google Scholar] [CrossRef] [PubMed]
- Wang, S.; Zhang, S.; Shen, Y. A joint modeling framework of responses and response times to assess learning outcomes. Multivar. Behav. Res. 2020, 55, 49–68. [Google Scholar] [CrossRef] [PubMed]
- Fox, J.P.; Marianti, S. Joint modeling of ability and differential speed using responses and response times. Multivar. Behav. Res. 2016, 51, 540–553. [Google Scholar] [CrossRef] [PubMed]
- Kingston, N.M.; Dorans, N.J. The effect of the position of an item within a test on item responding behavior: An analysis based on item response theory. ETS Res. Rep. Ser. 1982, 1982, i–26. [Google Scholar] [CrossRef]
- Reimann, P. Time is precious: Variable- and event-centred approaches to process analysis in CSCL research. Int. J. Comput.-Support. Collab. Learn. 2009, 4, 239–257. [Google Scholar] [CrossRef]
- Seifried, J.; Brandt, S.; Kögler, K.; Rausch, A. The computer-based assessment of domain-specific problem-solving competence: A three-step scoring procedure. Cogent Educ. 2020, 7, 1719571. [Google Scholar] [CrossRef]
- Han, Y.; Liu, H.; Ji, F. A sequential response model for analyzing process data on technology-based problem-solving tasks. Multivar. Behav. Res. 2022, 57, 960–977. [Google Scholar] [CrossRef]
- Peach, R.L.; Greenbury, S.F.; Johnston, I.G.; Yaliraki, S.N.; Lefevre, D.; Barahona, M. Data-driven modelling and characterisation of task completion sequences in online courses. arXiv 2020, arXiv:2007.07003. [Google Scholar] [CrossRef]
- Roque, F.V.; Junior, L.C.; Cechinel, C.; Marcon, M.Z.; Kuhnen, A.; Munoz, R.; Grellert, M. Learning Analytics for Virtual Industrial Labs: Performance Segmentation and Error Pattern Discovery via Sequential Mining. IEEE Access 2025, 13, 194401–194420. [Google Scholar] [CrossRef]
- Bulut, O.; Cormier, D.C.; Yildirim-Erbasli, S.N. Optimized screening for students at-risk in mathematics: A machine learning approach. Information 2022, 13, 400. [Google Scholar] [CrossRef]
- Wentworth, L.; Nagaoka, J. Early warning indicators in education: Innovations, uses, and optimal conditions for effectiveness. Teach. Coll. Rec. 2020, 122, 1–22. [Google Scholar] [CrossRef]
- Atif, A.; Richards, D.; Liu, D.; Bilgin, A.A. Perceived benefits and barriers of a prototype early alert system to detect engagement and support ‘at-risk’ students: The teacher perspective. Comput. Educ. 2020, 156, 103954. [Google Scholar] [CrossRef]
- Wang, Y.; Ding, Z.; Wu, X.; Sun, S.; Liu, N.; Zhai, X. Autoscore: Enhancing automated scoring with multi-agent large language models via structured component recognition. Proc. AAAI Conf. Artif. Intell. 2026, 40, 40898–40906. [Google Scholar] [CrossRef]
- Khalifa, A.; Tahhan, O.; Albazooni, M.; Saeed, M.; Hamdi, R.; Stanners, M.; Malik, A. Automated and artificial intelligence (AI)-derived performance assessment in surgical simulation: A systematic review. Cureus 2025, 17, e12. [Google Scholar] [CrossRef]
- Khine, M.S. Using AI for adaptive learning and adaptive assessment. In Artificial Intelligence in Education: A Machine-Generated Literature Overview; Springer Nature: Singapore, 2024; pp. 341–466. [Google Scholar]
- Kane, M.T. Current concerns in validity theory. J. Educ. Meas. 2001, 38, 319–342. [Google Scholar] [CrossRef]
- Gorgun, G.; Yildirim-Erbasli, S.N. Algorithmic bias in BERT for response accuracy prediction: A case study for investigating population validity. J. Educ. Meas. 2026, 63, e12420. [Google Scholar] [CrossRef]
- Khalil, M.; Prinsloo, P.; Slade, S. Fairness, trust, transparency, equity, and responsibility in learning analytics. J. Learn. Anal. 2023, 10, 1–7. [Google Scholar] [CrossRef]
- Bulut, O.; Gorgun, G.; Yildirim-Erbasli, S.N.; Wongvorachan, T.; Daniels, L.M.; Gao, Y.; Lai, K.W.; Shin, J. Standing on the shoulders of giants: Online formative assessments as the foundation for predictive learning analytics models. Br. J. Educ. Technol. 2022, 54, 19–39. [Google Scholar] [CrossRef]
- Ganesh, P.; Tran, C.; Shokri, R.; Fioretto, F. The data minimization principle in machine learning. In Proceedings of the 2025 ACM Conference on Fairness, Accountability, and Transparency, Athens, Greece, 23–26 June 2025; pp. 3075–3093. [Google Scholar]
- Kleinman, E.; Shergadwala, M.; Seif El-Nasr, M.; Teng, Z.; Villareale, J.; Bryant, A.; Zhu, J. Analyzing students’ problem-solving sequences: A human-in-the-loop approach. J. Learn. Anal. 2022, 9, 138–160. [Google Scholar] [CrossRef]
- Viberg, O.; Poquet, O.; Kovanovic, V.; Khosravi, H. Fostering Human Agency in Age of AI: A Learning Analytics Perspective. J. Learn. Anal. 2025, 12, 1–7. [Google Scholar] [CrossRef]
- Sedrakyan, G.; Malmberg, J.; Verbert, K.; Järvelä, S.; Kirschner, P.A. Linking learning behavior analytics and learning science concepts: Designing a learning analytics dashboard for feedback to support learning regulation. Comput. Hum. Behav. 2020, 107, 105512. [Google Scholar] [CrossRef]
- Mangaroska, K.; Giannakos, M. Learning analytics for learning design: A systematic literature review of analytics-driven design to enhance learning. IEEE Trans. Learn. Technol. 2018, 12, 516–534. [Google Scholar] [CrossRef]
- Asatryan, S.; Hakobyan, L.; Adamyan, N. The role of big data and learning analytics in the quality assurance process of higher education. Educ. 21st Century 2025, 7, 87–99. [Google Scholar] [CrossRef]
- Finn, B. Measuring motivation in low-stakes assessments. ETS Res. Rep. Ser. 2015, 2015, 1–17. [Google Scholar] [CrossRef]
- Wise, S.L.; Ma, L. Setting response time thresholds for a CAT item pool: The normative threshold method. In Proceedings of the Annual meeting of the National Council on Measurement in Education, Vancouver, BC, Canada, 14–16 April 2012. [Google Scholar]
- Wise, S.L.; Kong, X. Response time effort: A new measure of examinee motivation in computer-based tests. Appl. Meas. Educ. 2005, 18, 163–183. [Google Scholar] [CrossRef]
- Bulut, O.; Gorgun, G.; Wongvorachan, T.; Tan, B. Rapid guessing in low-stakes assessments: Finding the optimal response time threshold with random search and genetic algorithm. Algorithms 2023, 16, 89. [Google Scholar] [CrossRef]
- National Center for Education Statistics. NAEP 2017 Digitally Based Mathematics Assessment. The Nation’s Report Card. 2017. Available online: https://www.nationsreportcard.gov/math_2017/about/digitally-based-assessment/?grade=4 (accessed on 24 March 2026).
- Ogut, B.; Circi, R.; Huo, H.; Hicks, J.; Yin, M. Running Out of Time: Leveraging Process Data to Identify Students Who May Benefit from Extended Time. Int. Electron. J. Elem. Educ. 2025, 17, 253–265. [Google Scholar] [CrossRef]
| Dimension | Assessment Analytics | Learning Analytics |
|---|---|---|
| Primary Purpose | Support measurement and decision-making in assessment contexts | Support learning processes and instructional improvement |
| Core Question | What can be validly inferred about proficiency? | How can learning processes be understood and improved? |
| Theoretical Foundation | Educational measurement, psychometrics, validity theory | Learning sciences, educational data mining |
| Primary Data Sources | Assessment logs, item responses, timing data, navigation traces, item metadata | Clickstreams, discussion forums, assignments, engagement metrics |
| Unit of Analysis | Often items, tests or individuals | Often learners, courses, or cohorts |
| Interpretive Constraints | Less flexibility; focuses on construct validity, comparability, fairness, and standardization | More flexibility; focuses on usefulness for intervention and support |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Bulut, O.; Yildirim-Erbasli, S.N. Assessment Analytics in Digital Assessments. Encyclopedia 2026, 6, 81. https://doi.org/10.3390/encyclopedia6040081
Bulut O, Yildirim-Erbasli SN. Assessment Analytics in Digital Assessments. Encyclopedia. 2026; 6(4):81. https://doi.org/10.3390/encyclopedia6040081
Chicago/Turabian StyleBulut, Okan, and Seyma N. Yildirim-Erbasli. 2026. "Assessment Analytics in Digital Assessments" Encyclopedia 6, no. 4: 81. https://doi.org/10.3390/encyclopedia6040081
APA StyleBulut, O., & Yildirim-Erbasli, S. N. (2026). Assessment Analytics in Digital Assessments. Encyclopedia, 6(4), 81. https://doi.org/10.3390/encyclopedia6040081
