Validation Is a Methodology! Guideposts for Assessment Development and Validation
Abstract
1. Introduction
2. Related Literature
2.1. Language Matters
2.2. Validity and Validation: A Brief Primer
3. Validation as a Methodology
- Core tenet: Validation works with many theoretical perspectives.
- Core tenet: Validation scholarship seeks fairness and to minimize bias.
- Core tenet: Validation work is scholarship that should be shared publicly.
- Core Tenet: Validation involves trustworthiness and open communication.
- Core Tenet: Validation is collaborative.
- Core Tenet: Validation is not limited to one form of testing—it is something for all quantitative measurement.
- Conclusion
4. Guideposts for Validation Scholarship
- Guidepost: What theoretical perspective(s) guide the validation study?
- Guidepost: What experiences and knowledge does your team have?
- Guidepost: Claims can be shared a priori (in the beginning) or a posteriori (afterwards)
- Guidepost: There are numerous means for communicating a validity argument.
- Guidepost: Test users, test administrators, and respondents must be at the core of assessment development and validation.
- Guidepost: Test abstracts and test names are necessary.
- Guidepost: Technology is only as useful as the individuals using it during validation.
- Guidepost: Balance a need for a new test with the resources necessary to develop a robust measure.
4.1. A Recipe for a Validation Study
- Step 1: Thoughtfully consider partners that may be involved in your validation study. Determine available resources in the partnership.
- Step 2: Describe the construct to assess then thoroughly review the available literature.
- Step 3: Define the desired test format and test administration processes. Draft an associated validity argument and ensuing claims, as well as necessary validity evidence to support the claims and argument.
- Step 4: Develop a test item development process and implement it.
- Step 5: Align intended data collection practices to claims and validity evidence. Perform an alpha test by collecting and then analyzing data.
- Step 6: Refine items and the assessment based upon alpha testing results. Collect then analyze data as part of beta testing.
- Step 7: Analyze pilot test data. Share results from the analysis.
- Step 8: Create a ‘final’ test and administer it. Collect and analyze data. Disseminate results publicly through appropriate venues.
- Step 9: Return to any step to gather further validity evidence or make adjustments to the test.
- Step 10: Make the test or test development process and outcomes accessible to others.
4.2. Common Pitfalls in Validation Work
4.3. Remarks Following Test Development
4.4. Critiques to Come and Known Concerns
4.5. Final Remarks: A Call to Action and Collaboration
Funding
Institutional Review Board Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
| K-12 | Kindergarten-12th grade |
| STEM | Science, Technology, Engineering, and Mathematics |
| AI | Artificial Intelligence |
| VM2ED | Validity and Measurement in Mathematics Education |
References
- Al-Lawama, M., Altamimi, O., & Altamimi, E. (in press). Evaluating the ability of AI models to generate level-specific medical MCQs with variable difficulty. BMC Research Notes. [CrossRef]
- American Educational Research Association, American Psychological Association & National Council on Measurement in Education. (2014). Standards for educational and psychological testing. American Educational Research Association. [Google Scholar]
- American Educational Research Association, American Psychological Association, National Council on Measurement in Education & Joint Committee on Standards for Educational and Psychological Testing (U.S.). (1999). Standards for educational and psychological testing. American Educational Research Association. [Google Scholar]
- American Psychological Association. (2019). Understanding context of women’s and girls’ lives key to providing good psychological care, according to updated practice guidelines. Available online: https://www.apa.org/news/press/releases/2019/05/women-girls-psychological-care (accessed on 2 March 2026).
- American Psychological Association. (2025). Journal Article Reporting Standards (JARS). Available online: https://apastyle.apa.org/jars (accessed on 2 March 2026).
- Arjoon, J. A., Xu, X., & Lewis, J. E. (2013). Understanding the state of the art for measurement in chemistry education research: Examining the psychometric evidence. Journal of Chemical Education, 90(5), 536–545. [Google Scholar] [CrossRef]
- Bannon, L. (2011). Reimagining HCI: Toward a more human-centered perspective. Interactions, 18(4), 50–57. [Google Scholar] [CrossRef]
- Bostic, J. (2023). Engaging hearts and minds in assessment research. School Science and Mathematics Journal, 123(6), 217–219. [Google Scholar] [CrossRef]
- Bostic, J., Krupa, E., Carney, M., & Shih, J. (2019a). Reflecting on the past and thinking ahead in the measurement of students’ outcomes. In J. Bostic, E. Krupa, & J. Shih (Eds.), Quantitative measures of mathematical knowledge: Researching instruments and perspectives (pp. 205–229). Routledge. [Google Scholar]
- Bostic, J., Krupa, E., Folger, T., Bentley, B., & Stokes, D. (2022). Gathering validity evidence to support mathematics education scholarship. In A. Lischka, E. Dyer, R. Jones, J. Lovett, J. Strayer, & S. Drown (Eds.), Proceedings of the forty-fourth annual meeting of the North American chapter of the international group for the psychology of mathematics education (pp. 100–104). Middle Tennessee State University. [Google Scholar]
- Bostic, J., Lesseig, K., Sherman, M., & Boston, M. (2021a). Classroom observation and mathematics education research. Journal of Mathematics Teacher Education, 24, 5–31. [Google Scholar] [CrossRef]
- Bostic, J., Matney, G., & Sondergeld, T. (2019b). A lens on teachers’ promotion of the Standards for Mathematical Practice. Investigations in Mathematics Learning, 11(1), 69–82. [Google Scholar] [CrossRef]
- Bostic, J., May, T., Folger, T., Matney, G., Koskey, K., & Stone, G. (2024, July 7–14). Applying empathic principles to mathematics education assessment. International Council on Mathematics Education, Sydney, Australia. [Google Scholar]
- Bostic, J., & Sondergeld, T. (2015). Measuring sixth-grade students’ problem solving: Validating an instrument addressing the mathematics common core. School Science and Mathematics Journal, 115, 281–291. [Google Scholar] [CrossRef]
- Bostic, J., Sondergeld, T., Folger, T., & Kruse, L. (2017). PSM7 and PSM8: Validating two problem-solving measures. Journal of Applied Measurement, 18(2), 151–162. [Google Scholar]
- Bostic, J., Sondergeld, T., Matney, G., Stone, G., & Hicks, T. (2021b). Gathering response process data for a problem-solving measure through whole-class think alouds. Applied Measurement in Education, 34(1), 46–60. [Google Scholar] [CrossRef]
- Braat, M., Engelen, J., van Gemert, T., & Verhaegh, S. (2020). The rise and fall of behaviorism: The narrative and the numbers. History of Psychology, 23(3), 252. [Google Scholar] [CrossRef]
- Carney, M., Bostic, J., Krupa, E., & Shih, J. (2022). Interpretation and use statements for instruments in mathematics education. Journal for Research in Mathematics Education, 53(4), 334–340. [Google Scholar] [CrossRef]
- Carney, M. B., Cavey, L., & Hughes, G. (2017). Assessing teacher attentiveness to student mathematical thinking: Validity claims and evidence. The Elementary School Journal, 118(2), 281–309. [Google Scholar] [CrossRef]
- Confrey, J., Toutkoushian, E., & Shah, M. (2019). A validation argument from soup to nuts: Assessing progress on learning trajectories for middle-school mathematics. Applied Measurement in Education, 32(1), 23–42. [Google Scholar] [CrossRef]
- Cronbach, L. J. (1988). Five perspectives on validity argument. In H. Wainer, & H. Braun (Eds.), Test validity (pp. 3–17). Erlbaum. [Google Scholar]
- Crotty, M. (1998). The foundations of social research: Meaning and perspective in the research process. Sage. [Google Scholar]
- Cruz, M. L., Saunders-Smits, G. N., & Groen, P. (2020). Evaluation of competency methods in engineering education. European Journal of Engineering Education, 45(5), 729–757. [Google Scholar] [CrossRef]
- Decker, A., & McGill, M. M. (2019). A topical review of evaluation instruments for computing education. In Proceedings of the 50th ACM technical symposium on computer science education (pp. 558–564). ACM. [Google Scholar]
- Delandshere, G. (2002). Assessment as inquiry. Teachers College Record, 104(7), 1461–1484. [Google Scholar] [CrossRef]
- Edelen, D., Cook, K., Tripp, L. O., Jackson, C., Bush, S. B., Mohr-Schroeder, M. J., Schroeder, D. C., Roberts, T., Maiorca, C., Ivy, J., Burton, M., & Perrin, A. (2024). “No, this is not my boyfriend’s computer”: Elevating the voices of youth in STEM education research leveraging photo-elicitation. Journal for STEM Education Research, 7(3), 444–462. [Google Scholar] [CrossRef]
- Edelen, D., Roberts, O., Bostic, J., Young, J., Lin, K.-Y., & Roberts, A. (accepted). Methodological approaches that capture the nuances of STEM Education Research. In C. Johnson, M. Mohr-Schroeder, T. Moore, L. English, & C. Jackson (Eds.), Handbook of research on STEM education. Routledge. [Google Scholar]
- European Institute of Innovation & Technology. (2025). EIT launches €70 million call to boose STEM innovation and strength university cooperation. Available online: https://www.eit.europa.eu/news-events/news/eit-launches-eu70-million-call-boost-stem-innovation-and-strengthen-university (accessed on 2 March 2026).
- Fegely, A., Winslow, J., Lee, C. Y., & Setari, A. P. (2023). EdTech Align: A valid and reliable instrument for measuring teachers’ EdTech competencies aligned to professional standards. TechTrends. [Google Scholar] [CrossRef]
- Folger, T., Bostic, J., & Krupa, E. (2023). Defining test-score interpretation, use, and claims: Delphi study for the validity argument. Educational Measurement: Issues & Practice, 42(3), 22–38. [Google Scholar] [CrossRef]
- Gallagher, M. A., Folger, T. D., Walkowiak, T. A., Wilhelm, A. G., & Zelkowski, J. (2025). Measuring mathematics teaching quality: The state of the field and a call for the future. Education Sciences, 15(9), 1158. [Google Scholar] [CrossRef]
- Garcia, N. M., López, N., & Vélez, V. N. (2018). QuantCrit: Rectifying quantitative methods through critical race theory. Race Ethnicity and Education, 21(2), 149–157. [Google Scholar] [CrossRef]
- Ghosh, R., & May, T. A. (2025). Validation of identity-based mentoring scales for undergraduate minoritized students in engineering. Journal of College Student Development, 66(4), 454–462. [Google Scholar] [CrossRef]
- Glaser, B. G. (1965). The constant comparative method of qualitative analysis. Social Problems, 12(4), 436–445. [Google Scholar] [CrossRef]
- Herber, O. R., Bradbury-Jones, C., Okpokiri, C., & Taylor, J. (2025). Epistemologies, methodologies and theories used in qualitative global north health and social care research: A scoping review protocol. BMJ Open, 15, e100494. [Google Scholar] [CrossRef]
- Hill, H. C., & Shih, J. C. (2009). Research commentary: Examining the quality of statistical mathematics education research. Journal for Research in Mathematics Education, 40(3), 241–250. [Google Scholar] [CrossRef]
- H.R.4346—117th congress (2021–2022): CHIPS and science act. (2022). Available online: https://www.congress.gov/bill/117th-congress/house-bill/4346 (accessed on 17 March 2026).
- Ing, M., Kosko, K. W., Jong, C., & Shih, J. C. (2024). Validity evidence of the use of quantitative measures of students in elementary mathematics education. School Science and Mathematics, 124(6), 411–423. [Google Scholar] [CrossRef]
- Jackson, C., Mohr-Schroeder, M. J., Bush, S. B., Maiorca, C., Roberts, T., Yost, C., & Fowler, A. (2021). Equity-oriented conceptual framework for k-12 stem literacy. International Journal of STEM Education, 8, 38. [Google Scholar] [CrossRef]
- Jacobson, E., & Borowski, R. (2019). Measure validation as a research methodology for mathematics education. In J. Bostic, E. Krupa, & J. Shih (Eds.), Assessment in mathematics education contexts: Theoretical frameworks and new directions (pp. 40–62). Routledge. [Google Scholar]
- Jonson, J. L., & Geisinger, K. F. (2020). Fairness in educational and psychological testing. Examining theoretical, research, practice, and policy implications of the 2014 Standards. American Educational Research Association. [Google Scholar]
- Kane, M. T. (2013). Validating the interpretations and uses of test scores. Journal of Educational Measurement, 50(1), 1–73. [Google Scholar] [CrossRef]
- Kane, M. T. (2016). Explicating validity. Assessment in Education: Principles, Policy & Practice, 23(2), 198–211. [Google Scholar]
- Kane, M. T. (2021). Articulating a validity argument. In G. Fulcher, & L. Harding (Eds.), The Routledge handbook of language testing (2nd ed., pp. 32–47). Routledge. [Google Scholar]
- Kelley, T. R., & Knowles, J. G. (2016). A conceptual framework for integrated STEM education. International Journal of STEM Education, 3(1), 11. [Google Scholar] [CrossRef]
- Kelly, A. E. (2004). Yes, but is it methodological? Journal of the Learning Sciences, 13(1), 115–128. [Google Scholar] [CrossRef]
- Kosko, K. W. (2017). Reconsidering the role of disembedding in multiplicative concepts: Extending theory from the process of developing a quantitative measure. Investigations in Mathematics Learning, 10(1), 54–65. [Google Scholar] [CrossRef]
- Kosko, K. W. (2019). A multiplicative reasoning assessment for fourth and fifth grade students. Studies in Educational Evaluation, 60, 32–42. [Google Scholar] [CrossRef]
- Krupa, E., Bostic, J., Folger, T., & Burkett, K. (2024). Introducing a repository of quantitative measures used in mathematics education. In K. Kosko, J. Caniglia, S. Courtney, M. Zolfaghari, & G. Morris (Eds.), Proceedings of the 46th annual meeting of the North American chapter of the international group for the psychology of mathematics education (pp. 55–64). International Group for the Psychology of Mathematics Education. [Google Scholar]
- Krupa, E. E., Bostic, J. D., Bentley, B., Folger, T., Burkett, K. E., & VM²ED Community. (2024). Search. VM²ED Repository. [Google Scholar]
- Lavery, M. R., Bostic, J. D., Kruse, L., Krupa, E. E., & Carney, M. B. (2020). Argumentation surrounding argument-based validation: A systematic review of validation methodology in peer-reviewed articles. Educational Measurement: Issues and Practice, 39(4), 116–130. [Google Scholar] [CrossRef]
- Lawson, B., & Bostic, J. (2024). An investigation into two mathematics score reports: Problem solving measure (PSM) and measurement of academic progress (MAP). Mid-Western Educational Researcher, 36(1), 12. [Google Scholar] [CrossRef]
- Lesh, R., & Doerr, H. M. (2003). Foundations of a models and modeling perspective on mathematics teaching, learning, and problem solving. In Beyond constructivism (pp. 3–33). Routledge. [Google Scholar]
- Lincoln, Y. S., & Guba, E. G. (1985). Naturalistic inquiry. Sage. [Google Scholar]
- Maric, M., Glibetic, M., & Milinkovic, D. (2023). Measurement in STEM education research: A systematic literature review of trends in the psychometric evidence of scales. International Journal of STEM Education, 10(1), 39. [Google Scholar] [CrossRef]
- May, T. A., Fan, Y. K., Stone, G. E., Koskey, K. L. K., Sondergeld, C. J., Folger, T. D., Archer, J. N., Provinzano, K., & Johnson, C. C. (2025a). An effectiveness study of generative artificial intelligence tools used to develop multiple-choice test items. Education Sciences, 15(2), 144. [Google Scholar] [CrossRef]
- May, T. A., Johnson, C. C., Harold, S., & Walton, J. B. (2025b). The development and validation of a K-12 STEM engagement participant outcome instrument. Education Sciences, 15(3), 377. [Google Scholar] [CrossRef]
- Mayer, R. E. (2024). The past, present, and future of the cognitive theory of multimedia learning. Educational Psychology Review, 36(1), 8. [Google Scholar] [CrossRef]
- McMillan, J. H. (2026). Classroom assessment validation: Proficiency claims and uses. Educational Measurement: Issues and Practice, 45, E70014. [Google Scholar] [CrossRef]
- Merriam-Webster. (n.d.). Fairness. In Merriam-webster.com dictionary. Merriam-Webster. Available online: https://www.merriam-webster.com/dictionary/equity (accessed on 15 October 2025).
- Messick, S. (1989). Meaning and values in test validation: The science and ethics of assessment. Educational Researcher, 18(2), 5–11. [Google Scholar] [CrossRef]
- Minner, D., Martinez, A., & Freeman, B. (2012). Compendium of research instruments for STEM education. Part 1 & 2. Abt Associates; Education Development Center; CADRE website. [Google Scholar]
- National Research Council. (2001). Adding it up: Helping children learn mathematics. The National Academies Press. [Google Scholar]
- National Research Council. (2014). Developing assessments for the next generation science standards. The National Academies Press. [Google Scholar]
- Paul, A., Fakiyesi, V., Tahsin, M. U., Arinze, L. C., Moyaki, D., & Dunmoye, I. (2026). Exploring the impact of generative AI in engineering education: A scoping review of applications and innovations. In D. May, & M. E. Auer (Eds.), 2025 Yearbook emerging technologies in learning. Learning and analytics in intelligent systems (Vol. 59). Springer. [Google Scholar]
- Pellegrino, J., DiBello, L., & Goldman, S. (2016). A framework for conceptualizing and evaluating the validity of instructionally relevant assessments. Educational Psychologist, 51(1), 59–81. [Google Scholar] [CrossRef]
- Postma, C. E., Zwartkruis-Pelgrim, E., Daemen, E., & Du, J. (2012). Challenges of doing empathic design: Experiences from industry. International Journal of Design, 6(1), 59–70. [Google Scholar]
- Reeves, T. D., & Marbach-Ad, G. (2016). Contemporary test validity in theory and practice: A primer for discipline-based education researchers. CBE—Life Sciences Education, 15(1), rm1. [Google Scholar] [CrossRef]
- Riedl, M. O. (2019). Human-centered artificial intelligence and machine learning. Human Behavior and Emerging Technologies, 1(1), 33–36. [Google Scholar] [CrossRef]
- Roberts, T., Jackson, C., Mohr-Schroeder, M. J., Bush, S. B., Maiorca, C., Cavalcanti, M., Schroeder, D. C., Delaney, A., Putnam, L., & Cremeans, C. (2018). Students’ perceptions of STEM learning after participating in a summer informal learning experience. International Journal of STEM Education, 5(1), 35. [Google Scholar] [CrossRef]
- Roberts, T., Maiorca, C., Jackson, C., & Mohr-Schroeder, M. (2022). Integrated STEM as problem-solving practices. Investigations in Mathematics Learning, 14(1), 1–13. [Google Scholar] [CrossRef]
- Sablan, J. R. (2019). Can you really measure that? Combining critical race theory and quantitative methods. American Educational Research Journal, 56(1), 178–203. [Google Scholar] [CrossRef]
- Schilling, S. G., & Hill, H. C. (2007). Assessing measures of mathematical knowledge for teaching: A validity argument approach. Measurement: Interdisciplinary Research and Perspectives, 5(2–3), 70–80. [Google Scholar] [CrossRef]
- Schoenfeld, A. H., & Arcavi, A. (1988). On the meaning of variable. Mathematics Teacher, 81(6), 420–427. [Google Scholar] [CrossRef]
- Sireci, S., & Faulkner-Bond, M. (2014). Validity evidence based on test content. Psicothema, 26(1), 100–107. [Google Scholar] [CrossRef]
- Sireci, S. G. (2013). Agreeing on validity arguments. Journal of Educational Measurement, 50(1), 99–104. [Google Scholar] [CrossRef]
- Sireci, S. G., Suárez-Álvarez, J., Zenisky, A. L., & Oliveri, M. E. (2024). Evolving educational testing to meet students’ needs: Design-in-real-time assessment. Educational Measurement: Issues and Practice, 43, 112–118. [Google Scholar] [CrossRef]
- Sondergeld, T. A. (2020). Shifting sights on STEM education quantitative instrumentation development: The importance of moving validity evidence to the forefront rather than a footnote. School Science and Mathematics Journal, 120, 259–261. [Google Scholar] [CrossRef]
- Stylianides, A. J. (2007). Proof and proving in school mathematics. Journal for Research in Mathematics Education, 38(3), 289–321. [Google Scholar]
- Stylianides, G., Stylianides, A. J., & Weber, K. (2019). Research on the teaching and learning of proof: Taking stock and moving forward. In J. Cai (Ed.), Handbook for research in mathematics education (pp. 237–266). National Council of Teachers of Mathematics. [Google Scholar]
- Toulmin, S. E. (2003). The uses of argument. Cambridge University Press. [Google Scholar]
- U.S. President. (2025). Advancing artificial intelligence education for American youth (EO 14277). Federal register, 90 FR 17519. Available online: https://www.govinfo.gov/app/details/FR-2025-04-28/2025-07368 (accessed on 2 March 2026).
- Vygotsky, L. (1997). The collected works of L.S. Vygotsky, Vol. 4: The history of the development of higher mental functions (M. Hall, Trans.; R. Rieber, Ed.). Plenum. [Google Scholar]
- Walkowiak, T., Adams, E. R., & Berry, R. Q. (2019). Validity arguments for instruments that measure matheamtics teaching practice: Comparing the M-SCAn and IPL-M. In J. Bostic, E. Krupa, & J. Shih (Eds.), Assessment in mathematics education contexts: Theoretical frameworks and new directions (pp. 90–119). Routledge. [Google Scholar]
- Wilson, M. (2009). Measuring progressions: Assessment structures underlying a learning progression. Journal of Research in Science Teaching, 46(6), 716–730. [Google Scholar] [CrossRef]
- Wilson, M. (2023). Constructing measures: An item response modeling approach (2nd ed.). Routledge. [Google Scholar] [CrossRef]
- Wilson, M., & Mari, L. (2026). Mapping out the hexagon measurement framework as a blueprint underlying measurement in the human sciences. Journal of Educational Measurement, 63(1), e70036. [Google Scholar] [CrossRef]
- Wilson, M., & Wilmot, D. (2019). Gathering validity evidence using the BEAR assessment system (BAS): A mathematics assessment perspective. In J. Bostic, E. Krupa, & J. Shih (Eds.), Assessment in mathematics education contexts: Theoretical frameworks and new directions (pp. 63–89). Routledge. [Google Scholar]
- Wittgenstein, L. (1958). Philosophical investigations. Basil Blackwell, Ltd. [Google Scholar]
- Zhang, Y., Chen, Y., Kan, Z. C., & Xia, S. Y. (2026). The landscape of STEM education in China: Policies, practices, and pathways for integration. Research in Integrated STEM Education, 1, 1–43. [Google Scholar]
- Ziebarth, S., Fonger, N., & Kratky, J. (2014). Instruments for studying the enacted mathematics curriculum. In D. Thompson, & Z. Usiskin (Eds.), Enacted mathematics curriculum: A conceptual framework and needs (pp. 97–120). Information Age Publishing. [Google Scholar]
- Zolfaghari, M., Austin, C. K., & Kosko, K. W. (2021). Exploring teachers’ pedagogical content knowledge of teaching fractions. Investigations in Mathematics Learning, 13(3), 230–248. [Google Scholar] [CrossRef]
- Zolfaghari, M., Kosko, K., & Austin, C. (2024). Toward a better understanding of the nature of pedagogical content knowledge for fractions: The role of experience. Investigations in Mathematics Learning, 16(4), 263–280. [Google Scholar] [CrossRef]
- Zoltowski, C. B., Oakes, W. C., & Cardella, M. E. (2012). Students’ ways of experiencing human-centered design. Journal of Engineering Education, 101(1), 28–59. [Google Scholar] [CrossRef]
- Zumbo, B. (2014). What role does, and should, the test standards play outside of the United States of America? Educational Measurement: Issues and Practice, 33(4), 31–33. [Google Scholar] [CrossRef]
- Zumbo, B. D., & Hubley, A. M. (2016). Bringing consequences and side effects of testing and assessment to the foreground. Assessment in Education: Principles, Policy & Practice, 23(2), 299–303. [Google Scholar] [CrossRef]





| Source of Validity Evidence | Description |
|---|---|
| Test Content | Test content includes the wording and format of test items or tasks. Validity evidence based on test content would indicate that test items, or test content, align to the construct a test intends to measure. |
| Response Processes | Response processes describes the alignment between test takers’ performance or behavior and the construct a test intends to measure. In cases when a test relies on observers or judges to evaluate test takers, evidence may include “the extent to which the processes of observers or judges are consistent with the intended interpretation of scores” (American Educational Research Association et al., 2014, p. 15). |
| Internal Structure | Internal structure may indicate the degree to which test items conform to the construct a test intends to measure. Such evidence may be collected through analysis of test dimensionality and item interrelationships. |
| Relations to Other Variables | Relations to other variables examines the degree to which test scores are, or are not, related to some ancillary variable. The Standards describe several examples when relations to other variables may be of interest, such as: (a) hypothesized differences in group performance, (b) the degree to which test scores predict future performance, and (c) whether test scores from different tests measuring a similar construct produce a convergent association. |
| Consequences of Testing | Consequences of testing presents the intended and unintended consequences following the interpretation and use of test scores. Consequential evidence evaluates “the soundness of [test score] interpretations for their intended uses” (American Educational Research Association et al., 2014, p. 19). Unintended consequences warrant close examination, and consequential evidence may anticipate and proactively address unintended consequences. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Bostic, J.D. Validation Is a Methodology! Guideposts for Assessment Development and Validation. Educ. Sci. 2026, 16, 565. https://doi.org/10.3390/educsci16040565
Bostic JD. Validation Is a Methodology! Guideposts for Assessment Development and Validation. Education Sciences. 2026; 16(4):565. https://doi.org/10.3390/educsci16040565
Chicago/Turabian StyleBostic, Jonathan David. 2026. "Validation Is a Methodology! Guideposts for Assessment Development and Validation" Education Sciences 16, no. 4: 565. https://doi.org/10.3390/educsci16040565
APA StyleBostic, J. D. (2026). Validation Is a Methodology! Guideposts for Assessment Development and Validation. Education Sciences, 16(4), 565. https://doi.org/10.3390/educsci16040565
