Longitudinal Validation of EIP-Move for Assessing the Educational and Inclusive Potential of Physical Education and Sports Programs in Primary Schools
Abstract
1. Introduction
2. Study Objectives
3. Methods
3.1. Study Design
- Theoretical construction and content validity
- Pilot study
- Cross-sectional psychometric validation
- Longitudinal and predictive validation
3.2. Phase 1: Construction of the Instrument and Content Validity
3.2.1. Definition of the Construct and Specification of Dimensions
3.2.2. Generation of the Initial Pool of Items
3.2.3. Preliminary Qualitative Review (Face Validity)
3.2.4. Expert Panel: Composition and Selection Criteria
- Motor sciences and physical education teaching, with specific reference to primary school, in order to ensure that the items were consistent with the developmental characteristics of children and with practices that could actually be implemented in motor education contexts (Gallahue et al., 2012; Kirk, 2010).
- Special education and inclusive education, with experience in the design and evaluation of accessible educational contexts geared towards the participation of pupils with special educational needs, in line with the principles of educational inclusion and universal design for learning (Goodwin & Watkinson, 2000; UNESCO, 2017).
- Psychometrics and research methodology, with specific expertise in the construction, validation, and analysis of measurement tools, in order to ensure that the item evaluation process complied with the main methodological and psychometric standards (DeVellis, 2017; Streiner et al., 2015).
- (a)
- at least five years of professional or academic experience in the relevant field, or scientific output consistent with the topics of the study, a requirement considered essential to ensure informed judgments based on in-depth knowledge of the domain (Polit & Beck, 2006);
- (b)
- direct knowledge of physical education and basic sports contexts, considered essential for assessing the feasibility, applicability, and observability of the items in the real contexts of use of the tool (Patton, 2015; Kirk, 2010);
- (c)
- willingness to participate in one or more evaluation rounds, should critical issues arise that require further investigation or revision, in line with the iterative approach recommended in content validation processes (DeVellis, 2017; Streiner et al., 2015).
3.2.5. Rating Procedure and Assessment Tools
3.2.6. Calculation of the Content Validity Index (CVI) and Decision Criteria
- Acceptance, for items with I-CVI ≥ 0.78 and no significant qualitative issues emerging from the experts’ comments;
- Revision, for items with I-CVI ≥ 0.78 but accompanied by converging comments on aspects of clarity, formulation, or observability, or for items with values between 0.70 and 0.77 but characterized by strong theoretical consistency;
- Elimination, for items with I-CVI < 0.78 associated with comments indicating poor relevance, conceptual overlap with other items, or difficulty in reliable assessment in application contexts.
3.2.7. Item Review and Production of the Preliminary Version
- (a)
- elimination or merging of conceptually overlapping items;
- (b)
- reformulation of items characterised by excessive generality or risk of double-barrelledness;
- (c)
- maintaining a structural balance between the four theoretical dimensions.
3.3. Phase 2: Pilot Study
3.3.1. Sample and Context
3.3.2. Administration and Data Collection Procedure
- items perceived as ambiguous, too generic, or difficult to interpret unambiguously;
- items perceived as redundant or overlapping, especially within the same dimension;
- content considered relevant by respondents but not adequately represented in the tool;
- operational difficulties (e.g., need for additional examples, clarification on response anchors, or doubts about the reference period);
- perceived average completion time and sustainability in the specific context (e.g., immediate completion at the end of the lesson or at a later time).
3.3.3. Preliminary Analyses: Descriptive and Item Functioning
3.3.4. Criteria for Revision and Reduction of the Instrument
- Low variability and poor sensitivity: items characterised by marked floor or ceiling effects and reduced dispersion of responses. These items, being uninformative, limit the instrument’s ability to distinguish between different programmes. The decision to remove these items was taken mainly when the effect was also associated with redundancy with conceptually similar items.
- Interpretative ambiguity and generality: items perceived as too broad or interpretable in different ways (e.g., global judgements on climate or inclusion), or items that did not clearly indicate the observable object. In these cases, the priority was to reformulate them into more specific behavioural indicators.
- Poor observability: items considered difficult to assess due to a lack of evidence accessible to the compiler or because they depended on internal/not directly observable processes. In such cases, the item was revised to anchor it to verifiable behaviours or organisational conditions (e.g., routines, rules, adaptations, feedback).
- Intra-dimension redundancy: items with very high inter-item correlations (suggesting overlap) or reported as ‘repetitive’ by compilers. Redundancy was managed by retaining the item with the clearest, most observable and most discriminative wording, while still preserving coverage of the sub-aspects of the dimension.
- Low consistency with the theoretical dimension: items with weak or negative item-total correlations with respect to their own subscale, especially when accompanied by convergent qualitative feedback indicating uncertainty about meaning or difficulty in conceptual placement.
3.3.5. Output of the Pilot Phase
- it offered greater operational and interpretative clarity of the items, reducing the risk of ambiguity in compilation;
- it limited the risk of redundancy and conceptual overlap, improving the overall readability of the instrument;
- it showed better preliminary discriminative power, thanks to the removal of items characterised by marked floor or ceiling effects;
- it optimised application sustainability, reducing the compilation burden and making the instrument more easily usable in educational and sporting contexts.
3.4. Phase 3: Sample and Validation Procedure
3.4.1. Predictive Validity: Criterion Definition and Time Horizon
3.4.2. Sample Justification, Sampling Procedure, and Longitudinal Framework
Sample Size and Characteristics Across Measurement Waves
- T0 (baseline): beginning of the school or sports year, once organisational routines had stabilised;
- T1 (post-test): end of the school year or sports season;
- T2 (follow-up): approximately 12 months after T1, to assess temporal stability and predictive validity.
Sampling Procedure
- context of implementation (primary schools vs. grassroots sports clubs);
- organisational characteristics (group size, weekly frequency, session duration);
- pedagogical approaches (play-oriented, mixed, or sport-specific programmes);
- presence of pupils with special educational needs.
Justification of Sample Size for CFA and Longitudinal Analyses
3.4.3. Participants and Units of Analysis
3.4.4. Administration Procedure, Assessor Profile, and Inter-Rater Reliability
Assessor Profile and Role
- Teachers/coaches responsible for programme implementation, who completed the EIP-Move as an internal evaluation based on their continuous and direct involvement in the activities;
- Trained external observers, involved for methodological control purposes and for the estimation of inter-rater reliability.
Assessment Procedure and Observability of Items
Inter-Rater Reliability
3.4.5. Program Inclusion and Exclusion Criteria
3.4.6. Administration Procedure and Role of Assessors
- completion by teachers/coaches, as an internal evaluation of the program;
- evaluation by trained external observers, for the purposes of methodological control and reduction of the risk of bias related to social desirability or self-evaluation.
3.4.7. Timeline of the Surveys
- T0 (beginning of the year): the initial survey was conducted in the first weeks of the school or sports year, once the basic organization of activities had been stabilized (formation of groups, definition of schedules, general methodological approach). This survey served to describe the initial level of educational and inclusive potential of the program and to establish a baseline, useful both for cross-sectional analyses and as a reference point for subsequent change analyses.
- T1 (end of year): the post-survey was conducted at the end of the school year or sports season, at a stage when the program had reached full operational maturity and educational routines were well established. This time frame was chosen to assess any changes in EIP-Move scores associated with the ongoing implementation of educational and inclusive practices, allowing for analysis of the tool’s sensitivity to change in relation to program improvement, adaptation, or stabilization processes throughout the year.
- T2 (12-month follow-up): the follow-up survey was conducted approximately twelve months after the T1 survey, with the aim of verifying the temporal stability of the measure and exploring the predictive validity of EIP-Move with respect to relevant outcomes and the continuity of educational and inclusive practices over time. This allowed us to assess whether the levels of educational and inclusive potential previously identified were associated with the maintenance of quality programs, the consistency of pedagogical choices, and the persistence of conditions conducive to participation and inclusion.
3.5. Statistical Analyses
3.5.1. Data Preparation and Preliminary Checks
3.5.2. Exploratory Factor Analysis (EFA)
3.5.3. Confirmatory Factor Analysis (CFA)
3.5.4. Reliability: Internal Consistency, Temporal Stability, and Inter-Rater Agreement
3.5.5. Measurement Invariance (Multi-Group and Longitudinal)
- Configurational invariance, to verify that the factor structure was identical across groups;
- Metric invariance, imposing equality of factor saturations, a necessary condition for comparing relationships between latent variables;
- Scalar invariance, imposing equality of thresholds/intercepts (in an appropriate form for ordinal items), an essential requirement for comparing latent mean scores between groups.
3.5.6. Longitudinal Analyses and Multilevel Models
3.5.7. Predictive Validity
4. Results
4.1. Factor Structure
4.2. Reliability
4.3. Measurement Invariance
4.4. Sensitivity to Change
5. Discussion
6. Practical Implications
7. Limitations and Future Developments
8. Conclusions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Bailey, R., Armour, K., Kirk, D., Jess, M., Pickup, I., Sandford, R., & BERA Physical Education and Sport Pedagogy Special Interest Group. (2009). The educational benefits claimed for physical education and school sport: An academic review. Research Papers in Education, 24(1), 1–27. [Google Scholar] [CrossRef]
- Bollen, K. A., & Curran, P. J. (2006). Latent curve models: A structural equation perspective. Wiley. [Google Scholar]
- Brown, T. A. (2015). Confirmatory factor analysis for applied research (2nd ed.). Guilford Press. [Google Scholar]
- Chen, F. F. (2007). Sensitivity of goodness of fit indexes to lack of measurement invariance. Structural Equation Modeling, 14(3), 464–504. [Google Scholar] [CrossRef]
- Cheung, G. W., & Rensvold, R. B. (2002). Evaluating goodness-of-fit indexes for testing measurement invariance. Structural Equation Modeling, 9(2), 233–255. [Google Scholar] [CrossRef]
- De Bosscher, V., Shibli, S., Westerbeek, H., & Van Bottenburg, M. (2015). Successful elite sport policies: An international comparison of the sports policy factors leading to international sporting success (2nd ed.). Meyer & Meyer Sport. [Google Scholar]
- DeVellis, R. F. (2017). Scale development: Theory and applications (4th ed.). Sage. [Google Scholar]
- Dunn, T. J., Baguley, T., & Brunsden, V. (2014). From alpha to omega: A practical solution to the pervasive problem of internal consistency estimation. British Journal of Psychology, 105(3), 399–412. [Google Scholar] [CrossRef] [PubMed]
- Enders, C. K. (2010). Applied missing data analysis. Guilford Press. [Google Scholar]
- Flora, D. B., & Curran, P. J. (2004). An empirical evaluation of alternative methods of estimation for confirmatory factor analysis with ordinal data. Psychological Methods, 9(4), 466–491. [Google Scholar] [CrossRef] [PubMed]
- Gallahue, D. L., Ozmun, J. C., & Goodway, J. D. (2012). Understanding motor development: Infants, children, adolescents, adults (7th ed.). McGraw-Hill. [Google Scholar]
- Goodwin, D. L., & Watkinson, E. J. (2000). Inclusive physical education from the perspective of students with physical disabilities. Adapted Physical Activity Quarterly, 17(2), 144–160. [Google Scholar] [CrossRef]
- Haegele, J. A., & Sutherland, S. (2015). Perspectives of students with disabilities toward physical education: A qualitative inquiry review. Quest, 67(3), 255–273. [Google Scholar] [CrossRef]
- Hallgren, K. A. (2012). Computing inter-rater reliability for observational data: An overview and tutorial. Tutorials in Quantitative Methods for Psychology, 8(1), 23–34. [Google Scholar] [CrossRef] [PubMed]
- Holgado-Tello, F. P., Chacón-Moscoso, S., Barbero-García, I., & Vila-Abad, E. (2010). Polychoric versus pearson correlations in exploratory and confirmatory factor analysis of ordinal variables. Quality & Quantity, 44(1), 153–166. [Google Scholar]
- Hox, J. J., Moerbeek, M., & van de Schoot, R. (2018). Multilevel analysis: Techniques and applications (3rd ed.). Routledge. [Google Scholar]
- Kirk, D. (2010). Physical education futures. Routledge. [Google Scholar]
- Kline, R. B. (2016). Principles and practice of structural equation modeling (4th ed.). Guilford Press. [Google Scholar]
- Koo, T. K., & Li, M. Y. (2016). A guideline of selecting and reporting intraclass correlation coefficients for reliability research. Journal of Chiropractic Medicine, 15(2), 155–163. [Google Scholar] [CrossRef] [PubMed]
- Li, C. H. (2016). Confirmatory factor analysis with ordinal data: Comparing robust maximum likelihood and diagonally weighted least squares. Behavior Research Methods, 48(3), 936–949. [Google Scholar] [CrossRef] [PubMed]
- Little, R. J. A., & Rubin, D. B. (2002). Statistical analysis with missing data (2nd ed.). Wiley. [Google Scholar]
- Marsh, H. W., Hau, K. T., & Wen, Z. (2004). In search of golden rules: Comment on hypothesis-testing approaches to setting cutoff values for fit indexes. Structural Equation Modeling, 11(3), 320–341. [Google Scholar] [CrossRef]
- Muthén, L. K., & Muthén, B. O. (2017). Mplus user’s guide (8th ed.). Muthén & Muthén. [Google Scholar]
- Patton, M. Q. (2015). Qualitative research & evaluation methods (4th ed.). Sage. [Google Scholar]
- Polit, D. F., & Beck, C. T. (2006). The content validity index: Are you sure you know what’s being reported? Critique and recommendations. Research in Nursing & Health, 29(5), 489–497. [Google Scholar]
- Polit, D. F., Beck, C. T., & Owen, S. V. (2007). Is the CVI an acceptable indicator of content validity? Appraisal and recommendations. Research in Nursing & Health, 30(4), 459–467. [Google Scholar] [CrossRef] [PubMed]
- Putnick, D. L., & Bornstein, M. H. (2016). Measurement invariance conventions and reporting. Developmental Review, 41, 71–90. [Google Scholar] [CrossRef] [PubMed]
- Raudenbush, S. W., & Bryk, A. S. (2002). Hierarchical linear models: Applications and data analysis methods (2nd ed.). Sage. [Google Scholar]
- Raykov, T., & Marcoulides, G. A. (2011). Introduction to psychometric theory. Routledge. [Google Scholar]
- Singer, J. D., & Willett, J. B. (2003). Applied longitudinal data analysis: Modeling change and event occurrence. Oxford University Press. [Google Scholar]
- Snijders, T. A. B., & Bosker, R. J. (2012). Multilevel analysis: An introduction to basic and advanced multilevel modeling (2nd ed.). Sage. [Google Scholar]
- Streiner, D. L., Norman, G. R., & Cairney, J. (2015). Health measurement scales: A practical guide to their development and use (5th ed.). Oxford University Press. [Google Scholar]
- Tomporowski, P. D., McCullick, B., Pendleton, D. M., & Pesce, C. (2015). Exercise and children’s cognition: The role of exercise characteristics and a place for metacognition. Journal of Sport and Health Science, 4(1), 47–55. [Google Scholar] [CrossRef]
- UNESCO. (2017). A guide for ensuring inclusion and equity in education. UNESCO Publishing. [Google Scholar]
- Widaman, K. F., Ferrer, E., & Conger, R. D. (2010). Factorial invariance within longitudinal structural equation models: Measuring the same construct across time. Child Development Perspectives, 4(1), 10–18. [Google Scholar] [CrossRef] [PubMed]
| Framework/Instrument | Primary Focus | Level of Analysis | Inclusion Explicitly Assessed | Program-Level Evaluation | Standardized Psychometric Validation | Longitudinal Use |
|---|---|---|---|---|---|---|
| PETE frameworks | Teacher education and instructional quality | Teacher/training | Partial (implicit) | No | No | No |
| UNESCO (2017) Inclusive Education Model | Policy principles and values | System/policy | Yes (normative) | No | No | No |
| Teaching quality scales in PE | Instructional behaviours | Teacher/lesson | Limited | No | Yes (often) | Rare |
| Inclusion-focused PE instruments | Inclusion practices | Student or teacher | Yes | Partial | Variable | No |
| EIP-Move (present study) | Educational and inclusive potential | Programme | Yes (core dimension) | Yes | Yes | Yes |
| Dimension 1—Pedagogical Quality of the Programme |
|---|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| Dimension 2—Inclusion and Participation |
|---|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| Dimension 3—Relational Climate and Safety |
|---|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| Dimension 4—Equity and Valuing Differences |
|---|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| Dimension | Items Involved | Initial Criticality | Before Revision (Example) | After Review (Final Version) |
|---|---|---|---|---|
| Pedagogical quality of the programme | Items 2, 11 and 16 | Possible overlap between consistency, progression and teaching planning | “The activities are coherent and well organised.”/“The programme is planned.” | 2. The motor activities follow a structured teaching progression. 11. The activities are consistent and interconnected over time. 16. Teaching choices are intentional and planned. |
| Inclusion and participation | Items 17–19, 22 and 28 | Risk of overlap between adaptations, management of diversity and participation | ‘The programme is inclusive.’/‘The activities are adapted to the children.’ | 17. The programme includes systematic adaptations to promote inclusion. 18. Differences in ability among pupils are managed effectively. 19. The programme avoids the systematic assignment of marginal roles. 22. Activity rules are flexible and adaptable. 28. Activities allow for different levels of participation. |
| Relational climate and safety | Items 33, 34, 42 and 46 | Conceptual contiguity between emotional climate and psychological safety | “The climate is positive.”/“The children feel safe.” | 33. Children feel free to express themselves during physical activities. 34. Mistakes are accepted as part of the learning process. 42. Children perceive an emotionally safe environment. 46. The emotional climate encourages active participation. |
| Fairness and valuing differences | Items 48, 49, 50 and 56 | Overlap between evaluative fairness, equal opportunities and non-discrimination | “The programme is fair.”/“There is no discrimination.” | 48. The programme promotes equal learning opportunities. 49. The assessment criteria are fair and transparent. 50. Competitive comparison is managed in an educational manner. 56. The programme avoids selective or discriminatory practices. |
| Dimension | Item | Before Revision (Example) | Critical Issues Identified | After Revision (Final Version) |
|---|---|---|---|---|
| Pedagogical quality of the programme | 4 | The activities promote learning and improve pupils’ performance. | Double item (learning + performance) | The activities are designed to promote learning and not just task completion. |
| 10 | The adult provides useful and motivating feedback. | Dual content (educational + motivational function) | The feedback provided by the adult has a formative function. | |
| 13 | The programme values both the process and the result. | Conceptual ambiguity (process vs. result) | The learning process is valued more than the final result. | |
| 14 | The activities encourage cooperation and self-regulation. | Possible double-barrelled item | Activities promote the development of transversal skills (e.g., cooperation, self-regulation). | |
| Inclusion and participation | 18 | The programme is inclusive and suitable for everyone. | Formulation too generic | The programme provides for systematic adaptations to promote inclusion. |
| 19 | Differences between children are well managed. | Ambiguity and overall assessment | Differences in ability among pupils are managed effectively. | |
| 21 | All children are able to participate and succeed. | Dual content (participation + success) | All pupils have real opportunities for success. | |
| 28 | Activities are adapted to children’s levels. | Vague wording | The activities allow for different levels of participation. | |
| Relational climate and safety | 33 | The group atmosphere is positive. | Overall assessment not observable | The relational climate of the group is positive and collaborative. |
| 36 | The children feel safe and free to express themselves. | Double-barrelled (safety + expression) | The children feel free to express themselves during physical activities. | |
| 37 | Mistakes are not a problem. | Generic formulation | Mistakes are accepted as part of the learning process. | |
| 42 | The environment is safe. | Ambiguity (physical vs. emotional safety) | Children perceive an emotionally safe environment. | |
| Fairness and valuing differences | 48 | The programme is fair for everyone. | Overall assessment | The programme promotes equal learning opportunities. |
| 49 | The assessment is fair. | Vague wording | The assessment criteria are fair and transparent. | |
| 50 | Competition is not negative. | Ambiguity of interpretation | Competitive comparison is managed in an educational manner. | |
| 56 | The programme does not discriminate. | Negative and generic wording | The programme avoids selective or discriminatory practices. |
| Dimension | Original Items (n) | Items Deleted/Merged | Revision Criterion | Final Items (n) |
|---|---|---|---|---|
| Pedagogical quality | 16 | 4 deleted/merged (e.g., 6, 11, 15, 16) | Redundancy between consistency, formative assessment and intentionality | 12 |
| Inclusion and participation | 16 | 4 eliminated/merged (e.g., 21, 27, 31, 32) | Overlap between continuous participation, success and belonging | 12 |
| Relational climate and safety | 15 | 3 eliminated/merged (e.g., 34, 46, 47) | Contiguity between emotional climate, cohesion and psychological security | 12 |
| Fairness and appreciation | 15 | 3 eliminated/merged (e.g., 49, 56, 61) | Redundancy between evaluative fairness, non-discrimination and self-perception | 12 |
| Total | 62 | 14 | — | 48 |
| Dimension 1—Pedagogical Quality of the Programme (12 Items) |
|
|
|
|
|
|
|
|
|
|
|
|
| Dimension 2—Inclusion and participation (12 items) |
|
|
|
|
|
|
|
|
|
|
|
|
| Dimension 3—Relational climate and safety (12 items) |
|
|
|
|
|
|
|
|
|
|
|
|
| Dimension 4—Fairness and valuing differences (12 items) |
|
|
|
|
|
|
|
|
|
|
|
|
| Dim. | Item | Preliminary Version (48 Items) | Criterion Applied | Outcome |
|---|---|---|---|---|
| 1 | 3 | The programme offers a variety of meaningful motor experiences. | Ambiguity/generic nature | Reformulated → “The programme offers a variety of meaningful motor experiences for the development of pupils.” |
| 1 | 7 | The programme encourages autonomy and initiative in children. | Double-barrelled/observability | Rewritten → “The programme encourages children’s independence in carrying out activities.” |
| 1 | 11 | The learning process is valued more than the final result. | Low observability | Removed (redundant with Items 4 and 9, more observable) |
| 2 | 13 | All children actively participate in the proposed activities. | Ceiling effect | Reformulated → “Most children actively participate in the proposed activities.” |
| 2 | 15 | Differences in ability among pupils are managed effectively. | Generic | Reworded → “Differences in ability among pupils are managed through adaptations of activities.” |
| 2 | 21 | The programme prevents forms of implicit exclusion. | Poor observability | Reformulated → “The programme includes actions aimed at preventing forms of implicit exclusion.” |
| 2 | 24 | Participation time is distributed equally among children. | Intra-dimensional redundancy | Removed (overlaps with Items 22 and 16) |
| 3 | 25 | The relational climate of the group is positive and collaborative. | Overall assessment | Reformulated → “The group’s relational climate fosters collaboration and positive interactions.” |
| 3 | 30 | The programme prevents incidents of ridicule or exclusion. | Poor observability | Reformulated → “The programme adopts strategies to prevent episodes of ridicule or exclusion.” |
| 3 | 34 | Children perceive an emotionally safe environment. | Redundancy/low discriminative power | Removed (covered by Items 28 and 29) |
| 3 | 36 | Activities respect individual times and limits. | Intra-dimensional redundancy | Removed (overlaps with Items 8 and 22) |
| 4 | 40 | Competitive comparison is handled in an educational manner. | Ambiguity | Reformulated → “Competitive comparison, when present, is handled in an educational manner.” |
| 4 | 41 | Individual progress is valued more than performance. | Redundancy | Removed (conceptually overlapping with Items 38 and 46) |
| 4 | 46 | Activities enable each child to feel competent. | Genericity | Reformulated → “The activities allow each child to experience a sense of competence.” |
| 4 | 47 | The programme supports pupils’ self-esteem. | Low theoretical consistency | Removed (distal construct, not directly observable) |
| Dimension 1—Pedagogical Quality of the Programme |
| (Items 1–10) |
|
|
|
|
|
|
|
|
|
|
| Dimension 2—Inclusion and participation |
| (Items 11–20) |
|
|
|
|
|
|
|
|
|
|
| Dimension 3—Relational climate and safety |
| (Items 21–30) |
|
|
|
|
|
|
|
|
|
|
| Dimension 4—Fairness and valuing differences |
| (Items 31–40) |
|
|
|
|
|
|
|
|
|
|
| χ2 | df | χ2/df | CFI | TLI | RMSEA | 90% CI RMSEA | SRMR |
|---|---|---|---|---|---|---|---|
| 864.32 | 344 | 2.51 | 0.95 | 0.94 | 0.044 | [0.040–0.048] | 0.047 |
| Dimension | Items | Cronbach’s α | McDonald’s ω |
|---|---|---|---|
| Pedagogical Quality | 10 | 0.91 | 0.93 |
| Inclusion & Participation | 10 | 0.89 | 0.91 |
| Relational Climate & Safety | 10 | 0.90 | 0.92 |
| Equity & Valuing Differences | 10 | 0.88 | 0.90 |
| Dimension | ICC Model | ICC | 95% CI |
|---|---|---|---|
| Pedagogical Quality | ICC(2, 1) | 0.82 | [0.76–0.87] |
| Inclusion & Participation | ICC(2, 1) | 0.85 | [0.79–0.89] |
| Relational Climate & Safety | ICC(2, 1) | 0.80 | [0.73–0.85] |
| Equity & Valuing Differences | ICC(2, 1) | 0.77 | [0.69–0.83] |
| Predictor (T0) | Criterion Outcome | Time Point | β | SE | p |
|---|---|---|---|---|---|
| Total EIP-Move score | Active student participation | T1 | 0.41 | 0.06 | <0.001 |
| Total EIP-Move score | Relational climate quality | T1 | 0.38 | 0.07 | <0.001 |
| Total EIP-Move score | Psychological safety | T1 | 0.35 | 0.06 | <0.001 |
| Total EIP-Move score | Programme continuity | T2 | 0.33 | 0.08 | <0.001 |
| Inclusion & Participation | Active participation | T1 | 0.44 | 0.05 | <0.001 |
| Relational Climate & Safety | Psychological safety | T1 | 0.47 | 0.06 | <0.001 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Di Palma, D. Longitudinal Validation of EIP-Move for Assessing the Educational and Inclusive Potential of Physical Education and Sports Programs in Primary Schools. Educ. Sci. 2026, 16, 374. https://doi.org/10.3390/educsci16030374
Di Palma D. Longitudinal Validation of EIP-Move for Assessing the Educational and Inclusive Potential of Physical Education and Sports Programs in Primary Schools. Education Sciences. 2026; 16(3):374. https://doi.org/10.3390/educsci16030374
Chicago/Turabian StyleDi Palma, Davide. 2026. "Longitudinal Validation of EIP-Move for Assessing the Educational and Inclusive Potential of Physical Education and Sports Programs in Primary Schools" Education Sciences 16, no. 3: 374. https://doi.org/10.3390/educsci16030374
APA StyleDi Palma, D. (2026). Longitudinal Validation of EIP-Move for Assessing the Educational and Inclusive Potential of Physical Education and Sports Programs in Primary Schools. Education Sciences, 16(3), 374. https://doi.org/10.3390/educsci16030374