Next Article in Journal
Developing and Validating an EE–SEP Administration Model for Thai Primary Schools
Previous Article in Journal
Enhancing Teacher Educators’ Leadership Through Distributed Pedagogical Practice in Kenyan Preservice Education
Previous Article in Special Issue
The Critical Role of Science Teachers’ Readiness in Harnessing Digital Technology Benefits
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Teacher Professional Development and Student Mathematics Achievement: A Meta-Analysis of the Effects and Moderators

1
School of Teacher Education, College of Community Innovation and Education, University of Central Florida, Orlando, FL 32816, USA
2
Learning Sciences Research Institute, Seoul National University, Seoul 08826, Republic of Korea
3
Center on Teaching and Learning, University of Oregon, Eugene, OR 97403, USA
4
Department of Special Education, College of Education, The University of Texas at Austin, Austin, TX 78712, USA
*
Author to whom correspondence should be addressed.
Educ. Sci. 2025, 15(9), 1177; https://doi.org/10.3390/educsci15091177
Submission received: 5 June 2025 / Revised: 29 August 2025 / Accepted: 3 September 2025 / Published: 8 September 2025

Abstract

An essential element for increasing student mathematics achievement is providing teachers with professional development (PD) aimed at the design and delivery of high-quality mathematics instruction. To date, however, there is a lack of consistent data on the efficacy of PD on student outcomes; moreover, there is a need to explore PD characteristics as moderators of student outcomes. The purpose of this meta-analysis was to synthesize the effects of teacher PD on mathematics outcomes for students in PreK through 12th grade. Additionally, this study explored whether specific characteristics (i.e., grade level, format, PD focus, PD days, and inclusion of students with or at risk of disabilities) served as potential moderators of the effect of PD programs. The 20 studies included in the review investigated PD intended for in-service teachers who work with a full range of learners, including students with or at risk of disabilities. The results showed that there was a positive average effect of PD on student mathematics outcomes (g = 0.34, 95% PI = [−0.47, 1.15]), with wide heterogeneity of most effects ranging from −0.20 to 5.83. In addition, the five moderators examined in this meta-analysis were not significantly correlated with the relation between PD efficacy and student mathematics outcomes. Recommendations for improving the features of PD programs as well as exploring mechanisms of change hypothesized to improve student mathematics outcomes are discussed.

1. Introduction

With national efforts to improve mathematics achievement for students, professional development (PD) for teachers has been put forth as a central focus for mathematics education reform (National Governors Association Center for Best Practices & Council of Chief State School Officers, 2010; National Research Council, 2011). In this study, we operationalized PD as deliberate training, coaching, or professional learning communities (PLCs) to support in-service teachers’ content knowledge or pedagogical practices for teaching mathematics to the full range of learners in PreK–12 classrooms. PD can play a key role in improving teacher quality and effective delivery of instruction (Darling-Hammond et al., 2017; Koellner et al., 2024; Sims & Fletcher-Wood, 2021), which in turn may increase student mathematics achievement (Dash et al., 2012; Department for Education, 2016; K. K. Hill et al., 2017; Prast et al., 2018).
Despite the acknowledged importance of PD in teacher learning, a consensus is lacking in the literature about the overall effects of PD on student outcomes, as well as potential moderators of those effects (Garrett et al., 2019; Lynch et al., 2025; Kraft et al., 2018). Empirical evidence from studies focused on PD programs for mathematics teachers has returned mixed results regarding the effects of PD on student outcomes (e.g., for positive results see Brendefur et al., 2016, 2022; Copur-Gencturk et al., 2024; Jacobs et al., 2007; Roschelle et al., 2010; for null results see Garet et al., 2010; Jacob et al., 2017; Randel et al., 2016; Santagata et al., 2010). Thus, the purpose of this meta-analysis is to quantify the efficacy of PD on student mathematics achievement outcomes and examine features of PD programs and student-level characteristics as moderators of those outcomes.

1.1. Previous Reviews of Studies of PD on Student Mathematics Outcomes

To date, we identified only five systematic reviews or meta-analyses that summarized the effects of PD on student mathematics outcomes. Kennedy (1998) conducted a systematic review of 12 studies published between 1979 and 1996 that focused on examining the content of PD and its subsequent effects on student outcomes in mathematics and science. Eight of the 12 studies focused specifically on mathematics outcomes. Of the mathematics studies, Kennedy found the studies focusing on subject matter knowledge and on student learning of particular subject matter (e.g., number facts, computation) yielded a larger average effect (effect sizes range from 0.42 to 0.70) compared to studies focusing on teachers’ behaviors (e.g., teaching practices, changes in teacher knowledge; effect sizes range from −0.41 to 0.29).
Clewell et al. (2005) extended Kennedy (1998) by conducting a systematic review of 18 studies published between 1979 and 2004. They also focused on the effects of PD on student outcomes in mathematics and science. Clewell et al. looked at the types of content focus based on Kennedy’s classification, and reviewed the key PD characteristics (e.g., subject matter, grade span, participants, total contact hours, study duration, and effect sizes). They concluded that effective PD included content tied to curriculum. The authors also reported that effective PD included a minimum of 80 contact hours to observe a change in teachers’ instructional practices and a minimum of 160 contact hours to observe a change in the classroom environment.
Overall, the studies by Kennedy (1998) and Clewell et al. (2005) represented an important shift toward focusing on student learning outcomes in in-service teacher PD. Consequently, Blank and de las Alas (2009) conducted a meta-analysis of 16 studies, published between 1989 and 2007, synthesizing the effects of PD on student mathematics outcomes. Blank and de las Alas also investigated to what extent the characteristics of PD (e.g., content focus, duration, coherence, active learning, and collective participation of teachers) contributed to PD effectiveness. Their results suggested mathematics-focused PD yielded a significant and positive mean effect (pre-post design mean ES = 0.21; post-only design mean ES = 0.13) on student mathematics achievement. The authors reported larger average effects on student mathematics achievement for studies conducted at the elementary level, compared to middle and high school.
Yoon et al. (2007) and Gersten et al. (2014) investigated the effect of PD on student mathematics outcomes by applying the What Works Clearinghouse (WWC) research design standards. In a review of more than 1300 studies of PD (including science, mathematics, reading, and English/language arts) conducted between 1986 and 2006, Yoon et al. found that only nine studies demonstrated the methodological rigor necessary to meet the WWC standards. Of these nine studies, all focused on elementary grades and only two focused on mathematics. The sole negative effect across these nine studies was from a mathematics study (fraction computation). Gersten et al. extended the work of Yoon et al. by focusing specifically on PD in the area of mathematics in a review of 32 studies published between 2006 and 2012. Of five studies that met WWC design standards, Gersten et al. found two studies that demonstrated positive effects (ES range = 0.09–0.84) on student mathematics achievement (i.e., Perry & Lewis, 2011; Sample McMeeking et al., 2012). While Yoon et al. and Gersten et al. focused on investigating the quality of PD studies, the current study aims to fill a gap in the literature by using meta-analysis to determine the average effect of teacher PD on student mathematics achievement and the extent to which specific features of PD affect student outcomes. In short, previous studies have tended to focus on mathematics, science, and reading together, or have focused on study quality; the literature about mathematics specifically needs to be brought up to date.

1.2. Potential Moderators of PD Efficacy on Student Outcomes

Despite growing research on PD and its efficacy, there is a lack of consensus about which moderators affect the relationship between PD and student outcomes. Previous studies claimed that inconsistency among PD characteristics made it difficult for researchers to investigate the extent to which those characteristics determine PD effectiveness and in turn influence student outcomes (Darling-Hammond et al., 2017). Because of this variation, we limited our investigation to the following five moderating variables: (a) grade level, (b) format, (c) PD focus, (d) PD days, and (e) inclusion of students with or at risk of disabilities. Moderators (a) through (d) are malleable characteristics that appeared frequently as PD characteristics in our review of the literature. By focusing on these malleable characteristics, the results of the current meta-analysis will provide school administrators with targeted recommendations as they identify PD needs for teachers of mathematics. Although not a malleable characteristic, moderator (e) was selected to investigate if the effects of PD on student outcomes were different for students who have different learning needs. A discussion of the literature on these chosen variables and a rationale for exploring each one follows.

1.2.1. Grade Level

According to previous studies, PD may have different effect sizes on student outcomes depending on grade levels. Blank and de las Alas (2009) found that studies conducted in elementary grades had larger effects than those conducted in secondary grades. However, the results varied depending on when separate effect sizes were computed across grade levels (e.g., positive results for Grade 6, negative results for Grade 7, and null effects on Grade 8; META Associates, 2006). This finding has been replicated in more recent studies. For example, Taylor et al.’s (2018) analysis reported inconsistent associations between student outcomes and grade level. Similarly, Lynch et al. (2019) found no significant relation between these variables.

1.2.2. Format

Teacher PD can be delivered in various ways, in person or virtually, including training, workshops, coaching, PLCs, or a combination of these formats. As PD has evolved over time, it has moved beyond traditional classroom settings, leveraging online platforms and hybrid models to train in-service teachers (Darling-Hammond et al., 2017; Stevenson et al., 2015). The number of online PD options has grown due to its perceived benefits, including flexibility, access, networking, and cost effectiveness (Lay et al., 2020). PD can focus on a single component, such as in-person training, or it can be combined with other formats like integrating virtual training with coaching support (Piper et al., 2018). However, rigorous evidence to support effective PD design and delivery remains scarce (Sims & Fletcher-Wood, 2021; Yoon et al., 2007). Further, previous studies have concluded that the variety of formats used in PD makes it difficult to determine the extent to which format affects the relationship between PD and student outcomes (Didion et al., 2020).

1.2.3. PD Focus

The focus of PD can be subject-specific content, pedagogy, or a mixture of the two. In recent years, there has been a growing interest in how content-focused PD in mathematics deepens teachers’ content knowledge (Ball et al., 2008; Garet et al., 2016). Some studies indicated that content-focused PD had a positive impact on student achievement, although most of those results were not statistically significant (e.g., Jacobs et al., 2007; Jacob et al., 2017). Other researchers reported that PD highlighting a combination of content knowledge and pedagogical strategy was effective in improving student outcomes (Clewell et al., 2005; Ingvarson et al., 2005). This notion stems from the belief that teachers should have a solid foundation in both content knowledge and pedagogical strategies that can be adapted in their classrooms.

1.3. PD Days

PD days is defined as the number of hours spent on PD (Kennedy, 2016). Several researchers have focused on identifying the optimal number of days that teachers should attend PD to maximize student outcomes (Darling-Hammond et al., 2009; Desimone, 2009; Garet et al., 2010, 2016; Guskey & Yoon, 2009; Yoon et al., 2007). For example, Yoon et al. (2007) reported that PD duration should consist of more than 14 h because less than that demonstrated little to no effect on student achievement outcomes. Other scholars, however, have reported the need for greater intensities, including 20 h (Desimone, 2009), 30 h (Guskey & Yoon, 2009), 49 h (Darling-Hammond et al., 2009), and 68 h (Garet et al., 2010).

Inclusion of Students with or at Risk of Disabilities

There has been an increasing concern regarding a lack of studies demonstrating teachers’ abilities to support diverse student populations in their classrooms (Prast et al., 2018; Valiandes & Neophytou, 2018). Approximately 65% of students with disabilities spend more than 80% of their school day in general education classrooms (National Center for Education Statistics, 2019). Thus, it is vital to identify if PD effects vary when classrooms include students with disabilities. The provision of instructional support to students with a history of low achievement may be especially critical for them to benefit from rigorous mathematics instruction to minimize the observed widening of achievement gaps over time (Fuchs et al., 2021). The literature suggests that teachers are rarely supported with PD that focuses on students with disabilities (Darling-Hammond et al., 2009), creating difficulty when it comes to supporting these students and other struggling learners (e.g., students with or at risk for mathematics learning disabilities, students who are bilingual) in their classrooms (Allsopp & Haley, 2015). In conclusion, the five moderators that are the target of the current analysis have mixed findings across studies, indicating a need for further investigation with meta-analytic techniques.

1.4. Purpose of the Current Study

The present study fills a gap in the literature in the following ways. First, compared to the previous literature on PD (e.g., a combination of reading, mathematics, and science), the current study focuses exclusively on mathematics PD, as a potential method of remedying the decrease in mathematics achievement in the nation (National Center for Education Statistics, 2022). Second, the current study includes student-level characteristics as moderators, allowing us to investigate how these characteristics affect the relation between PD and student outcomes. Third, the current study adapts up-to-date meta-analytic procedures (i.e., meta-analysis with robust variance estimation using a correlated and hierarchical effects working model; Pustejovsky & Tipton, 2022) for better estimation of effect sizes and moderator effects. Fourth, research on PD has grown, leading to the need to summarize recent research findings by synthesizing the past 25 years of studies, between 2000 and 2024, to understand current practices. Taken together, a key aim of the current study is to add to the evidence that PD can enhance students’ mathematics achievement. The following research questions were addressed:
  • What are the main effects of PD on mathematics achievement for students in PreK–12?
  • To what extent do features of PD (i.e., grade level, format, PD focus, PD days, grade level, and inclusion of students with or at risk of disabilities) moderate its effects on student mathematics achievement?

2. Method

2.1. Search Procedures

We conducted a comprehensive review of the literature. First, we searched through four databases: Academic Search Complete, Education Source, ERIC, PsycINFO. The Boolean search string was: “professional development” OR “career development” OR “professional education” OR “teacher development” OR train* OR coach* OR “communit* of practice” OR PLC OR “professional learning communit*” OR “inservice teacher education” OR AND math* OR algebra OR geometry AND teachers. The search was completed within peer reviewed journals from January 2000 to August 2024. Second, we conducted a manual search of the following journals specializing in teacher education and mathematics: Review of Educational Research, Journal of Research on Educational Effectiveness, Journal of Mathematics Teacher Education, and Teaching and Teacher Education from January 2000 to August 2024. During the manual search, we reviewed the tables of contents of each journal to identify any additional studies that met this study’s inclusion criteria. Further, we conducted a backward and forward search of included studies to identify other relevant studies. Finally, we conducted a first author search of all included studies to investigate the first author’s other published studies.

2.2. Inclusion Criteria and Exclusion Criteria

To be included in this meta-analysis, studies had to meet seven inclusion criteria.
  • Topic of PD. Each study included teacher PD. The format of the PD in the studies varied, including training, workshops, coaching, or PLCs. PD learning goals, as defined by the studies, all focused on improving teacher knowledge as a way of increasing student mathematics achievement.
  • Population. Each study included in-service teachers. Studies including preservice teachers were excluded.
  • Grade. Each study included teachers who taught PreK–12.
  • Time. Each study was published in English between January 2000 and August 2024. We selected a start date of 2000, as this date aligned with the release of the National Council of Teachers of Mathematics (NCTM, 2000) Principles and Standards, which was a new impetus for altering the direction of mathematics standards in the United States. The 2000 NCTM Principles and Standards included recommendations for preschool learners not present in the initial 1989 NCTM standards, as well as more details for specific skills to be taught at each grade, which significantly impacted the broader standards movement in U.S. education.
  • Study design. Each study used a randomized controlled trial or a quasi-experimental design. Studies that used literature reviews, single-subject designs, qualitative methods, and quantitative methods that used descriptive analyses, correlational designs, or mixed methods were excluded. This choice was made because a meta-analysis requires the selection of certain types of study designs to extract necessary information (e.g., pretest and posttest scores for treatment and control group).
  • Type of publication. Each study was published in peer-reviewed educational journals in English. Gray literature (e.g., dissertations, book chapters, conference proposals, and technical reports) was excluded. In this study, our exclusion criterion sought to reduce the complexity and heterogeneity of gray literature, making the processes of search, analysis, and coding more manageable and replicable (Zhang et al., 2020). In addition, we made this decision to ensure that each study was reviewed by experts in the field as part of the peer review process.
  • Outcome measure. Each study included at least one student mathematics achievement outcome measure.

2.3. Screening Process and Study Identification

This study used two search methods to identify relevant studies (Figure 1). An electronic search identified a total of 12,768 studies. During this initial screening, 4314 studies were immediately removed because they were duplicates (k = 2302) or not peer-reviewed articles (k = 2012), yielding 8454 articles. The 8454 studies were then screened based on titles of the articles to identify if they were on the topic of PD in mathematics. A number of studies were removed as they were irrelevant to PD (k = 6998), irrelevant to math (k = 633), or identified as additional duplicates (k = 192); this title screening search yielded 631 articles. In the following step, the abstract search was reviewed to determine whether the research met the seven inclusion criteria. During this process, a number of studies were removed because they did not meet one or more of the following inclusion criteria: (a) research study (k = 28), (b) group design (k = 239), (c) preK-12 teachers (k = 33), (d) in-service teachers (k = 93), (e) about PD (k = 35), (f) focused on PD on student mathematics outcomes (k = 11). This electronic abstract screening search yielded 192 articles, qualifying for a full-text screening. In addition, 91 studies from the other search methods, including table of contents hand search (k = 46), forward and backward searches (k = 43), and first author searches (k = 2), qualified for full-text screening. The electronic and other search methods yielded 283 studies, which qualified for a full-text screening. During the full-text screening process, a number of studies were removed for the following reasons: (a) the studies did not use group designs (k = 89), (b) did not focus on preK-12 in-service teachers (k = 23), (c) did not focus on PD in mathematics (k = 54), (d) did not have outcome measures for students (k = 71), and (e) did not provide sufficient information to calculate effect sizes (k = 15), and (f) were duplicates (k = 11). At the end of the full-text screening process, 20 studies satisfied all the inclusion criteria.
The first and second authors led the screening process with two graduate research assistants (one master’s and one doctoral student in school psychology). The first author holds a PhD in special education and the second author holds a PhD in quantitative methods. Both authors have experience with conducting meta-analyses. The first author oriented the two research assistants through a 1 h screening training, an overview of inclusion criteria, instruction on the coding process, and practice opportunities with two articles. Then these two assistants coded two articles independently; the calculated interrater reliability (IRR) for this process was 91%. We calculated the IRR as [agreements/(agreements + disagreements) × 100]. Throughout the entire coding process, the first author held 1 h weekly meetings to address ongoing questions, discuss disagreements about coding, and provide continued support for the purpose of screening process reliability. The first author and the two graduate assistants double-coded 20% of the abstract, and the IRR was 89. The full screening was 20% double-coded, and the IRR was 84.5. All the discrepancies were discussed during these meetings and allowed all coders to reach 100% consensus. The second author double-checked the screening process by counting the number of studies that met eligibility for each stage, using the R program.

2.4. Coding Procedure

We developed a coding sheet and protocol based on the guidelines proposed by Cook et al. (2015) and the Institute of Education Sciences (2014). The coding sheet included the following study features: grade level, format, PD focus, PD days, and inclusion of students with or at risk of disabilities. Grade levels were categorized into a dichotomous variable: preK/kindergarten/elementary school (Grades 1–5) was coded as 0, and middle/high school (Grades 6–12) was coded as 1. Format refers to whether PD was provided via in-person workshop (coded as 0), or a combination of in-person workshop with another PD format such as online workshop, coaching, follow up meetings, and PLCs (coded as 1). PD focus of the content was coded as general pedagogical knowledge (coded as 0) or specific mathematics content knowledge (coded as 1), based on the primary focus. The combination of both pedagogical and content knowledge was coded as 2. PD days was coded as a continuous variable indicating days, such as 1 day or 10 days. A day was assumed to be 8 h long; if the study described the PD in hours, the variable was coded accordingly. For example, if the PD lasted only 4 h, PD days was coded as 0.5 days. Inclusion of students with or at risk of disabilities was coded as 1 if the PD included strategies for integrating these students into the classroom or presented data on this group. If PD did not include information on students with or at risk of disabilities or was focused on teaching strategies without any specific mention of these students, it was coded as 0.
The first author and second author tested the coding sheet using an example study. The IRR for this example study was 89%. After discussion, the two authors reached a 100% consensus, and the coding sheet was revised accordingly. To ensure the coding procedure was followed consistently, the first author independently coded all studies, and 20% of the studies were double-coded by the second author to establish IRR. These two coders held 1 h weekly meetings to resolve any discrepancies and discuss potential coding issues. Overall, the average IRR was 89.27% (range across categories = 81% to 96%). The IRR was 96.45% for grade level, for 90.15% format, 87.21% for PD focus, 81.56% for PD days, and 91% for the inclusion of students with disabilities. The first and second authors discussed any discrepancies and reached a consensus before the analysis stage; the final IRR was 100%.

2.5. Effect Size Calculation

We calculated the standardized mean of effect sizes of the PD for each study using Hedges’s g (Hedges, 1981). The effectiveness of PD reported by each study was based on the experimental or quasi-experimental group design, including independent groups in cross-sectional data (e.g., PD group and non-PD group), matched groups in repeated measures (e.g., pretest and posttest scores for one group), and independent groups of repeated measures (e.g., pretest and posttest scores for PD group and non-PD group). For studies that did not report the correlation between the pretest and posttest scores, we assumed the correlation of 0.6, considering the minimum correlation between repeated measures (Polly et al., 2017; Prast et al., 2018; WWC, 2020). If a study did not report sufficient descriptive statistics, we used other statistics to calculate the standardized mean difference. For example, we used F statistics in Brendefur et al. (2016) and McGatha et al. (2009), and Z statistics in Walker et al. (2012), with sample sizes to calculate the standardized mean difference.

2.6. Analysis

Before conducting meta-analyses, we examined the outliers in the raw effect sizes by using Tukey’s (1977) definition of outliers. If any values were less than the first quartile minus 1.5 × interquartile range (IQR) or greater than the third quartile plus 1.5 × IQR, we detected them as outliers and substituted ±1.5 × IQR. We conducted a sensitivity analysis with unadjusted and adjusted outliers to examine if the results changed significantly.
We employed a random effects meta-analysis with robust variance estimation (RVE) to examine the heterogeneity across the effect sizes. RVE addresses the issue of dependency among multiple effect sizes within a study, which could violate the assumption of the meta-analysis that each effect size should be independent (Hedges et al., 2010). In this case, RVE provides a robust standard error (SE) estimator even with the unknown covariance structure of the effect sizes. We assumed a 0.8 correlation for RVE and conducted a sensitivity analysis to examine if the effect size estimates differed across the correlation size. Additionally, a correlated and hierarchical effects working model (CHE; Pustejovsky & Tipton, 2022) was used while using RVE. This working model takes account of both within-study and between-study heterogeneity in true effect size, as in the hierarchical working model, while also handling correlation between the effect sizes within the study while assuming the constant correlation, as in the correlated working model.
Given the limited number of studies included, we employed small sample correction to adjust for inflated Type I errors in test statistics based on RVE using t tests (Tipton, 2015; Tipton & Pustejovsky, 2015). The meta-analysis was estimated using a restricted maximum likelihood estimation (REML). We reported tau-squared statistics, which assess heterogeneity among effect sizes, and I-squared statistics, the proportion of the observed variance that explains true differences in effect size estimates (Borenstein et al., 2011).
Following the meta-analysis, we conducted a meta-regression to evaluate the moderator effect on the relation between the PD and the student mathematics outcomes. We began by including all moderators in the model. Finally, we applied the Benjamini–Hochberg correction for the adjusted p value to control Type I errors across multiple statistical tests (Benjamini & Hochberg, 1995; Polanin & Pigott, 2015).
Lastly, we evaluated potential publication bias among effect size estimates without outliers. Publication bias may occur for several reasons, such as small-study effects or unpublished studies. To address publication bias, we used a funnel plot and a modified version of Egger’s regression test (Egger et al., 1997; Pustejovsky & Rodgers, 2019).
All the analyses were conducted in R 4.4.2 (R Core Team, 2024). We used the metafor (Viechtbauer, 2010), clubSandwich (Pustejovsky, 2017), and robumeta packages (Fisher & Tipton, 2015) to calculate effect sizes, conduct the meta-analysis, and demonstrate meta-analysis results. Only one value was missing for the intensity moderator, and we employed list-wise deletion to handle this missing datum. We followed the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines for reporting the results (Moher et al., 2009, 2015).

2.7. Examination of Outliers

According to Tukey’s (1977) definition of outliers, we detected seven outliers: two effect sizes from Saxe et al. (2001), three from Bruns et al. (2017), and two from Wang et al. (2013). We adjusted these outliers to the value of ±1.5 × IQR. The weighted averages of the effect of PD with adjusted and unadjusted outliers were both reported in the subsequent meta-analysis.

2.8. Publication Bias

Figure 2 illustrates the funnel plot that demonstrates the effect sizes included in the meta-analysis. The x axis of the graph indicates the effect sizes, and the y axis shows the SE. The smaller SE demonstrates the effect sizes with greater weights. This funnel plot shows the symmetry of the effect sizes, indicating no publication bias in the included studies. The result of Egger’s regression also showed that the included studies have no publication bias (t = −0.49, df = 85, p = 0.63). Thus, we concluded that the included studies provided representative overall effect sizes for the PD effect. The overall effect size was not adjusted to address potential publication bias.

3. Results

Within the 20 studies that examined the effect of PD, 87 effect sizes representing student achievement outcomes were identified. The number of effect sizes per study varied between one and 18 (M = 4.35 effect sizes per study). The total number of participating teachers across the studies was 21,271 (range: 14 to 5030 per study). The total number of participating students across the studies was 25,010 (range: 245 to 5658 per study). A list of study characteristics is shown in Table 1. The characteristics include title of PD, topics of PD, format, grade level, number of participants, number of included students with or at risk of disabilities, number of days, and measure of student mathematics outcomes.

3.1. Overall PD Effects on Student Mathematics Achievement

Figure 3 illustrates the distribution of the effect sizes in included studies using a forest plot. The size of the square symbols reflects the weight of each effect size, with larger squares representing effect sizes with greater weight. The variance of the effect sizes is shown as bars around the effect sizes, and the overall effect size is drawn with a diamond figure at the bottom of the forest plot. The details of the overall effect size values are also provided.
Table 2 shows the weighted average effect sizes of PD effects on student mathematics achievement. The average effect size with unadjusted outliers was 0.39 (SE = 0.11, 95% CI = [0.16, 0.61], 95% PI = [−0.63, 1.40], p = 0.001), and with adjusted outliers was 0.34 (SE = 0.10, 95% CI = [0.15, 0.53], 95% PI = [−0.47, 1.15], p < 0.001). The difference between the average effect size estimates from the two models was negligible. Both average effect sizes were large (Kraft, 2020) and statistically significant, indicating the positive weighted average of PD effects was statistically different from zero, regardless of the outlier adjustment. However, both prediction intervals contain zero, indicating that the true effect in future studies might not be different from zero. The sensitivity analysis showed negligible difference in the average effect size across different correlation values for RVE, ranging from 0 to 1. The estimated I 2 was 97.32 with τ 2 = 0.14 for within-study and 0.13 for between-study for the meta-analysis with unadjusted outliers. Also, the estimated I 2 was 95.82 with τ 2 of 0.09 for within-study and 0.08 for between-study for the same model with adjusted outliers, indicating a similar amount of heterogeneity across the studies between the two models. Therefore, we used the effect sizes that adjusted outliers throughout the analyses. To further determine the source of the heterogeneity, we conducted moderator analyses.

3.2. Moderator Analyses for PD Characteristics

We examined the moderating effects of five PD characteristics on the impact of teacher PD on student mathematics outcomes. The PD characteristics included (a) grade level, (b) format, (c) PD focus, (d) PD days, and (e) inclusion of students with or at risk of disabilities. Among the five moderators, none showed a significant effect (see Table 3). Specifically, the coefficient estimate for the PD focus was 1.30 larger for studies in which the PD focus was specific to math content compared to general instructional strategies (b = 1.30, SE = 0.63, 95% CI = [0.05, 2.56], p < 0.05). However, the Wald test indicated that the focus type was not a significant moderator overall. In addition, a combined focus compared to a general focus of PD did not significantly moderate the effect size (b = −0.29, SE = 0.39, 95% CI = [−1.07, 0.48], p = 0.45). The moderating effect of grade level (b = −0.79, SE = 0.57, 95% CI = [−1.93, 0.34], p = 0.17) did not significantly moderate the effect size. The moderating effect of format (combination vs. in-person workshop; b = 0.10, SE = 0.33, 95% CI = [−0.55, 0.75], p = 0.76) did not significantly moderate the effect size. The moderating effect of PD days (b = −0.01, SE = 0.01, 95% CI = [−0.03, 0.02], p = 0.54) did not significantly moderate the effect size. The moderating effect of inclusion of students with or at risk of disabilities (b = 0.55, SE = 0.40, 95% CI = [−0.25, 1.36], p = 0.17) did not significantly moderate the effect size.

4. Discussion

Many educational stakeholders would agree that high-quality PD is necessary for teachers to implement evidence-based practices and, in turn, positively impact student achievement in mathematics (Darling-Hammond et al., 2017). However, there is insufficient evidence for what constitutes effective PD. Therefore, we undertook a meta-analysis that explored features of teacher PD and student-level characteristics in mathematics and the impact they have on student achievement. We asked two research questions: (a) What are the main effects of PD on mathematics achievement for students in PreK–12? (b) To what extent do features of PD (i.e., grade level, format, PD focus, PD days, and inclusion of students with or at risk of disabilities) moderate its effects on student mathematics achievement? We concluded that teachers who received PD yielded a more positive impact on student mathematics outcomes when compared to teachers who did not receive PD. In addition, we found the five moderators did not significantly moderate the relation between PD efficacy and student mathematics outcomes.

4.1. Effect of PD on Student Mathematics Outcomes

Our primary research aim was to explore the effects of teacher PD on student mathematics outcomes. We included 20 studies in the meta-analysis; 98% of participants in PD were teachers of typically developing students from PreK to Grade 12. The results showed that there was a positive average effect of PD on student mathematics outcomes (g = 0.34), large enough to be of interest to policy makers in the field of education (Hedges & Hedberg, 2007). These findings add to the growing body of evidence suggesting that effective teacher PD has the potential to elevate student outcomes (e.g., Blank & de las Alas, 2009; Jacobs et al., 2007; Kraft et al., 2018; Roschelle et al., 2010). Despite the potential of training, prior research has indicated that evidence for the impact on student outcomes is still in the process of being established. For example, previous studies found the effect sizes for PD’s impact on student-level outcomes are relatively small, ranging from 0.14 to 0.37 (e.g., 0.14 for Egert et al., 2018; 0.21 for Markussen-Brown et al., 2017; 0.37 for Jung et al., 2018) compared to those reported in other meta-analyses that synthesized findings at the teacher level (e.g., 0.67 for Filderman et al., 2022; 0.57 for Gesel et al., 2021) and the classroom level (e.g., 0.45 for Egert et al., 2018). The current conversation regarding PD’s effect on student outcomes can be centered around either inconsistent findings of the effect of PD on student outcomes or a weak link between teacher impact on student outcomes compared to teacher outcomes. Although there is a strong belief that PD has a positive impact on student outcomes (Darling-Hammond et al., 2017), the consensus among researchers is that more research is needed to gather evidence of its impacts, particularly in mathematics. The number of PD programs focused on mathematics is relatively small compared to other content areas, such as language and literacy (Kraft et al., 2018; Piper et al., 2018). For instance, Brunsek et al. (2020), in their meta-analysis and systematic review in early childhood education, found the majority of PD focused on school readiness, social and emotional functioning, and language and literacy outcomes. The research investigating PD in mathematics is limited, and teachers receive insufficient PD to specifically enhance their mathematics content knowledge and pedagogical skills.
Our findings add value to the literature supporting the impact of PD on student outcomes and emphasizing the importance of PD. Approaches such as lesson study (Lewis & Perry, 2014, 2017) and teacher noticing (van Es & Sherin, 2002, 2021) are examples of long-standing PD programs in mathematics for K-12 teachers. However, continued investments in a range of instructional approaches have improved teachers’ competence and performance and students’ learning while also allowing for iterative adaptation. Thus, more PD opportunities for teachers are needed to explore effective ways of addressing instructional needs, particularly for students with low achievement or those with, or at risk of, disabilities. Given the low achievement trends in mathematics education (National Center for Education Statistics, 2022) and our study’s findings on the positive average effect of PD, providing teachers with more mathematics-specific PD could be a missing link. Thus, continued research on PD could investigate factors contributing to low achievement trends, supporting teachers to improve the trajectory of student mathematics achievement.

4.2. Moderator Analyses

With our second research question, we concluded that grade level did not significantly impact the overall efficacy of teacher PD on student mathematics outcomes. In other words, PD effects did not vary in studies examining teachers in early childhood or elementary schools compared to the studies with teachers in middle or high schools. This aligns with previous literature (e.g., Kennedy, 1998; Lynch et al., 2019; Taylor et al., 2018), which found no differences in intervention effects across grade levels. Although some studies (e.g., Blank & de las Alas, 2009) reported that grade level in elementary grades had larger effects than for secondary schools, our results did not find different effects across grades.
PD days was not found to moderate the effects of PD on student outcomes. These findings could be due to the variability of characteristics among the PD studied. Many researchers reported on the optimal number of days or contact hours of PD that teachers should receive to make significant differences in their practice to change student outcomes, at ranges including 20 h (Desimone, 2009), 30 h (Guskey & Yoon, 2009), 49 h (Darling-Hammond et al., 2009), and as much as 68 h (Garet et al., 2010). Although these studies claim that specific levels of intensity moderate PD effectiveness for student outcomes, our study aligns with the many meta-analyses showing that intensity does not moderate PD effectiveness (e.g., Blank & de las Alas, 2009; Kennedy, 2016; Kraft et al., 2018).
Likewise, our study found that format had no moderating effects on student mathematics outcomes. This may be partially due to the variety of PD types (e.g., coaching, PLC, online) found within the studies we examined. But, as with PD days, our finding aligns with previous studies that have attributed the lack of a moderating effect to the heterogeneity of PD formats (e.g., Blank & de las Alas, 2009; Didion et al., 2020; Yoon et al., 2007). For both PD days and format, more research needs to be conducted to more reliably explain the moderating effects of specific PD days and specific types of PD format. Until then, it is not possible to make definitive claims about which PD characteristics have the greatest moderating effects on student outcomes.
We also found no moderator effect for PD focus, which researchers have consistently found to include both specific subject content and pedagogical content (Blank & de las Alas, 2009; Markussen-Brown et al., 2017). This is somewhat unsurprising because, based on previous literature (Markussen-Brown et al., 2017), PD combining both subject and pedagogical content may be more beneficial for meeting the complexity of teachers’ needs than PD that only focuses on one or the other. This supports the assertions of Diamond and Powell (2011), who held that PD needs to offer sustained opportunities to understand specific content, recognize the challenges teachers encounter daily, and promote active learning. However, determining the specific effects of PD focus poses unique challenges for researchers, as the presence of both types of content makes it difficult to determine which component moderates the relation between PD and student outcomes (Didion et al., 2020). Another factor that makes this challenging is the ongoing debate about whether to use subject-specific observation instruments versus pedagogical content observation instruments for analyzing instructional quality (H. C. Hill & Grossman, 2013; Schoenfeld, 2018). Thus, it is likely that due to the complexity of PD and the varied research findings on its content focus, there is a paucity of solid evidence about effective PD design and implementation (Sims & Fletcher-Wood, 2021; Yoon et al., 2007).
Finally, the results of this meta-analysis indicate that the inclusion of students with or at risk of disabilities did not moderate effects. These findings align with Didion et al. (2020), who found no moderating effect of student characteristics related to disability status on the impact of PD on student reading outcomes. However, a null effect does not imply an absence of differences in the effectiveness of PD targeting specific student populations. This finding might be due to challenges in isolating the student population from the total sample, considering that many students with disabilities receive their education in inclusive settings. In fall 2021, according to the National Center for Education Statistics (2023), 95% of school-age children served under the Individuals with Disabilities Education Act were enrolled in regular schools. Despite this high enrollment in regular schools, teachers rarely receive PD tailored for supporting students with disabilities (Darling-Hammond et al., 2009). Because students with mathematics disabilities often have unique and persistent learning challenges and need individualized instruction, teachers in inclusive settings may require specific instructional strategies or intensified instruction to effectively support these students (Fuchs et al., 2021; Park et al., 2024). Further investigation is needed to determine if the effects of PD on student outcomes differ between students with disabilities and typically developing students.

4.3. Limitations and Implications for Research

There are several limitations to this study, and related key implications for research emerged from it. In this section, we provide suggestions for future research that may address these limitations. First, our meta-analysis only considered student mathematics outcomes, rather than teacher-level outcomes, as a way to measure the short-term effectiveness of PD in PreK–12. Our inclusion criteria focused only on studies that followed a randomized controlled trial or quasi-experimental design; other types of research design, such as qualitative studies, might have offered more contextualized information about the effectiveness of PD, information that could not be captured with student achievement outcomes only. While we considered student-level outcomes, researchers may also consider investigating the impact of teacher PD on teacher-level outcomes—for example, teachers’ knowledge, teaching efficacy, or self-esteem (Allsopp & Haley, 2015). The impact of PD on both teacher- and student-level outcomes, such as the improvement of content knowledge or the quality of delivering instruction, might bring different perspectives when designing effective PD. For example, prioritizing teachers’ knowledge, inquiry, and instructional resources might achieve sustained PD effects (e.g., Desimone, 2009; Lewis & Perry, 2017; Schoenfeld, 2018). Additionally, we only considered the short-term effects of teacher PD on student outcomes; researchers should further investigate the long-term effects. Given our specific focus on student-level outcomes and the short-term effects of PD on those outcomes, readers should be cautious about generalizing our results to teacher-level outcomes, non-school setting, and other content areas. Researchers may be able to refine the effectiveness of PD by examining how teacher- and student-level variables interact, investigating the effects of this interaction on teacher PD, and then applying the results when developing new PD.
Second, our analysis included mainly studies focusing on teachers of typically developing students, resulting in a limited view of PD effectiveness for students who receive special education services. Previous literature has indicated a lack of studies that show how PD can improve teachers’ abilities to support students with special needs, academic difficulties, or a mixed population of general and special education students in an inclusive setting (Darling-Hammond et al., 2009; Didion et al., 2020). There needs to be greater inclusion of students with disabilities in studies on teacher PD to create a more well-rounded view of student outcomes in PD studies (Clements et al., 2011; Griffin et al., 2018). To enhance student mathematics outcomes through PD, researchers should emphasize PD characteristics that best serve specific student populations, particularly those in inclusive settings. Thus, conducting more studies of PD specifically geared toward teachers who are instructing specific student populations may provide insights into important questions related to PD design; for instance, identifying characteristics of PD that are more effective for teachers instructing students with disabilities and designing the PD to emphasize those features may enhance student outcomes.
Third, it is important to note that the presence of the null effect of moderators does not mean that we can underestimate their implications for PD characteristics, because studies often describe PD in vague or incomplete terms. Because PD is complex and heterogeneous (Zhang et al., 2020), as noted above, researchers can learn more from PD with clearly defined components, allowing them to examine PD delivery in greater detail. However, in many of the studies in our meta-analysis, we found that PD design was often underdescribed, making it difficult to measure its effectiveness. An important implication was that due to the complexity of PD, it is crucial for researchers to provide detailed information on how the PD is conducted—for instance, implementation, content covered, and other relevant factors—to ensure quality, replicability, and applicability of findings on the impact of PD on student outcomes (Appelbaum et al., 2018; Desimone, 2009; Page et al., 2021). To provide more detailed information, future researchers who conduct PD meta-analyses may consider examining the type of information provided by authors such as quality of PD, research design, type of measures, and so on, to provide targeted recommendations for researchers who conduct and design PD studies.

5. Conclusions

This meta-analysis explored how effective teacher PD is for improving mathematics outcomes for students from PreK to Grade 12. The results of this study showed that PD had a positive average effect on student mathematics outcomes. In addition, we found that the five moderators explored in this meta-analysis had no significant impact on the relation between PD efficacy and student mathematics outcomes. It is important to note that while this meta-analysis found that the moderators were not statistically significant, features of PD recommended by the existing literature should not be ignored; researchers need to investigate the mechanism of effective PD characteristics on student outcomes. Thus, more studies on effective PD are needed to further articulate the relation between PD characteristics and student mathematics outcomes.

Author Contributions

Conceptualization, S.P., Y.R.L., G.N., and C.T.D.; methodology, S.P., Y.R.L., G.N., M.A.C., and C.T.D.; formal analysis, Y.R.L.; data curation, S.P., Y.R.L., G.N., and M.A.C.; writing—original draft preparation, S.P., Y.R.L., G.N., M.A.C., and C.T.D.; writing—review and editing, S.P., Y.R.L., G.N., M.A.C., and C.T.D.; project administration, S.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

References marked with an asterisk indicate studies included in this review.
  1. Allsopp, D. H., & Haley, K. C. (2015). A synthesis of research on teacher education, mathematics, and students with learning disabilities. Learning Disabilities: A Contemporary Journal, 13(2), 177–206. [Google Scholar]
  2. *Antoniou, P., & Kyriakides, L. (2013). A dynamic integrated approach to teacher professional development: Impact and sustainability of the effects on improving teacher behaviour and student outcomes. Teaching and Teacher Education, 29, 1–12. [Google Scholar] [CrossRef]
  3. Appelbaum, M., Cooper, H., Kline, R. B., Mayo-Wilson, E., Nezu, A. M., & Rao, S. M. (2018). Journal article reporting standards for quantitative research in psychology: The APA Publications and Communications Board task force report. American Psychologist, 73(1), 3–25. [Google Scholar] [CrossRef]
  4. Ball, D. L., Thames, M. H., & Phelps, G. (2008). Content knowledge for teaching: What makes it special? Journal of Teacher Education, 59(5), 389–407. [Google Scholar] [CrossRef]
  5. Benjamini, Y., & Hochberg, Y. (1995). Controlling the false discovery rate: A practical and powerful approach to multiple testing. Journal of the Royal Statistical Society: Series B (Methodological), 57(1), 289–300. [Google Scholar] [CrossRef]
  6. Blank, R. K., & de las Alas, N. (2009). Effects of teacher professional development on gains in student achievement: How meta-analysis provides scientific evidence useful to education leaders. Council of Chief State School Officers. Available online: https://files.eric.ed.gov/fulltext/ED544700.pdf (accessed on 28 August 2025).
  7. Borenstein, M., Hedges, L. V., Higgins, J. P., & Rothstein, H. R. (Eds.). (2011). Introduction to meta-analysis. John Wiley & Sons. [Google Scholar]
  8. *Brendefur, J., Champion, J., Strother, S., Thiede, K. W., & Osguthorpe, R. D. (2022). The effects of mathematics professional development on elementary student achievement. International Journal of Science and Mathematics Education, 20(6), 1079–1097. [Google Scholar] [CrossRef]
  9. *Brendefur, J., Strother, S., Thiede, K., Lane, C., & Surges-Prokop, M. J. (2013). A professional development program to improve math skills among preschool children in Head Start. Early Childhood Education Journal, 41(3), 187–195. [Google Scholar] [CrossRef]
  10. *Brendefur, J. L., Thiede, K. W., Strother, S., Jesse, D., & Sutton, J. (2016). The effects of professional development on elementary students’ mathematics achievement. Journal of Curriculum and Teaching, 5(2), 95–108. [Google Scholar] [CrossRef]
  11. *Bruns, J., Eichen, L., & Gasteiger, H. (2017). Mathematics-related competence of early childhood teachers visiting a continuous professional development course: An intervention study. Mathematics Teacher Education and Development, 19(3), 76–93. [Google Scholar]
  12. Brunsek, A., Perlman, M., McMullen, E., Falenchuk, O., Fletcher, B., Nocita, G., Kamkar, N., & Shah, P. S. (2020). A meta-analysis and systematic review of the associations between professional development of early childhood educators and children’s outcomes. Early Childhood Research Quarterly, 53, 217–248. [Google Scholar] [CrossRef]
  13. *Campbell, P. F., & Malkus, N. N. (2011). The impact of elementary mathematics coaches on student achievement. The Elementary School Journal, 111(3), 430–454. [Google Scholar] [CrossRef]
  14. Clements, D. H., Sarama, J., Spitler, M. E., Lange, A. A., & Wolfe, C. B. (2011). Mathematics learned by young children in an intervention based on learning trajectories: A large-scale cluster randomized trial. Journal for Research in Mathematics Education, 42(2), 127–166. [Google Scholar] [CrossRef]
  15. Clewell, B. C., Cosentino de Cohen, C., Campbell, P. B., Perlman, L., Deterding, N., & Manes, S. (2005). Review of evaluation studies of mathematics and science curricula and professional development models. Urban Institute. Available online: https://www.urban.org/research/publication/review-evaluation-studies-mathematics-and-science-curricula-and-professional-development-models (accessed on 28 August 2025).
  16. Cook, B. G., Buysse, V., Klingner, J., Landrum, T. J., McWilliam, R. A., Tankersley, M., & Test, D. W. (2015). CEC’s standards for classifying the evidence base of practices in special education. Remedial and Special Education, 36(4), 220–234. [Google Scholar] [CrossRef]
  17. Copur-Gencturk, Y., Li, J., Cohen, A. S., & Orrill, C. H. (2024). The impact of an interactive, personalized computer-based teacher professional development program on student performance: A randomized controlled trial. Computers & Education, 210, 104963. [Google Scholar] [CrossRef]
  18. Darling-Hammond, L., Hyler, M. E., & Gardner, M. (2017). Effective teacher professional development. Learning Policy Institute. Available online: https://learningpolicyinstitute.org/product/effective-teacher-professional-development-report (accessed on 28 August 2025).
  19. Darling-Hammond, L., Wei, R. C., Andree, A., Richardson, N., & Orphanos, S. (2009). Professional learning in the learning profession: A status report on teacher development in the United States and abroad. National Staff Development Council. [Google Scholar]
  20. *Dash, S., de Kramer, R. M., O’Dwyer, L. M., Masters, J., & Russell, M. (2012). Impact of online professional development or teacher quality and student achievement in fifth grade mathematics. Journal of Research on Technology in Education, 45(1), 1–26. [Google Scholar] [CrossRef]
  21. Department for Education. (2016). Standard for teachers’ professional development: Implementation guidance for school leaders, teachers, and organisations that offer professional development for teachers. Available online: https://assets.publishing.service.gov.uk/government/uploads/system/uploads/attachment_data/file/537031/160712_-_PD_Expert_Group_Guidance.pdf (accessed on 28 August 2025).
  22. Desimone, L. M. (2009). Improving impact studies of teachers’ professional development: Toward better conceptualizations and measures. Educational Researcher, 38(3), 181–199. [Google Scholar] [CrossRef]
  23. Diamond, K. E., & Powell, D. R. (2011). An iterative approach to the development of a professional development intervention for Head Start teachers. Journal of Early Intervention, 33(1), 75–93. [Google Scholar] [CrossRef]
  24. Didion, L., Toste, J. R., & Filderman, M. J. (2020). Teacher professional development and student reading achievement: A meta-analytic review of the effects. Journal of Research on Educational Effectiveness, 13(1), 29–66. [Google Scholar] [CrossRef]
  25. Egert, F., Fukkink, R. G., & Eckhardt, A. G. (2018). Impact of in-service professional development programs for early childhood teachers on quality ratings and child outcomes: A meta-analysis. Review of Educational Research, 88(3), 401–433. [Google Scholar] [CrossRef]
  26. Egger, M., Smith, G. D., Schneider, M., & Minder, C. (1997). Bias in meta-analysis detected by a simple, graphical test. BMJ, 315(7109), 629–634. [Google Scholar] [CrossRef]
  27. Filderman, M. J., Toste, J. R., Didion, L., & Peng, P. (2022). Data literacy training for K–12 teachers: A meta-analysis of the effects on teacher outcomes. Remedial and Special Education, 43(5), 328–343. [Google Scholar] [CrossRef]
  28. *Fisher, J. B., Schumaker, J. B., Culbertson, J., & Deshler, D. D. (2010). Effects of a computerized professional development program on teacher and student outcomes. Journal of Teacher Education, 61(4), 302–312. [Google Scholar] [CrossRef]
  29. Fisher, Z., & Tipton, E. (2015). Robumeta: An R-package for robust variance estimation in meta-analysis. arXiv, arXiv:1503.02220. [Google Scholar] [CrossRef]
  30. Fuchs, L. S., Fuchs, D., Hamlett, C. L., & Stecker, P. M. (2021). Bringing data-based individualization to scale: A call for the next-generation technology of teacher supports. Journal of Learning Disabilities, 54(5), 319–333. [Google Scholar] [CrossRef]
  31. Garet, M., Wayne, A., Stancavage, F., Taylor, J., Walters, K., Song, M., Brown, S., Hurlburt, S., Zhu, P., Sepanik, S., & Doolittle, F. (2010). Middle school mathematics professional development impact study: Findings after the first year of implementation (NCEE 2010-4009). National Center for Education Evaluation and Regional Assistance, Institute of Education Sciences, U.S. Department of Education. Available online: https://files.eric.ed.gov/fulltext/ED509306.pdf (accessed on 28 August 2025).
  32. Garet, M. S., Heppen, J. B., Walters, K., Parkinson, J., Smith, T. M., Song, M., Garrett, R., Yang, R., & Borman, G. D. (2016). Focusing on mathematical knowledge: The impact of content-intensive teacher professional development (NCEE 2016-4010). National Center for Education Evaluation and Regional Assistance, Institute of Education Sciences, U.S. Department of Education. Available online: https://files.eric.ed.gov/fulltext/ED569154.pdf (accessed on 28 August 2025).
  33. Garrett, R., Citkowicz, M., & Williams, R. (2019). How responsive is a teacher’s classroom practice to intervention? A meta-analysis of randomized field studies. Review of Research in Education, 43(1), 106–137. [Google Scholar] [CrossRef]
  34. Gersten, R., Taylor, M. J., Keys, T. D., Rolfhus, E., & Newman-Gonchar, R. (2014). Summary of research on the effectiveness of math professional development approaches. National Center for Education Evaluation and Regional Assistance. Regional Educational Laboratory Southeast at Florida State University. Available online: http://files.eric.ed.gov/fulltext/ED544681.pdf (accessed on 28 August 2025).
  35. Gesel, S. A., LeJeune, L. M., Chow, J. C., Sinclair, A. C., & Lemons, C. J. (2021). A meta-analysis of the impact of professional development on teachers’ knowledge, skill, and self-efficacy in data-based decision-making. Journal of Learning Disabilities, 54(4), 269–283. [Google Scholar] [CrossRef] [PubMed]
  36. Griffin, C. C., Dana, N. F., Pape, S. J., Algina, J., Bae, J., Prosser, S. K., & League, M. B. (2018). Prime online: Exploring teacher professional development for creating inclusive elementary mathematics classrooms. Teacher Education and Special Education, 41(2), 121–139. [Google Scholar] [CrossRef]
  37. Guskey, T. R., & Yoon, K. S. (2009). What works in professional development? Phi Delta Kappan, 90(7), 495–500. [Google Scholar] [CrossRef]
  38. Hedges, L. V. (1981). Distribution theory for Glass’s estimator of effect size and related estimators. Journal of Educational and Behavioral Statistics, 6(2), 107–128. [Google Scholar] [CrossRef]
  39. Hedges, L. V., & Hedberg, E. C. (2007). Intraclass correlation values for planning group-randomized trials in education. Educational Evaluation and Policy Analysis, 29(1), 60–87. [Google Scholar] [CrossRef]
  40. Hedges, L. V., Tipton, E., & Johnson, M. C. (2010). Robust variance estimation in meta-regression with dependent effect size estimates. Research Synthesis Methods, 1(1), 39–65. [Google Scholar] [CrossRef]
  41. Hill, H. C., & Grossman, P. (2013). Learning from teacher observations: Challenges and opportunities posed by new teacher evaluation systems. Harvard Educational Review, 83(2), 371–384. [Google Scholar] [CrossRef]
  42. Hill, K. K., Bicer, A., & Capraro, R. M. (2017). Effect of teachers’ professional development from Mathforward on students’ math achievement. International Journal of Research in Education and Science, 3(1), 67–74. [Google Scholar]
  43. *Hilton, A., Hilton, G., Dole, S., & Goos, M. (2016). Promoting middle school students’ proportional reasoning skills through an ongoing professional development programme for teachers. Educational Studies in Mathematics, 92(2), 193–219. [Google Scholar] [CrossRef]
  44. Ingvarson, L., Meiers, M., & Beavis, A. (2005). Factors affecting the impact of professional development programs on teachers’ knowledge, practice, student outcomes & efficacy. Education Policy Analysis Archives, 13(10), 1–26. [Google Scholar] [CrossRef]
  45. Institute of Education Sciences. (2014). What works clearinghouse procedures and standards handbook (Version 3.0). Available online: https://ies.ed.gov/ncee/wwc/Docs/referenceresources/wwc_procedures_v3_0_standards_handbook.pdf (accessed on 28 August 2025).
  46. Jacob, R., Hill, H., & Corey, D. (2017). The impact of a professional development program on teachers’ mathematical knowledge for teaching, instruction, and student achievement. Journal of Research on Educational Effectiveness, 10(2), 379–407. [Google Scholar] [CrossRef]
  47. *Jacobs, V. R., Franke, M. L., Carpenter, T. P., Levi, L., & Battey, D. (2007). Professional development focused on children’s algebraic reasoning in elementary school. Journal for Research in Mathematics Education, 38(3), 258–288. [Google Scholar] [CrossRef]
  48. Jung, P. G., McMaster, K. L., Kunkel, A. K., Shin, J., & Stecker, P. M. (2018). Effects of data–based individualization for students with intensive learning needs: A meta–analysis. Learning Disabilities Research & Practice, 33(3), 144–155. [Google Scholar] [CrossRef]
  49. Kennedy, M. (1998). Form and substance in inservice teacher education (Research Monograph No. 13). National Institute for Science Education. Available online: https://www.researchgate.net/publication/242434041_Form_and_Substance_in_Inservice_Teacher_Education (accessed on 28 August 2025).
  50. Kennedy, M. M. (2016). How does professional development improve teaching? Review of Educational Research, 86(4), 945–980. [Google Scholar] [CrossRef]
  51. Koellner, K., Seago, N., Riske, A., Placa, N., & Carlson, D. (2024). Teachers’ perceptions and uptake of professional development overtime. International Journal of Educational Research Open, 6, 100308. [Google Scholar] [CrossRef]
  52. Kraft, M. A. (2020). Interpreting effect sizes of education interventions. Educational Researcher, 49(4), 241–253. [Google Scholar] [CrossRef]
  53. Kraft, M. A., Blazar, D., & Hogan, D. (2018). The effect of teacher coaching on instruction and achievement: A meta-analysis of the causal evidence. Review of Educational Research, 88(4), 547–588. [Google Scholar] [CrossRef]
  54. Lay, C. D., Allman, B., Cutri, R. M., & Kimmons, R. (2020). Examining a decade of research in online teacher professional development. Frontiers in Education, 5, 573129. [Google Scholar] [CrossRef]
  55. Lewis, C., & Perry, R. (2014). Lesson study with mathematical resources: A sustainable model for locally-led teacher professional learning. Mathematics Teacher Education and Development, 16(1), n1. [Google Scholar]
  56. Lewis, C., & Perry, R. (2017). Lesson study to scale up research-based knowledge: A randomized, controlled trial of fractions learning. Journal for Research in Mathematics Education, 48(3), 261–299. [Google Scholar] [CrossRef]
  57. *Lindvall, J. (2017). Two large-scale professional development programs for mathematics teachers and their impact on student achievement. International Journal of Science and Mathematics Education, 15(7), 1281–1301. [Google Scholar] [CrossRef]
  58. Lynch, K., Gonzalez, K., Hill, H., & Merritt, R. (2025). A meta-analysis of the experimental evidence linking mathematics and science professional development interventions to teacher knowledge, classroom instruction, and student achievement. AERA Open, 11, 23328584251335302. [Google Scholar] [CrossRef]
  59. Lynch, K., Hill, H. C., Gonzalez, K. E., & Pollard, C. (2019). Strengthening the research base that informs STEM instructional improvement efforts: A meta-analysis. Educational Evaluation and Policy Analysis, 41(3), 260–293. [Google Scholar] [CrossRef]
  60. Markussen-Brown, J., Juhl, C. B., Piasta, S. B., Bleses, D., Højen, A., & Justice, L. M. (2017). The effects of language-and literacy-focused professional development on early educators and children: A best-evidence meta-analysis. Early Childhood Research Quarterly, 38, 97–115. [Google Scholar] [CrossRef]
  61. *McGatha, M. B., Bush, W. S., & Rakes, C. R. (2009). The effects of professional development in formative assessment on mathematics teaching performance and student achievement. Journal of Multidisciplinary Evaluation, 6(12), 32–43. [Google Scholar] [CrossRef]
  62. McIntosh, A. (2008). Understand and use numbers—A handbook. Nationellt centrum för matematikutbildning. [Google Scholar]
  63. META Associates. (2006, March). Northeast Front Range Math/science Partnership (MSP) to increase teacher competence in content. Final evaluation report: January 1, 2004–December 31, 2006. (Unpublished manuscript).
  64. Moher, D., Liberati, A., Tetzlaff, J., Altman, D. G., & The PRISMA Group. (2009). Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement. PLoS Medicine, 6(7), e1000097. [Google Scholar] [CrossRef]
  65. Moher, D., Shamseer, L., Clarke, M., Ghersi, D., Liberati, A., Petticrew, M., Shekelle, P., Stewart, L. A., & PRISMA-P Group. (2015). Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015 statement. Systematic Reviews, 4(1), 1. [Google Scholar] [CrossRef]
  66. National Center for Education Statistics. (2019). National Assessment of Educational Progress (NAEP): 2019 mathematics assessment. U.S. Department of Education, Institute of Education Sciences. Available online: https://www.nationsreportcard.gov/highlights/mathematics/2019/ (accessed on 28 August 2025).
  67. National Center for Education Statistics. (2022). National Assessment of Educational Progress (NAEP): 2022 mathematics assessment. U.S. Department of Education, Institute of Education Sciences. Available online: https://www.nationsreportcard.gov/highlights/mathematics/2022/ (accessed on 28 August 2025).
  68. National Center for Education Statistics. (2023). Students with disabilities. In Condition of education. U.S. Department of Education, Institute of Education Sciences. Available online: https://nces.ed.gov/programs/coe/indicator/cgg (accessed on 28 August 2025).
  69. National Council of Teachers of Mathematics. (2000). Principles and standards for school mathematics. Available online: https://www.nctm.org/Standards-and-Positions/Principles-and-Standards/ (accessed on 28 August 2025).
  70. National Governors Association Center for Best Practices & Council of Chief State School Officers. (2010). Common core state standards for mathematics. Available online: https://www.thecorestandards.org/Math/ (accessed on 28 August 2025).
  71. National Research Council. (2011). Successful K-12 STEM education: Identifying effective approaches in science, technology, engineering, and mathematics. National Academies Press. [Google Scholar] [CrossRef]
  72. Page, M. J., Moher, D., Bossuyt, P. M., Boutron, I., Hoffmann, T. C., Mulrow, C. D., Shamseer, L., Tetzlaff, J. M., Akl, E. A., Brennan, S. E., Chou, R., Glanville, J., Grimshaw, J. M., Hróbjartsson, A., Lalu, M. M., Li, T., Loder, E. W., Mayo-Wilson, E., McDonald, S., … McKenzie, J. E. (2021). PRISMA 2020 explanation and elaboration: Updated guidance and exemplars for reporting systematic reviews. BMJ, 372, 160. [Google Scholar] [CrossRef]
  73. Park, S., Stecker, P. M., & Powell, S. R. (2024). A teacher’s toolkit for assessment when implementing data-based individualization in mathematics. Intervention in School and Clinic, 59(4), 243–253. [Google Scholar] [CrossRef]
  74. Perry, R. R., & Lewis, C. C. (2011). Improving the mathematical content base of lesson study: Summary of results. Available online: https://tinyurl.com/2p92s7pf (accessed on 28 August 2025).
  75. *Piasta, S. B., Logan, J. A., Pelatti, C. Y., Capps, J. L., & Petrill, S. A. (2015). Professional development for early childhood educators: Efforts to improve math and science learning opportunities in early childhood classrooms. Journal of Educational Psychology, 107(2), 407–422. [Google Scholar] [CrossRef]
  76. Piper, B., Zuilkowski, S. S., Dubeck, M., Jepkemei, E., & King, S. J. (2018). Identifying the essential ingredients to literacy and numeracy improvement: Teacher professional development and coaching, student textbooks, and structured teachers’ guides. World Development, 106, 324–336. [Google Scholar] [CrossRef]
  77. Polanin, J. R., & Pigott, T. D. (2015). The use of meta-analytic statistical significance testing. Research Synthesis Methods, 6(1), 63–73. [Google Scholar] [CrossRef] [PubMed]
  78. *Polly, D., Wang, C., Lambert, R., Martin, C., McGee, J. R., Pugalee, D., & Lehew, A. (2017). Supporting kindergarten teachers’ mathematics instruction and student achievement through a curriculum-based professional development program. Early Childhood Education Journal, 45(1), 121–131. [Google Scholar] [CrossRef]
  79. *Prast, E. J., Van de Weijer-Bergsma, E., Kroesbergen, E. H., & Van Luit, J. E. (2018). Differentiated instruction in primary mathematics: Effects of teacher professional development on student achievement. Learning and Instruction, 54, 22–34. [Google Scholar] [CrossRef]
  80. Pustejovsky, J. (2017). clubSandwich: Cluster-robust (sandwich) variance estimators with small-sample corrections. R package version 0.2.3.
  81. Pustejovsky, J. E., & Rodgers, M. A. (2019). Testing for funnel plot asymmetry of standardized mean differences. Research Synthesis Methods, 10(1), 57–71. [Google Scholar] [CrossRef]
  82. Pustejovsky, J. E., & Tipton, E. (2022). Meta-analysis with robust variance estimation: Expanding the range of working models. Prevention Science, 23(3), 425–438. [Google Scholar] [CrossRef]
  83. Randel, B., Apthorp, H., Beesley, A. D., Clark, T. F., & Wang, X. (2016). Impacts of professional development in classroom assessment on teacher and student outcomes. The Journal of Educational Research, 109(5), 491–502. [Google Scholar] [CrossRef]
  84. R Core Team. (2024). R: A language and environment for statistical computing. R Foundation for Statistical Computing. Available online: https://www.R-project.org/ (accessed on 28 August 2025).
  85. *Roschelle, J., Shechtman, N., Tatar, D., Hegedus, S., Hopkins, B., Empson, S., Knudsen, J., & Gallagher, L. P. (2010). Integration of technology, curriculum, and professional development for advancing middle school mathematics: Three large-scale studies. American Educational Research Journal, 47(4), 833–878. [Google Scholar] [CrossRef]
  86. *Sample McMeeking, L. B. S., Orsi, R., & Cobb, R. B. (2012). Effects of a teacher professional development program on the mathematics achievement of middle school students. Journal for Research in Mathematics Education, 43(2), 159–181. [Google Scholar] [CrossRef]
  87. *Santagata, R., Kersting, N., Givvin, K. B., & Stigler, J. W. (2010). Problem implementation as a lever for change: An experimental study of the effects of a professional development program on students’ mathematics learning. Journal of Research on Educational Effectiveness, 4(1), 1–24. [Google Scholar] [CrossRef]
  88. *Saxe, G. B., Gearhart, M., & Nasir, N. S. (2001). Enhancing students’ understanding of mathematics: A study of three contrasting approaches to professional support. Journal of Mathematics Teacher Education, 4(1), 55–79. [Google Scholar] [CrossRef]
  89. Schoenfeld, A. H. (2018). Video analyses for research and professional development: The teaching for robust understanding (TRU) framework. ZDM Mathematics Education, 50(3), 491–506. [Google Scholar] [CrossRef]
  90. Sims, S., & Fletcher-Wood, H. (2021). Identifying the characteristics of effective teacher professional development: A critical review. School Effectiveness and School Improvement, 32(1), 47–63. [Google Scholar] [CrossRef]
  91. Stevenson, M., Stevenson, C., & Cooner, D. (2015). Improving teacher quality for Colorado science teachers in high need schools. Journal of Education and Practice, 6(3), 42–50. [Google Scholar]
  92. Taylor, J. A., Kowalski, S. M., Polanin, J. R., Askinas, K., Stuhlsatz, M. A., Wilson, C. D., & Wilson, S. J. (2018). Investigating science education effect sizes: Implications for power analyses and programmatic decisions. AERA Open, 4(3). [Google Scholar] [CrossRef]
  93. Tipton, E. (2015). Small sample adjustments for robust variance estimation with meta-regression. Psychological Methods, 20(3), 375–393. [Google Scholar] [CrossRef]
  94. Tipton, E., & Pustejovsky, J. E. (2015). Small-sample adjustments for tests of moderators and model fit using robust variance estimation in meta-regression. Journal of Educational and Behavioral Statistics, 40(6), 604–634. [Google Scholar] [CrossRef]
  95. Tukey, J. W. (1977). Exploratory data analysis. Addison-Wesley. [Google Scholar]
  96. Valiandes, S., & Neophytou, L. (2018). Teachers’ professional development for differentiated instruction in mixed-ability classrooms: Investigating the impact of a development program on teachers’ professional learning and on students’ achievement. Teacher Development, 22(1), 123–138. [Google Scholar] [CrossRef]
  97. van Es, E. A., & Sherin, M. G. (2002). Learning to notice: Scaffolding new teachers’ interpretations of classroom interactions. Journal of Technology and Teacher Education, 10(4), 571–596. [Google Scholar]
  98. van Es, E. A., & Sherin, M. G. (2021). Expanding on prior conceptualizations of teacher noticing. ZDM Mathematics Education, 53(1), 17–27. [Google Scholar] [CrossRef]
  99. Viechtbauer, W. (2010). Conducting meta-analyses in R with the metafor package. Journal of Statistical Software, 36(3), 1–48. [Google Scholar] [CrossRef]
  100. *Walker, A., Recker, M., Ye, L., Robertshaw, M. B., Sellers, L., & Leary, H. (2012). Comparing technology-related teacher professional development designs: A multilevel study of teacher and student impacts. Educational Technology Research and Development, 60(3), 421–444. [Google Scholar] [CrossRef]
  101. *Wang, C., Polly, D., Lehew, A., Pugalee, D., Lambert, R., & Martin, C. S. (2013). Supporting teachers’ enactment of elementary school student-centered mathematics pedagogies: The evaluation of a curriculum-focused professional development program. New Waves, 16(1), 76–91. [Google Scholar]
  102. What Works Clearinghouse. (2020). What works clearinghouse procedures handbook, Version 4.1. Department of Education, Institute of Education Sciences, National Center for Education Evaluation and Regional Assis-tance. Available online: https://ies.ed.gov/ncee/wwc/handbooks (accessed on 28 August 2025).
  103. Yoon, K. S., Duncan, T., Lee, S. W.-Y., Scarloss, B., & Shapley, K. L. (2007). Reviewing the evidence on how teacher professional development affects student achievement. National Center for Education Evaluation and Regional Assistance. Available online: https://eric.ed.gov/?id=ED498548 (accessed on 28 August 2025).
  104. Zhang, L., Basham, J. D., & Yang, S. (2020). Understanding the implementation of personalized learning: A research synthesis. Educational Research Review, 31, 1–15. [Google Scholar] [CrossRef]
Figure 1. PRISMA 2020 Flow Diagram. Note. From Page et al. (2021). For more information, visit http://www.prisma-statement.org.
Figure 1. PRISMA 2020 Flow Diagram. Note. From Page et al. (2021). For more information, visit http://www.prisma-statement.org.
Education 15 01177 g001
Figure 2. Funnel Plot of Meta-Analysis of PD on Student Outcome.
Figure 2. Funnel Plot of Meta-Analysis of PD on Student Outcome.
Education 15 01177 g002
Table 1. PD Characteristics of the Included Studies.
Table 1. PD Characteristics of the Included Studies.
Study
Authors (Year)
Title of PDTopics of PDFormatGrade LevelNo. of ParticipantsNo. of Included Students with Disabilities Number of Days (h)Measure of Student Mathematics Outcomes
Antoniou and Kyriakides (2013)The Dynamic Integrated Approach Critical reflection and focus on teaching skills of the dynamic model which correspond to teacher developmental stage and needsIn-personElementary130 teachers
2356 students
12 daysStudent achievement in mathematics (criterion-reference tests)
Brendefur et al. (2013)Professional Development and ActivitiesContent knowledge, active learning, coherenceIn-personPreK16 teachers
111 students
1 dayPrekindergarten – Primary Screener for Mathematics
Brendefur et al. (2016)Developing Mathematical ThinkingMathematics, student thinking, and pedagogyIn-person
Summer workshop
Ongoing follow-up PD
K-5993 teachers
(T = 424, C = 569)
3045 students
(T = 1457, C = 1588)
24218 daysIdaho State Achievement Test
Brendefur et al. (2022)Developing Mathematical ThinkingMathematics, student thinking, and pedagogyIn-personElementary184 teachers
(T = 98, C = 86)
4618 students
(T = 2470, C = 2148)
22 daysMeasures of Academic Progress
Bruns et al. (2017)Continuous Professional Development Course: EmMaCompetence-orientation, participant-orientation, case-relatedness, various instruction formats, stimulation cooperation and fostering (self-)reflectionIn-personEarly childhood99 teachers
(T = 51, C = 48)
100 hMathematical content knowledge test
Campbell and Malkus (2011)CoachingMathematical content, pedagogy, and curriculumIn-personElementary1593 teachers
24759 students
15 daysStatewide standardized achievement test
Dash et al. (2012)Online professional development programUsing models to understand fractions, algebraic thinking, and the complexities of measurementOnlineElementary79 teachers
1438 students
9 days
(70 h)
Researcher-developed assessment that measures fractions, algebraic thinking, and measurement
Fisher et al. (2010)A Computerized Professional Development ProgramA lesson plan, a blank Concept Diagram, students with whom to practice, and a coach to prompt their applicationHybridElementary59 teachers
(T = 30, C = 29)
2 daysStudent concept acquisition test
Hilton et al. (2016)An ongoing professional development program Proportional reasoning In-personMiddle 130 teachers 4 daysDiagnostic instrument
Jacob et al. (2017)Math Solutions Professional DevelopmentMathematics content knowledge, insight into individual learners through formative assessment, understanding of how children learn math, effective instructional strategiesIn-personElementary105 teachers
(T = 51, C = 54)
1523 students
(T = 780, C = 743)
13 daysState standardized assessment
Researcher-developed assessment
Jacobs et al. (2007)A professional development projectAlgebraic reasoningIn-personGrade 1–5180 teachers
(T = 89, C = 14)
3735 students
(T = 1827, C = 373)
8 days (16.5 h) Written Mathematics Tests
Lindvall (2017)Swedish PD programsFive mathematical competenciesIn-personElementary90 teachers
5000 students
9 daysMathematical tests (McIntosh, 2008)
McGatha et al. (2009)A year-long professional development programRational numberIn-personMiddle40 teachers
(T = 20, C = 20)
5 days in addition to 30 hNational Assessment of Educational Progress
Piasta et al. (2015)The professional development adapted from Core Knowledge Preschool SequenceIdentifying similarities and differences, classifying and sorting using one characteristic, classifying and sorting using more than one characteristic, identifying a pattern using only one alternating characteristic, and identifying and creating complex patterns involving at least two characteristicsIn-personEarly childhood65 teachers
(T = 31, C = 34)
385 students
(T = 191, C = 194)
10.5 days (64 h)Applied Problems subtest (Woodcock–Johnson Tests of Achievement III)

Tools for Early Assessment in Math
Polly et al. (2017)Curriculum-Based Professional Development ProgramExploring mathematical tasks, examining lessons in their curriculum, and modifying curriculum-based lessonsIn-personKindergarten15 teachers
245 students
80 hStudent achievement measure
Prast et al. (2018)A Teacher Professional Development ProgramDifferentiated instruction in primary mathematicsIn-personElementary76 teachers
5658 students
30 hCito Mathematics Tests
Roschelle et al. (2010)The SimCalc ApproachRate and proportionality, linear functionIn-personMiddle218 teachers
539 students
6 daysResearcher-developed assessment that measures rate, proportionality, and linear function
Sample McMeeking et al. (2012)A Teacher Professional Development ProgramA sequence of content-oriented and pedagogy-oriented structured coursesIn-personMiddle2319 students
(T = 1002, C = 1317)
2334 daysColorado Student Assessment Program
Santagata et al. (2010)A Teacher Professional Development ProgramFractions, ratio and proportion, and expressions and equationsHybridMiddle59 teachers
3900 students
1 dayDistrict-wide Quarterly Assessments and the California Standards Test
Saxe et al. (2001)The Educational Leaders in Mathematics ProjectSkills with fractions procedures and understandings of fractions conceptsIn-personElementary 23 teachers
(T = 17, C = 6)
5 daysResearcher developed test that contained both computation and more conceptually oriented items
Wang et al. (2013)Mathematics Science Partnership professional development projectTeachers’ knowledge of mathematics content and pedagogyIn-personElementary185 teachers
5070 students
9 daysEnd-of-unit assessments
Note. Dash indicates data were not reported. T = treatment group, C = control group.
Table 2. Overall Professional Development Effects on Student Mathematics Outcome.
Table 2. Overall Professional Development Effects on Student Mathematics Outcome.
Effectmk I 2 τ b e t w e e n 2 τ w i t h i n 2 gSE95% CI95% PIdfp
Unadjusted PD208797.320.130.140.390.11[0.16, 0.61][−0.63, 1.40]860.001
Adjusted PD208795.820.090.080.340.10[0.15, 0.53][−0.47, 1.15]86<0.001
Note. m indicates the number of studies and k indicates the number of effect sizes.
Table 3. Moderator Analyses of PD on Student Mathematics Outcome.
Table 3. Moderator Analyses of PD on Student Mathematics Outcome.
VariableEst.SE95% CIdfp
Secondary grade level (vs. Primary)−0.790.57[−1.93, 0.34]800.17
Combination format (vs. In-person)0.100.33[−0.55, 0.75]800.76
PD focus 0.17
Specific (vs. General)1.300.63[0.05, 2.56]800.04
Combination (vs. General)−0.290.39[−1.07, 0.48]800.45
PD days−0.010.01[−0.03, 0.02]800.54
Inclusion of students with or at risk of disabilities (vs. without these students)0.550.40[−0.25, 1.36]800.17
Note. The number of studies (m) = 20; the number of effect sizes (k) = 86; I 2 = 96.07; τ 2 = 0.14 for within-study and 0.14 for between-study; p-values were adjusted for multiple comparison using the Benjamini–Hochberg Correction.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Park, S.; Lee, Y.R.; Nelson, G.; Cook, M.A.; Doabler, C.T. Teacher Professional Development and Student Mathematics Achievement: A Meta-Analysis of the Effects and Moderators. Educ. Sci. 2025, 15, 1177. https://doi.org/10.3390/educsci15091177

AMA Style

Park S, Lee YR, Nelson G, Cook MA, Doabler CT. Teacher Professional Development and Student Mathematics Achievement: A Meta-Analysis of the Effects and Moderators. Education Sciences. 2025; 15(9):1177. https://doi.org/10.3390/educsci15091177

Chicago/Turabian Style

Park, Soyoung, Young Ri Lee, Gena Nelson, Madison A. Cook, and Christian T. Doabler. 2025. "Teacher Professional Development and Student Mathematics Achievement: A Meta-Analysis of the Effects and Moderators" Education Sciences 15, no. 9: 1177. https://doi.org/10.3390/educsci15091177

APA Style

Park, S., Lee, Y. R., Nelson, G., Cook, M. A., & Doabler, C. T. (2025). Teacher Professional Development and Student Mathematics Achievement: A Meta-Analysis of the Effects and Moderators. Education Sciences, 15(9), 1177. https://doi.org/10.3390/educsci15091177

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop