Next Article in Journal
The Spirituality–Resilience–Happiness Triad: A High-Powered Model for Understanding University Student Well-Being
Previous Article in Journal
Deficits in Duration Estimation in Individuals Aged 10–20 Years Old with Idiopathic Mild Intellectual Disability: The Role of Inhibition, Shifting, and Processing Speed
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

From Senses to Memory During Childhood: A Systematic Review and Bayesian Meta-Analysis Exploring Multisensory Processing and Working Memory Development

by
Areej A. Alhamdan
1,*,
Hayley E. Pickering
2,
Melanie J. Murphy
2 and
Sheila G. Crewther
2,3
1
Department of Psychology, Imam Muhammad Ibn Saud Islamic University (IMSIU), Riyadh 11564, Saudi Arabia
2
Department of Psychology, Counselling and Therapy, La Trobe University, Melbourne, VIC 3086, Australia
3
Centre for Mental Health and Brain Sciences, Swinburne University of Technology, Melbourne, VIC 3122, Australia
*
Author to whom correspondence should be addressed.
Eur. J. Investig. Health Psychol. Educ. 2025, 15(8), 157; https://doi.org/10.3390/ejihpe15080157
Submission received: 22 April 2025 / Revised: 10 July 2025 / Accepted: 24 July 2025 / Published: 12 August 2025

Abstract

Multisensory processing has long been recognized to enhance perception, cognition, and actions in adults. However, there is currently limited understanding of how multisensory stimuli, in comparison to unisensory stimuli, contribute to the development of both motor and verbally assessed working memory (WM) in children. Thus, the current study aimed to systematically review and meta-analyze the associations between the multisensory processing of auditory and visual stimuli, and performance on simple and more complex WM tasks, in children from birth to 15 years old. We also aimed to determine whether there are differences in WM capacity for audiovisual compared to unisensory auditory or visual stimuli alone after receptive and spoken language develop. Following PRISMA guidelines, a systematic search of PsycINFO, MEDLINE, Embase, PubMed, CINAHL and Web of Science databases identified that 21 out of 3968 articles met the inclusion criteria for Bayesian meta-analysis and the AXIS risk of bias criteria. The results showed at least extreme/decisive evidence for associations between verbal and motor reaction times on multisensory tasks and a variety of visual and auditory WM tasks, with verbal multisensory stimuli contributing more to verbally assessed WM capacity than unisensory auditory or visual stimuli alone. Furthermore, a meta-regression confirmed that age significantly moderates the observed association between multisensory processing and both visual and auditory WM tasks, indicating that verbal- and motor-assessed multisensory processing contribute differentially to WM performance, and to different age-determined extents. These findings have important implications for school-based learning methods and other educational activities where the implementation of multisensory stimuli is likely to enhance outcomes.

1. Introduction

Successful adaptation to constantly changing environments requires individuals to be able to attend to and process information from multiple senses simultaneously (Alais et al., 2010; Todd, 1912; Treisman & Gelade, 1980). From the day of birth, all five sensory systems (vision, sound, touch, taste, and smell) contribute to the emergence of conscious perception. with touch, taste, and smell dominating during the first few weeks of life (Maitre et al., 2020; Muir & Field, 1979), and vision dominating from around 10–16 weeks as the cortical binocular vision system becomes functionally active (Braddick & Atkinson, 2011; McCulloch & Skarf, 1991; Richards, 1985). However, it is the close temporal occurrence of multisensory stimuli, particularly vision and sound, that together facilitate the creation of a child’s perceptual world and conceptualization of their unique environment (Bahrick & Lickliter, 2000; Dionne-Dostie et al., 2015; Kopp, 2014; Neil et al., 2006). In particular, it is the regular multisensory combination of visual information in the environment and the hearing of consistently spoken words (e.g., a child’s own name) that leads to the rapid and early (within a few weeks) acquisition of aspects of basic receptive language by young animals and human infants (Mason et al., 2019). However, to date, there is little understanding of the neural pathways and factors affecting the developmental trajectory and achievement of adultlike multisensory motor integration and speech in children (Chartier et al., 2018; Kocsis et al., 2023; R. Wang et al., 2023). Thus, this review aims to explore the current literature relating to this issue.

1.1. Multisensory Development

Multisensory processing is typically defined in terms of the ability to establish neural connections between the neural networks associated with processing of temporally close sensory stimuli, which eventually enables the facilitation of more efficient motor and cognitive behaviour in young animals and children (Bremner et al., 2012; Stein et al., 2010), and adults (Todd, 1912; Treisman & Gelade, 1980). Indeed, as alluded to above, multisensory facilitation of sensorimotor information processing emerges early in the first year of life (such as a child turning upon hearing their name), long before the child gains adequate motor control to sit, walk, or speak (Mason et al., 2019). Neil et al. (2006) also observed multisensory facilitation of head and eye movements in response to spatially localized concurrent audiovisual stimuli in 2-month-old infants, and both Bahrick and Lickliter (2000); Kopp (2014) have reported enhanced responses to multisensory stimuli compared to unisensory stimuli in infants as young as six months old. Interestingly, it takes children another 18-24 months of physical and brain development, particularly biological development of the sensory, cognitive, and motor systems, to achieve the requisite basic speech and expressive language to adequately assess overall cognitive capacity, highlighting the complementary roles integration of motor control and sensory co-ordination play in multisensory processing and cognitive development (Chartier et al., 2018; Dionne-Dostie et al., 2015; D. J. Lewkowicz & Bremner, 2020; Soto-Faraco et al., 2012) especially verbally assessable working memory.
Multisensory development is also known to continue to facilitate faster and more precise visuomotor responses, even after formal schooling begins (Alhamdan et al., 2022; Barutchu et al., 2009; Brandwein et al., 2011; Nardini et al., 2016, 2008). Such enhanced multisensory motor reaction times have also been shown to be closely related to development of higher-order cognitive functions, such as memory and fluid intelligence (Barutchu et al., 2011; Denervaud et al., 2020b), and presumably also to the progressive development of motor control of speech pathways and verbal expressive language in young children (Alhamdan et al., 2023a). However, cognitive and neuropsychology research has seldom considered the early development and sophistication of the motor system required before either motor reaction times or verbal responses to multisensory stimuli can be measured. Thus, there remains a notable gap in the literature regarding understanding of how multisensory processing contributes to the development of motor reaction times (MRTs), speech multisensory processing, and working memory (WM) in children. It is also important to note that speech is a highly complex series of motor functions requiring multiple brain networks and areas (Guenther & Hickok, 2016; Kent, 2000; Maegherman et al., 2019; Simonyan & Horwitz, 2011), and potentially a limiting factor influencing the interpretation of the age at which various multisensory and cognitive tasks can be assessed. In fact, cognitive psychology has usually conceived of verbal speech as simple and largely automatic (Baddeley & Warrington, 1973), and hence not likely to require or impede the normally considered limited neural resources or receptive and expressive vocabulary available to individuals who differ with age. Therefore, this review aimed to systematically investigate the existing literature to better understand how concurrent audiovisual stimuli development contribute to both motor reaction times and verbal speech effect on the measurement of multisensory integration and verbally measured WM abilities across childhood.

1.2. Working Memory and Multisesnory Processing

Memory is an essential function of human cognition and daily life (Squire, 1992). As one of multiple memory networks, WM is traditionally defined as the ‘temporary maintenance and manipulation of a limited number’ of recently acquired (short-term memory; STM) or older long-term memory (LTM) information (Baddeley, 2003a, 2003b; Hitch & Baddeley, 1976) for a greater understanding of stimuli and their contribution to more complex semantic thinking (Buschman & Miller, 2022; Rossi et al., 2023; Weisberg & Reeves, 2013).
To date, cognitive research has almost exclusively focused on the role of auditory and verbal information in assessing development of WM (Baddeley, 2003a; Baddeley et al., 1998; Gathercole et al., 1992; Verhagen et al., 2019). Substantially less consideration has been given to development of visually based WM (H. E. Pickering et al., 2022), despite the known dominance of vision in driving attention and cognition, both subcortically and cortically in animals (Wiesel & Hubel, 1963), as well as the social interactions and language development of preverbal children (Moeskops et al., 2015; Padilla et al., 2015; Tzarouchi et al., 2009) and adult humans (Crewther et al., 1999; Goodale & Milner, 1992; Goldberg & Wurtz, 1972; Zackon et al., 1997). Indeed, the wealth of objective research available from primate electrophysiology post-1960s (Foxe & Simpson, 2005; Hubel & Wiesel, 1963; Goldberg & Wurtz, 1972), adult human psychophysics (Braddick & Atkinson, 2011; Crewther et al., 1999; Goodale & Milner, 1992), and brain imaging post 1990 (Corbetta et al., 1995) consistently demonstrate the dominance of visually driven attention and processing of more complex WM information, whether visually or auditorily derived via the parieto-frontal network (Corbetta & Shulman, 2002; D’Esposito, 2007).
Neuroimaging studies in adults have also consistently demonstrated the concurrent activation of the dorsolateral pre-frontal cortex via the parieto-frontal network and the neocerebellum during multisensory cognitive and sensory–motor tasks (Beauchamp et al., 2004; D’Esposito, 2007; Postle, 2006; D. Wang et al., 2025). Furthermore, in one of the few functional magnetic resonance imaging (fMRI) studies in children aged 8–13 years, Denervaud et al. (2020a) demonstrated that similar areas of blood-oxygen -level-dependent (BOLD) activation and functional connectivity associated with the visuo-parietal-frontal networks (Denervaud et al., 2020a). These findings align with the neural efficiency hypothesis, which proposes that multisensory input allows for more efficient cognitive processing by reducing redundancy and facilitating faster encoding and retrieval in adults (Haier et al., 1988; Stein & Stanford, 2008; Grady, 2012). Additionally, attentional control (Eysenck et al., 2007; Hopfinger & Slotnick, 2020) is enhanced when stimuli are multimodal (auditory and visual), as such stimuli are more effective in engaging and sustaining attention compared to unimodal stimuli (Macaluso & Driver, 2005; Shams & Seitz, 2008). Together, these findings suggest that multisensory integration not only affects relevant neural systems associated with WM but also provides cognitive advantages through binding, attention regulation, and WM load reduction during development (Denervaud et al., 2020b; Murray et al., 2016; Shimi et al., 2014).
As such, further investigation and clarification of the contributions from multisensory processing to auditory and visual attention and WM abilities should provide a more complete picture of the development of both domains in early childhood.
Substantial variability also exists in the literature regarding previously utilized paradigms of assessment of multisensory processing, and this may also reflect unconscious investigator prediction and perception of a child group’s maturity while impacting understanding of the developmental progression of the underlying skills. For example, Tremblay et al. (2007) have concluded that both verbally based multisensory processes as assessed with the McGurk Effect and nonverbal (flash–beep) multisensory motor responses follow different developmental time courses. This difference is likely associated with the ubiquitous observation that simple non-verbal motor multisensory processing reaches adult rates in early childhood (5–9-year-olds), whereas the complex neural motor pathways associated with both speech production (Chartier et al., 2018) and the cognitive development of language and extent of vocabulary continue throughout life (e.g., comparison of two older groups of 10–14 and 15–19 years; (Tremblay et al., 2007). Thus, the current review aims to incorporate the importance of considering the diversity into assessment paradigms when interpreting findings related to multisensory processing and WM development.

1.3. The Current Review

This review and meta-analysis aimed to systematically investigate current literature examining the association between multisensory processing and WM performance across childhood, with a focus on the type of multisensory tasks (i.e., manual motor vs. verbally assessed spoken multisensory tasks) and the ages of the participants. Previously published reviews of the existing multisensory literature in adults (Matusz et al., 2017; Quak et al., 2015) have consistently shown that multisensory presentation of information is generally associated with age improvement in accuracy in WM tasks, making meta-analysis necessary to determine these associations in schoolchildren (birth to 15-years). In addition, there is also considerable heterogeneity in the literature in terms of findings regarding the contribution of multisensory processing to WM capacity tasks in school age-children (Constantinidou & Evripidou, 2012; Gillam et al., 1998).
Therefore, the primary aims of the current review were as follows:
(i)
We aimed to examine the association between motor reaction times (MRTs) and verbal multisensory processing for both visual and auditory WM measures. On the basis of past findings, studies were grouped according to the type of multisensory task and parameter being assessed (i.e., MRTs, or verbal tasks thus we hypothesized that accurate responses on verbal multisensory tasks (Constantinidou et al., 2011), and faster MRTs (Alhamdan et al., 2023b) would be significantly associated with better visual and auditory WM performance of children old enough to speak.
(ii)
We sought to compare the contribution of unisensory (auditory-alone and/or visual-alone) to multisensory (audiovisual) stimuli to WM capacity across childhood. It was hypothesized, in line with past research in adults (Matusz et al., 2017; Tovar et al., 2020), that WM capacity (i.e., number of items able to be recalled) would be higher for audiovisual stimuli (i.e., cross-modal objects) than for unisensory stimuli (visual or/and auditory alone).

2. Method

2.1. Protocol and Registration

This review was conducted in accordance with the recommended protocols of the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guideline (Page et al., 2021) (see Supplementary Table S1 for PRISMA Checklist). Our protocol was pre-specified and registered in the International Prospective Register of Systematic Reviews (PROSPERO) under the existing registration number CRD42020148110 (Alhamdan et al., 2019), which is accessible https://www.crd.york.ac.uk/PROSPERO/view/CRD42020148110 (accessed on 29 July 2025). A summary of amendments to the protocol can be found in Supplementary Table S2.

2.2. Literature Search Strategy

The tailored search strategy was developed in collaboration with an academic librarian (HB) and executed in six electronic databases: PsycINFO (Ovid, from 1806), MEDLINE (Ovid, from 1946), EMBASE (Ovid, from 1947), PubMed, CINAHL (EBSCO), Web of Science (ISI) using MeSH terms as appropriate, as well as Google Scholar, for all available years. We also incorporated additional terms suggested by certain databases when they were deemed relevant (e.g., intersensory processing). The initial search in the database was conducted on the 29th and 30th of July 2019, with an updated the search using the same protocol conducted on the 11th of August 2023 to include newly published articles. The detailed search strategy employed for the MEDLINE (Ovid) database is presented in Table 1. The specific search strategy for all databases is presented in Supplementary Table S3, while Supplementary Figures S1–S6 provide the specific search strategies for each database.

2.3. Eligibility Criteria and Study Selection

To be eligible for inclusion, studies were required to meet the following criteria. Firstly, studies were required to be quantitative, peer-reviewed, and written in English, with no limitations on the publication year. Qualitative papers, case studies, training programmes, reviews, book chapters, and dissertations were excluded. Studies of children with a neurodevelopmental disorder (e.g., ADHD, ASD, Language Disorder) or hearing or visual impairment that did not include neurotypical children as controls were also excluded. Furthermore, studies were only included if they employed correlation analyses comparing a multisensory processing task and a measure of visual or auditory WM, or if they assessed WM capacity (i.e., number of items able to be recalled) using three different types of stimuli (audiovisual, auditory, and visual stimuli) within the same task to allow for effect size comparisons between conditions.
All studies retrieved from the databases matching the search strategy were imported into Covidence, an online software platform designed for high-quality systematic reviews and to assist researchers in efficiently tracking each document through the inclusion/exclusion criteria process (Veritas Health Innovation, 2019). For the initial search, two reviewers (AA and HP) independently screened all titles and abstracts from each identified record to decide whether to exclude the study or include it for further review. Subsequently, both reviewers (AA and HP) reviewed the method and results sections of each potentially relevant study to make a final decision regarding the inclusion and exclusion criteria. These steps were completed by one reviewer (AA) for the second search. In cases of disagreement between the two reviewers, consultation with a co-author (SC) was sought for resolution.
The decision to exclude a study was based on a custom hierarchy consisting of seven exclusion criteria, as outlined below:
  • Studies not published in English;
  • Study design that did not meet criteria (e.g., case studies, reviews, qualitative papers, training programs, book chapters or dissertations);
  • Sample/s that did not meet the inclusion criteria (e.g., participants over 15 years old and/or individuals with neurodevelopmental disorders, hearing impairments, or visual impairments);
  • No measure of multisensory processing (i.e., audiovisual stimuli);
  • No measure of WM;
  • Required data/statistics not reported (e.g., no correlation, or the standard deviation for each group could not be extracted from the provided data);
  • Study did not report the result on typically developing children separately from atypically developing children.
Where the full text of a study was unavailable, corresponding authors were contacted via email. Possible related articles were excluded if the full text could not be obtained, if there was insufficient information to determine if the study met the inclusion criteria, or if communication with the corresponding author proved ineffective in obtaining the required information.

2.4. Risk of Bias and Quality Assessment

The methodological quality of the included studies was evaluated using the AXIS tool (Appraisal Tool for Cross-Sectional Studies; Downes et al., 2016). The 20 items included in this appraisal are intended to achieve the following objectives: clarify the study aims and objectives, evaluate the appropriateness of the study design and methodology, assess the validity of the results, ascertain whether the discussion and conclusion are well-supported by the results, and review ethical considerations. To facilitate the comparison of study quality, we followed the procedures used by H. E. Pickering et al. (2022) by transforming outcomes into percentages, where a higher percentage indicated better study quality. These percentages were calculated based on the number of relevant items for each specific study, excluding items marked as ‘Not Applicable’. Studies achieving a rating of 75% or higher were categorized as high-quality, those with scores ranging from 50% to 74% were considered moderate quality, and studies with scores below 50% were categorized as poor quality. Only studies with a quality score of 50% or higher were included in the review. Table 2 presents the quality qssessment and summary of critical appraisal of the included studies using (AXIS) tool.

2.5. Data Extraction

The following information from each study was extracted: study details (author(s), year of publication), study purpose (aims as stated in the study), participant information (number of total participants, age [M, (SD), range]), measures of multisensory processing tasks (type of MS task, task description, and degree of complexity), measures of WM tasks (visual or auditory), and outcomes for both correlation coefficients and effect size analyses (see Table 3 and Table 4).
The first outcome extracted was the bivariate correlation r between a multisensory measure and WM task. We acknowledge that “r” is typically reported in experimental designs; however, in some studies where correlation is not the primary interest, it may not be included (Dunlap et al., 1996). Therefore, “r” may not always be available unless the raw data can be obtained. To address this issue Borenstein (2009) recommend estimating the correlation by referring to related studies and conducting sensitivity analyses using a range of reasonable correlation values. Another approach suggested by Hullett and Levine (2003) involves estimating “r” values using available t and f statistics. However, when these specific statistics were not accessible, effect size was directly estimated using means and standard deviations (Dunlap et al., 1996; Morris & DeShon, 2002). In the current review, effect sizes using Cohen’s d were computed for the comparison of WM capacity to audiovisual stimuli vs. auditory stimuli in the first analysis, and for WM capacity to audiovisual stimuli vs. visual stimuli in the second analysis.
To synthesize the correlation data, we categorized multisensory tasks into three groups:
  • Motor reaction times (i.e., measuring how quickly participants press a button);
  • Verbal multisensory tasks (i.e., assessing accuracy through the number of correct responses as reported verbally and without any time related data);
  • Motor non-timed tasks (i.e., tasks involving manual motor actions but focusing on the accuracy of responses rather than response time).
Two studies (Broadbent et al., 2018; Cleary et al., 2001) and two datapoints from Barutchu et al. (2020) did not fit this categorization of multisensory tasks, as they involved motor but non-timed tasks. Hence, they have been discussed in the narrative review, and their results are reported in Table 5.
For group difference (effect size) data synthesis, we categorized multisensory WM tasks based on two comparisons: (i) the contribution of audiovisual stimuli (AVS) vs. auditory stimuli (AS) to WM capacity, and (ii) the contribution of audiovisual stimuli (AVS) vs. visual stimuli (VS) to WM capacity. Complete data extractions, including means, N, and SD for each group to calculate effect sizes for group difference studies, are presented in Supplementary Table S4. In cases of unclear or missing information within the publications, we attempted to contact the authors for further clarification or additional data.

2.6. Data Analysis

Studies were grouped into two different analyses. First, correlation coefficients were used to investigate the association between multisensory tasks (including motor and verbal multisensory tasks) and WM. Fisher transformations, r-to-Z transformations, were employed to convert the correlation coefficients into effect size-based Fisher’s Z (Silver & Dunlap, 1987). Standard error was computed based on the sample size of each study. In some cases, studies presented multiple correlations of interest for a single sample. Effect size was categorized as weak (ranging from 0.1 to 0.3), moderate (ranging from 0.4 to 0.6), and strong (ranging from 0.7 to 0.9), following the classification proposed by (Akoglu, 2018).
Secondly, effect sizes were computed to examine the differences in WM capacity for audiovisual stimuli vs. auditory stimuli, as well as for audiovisual stimuli vs. visual stimuli (VS). Cohen’s d was computed using the standard deviation (SD) and sample size (n) for each group, allowing for the estimation of effect size across studies, with 95% credible intervals (CI). Cohen’s d was chosen as a robust measure of effect sizes for sample sizes larger than 20 (Borenstein et al., 2021). Multiple effect sizes for a single sample were extracted in some studies. Effect sizes of 0.30, 0.50, and 0.80 represented small, medium, and large effects, respectively (Cohen, 1988).
Bayesian meta-analyses were conducted using the free software JASP; Jeffreys’s Amazing Statistics Program 0.16.3.0 (JASP Team, 2022; http://www.jasp-stats.org/ (accessed on 30 Aguest 2022)). A random effect model is most appropriate for current data rather than fixed effect as it is well-suited for situations where sources beyond sampling error may contribute to the heterogeneity observed in effect sizes between studies (Langan et al., 2019; Thompson & Higgins, 2002). Bayesian analyses were selected to help in addressing the issue of variability underestimation, providing a more flexible and robust framework and straightforward interpretation (Hackenberger, 2020; Marsman & Wagenmakers, 2017). In addition, the Bayesian approach assists researchers to address publication bias and offers the advantage of sequential analysis, (Gronau et al., 2021), and this approach offers theoretical and practical advantages in the evaluation and interpretation of developmental data (Weaver & Hamada, 2016). For the effect size (μ), we used the default Cauchy prior distribution with scale of 1/ 2 ≈ 0.707, and the default Inverse-Gamma [1, 0.15] as the prior distribution on heterogeneity (τ) (Gronau et al., 2021). The prior (estimated) and posterior (observed) odds and 95% credible intervals (95% CI) are reported. Bayesian Factors (BF) are also reported, where higher BF10 values are interpreted as strength of evidence in favour of the alternative hypothesis (H1) against the null hypothesis (H0) testing. Heterogeneity (BFrf; variance between studies) was conducted to assess the presence or absence of between-study variability, with Bayes Factor estimates providing evidence for or against the null hypothesis of homogeneity.

3. Results

3.1. Study Selection

The database searches yielded 6070 records, of which 3968 were unique records (see Supplementary Material Figures S1–S6 for a detailed breakdown of records per database), and 61 studies were found via other sources (i.e., manual searching and Google Scholar). After individually assessing the titles and abstracts, 159 records met the criteria for full-text review. Of these, 138 articles were excluded during the full-text review, and one further duplicated article was detected. Thus, 21 studies were ultimately included in the systematic review and meta-analysis. The study selection process is illustrated in the PRISMA flow diagram in Figure 1.

3.2. Quality Assessment and Risk of Bias

The AXIS tool (Downes et al., 2016) was employed to assess the quality of the 21 selected studies. All studies met the required quality standards and were therefore included in the review and meta-analyses. Six studies achieved a moderate quality rating with an average percentage of 69% (60–73%), whereas fifteen studies demonstrated high quality with an average percentage of 83% (76–94%).
All studies included in the review had clear aims/objectives, employed appropriate study designs to meet their stated aims, and presented well-justified conclusions. However, only one of the 21 studies included statistical justification of sample size (Alhamdan et al., 2023b). Ethical approval and the use of consent forms were not described in eight studies (Badian, 1977; Cleary et al., 2001; Crawford & Fry, 1979; Field & Anderson, 1985; Gillam et al., 1998; Hatchette & Evans, 1983; Magimairaj et al., 2009; Montgomery et al., 2009). Seven studies also failed to include a discussion of limitations to their methodological approach (Badian, 1977; Barutchu et al., 2011; Cleary et al., 2001; Constantinidou et al., 2011; Crawford & Fry, 1979; Hatchette & Evans, 1983; Pillai & Yathiraj, 2017b).
In summary, the majority of the studies that were included in this review provided sufficient information and demonstrated a high level of quality. All 21 studies employed appropriate methodologies for data collection and featured samples that effectively represented the children they targeted. Consequently, all 21 studies were included in this systematic review and meta-analysis. A summary of the quality assessment is presented in Table 2.
Table 2. Summary of critical appraisal of the included studies using (AXIS) tool.
Table 2. Summary of critical appraisal of the included studies using (AXIS) tool.
Study Details IntroMethodResultsDiscussionOtherTotal
#(Author(s), Year)Study Design Contributed to This Meta-Analysis 1: Aims2: Study Design3: Sample Size Justification4: Target Population5: Sampling Frame6: Sample Selection7: Non-Responders8: Appropriate Measurement9: Reliable Measurement10: Statistical Significance11: Repeatability12: Basic Data13: Response Rate14: Non-Responders15: internal Consistency16: All Results Presented17: Conclusions Justified18: Limitations Discussed19: Funding a& COI Declared20: Ethics & Consent%
1(Alhamdan et al., 2023b)Correlation 94%
2(Badian, 1977)Correlation 69%
3(Barutchu et al., 2011)Correlation82%
4(Barutchu et al., 2019)Correlation87%
5(Barutchu et al., 2020)Correlation84%
6(Broadbent et al., 2018)Correlation79%
7(Choudhury et al., 2007)Group Differences85%
8(Cleary et al., 2001)Correlation70%
9(Constantinidou et al., 2011)Both78%
10(Constantinidou & Evripidou, 2012)Group Differences84%
11(Crawford & Fry, 1979)Correlation60%
12(Davies et al., 2019)Correlation80%
13(Denervaud et al., 2020b)Correlation84%
14(Field & Anderson, 1985)Group Differences76%
15(Gillam et al., 1998)Group Differences81%
16(Hatchette & Evans, 1983)Group Differences73%
17(Montgomery et al., 2009)Correlation76%
18(Magimairaj et al., 2009)Correlation84%
19(Ortiz-Mantilla et al., 2008)Group Differences89%
20(Pillai & Yathiraj, 2017a)Correlation79%
21(Pillai & Yathiraj, 2017b)Group Differences72%
Note. Green = yes; red = no; yellow = unclear; blue = not applicable; intro = introduction; COI = conflict of interest. Rating of “75% or higher” = high quality; “50% to 74%” = moderate quality; “50% and below” = poor quality (H. E. Pickering et al., 2022). AXIS appraisal tool for cross-sectional studies.

3.3. Study Characteristics and Data Synthesis

Overall, 19 of the 21 included studies contributed to the correlation and group differences meta-analyses. Specifically, 11 studies (58%) provided correlation data, which was categorized into two meta-analyses (based on MRTs and verbal multisensory tasks, see Table 3; for description of tasks). Additionally, 7 studies (36%) provided data for estimating effect sizes (to compare WM capacity for audiovisual stimuli versus unisensory stimuli). One study (6%) included both types of data (correlation and group differences) (Table 4). The remaining two studies (Broadbent et al., 2018; Cleary et al., 2001) and two datapoints from Barutchu et al. (2020) involved non-timed motor tasks and hence could not be included in the meta-analyses, providing six datapoints in a narrative synthesis presented in Table 5. A starburst visualization summarizing the various types of multisensory tasks, their corresponding assessment tools, and the contributing study sources is shown in Figure 2.
Table 3. Description of multisensory tasks in all included studies.
Table 3. Description of multisensory tasks in all included studies.
#(Author(s), Year) Type of MS TaskTask DescriptionAim of MS Task
1(Alhamdan et al., 2023b)MRTs
Non-verbal motor task
(Simple AV Task)
Multisensory processing was measured using motor reaction times to target detection. The targets included three types of stimuli: an auditory stimulus (AS; beep), a visual stimulus (VS; grey circle), and an audiovisual stimulus (AVS; beep and grey circle presented simultaneously) The children aged 5-10 years were instructed to press a button as rapidly as possibleMultisensory gain seen as measure of improvement in MRT (the time it took participants to respond in milliseconds)
2(Barutchu et al., 2011)MRTs
Non-verbal motor task
(Simple AV Task)
Participants aged 7–11 years were randomly presented with AS, VS, AVS, and blank stimuli and instructed to press a button when they saw a flash, when they heard a tone burst, or when both happened simultaneouslyMultisensory gain seen as measure of improvement in MRT (the time it took participants to respond in milliseconds)
3(Barutchu et al., 2020)MRTs
Non-verbal motor task
(Simple AV Task and Associate Learning Taks; ALT)
Simple detection task is similar to the task used by the same authors 2011, while ALTs required participants aged 8–11 years to learn an association between novel black symbols and either a novel auditory sound (novel-AV), or a verbal sound (verbal-AV) with a consonant-vowel-consonantMultisensory gain seen as measure of improvement in MRT (the time it took participants to respond in milliseconds)
4(Denervaud et al., 2020b) MRTs
Non-verbal motor task
(Simple AV Task)
Participants aged 4–15 years were randomly presented with AS, VS, AVS, and blank stimuli and instructed to press a button when they saw a flash, when they heard a tone burst, or when both happened simultaneouslyMultisensory gain seen as measure of improvement in MRT (the time it took participants to respond in milliseconds)
5 & 6 (Montgomery et al., 2009)
&
(Magimairaj et al., 2009)
MRTs
Non-verbal motor task
(Colour–word Recognition)
Participant aged 6–12 were asked to touch a coloured shape (e.g., a green square) as quickly as possible in response to hearing a familiar colour word (e.g., “green”). The task required them to make a simple, speeded decision about which visual object matched the spoken color name.Multisensory gain seen as measure of improvement in MRT (the time it took participants to respond in milliseconds)
7(Badian, 1977)Verbal MS TasksIn the verbal MS task, participants aged 8–11 years were presented with grouped sequences of the names of four geometric forms: square, circle, triangle, and cross. The auditory stimuli involved orally presenting these names, while the corresponding visual stimuli consisted of sequences of the geometric forms drawn in black outline on white cards. Participants were required to match or select the correct visual response based on the auditory inputMultisensory gain seen as measure of improvement in accuracy (the proportion of correct responses on trials)
8(Barutchu et al., 2019)Verbal (McGurk Effect)Participants aged 8–11 years were shown incongruous auditory syllables and visual syllables in four different videos (i.e., three matching videos and one incongruent audiovisual video with /ba-ba/ sounds and /ga-ga/ lip movements)Multisensory gain seen as the number of perceptions of the fused percept (e.g., “da” or “tha”)
9(Broadbent et al., 2018)Motor non-timed Task (Multisensory Attention Learning Task (MALT)The Multisensory Attention Learning Task (MALT) studies how sensory cues affect attention and learning. Participants aged 6–10 years view frog images, for example, paired with frog sounds in three conditions, visual, auditory, and combined cues for frog categorization. They respond to target frogs while disregarding others. Participants are instructed to press the space bar whenever a frog (the target animal) appears on the screenImprovement in accuracy
10(Cleary et al., 2001)Motor non-timed Task (Auditory-plus-visual–spatial memory game)Participants aged 8–9 years were instructed to press appropriate buttons on the response box when hearing sounds through the loudspeaker and seeing the buttons light up. For example, if they heard ‘blue’ followed by ‘green’ (indicating two buttons consecutively), they were expected to press the blue button first and then the green buttonImprovement in accuracy
11
&
12
(Constantinidou et al., 2011)
&
(Constantinidou & Evripidou, 2012)
Verbal Auditory, Visual, and Simultaneous Auditory visual TaskParticipants aged 7–13 years were presented words and pictures under three conditions: (a) auditory, (b) visual, and (c) auditory plus visual, and they asked to recall the items after each presentationMultisensory gain seen as measure of improvement in accuracy (the proportion of correct responses on trials)
13(Crawford & Fry, 1979)Verbal (Matching Auditory and Visual trigrams)Participants aged 5–9 years completed four matching tasks: two within the same senses (visual-to-visual, auditory-to-auditory) and two between senses (visual-to-auditory, auditory-to-visual). They matched cards with sounds and images, following instructions based on names given through cards or a tape recorderMultisensory gain seen as measure of improvement in accuracy (the proportion of correct responses on trials)
14(Davies et al., 2019)Audiovisual Change Detection TaskParticipants aged 5–6 years were presented with two sequential memory displays, each of which contained one shape and one sound. They were asked if the test shape-sound pairing was the same or different to either of the original pairings in the memory displaysMultisensory gain seen as measure of improvement in accuracy (the proportion of correct responses on trials)
15 & 16 (Choudhury et al., 2007)
&
(Ortiz-Mantilla et al., 2008)
Habituation Recognition Memory Taks
Non-verbal non-motor task
Infants (6–9 months) participated in a series of habituation-recognition memory tasks, including one task involving visual stimuli with familiarized and novel faces, and two tasks combining auditory and visual stimuli. These tasks featured abstract visual slides paired with tone-pairs (with durations of 70 ms and 300 ms, as described in Choudhury et al., 2007), as well as consonant-vowel (CV) syllables, as reported in Ortiz-Mantilla et al., 2008. During the habituation phase, infants underwent consecutive discrete trials based on their individual looking timesImprovement in looking time
17(Field & Anderson, 1985)Cued and Free Recall Measures of Television Program
Verbal multisensory task
Participants aged 5- and 9-years were presented with a 35 min colour videotape containing six short segments of story, including three types of modalities (visual, auditory and audiovisual content). Participants were then asked cued and free questions for each storyImprovement in accuracy and capacity of WM
18(Gillam et al., 1998)Audiovisual recall task-digits
Verbal multisensory task
Participants aged 8–12 years were presented with (a) auditory digits spoken in a male voice, (b) visual digits appearing on a computer screen, and (c) audiovisual digits appearing on a computer screen in synchrony with the auditory presentation. They were asked to recall the digit in the same order as they appeared in the presentationImprovement in accuracy and capacity of WM
19(Hatchette & Evans, 1983)Multisensory temporal/spatial Pattern Matching Taks
Verbal multisensory task
Participants aged 7–10 years were presented with four to nine black dot pictures with either one or two discontinuities in patterns, and auditory stimuli. In the multisensory condition, the visual and audio stimuli were presented simultaneously, and participants were asked to determine whether the items presented were the same or differentImprovement in accuracy and capacity of WM
20
&
21
(Pillai & Yathiraj, 2017a)
&
(Pillai & Yathiraj, 2017b)
Visual and Audiovisual Scores and Span TasksParticipants aged 7–9 years were instructed to repeat the words of the sequence presented auditory, visual and auditory–visual combined modalities and each memory skill was calculated using two different scoring procedures (score and span)Improvement in accuracy and capacity of WM
Note. MRTs = motor reaction times; WM = working memory.
Table 4. Characteristics of studies included in meta-analyses.
Table 4. Characteristics of studies included in meta-analyses.
Study DetailsParticipantsMeasuresMeta-Analysis Outcomes
#CitationAim as in StudyStudy DesignNAge M (SD); RangeMultisensory Task/sMemory Task/sMeta-Analysis Category
Fishers’ z/Cohens’ d [95% CI]
1(Alhamdan et al., 2023b)To investigate developmental changes in the associations between age, and visual and auditory short-term and WM, and multisensory processingCorrelation757.95 (1.70)
5–10 years
Simple Audiovisual Detection Task“VWM and AWM” using (Digit Span Tasks)1. AV−MRTs vs. VWM (Forwards)
−0.40 [−0.63, −0.17]
2. AV−MRTs vs. VWM (Backwards)
−0.55 [−0.78, −0.32]
3. AV−MRTs vs. AWM (Forwards)
−0.38 [−0.65, −0.15]
4. AV−MRTs vs. AWM (Backwards)
−0.45 [−0.68, −0.22]
2(Badian, 1977)To explore the nature of auditory–visual processing and memory performance Correlation309.71 (11.95)
8–11 years 0
Verbal Multisensory Tasks “AWM” Auditory Memory Task (AMT) and the Auditory Sequential Memory subtest (ITPA) 1. MS (verbal) vs. AMT
0.21 [−0.16, 0.59]
2. MS (verbal) vs. ITPA
0.29 [−0.09, 0.66]
3(Barutchu et al., 2011)To investigate the relationships between multisensory integration, auditory background noise, and the general intellectual abilities and AWMCorrelation859.7
7–11 years
Simple Audiovisual Detection Task“AWM”
WISC–IV WMI
AV−MRTs vs. AWM
−0.13 [−0.35, 0.09]
4(Barutchu et al., 2019)To investigate developmental relationships between attention, multisensory processes, and children’s general intellectual abilities including AWMCorrelation519.18 (1.61)
7–13 years
McGurk Effect/Illusion Task “AWM”
WISC–IV WMI
MS (McGurk) vs. AWM
−0.13 [−0.41, 0.15]
5(Barutchu et al., 2020)To investigate the developmental profile and relationships between uni- and multisensory processes and children’s intellectual abilities including AWMCorrelation419.85
8–11 years
Simple Audiovisual Detection Task Associative Learning Tasks (ALT)“AWM”
WISC–IV WMI
1. AV−MRTs vs. AWM
0.33 [0.00, 0.66]
2. AV−MRTs (ALT−Novel) vs. AWM
0.17 [−0.16, 0.50]
3. AV−MRTs (ALT−Verbal) vs. AWM
−0.06 [−0.39, 0.27]
6(Crawford & Fry, 1979)To explore the contributions of auditory and visual WM to audiovisual multisensory stimuliCorrelation526.75
5–9 years
Matching Auditory and Visual trigrams“VWM” Visual-Sequential Memory subtest
(ITPA), and “AWM” WISC (Digit Span)
1. Verbal (A−V) vs. Total−WM
0.55 [0.27, 0.83]
2. Verbal (V−A) vs. Total−WM
0.69 [0.41, 0.97]
3. (A−V) vs. VWM
0.52 [0.24, 0.80]
7(Davies et al., 2019)To test the developmental increase in audiovisual binding ability and its influence on retrieving bound audiovisual information and measures of verbal and visual complex WM spanCorrelation495.6 (3.25)
5–6 years
Audiovisual Change Detection Task“AWM” Listening Recall Task and “VWM” Odd One Out Task 1. Verbal (A−V binding) vs. AWM
0.65 [0.36, 0.94]
2. Verbal (A−V binding) vs. VWM
0.60 [0.32, 0.89]
8(Denervaud et al., 2020b)To investigate the relationship between multisensory
processing and higher-level cognition such as WM, and fluid intelligence in school children
Correlation778.1 (3.0) 4–15 yearsSimple Audiovisual Detection Task“AWM” WISC (Digit Span)AV−MRTs vs. AWM
−0.56 [−0.79, −0.33]
9(Montgomery et al., 2009)To examine the potential influence of three main components of WM, and processing speed (AV reaction time) on children’s ability to understand spoken narrativeCorrelation678.39 (1.69)
6–11 years
Auditory–visual MRTs Task (colour word recognition)“AWM” WISC (Digit Span), and Concurrent processing-storage (CPS)1. AV−MRTs vs. AWM (DS)
−0.46 [−0.70, −0.21]
2. AV−MRTs vs. AWM (CPS)
−0.51 [−0.76, −0.27]
10(Magimairaj et al., 2009)To investigate the relative contribution of storage, processing speed (AV reaction time), and attentional allocation on digits and complex span testsCorrelation658.6 (1.8)
6–12 years
Auditory–visual MRTs Task (colour word recognition)“AWM” WISC (Digit Span), and Complex Memory Span Task (CMST)1. AV−MRTs vs. AWM (CMST)
−0.44 [−0.68, −0.19]
2. AV−MRTs vs. AWM
−0.24 [−0.49, 0.00]
11(Pillai & Yathiraj, 2017a)To test whether there are associations between different memory skills and the auditory, visual, and combined modalitiesCorrelation288.1
7–8 years
Auditory, Visual and Audiovisual Scores and Span TasksRevised Auditory Memory and Sequencing Test (RAMST-IE)1. Verbal MS vs. AWM (RAMST−Span)
0.42 [0.03, 0.82]
2. Verbal MS vs. AWM (RAMST−Score)
0.89 [0.50, 1.28]
12(Constantinidou et al., 2011)To investigate the effects of stimulus presentation modality (i.e., auditory, visual and audiovisual) on WM performanceCorrelation & Group differences409.85 (1.03)
7–13 years
Auditory, Visual, and Simultaneous Auditory visual Task“AWM” CVLT-C (Listening Recall Task)Correlation
Verbal (A−V) vs. AWM
0.38 [0.06, 0.69
Group Differences
1. WMC−AVS vs. WMC−AS (M= 8 years)
0.32 [−0.28, 0.91]
2. WMC−AVS vs. WMC−AS (M = 10 years)
0.75 [0.07, 1.42]
3. WMC−AVS vs. WMC−VS (M = 8 years)
0.07 [−0.52, 0.67]
4. WMC−AVS vs. WMC−VS (M = 10 years)
−0.19 [−0.84, 0.47]
13(Constantinidou & Evripidou, 2012)To investigate the effects of stimulus presentation modality on WM performance
in children
Group differences2011.5 (5.07)
10–12
Auditory, Visual, and Simultaneous Auditory visual Task using AVLT (Auditory Verbal Learning Test Paradigm)1. WMC-AVS vs. WMC-AS
0.23 [−0.39, 0.86]
2. WMC-AVS vs. WMC-VS
−0.28 [−0.91, 0.34]
14(Choudhury et al., 2007)To determine if performance
on infant information processing measures differ as a function of family
history of specific language impairment or the particular demand characteristics of two different Rapid Auditory Processing paradigms used
Group differences298.97 mo (1.0)
6–9 mo
Auditory Visual Habituation Recognition Memory Task1. AVH/RM (70 ms) vs. VH/RM
0.25 [−0.27, 0.76]
2. AVH/RM (300 ms) vs. VH/RM
0.12 [−0.39, 0.64]
15(Ortiz-Mantilla et al., 2008)To determine whether performance differences on language and cognitive tasks were due to poorer global cognitive performance or to deficits in specific processing abilitiesGroup differences328.97 mo (0.58)
6 & 9 mo
Auditory Visual Habituation Recognition Memory Task1. AVH/RM (6 mo) vs. VH/RM
0.20 [−0.29, 0.70]
2. AVH/RM (9 mo) vs. VH/RM
−0.12 [−0.61, 0.37]
16(Field & Anderson, 1985)To investigate how children’s visual orientation and memory recall of a video clip are influenced by different factors such as the emphasis on different modalities (visual, auditory, or both)Group differences805.1 & 9.1
5 & 9 years
Cued and Free Recall Measures of Television Program1. WMC-AVS vs. WMC-AS (cued; 5 years)
0.67 [0.35, 0.98]
2. WMC-AVS vs. WMC-AS (cued; 9 years)
0.23 [−0.08, 0.54]
3. WMC-AVS vs. WMC-AS (free; 5 years)
0.25 [−0.06, 0.56]
4. WMC-AVS vs. WMC-AS (free; 9 years)
0.64 [0.33, 0.96]
5. WMC-AVS vs. WMC-VS (cued; 5 years)
0.46 [0.14, 0.77]
6. WMC-AVS vs. WMC-VS (cued; 9 years)
1.20 [0.86, 1.53]
7. WMC-AVS vs. WMC-AS (free; 5 years)
0.12 [−0.19, 0.43
8. WMC-AVS vs. WMC-AS (free; 9 years)
0.79 [0.47, 1.12]
17(Gillam et al., 1998)To test the immediate recall of digits presented visually, auditorily, or
audiovisually
Group differences169.8
(1.2)
8–12 years
Audio-Visual Digit Recall Task1. WMC-AVS vs. WMC-AS (WM)
−0.26 [−0.95, 0.44]
2. WMC-AVS vs. WMC-AS (STM)
−0.21 [−0.90, 0.49]
3. WMC-AVS vs. WMC-VS (WM)
−0.19 [−0.89, 0.50]
4. WMC-AVS vs. WMC-VS (STM)
0.18 [−0.52, 0.87]
18(Hatchette & Evans, 1983)To determine whether audiovisual processing (temporal and spatial integration) is the critical factor in accounting for the performance on pattern matching tasksGroup differences188.6
7.3–10 years
Multisensory temporal/spatial Pattern Matching Taks1. WMC-AVS vs. WMC-AS (WM)
0.46 [−0.20, 1.12]
2. WMC-AVS vs. WMC-VS (spatial)
0.49 [−0.18, 1.16]
3. WMC-AVS vs. WMC-VS (temporal)
−0.14 [−0.79, 0.51]
19(Pillai & Yathiraj, 2017b)To assess whether a correlation exists between two scoring of memory task (score and span) across three sensory conditions (auditory, visual, and auditory–visual)Group differences288.1
7–9
years
Visual and Audiovisual Scores and Span Tasks1. WMC-AVS vs. WMC-AS (Score)
0.34 [−0.18, 0.87]
2. WMC-AVS vs. WMC-AS (Span)
−0.44 [−0.97, 0.09]
3. WMC-AVS vs. WMC-VS (Score)
0.88 [0.33, 1.42
4. WMC-AVS vs. WMC-VS (Span)
0.46 [−0.07, 0.99]
Note. WM = working memory; VWM = visual working memory; AWM = auditory working memory; AV-MRTs = audiovisual motor reaction times; AMT = auditory memory task; ITPA = Illinois Test of Psycholinguistic Abilities; MS = multisensory; WISC–IV = Wechsler Memory Scale; McGurk = McGurk effect; ALT = Associative Learning Task; A-V = auditory–visual; V-A = visual–auditory; Total-WM = Total Score of VWM and AWM; DS = digit span; CMST = Complex Memory Span Task; RAMST = Revised Auditory Memory and Sequencing Test; CVLT-C = The California Verbal Learning Test for Children; WMC-AVS = Working Memory Capacity for Audiovisual Stimuli; WMC-AS = Working Memory Capacity for Auditory Stimuli; WMC-VS = Working Memory Capacity for Visual Stimuli; mo = Months; AVH/RM = Audiovisual Habituation Recognition Memory; VH/RM = Visual Habituation Recognition Memory; WM = working memory; STM = short-term memory.
Table 5. Characteristics and results of studies included in narrative synthesis.
Table 5. Characteristics and results of studies included in narrative synthesis.
Study DetailsParticipantsMeasuresResults
#CitationAim of StudyStudy DesignNAge M (SD); RangeMultisensory Task/sMemory Task/sCorrelation, pFishers’ z [95% CI]
1(Broadbent et al., 2018)To examine the role of multisensory information on incidental category learning during an attentional vigilance taskCorrelation1858.17 (0.41) 6–10 yearsMultisensory Attention Learning Task (MALT)“AWM” WISC (Digit Span Backward)r = 0.118, ns
r = 0.009 a, ns
0.12 [−0.03, 0.26]
0.01 [−0.14, 0.16]
2(Cleary et al., 2001)To examine the relationship between verbal digit span and each of the three different presentation conditions (visual, auditory and audiovisual)Correlation448.10
8–9 years
Auditory-plus-visual–spatial memory game“AWM” WISC (Digit Span Forward &Backward)r = 0.58, <0.01
r = 0.36, <0.05
0.66 [0.36, 0.97]
0.38 [0.07, 0.68]
3(Barutchu et al., 2020)See Table 2, study # 5 for further detailsAssociative Learning Tasks (ALT)“AWM”
WISC–IV WMI
r = 0.09, ns
r = 0.17, ns
0.09 [−0.24, 0.42]
0.18 [−0.16, 0.50]
Note. AWM = auditory working memory; WISC–IV = Wechsler Memory Scale; WMI = Working Memory Index. a Digit Span Backward T score; ns = not significant.

3.3.1. Participant Characteristics

A total of 1112 participants were drawn from the 21 studies, with 883 included for meta-analysis. Participant ages varied widely, ranging from 6 months to 15 years. However, the majority of studies (n = 18) focused on children aged 5 to 12 years, though the Denervaud et al. (2020b) study included children from 4 to 15 years, and two studies included infants (Choudhury et al., 2007; Ortiz-Mantilla et al., 2008).

3.3.2. Task Characteristics in Correlational Studies

As there was considerable variability in the assessment protocol and how participants responded to stimuli related to multisensory tasks in the correlation studies, this meta-analysis conducted subgroup analyses focusing on either multisensory task responded to by MRTs or verbal multisensory tasks. In the multisensory MRTs subgroup, the majority (4/6) of studies utilized a Simple Audiovisual Detection Task (e.g., Alhamdan et al., 2023b; Barutchu et al., 2011; Barutchu et al., 2020; Denervaud et al., 2020b). Two further studies in children 6–12 years (Magimairaj et al., 2009; Montgomery et al., 2009) employed MRTs in color word recognition tasks, where participants were asked to make a simple decision (press the button) about which object matched the heard colour name. These multisensory MRT tasks were usually correlated with auditory and visual WM tasks including Digit Span Tasks (Forwards and Backwards), and Complex Memory Span Tasks (see Table 3 and Table 4 for details).
In general, the verbal response to multisensory tasks focused on assessing how individuals perceive and process simultaneously presented visual and auditory information and required verbal responses (e.g., Badian, 1977; Barutchu et al., 2019; Constantinidou et al., 2011; Crawford & Fry, 1979). For example, participants (aged 7–13) were asked to verbally identify associations between visual symbols displayed on a screen and corresponding auditory cues. Only one study (Davies et al., 2019) used a Change Detection paradigm requiring participants (aged 5–6) to determine whether the shape-sound pairing was the same or different. These verbal multisensory tasks were correlated with a wider range of auditory and visual WM tasks including Digit Span Tasks, Listening Recall Tasks, Odd-One Out Tasks, and the Auditory and Visual Sequential Memory subtests of the Illinois Test of Psycholinguistic Abilities (ITPA) (see Table 3 and Table 4 for details).

3.3.3. Task Characteristics in Group Differences Studies

In group difference analysis, a single WM task was used with multi- and unisensory conditions, and effect sizes were extracted to assess the differences in WM capacity across three distinct modalities: audiovisual, auditory-alone, and visual-alone. In these tasks, participants (aged 5–12 years) were asked to verbally recall items (i.e., words or digits) presented under three conditions: (1) audiovisual, where the stimuli were presented simultaneously on a screen and verbalized, (2) auditory, where participants listened to verbalized names of objects or digits, and (3) visual, where participants were presented with visual objects or digits on a computer screen. Such tasks are generally considered as more complex than those used in correlation studies due to the involvement of multiple mental operations, including the need to integrate information from both auditory and visual modalities as well as the demands of encoding and verbal recall (see Table 5 for more details). Six of the 8 studies included in the group differences analyses involved school-age participants, while the two infants studies (Choudhury et al., 2007; Ortiz-Mantilla et al., 2008) assessed WM capacity in terms of looking time, i.e., in terms of visually driven attention (Goldberg & Wurtz, 1972) for two distinct modalities, i.e., audiovisual, and visual-alone. Further descriptions of the tasks are provided in Table 3.

3.4. Result of Bayesian Meta-Analyses

As previously mentioned, correlation data were grouped based on the type of assessment protocol of the multisensory task (multisensory MRTs, or verbal multisensory tasks). Following this, group difference data (i.e., effects sizes) were analysed to investigate the contribution of multisensory stimuli to WM capacity as compared to unisensory (auditory stimuli-alone and visual stimuli-alone). These categorisations were driven by the research aims and hypotheses of the review as well as to minimize result heterogeneity.

3.4.1. Multisensory MRTs Tasks

Six studies contributed 13 datapoints to an analysis of MRTs. Four studies used the Simple Audiovisual Detection Task, where participants were required to press a button when they heard a sound (e.g., a beep), or saw a picture (e.g., flashing or a circle), or both. Two studies (Montgomery et al., 2009; Magimairaj et al., 2009) required participants to match a spoken colour name with the corresponding visual shape by pressing the appropriate button on a touchscreen as quickly as possible. Four studies contributed more than one datapoint (e.g., Alhamdan et al., 2023b, 4 datapoints; Barutchu et al., 2020, 3 datapoints; Magimairaj et al., 2009; Montgomery et al., 2009, 2 datapoints) (see Table 3 for details). In all six studies examining school-aged children with a mean age of 8 to 9 years, MRT tasks were employed, and auditory WM was tested alone except in one study (Alhamdan et al., 2023b) where both aspects of auditory and visual Digit Span were used to assess WM capacity. The results of the Bayesian meta-analysis indicated strong evidence for the relationship between MRT and both auditory and visual WM, in favour of the alternative hypothesis (pooled correlation: μ = −0.30, 95% CI [−0.44, −0.14], BF10 = 26.84) (Figure 3). However, there was extreme evidence of heterogeneity (τ = 0.22, BF10 = 227.06, 95% CI [0.10, 0.38]).

3.4.2. Verbal Multisensory Tasks

Six studies contributed 11 datapoints to an analysis of verbally assessed multisensory tasks (see Table 3). Tasks were variable in all of these studies: Five studies required participants to verbalize the audiovisual stimuli (see Table 5 for details) (e.g., Badian, 1977; Barutchu et al., 2019; Constantinidou et al., 2011; Crawford & Fry, 1979; Pillai & Yathiraj, 2017a), and one study (Davies et al., 2019) employed a change detection audiovisual task in which participants were instructed to verbally respond ‘same’ or ‘different’ depending on whether the audiovisual pairing on the test display was the same as an old pairing. All six studies examined school-aged children from 5- to 13-years. Bayesian meta-analysis indicated extreme evidence supporting the alternative hypothesis for the association between verbal multisensory tasks and both visual and auditory WM (pooled correlation: μ = 0.45, 95% CI [0.29, 0.62], BF10 = 337.91) (Figure 4). There was very strong evidence of heterogeneity (τ = 0.21, BF10 = 33.77, 95% CI [0.08, 0.41]).
Thus, the meta-effect demonstrates that, while there is some degree of variability across studies, there is extreme evidence showing that verbal semantic multisensory tasks requiring adequate receptive language and verbal memory had a stronger association with the verbally assessed WM tasks than with MRT tasks. This likely to reflect the different cognitive demands of MRT versus verbal multisensory tasks, and may also support the existence of distinct developmental mechanisms and or more likely developmental timetables as suggested by recent brain imaging studies (Gunaydin et al., 2025). The variability in MRT outcomes may also be influenced by age-related changes in motor speed and executive control, whereas verbal tasks which involve integrated sensory, linguistic, and memory processes, may be more sensitive indicators of WM development in educational contexts.

3.4.3. The Contribution of Multisensory vs. Unisensory Stimuli to WM Capacity

As some studies compared WM capacity for multisensory stimuli within a single task, we analyzed the overall effect size for the contribution of multisensory stimuli to WM capacity in comparison to auditory- alone and visual- alone stimuli. Eight studies contributed 29 datapoints for inclusion in this meta-analysis. Six studies (Constantinidou et al., 2011; Constantinidou & Evripidou, 2012; Field & Anderson, 1985; Gillam et al., 1998; Hatchette & Evans, 1983; Pillai & Yathiraj, 2017b) required school-age participants (5–13 years) to verbally recall the presented stimuli (either items or digits), whilst two preverbal infant studies of children aged 6–9 months (Choudhury et al., 2007; Ortiz-Mantilla et al., 2008) employed a looking-time task involving only two distinct modalities: audiovisual and visual-alone to measure auditory–visual habituation/recognition memory tasks (see Table 3 for more details).
For the contribution of audiovisual stimuli versus auditory-only stimuli to WM capacity, Bayesian analysis showed moderate evidence of the alternative hypothesis, with a small effect (pooled correlation: μ = 0.29, 95% CI [0.08, 0.48], BF10 = 3.56) (Figure 5). There was moderate evidence of heterogeneity (τ = 0.22, BF10 = 3.72, 95% CI [0.05, 0.47]).
For the contribution of audiovisual stimuli versus visual-only stimuli to WM capacity, the Bayesian results of the overall effect size showed moderate evidence supporting the alternative hypothesis, with a small effect (pooled correlation: μ = 0.29, 95% CI [0.07, 0.49], BF10 = 3.06) (Figure 6; overall effect sizes). However, there was extreme evidence of heterogeneity (τ = 0.35, BF10 = 23,100, 95% CI [0.20, 0.56]).
Given the variability in the range of ages across studies (including two infant studies and the remainder involving school-age children), and the use of different measurement techniques, subgroup analyses were conducted. For infants studies (Choudhury et al., 2007; Ortiz-Mantilla et al., 2008) ‘looking time’ was used as the outcome, and results showed no evidence of differences between WM capacity for AVS compared to vs. in infants group (pooled correlation: μ = 0.10, 95% CI [−0.19, 0.40], BF10 = 0.21), with low heterogeneity (τ = 0.14, BF10 = 0.51) (Figure 6a). For school-aged children (5–13 years), when tasks required participants to verbally recall the presented stimuli (either items or digits), we found anecdotal evidence supporting the alternative hypothesis for differences between WM capacity for audiovisual stimuli compared to visual-only stimuli, with a medium effect (pooled correlation: μ = 0.33, 95% CI [0.06, 0.58], BF10 = 2.47). However, there was extreme heterogeneity (τ = 0.39, BF10 = 12,075.37) (Figure 6b).

3.5. Moderator Meta-Regression Analysis

Meta-regression analysis using traditional meta-analysis was conducted to explore whether accounting for mean age of participants as a moderator could explain the heterogeneity observed within subgroups. Results showed that the predictor variable “mean age” had a statistically significant effect on the correlation analysis between multisensory MRTs, verbal multisensory tasks, and visual and auditory WM (z = 5.66 p < 0.001, z = −2.59 p = 0.01, respectively). Heterogeneity was very low (τ2 = 8.329 × 10−6, Q (11) = 11.04, p = 0.44, I2 = 0%) in the first subgroup (i.e., multisensory MRTs), while there was moderate heterogeneity in the second subgroup (verbal multisensory tasks) (τ2 = 0.02, Q (9) = 16.62, p = 0.05, I2 = 45%). However, there was no statistically significant effect of the mean age on the group differences analysis (i.e., the contribution of multisensory vs. unisensory stimuli to WM capacity), with moderate heterogeneity (I2 65% −69%). I2 values were interpreted based on the guidelines by Higgins and Thompson (2002). Finally, we also included τ2 and the Q statistics as additional checks of heterogeneity (Borenstein et al., 2021).

3.6. Results of Narrative Synthesis

Two additional studies (Broadbent et al., 2018; Cleary et al., 2001) and two datapoints from Barutchu et al. (2020) met the inclusion criteria for this systematic review and provided correlation data, but the tasks used in these studies did not fit within the predefined categories of MRTs or verbal non motor multisensory tasks (i.e., Motor non-timed tasks (i.e., tasks involving manual motor actions but focusing on the accuracy of responses rather than response time, as defined in Section 2.5), thus these studies were not included in the meta-analysis. They are instead briefly summarized below (see Table 4).
Two of the three studies (Barutchu et al., 2020; Broadbent et al., 2018) employed Multisensory Learning Tasks, and required participants to respond to specific target stimuli by pressing the appropriate button. In particular, Barutchu et al. (2020) used an Associate Learning Task, where participants were required to learn associations between visual symbols and auditory sounds, while Broadbent et al. (2018) used a Multisensory Attention Learning Task (MALT), where the focus was on how sensory cues (visual, auditory, or combined) affect attention and learning, particularly in the context of animal categorization. Results across the two studies showed no correlation between multisensory learning tasks and digit span tasks. Lastly (Cleary et al., 2001) used a Multisensory Spatial Memory Game, where participants were asked to press appropriate buttons on a response box when presented with sounds (colour names) and or flashing buttons. Results of this study showed better more accurate performance on the multisensory spatial memory game that significantly correlated with both forward and backward digits. Importantly, all tasks only measured accuracy of response rather than speed of responses (i.e., how fast participants can identify the stimuli). Results of these studies were variable and inconclusive (see Table 4 for all correlation values). Further description of multisensory tasks can be found in Table 5.
To summarize, our results showed that verbal multisensory tasks were more strongly and consistently associated with verbally assessed auditory and visual WM than motor reaction time (MRT) tasks, suggesting distinct cognitive demands and developmental trajectories. The meta-analyses also revealed a modest advantage of multisensory over unisensory stimuli in supporting WM capacity. Moderator analysis indicated that age significantly influenced outcomes within both MRT and verbal multisensory subgroups, but not in group difference effects. Finally, studies that did not align with the predefined task categories were narratively reviewed and showed mixed, inconclusive results.

4. Discussion

The current systematic review and Bayesian meta-analyses aimed to assess the current literature examining the association between multisensory processing, including motor and verbal multisensory tasks, and measures of auditory and visual WM in children up to 15 years nothing that there is very little consideration of attention and WM in preverbal children (Teller et al., 1986; Teller, 1997). In fact, there are only a few mesures of timing response to either visual auditory or multisensory in infants expects for the two included studies (Choudhury et al., 2007; Ortiz-Mantilla et al., 2008) that investigated infant attention in terms of looking time. Based on the assumption that visual attention correlates with WM as in primates and older children, the review also aimed to explore the contribution of multisensory stimuli to WM capacity compared to unisensory auditory-only or visual-only stimuli in the same post verbal population. A total of 21 studies of moderate to high quality met the inclusion criteria, and 19 studies contributed to the correlation or group difference meta-analyses.
Findings from this review provide three main insights.
  • First, verbal multisensory processing was extremely/decisively associated with visual and auditory WM in children, while non-verbal motor reaction time multisensory tasks showed strong correlations with auditory and visual WM, suggesting verbal and MRTs multisensory processing contribute differently to WM performance, and might follow different developmental trajectories as suggested by recent brain imaging study (Gunaydin et al., 2025; Sydnor et al., 2025).
  • Furthermore, as expected, age was a significant moderator for the association between multisensory processing and visual and auditory WM, highlighting both the developmental maturation of cognitively complex WM and multisensory processing abilities.
  • Lastly, we found strong evidence that WM capacity for both verbal and MRT multisensory stimuli was higher than either visual or auditory stimuli alone. These results will first be discussed with reference to age and development, followed by the association between MRTs and verbal multisensory tasks and WM, and finally the contributions of multisensory stimuli to WM capacity will be considered.

4.1. Age Development in Motor and Verbal Multisensory Tasks and Their Relationship with WM

The previously noted a behavioural developmental study conducted by Tremblay et al. (2007) suggested that non-speech (MRT) and verbal aspects of multisensory processing may be dissociable and exhibit different developmental stages. This suggestion is supported by neuroimaging studies showing that in both adults and children, the site of activation of neurons during audiovisual processing varies based on the nature of the stimuli—verbal or nonverbal (Gkintoni et al., 2025; Hocking & Price, 2009; Thierry & Price, 2006).
Our findings of extreme/decisive evidence for an association between verbal multisensory tasks and visual and auditory WM, as opposed to a weak, though significant, associations between MRTs and visual and auditory WM, partially support a task function distinction and fit well with previous psychophysical and imaging studies (Hocking & Price, 2009; Okada & Hickok, 2009; Tremblay et al., 2007). This motor time dissociation has also been demonstrated well elsewhere, as both verbal and motor multisensory tasks have been shown to support cognitive functions such as learning and WM (Alhamdan et al., 2023b; Fifer et al., 2013; Heikkilä et al., 2015; Heikkilä & Tiippana, 2016). Such sensori-motor development with age is well known to follow different temporal trajectories and neural pathways both in terms of time and accuracy (Brandwein et al., 2011; Nardini et al., 2016). As noted earlier, simple hand motor function appears in infants at 4–5-months, yet there is much later maturation of the complex inter-relating motor pathways and muscles of diaphragm, throat, esophagus, tongue, lips, jaws, and cheeks, associated with sound production and the emergence of deliberate/conscious oral verbalization starting around 12-months (Chartier et al., 2018; Dionne-Dostie et al., 2015). Prior research has also shown that the time needed to access the verbal lexicon store necessary to allow individuals to retrieve and manipulate linguistic information is greater than time for motor responses alone (Bates et al., 1986; Bucci & Freedman, 1978). Furthermore, our findings highlight that the capacity to integrate multisensory information from different sensory sources, especially for verbal material, may enhance the efficient usage of information in more complex WM tasks. Such findings are consistent with the idea that multisensory integration of letter-speech sounds is associated with reading ability (Blau et al., 2010; Snowling et al., 2022), and the development of fluent reading skills (Bradley & Bryant, 1983; Elhassan et al., 2017; Froyen et al., 2011) in primary school grades. Further, Massaro et al. (1986) found a positive correlation in children between lip-reading ability and multisensory speech information, suggesting that lip-reading performance for audiovisual speech using McGurk effect reaches adult like levels sometime after the child’s sixth year (Massaro et al., 1986). Despite these insights, a significant developmental gap remains in the behavioural literature, particularly concerning children aged 2–4 years—a critical period for the emergence of receptive and expressive verbal language, multisensory processing and early WM capacity and time when testing compliance is difficult to maintain and an experimenter cannot be certain a child understands the task instructions of request or wishes to co-operate. Hence, Future studies should adopt age-appropriate, play-based paradigms using nonverbal cues (e.g., eye-tracking or preferential looking) to investigate these preverbal processes often on hand held device. Additionally, translating such findings into practice requires more targeted recommendations requiring audiovisual technology and devices for both teaching of motor skills and verbal language measurement. For example, using audiovisual cues to support early language comprehension in toddlers or differentiated multisensory instruction based on WM profiles in primary-aged children. Such efforts would enhance both the theoretical and applied contributions of multisensory research in developmental contexts.
Considering the growing evidence linking multisensory processing to cognitive development, practical integration into home and educational settings is both necessary and timely. Effective multisensory teaching strategies in infants include pairing spoken words with visual and tactile stimuli, then in early school age incorporating tactile ma-terials into reading and math instruction, and using interactive technologies that engage multiple senses simultaneously (Cuturi et al., 2022; Mathias et al., 2022; Shams & Seitz, 2008). These methods align with findings that multisensory input enhances neural effi-ciency i.e., the brain’s ability to process information more effectively (Park & Reu-ter-Lorenz, 2009) and learning outcomes, particularly in tasks that require or assess working memory and attention (Barutchu et al., 2020; Davies et al., 2019). Moreover, differentiated instruction based on developmental stages—for example, using more concrete sensory cues in early childhood and abstract multimodal tasks in later years can help accommodate individual learning profiles (Alhamdan et al., 2023b; Bremner et al., 2012). Embedding such strategies into everyday classroom activities holds promise for improving academic performance and supporting neurodiverse learners (Dobinson & Dockrell, 2021; Yi & Heidari Matin, 2025).
As alluded to above, the results from the various types of multisensory tasks, especially in the domains of verbal responses and motor processing, indicate varying strengths of correlations between performance of multisensory tasks and accuracy of visual and auditory WM performance, while also highlighting that different types of tasks and the requisite battery of cognitive skills to accomplish the task can affect research findings e.g., ability to articulate particularly long words (Gathercole & Baddeley, 1990). Indeed, all the tasks included in our meta-analysis required a range of other cognitive abilities, including the ability to attend to, process, and respond to the instructions, with an adequate level of receptive and expressive language skills needed to understand instructions prior to responding either verbally or with motor nonverbal responses. Thus, future research needs to consider these different aspects of multisensory tasks as well as a child’s abilities in receptive and expressive language and motor capability when assessing the association between multisensory processing and cognitive development.
In addition, our hypothesis that age and the rapid and extensive brain growth and significant changes during infancy and early childhood (Christian et al., 2017; Hedman et al., 2012) associated brain function should significantly moderate the association between multisensory tasks for both verbal and MRTs was well supported by our meta-regression. This is consistent with much previous research showing age-related improvement in multisensory processing (Barutchu et al., 2019; Brandwein et al., 2011) and cognitive abilities such as attention and visual WM (Buss et al., 2014; Buss et al., 2018). Indeed, our own research has demonstrated age-related improvements in multisensory MRTs that become faster and more efficient in older children, and then become slower in older adults later in life (Ebaid & Crewther, 2020); multisensory MRTs are also associated with better WM performance in both visual and auditory tasks (Alhamdan et al., 2022, 2023a, 2023b). Equally important is the fact that knowledge and experience may also provide further explanation for age-related improvement. Specifically, with the maturation of children, the growth of knowledge in using processing strategies and increasing familiarity with the task are considered essential factors contributing to the development of both multisensory and the visual and auditory aspects of WM (Anderson & Wagovich, 2010; S. J. Pickering, 2001; Stokes & Bors, 2001).

4.2. Contribution of Multisensory Stimuli to WM Capacity

In addition to examining the associations between multisensory tasks (MRTs and verbal) and visual and auditory WM, this review considered how WM capacity may improve when presented with multisensory stimuli compared to unisensory visual or auditory stimuli alone. Notably, tasks used in the group difference analyses in 6 out of 8 studies were considered more complex than those used in correlation meta-analyses due to the involvement of multiple mental operations, which required participants to integrate information from both auditory and visual modalities, retain this information in memory (i.e., items or digits), and orally recall the items to successfully complete articulation of such verbally based tasks. Overall, our findings demonstrated moderate evidence that multisensory stimuli contributed significantly to WM capacity more than unisensory visual or auditory stimuli alone. This suggests that more information can be retained and utilized with WM when children are provided with input from multiple senses. This finding is consistent with several past studies that have demonstrated WM capacity is usually greater for audiovisual stimuli than unimodal stimuli in both adults and children aged 7 to 13 years (Heikkilä et al., 2015; Heikkilä & Tiippana, 2016).
Although the contribution of age was not a significant moderator of these group difference analyses, the subgroups analyses (infants and school-age children) suggest that age-related effects do occur, given that in infant studies (Choudhury et al., 2007; Ortiz-Mantilla et al., 2008) the results indicated that both multisensory and visual-only stimuli showed comparable/similar WM attentional performances. However, our findings from school-age children did show anecdotal evidence of differences in the contributions of multisensory stimuli to both auditory and visual WM. Whilst previous studies have suggested that multisensory facilitation (i.e., matching and transferring information across senses) appears early in development of young infants (D. J. Lewkowicz & Lickliter, 1994; Bahrick & Lickliter, 2012; Flom & Bahrick, 2007; G. Lewkowicz, 2009), we did not observe any differences between multisensory and visual stimuli-alone, although this finding may be due to the limited number of studies included in the infant subgroup, which necessarily only included pre-verbal infants. Furthermore, there is evidence that the actual combination and integration of multisensory stimuli appears to develop later in childhood (Alhamdan et al., 2022; Barutchu et al., 2009; Brandwein et al., 2011). This is likely because these processes are dependent on the maturation of cortical brain regions and attention capabilities which continue to develop throughout later childhood and early adolescence (Corbetta & Shulman, 2002; Kanaka et al., 2008; Mabbott et al., 2006; Smith & Chatterjee, 2008). Importantly, however, the inclusion of preverbal infant studies alongside those with older children shows a conceptual gap in maturity that must be acknowledged but not the total lack of such neural and cognitive mechanisms in infant audiovisual processing (e.g., perceptual binding, attentional orienting of young infants’ eyes (first few days), mouth towards nipple for drinking and head (6–8 weeks and to mother’s voice few days). A considerable body of literature relating to visually driven eye movements and attention and multisensory (visual to auditory sounds) very early in life fundamentally the same but less accurate from the executive and memory-based processes underlying WM performance in school-aged children. Thus, although both age groups engage in multisensory processing, the functions and outcomes of that processing differ in efficiency substantially, they should be interpreted as points along a single developmental continuum. Future research should aim to better define the temporal trajectory of these mechanisms preverbally, possibly by adopting age old ‘looking time” and habituation (Horowitz et al., 1972) and animal research (Goldberg & Wurtz, 1972) with preverbal monkeys’ complementary eye movement measures of attention shifting and looking behaviours early neurophysiological and behavioural approaches tailored to the developmental stage. Taken together, these results show that age-related differences may explain the different results between infants and school-age children in the context of brain development and all forms of cognition including multisensory WM. Furthermore, considering the limited number of studies considering the development of multisensory processing during WM tasks in children, further research using psychophysics and both electrical and magnetic brain imaging is needed in this area to gain a more comprehensive understanding of both anatomical and functional age-related factors. Moreover, as the visually presented items in many of these tasks (e.g., digits, coloured stimuli) can be readily verbalized, it remains unclear whether the children, who are acquainted with the objects, employed additional verbal encoding (i.e., multisensory processing) to enhance their performance on visual-only WM tasks. Therefore, future research should address the potential confound of potentially verbalizable visual stimuli on the assessment of multisensory WM.

4.3. Limitations and Future Direction

The strength of this review has been the use of Bayesian probability statistics in accordance with recent analytical recommendations (Bartoš et al., 2021; Bartoš et al., 2023) to provide robust assessments of the strength of evidence for the alternative hypothesis. This approach was employed in two different analyses (correlational and group differences Bayesian meta-analysis) to enhance the precision and depth of our investigation into the association and contributions of multisensory stimuli to WM performance. However, there are several limitations of the literature we reviewed that also need to be considered. First, there were very few studies available for inclusion involving infants (only 2 out of 21), and none included children aged 2- to 4-years who were also noticeably absent from our meta-analysis. Consequently, the developmental trajectory in this critical early period remains understudied, and further investigation is needed to investigate the effects of multisensory processing to cognitive development, such as WM and intelligence, during these formative years. Since our review included studies with a wide age range, we attempted to address potential age-related effects by using the mean age of participants as a moderator. However, it is crucial to acknowledge that mean age might not provide a complete representation of the sample within each study, particularly those that cover a large age range (e.g., Denervaud et al., 2020b [4–15-year-olds]). Therefore, more robust research designs, with prior consideration for intelligence (both verbal and non-verbal) and working memory ability at several time points, are needed to determine age-related changes and understanding of the longitudinal developmental trajectory of multisensory and WM relations.
Furthermore, the scarcity of studies that measured visual WM in our review presents another severe limitation given that vision is recognized as the primary information channel of the primate brain (H. E. Pickering et al., 2022). Most included studies exclusively focussed on traditional auditory-verbal WM digit tasks, in line with the prominent multicomponent view of language as the basis of WM as a cognitive psychological issue (Baddeley, 2007, 2012 and the WISC and WAIS tests), while giving comparatively little consideration to more primate like visual-based WM. However, electrophysiological and brain imaging studies have consistently shown that vision drives attention and eye movements, and that the parieto-frontal networks of the dorsal visual stream of the brain carry both spatial and temporal information fastest and hence dominate attention and working memory manipulations from the time of birth (e.g., Johnson, 1990; Nelson, 1995; Laycock et al., 2022). Therefore, including assessments of both visual and auditory WM would provide a more comprehensive view of how multisensory processing affects different aspects of WM. Additionally, the visually presented digit span tasks employed in some of the included studies could be easily verbalized, meaning children may have employed additional verbal encoding to enhance their task performance, which may confound the ability to accurately compare visual and multisensory WM abilities. Thus, it will be important for future research considering visual WM to also include tasks with stimuli that are less conducive to verbalization, as this will help to minimize unintentional multisensory processing for visual tasks. Moreover, other individual and contextual factors such as socioeconomic status, bilingualism, lip reading (Massaro et al., 1986) and cultural variation in sensory processing were rarely considered in the included studies. These variables may influence attentional control, language processing, and familiarity with task stimuli, and should be accounted for in future research to enhance the ecological validity and generalizability of findings across diverse populations.

4.4. Conclusion and Implications

This current review aimed to explore whether multisensory processing using different types of tasks (verbal and MRTs) contributes to visual and auditory WM performance. Overall, our findings provide preliminary insight into the impact of multisensory tasks such as verbal and motor RT tasks that contribute differentially to WM performance. The main findings from our Bayesian meta-analysis revealed that multisensory processing (assessed verbally) was decisively associated with visual and auditory WM, while MRTs of multisensory processing also showed a strong association with visual and auditory WM tasks. These findings suggest that performances on verbal and motor multisensory tasks contribute differentially to WM performances, which supports observations that these abilities appear to have different developmental trajectories, presumably due to how motor and visuomotor function develops in childhood (Alhamdan et al., 2022, 2023a, 2023b). Furthermore, our results confirm that multisensory stimuli, when compared to unisensory stimuli, resulted in a higher WM capacity in school-aged children and adolescents, suggesting that multisensory information may enhance retention or manipulation of information in mind. Additionally, results from our current review also indicate that age is an important factor that moderates the association between multisensory tasks and visual and auditory WM, contributing to the expected increase in cognitive abilities, and the faster and more efficient multisensory processing seen with age. However, it is important to note that the consideration of other factors, such as years of schooling, intellectual ability, and language skills, is crucial for future studies and requires further investigation. This review has also highlighted the dearth of evidence regarding the development of such processes and abilities in pre expressively verbal toddlers aged 2 to 4 years of age.
This review also has important implications for cognitive and neuropsychological assessment of motor and speech-based tasks, in general terms of brain function and in school-based use of multisensory information in learning methods and educational activities. For example, the use of multisensory information could prove highly advantageous in enhancing learning outcomes, particularly where information retention is crucial. Moreover, this review highlighted the potential importance of multisensory processing to WM and learning skills in neurotypical children; however, it is important to acknowledge that multisensory teaching strategies may also more particularly be employed to support students with neurodevelopmental disorders such as those with learning difficulties such as dyslexia (Eroğlu et al., 2022; Hahn et al., 2014; Nijakowska, 2008, 2013) and autism spectrum disorder (ASD) (Molholm et al., 2020) or mutism. Specific investigation in these populations is necessary in terms of future educational and psychological research.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/ejihpe15080157/s1, Figure S1: PsyINFO (Ovid) Database; Figure S2: MEDLINE (Ovid) Database; Figure S3: EMBASE (Ovid) Database; Figure S4: CINAHL (EBSCO) Database; Figure S5: PubMed Database; Figure S6: Web of sciences (ISI) Database; Table S1: PRISMA Checklist (Page et al., 2021); Table S2: Summary of Amendments to PROSPERO Protocol (CRD42020148110; Alhamdan et al., 2019) (D. J. Lewkowicz & Lickliter, 1994; Barutchu et al., 2009; Brandwein et al., 2011; Nardini et al., 2008; JASP Team, 2022); Table S3: Search Strategy for all Databases; Table S4: Complete data extractions, including means, N, and SD for each group to calculate effect sizes for group difference studies.

Author Contributions

A.A.A.: Conceptualization, Formal analysis, Investigation, Methodology, Project administration, Writing—Original Draft. H.E.P.: Investigation, Validation, Writing—Review and Editing. M.J.M.: Methodology, Validation, Review and Editing. S.G.C.: Conceptualization, Methodology, Validation, Writing—Review and Editing, Supervision. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) (grant number IMSIU-DDRSP2501).

Acknowledgments

The authors are grateful for the facilities and other support given by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU) (grant number IMSIU-DDRSP2501). Authors would also like to thank Hannah Buttery (Senior Research Adviser/Library/La Trobe University) for her assistance with databases searching.

Conflicts of Interest

The authors declare that there are no conflicts of interest.

References

  1. Akoglu, H. (2018). User’s guide to correlation coefficients. Turkish Journal of Emergency Medicine, 18(3), 91–93. [Google Scholar] [CrossRef]
  2. Alais, D., Newell, F., & Mamassian, P. (2010). Multisensory processing in review: From physiology to behaviour. Seeing and Perceiving, 23(1), 3–38. [Google Scholar] [CrossRef]
  3. Alhamdan, A. A., Murphy, M., & Crewther, S. G. (2022). Age-related decrease in motor contribution to multisensory reaction times in primary school children. Frontiers in Human Neuroscience, 16, 967081. [Google Scholar] [CrossRef]
  4. Alhamdan, A. A., Murphy, M. J., & Crewther, S. G. (2023a). Visual motor reaction times predict receptive and expressive language development in early school-age children. Brain Sciences, 13(6), 965. [Google Scholar] [CrossRef]
  5. Alhamdan, A. A., Murphy, M. J., Pickering, H. E., & Crewther, S. G. (2019). A systematic review of the contribution of multisensory processing to memory performance in children (0–12-years). PROSPERO international prospective register of systematic reviews. Available online: https://www.crd.york.ac.uk/PROSPERO/display_record.php?RecordID=148110 (accessed on 29 July 2025).
  6. Alhamdan, A. A., Murphy, M. J., Pickering, H. E., & Crewther, S. G. (2023b). The contribution of visual and auditory working memory and non-verbal IQ to motor multisensory processing in elementary school children. Brain Sciences, 13(2), 270. [Google Scholar] [CrossRef] [PubMed]
  7. Anderson, J. D., & Wagovich, S. A. (2010). Relationships among linguistic processing speed, phonological working memory, and attention in children who stutter. Journal of Fluency Disorders, 35(3), 216–234. [Google Scholar] [CrossRef] [PubMed]
  8. Baddeley, A. (2003a). Working memory and language: An overview. Journal of Communication Disorders, 36(3), 189–208. [Google Scholar] [CrossRef] [PubMed]
  9. Baddeley, A. (2003b). Working memory: Looking back and looking forward. Nature Reviews Neuroscience, 4(10), 829–839. [Google Scholar] [CrossRef]
  10. Baddeley, A. (2007). Working memory, thought, and action (Vol. 45). OuP Oxford. [Google Scholar]
  11. Baddeley, A. (2012). Working memory: Theories, models, and controversies. Annual Review of Psychology, 63(1), 1–29. [Google Scholar] [CrossRef]
  12. Baddeley, A., Gathercole, S., & Papagno, C. (1998). The phonological loop as a language learning device. Psychological Review, 105(1), 158–173. [Google Scholar] [CrossRef]
  13. Baddeley, A., & Warrington, E. K. (1973). Memory coding and amnesia. Neuropsychologia, 11(2), 159–165. [Google Scholar] [CrossRef] [PubMed]
  14. Badian, N. A. (1977). Auditory-visual integration, auditory memory, and reading in retarded and adequate readers. Journal of Learning Disabilities, 10(2), 108–114. [Google Scholar] [CrossRef]
  15. Bahrick, L. E., & Lickliter, R. (2000). Intersensory redundancy guides attentional selectivity and perceptual learning in infancy. Developmental Psychology, 36(2), 190. [Google Scholar] [CrossRef]
  16. Bahrick, L. E., & Lickliter, R. (2012). The role of intersensory redundancy in early perceptual, cognitive, and social development. In A. J. Bremner, D. J. Lewkowicz, & C. Spence (Eds.), Multisensory development (pp. 183–206). Oxford University Press. [Google Scholar]
  17. Bartoš, F., Maier, M., Wagenmakers, E.-J., Doucouliagos, H., & Stanley, T. (2021). No need to choose: Robust Bayesian meta-analysis with competing publication bias adjustment methods. PsyArXiv Prepr. Available online: https://www.researchgate.net/publication/352502561_No_Need_to_Choose_Robust_Bayesian_Meta-Analysis_with_Competing_Publication_Bias_Adjustment_Methods (accessed on 28 July 2025).
  18. Bartoš, F., Maier, M., Wagenmakers, E.-J., Doucouliagos, H., & Stanley, T. D. (2023). Robust Bayesian meta-analysis: Model-averaging across complementary publication bias adjustment methods. Research Synthesis Methods, 14(1), 99–116. [Google Scholar] [CrossRef]
  19. Barutchu, A., Crewther, D. P., & Crewther, S. G. (2009). The race that precedes coactivation: Development of multisensory facilitation in children. Developmental Science, 12(3), 464–473. [Google Scholar] [CrossRef]
  20. Barutchu, A., Crewther, S. G., Fifer, J., Shivdasani, M. N., Innes-Brown, H., Toohey, S., Danaher, J., & Paolini, A. G. (2011). The relationship between multisensory integration and IQ in children [Empirical Study; Quantitative Study]. Developmental Psychology, 47(3), 877–885. [Google Scholar] [CrossRef]
  21. Barutchu, A., Fifer, J. M., Shivdasani, M. N., Crewther, S. G., & Paolini, A. G. (2020). The interplay between multisensory associative learning and IQ in children. Child Development, 91(2), 620–637. [Google Scholar] [CrossRef]
  22. Barutchu, A., Toohey, S., Shivdasani, M. N., Fifer, J. M., Crewther, S. G., Grayden, D. B., & Paolini, A. G. (2019). Multisensory perception and attention in school-age children. Journal of Experimental Child Psychology, 180, 141–155. [Google Scholar] [CrossRef] [PubMed]
  23. Bates, E., O’Connell, B., Vaid, J., Sledge, P., & Oakes, L. (1986). Language and hand preference in early development. Developmental Neuropsychology, 2(1), 1–15. [Google Scholar] [CrossRef]
  24. Beauchamp, M. S., Lee, K. E., Argall, B. D., & Martin, A. (2004). Integration of auditory and visual information about objects in superior temporal sulcus. Neuron, 41(5), 809–823. [Google Scholar] [CrossRef]
  25. Blau, V., Reithler, J., van Atteveldt, N., Seitz, J., Gerretsen, P., Goebel, R., & Blomert, L. (2010). Deviant processing of letters and speech sounds as proximate cause of reading failure: A functional magnetic resonance imaging study of dyslexic children. Brain, 133(3), 868–879. [Google Scholar] [CrossRef]
  26. Borenstein, M. (2009). Effect sizes based on means. In Introduction to meta-analysis (pp. 21–32). John Wiley & Sons. [Google Scholar] [CrossRef]
  27. Borenstein, M., Hedges, L. V., Higgins, J. P., & Rothstein, H. R. (2021). Introduction to meta-analysis. John Wiley & Sons. [Google Scholar]
  28. Braddick, O., & Atkinson, J. (2011). Development of human visual function. Vision Research, 51(13), 1588–1609. [Google Scholar] [CrossRef] [PubMed]
  29. Bradley, L., & Bryant, P. E. (1983). Categorizing sounds and learning to read—A causal connection. Nature, 301(5899), 419–421. [Google Scholar] [CrossRef]
  30. Brandwein, A. B., Foxe, J. J., Russo, N. N., Altschuler, T. S., Gomes, H., & Molholm, S. (2011). The development of audiovisual multisensory integration across childhood and early adolescence: A high-density electrical mapping study. Cerebral Cortex, 21(5), 1042–1055. [Google Scholar] [CrossRef]
  31. Bremner, A. J., Lewkowicz, D. J., & Spence, C. (2012). The multisensory approach to development. In Multisensory development (pp. 1–26). Oxford University Press. [Google Scholar] [CrossRef]
  32. Broadbent, H. J., White, H., Mareschal, D., & Kirkham, N. Z. (2018). Incidental learning in a multisensory environment across childhood. Developmental Science, 21(2), e12554. [Google Scholar] [CrossRef]
  33. Bucci, W., & Freedman, N. (1978). Language and hand: The dimension of referential competence. Journal of Personality, 46(4), 594–622. [Google Scholar] [CrossRef]
  34. Buschman, T. J., & Miller, E. K. (2022). Working memory is complex and dynamic, like your thoughts. Journal of Cognitive Neuroscience, 35(1), 17–23. [Google Scholar] [CrossRef]
  35. Buss, A. T., Fox, N., Boas, D. A., & Spencer, J. P. (2014). Probing the early development of visual working memory capacity with functional near-infrared spectroscopy. NeuroImage, 85, 314–325. [Google Scholar] [CrossRef] [PubMed]
  36. Buss, A. T., Ross-Sheehy, S., & Reynolds, G. D. (2018). Visual working memory in early development: A developmental cognitive neuroscience perspective. Journal of Neurophysiology, 120(4), 1472–1483. [Google Scholar] [CrossRef] [PubMed]
  37. Chartier, J., Anumanchipalli, G. K., Johnson, K., & Chang, E. F. (2018). Encoding of articulatory kinematic trajectories in human speech sensorimotor cortex. Neuron, 98(5), 1042–1054.e1044. [Google Scholar] [CrossRef] [PubMed]
  38. Choudhury, N., Leppanen, P. H., Leevers, H. J., & Benasich, A. A. (2007). Infant information processing and family history of specific language impairment: Converging evidence for RAP deficits from two paradigms. Developmental Science, 10(2), 213–236. [Google Scholar] [CrossRef]
  39. Christian, K. T., Megan, M. H., Anne-Lise, G., Rosa, M., Sarah-Jayne, B., Ronald, E. D., Berna, G., Armin, R., Elizabeth, R. S., Eveline, A. C., & Kathryn, L. M. (2017). Development of the cerebral cortex across adolescence: A multisample study of inter-related longitudinal changes in cortical volume, surface area, and thickness. The Journal of Neuroscience, 37(12), 3402. [Google Scholar] [CrossRef]
  40. Cleary, M., Pisoni, D. B., & Geers, A. E. (2001). Some measures of verbal and spatial working memory in eight-and nine-year-old hearing-impaired children with cochlear implants. Ear and Hearing, 22(5), 395. [Google Scholar] [CrossRef] [PubMed]
  41. Cohen, J. (1988). Statistical power analysis for the behavioral sciences. Taylor & Francis. [Google Scholar]
  42. Constantinidou, F., Danos, M. A., Nelson, D., & Baker, S. (2011). Effects of modality presentation on working memory in school-age children: Evidence for the pictorial superiority hypothesis. Child Neuropsychology, 17(2), 173–196. [Google Scholar] [CrossRef]
  43. Constantinidou, F., & Evripidou, C. (2012). Stimulus modality and working memory performance in Greek children with reading disabilities: Additional evidence for the pictorial superiority hypothesis. Child Neuropsychology, 18(3), 256–280. [Google Scholar] [CrossRef]
  44. Corbetta, M., & Shulman, G. L. (2002). Control of goal-directed and stimulus-driven attention in the brain. Nature Reviews Neuroscience, 3(3), 201–215. [Google Scholar] [CrossRef]
  45. Corbetta, M., Shulman, G. L., Miezin, F. M., & Petersen, S. E. (1995). Superior parietal cortex activation during spatial attention shifts and visual feature conjunction. Science, 270(5237), 802–805. [Google Scholar] [CrossRef] [PubMed]
  46. Crawford, J. H., & Fry, M. A. (1979). Trait-Task interaction in intra-and intermodal matching of auditory and visual trigrams. Contemporary Educational Psychology, 4(1), 1–10. [Google Scholar] [CrossRef]
  47. Crewther, S. G., Crewther, D. P., Klistorner, A., & Kiely, P. M. (1999). Development of the magnocellular VEP in children: Implications for reading disability. Electroencephalography and Clinical Neurophysiology. Supplement, 49, 123–128. Available online: http://europepmc.org/abstract/MED/10533097.
  48. Cuturi, L. F., Cappagli, G., Yiannoutsou, N., Price, S., & Gori, M. (2022). Informing the design of a multisensory learning environment for elementary mathematics learning. Journal on Multimodal User Interfaces, 16, 155–171. [Google Scholar] [CrossRef]
  49. Davies, S., Bourke, L., & Harrison, N. (2019). Does audio-visual binding as an integrative function of working memory influence the early stages of learning to write? Reading and Writing, 33(4), 835–857. [Google Scholar] [CrossRef]
  50. Denervaud, S., Gentaz, E., Matusz, P. J., & Murray, M. M. (2020a). An fMRI study of error monitoring in Montessori and traditionally-schooled children. NPJ Science of Learning, 5(1), 11. [Google Scholar] [CrossRef] [PubMed]
  51. Denervaud, S., Gentaz, E., Matusz, P. J., & Murray, M. M. (2020b). Multisensory gains in simple detection predict global cognition in schoolchildren. Scientific Reports, 10(1), 1394. [Google Scholar] [CrossRef] [PubMed]
  52. D’Esposito, M. (2007). From cognitive to neural models of working memory. Philosophical Transactions of the Royal Society B: Biological Sciences, 362(1481), 761–772. [Google Scholar] [CrossRef]
  53. Dionne-Dostie, E., Paquette, N., Lassonde, M., & Gallagher, A. (2015). Multisensory integration and child neurodevelopment. Brain Sciences, 5(1), 32–57. [Google Scholar] [CrossRef] [PubMed]
  54. Dobinson, K. L., & Dockrell, J. E. (2021). Universal strategies for the improvement of expressive language skills in the primary classroom: A systematic review. First Language, 41(5), 527–554. [Google Scholar] [CrossRef]
  55. Downes, M. J., Brennan, M. L., Williams, H. C., & Dean, R. S. (2016). Development of a critical appraisal tool to assess the quality of cross-sectional studies (AXIS). BMJ Open, 6(12), e011458. [Google Scholar] [CrossRef]
  56. Dunlap, W. P., Cortina, J. M., Vaslow, J. B., & Burke, M. J. (1996). Meta-analysis of experiments with matched groups or repeated measures designs. Psychological Methods, 1(2), 170–177. [Google Scholar] [CrossRef]
  57. Ebaid, D., & Crewther, S. G. (2020). The contribution of oculomotor functions to rates of visual information processing in younger and older adults. Scientific Reports, 10(1), 10129. [Google Scholar] [CrossRef]
  58. Elhassan, Z., Crewther, S. G., & Bavin, E. L. (2017). The contribution of phonological awareness to reading fluency and its individual sub-skills in readers aged 9- to 12-years. Frontiers in Psychology, 8, 533. [Google Scholar] [CrossRef] [PubMed]
  59. Eroğlu, G., Teber, S., Ertürk, K., Kırmızı, M., Ekici, B., Arman, F., Balcisoy, S., Özcan, Y. Z., & Çetin, M. (2022). A mobile app that uses neurofeedback and multi-sensory learning methods improves reading abilities in dyslexia: A pilot study. Applied Neuropsychology: Child, 11(3), 518–528. [Google Scholar] [CrossRef]
  60. Eysenck, M. W., Derakshan, N., Santos, R., & Calvo, M. G. (2007). Anxiety and cognitive performance: Attentional control theory. Emotion, 7(2), 336. [Google Scholar] [CrossRef]
  61. Field, D. E., & Anderson, D. R. (1985). Instruction and modality effects on children’s television attention and comprehension. Journal of Educational Psychology, 77(1), 91. [Google Scholar] [CrossRef]
  62. Fifer, J. M., Barutchu, A., Shivdasani, M. N., & Crewther, S. G. (2013). Verbal and novel multisensory associative learning in adults. F1000Research, 2, 34. [Google Scholar] [CrossRef] [PubMed]
  63. Flom, R., & Bahrick, L. E. (2007). The development of infant discrimination of affect in multimodal and unimodal stimulation: The role of intersensory redundancy. Developmental Psychology, 43(1), 238. [Google Scholar] [CrossRef]
  64. Foxe, J. J., & Simpson, G. V. (2005). Biasing the brain’s attentional set: II. Effects of selective intersensory attentional deployments on subsequent sensory processing. Experimental Brain Research, 166(3), 393–401. [Google Scholar] [CrossRef]
  65. Froyen, D., Willems, G., & Blomert, L. (2011). Evidence for a specific cross-modal association deficit in dyslexia: An electrophysiological study of letter–speech sound processing. Developmental Science, 14(4), 635–648. [Google Scholar] [CrossRef] [PubMed]
  66. Gathercole, S. E., & Baddeley, A. D. (1990). Phonological memory deficits in language disordered children: Is there a causal connection? Journal of Memory and Language, 29(3), 336–360. [Google Scholar] [CrossRef]
  67. Gathercole, S. E., Willis, C. S., Emslie, H., & Baddeley, A. D. (1992). Phonological memory and vocabulary development during the early school years: A longitudinal study. Developmental Psychology, 28(5), 887–898. [Google Scholar] [CrossRef]
  68. Gillam, R. B., Cowan, N., & Marler, J. A. (1998). Information processing by school-age children with specific language impairment: Evidence from a modality effect paradigm. Journal of Speech, Language, and Hearing Research, 41(4), 913–926. [Google Scholar] [CrossRef] [PubMed]
  69. Gkintoni, E., Vassilopoulos, S. P., & Nikolaou, G. (2025). Brain-inspired multisensory learning: A systematic review of neuroplasticity and cognitive outcomes in adult multicultural and second language acquisition. Biomimetics, 10(6), 397. [Google Scholar] [CrossRef]
  70. Goldberg, M. E., & Wurtz, R. H. (1972). Activity of superior colliculus in behaving monkey. II. Effect of attention on neuronal responses. Journal of Neurophysiology, 35(4), 560–574. [Google Scholar] [CrossRef]
  71. Goodale, M. A., & Milner, A. D. (1992). Separate visual pathways for perception and action. Trends in Neurosciences, 15(1), 20–25. [Google Scholar] [CrossRef]
  72. Grady, C. L. (2012). The cognitive neuroscience of ageing. Nature Reviews Neuroscience, 13(7), 491–505. [Google Scholar] [CrossRef]
  73. Gronau, Q. F., Heck, D. W., Berkhout, S. W., Haaf, J. M., & Wagenmakers, E.-J. (2021). A primer on Bayesian model-averaged meta-analysis. Advances in Methods and Practices in Psychological Science, 4(3), 25152459211031256. [Google Scholar] [CrossRef]
  74. Guenther, F. H., & Hickok, G. (2016). Chapter 58—Neural models of motor speech control. In G. Hickok, & S. L. Small (Eds.), Neurobiology of language (pp. 725–740). Academic Press. [Google Scholar] [CrossRef]
  75. Gunaydin, G., Moran, J. K., Rohe, T., & Senkowski, D. (2025). Causal inference shapes crossmodal postdictive perception within the temporal window of multisensory integration. bioRxiv. [Google Scholar] [CrossRef]
  76. Hackenberger, B. K. (2020). Bayesian meta-analysis now—let’s do it. Croatian Medical Journal, 61(6), 564–568. [Google Scholar] [CrossRef] [PubMed]
  77. Hahn, N., Foxe, J. J., & Molholm, S. (2014). Impairments of multisensory integration and cross-sensory learning as pathways to dyslexia. Neuroscience & Biobehavioral Reviews, 47, 384–392. [Google Scholar] [CrossRef] [PubMed]
  78. Haier, R. J., Siegel, B. V., Nuechterlein, K. H., Hazlett, E., Wu, J. C., Paek, J., Browning, H. L., & Buchsbaum, M. S. (1988). Cortical glucose metabolic rate correlates of abstract reasoning and attention studied with positron emission tomography. Intelligence, 12(2), 199–217. [Google Scholar] [CrossRef]
  79. Hatchette, R. K., & Evans, J. R. (1983). Auditory-visual and temporal-spatial pattern matching performance of two types of learning-disabled children. Journal of Learning Disabilities, 16(9), 537–541. [Google Scholar] [CrossRef] [PubMed]
  80. Hedman, A. M., van Haren, N. E. M., Schnack, H. G., Kahn, R. S., & Hulshoff Pol, H. E. (2012). Human brain changes across the life span: A review of 56 longitudinal magnetic resonance imaging studies. Human Brain Mapping, 33(8), 1987–2002. [Google Scholar] [CrossRef]
  81. Heikkilä, J., Alho, K., Hyvönen, H., & Tiippana, K. (2015). Audiovisual semantic congruency during encoding enhances memory performance. Experimental Psychology, 62(2), 123–130. [Google Scholar] [CrossRef]
  82. Heikkilä, J., & Tiippana, K. (2016). School-aged children can benefit from audiovisual semantic congruency during memory encoding. Experimental Brain Research, 234(5), 1199–1207. [Google Scholar] [CrossRef]
  83. Higgins, J. P., & Thompson, S. G. (2002). Quantifying heterogeneity in a meta-analysis. Statistics in Medicine, 21(11), 1539–1558. [Google Scholar] [CrossRef]
  84. Hitch, G. J., & Baddeley, A. (1976). Verbal reasoning and working memory. The Quarterly Journal of Experimental Psychology, 28(4), 603–621. [Google Scholar] [CrossRef]
  85. Hocking, J., & Price, C. J. (2009). Dissociating verbal and nonverbal audiovisual object processing. Brain and Language, 108(2), 89–96. [Google Scholar] [CrossRef] [PubMed]
  86. Hopfinger, J. B., & Slotnick, S. D. (2020). Attentional control and executive function. Cognitive Neuroscience, 11(1–2), 1–4. [Google Scholar] [CrossRef] [PubMed]
  87. Horowitz, F. D., Paden, L., Bhana, K., & Self, P. (1972). An infant-control procedure for studying infant visual fixations. Developmental Psychology, 7(1), 90. [Google Scholar] [CrossRef]
  88. Hubel, D. H., & Wiesel, T. N. (1963). Receptive fields of cells in striate cortex of very young, visually inexperienced kittens. Journal of Neurophysiology, 26(6), 994–1002. [Google Scholar] [CrossRef]
  89. Hullett, C. R., & Levine, T. R. (2003). The overestimation of effect sizes from f values in meta-analysis: The cause and a solution. Communication Monographs, 70(1), 52–67. [Google Scholar] [CrossRef]
  90. JASP Team. (2022). JASP (Version 0.16.3) [Computer Software]. Available online: https://jasp-stats.org/ (accessed on 30 August 2022).
  91. Johnson, M. H. (1990). Cortical maturation and the development of visual attention in early infancy. Journal of Cognitive Neuroscience, 2(2), 81–95. [Google Scholar] [CrossRef]
  92. Kanaka, N., Matsuda, T., Tomimoto, Y., Noda, Y., Matsushima, E., Matsuura, M., & Kojima, T. (2008). Measurement of development of cognitive and attention functions in children using continuous performance test. Psychiatry and Clinical Neurosciences, 62(2), 135–141. [Google Scholar] [CrossRef]
  93. Kent, R. D. (2000). Research on speech motor control and its disorders: A review and prospective. Journal of Communication Disorders, 33(5), 391–428. [Google Scholar] [CrossRef] [PubMed]
  94. Kocsis, Z., Jenison, R. L., Taylor, P. N., Calmus, R. M., McMurray, B., Rhone, A. E., Sarrett, M. E., Deifelt Streese, C., Kikuchi, Y., Gander, P. E., Berger, J. I., Kovach, C. K., Choi, I., Greenlee, J. D., Kawasaki, H., Cope, T. E., Griffiths, T. D., Howard, M. A., & Petkov, C. I. (2023). Immediate neural impact and incomplete compensation after semantic hub disconnection. Nature Communications, 14(1), 6264. [Google Scholar] [CrossRef] [PubMed]
  95. Kopp, F. (2014). Audiovisual temporal fusion in 6-month-old infants. Developmental Cognitive Neuroscience, 9, 56–67. [Google Scholar] [CrossRef]
  96. Langan, D., Higgins, J. P. T., Jackson, D., Bowden, J., Veroniki, A. A., Kontopantelis, E., Viechtbauer, W., & Simmonds, M. (2019). A comparison of heterogeneity variance estimators in simulated random-effects meta-analyses. Research Synthesis Methods, 10(1), 83–98. [Google Scholar] [CrossRef] [PubMed]
  97. Laycock, R., Stein, J. F., & Crewther, S. G. (2022). Pathways for rapid visual processing: Subcortical contributions to emotion, threat, biological relevance, and motivated behavior. Frontiers in Behavioral Neuroscience, 16, 960448. [Google Scholar] [CrossRef]
  98. Lewkowicz, D. J., & Bremner, A. J. (2020). Chapter 4—The development of multisensory processes for perceiving the environment and the self. In K. Sathian, & V. S. Ramachandran (Eds.), Multisensory perception (pp. 89–112). Academic Press. [Google Scholar] [CrossRef]
  99. Lewkowicz, D. J., & Lickliter, R. (1994). The development ofintersensory perception: Comparative perspectives. Psychology Press. [Google Scholar]
  100. Lewkowicz, G. (2009). The emergence of multisensory systems through perceptual narrowing. Trends in Cognitive Sciences, 13(11), 470–478. [Google Scholar] [CrossRef]
  101. Mabbott, D. J., Noseworthy, M., Bouffet, E., Laughlin, S., & Rockel, C. (2006). White matter growth as a mechanism of cognitive development in children. NeuroImage, 33(3), 936–946. [Google Scholar] [CrossRef]
  102. Macaluso, E., & Driver, J. (2005). Multisensory spatial interactions: A window onto functional integration in the human brain. Trends in Neurosciences, 28(5), 264–271. [Google Scholar] [CrossRef]
  103. Maegherman, G., Nuttall, H. E., Devlin, J. T., & Adank, P. (2019). Motor imagery of speech: The involvement of primary motor cortex in manual and articulatory motor imagery. Frontiers in Human Neuroscience, 13, 195. [Google Scholar] [CrossRef]
  104. Magimairaj, B., Montgomery, J., Marinellie, S., & McCarthy, J. (2009). Relation of three mechanisms of working memory to children’s complex span performance. International Journal of Behavioral Development, 33(5), 460–469. [Google Scholar] [CrossRef]
  105. Maitre, N. L., Key, A. P., Slaughter, J. C., Yoder, P. J., Neel, M. L., Richard, C., Wallace, M. T., & Murray, M. M. (2020). Neonatal multisensory processing in preterm and term infants predicts sensory reactivity and internalizing tendencies in early childhood. Brain Topography, 33(5), 586–599. [Google Scholar] [CrossRef] [PubMed]
  106. Marsman, M., & Wagenmakers, E.-J. (2017). Bayesian benefits with JASP. European Journal of Developmental Psychology, 14(5), 545–555. [Google Scholar] [CrossRef]
  107. Mason, G. M., Goldstein, M. H., & Schwade, J. A. (2019). The role of multisensory development in early language learning. Journal of Experimental Child Psychology, 183, 48–64. [Google Scholar] [CrossRef]
  108. Massaro, D. W., Thompson, L. A., Barron, B., & Laren, E. (1986). Developmental changes in visual and auditory contributions to speech perception. Journal of Experimental Child Psychology, 41(1), 93–113. [Google Scholar] [CrossRef]
  109. Mathias, B., Andrä, C., Schwager, A., Macedonia, M., & von Kriegstein, K. (2022). Twelve- and fourteen-year-old school children differentially benefit from sensorimotor- and multisensory-enriched vocabulary training. Educational Psychology Review, 34(3), 1739–1770. [Google Scholar] [CrossRef]
  110. Matusz, P. J., Wallace, M. T., & Murray, M. M. (2017). A multisensory perspective on object memory. Neuropsychologia, 105, 243–252. [Google Scholar] [CrossRef]
  111. McCulloch, D. L., & Skarf, B. (1991). Development of the human visual system: Monocular and binocular pattern VEP latency. Investigative Ophthalmology & Visual Science, 32(8), 2372–2381. [Google Scholar]
  112. Moeskops, P., Benders, M. J. N. L., Kersbergen, K. J., Groenendaal, F., de Vries, L. S., Viergever, M. A., & Išgum, I. (2015). Development of cortical morphology evaluated with longitudinal MR brain images of preterm infants. PLoS ONE, 10(7), e0131552. [Google Scholar] [CrossRef] [PubMed]
  113. Molholm, S., Murphy, J. W., Bates, J., Ridgway, E. M., & Foxe, J. J. (2020). Multisensory audiovisual processing in children with a sensory processing disorder (I): Behavioral and electrophysiological indices under speeded response conditions. Frontiers in Integrative Neuroscience, 14, 4. [Google Scholar] [CrossRef]
  114. Montgomery, J. W., Polunenko, A., & Marinellie, S. A. (2009). Role of working memory in children’s understanding spoken narrative: A preliminary investigation. Applied Psycholinguistics, 30(3), 485–509. [Google Scholar] [CrossRef]
  115. Morris, S. B., & DeShon, R. P. (2002). Combining effect size estimates in meta-analysis with repeated measures and independent-groups designs. Psychological Methods, 7(1), 105. [Google Scholar] [CrossRef]
  116. Muir, D., & Field, J. (1979). Newborn infants orient to sounds. Child Development, 431–436. Available online: https://www.jstor.org/stable/1129419 (accessed on 5 July 2021). [CrossRef]
  117. Murray, M. M., Thelen, A., Thut, G., Romei, V., Martuzzi, R., & Matusz, P. J. (2016). The multisensory function of the human primary visual cortex. Neuropsychologia, 83, 161–169. [Google Scholar] [CrossRef]
  118. Nardini, M., Bales, J., & Mareschal, D. (2016). Integration of audio-visual information for spatial decisions in children and adults. Developmental Science, 19(5), 803–816. [Google Scholar] [CrossRef] [PubMed]
  119. Nardini, M., Jones, P., Bedford, R., & Braddick, O. (2008). Development of cue integration in human navigation. Current biology, 18(9), 689–693. [Google Scholar] [CrossRef]
  120. Neil, P. A., Chee-Ruiter, C., Scheier, C., Lewkowicz, D. J., & Shimojo, S. (2006). Development of multisensory spatial integration and perception in humans. Developmental Science, 9(5), 454–464. [Google Scholar] [CrossRef]
  121. Nelson, C. A. (1995). The ontogeny of human memory: A cognitive neuroscience perspective. Developmental Psychology, 31(5), 723–738. [Google Scholar] [CrossRef]
  122. Nijakowska, J. (2008). An experiment with direct multisensory instruction in teaching word reading and spelling to Polish dyslexic learners of English. In Language learners with special needs: An international perspective (pp. 130–157). Available online: https://scholar.google.com/scholar_lookup?title=An+Experiment+with+Direct+Multisensory+Instruction+in+Teaching+Word+Reading+and+Spelling+to+Polish+Dyslexic+Learners+of+English&author=Nijakowska,+Joanna&publication_year=2008&pages=130–57 (accessed on 28 July 2025).
  123. Nijakowska, J. (2013). Multisensory structured learning approach in teaching foreign languages to dyslexic learners. In D. Gabryś-Barker, E. Piechurska-Kuciel, & J. Zybert (Eds.), Investigations in teaching and learning languages: Studies in honour of Hanna Komorowska (pp. 201–215). Springer International Publishing. [Google Scholar] [CrossRef]
  124. Okada, K., & Hickok, G. (2009). Two cortical mechanisms support the integration of visual and auditory speech: A hypothesis and preliminary data. Neuroscience Letters, 452(3), 219–223. [Google Scholar] [CrossRef]
  125. Ortiz-Mantilla, S., Choudhury, N., Leevers, H., & Benasich, A. A. (2008). Understanding language and cognitive deficits in very low birth weight children. Developmental Psychobiology: The Journal of the International Society for Developmental Psychobiology, 50(2), 107–126. [Google Scholar] [CrossRef] [PubMed]
  126. Padilla, N., Alexandrou, G., Blennow, M., Lagercrantz, H., & Ådén, U. (2015). Brain growth gains and losses in extremely preterm infants at term. Cerebral Cortex, 25(7), 1897–1905. [Google Scholar] [CrossRef] [PubMed]
  127. Page, M. J., McKenzie, J. E., Bossuyt, P. M., Boutron, I., Hoffmann, T. C., Mulrow, C. D., Shamseer, L., Tetzlaff, J. M., & Moher, D. (2021). Updating guidance for reporting systematic reviews: Development of the PRISMA 2020 statement. Journal of Clinical Epidemiology, 134, 103–112. [Google Scholar] [CrossRef] [PubMed]
  128. Park, D. C., & Reuter-Lorenz, P. (2009). The adaptive brain: Aging and neurocognitive scaffolding. Annual Review of Psychology, 60, 173–196. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  129. Pickering, H. E., Peters, J. L., & Crewther, S. G. (2022). A role for visual memory in vocabulary development: A systematic review and meta-analysis. Neuropsychology Review, 33(4), 803–833. [Google Scholar] [CrossRef]
  130. Pickering, S. J. (2001). The development of visuo-spatial working memory. Memory, 9(4–6), 423–432. [Google Scholar] [CrossRef]
  131. Pillai, R., & Yathiraj, A. (2017a). Auditory, visual and auditory-visual memory and sequencing performance in typically developing children. International Journal of Pediatric Otorhinolaryngology, 100, 23–34. [Google Scholar] [CrossRef]
  132. Pillai, R., & Yathiraj, A. (2017b). Two scoring procedures to evaluate memory and sequencing in auditory, visual and auditory-visual combined modalities. Hearing, Balance and Communication, 15(4), 214–220. [Google Scholar] [CrossRef]
  133. Postle, B. R. (2006). Working memory as an emergent property of the mind and brain. Neuroscience, 139(1), 23–38. [Google Scholar] [CrossRef]
  134. Quak, M., London, R. E., & Talsma, D. (2015). A multisensory perspective of working memory. Frontiers in Human Neuroscience, 9, 197. [Google Scholar] [CrossRef]
  135. Richards, J. E. (1985). The development of sustained visual attention in infants from 14 to 26 weeks of age. Psychophysiology, 22(4), 409–416. [Google Scholar] [CrossRef]
  136. Rossi, C., Vidaurre, D., Costers, L., Akbarian, F., Woolrich, M., Nagels, G., & Van Schependom, J. (2023). A data-driven network decomposition of the temporal, spatial, and spectral dynamics underpinning visual-verbal working memory processes. Communications Biology, 6(1), 1079. [Google Scholar] [CrossRef] [PubMed]
  137. Shams, L., & Seitz, A. R. (2008). Benefits of multisensory learning. Trends in Cognitive Sciences, 12(11), 411–417. [Google Scholar] [CrossRef]
  138. Shimi, A., Nobre, A. C., Astle, D., & Scerif, G. (2014). Orienting attention within visual short-term memory: Development and mechanisms. Child Development, 85(2), 578–592. [Google Scholar] [CrossRef] [PubMed]
  139. Silver, N. C., & Dunlap, W. P. (1987). Averaging correlation coefficients: Should Fisher’s z transformation be used? Journal of Applied Psychology, 72(1), 146. [Google Scholar] [CrossRef]
  140. Simonyan, K., & Horwitz, B. (2011). Laryngeal motor cortex and control of speech in humans. Neuroscientist, 17(2), 197–208. [Google Scholar] [CrossRef]
  141. Smith, S. E., & Chatterjee, A. (2008). Visuospatial attention in children. Archives of Neurology, 65(10), 1284–1288. [Google Scholar] [CrossRef]
  142. Snowling, M. J., Hulme, C., & Nation, K. (2022). The science of reading: A handbook. John Wiley & Sons. [Google Scholar]
  143. Soto-Faraco, S., Calabresi, M., Navarra, J., Werker, J. F., & Lewkowicz, D. J. (2012). The development of audiovisual speech perception. In Multisensory development (pp. 207–228). Oxford University Press. [Google Scholar] [CrossRef]
  144. Squire, L. R. (1992). Declarative and nondeclarative memory: Multiple brain systems supporting learning and memory. Journal of Cognitive Neuroscience, 4(3), 232–243. [Google Scholar] [CrossRef]
  145. Stein, B. E., Burr, D., Constantinidis, C., Laurienti, P. J., Alex Meredith, M., Perrault, T. J., Jr., Ramachandran, R., Röder, B., Rowland, B. A., & Sathian, K. (2010). Semantic confusion regarding the development of multisensory integration: A practical solution. European Journal of Neuroscience, 31(10), 1713–1720. [Google Scholar] [CrossRef] [PubMed]
  146. Stein, B. E., & Stanford, T. R. (2008). Multisensory integration: Current issues from the perspective of the single neuron. Nature Reviews Neuroscience, 9(4), 255–266. [Google Scholar] [CrossRef]
  147. Stokes, T. L., & Bors, D. A. (2001). The development of a same–different inspection time paradigm and the effects of practice. Intelligence, 29(3), 247–261. [Google Scholar] [CrossRef]
  148. Sydnor, V. J., Bagautdinova, J., Larsen, B., Arcaro, M. J., Barch, D. M., Bassett, D. S., Alexander-Bloch, A. F., Cook, P. A., Covitz, S., Franco, A. R., Gur, R. E., Gur, R. C., Mackey, A. P., Mehta, K., Meisler, S. L., Milham, M. P., Moore, T. M., Müller, E. J., Roalf, D. R., … Satterthwaite, T. D. (2025). Human thalamocortical structural connectivity develops in line with a hierarchical axis of cortical plasticity. Nature Neuroscience. Advance online publication. [Google Scholar] [CrossRef] [PubMed]
  149. Teller, D. Y. (1997). First glances: The vision of infants. the Friedenwald lecture. Investigative Ophthalmology & Visual Science, 38(11), 2183–2203. [Google Scholar]
  150. Teller, D. Y., McDonald, M. A., Preston, K., Sebris, S. L., & Dobson, V. (1986). Assessment of visual acuity in infants and children; the acuity card procedure. Developmental Medicine & Child Neurology, 28(6), 779–789. [Google Scholar] [CrossRef]
  151. Thierry, G., & Price, C. J. (2006). Dissociating verbal and nonverbal conceptual processing in the human brain. Journal of Cognitive Neuroscience, 18(6), 1018–1028. [Google Scholar] [CrossRef]
  152. Thompson, S. G., & Higgins, J. P. T. (2002). How should meta-regression analyses be undertaken and interpreted? Statistics in Medicine, 21(11), 1559–1573. [Google Scholar] [CrossRef] [PubMed]
  153. Todd, J. W. (1912). Reaction to multiple stimuli. Science Press. [Google Scholar]
  154. Tovar, D. A., Murray, M. M., & Wallace, M. T. (2020). Selective Enhancement of object representations through multisensory integration. The Journal of Neuroscience, 40(29), 5604–5615. [Google Scholar] [CrossRef] [PubMed]
  155. Treisman, A. M., & Gelade, G. (1980). A feature-integration theory of attention. Cognitive Psychology, 12(1), 97–136. [Google Scholar] [CrossRef]
  156. Tremblay, C., Champoux, F., Voss, P., Bacon, B. A., Lepore, F., & Théoret, H. (2007). Speech and non-speech audio-visual illusions: A developmental study. PLoS ONE, 2(8), e742. [Google Scholar] [CrossRef]
  157. Tzarouchi, L. C., Astrakas, L. G., Xydis, V., Zikou, A., Kosta, P., Drougia, A., Andronikou, S., & Argyropoulou, M. I. (2009). Age-related grey matter changes in preterm infants: An MRI study. NeuroImage, 47(4), 1148–1153. [Google Scholar] [CrossRef]
  158. Verhagen, J., Boom, J., Mulder, H., de Bree, E., & Leseman, P. (2019). Reciprocal relationships between nonword repetition and vocabulary during the preschool years. Developmental Psychology, 55(6), 1125–1137. [Google Scholar] [CrossRef]
  159. Veritas Health Innovation. (2019). Covidence systematic review software. Available online: www.covidence.org (accessed on 11 November 2020).
  160. Wang, D., Marcantoni, E., Shapiro, K. L., & Hanslmayr, S. (2025). Pre-stimulus alpha power modulates trial-by-trial variability in theta rhythmic multisensory entrainment strength and theta-induced memory effect. bioRxiv. [Google Scholar] [CrossRef]
  161. Wang, R., Chen, X., Khalilian-Gourtani, A., Yu, L., Dugan, P., Friedman, D., Doyle, W., Devinsky, O., Wang, Y., & Flinker, A. (2023). Distributed feedforward and feedback cortical processing supports human speech production. Proceedings of the National Academy of Sciences of the United States of America, 120(42), e2300255120. [Google Scholar] [CrossRef] [PubMed]
  162. Weaver, B. P., & Hamada, M. S. (2016). Quality quandaries: A gentle introduction to Bayesian statistics. Quality Engineering, 28(4), 508–514. [Google Scholar] [CrossRef]
  163. Weisberg, R. W., & Reeves, L. M. (2013). Cognition: From memory to creativity. John Wiley & Sons. [Google Scholar]
  164. Wiesel, T. N., & Hubel, D. H. (1963). Single-cell responses in striate cortex of kittens deprived of vision in one eye. Journal of Neurophysiology, 26(6), 1003–1017. [Google Scholar] [CrossRef] [PubMed]
  165. Yi, Y. J., & Heidari Matin, N. (2025). Perception and experience of multisensory environments among neurodivergent people: Systematic review. Architectural Science Review, 1–19. [Google Scholar] [CrossRef]
  166. Zackon, D. H., Casson, E. J., Stelmach, L., Faubert, J., & Racette, L. (1997). Distinguishing subcortical and cortical influences in visual attention. Subcortical attentional processing. Investigative Ophthalmology & Visual Science, 38(2), 364–371. [Google Scholar]
Figure 1. PRISMA flow diagram of study selection process for review.
Figure 1. PRISMA flow diagram of study selection process for review.
Ejihpe 15 00157 g001
Figure 2. A starburst visualization representing multisensory processing improvements, assessment tasks used to assess skills, and study sources identified following systematic review and Bayesian meta-analysis. This figure presents a hierarchical visualization of findings from a Bayesian meta-analysis examining visual, auditory, audiovisual multisensory, and working memory abilities in childhood. The inner ring represents the types of multisensory task where improvements were observed. The middle ring depicts the specific tasks used to assess these multisensory processing. The outer ring reports the study authors reporting the observed patterns. This layered representation facilitates an integrated understanding of the relationships between multisensory processing skills and WM performance, their operationalisation, and the supporting evidence base. (Source studies: Alhamdan et al., 2023b; Barutchu et al., 2011, 2019, 2020; Badian, 1977; Crawford & Fry, 1979; Choudhury et al., 2007; Constantinidou & Evripidou, 2012; Constantinidou et al., 2011; Davies et al., 2019; Denervaud et al., 2020b; Field & Anderson, 1985; Gillam et al., 1998; Hatchette & Evans, 1983; Magimairaj et al., 2009; Montgomery et al., 2009; Ortiz-Mantilla et al., 2008; Pillai & Yathiraj, 2017a, 2017b).
Figure 2. A starburst visualization representing multisensory processing improvements, assessment tasks used to assess skills, and study sources identified following systematic review and Bayesian meta-analysis. This figure presents a hierarchical visualization of findings from a Bayesian meta-analysis examining visual, auditory, audiovisual multisensory, and working memory abilities in childhood. The inner ring represents the types of multisensory task where improvements were observed. The middle ring depicts the specific tasks used to assess these multisensory processing. The outer ring reports the study authors reporting the observed patterns. This layered representation facilitates an integrated understanding of the relationships between multisensory processing skills and WM performance, their operationalisation, and the supporting evidence base. (Source studies: Alhamdan et al., 2023b; Barutchu et al., 2011, 2019, 2020; Badian, 1977; Crawford & Fry, 1979; Choudhury et al., 2007; Constantinidou & Evripidou, 2012; Constantinidou et al., 2011; Davies et al., 2019; Denervaud et al., 2020b; Field & Anderson, 1985; Gillam et al., 1998; Hatchette & Evans, 1983; Magimairaj et al., 2009; Montgomery et al., 2009; Ortiz-Mantilla et al., 2008; Pillai & Yathiraj, 2017a, 2017b).
Ejihpe 15 00157 g002
Figure 3. Forest plot showing pooled correlations (Fisher’s Z) for multisensory MRTs and auditory tasks as measures of WM capacity. Note. Observed per-study effect sizes (i.e., Fisher’s z) with 95% confidence intervals are shown in black. Estimated: the mean prior beliefs about the distribution per-study effect sizes and 95% credible intervals are shown in grey. All measures of WM capacity were auditory digits span, except that of (Alhamdan et al., 2023b), which included both visual and auditory digit span. (Source studies: Alhamdan et al., 2023b; Barutchu et al., 2011, 2020; Denervaud et al., 2020b; Magimairaj et al., 2009; Montgomery et al., 2009). Superscript numbers (1, 2, 3 & 4) after references in the fIgure indicate multiple effect sizes reported from the same source study (e.g., from different tasks, samples, or conditions).
Figure 3. Forest plot showing pooled correlations (Fisher’s Z) for multisensory MRTs and auditory tasks as measures of WM capacity. Note. Observed per-study effect sizes (i.e., Fisher’s z) with 95% confidence intervals are shown in black. Estimated: the mean prior beliefs about the distribution per-study effect sizes and 95% credible intervals are shown in grey. All measures of WM capacity were auditory digits span, except that of (Alhamdan et al., 2023b), which included both visual and auditory digit span. (Source studies: Alhamdan et al., 2023b; Barutchu et al., 2011, 2020; Denervaud et al., 2020b; Magimairaj et al., 2009; Montgomery et al., 2009). Superscript numbers (1, 2, 3 & 4) after references in the fIgure indicate multiple effect sizes reported from the same source study (e.g., from different tasks, samples, or conditions).
Ejihpe 15 00157 g003
Figure 4. Forest plot showing pooled correlations (Fisher’s Z) for verbal multisensory measures and a variety of auditory and visual tasks as measures of WM capacity. Note. Observed per-study effect sizes (i.e., Fisher’s z) with 95% confidence intervals are shown in black. Estimated: the mean prior beliefs about the distribution per-study effect sizes and 95% credible intervals are shown in grey. (Source studies: Barutchu et al., 2019; Badian, 1977; Crawford & Fry, 1979; Constantinidou et al., 2011; Davies et al., 2019; Pillai & Yathiraj, 2017a). Superscript numbers (1, 2 & 3) after references in the figure indicate multiple effect sizes reported from the same source study (e.g., from different tasks, samples, or conditions).
Figure 4. Forest plot showing pooled correlations (Fisher’s Z) for verbal multisensory measures and a variety of auditory and visual tasks as measures of WM capacity. Note. Observed per-study effect sizes (i.e., Fisher’s z) with 95% confidence intervals are shown in black. Estimated: the mean prior beliefs about the distribution per-study effect sizes and 95% credible intervals are shown in grey. (Source studies: Barutchu et al., 2019; Badian, 1977; Crawford & Fry, 1979; Constantinidou et al., 2011; Davies et al., 2019; Pillai & Yathiraj, 2017a). Superscript numbers (1, 2 & 3) after references in the figure indicate multiple effect sizes reported from the same source study (e.g., from different tasks, samples, or conditions).
Ejihpe 15 00157 g004
Figure 5. Forest plot showing pooled effect sizes (Cohen’s d) for the response to multisensory stimuli vs. auditory stimuli to WM capacity. Note. Observed per-study effect sizes (i.e., Cohen’s d) with 95% confidence intervals are shown in black. Estimated: the mean prior beliefs about the distribution per-study effect sizes and 95% credible intervals are shown in grey. (Source studies: Constantinidou & Evripidou, 2012; Constantinidou et al., 2011; Field & Anderson, 1985; Gillam et al., 1998; Hatchette & Evans, 1983; Pillai & Yathiraj, 2017b). Superscript numbers (1, 2, 3 & 4) after references in the fIgure indicate multiple effect sizes reported from the same source study (e.g., from different tasks, samples, or conditions).
Figure 5. Forest plot showing pooled effect sizes (Cohen’s d) for the response to multisensory stimuli vs. auditory stimuli to WM capacity. Note. Observed per-study effect sizes (i.e., Cohen’s d) with 95% confidence intervals are shown in black. Estimated: the mean prior beliefs about the distribution per-study effect sizes and 95% credible intervals are shown in grey. (Source studies: Constantinidou & Evripidou, 2012; Constantinidou et al., 2011; Field & Anderson, 1985; Gillam et al., 1998; Hatchette & Evans, 1983; Pillai & Yathiraj, 2017b). Superscript numbers (1, 2, 3 & 4) after references in the fIgure indicate multiple effect sizes reported from the same source study (e.g., from different tasks, samples, or conditions).
Ejihpe 15 00157 g005
Figure 6. Forest plot showing pooled effect sizes (Cohen’s d) for the response to multisensory stimuli vs. visual stimuli to WM capacity. Infant studies assessed WM in terms of looking time that has been known to correlate and reflect shift with visually driven attention (Goldberg & Wurtz, 1972). (Source studies: Choudhury et al., 2007; Constantinidou & Evripidou, 2012; Constantinidou et al., 2011; Field & Anderson, 1985; Gillam et al., 1998; Hatchette & Evans, 1983; Ortiz-Mantilla et al., 2008; Pillai & Yathiraj, 2017b). Superscript numbers (1, 2, 3 & 4) after references in the fIgure indicate multiple effect sizes reported from the same source study (e.g., from different tasks, samples, or conditions).
Figure 6. Forest plot showing pooled effect sizes (Cohen’s d) for the response to multisensory stimuli vs. visual stimuli to WM capacity. Infant studies assessed WM in terms of looking time that has been known to correlate and reflect shift with visually driven attention (Goldberg & Wurtz, 1972). (Source studies: Choudhury et al., 2007; Constantinidou & Evripidou, 2012; Constantinidou et al., 2011; Field & Anderson, 1985; Gillam et al., 1998; Hatchette & Evans, 1983; Ortiz-Mantilla et al., 2008; Pillai & Yathiraj, 2017b). Superscript numbers (1, 2, 3 & 4) after references in the fIgure indicate multiple effect sizes reported from the same source study (e.g., from different tasks, samples, or conditions).
Ejihpe 15 00157 g006
Table 1. Sample MEDLINE (Ovid) search strategy.
Table 1. Sample MEDLINE (Ovid) search strategy.
NoSearch Term (S)
1.(sound* ADJ2 (vision OR visual OR sight*)).mp
2.(multisensory ADJ2 (integration OR perspective OR processing)).mp
3.(sensory ADJ2 (integration OR cross-modal OR multiple)).mp
4.Intersensory processing/OR sensory integration/
5.((Audiovisual OR audio-visual) ADJ2 (synchrony OR information)).mp
6.Visual auditory integration.mp
7.Visual perception.mp OR exp Visual perception/
8.Auditory perception.mp OR exp Auditory perception/
9.7 and 8
10.1 or 2 or 3 or 4 or 5 or 6 or 7 or 8 or 9
11.((working OR short-term OR long-term OR verbal OR visual OR spatial) ADJ2 memory*).mp
12.memory/ or long-term memory/ or short-term memory/ or spatial memory/ or verbal memory/ or visual memory/
13.11 or 12
14.(Child* ADJ2(development OR early OR school age* OR pre-school OR young)).mp
15.((preschool OR pre-school OR school) ADJ2 student*).mp
16.Birth OR infant OR toddler OR newborn OR child*
17.elementary school students/or preschool students/childhood development/or early childhood development/exp Infant Development/
18.14 or 15 or 16 or 17
19.10 and 13 and 18
Note. ADJ is a proximity operator that finds words within a certain distance of each other. ADJ2 finds each word within 2 words of the other. This can be represented differently in different databases.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alhamdan, A.A.; Pickering, H.E.; Murphy, M.J.; Crewther, S.G. From Senses to Memory During Childhood: A Systematic Review and Bayesian Meta-Analysis Exploring Multisensory Processing and Working Memory Development. Eur. J. Investig. Health Psychol. Educ. 2025, 15, 157. https://doi.org/10.3390/ejihpe15080157

AMA Style

Alhamdan AA, Pickering HE, Murphy MJ, Crewther SG. From Senses to Memory During Childhood: A Systematic Review and Bayesian Meta-Analysis Exploring Multisensory Processing and Working Memory Development. European Journal of Investigation in Health, Psychology and Education. 2025; 15(8):157. https://doi.org/10.3390/ejihpe15080157

Chicago/Turabian Style

Alhamdan, Areej A., Hayley E. Pickering, Melanie J. Murphy, and Sheila G. Crewther. 2025. "From Senses to Memory During Childhood: A Systematic Review and Bayesian Meta-Analysis Exploring Multisensory Processing and Working Memory Development" European Journal of Investigation in Health, Psychology and Education 15, no. 8: 157. https://doi.org/10.3390/ejihpe15080157

APA Style

Alhamdan, A. A., Pickering, H. E., Murphy, M. J., & Crewther, S. G. (2025). From Senses to Memory During Childhood: A Systematic Review and Bayesian Meta-Analysis Exploring Multisensory Processing and Working Memory Development. European Journal of Investigation in Health, Psychology and Education, 15(8), 157. https://doi.org/10.3390/ejihpe15080157

Article Metrics

Back to TopTop