Abstract
A number of different language measures are used in child language acquisition studies. This raises the issue of comparability across tasks, and whether this comparability diverges depending on the specific language domain or the language population (e.g., monolinguals versus bilinguals). The current study investigates the comparability across tasks in the domains of vocabulary, morphology, and syntax in primary-school-aged sequential bilingual children with L1 Arabic/L2 English (N = 40, 5;7–12;2) and age-matched monolinguals (N = 40). We collected narrated speech samples to produce measures across language domains, and additional measures from separate vocabulary, morphology, and syntax assessments. Using a logistic regression analysis, we find a correspondence between syntax measures in monolinguals; however, we find no further correspondences in the other domains for monolinguals, and no correspondences at all in bilinguals. This suggests that assessment measures are highly task-dependent, and that a given assessment measure is not necessarily indicative of language as a whole, or even of language within a domain. We also find selective effects of age for monolinguals and both age and length of exposure (LOE) for bilinguals; in particular, while LOE predicts variation between the first and second language, age effects reflect underlying similarity across languages. We consider the implications of these effects for language assessments across populations.
1. Introduction
Different types of tasks use different metrics and measures to assess child language. This raises a key question of comparability across the different measures—that is, do they tap into the same underlying competence? This issue of comparability has important implications for the validity of different language assessments: if different assessments are designed to measure the same construct but produce different results, then the results cannot be linked to the intended construct (). This condition generalises across language assessments, but also across language experience: for example, two comparable tasks should produce similar results for monolinguals and similar results for bilinguals (; ).
Many studies on monolinguals have investigated the relationship between tasks for assessing different constructs, e.g., the relationship between ‘lower-level’ vocabulary and grammar and ‘higher-level’ narrative macrostructure (). These tasks tend to be normed with monolingual populations, without a formal comparison with bilinguals (; ); meanwhile, studies on bilinguals have focused more on relations between the assessments and individual difference factors (e.g., socio-economic status, home language) (; ). Fewer studies have directly investigated the comparability of different measures within a single language domain (e.g., vocabulary, morphology, syntax), and for both monolingual and bilingual populations. In this study, we aim to address the issue of comparability by evaluating different language assessments which are designed to measure the same underlying linguistic competence in monolinguals and bilinguals.
For each language domain, we ask whether two tasks produce measures which correspond with each other for monolingual English-speaking children and for age-matched sequential bilingual children with L1 Arabic and L2 English. We compare the children’s behaviour on a series of (English) tasks across domains: the Renfrew Word Finding Test for vocabulary (), TEGI morphology (), Colouring Book syntax (), narrative Type–Token Frequency, VOCD, mean length of utterance, and syntactic complexity. Using logistic regression analyses, we find that the correspondence between tasks is generally absent—even within a language domain—for both populations. Rather, children’s behaviour is more consistently predicted by their age, or—for bilinguals—by their length of exposure to the second language (i.e., the language of assessment). This variability across tasks may be related to differences in how tasks measure the particular features, but are also associated with ceiling effects, particularly for monolinguals.
The article is organised as follows: in Section 2 we explore how children’s lexical, morphological, and syntactic proficiency has been measured in language acquisition research, as well as comparisons of children’s behaviour across tasks. Section 3 then explores these comparisons with bilingual populations. In Section 4, we present the current research questions and hypotheses, and Section 5 describes the participants and the different language tasks. In Section 6 and Section 7, we present the results and the discussion, and Section 8 concludes the study and raises possibilities for future research.
2. Language Assessment Measures
Approaches to language testing may be conceptualised as being on a continuum ranging from item-based tasks which aim to isolate discrete language skills or ‘single domains’ (such as vocabulary, morphology, and syntax), through to highly naturalistic language sampling, for example, recording the language used by children in conversation or free-play situations. Between these is narrative testing. Narrative testing may be described as ‘multi-domain’ testing. As in conversation and free play, children must simultaneously integrate aspects of vocabulary, morphology, and syntax within a single task; however, as with single-domain measures, potential content and structure can be purposefully targeted and, to some degree, controlled for.
To compare children’s behaviour on different assessments across language domains, the current study focuses on a single domain at a time and respective narrative tools/approaches to assessment. These approaches are outlined in the following sections for different language domains, followed by a review of previous comparisons across measures.
2.1. Single-Domain Testing
Multiple standardised and experimental tests of vocabulary, morphology, and syntax have been developed. These single-domain tests can be broadly classified as receptive or expressive (also categorised as comprehension and production, respectively). Receptive tests tap into the child’s understanding of the language target, and expressive, their production. The following sub-sections briefly outline options in each area.
2.1.1. Vocabulary
Standardised receptive options for primary-school-aged children include the British Picture Vocabulary Scale (BPVS; ) and the Peabody Picture Vocabulary Test (PPVT; ). Receptive vocabulary tasks typically involve inferential reasoning and eliminating options from a test array that contains a closed set of possible responses. Correct responses on receptive tasks may also be possible via chance and/or weak or partial phonological or semantic word knowledge ().
Standardised expressive options include the Renfrew Expressive Vocabulary Test () and the Expressive Vocabulary Test (EVT-3; ). Assessment typically involves asking children to name singular, pictorially represented concrete objects or action events without the above cues (; ; ). As standardised tasks test the same lexical items with each child, they are a generally reliable test of vocabulary proficiency across a participant cohort; however, this reliability may vary across cohorts with different acquisition experience.
For example, variation in the quantity of vocabulary in children’s linguistic input over the first 2 years of life was observed by () depending on socio-economic status (SES). This effect, which varied by up to 30 million words, has been widely cited as a source of variation by SES, with knock-on effects beyond vocabulary (; ). However, the validity of this figure has been debated, depending on the operationalisation of linguistic input to children (; ; ). In addition, the quantitative deficit-based framing of the issue has also been criticised for overgeneralising assumptions about development from Western, Educated, Industrialised, Rich, and Democratic (WEIRD) populations (for a review, see ). These critiques are beyond the scope of the current study, although we return to the issue of the linguistic input in Section 7.3.2. in the context of input in the second language.
2.1.2. Morphology
Tests of receptive morphology, independent of syntax, are limited particularly for early language development. For example, the standardised Test for Reception of Grammar (TROG-2; ) may be used with children aged 4 years and above, and uses the same multiple-choice format outlined above for vocabulary. In the TROG-2, children select which of four images correspond with a sentence read aloud by the examiner, although only a small number of test items (e.g., those targeting plural noun inflections and negation) rely on accurate processing of morphological contrasts alone. Additional experimental measures have also been developed to assess tense morphology in 3-to-5-year-olds, with a similar forced-choice paradigm between dynamic event scenes ().
Expressive morphology (again, independent of syntax) is most commonly measured with elicitation tasks designed to provide obligatory contexts for use of a given morpheme (; ; ; ; ; ; ). Sentence cloze elicitation tasks feature in both the Word Structure sub-test of the Clinical Evaluation of Language Fundamentals—Fifth Edition (CELF-5; ) and criterion-referenced tools like the Test of Early Grammatical Impairment (TEGI; ), which create obligatory elicitation contexts as in (1):
This controlled elicitation context provides a direct test of a closed set of specific morphemes which may not occur naturally within samples of children’s spontaneous language. However, conversely, the controlled context may limit children to use of only the target morphemes and not reveal the true range of some children’s abilities. Furthermore, children may be successful in the controlled contexts due to the indirect modelling provided by sentence-cloze-type tasks, and because overall linguistic demands are reduced relative to natural discourse where children must integrate skills and knowledge across multiple linguistic levels (vocabulary, morphology, and syntax) ().
| (1) | ‘this baby eats, this baby ___’ |
| (target = sleeps) |
2.1.3. Syntax
Standardised tests of syntax may also be classified as receptive or expressive. For example, the expressive task of sentence repetition (also described as sentence imitation) is used commonly in clinical contexts, with monolingual and bilingual populations (e.g., ; ). Sentence repetition has many advantages in that it is simple and quick to use, and target language structures can be constructed exactly as required (). The task draws on children’s ability to repeat verbatim what was just heard (). However, this verbatim repetition still requires the language user to employ an analysis and reconstruction of the utterance rather than passively imitating the structure (). These processes involve linguistic knowledge of the structure in question; however, they also rely on short-term memory, making it more difficult to disentangle the measurement of cognitive ability and language ability (; ).
Receptive assessments of children’s syntax may alleviate some of this cognitive load (the ‘production–comprehension asymmetry’; for a review see ). For example, picture selection is utilised in both standardised and experimental measures of children’s syntax, including the standardised Diagnostic Evaluation of Language Variation (DELV, ), which involves a choice picture-matching task. For this task, which evaluates comprehension across a range of syntactic structures, participants are asked to choose the picture which best represents the sentence heard ().
Variation is also observed across picture selection tasks, however, which has been attributed to the specific demands of this task relative to other syntactic assessments (; ; ). In particular, subjects must hold a sentence in mind while comparing the sentence with the actions in a number of different pictures (). Therefore, picture-selection tasks will be easier for individuals with greater memory capacity than for those with reduced memory capacity; thus—like for sentence repetition—working memory may prove a confounding variable ().
These demands may be mitigated in part by receptive tasks which use a version of the picture-selection task with just one picture for each test sentence (e.g., ; ; ), thereby reducing the cognitive costs involved in processing multiple pictures per sentence. Instead of selecting between distinct pictures, such tasks involve a choice between items within the single picture, resulting in a lower processing demands due to fewer events in the task context (; ; ). We therefore use this type of syntax assessment for the current study.
2.2. Narrative
Very broadly, children’s narratives or ‘stories’ can be personal or fictional. Fictional narrative tasks include both narrative generation and narrative retell. In retell tasks, e.g., ‘The Bus Story Test’ () and the ‘Frog Stories’ (see ), children are tasked with retelling a story immediately after an initial narration by the examiner. In generation (or elicitation) tasks, children are tasked with creating their own story; examples include the ‘Peter and the Cat’ Narrative Assessment () and the Edmonton Narrative Norms Instrument (ENNI; ). Within both types of tasks, narration may be supported by a series of illustrations or a wordless picture book, which control, to some degree, for content. This can be useful in reducing variation between subjects and enabling more valid comparisons to be made across a cohort, particularly for smaller speech samples (; ).
Relative to assessment targeting other linguistic domains, narrative generation ‘storytelling’ tasks have high clinical validity (e.g., in the identification of language disorder; ; ), and are most representative of the majority of spontaneous language used by children in naturalistic contexts (; ). Storytelling, as an interaction pattern, is also more likely to be familiar to children from a range of cultural and linguistic backgrounds (), making it particularly suitable for bilingual children. Narration also requires children to draw on and integrate the aforementioned linguistic features and domains of vocabulary, morphology, and syntax (; ) and can be efficiently used to assess for multiple linguistic features within a single task—i.e., it is a multi-domain assessment tool.
For the purposes of assessing across children and populations, fictional narrative generation may provide a good compromise between highly individualised but ecologically valid personal narratives elicited during free play or conversation, and highly controlled but decontextualised standardised and item-based assessment of vocabulary, morphology, and syntax.
Measures of vocabulary, morphology, and syntax may be elicited from narrative samples through completing a microstructural-level analysis. Common measures for vocabulary include Number of Different Words (NDW), Type–Token Ratio (TTR), and VOCD. Measures of morphology and syntax include mean length of utterance (MLU), percent grammatical utterances (PGU), and complexity index (CI). For this study, we include a comparison of TTR, VOCD, MLU, and CI, discussed further in Section 5.2.2.
Although the advantages are many and now well-documented, narrative elicitation techniques also have some limitations. More confident and extroverted children may be able to give fuller descriptions of the story pictures, therefore affording them higher scores. Typical developmental trajectories have not yet been established for all of the above measures, and opportunities for assessing comprehension or more complex sentence structures may be limited. Although language production measures can tap the underlying understanding of the target structures to some extent, they may not provide explicit presentation of the comprehension of sentence structures. Therefore, production tasks may not provide a full depiction of the individuals’ linguistic proficiency ().
2.3. Relations Between Measures
With a range of tasks available for assessing proficiency in different language domains, a key question is comparability across measures—that is, are measures of vocabulary, morphology, and syntax elicited via targeted items within single-domain assessment comparable to those elicited via a multi-domain narrative assessment? This issue has been addressed for the domain of vocabulary in particular, and also for relations between different tasks across domains (). However, fewer studies have directly investigated how measures of vocabulary, morphology, and syntax might differ based on the elicitation task type—that is, how standardised item-based testing such as that outlined in Section 2.1 versus more naturalistic narrative testing such as that outlined in Section 2.2 might impact results ().
In an early study, () investigated the concurrent relationship(s) between vocabulary measures obtained from four standardised, item-based vocabulary measures (the PPVT-3, the EVT, the ROWPVT, and the EOWPVT; see Section 2.1.1) and a three-part language sample for children aged almost 4 to 6 years. The language sample comprised conversation, expository and narrative elements; the latter elicited using a wordless picture book from the frog series (). The strength of relationships between each of the tests and the language sample vocabulary measure, NDW, ranged from weak to moderate. Across-assessment differences were attributed, in part, to the role of context in performance. This was not seen as undesirable; rather, findings highlighted the value of bringing together different types of vocabulary assessment in determining children’s ability to adapt to situational variation. Thus, different tasks may assess different aspects of a construct. Moving up linguistic levels, () focused on the morphosyntactic level with a comparison of PGU calculated from ENNI () transcripts and CELF-3/CELF-P Recalling Sentences (in Context) sub-test raw scores. Significant and strong positive correlations were reported for children aged 4 to 9 years, suggesting that grammar may be less susceptible to task and situational influences.
Looking across domains, a number of correspondences have been observed for different tasks. For example, () investigated two groups, children aged 6 to 8 years and children aged 9 to 12 years. Children were from more diverse backgrounds, and language abilities ranged from very impaired to within the average range. Narratives were again obtained from the frog series. Standardised metrics were 11 norm-referenced sub-tests (e.g., the Formulating Sentences sub-test of the CELF-4) and categorised by the authors as word-, sentence-, or discourse-level. Different results were reported for the younger versus the older children. For the older children, the only significant correlation was for narrative sample TNW and discourse-level GORT (reading) comprehension scores, while for the younger children, narrative sample MLUs and subordination index (SI) were significantly correlated with multiple word- and sentence-level standard scores. Consistent with (), grammatical errors were also comparable across task types. Meanwhile, strong word-level skills supported a range of sentence-level tasks including sentence repetition and construction. Changing findings as a result of age were posited to be due to ceiling effects on the narrative task and highlighted the importance of selecting a task type that did not constrain children’s abilities. () subsequently compared MLUs, TNW, Clauses Per Sentence (CPS), and Words Per Sentence (WPS)—elicited via a conversational language sample—with the results of three CASL () sub-tests for children aged ~3 to 7 years. Statistically significant relationships were reported for both MLUs and WPS and the Syntactic Construction (SC) sub-test.
Similarly, () compared vocabulary and grammar measures obtained from a narrative sample (the Test of Narrative Language, TNL; ) to those obtained from norm-referenced tests (including the PPVT and the TEGI Past Tense and Third Person Singular probes; see Section 2.1.2 above for a description) for children aged ~6 years. Significant relationships were reported for narrative sample MLUs, NDW, and PGU and all norm-referenced test scores; however, the strongest across-task relationships were for those using the PGU narrative measure. PGU may be particularly useful in capturing a range of syntactic and morphological difficulties and therefore potential language disorder.
Finally, () took a broader approach in comparing composite CELF-P2 core language scores, with a range of measures obtained from a novel, digitally presented oral narrative (story retell) task administered to 5-year-old children. Adding to the robustness of morphosyntactic measures, narrative error composite scores—omitted morphemes and word-level errors such as pronoun errors—were moderately correlated with the CELF-P2 scores.
3. Monolingual and Bilingual Differences
The studies outlined in the previous section differ in terms of participant cultural and linguistic background; however, this aspect was not specifically controlled for, and studies were overall skewed toward monolingual cohorts. Worldwide, more than 50% of children now speak two or more languages (), meaning that a more nuanced understanding of bilingual children’s language test profiles is warranted. For example, comparable measures across tasks must be comparable across populations. The current study was therefore also interested in how an additional variable—bilingualism—might interact with task type and how this might vary or hold across the domains of vocabulary, morphology, and syntax.
Multiple studies (e.g., ; ; ) discuss the limitations and risks of using standardised assessments, normed on monolingual populations, with bilingual children. For example, standardised tests of vocabulary which limit bilingual children to use of just one language do not give credit for distributed vocabulary, i.e., where a word label is known in one language but not the other (). Numerous studies (e.g., ) purport that narrative elicitation contexts are less likely to disadvantage bilingual learners in this way because they give more freedom to select from a range of suitable vocabulary. Scoring guidelines for standardised tests of morphology and syntax do not typically allow for cross-linguistic influences or dialectal variation, meaning that language differences may be wrongly interpreted as ‘errors’, resulting in lower scores in this domain also (). The general format of item-based testing procedures—highly structured and based on the interaction pattern of pointing to pictures—may also be less familiar to some bilingual children, while storytelling tends to be common across languages (). Furthermore, the majority of English tests do not include bilingual children in the norming sample (). Collectively, these factors may result in different patterns of across-task performance as a function of language-learning status (i.e., monolingual or bilingual).
A growing number of studies have compared bilingual children’s performance across task types (i.e., single-domain vs. multi-domain narrative). These have been primarily targeted at enhancing the ecological, criterion, and/or diagnostic validity (i.e., classification accuracy) of a range of narrative assessment measures (e.g., ; ) for bilingual children with language disorder (LD) (e.g., ). Few to none to date have used a two-group/four-factor design and compared monolingual and bilingual children on task type for vocabulary and morphosyntax measures specifically within the same study.
Focusing on bilingual children only, () reported a strong positive correlation between NDW used by Spanish–English bilingual children during English retells of animated stories and PPVT-4 raw scores. A standardised expressive vocabulary test that would have allowed for more direct comparisons within the vocabulary domain was not administered. () somewhat replicated ’s () findings with Spanish–English bilingual 5- to 12-year-olds. Moderate positive associations between standardised test results and narrative sample measures of vocabulary and morphosyntax were found for younger, but not older children.
Given the well-documented need for caution around use of standardised assessments with bilingual populations, these reported correlations across assessment type, and how they might vary as a function of age, are of much interest. This review of different language measures demonstrates how language tasks can vary greatly in terms of how the feature is measured. These differences can influence the specific measure of language outcomes, and therefore, divergent results may manifest depending on the language measure used. The current study aims to contribute to this understanding by directly comparing language measures and by investigating measures across domains as well as language populations.
4. The Current Study
4.1. Research Questions, Hypotheses, and Predictions
In this study, our general research question concerns the link between language measures and linguistic competence; in particular, to what extent do different measures of children’s language proficiency tap into the same underlying linguistic competence? We focus on semi-structured narratives and highly structured item-based assessments, and apply this general question to two key specific contexts:
- To what extent does the construct validity of semi-structured narratives vs. highly structured item-based assessments vary depending on the language domain (vocabulary, morphology, syntax)?
- To what extent does the correspondence (or lack thereof) between language assessments highlight a difference between language populations (sequential bilingual versus monolingual)?
To address these two questions, we will consider two sets of hypotheses and respective predictions.
4.1.1. Hypothesis Set 1: Language Domains
One hypothesis is that the different measures of children’s language proficiency in a particular domain do indeed measure the same underlying grammatical knowledge for both monolinguals and bilinguals. If so, this reliability may be observed in just one domain or for multiple domains. This predicts that—within a given domain—children’s proficiency on narrative measures will correspond with their proficiency on item-based measures. In particular:
- lexical diversity in a narrative task will predict lexical proficiency in the standardised Renfrew task, and
- measures of morphological complexity (e.g., MLU) in the narrative task will predict accuracy on the standardised TEGI task, and
- measures of syntactic complexity (e.g., complexity index) in the narrative task will predict children’s syntax comprehension in experimental tasks.
Alternatively, these measures of language proficiency measure the same knowledge for some domains, but not others—again, with the same pattern for monolinguals and bilinguals. In this case, we would predict one or two of the above associations, but not all three.
Finally, if the different assessments do not measure the same underlying knowledge for any language domain, then we predict none of the above relations between assessments—neither for monolinguals nor bilinguals.
4.1.2. Hypothesis Set 2: Language Populations
The above hypotheses may be true for monolingual language acquisition, sequential bilingual acquisition, or across both types of acquisition contexts. In these cases, we would predict the above correspondences for monolinguals, sequential bilinguals, or both populations, respectively. However, if the assessments do not measure the same underlying knowledge for either population, then none of the above correspondences are predicted for any of the three domains.
5. Materials and Methods
5.1. Participants
The participants were 40 typically developing sequential bilingual children (5;7–12;2, mean = 8;5.22, SD = 1;10.26) from L1 Arabic-speaking backgrounds acquiring English as an additional language, and 40 monolingual English-speaking children (5;3–10;6, mean = 8;5.26, SD = 1;7.10). Participants were residents in Northern Ireland and attended mainstream school with English as the language of instruction. Age of L2 onset for the bilingual children ranged from 5 months to 10 years old (mean = 4;10.08, SD = 2;7.03), and length of L2 exposure ranged from 7 months to 10;6 years (mean = 3;6.29, SD = 2;1.15). The bilingual participants spoke Arabic as a main language at home (as reported by their own parents or teachers), and they, or their parents, originated from thirteen different Arabic-speaking countries.
5.2. Procedures
To compare children’s behaviour across methodologies and language domains, we used item-based and narrative tasks across the three language domains (vocabulary, morphology, and syntax). We describe the item-based and narrative tasks in the following sections: these include three separate item-based tasks (i.e., one domain per task) and one narrative task which allowed for measures across domains. Children completed all four tasks in a single testing session with a constant order of presentation, with breaks taken as needed. The experimenter, a native speaker of Northern Ireland English, facilitated all four tasks across all participants.
5.2.1. Item-Based Tasks
Vocabulary
The item-based assessment for vocabulary was the Renfrew Word Finding Vocabulary Test (). This is a standardised test evaluating children’s expressive vocabulary skills. This assessment employs a discrete, convergent, picture-naming method, which is a common instrument type used in evaluating children’s vocabulary size and involves testing responses to a sample of words which occur within a specific frequency range (; ). The Word Finding Vocabulary Test comprises 50 line-drawn pictures of objects (nouns) arranged in order of difficulty. The pictures are shown to the participant one by one, and they are asked to name each item, after which responses are scored. The discrete approach employed by this assessment is selective and measures the individual’s knowledge of specific word meanings as independent constructs separate from other features of language competence and without contextual information (). Participants were scored as either correct (1) or incorrect (0) for each of the 50 lexical items tested. Each trial was entered into the dataset individually (i.e., not as an average or composite score).
Morphology
For morphology, we used the Test of Early Grammatical Impairment (TEGI; ) to assess children’s production of the third person singular and regular and irregular past-tense morphology in English via elicitation. In particular, the TEGI assessment uses picture elicitation probes requiring participants to produce sentences using words with the target morphological features.
First, the third person singular probe evaluates the participants’ use of /-s/ or /-z/ on present-tense verbs. This probe includes a practice item followed by 10 test items. Each item shows a picture of a person who has a specific occupation (e.g., a teacher) and who is carrying out their job in the picture, with a prompt as in (2):
The target response in (2) includes a correct use of the third person singular, while incorrect uses include omission (e.g., ‘teach’) and incorrect forms (e.g., ‘teacher’). Responses were scored as correct (1) or incorrect (0) depending on whether the correct suffix was used.
| (2) | Investigator: Here is a teacher. Tell me what a teacher does. |
| Target response: A teacher teaches. |
Similarly, the past-tense probe evaluates the subject’s use of both regular (/-d/ or /-t/ final word phonemes) and irregular past-tense forms of verbs. This probe includes two practice items followed by 18 test items (10 regular, 8 irregular). Each item presents two pictures showing the same character but in two different scenarios: in the first, a character is doing an action; in the second, the character has completed the action. The past tense is elicited with a prompt as in (3):
As for the third person singular, participants’ responses were scored as either correct (1) or incorrect (0), and each trial was entered into the dataset individually.
| (3) | Investigator: Here, the girl jumps in the puddle. Now she is done. Tell me what she did. |
| Target response: She jumped in the puddle. |
Syntax
While the item-based tasks for vocabulary and morphology involved production measures, we used a comprehension measure to assess a range of syntactic structures: the Colouring Book Task, a digital colouring paradigm completed on a touchscreen PC (; ). In the Colouring Book Task, target structures are assessed by using the test sentence as a prompt to colour in a black and white picture; participants’ interpretations are then inferred based on how they colour in the pictures.
For each trial, children were presented with a black and white picture (Figure 1) with three animal characters performing the same action on each other, with one character as an agent (doing the action), and one as a patient (receiving the action), and one character as both an agent and patient.
Figure 1.
Sample item from the colouring task.
Children indicated their interpretations by colouring one character, based on the test sentences in (4) and (5).
The sentences in (4) and (5) vary in their clausal embedding (with embedding in (5)), and in canonical word order (with non-canonical order in (4b) and (5b)). This second contrast—between the (a) sentences and (b) sentences—is generally attributed to complexity (e.g., ; ; ; ; ; ; ). However, this complexity across sentence types may be attributed to various sources, which may have features in common across the structures (e.g., ; ; ; ). While the source of this complexity is beyond the scope of this study, the sentence conditions in (4) and (5) allow for two points of comparison:
| (4) | a. | Active voice: The cow washed the blue sheep. |
| b. | Passive voice: The cow was washed by the blue sheep. | |
| (5) | Experimenter: Something here is blue. | |
| [Child selects blue] | ||
| a. | Subject relative clause: There’s the sheep that washed the cow. | |
| b. | Object relative clause: There’s the sheep that the cow washed. | |
- Active/passive (in (4)) compared with subject/object relative clauses (in (5)), i.e., the sentence structure
- Simple structures ((4a) and (5a)) compared with complex structures ((4b) and (5b)), i.e., sentence complexity
The main test items consisted of 32 trials, with 8 items for each sentence structure (active voice, passive voice, subject relative clauses, object relative clauses). The test items were alternated with fillers comprising simple present-tense sentences with an animate subject and inanimate object, as in (6):
Participants were scored as either correct (1) or incorrect (0) for each of the trials, and each trial was entered into the dataset individually.
| (6) | a. Experimenter: Something here is blue. |
| [Child selects blue] | |
| b. It drives a bus. |
5.2.2. Narrative Task
To evaluate participants’ narrative language production, we used the Edmonton Narrative Norms Instrument (ENNI, )—a wordless picture narrative storytelling task. Preceding the main narrative task, participants completed a training story—a very short storytelling task—to facilitate their understanding for the narrative task. Participants were first instructed to only look at each picture one by one, and were given a few seconds to view each picture before moving on to the next one. Once the pictures were viewed, they were then presented to the children again, who were told that the examiner could not see the pictures and asked to tell the story. The children’s sentences were recorded and transcribed in the standard CHILDES format () to collect assessment measures across language domains.
Vocabulary
Children’s vocabulary in the narrative speech samples was evaluated using two measures of lexical diversity, i.e., the degree of variation in the lexical items that children produced. The first of these, Type–Token Ratio (TTR), is based on a quantitative measure of the lexical types in the sample, i.e., unique vocabulary items, and tokens—the overall number of items, including repetitions. Types and tokens reflect different aspects of the sample: while type represents the variation in lexical items within the sample, token reflects the overall quantity of items. The Type–Token Ratio (TTR) therefore is the ratio of different types to the total number of tokens in a language sample, varying between 0 and 1, with greater lexical diversity reflected by values closer to 1. For example, a child who produces 100 tokens of the same type will have a lower TTR than a child who produces 100 tokens of different types.
This calculation of lexical diversity, however, can also be impacted by sample size. Specifically, a larger language sample with a larger number of tokens often results in lower values for the TTR, while samples containing smaller numbers of tokens often result in a higher TTR (; ). To address this concern with varied sample lengths, an additional measure—the VOCD lexical diversity measure—was also included in the analysis. VOCD—from “vocabulary diversity”—is based on an analysis of the probability of new vocabulary being introduced into longer samples of speech, yielding a mathematical model of how TTR varies with token size. Therefore, the key advantage of VOCD is that it is not a function of the number of words in the sample, as it represents how TTR varies over a range of token size (). For both TTR and VOCD, we used the kideval command in CLAN, which includes these calculations (). With one narrative speech sample per participant, we collected a single value for each participant for TTR, and a single value for VOCD. Note that this contrasts with the item-based assessments, which included a value for each trial of the assessment.
Morphology
While the item-based assessment allowed for elicited production of specific morphemes—the past tense and third person singular—these same morphemes were not consistently produced in the children’s narrative speech samples. To avoid floor effects for these specific morphemes, we therefore used the more comprehensive measure of mean length of utterance (MLU) as the narrative measure of children’s morphology. MLU was measured in the narrative speech samples as an average of the number of morphemes in each utterance, and was also calculated for each participant from the kideval command in CLAN (). As for each of the vocabulary measures, a single MLU value was produced for each participant.
Syntax
The final measure calculated from the narrative task was the production of complex sentence structures, as a measure of children’s syntax. This included subordinate clauses (relative, adverbial, and wh- clauses, and direct and indirect quotations) and non-finite clauses (infinitive, wh- infinitive, and gerund clauses). Complex sentence structures were analysed based on the ENNI complexity index, which views a complex sentence as one which contains an independent or main clause and one or more dependent clauses. Dependent clauses are subordinate or non-finite clauses with a verb.
To calculate the complexity index score for narrative production, the total number of clauses (independent and dependent) was divided by the number of independent clauses. For example, the complexity index score for a narrative with 13 independent clauses and 15 dependent clauses would be calculated as in Equation (1):
Examples of narrative utterances and complexity index scoring are presented in Table 1.
Table 1.
Examples of complex and non-complex utterances from the narrative speech samples, with dependent clauses in bold.
As for vocabulary and morphology, the complexity index formula in Equation (1) produced a single measure for each participant. This resulted in a total of four individual measures for each participant from the narrative task, along with the accuracy values from each trial for the item-based tasks. To analyse these measures in the following sections, we ask whether the measures from the narrative tasks can predict children’s accuracy on the item-based tasks, for each language domain.
6. Results
In this section, we first describe the results of each task for monolinguals and bilinguals, followed by a correlation analysis across all tasks. This is followed by our regression analysis for each language domain. In summary, we find that the narrative measures generally do not predict measures on the item-based assessments, for any language domain—especially for bilinguals. Moreover, we find no correspondence by task type, i.e., children’s accuracy on item-based assessments is generally not predicted by other item-based assessments, with a similar result for the narratives. While some correspondences do emerge, the result is more nuanced, and we consider the implications across language domains in the discussion section.
6.1. Descriptive Statistics
The results of each task for monolinguals and bilinguals are presented in Figure 2, Figure 3, Figure 4, Figure 5 and Figure 6. Corresponding numerical values are provided in Appendix A.
Figure 2.
Accuracy for monolinguals and bilinguals on the Renfrew vocabulary task.
First, accuracy for monolinguals and bilinguals on the Renfrew vocabulary task is presented in Figure 2. As expected, accuracy is higher for monolinguals, with a broader distribution for bilinguals. Importantly, neither population is at floor or ceiling accuracy, which will allow for more meaningful comparisons with the other language measures.
Next, accuracy on the TEGI tasks is presented in Figure 3. This accuracy is generally higher for both populations than on the Renfrew vocabulary task, with largely ceiling accuracy for monolinguals. While this high accuracy is not unsurprising for monolinguals in the target age range (5;7–12;2), it should be taken into consideration for any correspondences in the regression analysis.
Figure 3.
Accuracy for monolinguals and bilinguals on the TEGI morphology task.
For the final item-based task, a broad range of variation is observed across populations on the Colouring Book Task as an assessment of children’s syntax (Figure 4). First, accuracy is generally at ceiling for monolinguals on all structures except for object relative clauses, meaning that the same caution as for the TEGI tasks should also be exercised for any correspondences for these structures in monolinguals.
Figure 4.
Accuracy on each syntactic structure for monolinguals and bilinguals in the Colouring Book Task.
In contrast, broader distributions are observed across all structures for bilinguals, along with lower accuracy. This contrast is especially apparent for the complex structures—passives and object relative clauses—despite the relatively lower accuracy on the latter for monolinguals.
While clear contrasts emerged between monolinguals and bilinguals on the item-based tasks, the results for the narrative measures are more mixed. First, for both lexical diversity measures—TTR and VOCD—the same distribution is observed for both populations (Figure 5). While the results are broadly distributed for both measures, the similar distributions between monolinguals and bilinguals suggest that the variation involves processes which are not dependent on specific language experience—an unexpected result for vocabulary.
Figure 5.
Type–token ratio and VOCD for monolinguals and bilinguals in the narrative task.
Finally, the narrative measure for morphology (MLU) and the measure for syntax (complexity based on embedded clauses) are both presented in Figure 6. Notably, the narrative complexity/syntax measure has the same distribution as for the measures of lexical diversity—i.e., with no difference between monolinguals and bilinguals. Rather, MLU stands out as the only narrative measure with a noticeable contrast between populations, although still with a broad enough range to allow for meaningful correspondences.
Figure 6.
MLU and complexity index for monolinguals and bilinguals in the narrative task.
6.2. Statistical Analysis
Before analysing children’s behaviour within each language domain, we conducted a correlation analysis to compare the different measures across tasks. These correlations are reported in Table 2 for monolinguals and Table 3 for bilinguals.
Table 2.
R correlation coefficients between age and language measures for monolinguals (cells with correspondences within domains are shaded).
Table 3.
R correlation coefficients between age and language measures for bilinguals (cells with correspondences within domains are shaded).
In addition to the task measures, we included children’s age as a continuous measure in the correlation analysis. As an indirect measure of overall language experience, including age as a factor may allow us to identify relations above and beyond this experience. Indeed, the correlation analysis in Table 2 and Table 3 revealed that many of the language measures were correlated with age both for monolinguals (Renfrew vocabulary; VOCD, TEGI past; and Colouring active, passive, subject relative clauses, and object relative clauses) and for bilinguals (TTR; TEGI past; and Colouring passive, subject relative clauses, and object relative clauses). However, several other correlations were observed both within and across domains, thus highlighting the need for a regression analysis to control for covarying factors.
We designed the regression analysis to determine the comparability between language measures, using logistic regression modelling with R () and lme4 (). In particular, we aimed to establish whether the language measures from the item-based assessments (i.e., standardised and experimental tasks) were predicted by the narrative measure within the same language domain, over and above the measures in other domains. Thus, within-domain relations should be indicative of common processes in tasks within the same language domain, while cross-domain relations will identify common processes across tasks, independent of the domain. Finally, to address the correlations observed with age for both bilinguals and monolinguals, we included age as an additional factor in the analysis. The analysis is presented in the following sections, by language domain.
6.2.1. Logistic Regression Analyses for Vocabulary
With the logistic regression for vocabulary, we investigated whether children’s accuracy on the Renfrew Word Finding Vocabulary Task could be predicted by either of the narrative measures for lexical diversity. Each individual trial on the Renfrew task was entered into the model with a binary accuracy variable coded in the dataset as 1 (correct) or 0 (incorrect). The continuous narrative measures—Type–Token Ratio (TTR) and VOCD—were both calculated using CLAN and centred before being entered into the regression model.
In addition to the narrative measures, we included the following fixed effects:
- Monolingual/bilingual, to address our aim of investigating language measures across populations
- Age (in years, centred at 0), to assess the role of overall language experience as mentioned above
- The following measures from the other language domains (each centred at 0):
- ○
- Narrative MLU, as the narrative measure of morphological complexity
- ○
- Narrative embedded clauses, as the narrative measure of syntactic complexity
- ○
- Average accuracy on the TEGI (both past tense and 3rd person singular)
- ○
- Average accuracy on each of the four Colouring Book structures (active, passive, subject relative clause, and object relative clause)
We also included random effects of participant and item for the Renfrew vocabulary task to account for variation across individual participants and items, respectively. The results for this model are presented in Table 4.
Table 4.
Model for vocabulary (dependent measure = Renfrew vocabulary task accuracy).
The within-domain effects in Table 4 involve TTR and VOCD (in bold), which would indicate a correspondence between accuracy on the Renfrew vocabulary task and the narrative measures of lexical diversity. However, while the main effects of TTR and VOCD are both marginally significant—(β = −4.33, Z = −1.91, p = 0.061) and (β = 0.03, Z = 1.87, p = 0.061), respectively—both interactions with monolingual/bilingual are not significant. This suggests that the Renfrew task and narrative measures of lexical diversity measure different underlying processes, or that any shared processes play a relatively minor role in predicting children’s behaviour.
Meanwhile, other factors did predict children’s accuracy on the Renfrew task. Significant factors include age (β = −0.25, Z = −2.77, p = 0.006) and monolingual/bilingual (β = −2.58, Z = −7.16, p < 0.001), as well as an interaction between age and monolingual/bilingual (β = 0.58, Z = 3.20, p = 0.001); this suggests that the age effect is selective to one of the two populations. In addition, correspondences across domains are indicated by the interactions between the TEGI past tense and monolingual/bilingual (β = −3.97, Z = −2.14, p = 0.032) and between the Colouring Book active structure and monolingual/bilingual (β = −3.86, Z = −2.29, p = 0.022).
To explore these contrasts in main effects across populations, we conducted post hoc tests with the joint_tests function from the emmeans package in R (). The results of these tests are presented in Table 5, with significant effects in bold.
Table 5.
Vocabulary post hoc tests (dependent measure = Renfrew vocabulary task accuracy).
While the post hoc tests confirm the null effects for the narrative vocabulary measures, all three of the interactions in Table 4 are reflected in Table 5. Notably, the post hoc tests reveal different contrasts across populations. While Renfrew accuracy is predicted in monolinguals by age (F = 15.79, p < 0.001) and the Colouring Book active structure (F = 5.24, p = 0.022), accuracy in bilinguals is predicted by the TEGI past tense (F = 6.54, p = 0.011) and the TEGI 3rd person singular (F = 6.06, p = 0.014). These contrasting profiles are illustrated in Figure 7.
Figure 7.
Age, Colouring Book Active, TEGI past tense, and TEGI 3rd person singular by accuracy on the Renfrew vocabulary task for monolinguals and bilinguals.
All four of the graphs in Figure 7 clearly show the main effect of monolingual/bilingual, with a significantly higher accuracy observed for monolinguals. Meanwhile, the effects reported in Table 5 are also evident, most notably:
- Greater accuracy with age is observed for monolinguals, but not for bilinguals (top left plot).
- Greater accuracy is predicted by both TEGI tasks for bilinguals, but not monolinguals (bottom row plots).
These effects contrast with the correspondence for monolinguals in Table 5 between vocabulary and the Colouring Book Active structure. This relation was not expected, both conceptually and given that performance in monolinguals for the active structure was generally at ceiling (c.f. Figure 4). Figure 7 (top right plot) reveals that this relation for monolinguals is likely amplified by outliers.
In sum, there is a clear contrast between the profiles of monolinguals and bilinguals for the domain of vocabulary: for bilinguals, a correspondence was observed across item-based tasks—between the Renfrew vocabulary assessment and both TEGI tasks—while for monolinguals, the Renfrew accuracy was mainly predicted by age. This age effect for monolinguals but not bilinguals raises a key issue regarding the role of language experience/input: the factor of age reflects experience with English for monolinguals, but experience with multiple languages for bilinguals. Instead, experience with English is measured for bilinguals by a different factor—the length of exposure to English, measured as the difference between the current age and the age of initial English exposure. This measure is equal to age for monolinguals, meaning that age effects in monolinguals may be due to experience with English alone, or to experience with language in general. We revisit this distinction in Section 6.2.4, after exploring whether this same contrast is observed for other language domains.
6.2.2. Logistic Regression Analyses for Morphology
For morphology, the item-based tasks included the past-tense and 3rd-person-singular components of the TEGI. The narrative measure in this domain was MLU—a measure of morphological complexity which increases with a richer inflectional inventory. To compare these measures, we developed a logistic regression model with TEGI accuracy as a dependent measure, with each trial entered individually in the dataset as 1 (correct) or 0 (incorrect). To distinguish between the two tasks, we included a binary fixed effect of TEGI task (third person singular/past tense), while narrative MLU was included as the within-domain measure of morphological complexity, centred at 0.
As for vocabulary, we included the binary fixed-effect monolingual/bilingual to compare populations, and age as a continuous fixed effect centred at 0. We also included the following (centred) continuous measures:
- Narrative TTR and narrative VOCD, as narrative measures of lexical complexity
- Narrative embedded clauses, as the narrative measure of syntactic complexity
- Average accuracy on the Renfrew vocabulary task
- Average accuracy on each of the four Colouring Book structures (active, passive, subject relative clause, and object relative clause)
Finally, random effects of participant and item were included for the TEGI morphology task to account for variation across these factors. The full set of results for this model is reported in Appendix B, with the significant and marginal effects in Table 6.
Table 6.
Model for morphology (dependent measure = TEGI task accuracy); see also Appendix B.
While the main effect of MLU was not significant (β = 0.12, Z = 0.34, p = 0.732), there are two key interactions in Table 6 with narrative MLU (in bold). First, there is an interaction between MLU and monolingual/bilingual (β = −2.45, Z = −3.63, p < 0.001), suggesting that MLU predicted TEGI accuracy for just one of the populations. However, the three-way interaction between MLU, TEGI task, and monolingual/bilingual (β = 2.35, Z = 3.33, p = 0.001) suggests that this effect on TEGI accuracy is further limited to just one of the TEGI tasks.
Post hoc tests are needed to identify which task, and which population—not just for MLU but also for the other three-way interactions in Table 5, which suggest further relations than just the within-domain factor of MLU. We conducted these post hoc tests again with the joint_tests function from the R emmeans package (), by monolingual/bilingual and TEGI task; the results of these tests are presented in Table 7 and Table 8.
Table 7.
Morphology post hoc tests for monolinguals on the TEGI tasks.
Table 8.
Morphology post hoc tests for bilinguals on the TEGI tasks.
The first result to note in Table 7 and Table 8 is the significant effect of narrative MLU on TEGI past tense accuracy in both monolinguals (F = 5.23, p = 0.022) and bilinguals (F = 10.94, p = 0.001). While both effects are significant, we expected a difference between them based on the three-way interaction in Table 5 between MLU, TEGI task, and monolingual/bilingual. Indeed, Figure 8 reveals a stronger effect of MLU on TEGI past tense accuracy in bilinguals than in monolinguals—the latter of which was likely amplified by outliers with a higher MLU. Thus, a within-domain correspondence for morphology is clear for bilinguals, but less so for monolinguals. However, this correspondence must be interpreted in the context of the other significant effects in Table 7 and Table 8.
Figure 8.
TEGI past tense by Narrative MLU.
The first of these other effects bears a striking resemblance to the result for vocabulary—for monolinguals, age also predicts accuracy on the TEGI past tense (F = 12.31, p = 0.001). A marginal effect is also observed for bilinguals, however (F = 2.95, p = 0.086). This contrast is illustrated in Figure 9, which shows the same trend across both populations, but with more variation for bilinguals. Meanwhile, no age effect is observed in either population for the TEGI 3rd person singular.
Figure 9.
TEGI tasks by Age.
Next, while the prediction for a within-domain correspondence is borne out for morphology—particularly for bilinguals—effects are also observed for several cross-domain language measures. In particular, TEGI past tense accuracy was predicted by the narrative complexity measure in both monolinguals (F = 7.07, p = 0.008) and bilinguals (F = 7.32, p = 0.007), and by both narrative TTR (F = 7.94, p = 0.005) and narrative VOCD (F = 5.94, p = 0.015) in bilinguals.
The first of these—narrative complexity—patterns similarly to MLU (Figure 10). Moreover, like MLU, there is a clear effect for bilinguals, while the effect for monolinguals is more likely to be driven by outliers.
Figure 10.
TEGI past tense by Narrative Complexity.
Next, while TEGI past tense accuracy is predicted in bilinguals by both narrative TTR and narrative VOCD, the interpretation of these effects is problematic, as illustrated by Figure 11. In particular, for both TTR and VOCD, higher values correspond with greater lexical diversity. However, the relation between TTR and the TEGI past tense for bilinguals in Figure 11 is negative: a higher TTR corresponds with lower TEGI accuracy. This relation is therefore more likely to reflect an issue with the assessment measure than a true relation, and we revisit this issue in the discussion. Meanwhile, no trend is apparent in Figure 11 between VOCD and the TEGI past tense for bilinguals, despite the significant effect.
Figure 11.
TEGI past tense by Narrative TTR and Narrative VOCD.
Finally, accuracy on both TEGI tasks is predicted by bilinguals’ accuracy on the Renfrew vocabulary task—past tense (F = 6.16, p = 0.013) and 3rd person singular (F = 8.55, p = 0.004). This relation is the same as the one observed for vocabulary above, where Renfrew accuracy for bilinguals was predicted by both TEGI tasks (Figure 7). Note, however, that while Renfrew accuracy was predicted only by the TEGI tasks, the TEGI past tense was also predicted by MLU and narrative complexity. Importantly, these effects provide further context to the correspondence above between morphology assessments; moreover, this context requires an analysis with measures across different language domains. The following section applies this analysis to the final language domain, syntax.
6.2.3. Logistic Regression Analyses for Syntax
For syntax, we used the Colouring Book Task described in Section 5.2.1 to assess children’s comprehension of the active voice, passive voice, subject relative clauses, and object relative clauses. Of these four structures, the passive voice and object relative clauses are expected to cause greater difficulty; this difficulty has been attributed to a range of sources (e.g., ; ; ; ; ; ). Importantly, while we do not expect the exact same sources of difficulty for passives and object relative clauses, we do expect a lower accuracy in the Colouring Book Task for both of these structures than for actives and subject relative clauses.
To include all four structures in the same analysis with accuracy as the dependent measure, we constructed a similar regression model structure as for morphology above. However, to distinguish both between complexity (passives and object relative clauses compared to actives and subject relative clauses) and between structures (actives and passives compared to subject and object relative clauses), we expanded the single-factor “task” into these two separate binary fixed effects: Complex and Structure. As for vocabulary and morphology in the previous analyses, accuracy for each of these syntactic structures was entered for each trial as 1 (correct) or 0 (incorrect).
To compare assessments within the domain of syntax, we used the narrative measure of syntactic complexity based on the number of embedded clauses produced, as described in Section 5.2.2. The measure was calculated for each narrative and centred at 0.
Next, as for vocabulary and morphology, we included the binary fixed effect monolingual/bilingual to compare populations, and age as a continuous fixed effect centred at 0. We also included the following (centred) continuous measures:
- Narrative TTR and narrative VOCD, as narrative measures of lexical complexity
- Narrative MLU, as the narrative measure of morphological complexity
- Average accuracy on the Renfrew vocabulary task
- Average accuracy on the TEGI (both past tense and 3rd person singular)
Finally, random effects were included for participant and item in the Colouring Book Task, to account for variation across individual participants and items, respectively. The full set of results for this model is reported in Appendix B, with the significant and marginal effects in Table 9.
Table 9.
Model for syntax (dependent measure = Colouring Book Task accuracy); see also Appendix B.
Key effects to focus on in Table 9 for the within-domain (syntax) correspondence involve narrative complexity (in bold). For example, while the main effect of narrative complexity is not significant (β = 3.16, Z = 2.96, p = 0.286), there is a significant interaction between narrative complexity and monolingual/bilingual (β = −13.81, Z = −2.36, p = 0.018). This interaction suggests that this narrative complexity measure predicts accuracy on the Colouring Book Task for just one of the two populations; however, a further interaction arises between narrative complexity, Complex, and monolingual/bilingual (β = 12.00, Z = 1.99, p = 0.046). This second interaction indicates that the effect of narrative complexity is also selective to just complex or non-complex structures for this population, rather than all structures in the Colouring Book Task. Given that the narrative complexity measure is based on embedded clauses, we would predict this effect to result from a correspondence with complex rather than non-complex structures, although post hoc tests are required to confirm this prediction. The interactions with vocabulary and morphology measures in Table 9 also suggest that further correspondences are present across domains, which will be similarly clarified by the post hoc tests.
We conducted these post hoc tests again with the joint_tests function from the R emmeans package (), by monolingual/bilingual, Complex, and Structure, allowing us to isolate the effects across populations and the four Colouring Book structures (e.g., “Active” is Structure = Active/passive and Complex = non-complex); the results of these tests are presented in Table 10 and Table 11.
Table 10.
Syntax post hoc tests for monolinguals on the Colouring Book Task.
Table 11.
Syntax post hoc tests for bilinguals on the Colouring Book Task.
The first result to note in Table 10 and Table 11 is in relation to the effect of narrative complexity. Based on the three-way interaction in Table 9 between narrative complexity, Complex, and monolingual/bilingual, we expected this effect to be selective to just one population, for specific structures. Indeed, the effect of narrative complexity is observed only for monolinguals, in the object relative clause structure (F = 6.08, p = 0.014). This contrast between monolinguals and bilinguals is reflected in Figure 12.
Figure 12.
Accuracy for object relative clauses by narrative complexity.
This effect of narrative complexity for monolinguals contrasts with the pattern observed for morphology, where narrative complexity predicted accuracy on the TEGI past tense for bilinguals. Moreover, the inverse result is also observed for age: while this measure predicted accuracy on the TEGI past tense for monolinguals, age does not predict monolinguals’ accuracy for any of the syntactic structures. Rather, accuracy in the syntactic domain is predicted by age for bilinguals—both for subject relative clauses (F = 5.38, p = 0.020) and object relative clauses (F = 4.15, p = 0.042). This contrast in age effects between monolinguals and bilinguals for subject and object relative clauses is illustrated in Figure 13.
Figure 13.
Accuracy for subject and object relative clauses by Age.
For subject relative clauses, monolinguals’ ceiling performance can account for the lack of any age effect; this contrasts with bilinguals, who do show a subtle effect of age. Meanwhile, both populations appear to exhibit age effects for object relative clauses. The lack of a significant effect for monolinguals may therefore be due to collinearity with other predictors.
Table 10 and Table 11 contain four additional significant results. First, significant effects of narrative TTR are reported on monolinguals’ accuracy for both the active (F = 5.08, p = 0.024) and subject relative clause (F = 4.42, p = 0.036) structures. These effects are illustrated in Figure 14, and reflect the ceiling effect for both structures in monolinguals. They are therefore likely driven by outliers for both structures, rather than being meaningful effects.
Figure 14.
Accuracy for actives and subject relative clauses, by Narrative TTR.
Finally, accuracy on the active structure is also predicted for monolinguals by the Renfrew vocabulary task and the TEGI 3rd person singular. The first of these was also observed with the Renfrew task as the dependent variable, in Figure 7 above, while the second is illustrated below in Figure 15. However, as for the relation with narrative TTR, the ceiling accuracy for monolinguals with the active structure suggests that these are not meaningful correspondences, and are likely driven by outliers.
Figure 15.
Accuracy for actives, by TEGI 3rd person singular.
In summary, for syntax, we observe a contrasting profile from the patterns for vocabulary and morphology: in particular, a within-domain correspondence is observed for monolinguals—between narrative complexity and object relative clauses—while accuracy for bilinguals is predicted only by age, for both subject and object relative clauses.
This age effect is unexpected for bilinguals, particularly in a context where no corresponding effect is observed for monolinguals. However, some additional context is missing from an effect of age. For monolinguals, age effects may reflect the role of different factors which develop in parallel. For example, while linguistic experience increases with age, so does the development of other domain-general cognitive processes which may interface with language. Meanwhile, age effects in bilinguals provide some opportunities to tease apart these factors: while age can reflect overall linguistic experience across multiple languages, experience with the second language—i.e., English in the current study—will depend on the length of exposure to the second language, rather than overall age. For monolinguals, these measures are one and the same. However, for bilinguals, these measures were not correlated (R = 0.18, p = 0.268), as illustrated in Figure 16.
Figure 16.
Length of exposure to English for bilinguals, by age.
Thus, language measures which are more dependent on specific language experience should be predicted in bilinguals by the factor of length of exposure, while overall language experience will be more closely reflected by age. In the following section, we tease apart these effects by revisiting the analysis of each language domain for bilinguals, with the addition of length of exposure as a predicting factor.
6.2.4. Bilinguals Revisited: Length of Exposure
In the previous sections, we found that age did not predict accuracy in bilinguals for vocabulary and morphology, but an age effect was observed for subject and object relative clauses. However, this pattern must be interpretated in the context of the length of exposure (LOE). This context includes four possibilities:
- With no age effect (i.e., vocabulary and morphology), we may observe:
- an effect of LOE, suggesting that the assessment involves processes that depend on exposure to the specific language.
- no effect of LOE, suggesting that the assessment is not dependent on the input.
- With an age effect, we may observe:
- an effect of LOE, suggesting that the assessment involves processes that depend both on overall language experience (reflecting the effect of age) and on experience with the specific language (reflecting LOE).
- no effect of LOE, suggesting that the assessment depends entirely on overall language experience.
We tease apart these hypotheses in the following sections by comparing a model with LOE as a fixed effect to a model without LOE, for each language domain.
Vocabulary
For vocabulary, the models were identical to the one in Section 6.2.1, with two exceptions: first, as only the bilinguals were included in the analysis, monolingual/bilingual was not included as a fixed effect. Second, the LOE model included this factor as an additional fixed effect; this continuous measure was centred before being entered into the model.
There was a significant difference between the model with LOE and the model without LOE (χ2(1) = 12.166, p < 0.001), with a lower Akaike Information Criterion (AIC) for the model with LOE (865) than for the model without LOE (875). This suggests that children’s responses for the Renfrew vocabulary task were predicted by LOE, above and beyond the other measures. The results for this model are presented in Table 12.
Table 12.
Model for vocabulary, with Length of Exposure (bilinguals only).
The key result in Table 12 is the strong effect of LOE (β = −0.43, Z = −3.88, p < 0.001), illustrated in Figure 17. With no additional effect of age, this suggests that the Renfrew task involves processes that are specific to English—an intuitive result for a vocabulary assessment. In contrast, the narrative measures for vocabulary—TTR and VOCD—remain non-significant, consistent with no within-domain correspondence for the vocabulary assessments. We return to the implications of this result in the discussion section.
Figure 17.
Renfrew vocabulary accuracy, by Length of exposure (LOE) for bilinguals.
Morphology
For morphology, the difference between the model with LOE and the model without LOE was not significant (χ2(2) = 3.70, p = 0.158), meaning that adding LOE as a fixed effect did not improve the model fit. Therefore, with no age effects observed in bilinguals in the original analysis for morphology, this suggests that accuracy on the TEGI morphology tasks for bilinguals is not related to language experience. Rather, as discussed in Section 6.2.2, these tasks were predicted for bilinguals by accuracy on the Renfrew vocabulary task, while the TEGI past tense was also predicted for bilinguals by the narrative measures for both morphology and syntax (i.e., the narrative measures of complexity). Note that Renfrew accuracy was predicted by LOE; we return to this contrast in the discussion section.
Syntax
Finally, for syntax, there was a significant difference between the model with LOE and the model without LOE (χ2(4) = 26.19, p < 0.001), with a lower AIC for the model with LOE (637) than for the model without LOE (655). This suggests that children’s responses for the Colouring Book Task were predicted by LOE, above and beyond the other measures. As this model included the four different syntactic structures differentiated by the fixed effects Structure and Complex, we conducted post hoc tests by these factors using the joint_tests function in R () to identify the relevant effects for each structure, as shown in Table 13. Notably, the effect of Length of Exposure in Table 13 is significant across all four syntactic structures. This result is illustrated in Figure 18.
Table 13.
Post hoc tests for bilinguals on the Colouring Book Task, with Length of Exposure.
Figure 18.
Accuracy for bilinguals on the Colouring Book Task, by Length of Exposure.
While the LOE effects appear to vary in Figure 18, the interactions were not significant between LOE and Structure (β = −0.49, Z = −1.29, p = 0.196), LOE and Complex (β = 0.14, Z = −0.38, p = 0.703), or LOE, Structure, and Complex (β = −0.04, Z = 0.10, p = 0.919). Meanwhile, in addition to LOE, subject and object relative clauses were also predicted by age. This combination of age and LOE effects for subject and object relative clauses is consistent with the hypothesis above that accuracy for these structures in the Colouring Book Task depends both on overall language experience (reflecting the effect of age) and on experience with the specific language (reflecting LOE). This contrasts with the Renfrew vocabulary task, which was predicted by LOE but not by age.
The results discussed in the previous sections are summarised in Table 14, with the two within-domain correspondences in bold. We consider the implications of these results in the final discussion section.
Table 14.
Summary for monolinguals (L1) and bilinguals (L2); within-domain effects in bold1.
7. Discussion
The aim of this research was to examine whether different measures of language proficiency generate comparable results, and also whether these results vary depending on the language domain (vocabulary, morphology, and syntax) and language population (sequential bilingual and monolingual). To answer this question, 40 sequential bilingual children and 40 monolingual primary-school-aged children completed three item-based language assessments—one for each language domain—and a narrative task which produced narrative language measures for each domain.
To compare different language measures across tasks and populations, we used the language assessment measures to predict children’s accuracy on the item-based assessments for each language domain. We found that the measures did not correspond within a given domain, with two exceptions:
- MLU predicted accuracy on the TEGI past tense for bilinguals. However, accuracy on this same assessment was also predicted by the syntactic measure of narrative complexity, as well as the average accuracy on the Renfrew vocabulary task.
- Narrative complexity predicted accuracy with object relative clauses for monolinguals. However, this effect was not also observed for bilinguals.
Children’s accuracy also was predicted by age, and—for bilinguals—by length of exposure to English, depending on the domain, above and beyond the effects observed for specific language assessment. In the following sections, we review these effects for each language domain and consider their implications for language assessment.
7.1. Vocabulary: Effects of English Language Experience
We assessed vocabulary with the Word Finding Vocabulary Test from the Renfrew Language Scales (), and with two lexical diversity measures calculated from the participants’ narrative productions—TTR and VOCD. A key consideration for these vocabulary measures is the role of language experience. For monolinguals, this experience can be generally reflected by their age, which in turn predicted their accuracy on the Renfrew task. In contrast, age for bilinguals is more likely to represent language experience in both the L1 and the L2 than in just English (the L2). This contrast between age for monolinguals and age for bilinguals is supported by the significant effect of Length of Exposure (LOE) for bilinguals on Renfrew accuracy. As the populations were age-matched, the monolinguals had more exposure to English, which in turn predicts the difference in overall accuracy on the Renfrew task between monolinguals and bilinguals (Figure 2).
Meanwhile, no difference was observed between monolinguals and bilinguals in TTR or VOCD (Figure 5). Moreover, the correspondence for bilinguals between TTR and Renfrew accuracy was in the reverse direction of that which would otherwise be predicted: bilingual children with higher accuracy on the Renfrew task had lower narrative TTRs. In contrast, bilingual children with higher accuracy on the Renfrew task had higher VOCD scores. While this relation would be consistent with similar processes in both vocabulary measures, this is inconsistent with the lack of a difference in VOCD between bilinguals and monolinguals. That is, if Renfrew accuracy depends on specific language (e.g., English) experience, and the same processes are involved in both the Renfrew task and in VOCD, then VOCD should also depend on specific language experience, thus predicting a difference between monolinguals and bilinguals in VOCD.
The results therefore do not clearly provide evidence for shared processes between the different vocabulary assessments, aligning with previous work by ()—for bilinguals in addition to monolinguals. Rather, while specific language experience may play a role for the Renfrew task, this experience is less clearly linked to TTR and VOCD as measures of lexical diversity.
7.2. Morphology: No Single Common Source
For morphology, we assessed children’s production of the past tense and third person singular morphemes in the standardised TEGI task, and the mean length of utterance (MLU) in the children’s narratives. As for the Renfrew assessment task, we observed a contrast between bilinguals and monolinguals for both measures: monolinguals were more accurate than bilinguals on the TEGI tasks, and produced narratives with higher MLUs. However—also as for the Renfrew task—accuracy for monolinguals on the TEGI was predicted only by age, with no correspondence for monolinguals between either TEGI task and the other language measures.
This suggests that the TEGI tasks—like the Renfrew vocabulary task—would depend on language experience, either overall or specific to English. However, the results for bilinguals are not consistent with this: unlike in the Renfrew vocabulary task, accuracy for the TEGI in bilinguals was not predicted by LOE. Instead, we observed a correspondence between TEGI accuracy for bilinguals and accuracy on the Renfrew vocabulary task, MLU, and narrative complexity. While the correspondence between the TEGI and MLU is within domains, the two additional correspondences (with the Renfrew task and narrative complexity) could be indicative of some more general common process which affects performance across domains. For example, domain-general executive functions which interface with language may support both stronger performance in the Renfrew task and the procedures involved in planning and producing complex utterances—both morphologically and syntactically.
One consideration for this common source, however, is that TEGI accuracy was not predicted by any of the structures in the Colouring Book Task. This rules out generalisations across all item-based assessments. Moreover, since TEGI accuracy was predicted by MLU and narrative complexity, but not by TTR or VOCD, this rules out broader generalisations about correspondences with the narrative assessment measures. Rather, the three different correspondences for bilinguals between TEGI accuracy and accuracy on the Renfrew task, MLU, and narrative complexity are more likely to reflect different common processes (), to be explored further in future research.
7.3. Syntax: A Contrast Between the Different Structures
The assessments for syntax were the Colouring Book comprehension task of complex and non-complex structures (), and a syntactic complexity index for the narratives, based on embedded clauses. As for vocabulary and morphology, the correspondence between these tasks was limited. However, we also observed effects of Length of Exposure in bilinguals, which provide additional context for the LOE effects on vocabulary.
7.3.1. Assessments of Complex Syntax
Importantly, a measure of complex syntax is produced by both the Colouring Book Task and the narrative assessment—the Colouring Book Task via accuracy on complex structures (passives and object relative clauses), and the narrative complexity measure based on children’s use of embedded clauses. However, the narrative complexity measure predicted accuracy only for monolinguals—for object relative clauses—and not for bilinguals. This has important implications for the interpretation of the narrative complexity measure in both populations.
First, the Colouring Book Task included two complex structures: the passive voice and object relative clauses. However, the passive voice was not predicted by any other language measure, including the narrative complexity measure. The effect of narrative complexity for object relative clauses but not for passives is consistent with different sources of difficulty for the two structures—that is, object relative clause accuracy and the use of complex syntax in the narratives may reflect the deployment of similar processes. For example, one such process could relate to the use of embedded structures that is shared by both complex narrative structures and object relative clauses. This contrasts with the passive voice, which involves a different type of syntactic complexity.
This interpretation for narrative complexity is specific to monolinguals, however, as the same correspondence was not observed for bilinguals. That is, while the narrative complexity measure may reflect underlying competence for monolinguals (based on the correspondence with object relative clause accuracy), this same inference is not available for bilinguals. This is a key consideration for the narrative complexity measure, as a measure of complexity: complexity itself may be operationalised in terms of linguistic competence (e.g., based on structural complexity) or performance (based on the deployment of complex parsing procedures). The results for bilinguals therefore suggest that the two tasks produce two fundamentally different measures of complex syntax, which do not involve a shared component at the level of syntactic competence. That is, while comprehension of the complex structures in the Colouring Book Task requires explicit knowledge of these structures, the narrative complexity measure includes a wider range of structures. In assessing children’s complex syntax, low accuracy for comprehension of passives and object relative clauses is therefore not an indicator that a child will necessarily produce few complex structures in a narrative task. Correspondingly, a narrative with few complex structures will not necessarily predict incorrect comprehension of passives and relative clauses. Further implications of this result for selecting and interpreting language assessments are discussed in Section 7.4.
7.3.2. Sources of Age Effects
For the Colouring Book comprehension task—unlike for the Renfrew (vocabulary) and TEGI (morphology) tasks, we did not find an effect of age on response accuracy in monolinguals. Meanwhile, age effects were observed in bilinguals for subject and object relative clauses. In addition to these age effects, however, we also observed an effect of Length of Exposure (LOE).
For vocabulary, the effect of LOE suggested that the relevant processes in the Renfrew tasks depended on specific experience with English. However, with syntax, the dual effects of age and LOE suggest a more complex picture. This seems to indicate that knowledge of syntax is dependent both on the overall language experience of bilinguals (measured by age), and on more specific language experience (measured by length of exposure to English). This is consistent with the LOE effects observed across all four structures in the Colouring Task, which vary between English and Arabic, i.e., the L2 and L1 of the bilingual participants in this study, respectively.
For example, we observed LOE effects for the active structure, which occurs uniformly in English with a subject–verb–object order (SVO). However, the word order in Arabic includes both VSO and SVO orders, as in (7) ():
Next, we observed LOE effects for the passive structure, which includes an optional “by” phrase in English, as in (8):
However, the equivalent of a by-phrase is not available in Arabic; only short passives are possible, as in (9) ():
Finally, variation in the form of relative clauses is observed between English and Arabic: while English requires a gap in the place of the moved element, Arabic allows an optional resumptive element, shown here in bold ():
| (7) | a. | raʔa-a | l-ʔawlaad-u | Zayd-an |
| saw | The-boys-NOM | Zayd-ACC | ||
| ‘The boys saw Zayd’ (VSO) | ||||
| b. | l-ʔawlaad-u | raʔa-a | Zayd-an | |
| The-boys-NOM | saw | Zayd-ACC | ||
| ‘The boys saw Zayd’ (SVO) | ||||
| (8) | The cow was washed (by the sheep). |
| (9) | a. | Active: | ʔa-ktubu | d-dars-a |
| 1.act.def-write | the-lesson-ACC | |||
| ‘You write the lesson.’ | ||||
| b. | Passive: | y-u-ktabu | d-dars-u | |
| 3-ind.pass-write | the-lesson-NOM | |||
| The lesson is written. | ||||
| (10) | mnaʕrif | l-mara | lli | raħ | yʔeebəl-a | Saami |
| know-1p | the-woman | that | will | meet-her | Sami | |
| ‘We know the woman that Sami will meet’ | ||||||
However, this variation in form for relative clauses (i.e., the presence or absence of a resumptive pronoun) contrasts with the structural variation for actives and passives. That is, the same underlying structure is observed for relative clauses in both English and Arabic, and this structural similarity is consistent with the additional age effects that were observed in bilinguals for relative clauses, but not actives and passives. In particular, competence with relative clause structures is also dependent on overall language experience, independent of the specific language. This reliability of the Colouring Book Task for relative clauses as an assessment of underlying syntax has broader implications for language assessments in bilinguals, considered further in the following section.
7.4. Implications for Assessment
In this study, we compared children’s behaviour across different language assessments for vocabulary, morphology, and syntax. The results include two key takeaways. First, the assessments involve linguistic knowledge from each of these domains respectively, but for the most part we observed no correspondence between different measures from the same domain. Second, the correspondences which did arise were different between monolinguals and bilinguals.
The first takeaway highlights the limitations of generalising beyond a single assessment measure to other aspects of linguistic competence—even within the same domain. This limitation stems from the extent to which different processes are shared across different tasks. For example, the correspondence for bilinguals between the TEGI morphology and Renfrew vocabulary is unlikely to have the same source as the correspondence between the TEGI and narrative MLU, as discussed in Section 7.2. Thus, while the TEGI corresponds with both the Renfrew vocabulary measure and MLU, the underlying processes involved in the Renfrew vocabulary task are less likely to generalise to MLU. More generally, for an assessment of children’s language within a given domain, a single measure may provide an incomplete picture, while a complete language profile will require a battery of expressive and receptive measures across language domains.
Next, the different correspondences observed between monolinguals and bilinguals illustrate some of the qualitative differences between the two populations. For example, as discussed in Section 7.3.1, the narrative complexity measure predicted object relative clauses for monolinguals, but not for bilinguals; this contrast suggests that the narrative complexity measure is indicative of different constructs for monolinguals and bilinguals. In addition, the distributions of language measures were different for the two populations on most of the assessments, highlighting the notion that the bilingual speaker is not simply two monolinguals in one (). Rather, separate norms are needed for each population, to reflect the different expected trajectories. These norms are developed for the purposes of comparisons within the normed population; therefore, comparing bilingual assessments with monolingual norms defeats the very purpose of these norms. To do so discounts the differences between these populations (). Previous research on norms for bilingual populations has focused largely in this domain on vocabulary (e.g., ; ), while the mixed effects of age and LOE in the current study suggest a more complex profile, particularly for syntax.
8. Conclusions
The findings from this study suggest that not all language proficiency measures provide comparable results, and that results vary depending on the language domain/feature and the language population under investigation. In addition, we observed different correspondences in monolinguals and bilinguals, highlighting the need for a range of assessments and norms for both populations. Future research may also focus on different L1/L2 combinations which vary in their contrasting features (in structure and/or form), as well as a broader range of expressive and receptive measures, to gain a more complete picture of variation across assessment type.
Author Contributions
Conceptualisation, S.L., C.S., R.F. and J.G.; methodology, S.L.; formal analysis, S.L. and J.G.; investigation, S.L.; resources, S.L.; data curation, S.L.; writing—original draft preparation, S.L., V.S. and J.G.; writing—review and editing, J.G., V.S., C.S. and R.F.; visualisation, J.G.; supervision, J.G., C.S. and R.F. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by a Department for the Economy Postgraduate Studentship to S.L.
Institutional Review Board Statement
The study was conducted in accordance with the Declaration of Helsinki, and approved by the Ulster University Risk and Ethics Committee Institutional Review Board (protocol code REC.17.0095 on 24 May 2018) for studies involving humans.
Informed Consent Statement
Informed consent was obtained from all subjects involved in the study.
Data Availability Statement
The original data presented in the study are openly available at https://osf.io/va47z.
Acknowledgments
We would like to thank the families and schools that participated in our study, and the audiences at BUCLD 44 and LAGB 2019 for valuable feedback on the issues in this paper. We would also like to thank Shalom Zuckerman and Manuela Pinto for support on the Colouring Book task, and Dora Alexopoulou for helpful discussions throughout the project.
Conflicts of Interest
The authors declare no conflicts of interest.
Appendix A
Table A1.
Numerical scores for item-based assessments.
Table A1.
Numerical scores for item-based assessments.
| Task Type | Task | Population | Mean Accuracy | SD | Range |
|---|---|---|---|---|---|
| Renfrew vocabulary | Vocabulary | Monolingual | 0.84 | 0.09 | 0.62–0.96 |
| Bilingual | 0.57 | 0.18 | 0.18–0.86 | ||
| TEGI morphology | Past tense | Monolingual | 0.93 | 0.13 | 0.44–1 |
| Bilingual | 0.81 | 0.27 | 0.06–1 | ||
| 3rd person singular | Monolingual | 0.95 | 0.11 | 0.4–1 | |
| Bilingual | 0.73 | 0.32 | 0–1 | ||
| Colouring Book (Syntax) | Active | Monolingual | 0.94 | 0.14 | 0.38–1 |
| Bilingual | 0.82 | 0.24 | 0–1 | ||
| Passive | Monolingual | 0.93 | 0.12 | 0.38–1 | |
| Bilingual | 0.63 | 0.34 | 0–1 | ||
| Subject relative clause | Monolingual | 0.97 | 0.07 | 0.75–1 | |
| Bilingual | 0.87 | 0.16 | 0.38–1 | ||
| Object relative clause | Monolingual | 0.85 | 0.24 | 0.12–1 | |
| Bilingual | 0.62 | 0.31 | 0–1 |
Table A2.
Numerical scores for narrative assessment.
Table A2.
Numerical scores for narrative assessment.
| Task Type | Task | Population | Mean Score | SD | Range |
|---|---|---|---|---|---|
| Vocabulary | TTR | Monolingual | 0.53 | 0.12 | 0.32–0.84 |
| Bilingual | 0.56 | 0.12 | 0.4–0.77 | ||
| VOCD | Monolingual | 31.75 | 11.6 | 15.68–65.82 | |
| Bilingual | 29.72 | 11.1 | 9.2–62.47 | ||
| Morphology | MLU | Monolingual | 11.52 | 3.97 | 4.88–23.08 |
| Bilingual | 9.51 | 3.29 | 2.89–16 | ||
| Syntax | Complexity index | Monolingual | 1.42 | 0.23 | 1–2.15 |
| Bilingual | 1.32 | 0.34 | 0–1.88 |
Appendix B
Table A3.
Full model for morphology; significant effects in bold.
Table A3.
Full model for morphology; significant effects in bold.
| Fixed Effect | β | SE | Z | p | |
|---|---|---|---|---|---|
| Intercept | 4.67 | 0.73 | 6.40 | <0.001 | *** |
| Age | −0.52 | 0.25 | −2.05 | 0.040 | * |
| Narr TTR | −8.09 | 10.63 | −0.76 | 0.447 | |
| Narr VOCD | 0.11 | 0.05 | 2.14 | 0.033 | * |
| Narr MLU | 0.12 | 0.34 | 0.34 | 0.732 | |
| Narr Complexity | 0.95 | 1.63 | 0.58 | 0.561 | |
| Renfrew vocab | −5.71 | 4.18 | −1.37 | 0.172 | |
| CB Active | 0.09 | 1.68 | 0.06 | 0.955 | |
| CB Passive | −1.53 | 2.02 | −0.76 | 0.448 | |
| CB Subject RC | −4.40 | 2.82 | −1.56 | 0.119 | |
| CB Object RC | −0.43 | 1.52 | −0.29 | 0.775 | |
| TEGI task | −0.21 | 1.28 | −0.17 | 0.868 | |
| Monolingual/bilingual | −2.41 | 1.44 | −1.68 | 0.094 | . |
| Age: TEGI task | 0.50 | 0.29 | 1.76 | 0.079 | . |
| Narr TTR: TEGI task | 23.20 | 13.90 | 1.67 | 0.095 | . |
| Narr VOCD: TEGI task | −0.17 | 0.07 | −2.54 | 0.011 | * |
| Narr MLU: TEGI task | −0.17 | 0.35 | −0.48 | 0.634 | |
| Narr Complexity: TEGI task | −0.35 | 2.22 | −0.16 | 0.876 | |
| Renfrew vocab: TEGI task | −2.67 | 5.40 | −0.49 | 0.622 | |
| CB Active: TEGI task | −1.06 | 2.41 | −0.44 | 0.661 | |
| CB Passive: TEGI task | 3.58 | 2.50 | 1.43 | 0.153 | |
| CB Subject RC: TEGI task | 5.76 | 4.71 | 1.22 | 0.222 | |
| CB Object RC: TEGI task | −0.53 | 1.89 | −0.28 | 0.779 | |
| Age: Monolingual/bilingual | 1.97 | 0.51 | 3.88 | 0.000 | *** |
| Narr TTR: Monolingual/bilingual | −56.35 | 21.27 | −2.65 | 0.008 | ** |
| Narr VOCD: Monolingual/bilingual | 0.14 | 0.11 | 1.35 | 0.177 | |
| Narr MLU: Monolingual/bilingual | −2.45 | 0.68 | −3.63 | 0.000 | *** |
| Narr Complexity: Monolingual/bilingual | 12.23 | 3.25 | 3.76 | 0.000 | *** |
| Renfrew vocab: Monolingual/bilingual | −2.86 | 8.36 | −0.34 | 0.733 | |
| CB Active: Monolingual/bilingual | −1.37 | 3.37 | −0.41 | 0.685 | |
| CB Passive: Monolingual/bilingual | 2.90 | 4.04 | 0.72 | 0.473 | |
| CB Subject RC: Monolingual/bilingual | 5.21 | 5.65 | 0.92 | 0.356 | |
| CB Object RC: Monolingual/bilingual | 4.39 | 3.04 | 1.44 | 0.149 | |
| TEGI task: Monolingual/bilingual | 5.81 | 2.53 | 2.30 | 0.021 | * |
| Age: TEGI task: Monolingual/bilingual | −1.36 | 0.57 | −2.37 | 0.018 | * |
| Narr TTR: TEGI task: Monolingual/bilingual | 85.23 | 27.80 | 3.07 | 0.002 | ** |
| Narr VOCD: TEGI task: Monolingual/bilingual | −0.25 | 0.13 | −1.91 | 0.057 | . |
| Narr MLU: TEGI task: Monolingual/bilingual | 2.35 | 0.71 | 3.33 | 0.001 | *** |
| Narr Complexity: TEGI task: Monolingual/bilingual | −18.92 | 4.43 | −4.27 | 0.000 | *** |
| Renfrew vocab: TEGI task: Monolingual/bilingual | −11.62 | 10.80 | −1.08 | 0.282 | |
| CB Active: TEGI task: Monolingual/bilingual | 4.81 | 4.83 | 1.00 | 0.319 | |
| CB Passive: TEGI task: Monolingual/bilingual | −7.53 | 5.01 | −1.50 | 0.133 | |
| CB Subject RC: TEGI task: Monolingual/bilingual | 4.77 | 9.42 | 0.51 | 0.613 | |
| CB Object RC: TEGI task: Monolingual/bilingual | −2.71 | 3.77 | −0.72 | 0.472 |
Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1.
Table A4.
Full model for syntax; significant effects in bold.
Table A4.
Full model for syntax; significant effects in bold.
| Fixed Effects | β | SE | Z | p | |
|---|---|---|---|---|---|
| Intercept | 3.33 | 1.02 | 3.25 | 0.001 | ** |
| Age | −1.15 | 0.61 | −1.90 | 0.057 | . |
| Narr TTR | −27.05 | 10.20 | −2.65 | 0.008 | ** |
| Narr VOCD | 0.10 | 0.05 | 1.85 | 0.064 | . |
| Narr MLU | −0.39 | 0.26 | −1.52 | 0.129 | |
| Narr Complexity | 3.16 | 2.96 | 1.07 | 0.286 | |
| Renfrew vocab | −25.21 | 7.22 | −3.49 | 0.000 | *** |
| TEGI Past tense | −6.05 | 2.87 | −2.11 | 0.035 | * |
| TEGI 3rd person singular | 15.35 | 7.35 | 2.09 | 0.037 | * |
| Complex | −1.99 | 1.03 | −1.93 | 0.054 | . |
| Structure | 0.07 | 1.21 | 0.06 | 0.956 | |
| Monolingual/bilingual | −2.50 | 2.05 | −1.22 | 0.222 | |
| Age: Complex | 1.32 | 0.62 | 2.12 | 0.034 | * |
| Narr TTR: Complex | 25.01 | 10.48 | 2.39 | 0.017 | * |
| Narr VOCD: Complex | −0.09 | 0.05 | −1.58 | 0.114 | |
| Narr MLU: Complex | 0.44 | 0.27 | 1.65 | 0.099 | . |
| Narr Complexity: Complex | −2.80 | 3.06 | −0.91 | 0.361 | |
| Renfrew vocab: Complex | 20.67 | 7.32 | 2.83 | 0.005 | ** |
| TEGI Past tense: Complex | 2.84 | 3.09 | 0.92 | 0.358 | |
| TEGI 3rd person singular: Complex | −18.17 | 7.41 | −2.45 | 0.014 | * |
| Age: Structure | 0.11 | 0.68 | 0.16 | 0.873 | |
| Narr TTR: Structure | 11.88 | 11.61 | 1.02 | 0.306 | |
| Narr VOCD: Structure | −0.04 | 0.07 | −0.56 | 0.579 | |
| Narr MLU: Structure | 0.10 | 0.32 | 0.31 | 0.755 | |
| Narr Complexity: Structure | −2.48 | 3.52 | −0.71 | 0.480 | |
| Renfrew vocab: Structure | 20.46 | 7.95 | 2.57 | 0.010 | * |
| TEGI Past tense: Structure | 2.11 | 3.45 | 0.61 | 0.540 | |
| TEGI 3rd person singular: Structure | −12.49 | 8.01 | −1.56 | 0.119 | |
| Complex: Structure | −0.33 | 1.33 | −0.25 | 0.806 | |
| Age: Monolingual/bilingual | 2.02 | 1.22 | 1.65 | 0.099 | . |
| Narr TTR: Monolingual/bilingual | 23.22 | 20.17 | 1.15 | 0.250 | |
| Narr VOCD: Monolingual/bilingual | 0.01 | 0.11 | 0.06 | 0.954 | |
| Narr MLU: Monolingual/bilingual | 0.86 | 0.51 | 1.69 | 0.092 | . |
| Narr Complexity: Monolingual/bilingual | −13.81 | 5.85 | −2.36 | 0.018 | * |
| Renfrew vocab: Monolingual/bilingual | 57.26 | 14.23 | 4.02 | 0.000 | *** |
| TEGI Past tense: Monolingual/bilingual | 7.82 | 5.75 | 1.36 | 0.174 | |
| TEGI 3rd person singular: Monolingual/bilingual | −34.85 | 14.69 | −2.37 | 0.018 | * |
| Complex: Monolingual/bilingual | 2.04 | 2.07 | 0.99 | 0.325 | |
| Structure: Monolingual/bilingual | 0.52 | 2.45 | 0.21 | 0.830 | |
| Age: Complex: Structure | −0.71 | 0.73 | −0.97 | 0.331 | |
| Narr TTR: Complex: Structure | −9.65 | 12.78 | −0.76 | 0.450 | |
| Narr VOCD: Complex: Structure | 0.02 | 0.08 | 0.23 | 0.822 | |
| Narr MLU: Complex: Structure | 0.15 | 0.36 | 0.41 | 0.679 | |
| Narr Complexity: Complex: Structure | 4.85 | 3.91 | 1.24 | 0.214 | |
| Renfrew vocab: Complex: Structure | −19.82 | 8.57 | −2.31 | 0.021 | * |
| TEGI Past tense: Complex: Structure | −1.36 | 4.12 | −0.33 | 0.741 | |
| TEGI 3rd person singular: Complex: Structure | 16.99 | 8.36 | 2.03 | 0.042 | * |
| Age: Complex: Monolingual/bilingual | −2.62 | 1.26 | −2.08 | 0.037 | * |
| Narr TTR: Complex: Monolingual/bilingual | −36.05 | 20.33 | −1.77 | 0.076 | . |
| Narr VOCD: Complex: Monolingual/bilingual | 0.01 | 0.11 | 0.09 | 0.926 | |
| Narr MLU: Complex: Monolingual/bilingual | −0.80 | 0.53 | −1.53 | 0.126 | |
| Narr Complexity: Complex: Monolingual/bilingual | 12.00 | 6.02 | 1.99 | 0.046 | * |
| Renfrew vocab: Complex: Monolingual/bilingual | −47.68 | 14.50 | −3.29 | 0.001 | ** |
| TEGI Past tense: Complex: Monolingual/bilingual | −2.44 | 6.16 | −0.40 | 0.692 | |
| TEGI 3rd person singular: Complex: Monolingual/bilingual | 34.38 | 14.77 | 2.33 | 0.020 | * |
| Age: Structure: Monolingual/bilingual | −1.24 | 1.38 | −0.90 | 0.371 | |
| Narr TTR: Structure: Monolingual/bilingual | 1.87 | 22.49 | 0.08 | 0.934 | |
| Narr VOCD: Structure: Monolingual/bilingual | −0.04 | 0.14 | −0.29 | 0.771 | |
| Narr MLU: Structure: Monolingual/bilingual | −0.42 | 0.63 | −0.67 | 0.504 | |
| Narr Complexity: Structure: Monolingual/bilingual | 8.40 | 6.80 | 1.24 | 0.217 | |
| Renfrew vocab: Structure: Monolingual/bilingual | −38.54 | 15.81 | −2.44 | 0.015 | * |
| TEGI Past tense: Structure: Monolingual/bilingual | −4.76 | 6.93 | −0.69 | 0.493 | |
| TEGI 3rd person singular: Structure: Monolingual/bilingual | 25.20 | 15.97 | 1.58 | 0.115 | |
| Complex: Structure: Monolingual/bilingual | −1.24 | 2.67 | −0.47 | 0.642 | |
| Age: Complex: Structure: Monolingual/bilingual | 1.80 | 1.47 | 1.22 | 0.222 | |
| Narr TTR: Complex: Structure: Monolingual/bilingual | 4.01 | 24.25 | 0.17 | 0.869 | |
| Narr VOCD: Complex: Structure: Monolingual/bilingual | 0.03 | 0.16 | 0.21 | 0.835 | |
| Narr MLU: Complex: Structure: Monolingual/bilingual | 0.38 | 0.70 | 0.54 | 0.590 | |
| Narr Complexity: Complex: Structure: Monolingual/bilingual | −12.17 | 7.57 | −1.61 | 0.108 | |
| Renfrew vocab: Complex: Structure: Monolingual/bilingual | 34.64 | 17.02 | 2.04 | 0.042 | * |
| TEGI Past tense: Complex: Structure: Monolingual/bilingual | 0.11 | 8.24 | 0.01 | 0.989 | |
| TEGI Third person singular: Complex: Structure: Monolingual/bilingual | −29.95 | 16.64 | −1.80 | 0.072 | . |
Signif. codes: 0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1.
Note
| 1 | Note that monolinguals were at ceiling on both TEGI tasks, and for all Colouring Book structures other than object relative clauses. The correspondences observed for these structures are therefore not included in Table 14 due to the likelihood that they were driven by outliers. |
References
- Acheson, D. J., & MacDonald, M. C. (2009). Verbal working memory and language production: Common approaches to the serial ordering of verbal information. Psychological Bulletin, 135(1), 50. [Google Scholar] [CrossRef]
- Adani, F. (2011). Rethinking the acquisition of relative clauses in Italian: Towards a grammatically based account. Journal of Child Language, 38(01), 141–165. [Google Scholar] [CrossRef] [PubMed]
- Anaya, J. B., Peña, E. D., & Bedore, L. M. (2018). Conceptual Scoring and Classification Accuracy of Vocabulary Testing in Bilingual Children. Language, Speech, and Hearing Services in Schools, 49(1), 85–97. [Google Scholar] [CrossRef]
- Aoun, J. E., Benmamoun, E., & Choueiri, L. (2009). The syntax of Arabic. Cambridge University Press. [Google Scholar]
- Armon-Lotem, S., de Jong, J., & Meir, N. (2015). Assessing multilingual children: Disentangling bilingualism from language impairment (Vol. 13). Multilingual Matters. [Google Scholar]
- Baker, C. (2011). Foundations of bilingual education and bilingualism. Multilingual Matters. [Google Scholar]
- Bates, D., Maechler, M., Bolker, B., & Walker, S. (2015). Fitting Linear Mixed-Effects Models Using lme4. Journal of Statistical Software, 67(1), 1–48. [Google Scholar] [CrossRef]
- Bedore, L. M., & Peña, E. D. (2008). Assessment of Bilingual Children for Identification of Language Impairment: Current Findings and Implications for Practice. International Journal of Bilingual Education and Bilingualism, 11(1), 1–29. [Google Scholar] [CrossRef]
- Berko, J. (1958). The child’s learning of English morphology. Word, 14(2–3), 150–177. [Google Scholar] [CrossRef]
- Bishop, D. V. (2003). Test for reception of grammar: TROG-2: Manual. Pearson. [Google Scholar]
- Blom, E., & Bosma, E. (2016). The sooner the better? An investigation into the role of age of onset and its relation with transfer and exposure in bilingual Frisian–Dutch children. Journal of Child Language, 43(3), 581–607. [Google Scholar] [CrossRef]
- Blom, E., & Paradis, J. (2015). Sources of individual differences in the acquisition of tense inflection by English second language learners with and without specific language impairment. Applied Psycholinguistics, 36(4), 953–976. [Google Scholar] [CrossRef]
- Boerma, T., Leseman, P., Timmermeister, M., Wijnen, F., & Blom, E. (2016). Narrative abilities of monolingual and bilingual children with and without language impairment: Implications for clinical practice: A narrative as diagnostic tool. International Journal of Language & Communication Disorders, 51(6), 626–638. [Google Scholar] [CrossRef]
- Bonifacci, P., Atti, E., Casamenti, M., Piani, B., Porrelli, M., & Mari, R. (2020). Which Measures Better Discriminate Language Minority Bilingual Children with and Without Developmental Language Disorder? A Study Testing a Combined Protocol of First and Second Language Assessment. Journal of Speech, Language, and Hearing Research, 63(6), 1898–1915. [Google Scholar] [CrossRef] [PubMed]
- Botting, N. (2002). Narrative as a tool for the assessment of linguistic and pragmatic impairments. Child Language Teaching and Therapy, 18(1), 1–21. [Google Scholar] [CrossRef]
- Bowerman, M. (1979). The acquisition of complex sentences. In P. Fletcher, & M. Garman (Eds.), Language acquisition: Studies in first language development (pp. 285–305). Cambridge University Press. [Google Scholar]
- Boyle, W., Lindell, A. K., & Kidd, E. (2013). Investigating the Role of Verbal Working Memory in Young Children’s Sentence Comprehension: Verbal WM and Sentence Comprehension in Children. Language Learning, 63(2), 211–242. [Google Scholar] [CrossRef]
- Bulté, B., & Housen, A. (2012). Defining and operationalising L2 complexity. Dimensions of L2 Performance and Proficiency: Complexity, Accuracy and Fluency in SLA, 32, 21. [Google Scholar]
- Caplan, D., & Waters, G. (2002). Working memory and connectionist models of parsing: A reply to MacDonald and Christiansen (2002). Psychological Review, 109(1), 66–74. [Google Scholar] [CrossRef]
- Carr, L., & Johnston, J. (2001). Morphological cues to verb meaning. Applied Psycholinguistics, 22(4), 601–618. [Google Scholar] [CrossRef]
- Carrow-Woolfolk, E. (1999). CASL: Comprehensive assessment of spoken language. American Guidance Services. Available online: https://www.proedinc.com/Products/14564/casl2-comprehensive-assessment-of-spoken-languagesecond-edition.aspx?bCategory=OLA!COMP (accessed on 20 March 2024).
- Cheung, H. (1996). Aspects of working memory in L2 learning. Developmental Psychology, 32, 867–873. [Google Scholar] [CrossRef]
- Cheung, S., Kan, P. F., Winicour, E., & Yang, J. (2019). Effects of home language input on the vocabulary knowledge of sequential bilingual children. Bilingualism: Language and Cognition, 22(5), 986–1004. [Google Scholar] [CrossRef]
- Chiat, S., Armon-Lotem, S., Marinis, T., Polisenska, K., Roy, P., & Seeff-Gabriel, B. (2013). The potential of sentence imitation tasks for assessment of language abilities in sequential bilingual children. In V. Mueller-Gathercole (Ed.), Issues in the Assessment of Bilinguals (pp. 56–89). Multilingual Matters. [Google Scholar]
- Chondrogianni, V., & Marinis, T. (2011). Differential effects of internal and external factors on the development of vocabulary, tense morphology and morpho-syntax in successive bilingual children. Linguistic Approaches to Bilingualism, 1(3), 318–345. [Google Scholar] [CrossRef]
- Coughler, C., Bardell, T., Schouten, M. A., Smith, K., & Archibald, L. M. D. (2023). Narrative Retell Assessment Using “Frog” Stories: A Practice-Based Research Speech-Language Pathology Partnership Exploring Story Equivalency. Language, Speech, and Hearing Services in Schools, 54(4), 1249–1266. [Google Scholar] [CrossRef] [PubMed]
- Cowley, J., & Glasgow, C. (1994). The renfrew bus story. The Centreville School. [Google Scholar]
- Dailey, S., & Bergelson, E. (2022). Language input to infants of different socioeconomic statuses: A quantitative meta-analysis. Developmental Science, 25(3), e13192. [Google Scholar] [CrossRef]
- de Cat, C., & Unsworth, S. (2023). So many variables, but what causes what? Journal of Child Language, 50(4), 832–836. [Google Scholar] [CrossRef] [PubMed]
- De Lamo White, C., & Jin, L. (2011). Evaluation of speech and language assessment approaches with bilingual children. International Journal of Language & Communication Disorders, 46(6), 613–627. [Google Scholar] [CrossRef]
- Devescovi, A., & Caselli, M. C. (2007). Sentence repetition as a measure of early grammatical development in Italian. International Journal of Language & Communication Disorders, 42(2), 187–208. [Google Scholar] [CrossRef]
- Diessel, H. (2004). The acquisition of complex sentences (Vol. 105). Cambridge University Press. [Google Scholar]
- Dunn, L. M., & Dunn, D. M. (2007). PPVT-4: Peabody picture vocabulary test. Pearson Assessments. [Google Scholar]
- Dunn, L. M., Dunn, L. M., Whetton, C., & Burley, J. (1997). The British picture vocabulary scale (2nd ed.). NFER-Nelson. [Google Scholar]
- Ebert, K. D., & Pham, G. (2017). Synthesizing Information from Language Samples and Standardized Tests in School-Age Bilingual Assessment. Language, Speech, and Hearing Services in Schools, 48(1), 42–55. [Google Scholar] [CrossRef] [PubMed]
- Ebert, K. D., & Scott, C. M. (2014). Relationships Between Narrative Language Samples and Norm-Referenced Test Scores in Language Assessments of School-Age Children. Language, Speech, and Hearing Services in Schools, 45(4), 337–350. [Google Scholar] [CrossRef] [PubMed]
- Figueroa, M. (2024). Decolonizing (Psycho)linguistics Means Dropping the Language Gap Rhetoric. In A. H. Charity Hudley, C. Mallinson, & M. Bucholtz (Eds.), Decolonizing linguistics (pp. 157–174). Oxford University Press. [Google Scholar] [CrossRef]
- Friedmann, N., Belletti, A., & Rizzi, L. (2009). Relativized relatives: Types of intervention in the acquisition of A-bar dependencies. Lingua, 119(1), 67–88. [Google Scholar] [CrossRef]
- Frizelle, P., Thompson, P., Duta, M., & Bishop, D. V. M. (2019). Assessing Children’s Understanding of Complex Syntax: A Comparison of Two Methods. Language Learning, 69(2), 255–291. [Google Scholar] [CrossRef]
- Gerard, J. (2022). The extragrammaticality of the acquisition of adjunct control. Language Acquisition, 29(2), 107–134. [Google Scholar] [CrossRef]
- Gerard, J., Lidz, J., Zuckerman, S., & Pinto, M. (2018). The acquisition of adjunct control is colored by the task. Glossa: A Journal of General Linguistics, 3(1), 75. [Google Scholar] [CrossRef]
- Gillam, R. B., & Pearson, N. A. (2004). Test of narrative language. Pro-ed. [Google Scholar]
- Gillon, G., McNeill, B., Scott, A., Gath, M., & Westerveld, M. (2023). Retelling stories: The validity of an online oral narrative task. Child Language Teaching and Therapy, 39(2), 150–174. [Google Scholar] [CrossRef]
- Golinkoff, R. M., Hoff, E., Rowe, M. L., Tamis-LeMonda, C. S., & Hirsh-Pasek, K. (2019). Language Matters: Denying the Existence of the 30-Million-Word Gap Has Serious Consequences. Child Development, 90(3), 985–992. [Google Scholar] [CrossRef]
- Gordon, P. C., Hendrick, R., & Johnson, M. (2001). Memory interference during language processing. Journal of Experimental Psychology: Learning, Memory, and Cognition, 27(6), 1411. [Google Scholar] [PubMed]
- Granger, S. (2011). How to use foreign and second language learner corpora. In A. Mackey, & S. M. Gass (Eds.), Research methods in second language acquisition: A practical guide (pp. 5–29). Wiley Online Library. [Google Scholar]
- Grosjean, F. (1989). Neurolinguists, beware! The bilingual is not two monolinguals in one person. Brain and Language, 36, 3–15. [Google Scholar] [CrossRef] [PubMed]
- Guo, L.-Y., Eisenberg, S., Schneider, P., & Spencer, L. (2019). Percent Grammatical Utterances Between 4 and 9 Years of Age for the Edmonton Narrative Norms Instrument: Reference Data and Psychometric Properties. American Journal of Speech-Language Pathology, 28(4), 1448–1462. [Google Scholar] [CrossRef]
- Hadjadj, O., Kehoe, M., & Delage, H. (2022). Dynamic Assessment Identifies Morphosyntactic Deficits in Mono-and Bilingual Children with Developmental Language Disorder. Languages, 7(4), 295. [Google Scholar] [CrossRef]
- Hart, B., & Risley, T. R. (1995). Meaningful differences in the everyday experience of young American children. Paul H Brookes Publishing. [Google Scholar]
- Henderson, D. E., Restrepo, M. A., & Aiken, L. S. (2018). Dynamic Assessment of Narratives Among Navajo Preschoolers. Journal of Speech, Language, and Hearing Research, 61(10), 2547–2560. [Google Scholar] [CrossRef]
- Huang, Y. T., Zheng, X., Meng, X., & Snedeker, J. (2013). Children’s assignment of grammatical roles in the online processing of Mandarin passive sentences. Journal of Memory and Language, 69(4), 589–606. [Google Scholar] [CrossRef] [PubMed]
- Just, M. A., & Carpenter, P. A. (1992). A capacity theory of comprehension: Individual differences in working memory. Psychological Review, 99(1), 122. [Google Scholar] [CrossRef]
- Karem, R. W., & Washington, K. N. (2021). The Cultural and Diagnostic Appropriateness of Standardized Assessments for Dual Language Learners: A Focus on Jamaican Preschoolers. Language, Speech, and Hearing Services in Schools, 52(3), 807–826. [Google Scholar] [CrossRef]
- Kuchirko, Y., & Nayfeld, I. (2020). Language Gap: Cultural Assumptions and Ideologies. In C.-A. Huertas-Abril, & M. E. Gómez-Parra (Eds.), Advances in Linguistics and communication studies (pp. 32–53). IGI Global. [Google Scholar] [CrossRef]
- Laing, S. P., & Kamhi, A. (2003). Alternative Assessment of Language and Literacy in Culturally and Linguistically Diverse Populations. Language, Speech, and Hearing Services in Schools, 34(1), 44–55. [Google Scholar] [CrossRef] [PubMed]
- Leitão, S., & Allan, L. (2003). Peter and the cat: Narrative assessment. Black Sheep Press. [Google Scholar]
- Lenth, R. V., Buerkner, P., Giné-Vázquez, I., Herve, M., Jung, M., Love, J., Miguez, F., Riebl, H., & Singmann, H. (2022). Emmeans: Estimated marginal means, aka least-squares means. (Version 1.8.2) [Computer software]. Available online: https://CRAN.R-project.org/package=emmeans (accessed on 2 December 2022).
- Leonard, L. B. (2009). Is Expressive Language Disorder an Accurate Diagnostic Category? American Journal of Speech-Language Pathology, 18(2), 115–123. [Google Scholar] [CrossRef]
- Lewis, R. L., & Vasishth, S. (2005). An activation-based model of sentence processing as skilled memory retrieval. Cognitive Science, 29(3), 375–419. [Google Scholar] [CrossRef]
- Lust, B., Foley, C., & Dye, C. D. (2009). The first language acquisition of complex sentences. In E. L. Bavin (Ed.), The Cambridge Handbook of Child Language. Cambridge Handbooks in Language and Linguistics (pp. 237–258). Cambridge University Press. [Google Scholar]
- MacWhinney, B. (2000). The CHILDES project: The database (Vol. 2). Psychology Press. [Google Scholar]
- Malvern, D., Richards, B., Chipere, N., & Durán, P. (2004). Lexical diversity and language development. Palgrave Macmillan UK. [Google Scholar] [CrossRef]
- Marinis, T. (2010). Using on-line processing methods in language acquisition research. In E. Blom, & S. Unsworth (Eds.), Experimental methods in language acquisition research (pp. 139–162). John Benjamins Publishing Company. [Google Scholar] [CrossRef]
- Mayer, M. (2003). Frog, where are you? Penguin. [Google Scholar]
- Murphy, K. A., Springle, A. P., Sultani, M. J., McIlraith, A., & Language and Reading Research Consortium (LARRC). (2022). Predicting Language Performance from Narrative Language Samples. Journal of Speech, Language, and Hearing Research, 65(2), 775–784. [Google Scholar] [CrossRef] [PubMed]
- Nippold, M. A. (2016). Later language development: School-age Children, adolescents, and young adults. PRO-ED. [Google Scholar]
- Norris, J. M., & Ortega, L. (2009). Towards an organic approach to investigating CAF in instructed SLA: The case of complexity. Applied Linguistics, 30(4), 555–578. [Google Scholar] [CrossRef]
- O’Toole, C. (2022). Developing an assessment of vocabulary proficiency for Irish-English bilingual children: The Irish crosslinguistic lexical tasks. Journal of Multilingual and Multicultural Development, 45(6), 2023–2038. [Google Scholar] [CrossRef]
- Ouhalla, J. (1999). Focus and Arabic clefts. In The Grammar of Focus (pp. 335–359). John Benjamins. [Google Scholar] [CrossRef]
- Owens, R. E., & Pavelko, S. L. (2017). Relationships among conversational language samples and norm-referenced test scores. Clinical Archives of Communication Disorders, 2(1), 43–50. [Google Scholar] [CrossRef]
- Paradis, J. (2023). Sources of individual differences in the dual language development of heritage bilinguals. Journal of Child Language, 50(4), 793–817. [Google Scholar] [CrossRef] [PubMed]
- Paradis, J., & Blom, E. (2016). Do early successive bilinguals show the English L2 pattern of precocious BE acquisition? Bilingualism: Language and Cognition, 19(3), 630–635. [Google Scholar]
- Peterson, C. (1990). The who, when and where of early narratives. Journal of Child Language, 17(2), 433–455. [Google Scholar] [CrossRef] [PubMed]
- Pinto, M., & Zuckerman, S. (2019). Coloring Book: A new method for testing language comprehension. Behavior Research Methods, 51(6), 2609–2628. [Google Scholar] [CrossRef] [PubMed]
- Polišenská, K., Chiat, S., & Roy, P. (2015). Sentence repetition: What does the task measure? International Journal of Language & Communication Disorders, 50(1), 106–118. [Google Scholar] [CrossRef]
- Potter, M. C., & Lombardi, L. (1990). Regeneration in the short-term recall of sentences. Journal of Memory and Language, 29(6), 633–654. [Google Scholar] [CrossRef]
- R Core Team. (2015). R: A language and environment for statistical computing. [Computer software]. R Foundation for Statistical Computing.
- Read, J. (2000). Assessing vocabulary. Cambridge University Press. [Google Scholar]
- Read, J. (2007). Second language vocabulary assessment: Current practices and new directions. International Journal of English Studies, 7(2), 105–126. [Google Scholar]
- Renfrew, C. E. (1995). Word finding vocabulary test. Winslow. [Google Scholar]
- Renfrew, C. E. (2023). Expressive vocabulary test. Taylor & Francis Ltd. [Google Scholar]
- Rice, M. L., & Wexler, K. (2001). Test of Early Grammatical Impairment (TEGI). Pearson Publishing. [Google Scholar]
- Richards, B. (1987). Type/token ratios: What do they really tell us? Journal of Child Language, 14(2), 201–209. [Google Scholar] [CrossRef]
- Rocca, S. (2007). Child second language acquisition: A bi-directional study of English and Italian tense-aspect morphology (Vol. 35). John Benjamins Publishing. [Google Scholar]
- Roesch, A. D., & Chondrogianni, V. (2016). ‘Which mouse kissed the frog?’ Effects of age of onset, length of exposure, and knowledge of case marking on the comprehension of wh-questions in German-speaking simultaneous and early sequential bilingual children. Journal of Child Language, 43(3), 635–661. [Google Scholar] [CrossRef] [PubMed]
- Romeo, R. R., Leonard, J. A., Robinson, S. T., West, M. R., Mackey, A. P., Rowe, M. L., & Gabrieli, J. D. E. (2018). Beyond the 30-Million-Word Gap: Children’s Conversational Exposure Is Associated with Language-Related Brain Function. Psychological Science, 29(5), 700–710. [Google Scholar] [CrossRef] [PubMed]
- Rothman, J., Bayram, F., DeLuca, V., Di Pisa, G., Dunabeitia, J. A., Gharibi, K., Hao, J., Kolb, N., Kubota, M., & Kupisch, T. (2023). Monolingual comparative normativity in bilingualism research is out of “control”: Arguments and alternatives. Applied Psycholinguistics, 44(3), 316–329. [Google Scholar] [CrossRef]
- Schneider, P., Hayward, D., & Dubé, R. V. (2006). Storytelling from pictures using the Edmonton narrative norms instrument. Journal of Speech Language Pathology and Audiology, 30(4), 224. [Google Scholar]
- Seymour, H., Roeper, T., de Villiers, J., & de Villiers, P. A. (2005). The DELV-NR.(norm-referenced version) the diagnostic evaluation of language variation. The Psychological Corporation. [Google Scholar]
- Shormani, M. (2017). A person approach to personal passive in standard Arabic. Studia Linguistica. Available online: https://www.vitaminedz.com/articlesfiche/7254/7254368.pdf (accessed on 20 March 2024).
- Silva, M., & Cain, K. (2019). The use of questions to scaffold narrative coherence and cohesion. Journal of Research in Reading, 42(1), 1–17. [Google Scholar] [CrossRef]
- Snow, C. E., & Hoefnagel-Hohle, M. (1978). The critical period for language acquisition: Evidence from second language learning. Child Development, 49(4), 1114–1128. [Google Scholar] [CrossRef]
- Snyder, W., & Hyams, N. (2015). Minimality effects in children’s passives. In E. Di Domenico, C. Hamann, & S. Matteini (Eds.), Structures, Strategies and Beyond: Essays in Honour of Adriana Belletti (Linguistik Aktuell/Linguistics Today, 223) (pp. 343–368). John Benjamins. [Google Scholar]
- Sperry, D. E., Sperry, L. L., & Miller, P. J. (2019). Re-examining the Verbal Environments of Children from Different Socioeconomic Backgrounds. Child Development, 90(4), 1303–1318. [Google Scholar] [CrossRef] [PubMed]
- Tamis-LeMonda, C. S., Kachergis, G., Masek, L. R., Gonzalez, S. L., Soska, K. C., Herzberg, O., Xu, M., Adolph, K. E., Gilmore, R. O., Bornstein, M. H., Casasola, M., Fausey, C. M., Frank, M. C., Goldin-Meadow, S., Gros-Louis, J., Hirsh-Pasek, K., Iverson, J., Lew-Williams, C., MacWhinney, B., … Yurovsky, D. (2024). Comparing apples to manzanas and oranges to naranjas: A new measure of English-Spanish vocabulary for dual language learners. Infancy, 29(3), 302–326. [Google Scholar] [CrossRef]
- Theodorou, E., Kambanaros, M., & Grohmann, K. K. (2017). Sentence repetition as a tool for screening morphosyntactic abilities of bilectal children with SLI. Frontiers in Psychology, 8, 2104. [Google Scholar] [CrossRef] [PubMed]
- Tsimpli, I. M., & Dimitrakopoulou, M. (2007). The Interpretability Hypothesis: Evidence from wh-interrogatives in second language acquisition. Second Language Research, 23(2), 215–242. [Google Scholar] [CrossRef]
- Ukrainetz, T. A., & Blomquist, C. (2002). The criterion validity of four vocabulary tests compared with a language sample. Child Language Teaching and Therapy, 18(1), 59–78. [Google Scholar] [CrossRef]
- Unsworth, S. (2016). Early child L2 acquisition: Age or input effects? Neither, or both? Journal of Child Language, 43(3), 608–634. [Google Scholar] [CrossRef] [PubMed]
- Valian, V. (2016). When Children Don’t Say What They Know. In D. Barner, & A. S. Baron (Eds.), Core knowledge and conceptual change (pp. 261–276). Oxford University Press. [Google Scholar] [CrossRef]
- Van Dyke, J. A., & Lewis, R. L. (2003). Distinguishing effects of structure and decay on attachment and repair: A cue-based parsing account of recovery from misanalyzed ambiguities. Journal of Memory and Language, 49(3), 285–316. [Google Scholar] [CrossRef]
- Van Hell, J. G., Bosman, A. M. T., Wiggers, I., & Stoit, J. (2003). Children’s cultural background knowledge and story telling performance. International Journal of Bilingualism, 7(3), 283–303. [Google Scholar] [CrossRef]
- Vinther, T. (2002). Elicited imitation: A brief overview. International Journal of Applied Linguistics, 12(1), 54–73. [Google Scholar] [CrossRef]
- Wiig, E. H., Secord, W. A., & Semel, E. (2017). Clinical evaluation of language fundamentals: Australian and New Zealand standardised edition (CELF-5 A&NZ). Pearson Clinical Assessment. [Google Scholar]
- Williams, K. T. (2019). Expressive Vocabulary Test (3rd ed.). NCS Pearson. [Google Scholar]
- Wood, C., Hoge, R., Schatschneider, C., & Castilla-Earls, A. (2021). Predictors of item accuracy on the Test de Vocabulario en Imagenes Peabody for Spanish-English speaking children in the United States. International Journal of Bilingual Education and Bilingualism, 24(8), 1178–1192. [Google Scholar] [CrossRef] [PubMed]
- Zuckerman, S., Pinto, M., Koutamanis, E., & Spijk, Y. (2016). A New Method for Testing Language Comprehension Reveals Better Performance on Passive and Principle B Constructions. In J. Scott, & D. Waughtal (Eds.), BUCLD 40: Proceedings of the 40th annual Boston University Conference on Language Development (pp. 443–456). Cascadilla Press. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
