1. Introduction
Early literacy skills are among the most critical skills that determine a child’s educational achievement and overall quality of life. These skills represent a multidimensional network of interrelated cognitive competencies, ranging from basic letter recognition to phonological processing, rapid naming, print awareness, and oral language proficiency (
Ne’eman and Shaul 2023). Early literacy develops through a dynamic process that encompasses all the interactions and awareness children have with written and spoken language in their environment before they receive formal reading and writing instruction (
Bonifacci et al. 2022). Subcomponents such as phonological awareness, alphabet knowledge, print awareness, and vocabulary acquired during this period serve as a strong prerequisite for the development of fundamental literacy skills that begin in elementary school (
Pham et al. 2025).
Early literacy skills encompass the knowledge, skills, and attitudes that children must acquire before reading and writing, including critical components such as phonological awareness (recognizing, discriminating, and manipulating sounds), letter knowledge, vocabulary, and print awareness (understanding how books and print function) (
Derby et al. 2020;
Ergin et al. 2025). Acquiring these skills during early childhood has significant and long-term effects on an individual’s cognitive, socio-emotional, and academic development. In particular, phonological awareness has been found to directly enhance reading success, expand vocabulary through reading activities, strengthen grammatical skills, and improve oral expression (
De Abreu et al. 2020). Moreover, reading experiences support higher-order cognitive processes such as attention, memory, problem-solving, and planning, thereby contributing to the development of executive functions (
Blair and Raver 2014;
Ne’eman and Shaul 2023). Children who develop literacy skills at an early age demonstrate greater success in reading and comprehension skills in later years (
Inoue et al. 2023) and facilitate learning in other academic areas by enhancing their overall school achievement (
Pham et al. 2025). Research has also shown that children with this strong foundation are more likely to succeed in school, pursue further education, and are less likely to drop out (
Hernandez 2011).
Neuroscientific research demonstrates that early literacy experiences have positive effects on brain development and strengthen the neural connections among regions associated with reading, enabling these skills to become faster and more automatic (
Benischek et al. 2020;
Mateu-Estivill et al. 2020). Strong early literacy skills positively influence an individual’s academic success not only in elementary school but also in middle and high school, as well as their access to better educational opportunities and upward social mobility later in life (
Shahaeian et al. 2018). Moreover, sufficient reading skills contribute to children’s social and emotional development by enhancing their self-confidence and fostering a positive attitude toward learning; children with strong reading and social skills exhibit better academic performance (
Carpendale et al. 2025;
Cooper et al. 2014). Conversely, children with reading difficulties are more likely to experience attention deficits, poor self-regulation, and problems with social adjustment resulting from their learning difficulties (
Montroy et al. 2014). The lack of reading skills is strongly associated with academic failure and can negatively affect children’s self-esteem, sense of self-efficacy, and overall psychological well-being, leading to feelings of inadequacy (
Humphrey 2003;
Ne’eman and Shaul 2023). Such individuals have been found to display lower academic achievement and reduced workforce participation rates in adulthood (
James et al. 2024). Therefore, assessing these skills with valid and reliable methods at the beginning of schooling is a crucial step toward identifying at-risk children and developing evidence-based preventive interventions.
Assessing early literacy skills goes beyond functioning as a screening tool and serves a strategic role in promoting educational equity. Research indicates that environmental factors such as socioeconomic conditions shape the neurocognitive systems underlying literacy development from an early age (
Noble et al. 2021). This highlights the need for assessment instruments that are standardized across all children yet sensitive to cultural and linguistic contexts. A variety of assessment tools are available internationally for evaluating the early literacy skills of preschool children. These tools generally target fundamental components such as phonological awareness, letter knowledge, vocabulary, and reading comprehension. For instance, the STAR Early Literacy Scale comprehensively measures these skills through a computer-adaptive system (
Renaissance Institute 2001), whereas the Preschool Early Literacy Indicators (PELIs), developed by
Kaminski et al. (
2014), assess vocabulary and comprehension skills alongside subdimensions of early literacy, including alphabet and print awareness. Similarly,
Barringer (
2009) evaluated pre-reading skills such as phonological awareness and grammar using the Early Literacy Knowledge and Skills (ELKS) scale. More specifically,
McBride’s (
2015) Observation Survey of Early Literacy Achievement scale aims to identify what children already know and what they need to learn at the beginning of the reading and writing acquisition process.
The pace and nature of reading and writing acquisition are significantly influenced by the orthographic depth of the language being learned (
Ellis et al. 2004). The transparent orthographic structure of Turkish, characterized by high letter–sound consistency, enables children to acquire decoding skills more rapidly than in languages with opaque orthographies, such as English. This structural difference implies that direct translation or adaptation of tests developed for other languages cannot accurately capture the unique developmental profiles and skill hierarchies of children in Türkiye. Therefore, developing measurement instruments that are specific to Türkiye’s linguistic structure, psychometrically robust, and based on nationally normative data is a critical need. Indeed, although limited, several studies have been conducted in this area. The Early Literacy Skills Assessment Instrument (EOBDA), developed by
Karaman and Aytar (
2016), assesses phonological awareness, print awareness, story comprehension, visual matching, and prewriting skills among children aged 48–77 months. Similarly,
Güven and Topbaş (
2014) adapted the Test of Early Language Development (TELD-3) for children in Türkiye.
While a limited number of instruments exist in Türkiye to assess children’s reading readiness within the field of early literacy, these tools remain inadequate due to two fundamental limitations. First, most of them focus on specific pre-reading subskills and fail to comprehensively measure the current proficiency levels of students who have already learned to read and write by the time they enter elementary school—that is, their pre-literacy skills. It is noteworthy that some children in Türkiye begin first grade already knowing how to read and write. However, current random placement practices and standardized phonics instruction often lead to boredom, loss of motivation, and slower progress among these advanced students, thereby reducing the overall effectiveness of instruction. The second and more critical limitation is that many existing local instruments do not meet modern psychometric standards (e.g., testing measurement invariance through multigroup confirmatory factor analysis [MGCFA]). This underscores the need for an instrument that assesses various dimensions of writing skills—particularly those involving both motor and cognitive processes, such as copying and dictation—within a holistic model and rigorously examines their construct validity. In this context, developing a valid and reliable Pre-Literacy Test (PLT) is essential for accurately identifying students’ initial literacy levels upon entering first grade and for providing appropriate placement and individualized instruction, thereby promoting educational equity and supporting individual development. Based on this need, the present study aims to develop the PLT and to comprehensively examine its psychometric properties in order to assess the pre-literacy skills of children transitioning from preschool to elementary school.
The primary objective of this study is to address the critical shortage of literacy assessment tools in Türkiye while challenging the predominantly English-centric paradigm in the international literature. Developed within the context of Turkish transparent orthography, the PLT serves as a pivotal testing ground for the orthographic depth hypothesis and universal theories of literacy acquisition. By providing comparative data from a transparent script, the study aims to offer significant, cross-linguistic evidence that broadens the scope of existing literacy research.
Furthermore, the PLT advances theoretical understanding by operationalizing early writing as distinct cognitive processes—specifically copying and dictation—rather than a unitary construct, offering empirical support for frameworks such as the Not-So-Simple View of Writing and DIEW. Practically, the instrument moves beyond global tools that traditionally focus on precursor skills; instead, it directly measures the holistic proficiencies children demonstrate at school entry. This approach fills a critical gap in the field, enabling educators to look beyond mere readiness to identify specific student needs and implement targeted instruction.
2. Materials and Methods
This study employed a quantitative design aimed at developing a measurement instrument to assess the pre-literacy skills of children who have completed preschool and are about to enter first grade. The methodological framework of the study was based on a multistage and comprehensive hierarchy of construct validity and reliability analyses, designed to ensure the psychometric soundness of the scale.
2.1. Participants
The study group consisted of children aged 6–7 who were attending preschool preparatory classes and transitioning to first grade in elementary school during the 2024–2025 academic year. Convenience sampling was employed, prioritizing accessibility and suitability for the study’s purpose (
Fraenkel Jack et al. 2012). The study was conducted in three subgroups based on the type and stage of analyses. The first study group included 1911 children (49.8% female, 50.2% male); the second comprised 1644 children (47.3% female, 52.7% male); and the third consisted of 2411 children (48.6% female, 51.4% male). Most participants were from families identified as having middle to upper socioeconomic status.
2.2. Data Analysis
Data were analyzed using Jamovi 2.6.44. The analyses were conducted in three stages. In the first stage, an Exploratory Factor Analysis (EFA) was performed on data from 1911 participants; in the second stage, a Confirmatory Factor Analysis (CFA) was conducted on data from 1644 participants; and in the third stage, item analyses were carried out on data from 2411 participants. Importantly, each analysis was conducted on a different sample. Using separate samples for EFA, CFA, and item analyses is a methodological requirement to prevent overfitting, ensure the independence of analyses, and increase the generalizability and robustness of the psychometric findings. Therefore, the data for each analytical stage were collected from distinct participant groups.
Before conducting the EFA, the suitability of the data for factor analysis was examined using Bartlett’s Test of Sphericity and the Kaiser–Meyer–Olkin (KMO) Measure Of Sampling Adequacy. After confirming that the sample met the necessary assumptions, an EFA was performed to determine the construct validity of the scale. Principal Axis Factoring (PAF) was employed as the factor extraction method. Assuming intercorrelations among the factors, an oblique Oblimin rotation was applied (
Tabachnick and Fidell 2013). To determine the optimal number of factors to be retained, a Parallel Analysis was conducted, as it provides more reliable results (
Horn 1965).
CFA were conducted using the Diagonal Weighted Least Squares (DWLS) estimation method, which is appropriate for ordinal data, along with robust standard errors. Because the χ
2 test tends to be significant in large samples, the evaluation of model fit primarily focused on comparative fit indices: CFI and TLI (≥0.95 indicating excellent fit) and SRMR and RMSEA (≤0.08 indicating good fit) (
Hu and Bentler 1999). To demonstrate the structural integrity of the model, convergent validity was tested using item factor loadings (β), Composite Reliability (CR), and Average Variance Extracted (AVE), whereas discriminant validity was assessed using the Fornell–Larcker criterion (
Fornell and Larcker 1981). Furthermore, to determine whether the test measured the same construct in the same way across groups, measurement invariance was evaluated across gender using (MGCFA). Following a hierarchical procedure consistent with the ordinal data structure, three sequential models—configural, metric, and scalar—were compared. Establishing scalar invariance ensured the reliability of factor mean comparisons across groups. A ΔCFI value of ≤0.01 was used as the criterion for evaluating model fit between successive models (
Cheung and Rensvold 2002).
For item analyses, item difficulty indices were calculated, and item discrimination was evaluated using point-biserial correlation coefficients within the framework of Classical Test Theory. Item difficulty is interpreted as “difficult” as it approaches 0 and “easy” as it approaches 1. Items with high discriminatory power are expected to have point-biserial correlation values above 0.30 (
Ebel and Frisbie 1991). Finally, to determine the reliability of the test, Cronbach’s alpha, McDonald’s Omega, Kuder–Richardson 20 (KR-20) coefficients, and split-half reliability values were calculated.
2.3. Data Collection Tools
Demographic Information Form. This form was developed to collect information on the participants’ age, gender, and socioeconomic status.
Pre-Literacy Test (PLT). The PLT was developed by a researcher with extensive academic experience—both theoretical and practical—in the field of reading and writing, taking into account the relevant body of literature. The PLT is designed to measure the basic literacy skills that children acquire prior to school entry, whether through formal or informal learning experiences. It consists of two main sections: reading and writing. The reading section comprises five domains that assess children’s phonics, letter knowledge, spelling, word reading, and simple text reading, including a total of six items. The writing section similarly covers five domains with seven items designed to evaluate letter/sound writing, syllable writing, word writing, and simple sentence writing skills (both copying and dictation).
Specific indicators and items (e.g., recognizing or writing specific letters and sounds; reading or writing particular syllables and words; reading or writing simple sentences) were developed to assess student performance in each domain of the PLT. During the test development process, relevant literature on early literacy, literacy development, and educational measurement was reviewed, expert opinions were obtained, and pilot studies were conducted. An Administration Guide was prepared to support the implementation, scoring, and interpretation of the PLT. This guide outlines the expected basic skills, scoring criteria, and sample student responses for each domain.
The PLT consists of three subtests—Reading Skills, Writing Skills (Dictation), and Writing Skills (Copying)—each scored on a 0–4 scale, where 0 = unacceptable, 1 = inadequate, 2 = moderate, 3 = good, and 4 = excellent. The Reading Skills subtest includes six items assessing letter, closed syllable, open syllable, word, sentence, and text reading. For example, in the word-reading item, children are presented with a page containing various words and are instructed to “read the indicated word.” A score of 4 is assigned when the child reads the target word automatically and fluently; a score of 3 when the child reads the word with slight hesitation; a score of 2 when the reading is syllabic or fragmented; a score of 1 when the child attempts to read by sounding out individual letters; and a score of 0 when the child recognizes some letters but cannot read the word either as a whole or in parts. Similarly, in the Writing Skills (Dictation) subtest, children write letters, syllables, words, or sentences dictated by the examiner, while in the Writing Skills (Copying) subtest, they copy visually presented items of the same types. Each response is scored on the same 0–4 scale based on accuracy, completeness, and correctness. For each subtest, the total score is divided by the number of items to obtain a final score between 0 and 4, which is interpreted using the same performance categories (0 = unacceptable, 1 = inadequate, 2 = moderate, 3 = good, 4 = excellent).
2.4. Data Collection Process
Ethical approval for the study was obtained from Istanbul University-Cerrahpaşa. Data were collected by classroom teachers working in elementary schools. Prior to data collection, the researcher provided a one-day training program for the teachers, which introduced the instrument and covered the procedures for administration and scoring. In addition, sample application and scoring videos were prepared and shared with the participating teachers.
Data were collected in June and July of 2025, when the children were completing the preschool preparation program and transitioning to first grade. The PLT was administered by classroom teachers in accordance with the procedures outlined in the Administration Guide. Teachers evaluated each child’s reading and writing performance based on their responses to the test items and recorded the scores on the administration form.
4. Discussion
This study comprehensively examined the psychometric properties of the PLT, which was developed to assess the pre-literacy skills of children transitioning from preschool to elementary school. The key findings provide strong evidence that the PLT is a psychometrically robust, valid, and reliable instrument that not only meets local educational needs but also makes a significant contribution to the international early literacy literature.
One of the most striking findings of the study was the three-factor structure revealed by the EFA and CFA: Reading Skills, Writing Skills (Dictation), and Writing Skills (Copying). This structure supports the general notion that reading and writing are highly correlated yet separable constructs (
Kim et al. 2023). Although reading and writing skills are often linked within models of literacy development, recent research suggests that these skills can be differentiated and are associated with at least partially distinct cognitive determinants (
Sigmund et al. 2024). The present study’s findings are consistent with this view. Moreover, the current study empirically demonstrated that early writing is not a unidimensional construct. The distinction between dictation and copying as separate dimensions of early writing is increasingly supported by empirical research across diverse orthographies. Recent studies highlight that copying and dictation engage distinct cognitive and motor processes, reinforcing the multidimensional nature of early writing development. Copying tasks primarily depend on transcription-related competencies such as visual perception, fine motor coordination, and visual–motor integration, as demonstrated in both alphabetic and logographic systems (
Puranik and Al Otaiba 2012). For example, in Arabic and Chinese, graphomotor and fine-motor skills significantly influence copying performance, while orthographic knowledge is more critical for dictation accuracy (
Salameh-Matar et al. 2024;
Ye et al. 2021a,
2021b). Dictation, in contrast, requires more complex cognitive processes, including phoneme–grapheme mapping, orthographic knowledge, and auditory short-term memory (
Berninger and Winn 2006). Studies in transparent orthographies like Spanish and Italian show that dictation tasks are more sensitive to orthographic complexity and phonological processing demands than copying tasks, and children tend to make more errors in dictation than in copying. This pattern is consistent across languages, suggesting that the cognitive load of dictation is higher due to the need for sound–letter mapping and memory retrieval (
Afonso et al. 2020;
Arfé et al. 2019).
Theoretical models such as the “Not-So-Simple View of Writing” and the Direct and Indirect Effects Model of Writing (DIEW) emphasize the interplay between lower-order transcription skills (copying, handwriting) and higher-order cognitive–linguistic skills (spelling, phonological awareness) in writing development (
Barnett et al. 2019;
Kim and Schatschneider 2017). Empirical evidence from cross-linguistic studies further demonstrates that orthographic transparency modulates the relative contribution of these skills, with transparent orthographies placing greater emphasis on phonological processing during dictation (
Afonso et al. 2020;
Iniesta et al. 2022;
Iniesta et al. 2021).
This finding holds critical importance for clinicians and educators, particularly for the early diagnosis of writing difficulties (dysgraphia) and in distinguishing whether the source of the problem lies in motor–coordination or phonological–orthographic processing. Recognizing dictation and copying as distinct dimensions allows for more targeted assessment and intervention. For instance, interventions that separately address graphomotor skills for copying and phonological/orthographic skills for dictation have shown promise in improving early writing outcomes. This nuanced understanding enables educators to better identify specific writing difficulties and tailor instruction accordingly, thereby bridging theoretical insights with classroom practice (
Alves et al. 2016;
Barnett et al. 2019;
McMaster et al. 2017).
The PLT demonstrated strong psychometric properties that align with modern measurement standards. High internal consistency coefficients for both the subscales and the total test indicate that the instrument reliably measures the intended constructs. Convergent and discriminant validity analyses confirmed the structural integrity of the three-factor model based on the Fornell–Larcker criterion. Notably, the high CFI and TLI values observed in the CFA indicate that the proposed model provides an excellent fit to the data. Although the RMSEA value exceeded the conventional threshold—a common occurrence in large samples and with the DWLS estimator—this does not undermine the model’s validity, given the excellent performance of the other fit indices (
Shi and Maydeu-Olivares 2020).
One of the most important methodological contributions of the study is the establishment of full scalar measurement invariance across gender groups. This finding demonstrates that the PLT measures the same psychological construct for both boys and girls, and that the items neither advantage nor disadvantage either group (
Vandenberg and Lance 2000). Establishing scalar invariance ensures that mean score differences between groups reflect genuine skill differences rather than measurement bias. This enables fair and valid comparisons in research and educational policy aimed at examining gender differences in early literacy skills.
The PLT has important practical implications for the Turkish education system. While traditional early literacy tests often focus on precursor skills (e.g., phonological awareness), the PLT is designed to directly assess the pre-literacy skills that children possess at the beginning of elementary school. Item analysis results show that the test includes both easy and difficult items, with an overall moderate level of difficulty. This feature enables the PLT to effectively identify not only at-risk students who require support but also advanced students who have already learned to read and write. This is a critical step in preventing the “Matthew Effect,” conceptualized by
Stanovich (
1986), in which strong readers continue to improve while weaker readers fall further behind. Accurately determining students’ initial literacy levels during the first days of school allows teachers to implement differentiated instruction and response-to-intervention models rather than applying a uniform curriculum (
Fuchs et al. 2012). This approach prevents advanced students from losing motivation while enabling early and targeted interventions for those who need additional support. In this respect, the PLT stands out as an evidence-based tool that promotes equal opportunity and individualized learning in education.
Beyond classroom practice, the PLT also offers broader practical implications for schools and policymakers. Because the test provides criterion-based and developmentally appropriate information on students’ literacy readiness, schools can use PLT results to plan resource allocation, determine the need for remedial programs, and develop evidence-based literacy initiatives. At the policy level, the PLT can contribute to early screening frameworks, ensuring that literacy difficulties are detected before gaps widen. Furthermore, researchers and curriculum developers can utilize PLT data to monitor cohort-level literacy trends, evaluate the effectiveness of instructional programs, and design interventions tailored to regional or demographic needs. In this respect, the PLT stands out as an evidence-based tool that not only supports individualized learning in classrooms but also informs decision-making processes that promote equity and early prevention in the education system.
5. Limitations and Recommendations for Future Research
This study should be interpreted in light of several limitations. First, participants were selected using a convenience sampling method and predominantly came from families with middle- to upper-socioeconomic backgrounds. This sample composition may limit the generalizability of the factor structure, item parameters, and normative characteristics of the PLT. Because socioeconomic context is known to influence early literacy exposure and pre-literacy skills, the current factor structure and item functioning may not fully represent children from lower socioeconomic backgrounds. Therefore, the findings should be interpreted with caution, and overgeneralization should be avoided until the PLT is validated with more socioeconomically diverse samples. Establishing national norms using a larger, more heterogeneous, and representative dataset is an important direction for future research.
Second, although convergent and discriminant validity were examined using statistical indices such as Composite Reliability (CR), Average Variance Extracted (AVE), and the square root of AVE, these analyses did not include external instruments. True convergent and discriminant validity should ideally be assessed using well-established measures in the field. Accordingly, further research is needed to examine how PLT scores relate to other validated pre-literacy and early literacy assessments.
Third, the present study employed a cross-sectional design, and the predictive validity of the PLT—for example, the extent to which PLT scores predict reading achievement at the end of first grade—was not examined. Longitudinal studies are required to determine how effectively the test predicts long-term academic outcomes.
Finally, additional avenues for research should include examining measurement invariance across key demographic and educational groups beyond those included in the current study. Testing invariance across variables such as socioeconomic status, geographical region, and bilingual status would provide stronger evidence for the robustness and fairness of the PLT across diverse populations. Moreover, cross-linguistic comparative studies—particularly within languages that vary in orthographic transparency—would further enrich theoretical models of literacy acquisition and help clarify the extent to which the multidimensional structure of early writing (e.g., copying and dictation) generalizes across orthographies. Such work would both deepen the theoretical contributions of the PLT and enhance its relevance to the broader international literature.