Next Article in Journal
The Relationship between Executive Functions, Working Memory, and Intelligence in Kindergarten Children
Next Article in Special Issue
Eye Gaze Patterns during Reasoning Provide Insights Regarding Individual Differences in Underlying Cognitive Abilities
Previous Article in Journal
Reducing Black–White Racial Differences on Intelligence Tests Used in Hiring for Public Safety Jobs
Previous Article in Special Issue
William Stern: The Relevance of His Program of ‘Differential Psychology’ for Contemporary Intelligence Measurement and Research
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Faster ≠ Smarter: Children with Higher Levels of Ability Take Longer to Give Incorrect Answers, Especially When the Task Matches Their Ability

by
Martin Tancoš
1,*,†,
Edita Chvojka
1,2,†,
Michal Jabůrek
1 and
Šárka Portešová
1
1
Psychology Research Institute, Faculty of Social Studies, Masaryk University, 602 00 Brno, Czech Republic
2
Departments of Methodology and Statistics and Interdisciplinary Social Science, Faculty of Social and Behavioral Sciences, University of Utrecht, 3584 CH Utrecht, The Netherlands
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Intell. 2023, 11(4), 63; https://doi.org/10.3390/jintelligence11040063
Submission received: 10 February 2023 / Revised: 18 March 2023 / Accepted: 23 March 2023 / Published: 29 March 2023
(This article belongs to the Special Issue Differential Psychology and Individual Differences in Intelligence)

Abstract

:
The stereotype that children who are more able solve tasks quicker than their less capable peers exists both in and outside education. The F > C phenomenon and the distance–difficulty hypothesis offer alternative explanations of the time needed to complete a task; the former by the response correctness and the latter by the relative difference between the difficulty of the task and the ability of the examinee. To test these alternative explanations, we extracted IRT-based ability estimates and task difficulties from a sample of 514 children, 53% girls, M(age) = 10.3 years; who answered 29 Piagetian balance beam tasks. We used the answer correctness and task difficulty as predictors in multilevel regression models when controlling for children’s ability levels. Our results challenge the ‘faster equals smarter’ stereotype. We show that ability levels predict the time needed to solve a task when the task is solved incorrectly, though only with moderately and highly difficult items. Moreover, children with higher ability levels take longer to answer items incorrectly, and tasks equal to children’s ability levels take more time than very easy or difficult tasks. We conclude that the relationship between ability, task difficulty, and answer correctness is complex, and warn education professionals against basing their professional judgment on students’ quickness.

1. Introduction

We hear school men very authoritatively saying that the fast students make the best grades and the slow ones the poorest. Statements of this kind are usually based on the assumption that if a student knows the subject in which he is being tested it should follow that he requires but a short time to make his answer. Needless to say, this assumption merits confirmation
(Longstaff and Porter 1928, p. 638; as cited in Gernsbacher et al. 2020).
It is highly likely that, when asked to imagine a straight-A student, many teachers would picture a kid who can answer all their questions without delay and always raises their hand first. This stereotype is so firm that it has paved the way into the common language. Many synonyms for the word ‘clever’ have something to do with speed: ‘quick’, ‘nimble’, or ‘fly’ are but some examples of this.
If the stereotype was valid, it could turn into a good heuristic for identifying gifted pupils or students that may be challenged to fulfil the study requirements. However, adhering to the stereotype may be dangerous when shown to be invalid, as teachers’ beliefs influence classroom practice (Brighton 2003; Cross 2009; Savasci and Berlin 2012). A significant amount of literature tackles the stereotype in mathematics. For example, Seeley (2009, 2016) claims that many teachers have internalised the stereotype. As a result, they see fast recall and computation as signs of high mathematical achievement and often neglect conceptual understanding. Stipek et al. (2001) found that maths teachers with more traditional values (mathematics being a solid body of knowledge that can be efficiently mastered without knowing what the symbols the pupils deal with represent; Thompson 1992) emphasised speed in completing a task as a necessary condition for being good at maths. When Thompson’s study was published, most American maths teachers held such traditional values. Independently of the ability in question, gifted children should be more likely to achieve above-average results. Many definitions of giftedness also include a time component (see, e.g., Subhi-Yamin 2009). Similarly, scales for giftedness ratings often have items that ask how quick and efficient the child is when dealing with a task (Pfeiffer and Jarosewich 2003; Renzulli 2021; Ryser and McConnell 2004).
Similar to Longstaff and Porter (1928; as cited in Gernsbacher et al. 2020), we think that the assumption merits confirmation even today. Most of the research carried out in the area is quickly becoming outdated. Moreover, some approaches to teaching and learning that are at odds with the stereotype (e.g., constructivist learning) are becoming more prevalent in education (Gravemeijer 2020; Steffe and Thompson 2000; Voskoglou and Salem 2020). To inspect the issue further, we may approach it from a psychological point of view.

1.1. The Uncertain Role of Intelligence

One of the most prominent frameworks for explaining the relationship between abilities and the time needed to complete a particular task is the Cattell–Horn–Carroll (CHC; Schneider and McGrew 2018) theory of intelligence. The CHC partitions general intelligence into three layers: the general intelligence itself (g; as the third-order factor), broad intellectual abilities (as second-order factors), and, finally, narrow intellectual abilities (as first-order factors). This common factor model (CFM; van Bork et al. 2017) implies that the correlation between processing speed and quantitative reasoning (ranging from 0.21 to 0.42, in the current technical manual of the WJ IV Test of Cognitive Abilities depending on age and specific subtests; McGrew et al. 2014) has arisen due to g. Thus, children that are both fast and capable simultaneously are so because of their general intelligence. Alternatively, Jensen (2011) equates g to the periodicity of neural oscillation. As such, differences in Jensen’s g would manifest through differences in reaction times. Therefore, both Jensen’s theory and CHC predict that children with higher levels of ability will take less time to finish a task, albeit with different mechanisms.
Nevertheless, alternative frameworks exist for modelling and understanding intelligence and related abilities. van der Maas et al. (2006, 2021) have shown that intelligence may be a dynamic system that emerges through mutual causal interactions of its components. Dynamical systems have been modelled as networks of partial correlations (or their equivalents). This allows for uncovering direct and indirect dependencies between the individual variables. This way, processing speed and particular abilities can still correlate, yet this dependency could only be induced by relationships with other variables in the system. Kan et al. (2019) compared factor and network models computed on subtests from WAIS-IV (Wechsler 2008). They found not only that the mutualistic network model explained the data better, but also that scores from the arithmetic reasoning subtest were weakly related only to one of the processing speed subtests (symbol search) and even had a direct negative relationship (albeit very small) with another processing speed subtest (coding). The mutualistic model tells a different story—students who are both fast and capable can be so for two reasons. Firstly, there is a direct relationship between processing speed (or some of its facets) and particular ability. Secondly, the students are both fast and capable because they are also good at another (so-called bridging) ability. Such a bridging ability would connect processing speed with the ability in question.
Regardless of the overarching theoretical framework, the character of the task might moderate the relationship between people’s ability levels and their response times. Goldhammer et al. (2014) showed that, when solving complex problems, students who take longer also perform better. However, when dealing with routine tasks (like basic reading), students who take longer are likely to be less successful than the faster ones. Scherer et al. (2015) observed that students who spend more time also tend to score higher on complex problem solving. To sum up, there is evidence against a direct relationship between the ability to process cognitive challenges quickly and other abilities relevant to the educational context.

1.2. Refuting the Stereotype

The relationship of cognitive speed with either intelligence or other more specific abilities is not straightforward and may be relatively challenging to uncover. Firstly, it is difficult to define what ‘an ability’ really is. Many theories aim to explain the same phenomenon but define abilities in vastly different ways (see the difference above between the CHC and Jensen’s g, where the former is an ability, yet the latter one ‘only’ causes variance in abilities). Consequently, it is not easy to obtain a good proxy for abilities. Grades do not explain enough variance in abilities on their own (Cucina et al. 2016), and standardised ability tests are not a common part of the curriculum. One way of empirically obtaining ability and difficulty estimates is via the item response theory (IRT; de Ayala et al. 2022). IRT is the most prevalent approach in modern psychometrics and is routinely used in test construction (Borsboom and Mellenbergh 2004). IRT models allow for estimating a so-called latent ability of each participant directly from answers to a set of items.
Modelling response times (through which cognitive speed is often operationalised) has a long tradition in psychology. Thurstone (1937) tried to formalise the relationship between reaction time and difficulty (defined as a ratio of people who were and were not able to perform a task successfully). van der Linden (2009) proposed a model of processing speed analogous to the model of speed in physics: processing speed equals the ratio of mental labour to time. In this paper, we will combine the two approaches. We will first use IRT modelling to extract empirical estimates of the abilities and difficulties of individual children and tasks, respectively. Then we will use these estimates to model the relationship between ability levels and the time needed to complete a task as a system of multilevel regressions.
The structure of this paper is as follows. First, we introduce the F > C phenomenon, which states that incorrect answers take more time to complete (Beckmann 2000; Beckmann et al. 1997). This phenomenon provides a conceptual basis for the model we propose. Then, we introduce Thissen’s model (Thissen 1983), which formalises the distance–difficulty hypothesis. This hypothesis states that the time needed to solve a task increases as the person’s ability nears the task’s difficulty. We then combine these two approaches in a model where the time needed to solve a task is a dependent variable. This model implies that children take the most time to solve a task whose difficulty matches their ability level, and the amount of time differs between correctly and incorrectly solved items1. We then test all models on data from a game-based test of logical thinking, controlling for children’s ability levels and varying task difficulties. We report how we determined our sample size, all data exclusions (if any), all manipulations, and all measures in the study (Simmons et al. 2012). We put our findings into context in the Discussion, with recommendations for future research and educational practice.

1.3. The F > C Phenomenon

The F > C phenomenon (Beckmann 2000; Beckmann et al. 1997) implies that incorrect responses take more time than correct ones. The F > C phenomenon can be formally expressed by Equation (1):
tij = μ + γ FCij + εij,
where tij is the response time of person j on item i, μ is an intercept (average time across each item and person), FCij is a binary variable that indicates whether the answer to item i of person j was false or correct, γ is the unstandardised regression coefficient that could be interpreted as a mean difference in response time between the false and correct answers, and εij is normally distributed residuals of the model.
The F > C phenomenon is robust (Beckmann 2000; Beckmann et al. 1997; Preckel and Freund 2005; Troche and Rammsayer 2005), but, at the same time, the authors admit that the emergent difference may be an artefact of empirical aggregation, as incorrect answers are more likely when dealing with the most difficult items in a test. Moreover, Beckmann and Beckmann (2005) also suggest that the magnitude of the F > C phenomenon may differ concerning the examinee’s ability. According to the authors, people who performed worse gave incorrect responses faster than examinees who were more successful. Preckel and Freund controlled for the ability, which did not have a significant effect, though due to their relatively small sample we might also attribute this to sampling error. Our study will examine the effects of both ability and difficulty.

1.4. The Distance–Difficulty Hypothesis

Thissen’s model was initially proposed in 1983. It was later revised by Ferrando and Lorenzo-Seva (2007), who perceived the model as a formal representation of the distance–difficulty hypothesis. The hypothesis states that the response time for a task decreases with the distance between the person’s ability level (θj) and the item’s difficulty (bi). In other words, people should take more time to solve tasks closer to their ability level. Conversely, people should spend less time on tasks that are substantially easy or difficult for them. Since the formal representation of the hypothesis includes an absolute value, the model implies that the predictive time differences should be the same and symmetrical. Thissen’s model is formally represented by Equation (2), which assumes a person who answers a set of items in a test within a certain time:
ln tij = μ + τj + βi − γ|θjbi| + εij,
where ln tij is a logarithmic transformation of the response time of person j spent on item i (the transformation is used to achieve normally distributed errors, εij, since the response times are assumed to be log-normally distributed); μ is the intercept, which could be interpreted as the mean time spent on all items among the whole sample; τj is a parameter for the general speediness of person j (how much the person spent on the items on average); βi is the time required to answer item i by the person of average ability; and γ is the magnitude of the linear relationship between the ability (θj) and difficulty (bi) absolute distance and the response time (expected to be negative by definition).

1.5. The Proposed Model

We will build a new model in two steps, testing two expectations. First, we will verify empirically whether the F > C phenomenon (Beckmann 2000; Beckmann et al. 1997) holds when controlling for person’s task difficulties, as the authors suggested. We expect the phenomenon to hold regardless of the item’s difficulty. We will also control for participants’ ability levels.
Secondly, the distance–difficulty hypothesis (Ferrando and Lorenzo-Seva 2007; Thissen 1983) implies that the relationship between the distance from the person’s level of ability and the time needed is symmetrical, no matter the direction of the difference. We aim to replicate this hypothesis. If both the distance–difficulty hypothesis, and the F > C phenomenon hold, we will extend Thissen’s model by the response correctness parameter (represented by the binary F > C term in Equation (1)). This parameter encodes whether the item was answered correctly. We expect this model to explain more variance than the original Thissen’s model. We will inspect whether there is a significant interaction between the ability–difficulty distance and response correctness. A significant interaction would suggest that the relationship between the time needed to solve a task and the ability-difficulty distance differs for correctly and incorrectly answered items.

2. Materials and Methods

2.1. Participants

The sample consisted of 514 children, 53% girls. The average age was 10.3 years (SD = 0.8 years), the youngest participant was eight years and two months old, and the oldest one was twelve years and four months old. Children were recruited in 16 Czech elementary schools willing to participate in a broader validation study of a giftedness-screening system. The schools selected the classes that would participate in the study. We worked with all children in a class whose legal guardians gave informed consent.
Since this research was a secondary data analysis, we performed an a posteriori sensitivity analysis in the mixedpower R package (version 0.1.0) to determine whether the sample was sufficient to detect small effects, as proposed by Kumle et al. (2021). The required sample to reach the sufficient power of 0.8 for all significant parameters in the two most complex models was 400. The setup script of our sensitivity analysis is in the Supplementary materials.

2.2. Measures

Triton and the Hungry Ocean

The data were gathered during several group sessions of pilot testing of a game-based assessment application for the identification of gifted students. The game’s concept is similar to MathGarden (Straatemeier 2014). Triton and the Hungry Ocean (referred to here as Triton) is based on the ‘balance beam task’ of Inhelder and Piaget (1958) that was later adopted by other authors and is currently referred to as Figure Weights (McGrew et al. 2014; Wechsler 2008). The objective is to choose a set of weights for one balance beam to counterbalance the weights on the other. Triton re-uses this principle in a submarine setting, including some novel features. The game uses cartoon-like graphics and simple sounds. There is no time limit for individual tasks.
A sample task is shown in Figure 1, where the individual features of the game are highlighted. Two circles are outlined by bubbles surrounding a hook (Feature 1). On the left side, the circle contains a certain number of animals (Feature 2), and the right-side circle is empty (further referred to as a slot; Feature 3). The player is supposed to fill this empty slot with one of the five groups of animals from the bottom part of the screen (Feature 4) to balance out the strength of the sea-creature group on the left side (Feature 2). Creatures of the same colour, shape, and number have the same strength. In more complex tasks, the strength of individual animals is expressed via so-called conditions: shorter hooks with both sides already occupied and balanced (Feature 5). These conditions imply the relative strength of specific animal types.
Besides moving groups of animals from the bottom part of the screen to the slots and back, the player is allowed to reset the task (i.e., return all the features to the original state) by pushing the reset button (Feature 6). They move to the next task by pushing the play button (Feature 7).
To solve the task, the player needs to deduce the relative strength of the individual animals, applying primarily logical reasoning. In general, the abilities here fall into fluid reasoning (within the CHC framework): logical reasoning intentionally and purposefully aimed at solving novel ‘on-the-spot’ problems. Such problems cannot be solved using previously learned habits, schemas, or scripts (Schneider and McGrew 2018). To solve the task, the children apply simple addition, subtraction, multiplication, and division). Therefore, they apply logical thinking within a mathematical context. This narrow ability is termed quantitative reasoning within the CHC model. However, we want to emphasise that the tasks require no advanced mathematical knowledge, and logical reasoning explains the most variance in the scores.
The game consisted of 29 tasks, the complexity of which gradually increased, as did the number of game mechanics involved. At the game’s start, three trial items did not contribute to participants’ scores. A description of these mechanics is available online in the supplemental material. A narrated video demonstration showcasing a task introduced each new game mechanic. The response time for each item was recorded from the moment children were exposed to the item task until they pushed the play button. In this study, we administered the game on PCs in groups of several children. Each child worked individually.

2.3. Data Management

We worked with data on the correctness of the solution for each task and the time spent on solving each task recorded by the Triton and the Hungry Ocean app. Correct responses were scored as “1”, and false ones as “0”. Response time was recorded in seconds with 0.5 s intervals. We also collected the participants’ data on gender, age, grade, and the school they attended.
No participants were excluded due to missing answers. All participants answered 22 tasks; even the last task, 29, had only 3.5% (n = 18) of missing values. All tasks left unanswered were coded as incorrect, and the corresponding time records were left missing.

2.4. Analysis plan

2.4.1. Preliminary IRT Models

We first needed to establish a well-fitting IRT model using the item response correctness data as an input. This IRT model allowed us to extract children’s ability levels (θj) and each item’s difficulty parameter (bi). We initially estimated the dichotomous Rasch model (Bond and Fox 2013) in R version 4.2.2 (R Core Team 2021) using package mirt (version 1.37.1; Chalmers 2012). This model estimates the probability of solving an item as a function of the participant’s ability. The dichotomous Rasch model is defined by Equation (3) as:
P i j = e   ( θ j   -   b i ) 1 + e   ( θ j -   b i ) ,
where Pij is the probability of the correct answer of person j to item i, θj is the ability of person j, and bi is the difficulty parameter of item i.
However, the Rasch model (Bond and Fox 2013) did not fit the data well. Since the design of Triton is based on selecting at least one answer from a set of five options, it is possible to select a correct answer by guessing. We extended the model by fixing the guessing parameter (the lower asymptote of the item characteristic curve) of each task by a reciprocal of all its possible solutions (e.g., if the task has five possible solutions, the guessing parameter would be fixed to value 0.2). Fixing the parameter led to the so-called dichotomous quasi-Rasch model with guessing (Linacre 2002), defined by Equation (4) as:
P ij = c i + ( 1 c i )   e   ( θ j - b i ) 1 + e   ( θ j - b i ) ,
where Pij is the probability of the correct answer of person j to item i, θj is the ability of person j, bi is the difficulty parameter of item i, and ci is the guessing parameter of item i.
This model had an acceptable fit. Thus, we extracted every item’s difficulty parameter (bi) and the ability parameter (θj) of every participant under this model and used these in the main analyses.

2.4.2. Main Analyses

We built two series of nested linear multilevel regression models. The first set of models tested the validity of the F > C phenomenon and controlled for task difficulty and participants’ ability levels. The second set tested the distance–difficulty hypothesis and its interaction with the F > C phenomenon. We used the lme4 (version 1.1-31; Bates et al. 2015) R package to estimate these models.
First, we defined a null model (Model 0) as a baseline for both series. Model 0 included only fixed and random intercept terms for participants and items to reflect that all observations were nested in participants who answered the same items. All models in the series used logarithmic transformation of response time as a dependent variable. This transformation linearised the relationship between the predictors and the response time. This is one of the main changes from previous studies on the F > C phenomenon. Beckmann et al. (Beckmann 2000; Beckmann et al. 1997; Beckmann and Beckmann 2005) did not check distributional assumptions and used the absolute item response time. This is problematic, as their analyses assumed normal distribution, but response times are log-normally distributed (van der Linden 2009). They also did not include any random parameters in their studies, which would allow for modelling systematic differences between the items and participants answering the same set of items.
The first model series focused on the F > C phenomenon. At first (Model A1), we included only a binary predictor for answer correctness (FCij). In the next step (Model A2), we added the predictors of item difficulty (bi) and person’s ability (θj) as control variables. Finally (Model A3), we added the interaction term of the answer correctness and the person’s ability to investigate whether the effect of the F > C phenomenon increases or decreases with higher/lower levels of participant’s ability. Model A3 is represented by Equation (5):
ln tij = μ + τj + βi + γ1 FCij + γ2 bi + γ3 θj + γ13 FCij θj + εij,
where ln tij is a logarithmic transformation of the response time of person j on item I; μ is the fixed intercept (response time of average-ability person spent on average-difficulty item); τj is the random intercept for each person (general speediness of each person); βi is the random intercept for each item (average time required to answer each item); γ1, γ2, γ3, and γ13 are the fixed effects of corresponding predictors; and εij is normally distributed residuals. The previous models, A1 and A2, could be derived from this equation by setting select regression coefficients to zero (see Appendix A).
Based on one reviewer’s suggestion, we estimated an additional Model A4, which included the interaction of a person’s ability and an item difficulty besides all terms from Model A3, as an exploratory feature. This interaction term tests whether the relationship between the children’s ability and response time varies according to the item’s difficulty. The equation of this model is in Appendix A.
The second model series assessed the distance–difficulty hypothesis and its incremental validity over the F > C phenomenon. To test the distance–difficulty hypothesis, we included only the absolute difference between the child’s ability and the difficulty of each item (|θjbi|) in the first model (Model B1). We extended the second model (Model B2) by a binary variable representing the F > C phenomenon (FCij) to assess the incremental validity of both concepts against each other. The model series ended with the last model (Model B3), where we added the interaction term of the distance and answer correctness (FCij). By including this term, we examined whether the distance difficulty effect followed a different pattern with correct and incorrect responses. Model B3 is represented by Equation (6) as follows:
ln tij = μ + τj + βi + γ4jbi| + γ1 FCij + γ14 FCijjbi| + εij,
where ln tij is a logarithmic transformation of the response time of person j on the item i; μ is the fixed intercept; τj is the random intercept for each person; βi is the random intercept for each item; γ1, γ4, and γ14 are the fixed effects of corresponding predictors; and εij is a normally distributed residual2. The previous models, B1 and B2, could be derived from this equation by setting select regression coefficients to zero (see Appendix A).

3. Results

3.1. Ability Estimates

As previously mentioned, we first estimated the dichotomous Rasch model (Bond and Fox 2013) to obtain children’s ability (θj) and item difficulty (bi) parameters for the main analyses. The Rasch model did not fit the item data well (M2(405) = 1255.32, p < .001, RMSEA = 0.065, SRMSR = 0.069, TLI = 0.867, AIC = 14,200.33, and BIC = 14,327.60). The empirical reliability of the sum score was rather high (r = 0.847).
We circumvented the issue by modifying the model to the quasi-Rasch model with guessing (Linacre 2002). The quasi-Rasch model showed an acceptable fit (M2(405) = 888.39, p < .001, RMSEA = 0.049, SRMSR = 0.075, TLI = 0.924, AIC = 14,000.78, and BIC = 14,128.04). Moreover, the empirical reliability of this model, r = 0.871, was slightly higher than that of the previous Rasch model. Item descriptive statistics with fixed guessing parameters and estimated difficulty parameters are listed in Table A1 (in Appendix B). The parameters from this model were used in multilevel regression models.

3.2. Null Model

The null model, listed in Table 1 and Table 2 as Model 0, provided a baseline for all subsequent models. The fixed intercept was significant (μ = 3.16, 95% CI [3.00, 3.33]). This parameter can be interpreted as the response time of an average-ability person on an average-difficulty item. Transforming the parameter from its logarithmic form, we obtained an average response time of 23.66 s.
We also found that random intercept terms (participants and items) explained a significant proportion of response time variance. The differences between the item intercept (var(βi) = 0.20, 95% CI [0.12, 0.33]) explained more variance of the response time than the individual differences of children in that characteristic (var(τj) = 0.07, 95% CI [0.06, 0.08]).

3.3. Models Assessing the F > C Phenomenon

In Model A1, we included a binary predictor reflecting the correctness of the answer on an item. In accordance with the F > C phenomenon, we found that the false answers took children significantly more time than the correct answers (γ1 = −0.04, 95% CI [−0.06, −0.01]). However, the effect size was relatively small, and the transformed parameter indicated that the expected average difference in response times between wrongly and correctly answered items was 0.87 s.
In Model A2, the F > C phenomenon effect remained significant once controlled for item difficulty and children’s ability (γ1 = −0.05, 95% CI [−0.08, −0.03]). In addition, we found that response time was significantly higher in children with higher ability (γ2 = 0.06, 95% CI [0.05, 0.07]). On the other hand, response time did not have a significant relationship with item difficulty (γ3 = 0.05, 95% CI [0.00, 0.12]).
In Model A3, we added an interaction of answer correctness and children’s ability, which was significant (γ13 = −0.12, 95% CI [−0.13, −0.11]). Adding the interaction also slightly suppressed the F > C phenomenon effect (γ1 = −0.07, 95% CI [−0.09, −0.04]), as well as the relationship of children’s ability with response time (γ3 = 0.12, 95% CI [0.11, 0.14]). All effects combined, the relationship between children’s ability and response time was negligible when the item was answered correctly. However, response time increased with higher children’s ability in case of false answers. Figure 2 illustrates these patterns.
As shown in Table 1, Model A3 was the best, and the information criteria also supported this model as the best one. All models fitted significantly better than the null model, and each model had a significantly better fit than the previous models in the sequence.
Table 1. Parameters of the models assessing the F > C phenomenon (interacting with a person’s ability).
Table 1. Parameters of the models assessing the F > C phenomenon (interacting with a person’s ability).
Model 0Model A1Model A2Model A3
95% CI 95% CI 95% CI 95% CI
coef.est. LLULest. LLULest. LLULest. LLUL
Fixed effects
interceptμ3.16***3.003.333.18***3.023.353.15***2.993.313.19***3.033.36
correct answer (FC)γ1 −0.04**−0.06−0.01−0.05***−0.08−0.03−0.07***−0.09−0.04
item difficultyγ2 0.05 0.000.120.05 0.000.11
person abilityγ3 0.06***0.050.070.12***0.110.14
FC × abilityγ13 −0.12***−0.13−0.11
Random effects
person intercept variancevar(τj)0.07***0.060.080.07***0.060.080.06***0.050.070.06***0.050.07
item intercept variancevar(βi)0.20***0.120.330.19***0.110.320.17***0.090.280.17***0.100.29
residual variancevar(εij)0.37***0.360.380.37***0.360.380.37***0.360.380.36***0.350.37
Goodness of fit
conditional R2 0.417 0.415 0.422 0.439
marginal R2 0.000 0.001 0.055 0.070
log-likelihood −14,193 −14,189 −14,154 −14,008
AIC 28,395 28,389 28,323 28,031
BIC 28,425 28,427 28,376 28,092
Δχ2 (df) 8.09 (1)** 70.08 (2)*** 293.23 (1)***
Notes. coef.—coefficient, est.—estimate, CI—confidence interval, LL—lower limit, UL—upper limit, var—variance; ** p < .010, *** p < .001.
In Model A4, we included the interaction of a person’s ability and an item difficulty. As we did not hypothesise this exploratory model, we describe and interpret it separately from Models A1–A3. As seen in Table 2, the ability–difficulty interaction was significant (γ23 = 0.04, 95% CI [0.03, 0.04]). This means the relationship between the ability and response time was stronger for more difficult items. Adding this information also significantly improved the model fit in comparison with the previous Model A3.
Including the interaction term also partially explained the F > C phenomenon, whose main effect became non-significant (γ1 = −0.02, 95% CI [−0.04, 0.01]). The strength of the interaction between answer correctness and children’s ability also noticeably decreased (γ13 = −0.03, 95% CI [−0.04, −0.01]). Figure 3 aids the interpretation of the additional term. It further expands the interpretation of Model A3, indicating that the effect of the ability on time required to answer incorrectly answered items applies only to moderately and highly difficult items.
Table 2. Parameters of the exploratory Model A4 (including the interaction of item difficulty and person ability).
Table 2. Parameters of the exploratory Model A4 (including the interaction of item difficulty and person ability).
95% CI
coef.est. LLUL
Fixed effects
interceptμ3.14***2.983.30
correct answer (FC)γ1−0.02 −0.040.01
item difficultyγ20.06 0.000.12
person abilityγ30.05***0.030.06
FC × abilityγ13−0.03***−0.04−0.01
difficulty × abilityγ230.04***0.030.04
Random effects
person intercept variancevar(τj)0.06***0.050.07
item intercept variancevar(βi)0.17***0.100.28
residual variancevar(εij)0.34***0.330.35
Goodness of fit
conditional R2 0.466
marginal R2 0.097
log-likelihood −13,606
AIC 27,231
BIC 27,300
Δχ2 (df) 802.28 (1)***
Notes. coef.—coefficient, est.—estimate, CI—confidence interval, LL—lower limit, UL—upper limit, var—variance; *** p < .001; goodness of fit comparison with Model A3.

3.4. Models Assessing the Distance–Difficulty Hypothesis

In model B1, we tested the effect of absolute distance between the children’s ability and item difficulty. This effect was significant (γ13 = −0.13, 95% CI [−0.14, −0.12]). This means that the response time decreased 1.14 times with each logit unit of the absolute ability–difficulty distance, which is a moderately strong effect. The fixed intercept showed the average response time of 33.14 s for zero distance (the item difficulty equivalent to the person’s ability), where the time was at its maximum. The estimated response time decreases to 29.03 when the ability–difficulty distance is one logit unit, to 25.45 when the distance is two logit units, and so on.
Combining ability–difficulty distance with the F > C phenomenon led to Model B2, where the F > C phenomenon did not show a significant effect (γ1 = −0.02, 95% CI [−0.04, 0.01]); the ability–difficulty distance effect remained unchanged (γ4 = −0.13, 95% CI [−0.14, −0.12]).
In Model B3, we extended the previous model by the interaction of the ability–difficulty distance effect and the F > C phenomenon, which was significant (γ14 = 0.05, 95% CI [0.04, 0.07]). Including the interaction also suppressed the F > C phenomenon effect, which became significant (γ1 = −0.10, 95% CI [−0.13, −0.07]); small suppression was also visible in the ability–difficulty distance effect (γ4 = −0.16, 95% CI [−0.17, −0.15]). Interpretation-wise, the interaction effect means that, for false answers, the negative relationship between ability–difficulty distance and response time is stronger than for correct answers. A graphical description of these effects is in Figure 4.
This sequence of models also showed that Model B3 outperformed the preceding models in fit and information criteria—as listed in Table 3. All three models were significantly better than the null model; nevertheless, Model B2 did not fit significantly better than Model B1 (Δχ2(1) = 2.14, p = 0.144).

4. Discussion

In this study, we wanted to test the validity of the ‘faster equals smarter’ stereotype. More precisely, our aim was twofold. First, we wanted to inspect whether the F > C phenomenon holds when controlling for item difficulty and a person’s ability, as Beckman et al. (Beckmann 2000; Beckmann et al. 1997; Beckmann and Beckmann 2005) mentioned that the phenomenon might be a byproduct of empirical aggregation. In our data, the F > C phenomenon remained significant after controlling for ability and difficulty, albeit with a small effect size. Item difficulty did not influence the time needed to solve a task, but children’s ability did, though only with moderately and highly difficult items. Moreover, we found an interaction between ability level and correctness. Together, these results mean that children with higher levels of ability take longer to give incorrect answers on moderately difficult and difficult items than their peers with lower levels of ability. With correct answers, there is no relationship between ability and time. If the ‘faster equals smarter’ stereotype is true, it would imply that children with higher levels of ability answer more quickly in general. Our results, therefore, go against the stereotype, as they imply either no difference at all or longer response times of children with higher ability levels.
There are several explanations for this phenomenon. High-ability children may, on average, possess better meta-cognitive skills (Swanson 1992). They may invest some time to devise a strategy before turning in an answer, regardless of its correctness. These children may also put more effort into problem solving because of their better self-regulation and thus show greater persistence (Howard and Vasseleu 2020). High-ability children are also likely to see difficult tasks as challenging. This positive framing can make them spend more time on a task, as opposed to their peers that may see difficult tasks as too demanding (Bouffard-Bouchard et al. 1993). Finally, it is worth noting that the causality can also flow in the reverse way: high-ability children may be persistent and determined in the first place, which may, in turn, accelerate their development and make them score high in ability tests.
The small effect size may reflect the true phenomenon’s strength, as Beckman et al. (Beckmann 2000; Beckmann et al. 1997; Beckmann and Beckmann 2005) violated the distributional assumptions of their analysis, which likely inflated the estimated effect. The parametric t-test performed in their study assumed a normal distribution, but response times are commonly assumed to be log-normally distributed (van der Linden 2009). Moreover, in alignment with the original authors’ proposition, part of the variance in time previously explained by the correctness as the only parameter has indeed been explained by children’s ability, which attenuates the main effect of correctness.
Since the F > C phenomenon held, we proceeded to the study’s second aim: to test the distance–difficulty hypothesis and build a new model that included the relative difference between the examinee’s ability level and the task difficulty, and the F > C phenomenon. This allowed us to assess whether the phenomena were incrementally valid over one another; in other words, whether the relationship between the distance–difficulty term and the time needed to solve an item differed between correctly and incorrectly solved tasks. We supported the distance–difficulty hypothesis and found its interaction with the F > C phenomenon. This means that items that match the children’s ability take the longest to solve, and the time is even longer for incorrectly answered items. As the difference between the ability and item difficulty grows (when the examinee works with tasks that are very easy or difficult for them), the difference narrows to the point when it changes direction. Tasks that are too easy or too difficult take more time when answered correctly than incorrectly, as the hypothesis operates with the absolute difference value.
This seems intuitive with items whose difficulty surpasses the child’s ability, as solving such tasks would require considerable cognitive effort, which would ultimately increase the time needed to produce a correct answer. Since the order of the items in Triton was roughly arranged according to their difficulty, the most difficult items were administered towards the end of the test. In such a setting, children may have been frustrated, bored or tired. Experiencing the discomfort could have made them simply give up on the task, guess, or produce an erroneous solution when trying to turn the answer in as soon as possible.
However, when the tasks were far below their ability levels, children also took longer to answer correctly. As opposed to very difficult tasks, this may seem surprising. In this situation, for instance, children could have hesitated about whether the task was that easy or whether they may have overlooked something or faced a trick question, which would have made them ponder about the answer and submit it later. Moreover, the solutions to the easiest task may have seemed obvious, leading children to rush the answer and answer incorrectly. Since the easiest items were administered at the start of the test, it is also possible that some children had not understood the instructions fully and underestimated the task, leading to erroneous answers.
It also is worth noting that the main effect of correctness only became significant once we added the interaction into the model. The interaction between the distance–difficulty term and the correctness of the answer removed irrelevant variance of the main term, strengthening its relationship with response time.
These findings may influence how we perceive time demands in educational practice. According to our results, solving tasks on newly acquired material, whose difficulty parallels a child’s ability, takes the most time. This is in line with classic theories of learning acquisition, such as Vygotsky’s proximal development zone (Eun 2019; Roth 2020), which builds upon the thesis that learning primarily happens when dealing with tasks whose difficulty is close to the child’s ability. Mediated learning experience (Tzuriel 2021) also works with the mechanism. The mediators adjust the difficulty of the task to provide the child with appropriate learning stimuli. When facing complex tasks, the mediator should offer substantial support. On the other hand, with simple tasks, the mediator should withdraw. Children that seem to take too long to solve some tasks may need this time to assimilate new knowledge. The need for time may signalise ongoing learning instead of poor performance.
To our knowledge, this is the first study that combines answer correctness and the distance between the ability and difficulty to explain variance in response times. We have shown that both the F > C phenomenon and the distance–difficulty hypothesis hold and are incrementally valid over one another. Our findings also go against the stereotype that children who are more able solve tasks quickly. However, our results must be interpreted with caution, as overall our estimated effects are not of great magnitude. Moreover, it is essential to remember that our findings concern isolated individual tasks and may differ when dealing with a whole test.
To illustrate this problem, imagine two students, Patt and Matt, and a test of 10 items ordered by difficulty. The expected response time for a correct answer (regardless of difficulty) is larger than for an incorrect answer. Patt’s ability is higher than Matt’s. Patt answers seven items before she starts struggling and giving incorrect answers that take longer. Matt already struggles with the fourth item. There are settings when Patt will finish sooner than Matt, as she takes longer to give incorrect answers but produces fewer of them. There are also configurations when the difference may be negligible. Since the effect and parameters estimated in our study were extracted directly from the test data, our results are not a universal predictive guide. The difference in the total time needed to complete a specific test between two children with different ability levels depends on the configuration of the differences between their abilities and the differences in the time needed for giving correct vs incorrect answers. We do not want to imply that children who are more able will always finish later. However, we do want to refute the expectation that a more capable child will automatically finish sooner.
Furthermore, both the F > C phenomenon and the distance–difficulty hypothesis assume that the test in question is not a timed test. Timed tests are more prevalent in the educational context, and it is reasonable to assume that, with a time limit, additional factors need to be considered, for instance, the individual strategies of the test-takers. Cultural context and school climate may play a role as well. Despite recent developments, the Czech school system has been mostly seen as authoritarian and focused on memorising (Perry 2005). It prioritises the successful transmission of knowledge over creativity and positive learning experience (Straková and Simonová 2013). Children educated in such a system may not be motivated to keep working on a task without knowing the correct answer right away. The effects we found may be less or more pronounced in different school systems; for instance, based on how the school system treats failure or how strongly the teachers have internalised the ‘faster equals smarter’ stereotype. Finally, while we think that our computational approach allowed us to remedy some methodological issues of the studies we built upon, it is not perfect. Our two-step estimation (first IRT and then multilevel regression model) leads to two sets of errors. Certain modelling approaches, such as generalised mixed modelling for explanatory item response analyses (De Boeck and Wilson 2004; Baayen et al. 2008), allow for incorporating the person- and item-level effects in a one-step estimation that also takes the clustered data structure into account. It may be interesting to re-analyse our or similar data using these techniques to get even more accurate estimates of the effects reported in our study. We hope that these limitations can inform future research on related topics.
Further research can also explore whether our results also hold with knowledge-based tasks, which would tap into crystallised, as opposed to fluid, intelligence. Teachers often observe that students who can quickly answer questions tend to have better knowledge accessibility, suggesting a strong understanding of the subject. Goldhammer et al. (2014) have shown that students who take longer perform better with complex tasks, but with routine tasks (like basic reading) the opposite is true. We should investigate and separate the contexts where we can refute the stereotype from these where it still may apply to some extent. Overall, the relationship between the time needed to solve a task and children’s ability, task difficulty, and answer correctness is complex, and the influence of other variables is yet to be examined. Nevertheless, educational professionals should avoid basing their professional judgement on how quickly a child processes a task and be aware that tasks with great learning potential are likely to take a long time.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/jintelligence11040063/s1; Table S1: game mechanics; R script S1: Analytic script; R script S2: Plot script; R script S3: Sensitivity analysis (with results).

Author Contributions

Conceptualisation, M.T., E.C., M.J. and Š.P.; methodology, M.T. and E.C.; coding, M.T.; formal analysis, M.T.; investigation, M.T., E.C. and M.J.; resources, Š.P.; data curation, M.T. and M.J.; writing—original draft preparation, E.C.; writing—review and editing, E.C. and Š.P.; visualisation, M.T.; supervision, Š.P.; project administration, M.J.; funding acquisition, Š.P. All authors have read and agreed to the published version of the manuscript.

Funding

This study is part of the project “MUNI/A/1473/2022”, and it received support and the APC from the Grant Projects of Specific Research Program at the Masaryk University, Faculty of Social Studies.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Ethics Committee of Masaryk University (protocol code EKV-2021-117, date 12 February 2022).

Informed Consent Statement

Informed consent was obtained from legal guardians of all subjects involved in the study.

Data Availability Statement

The data, together with analytic scripts, are freely available on the Open Science Framework repository under the following URL: https://osf.io/dhu7f/ (accessed on 22 March 2023).

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Appendix A

In this section, all model equations used in this study are listed.
Null model (Model 0):
ln tij = μ + τj + βi + εij,
Model A1:
ln tij = μ + τj + βi + γ1 FCij + εij,
Model A2:
ln tij = μ + τj + βi + γ1 FCij + γ2 bi + γ3 θj + εij,
Model A3:
ln tij = μ + τj + βi + γ1 FCij + γ2 bi + γ3 θj + γ13 FCij θj + εij,
Model A4:
ln tij = μ + τj + βi + γ1 FCij + γ2 bi + γ3 θj + γ13 FCij θj + γ23 bi θj + εij,
Model B1:
ln tij = μ + τj + βi + γ4jbi| + εij,
Model B2:
ln tij = μ + τj + βi + γ4jbi| + γ1 FCij + εij,
Model B3:
ln tij = μ + τj + βi + γ4jbi| + γ1 FCij + γ14 FCijjbi| + εij,

Appendix B

Table A1. Item descriptive statistics and quasi-Rasch model parameters.
Table A1. Item descriptive statistics and quasi-Rasch model parameters.
Response CorrectnessResponse Time (in Seconds)
ItemSampleGuessingDifficultyMSDMSDMdn
item 15140.200−1.280.750.4321.1924.5715.00
item 25140.200−1.910.810.3919.5217.8014.50
item 35140.067−3.350.910.2831.7524.6824.00
item 45140.067−3.120.900.3020.8021.4916.50
item 55140.050−1.260.710.4647.7138.1835.50
item 65140.067−2.320.830.3720.0715.5915.50
item 75140.200−1.970.840.3715.9625.8810.00
item 85140.100−2.540.860.3524.8121.5518.00
item 95140.2000.170.590.4913.5210.5710.00
item 105140.050−1.620.760.4334.3819.4628.50
item 115140.200−0.540.680.4722.8918.0417.50
item 125140.200−1.550.790.4110.878.868.50
item 135140.2002.910.250.4415.9216.1411.00
item 145140.050−0.170.560.5051.2230.0544.20
item 155140.2000.990.490.5027.2319.4321.00
item 165140.0170.310.460.5050.4033.9842.50
item 175140.2001.400.440.5031.6324.0526.00
item 185140.2001.290.420.4925.9320.6920.00
item 195140.0500.350.480.5042.4428.5234.25
item 205140.2001.740.380.4942.1932.5033.75
item 215140.2002.190.320.4729.1423.7622.50
item 225140.0673.150.190.3961.9353.7445.75
item 235140.2001.440.400.4925.4519.6019.50
item 245100.1003.600.160.3743.2136.1631.75
item 255070.0503.090.150.3660.8152.3846.00
item 265040.2004.880.150.3639.1542.5326.50
item 275020.2003.710.280.4531.2728.1222.50
item 285000.0505.230.060.2444.3337.5533.50
item 294960.0175.800.040.1945.7040.6035.00
Note. Correct response was scored as “1”, false as “0”, thus, response correctness mean represents the ratio of correct responses.

Notes

1
In this study, we use the terms ‘item’ and ‘task’ semi-interchangeably. The word ‘item’ refers to a clearly demarcated part of the test whose psychometric difficulty can be empirically extracted. The word ‘task’ refers to the content of the item. In the case of Triton, children solve the same ‘task’ (balance both sides of the equation) many times, though are administered ‘items’ of varying difficulty.
2
Please note that the individual effects of item difficulty (bi) and a person’s ability (θj) are not included in the model, as they are already used to form the distance–difficulty difference term.

References

  1. Baayen, R. Harald, Douglas J. Davidson, and Douglas M. Bates. 2008. Mixed-effects modeling with crossed random effects for subjects and items. Journal of Memory and Language 59: 390–412. [Google Scholar] [CrossRef]
  2. Bates, Douglas, Martin Mächler, Ben Bolker, and Steve Walker. 2015. Fitting Linear Mixed-Effects Models Using Lme4. Journal of Statistical Software 67: 1–48. [Google Scholar] [CrossRef]
  3. Beckmann, Jens F. 2000. Differentielle Latenzzeiteffekte Bei Der Bearbeitung von Reasoning-Items [Differential Effects on Latencies in Solving Reasoning Items]. Diagnostica 46: 124–29. [Google Scholar] [CrossRef]
  4. Beckmann, Jens F., and Nadin Beckmann. 2005. Effects of Feedback on Performance and Response Latencies in Untimed Reasoning Tests. Psychology Science 47: 262–78. [Google Scholar]
  5. Beckmann, Jens F., Jürgen Guthke, and Hans Vahle. 1997. Analysen Zum Zeitverhalten Bei Computergestützten Adaptiven Intelligenz-Lerntests. Diagnostica 43: 40–62. [Google Scholar]
  6. Bond, Trevor G., and Christine M. Fox. 2013. Applying the Rasch Model: Fundamental Measurement in the Human Sciences. London: Psychology Press. [Google Scholar]
  7. Borsboom, Denny, and Gideon J. Mellenbergh. 2004. Why Psychometrics Is Not Pathological: A Comment on Michell. Theory & Psychology 14: 105–20. [Google Scholar] [CrossRef]
  8. Bouffard-Bouchard, Thérèse, Sophie Parent, and Serge Lavirée. 1993. Self-regulation on a concept-formation task among average and gifted students. Journal of Experimental Child Psychology 56: 115–34. [Google Scholar] [CrossRef]
  9. Brighton, Catherine M. 2003. The Effects of Middle School Teachers’ Beliefs on Classroom Practices. Journal for the Education of the Gifted 27: 177–206. [Google Scholar] [CrossRef]
  10. Chalmers, R. Philip. 2012. Mirt: A Multidimensional Item Response Theory Package for TheREnvironment. Journal of Statistical Software 48: 1–29. [Google Scholar] [CrossRef]
  11. Cross, Dionne I. 2009. Alignment, Cohesion, and Change: Examining Mathematics Teachers’ Belief Structures and Their Influence on Instructional Practices. Journal of Mathematics Teacher Education 12: 325–46. [Google Scholar] [CrossRef]
  12. Cucina, Jeffrey M., Sharron T. Peyton, Chihwei Su, and Kevin A. Byle. 2016. Role of Mental Abilities and Mental Tests in Explaining High-School Grades. Intelligence 54: 90–104. [Google Scholar] [CrossRef]
  13. de Ayala, R. J., Bruno D. Zumbo, David J. Weiss, Mark Reckase, and Scott Hofer. 2022. The Theory and Practice of Item Response Theory the Theory and Practice of Item Response Theory. London: Guilford Press. [Google Scholar]
  14. De Boeck, Paul, and Mark Wilson, eds. 2004. Explanatory Item Response Models: A Generalised Linear and Nonlinear Approach. New York: Springer, vol. 10. [Google Scholar]
  15. Eun, Barohny. 2019. The Zone of Proximal Development as an Overarching Concept: A Framework for Synthesizing Vygotsky’s Theories. Educational Philosophy and Theory 51: 18–30. [Google Scholar] [CrossRef]
  16. Ferrando, Pere J., and Urbano Lorenzo-Seva. 2007. An Item Response Theory Model for Incorporating Response Time Data in Binary Personality Items. Applied Psychological Measurement 31: 525–43. [Google Scholar] [CrossRef]
  17. Gernsbacher, Morton Ann, Raechel N. Soicher, and Kathryn A. Becker-Blease. 2020. Four Empirically Based Reasons Not to Administer Time-Limited Tests. Translational Issues in Psychological Science 6: 175–90. [Google Scholar] [CrossRef] [PubMed]
  18. Goldhammer, Frank, Johannes Naumann, Annette Stelter, Krisztina Tóth, Heiko Rölke, and Eckhard Klieme. 2014. The time on task effect in reading and problem solving is moderated by task difficulty and skill: Insights from a computer-based large-scale assessment. Journal of Educational Psychology 106: 608. [Google Scholar] [CrossRef]
  19. Gravemeijer, Koeno. 2020. A Socio-Constructivist Elaboration of Realistic Mathematics Education. In National Reflections on the Netherlands Didactics of Mathematics. Cham: Springer, pp. 217–33. [Google Scholar]
  20. Howard, Steven James, and Elena Vasseleu. 2020. Self-regulation and executive function longitudinally predict advanced learning in preschool. Frontiers in Psychology 11: 49. [Google Scholar] [CrossRef]
  21. Inhelder, Bärbel, and Jean Piaget. 1958. The Growth of Logical Thinking from Childhood to Adolescence: An Essay on the Construction of Formal Operational Structures. London: Psychology Press, vol. 22. [Google Scholar]
  22. Jensen, Arthur R. 2011. The Theory of Intelligence and Its Measurement. Intelligence 39: 171–77. [Google Scholar] [CrossRef]
  23. Kan, Kees-Jan, Han L. J. van der Maas, and Stephen Z. Levine. 2019. Extending Psychometric Network Analysis: Empirical Evidence against g in Favor of Mutualism? Intelligence 73: 52–62. [Google Scholar] [CrossRef]
  24. Kumle, Levi, Melissa L.-H. Võ, and Dejan Draschkow. 2021. Estimating Power in (Generalized) Linear Mixed Models: An Open Introduction and Tutorial in R. Behavior Research Methods 53: 2528–43. [Google Scholar] [CrossRef]
  25. Linacre, John Michael. 2002. Dichotomous Quasi-Rasch Model with Guessing. Rasch Measurement Transactions 15: 856. [Google Scholar]
  26. McGrew, Kevin S., Erica M. LaForte, and Fredrick A. Schrank. 2014. Technical Manual: Woodcock-Johnson IV. Rolling Meadows: Riverside. [Google Scholar]
  27. Perry, Laura B. 2005. The Seeing and the Seen: Contrasting Perspectives of Post-communist Czech Schooling. Compare: A Journal of Comparative and International Education 35: 265–83. [Google Scholar] [CrossRef]
  28. Pfeiffer, Steven I., and Tania Jarosewich. 2003. GRS: Gifted Rating Scales. San Antonio: Psychological Corporation. [Google Scholar]
  29. Preckel, Franzis, and Philipp Alexander Freund. 2005. Accuracy, Latency, and Confidence in Abstract Reasoning: The Influence of Fear of Failure and Gender. Psychology Science 47: 230. [Google Scholar]
  30. R Core Team. 2021. R: A Language and Environment for Statistical Computing. Vienna: R Foundation for Statistical Computing. Available online: https://www.R-project.org/ (accessed on 22 March 2023).
  31. Renzulli, Joseph. 2021. Scales for Rating the Behavioral Characteristics of Superior Students: Technical and Administration Manual. London: Routledge. [Google Scholar]
  32. Roth, Wolff-Michael. 2020. Zone of Proximal Development in Mathematics Education. In Encyclopedia of Mathematics Education. Cham: Springer, pp. 913–16. [Google Scholar]
  33. Ryser, Gail R., and Kathleen McConnell. 2004. Scales for Identifying Gifted Students. Waco: Prufrock Press Inc. [Google Scholar]
  34. Savasci, Funda, and Donna F. Berlin. 2012. Science Teacher Beliefs and Classroom Practice Related to Constructivism in Different School Settings. Journal of Science Teacher Education 23: 65–86. [Google Scholar] [CrossRef]
  35. Scherer, Ronny, Samuel Greiff, and Jarkko Hautamäki. 2015. Exploring the relation between time on task and ability in complex problem solving. Intelligence 48: 37–50. [Google Scholar] [CrossRef]
  36. Schneider, W. Joel, and Kevin S. McGrew. 2018. The Cattell–Horn–Carroll Theory of Cognitive Abilities. In Contemporary Intellectual Assessment: Theories, Tests, and Issues, 4th ed. New York: The Guilford Press, pp. 73–163. [Google Scholar]
  37. Seeley, Cathy L. 2009. Faster Isn’t Smarter: Messages about Math, Teaching, and Learning in the 21st Century: A Resource for Teachers, Leaders, Policy Makers, and Families. Sausalito: Math Solutions. [Google Scholar]
  38. Seeley, Cathy L. 2016. Building a Math-Positive Culture: How to Support Great Math Teaching in Your School (ASCD Arias). Denver: ASCD. [Google Scholar]
  39. Simmons, Joseph P., Leif D. Nelson, and Uri Simonsohn. 2012. A 21 Word Solution. SSRN 2160588. Available online: http://ssrn.com/abstract=2160588 (accessed on 9 February 2023).
  40. Steffe, Leslie P., and Patrick W. Thompson. 2000. Teaching Experiment Methodology: Underlying Principles and Essential Elements. In Handbook of Research Design in Mathematics and Science Education. Mahwah: Lawrence Erlbaum Associates, pp. 267–306. [Google Scholar]
  41. Stipek, Deborah J., Karen B. Givvin, Julie M. Salmon, and Valanne L. MacGyvers. 2001. Teachers’ Beliefs and Practices Related to Mathematics Instruction. Teaching and Teacher Education 17: 213–26. [Google Scholar] [CrossRef]
  42. Straatemeier, Marthe. 2014. Math Garden: A New Educational and Scientific Instrument. Amsterdam: Universiteit van Amsterdam. [Google Scholar]
  43. Straková, Jana, and Jaroslava Simonová. 2013. Assessment in the School Systems of the Czech Republic. Assessment in Education: Principles, Policy & Practice 20: 470–90. [Google Scholar] [CrossRef]
  44. Subhi-Yamin, Taisir. 2009. Gifted Education in the Arabian Gulf and the Middle Eastern Regions: History, Current Practices, New Directions, and Future Trends. In International Handbook on Giftedness. Berlin/Heidelberg: Springer, pp. 1463–90. [Google Scholar]
  45. Swanson, H. Lee. 1992. The relationship between metacognition and problem solving in gifted children. Roeper Review 15: 43–48. [Google Scholar] [CrossRef]
  46. Thissen, D. 1983. 9—Timed Testing: An Approach Using Item Response Theory. In New Horizons in Testing. Edited by David J. Weiss. San Diego: Academic Press, pp. 179–203. [Google Scholar] [CrossRef]
  47. Thompson, Alba G. 1992. Teachers’ Beliefs and Conceptions: A Synthesis of the Research. In Handbook of Research on Mathematics Teaching and Learning: A Project of the National Council of Teachers of Mathematics. New York: Macmillan Publishing Co, Inc., pp. 127–46. [Google Scholar]
  48. Thurstone, Louis Leon. 1937. Ability, Motivation, and Speed. Psychometrika 2: 249–54. [Google Scholar] [CrossRef]
  49. Troche, Stefan, and Thomas Rammsayer. 2005. The “False > Correct-Phenomenon” and Subjective Confidence: Two Distinct Phenomena Influencing Response Latencies in Psychological Testing. Psychology Science 47: 246. [Google Scholar]
  50. Tzuriel, David. 2021. The Theory of Structural Cognitive Modifiability and Mediated Learning Experience (SCM-MLE). In Mediated Learning and Cognitive Modifiability. Edited by David Tzuriel. Social Interaction in Learning and Development. Cham: Springer International Publishing, pp. 13–52. [Google Scholar] [CrossRef]
  51. van Bork, Riet, Lisa D. Wijsen, and Mijke Rhemtulla. 2017. Toward a Causal Interpretation of the Common Factor Model. Disputatio 9: 581–601. [Google Scholar] [CrossRef]
  52. van der Linden, Wim J. 2009. Conceptual Issues in Response-Time Modeling. Journal of Educational Measurement 46: 247–72. [Google Scholar] [CrossRef]
  53. van der Maas, Han L. J., Alexander O. Savi, Abe Hofman, Kees Jan Kan, and Maarten Marsman. 2021. The Network Approach to General Intelligence. PsyArXiv. [Google Scholar] [CrossRef]
  54. van der Maas, Han L. J., Conor V. Dolan, Raoul P. P. P. Grasman, Jelte M. Wicherts, Hilde M. Huizenga, and Maartje E. J. Raijmakers. 2006. A Dynamical Model of General Intelligence: The Positive Manifold of Intelligence by Mutualism. Psychological Review 113: 842–61. [Google Scholar] [CrossRef] [PubMed]
  55. Voskoglou, Michael Gr, and Abdel-Badeeh M. Salem. 2020. Benefits and Limitations of the Artificial with Respect to the Traditional Learning of Mathematics. Mathematics 8: 611. [Google Scholar] [CrossRef]
  56. Wechsler, David. 2008. Wechsler Adult Intelligence Scale, 4th ed. Washington, DC: American Psychological Association (APA). [Google Scholar] [CrossRef]
Figure 1. Sample task and individual game features of Triton.
Figure 1. Sample task and individual game features of Triton.
Jintelligence 11 00063 g001
Figure 2. Predicted response time values according to Model A3 for correct (green) and incorrect (red) responses depending on children’s ability. With correctly answered items (green line), there is no substantial relationship between a person’s ability and the time needed to solve an item. On the other hand, with incorrectly answered items (red line), the time required to answer an item increases with the ability level. Children with greater ability, therefore, take longer to answer an item incorrectly.
Figure 2. Predicted response time values according to Model A3 for correct (green) and incorrect (red) responses depending on children’s ability. With correctly answered items (green line), there is no substantial relationship between a person’s ability and the time needed to solve an item. On the other hand, with incorrectly answered items (red line), the time required to answer an item increases with the ability level. Children with greater ability, therefore, take longer to answer an item incorrectly.
Jintelligence 11 00063 g002
Figure 3. Predicted response time values according to the exploratory model A3 for correct (green line) and incorrect (red line) responses depending on children’s ability. The graph is divided into three panels based on different item difficulty levels. The observed difference in slopes of correct and incorrect responses is visibly weaker in Model A4. This further expands Model A3, as the proposition that the time required to answer incorrectly answered items (red line) increases with ability level is applicable only for moderately (e.g., difficulty = 0) and highly (e.g., difficulty = 2) difficult items.
Figure 3. Predicted response time values according to the exploratory model A3 for correct (green line) and incorrect (red line) responses depending on children’s ability. The graph is divided into three panels based on different item difficulty levels. The observed difference in slopes of correct and incorrect responses is visibly weaker in Model A4. This further expands Model A3, as the proposition that the time required to answer incorrectly answered items (red line) increases with ability level is applicable only for moderately (e.g., difficulty = 0) and highly (e.g., difficulty = 2) difficult items.
Jintelligence 11 00063 g003
Figure 4. Predicted response time values according to Model B3 as a function of ability–difficulty distance separated by whether the answer was correct (green) or incorrect (red). Regardless of the response correctness, items whose difficulty matches the participant’s ability level take the longest time to solve (with incorrect answers taking the longest). The relationship changes for very difficult and easy items—with these, correctly answered items take longer than the incorrect ones.
Figure 4. Predicted response time values according to Model B3 as a function of ability–difficulty distance separated by whether the answer was correct (green) or incorrect (red). Regardless of the response correctness, items whose difficulty matches the participant’s ability level take the longest time to solve (with incorrect answers taking the longest). The relationship changes for very difficult and easy items—with these, correctly answered items take longer than the incorrect ones.
Jintelligence 11 00063 g004
Table 3. Parameters of the models assessing distance–difficulty hypothesis interacting with the F > C phenomenon.
Table 3. Parameters of the models assessing distance–difficulty hypothesis interacting with the F > C phenomenon.
Model 0Model B1Model B2Model B3
95% CI 95% CI 95% CI 95% CI
coef.est. LLULest. LLULest. LLULest. LLUL
Fixed effects
interceptμ3.16***3.003.333.50***3.313.693.51***3.323.703.56***3.373.75
correct answer (FC)γ1 −0.02 −0.040.01−0.10***−0.13−0.07
ability–difficulty distanceγ4 −0.13***−0.14−0.12−0.13***−0.14−0.12−0.16***−0.17−0.15
distance × FCγ14 0.05***0.040.07
Random effects
person intercept variancevar(τj)0.07***0.060.080.07***0.060.080.07***0.060.080.07***0.060.08
item intercept variancevar(βi)0.20***0.120.330.25***0.150.420.25***0.150.420.26***0.160.45
residual variancevar(εij)0.37***0.360.380.34***0.330.350.34***0.330.350.34***0.330.34
Goodness of fit
conditional R2 0.417 0.527 0.525 0.542
marginal R2 0.000 0.082 0.081 0.092
log-likelihood −14,193 −13,563 −13,562 −13,537
AIC 28,395 27,136 27,136 27,088
BIC 28,425 27,174 27,181 27,142
Δχ2 (df) 1260.82 (1)*** 2.14 (1) 49.49 (1)***
Notes. coef.—coefficient, est.—estimate, CI—confidence interval, LL—lower limit, UL—upper limit, var—variance; *** p < .001.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tancoš, M.; Chvojka, E.; Jabůrek, M.; Portešová, Š. Faster ≠ Smarter: Children with Higher Levels of Ability Take Longer to Give Incorrect Answers, Especially When the Task Matches Their Ability. J. Intell. 2023, 11, 63. https://doi.org/10.3390/jintelligence11040063

AMA Style

Tancoš M, Chvojka E, Jabůrek M, Portešová Š. Faster ≠ Smarter: Children with Higher Levels of Ability Take Longer to Give Incorrect Answers, Especially When the Task Matches Their Ability. Journal of Intelligence. 2023; 11(4):63. https://doi.org/10.3390/jintelligence11040063

Chicago/Turabian Style

Tancoš, Martin, Edita Chvojka, Michal Jabůrek, and Šárka Portešová. 2023. "Faster ≠ Smarter: Children with Higher Levels of Ability Take Longer to Give Incorrect Answers, Especially When the Task Matches Their Ability" Journal of Intelligence 11, no. 4: 63. https://doi.org/10.3390/jintelligence11040063

APA Style

Tancoš, M., Chvojka, E., Jabůrek, M., & Portešová, Š. (2023). Faster ≠ Smarter: Children with Higher Levels of Ability Take Longer to Give Incorrect Answers, Especially When the Task Matches Their Ability. Journal of Intelligence, 11(4), 63. https://doi.org/10.3390/jintelligence11040063

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop