Abstract
Although previous research has consistently reported a positive association between mental speed and reasoning ability, it remains unclear whether the magnitude of this association depends on whether the reasoning test is administered with or without a time limit. In addition, it is unknown how mental speed task complexity affects the mental speed–reasoning association when the effects of time limitations in the reasoning test (labeled “speededness”) are controlled for. The present study examined these questions in a sample of 200 participants who completed the time-limited Culture Fair Test (CFT) and a Hick task with three levels of complexity to measure mental speed. Results showed that the latent correlation between mental speed and reasoning was slightly lower when the effect of speededness in reasoning was statistically controlled for. However, for both controlled and uncontrolled reasoning, the correlation with mental speed was of medium size and statistically significant. When reasoning was controlled for the effects of speededness, only complexity-related mental speed aspects were correlated with reasoning, whereas basic mental speed aspects were correlated with the speededness factor and unrelated to reasoning. These findings demonstrate that time limitations in reasoning tests and complexity in mental speed tasks affect the magnitude of the mental speed–reasoning association.
1. How Speededness of a Reasoning Test and the Complexity of Mental Speed Tasks Influence the Relation between Mental Speed and Reasoning Ability
Individuals with higher psychometric intelligence have been consistently reported to process information faster than individuals with lower intelligence scores (). Although current models of intelligence differentiate between “reaction and decision speed” on the one hand, and “processing speed” on the other one (), there is some evidence for the notion that both aspects of speed tap the same construct and their differentiation is primarily caused by the assessment through reaction times and number of items processed within a given time ().
Commonly, reaction/decision speed is measured by elementary cognitive tasks (ECTs), which are computer-based and involve a limited number of specific mental operations, placing low demands on motor preparation and execution (; ). Reaction times (RTs) in these ECTs are used as indicators for mental speed and show a quite consistent negative association with psychometric intelligence (; ). This relation seems to increase with increasing complexity of a given ECT (; ). However, the strength of the association and its role in explaining individual differences in intelligence remain a source of continuous debate (; ; ; ).
Many studies in this field use time-limited tests to measure intelligence or, more specifically, reasoning ability as a well-established proxy of general intelligence (; ). Such a time limitation is common for psychometric intelligence tests as it increases test efficiency and facilitates the testing of groups of participants in research. Surprisingly, however, only a very limited number of studies (described in detail below) investigated whether the time limitation and the resulting speededness of the test influence the relation between mental speed and intelligence (; ). In this context, speededness refers to the degree that test performance is affected by a limited testing time, i.e., when (at least some) participants do not have enough time to attempt all items or guess, especially on the last items of a test (; ). Hence, the relationship between psychometric intelligence and RT measures in ECTs could be overestimated when speed of information processing influences both the RT measures and, as a kind of method effect, the measure of intelligence ().
In one of the few existing studies examining this question, () asked participants to complete a set of computer-based ECTs with varying complexity. Additionally, one half of the sample completed a battery of time-limited intelligence tests, while the other half completed a battery of time-unlimited intelligence tests. The correlation between a mental speed factor extracted from the ECTs and intelligence was slightly (but not significantly) higher for time-unlimited than for time-limited intelligence tests (see also ). Thus, in this study, time limitations of the intelligence tests had no influence on the correlational relationship between mental speed and intelligence. It should be noted, however, that time-unlimited intelligence tests correlated most strongly with RT in the most complex ECTs, defined as those ECTs with the highest mean RTs. In contrast, the time-limited intelligence tests correlated most strongly with RTs in those ECTs that had the highest loadings on a general mental speed factor and presumably captured the largest variety of basic information-processing operations ().
In another study, a sample of 367 participants completed a battery of time-limited and a battery of time-unlimited reasoning tasks along with a battery of paper–pencil-based clerical speed tests (). Latent variables representing time-limited and time-unlimited reasoning ability as well as mental speed were extracted from the respective tests. Both time-limited and time-unlimited reasoning ability were related to mental speed. In contrast to (), however, the relation to mental speed was stronger for time-limited compared to time-unlimited reasoning ability. () concluded that time-limited reasoning scores are a compound of reasoning ability and mental speed, which leads to an artificial overestimation of the true relation between reasoning ability and mental speed.
The above-mentioned studies were very demanding in terms of data collection, with () requiring two large groups to complete either time-unlimited or time-limited tests and with () administering a very large battery of time-limited and time-unlimited tests to one group. A more parsimonious approach to isolate test speededness recently introduced by Schweizer and colleagues (, ; ) is based on confirmatory factor analysis (CFA). This approach uses bifactor modeling to separate a latent variable representing the effect of speededness from a latent variable representing reasoning as a measure of intelligence (see method section for a detailed statistical explanation). The advantage of this approach is that the intelligence assessment does not need to be presented under both time-limited and time-unlimited conditions. Since the effect of speededness is statistically controlled for in such a CFA model, reasoning ability can be estimated in a way that is purified from the effect of speededness on test performance ().
In one of the initial studies using this approach, () used Raven’s Advanced Progressive Matrices as a measure of reasoning ability with a time limitation and separated reasoning ability from the effect of speededness. An additional paper–pencil-based clerical speed measure, where specific target items presented among distractors had to be identified as fast as possible, correlated with the effect of speededness but not with reasoning ability. Thus, similar to ’s () results, this finding suggests that the time limitation of reasoning tests leads to an overestimation of the relation between mental speed and reasoning ability or intelligence more generally.
However, it is still not clear whether clerical speed measures in paper–pencil tests (e.g., the number of correctly identified targets among distractors on a paper sheet in a given time) functionally tap the same ability as RT measures in ECTs (; ). Therefore, the first research question of the present study is to assess how the effect of speededness affects the relationship between reasoning ability and mental speed represented by RTs in the Hick task as a well-established ECT. For this purpose, the relationship between reasoning ability in Cattell’s time-limited Culture Fair Test (CFT) and mental speed is computed twice. Initially, the effect of speededness on CFT performance is not controlled for; however, subsequently, it is statistically controlled for by means of a bifactor CFA model. Based on the findings by () and (), we expect a weaker association between reasoning and mental speed when speededness is controlled for.
The second research question of the present study is to assess how increasing the complexity of the ECT (operationalized through an increasing number of choices in the Hick task) affects the relationship between mental speed and reasoning ability (measured with the CFT) when reasoning ability is controlled for the effect of speededness. To this end, we use a second bifactor CFA model to statistically separate basic and complexity-related mental speed aspects from the RTs in the Hick task. This model is then combined with the bifactor model separating reasoning ability and speededness as explained above. () reported that RTs from more complex ECTs were more closely related to time-unlimited intelligence scores than RTs from less complex ECTs. Proceeding from this finding, we expect that complexity-related aspects of mental speed are more highly correlated with reasoning (controlled for the effect of speededness) than basic mental speed.
Taken together, the present study seeks to clarify the role of mental speed in psychometric intelligence, which has been a longstanding issue in the field. In addition, the study is relevant for applied settings in which intelligence is typically assessed, as it may affect the interpretation of test results.
2. Method
2.1. Participants
Participants were 67 men and 133 women ranging in age from 17 to 30 years (M = 22.6, SD = 2.5 years), who were also subjects in the study by () on another research question. A total of 149 participants were university students and 51 participants did not have secondary education. All participants reported normal or corrected-to-normal vision. The study was approved by the local ethics committee. Prior to testing, participants gave informed consent.
2.2. Measure of Intelligence
The German adaptation of Cattell’s Culture Fair Test 20-R (CFT; ) was used with a reliability of .92 (). The CFT is a (not culturally independent) measure of general reasoning ability and fluid intelligence (; ; ) and consists of four subtests presented in a multiple-choice format with five response options for each item.
In the subtest Series, participants chose the option best completing a series of three figures. In the Classifications subtest, participants were asked to identify the deviant in five figures. In the Matrices subtest, the appropriate figure to complete a matrix had to be found. In the Topologies subtest, participants identified the one figure in which a given number of dot(s) could be placed with the same topological relationship to the other parts as given in a reference configuration.
The first three subtests consisted of 15 items and Topologies consisted of 11 items. In accordance with the standard instructions provided in the test manual, the time limit was set to four minutes for Series and Classifications, and to three minutes for Matrices and Topologies; all items and subtests were presented in the same order for each participant. Responses on all items were scored as 1 (correct option chosen) or 0 (incorrect or no option chosen). In addition, IQ scores were calculated based on the norms in the CFT () in order to describe the intelligence level of the sample.
2.3. Measurement of Mental Speed
The Hick task was adapted from () and programmed in E-Prime Professional 2.0 (). Stimuli were rectangles with a height of 1.5 cm and a width of 1.9 cm as well as a plus sign with a height and a width of 0.8 cm. They were presented in white against a black background on a 19-inch monitor screen (Belinea, Model 101902) with a refresh rate of 60 Hz. The distance between the participants’ heads and the monitor screen was about 60 cm. The response pad was a Cedrus RB Box, allowing us to record RTs with a resolution of ±1 ms.
The task consisted of three conditions (0-bit, 1-bit, and 2-bit conditions), presented in three blocks of increasing task demands. Each condition consisted of 32 trials preceded by 8 practice trials. In each trial of the 0-bit condition, a black screen was presented for one second. Then, a rectangle appeared in the center of the monitor screen and after a foreperiod of 1000, 1333, 1666, or 2000 ms, a plus sign was presented in the center of the rectangle. Participants were asked to press a designated key on the response pad as soon as the plus sign appeared. The plus sign remained on the screen up until the key press and then disappeared.
The trials of the 1-bit and 2-bit conditions were similar to the 0-bit condition, but two rectangles were presented in the 1-bit condition, and four rectangles in the 2-bit condition. In each trial, the plus sign appeared in one of the rectangles according to a random sequence, which was identical for all participants. Participants were instructed to press a designated key corresponding to the position of the plus sign as soon as the sign appeared. In all conditions, key presses before stimulus appearance and incorrect responses were registered as errors and followed by a 200 ms feedback tone.
The instructions were presented on the screen and encouraged participants to respond as fast as possible but not to sacrifice accuracy. As the dependent variable, the mean latency of correct responses was recorded for each task condition. To minimize the probability of outliers, for each participant, the very first item and the item with the longest response latency were excluded from analysis. Thereafter, a visual outlier control was performed, identifying response latencies under 100 ms and over 1100 ms as outliers and excluding them from further analysis. Less than 0.21% of responses per condition were outliers. In addition, to avoid outliers in mean response latencies, after a visual inspection, mean response latencies were winsorized to 450 ms (n = 4) in the 0-bit condition, to 500 ms in the 1-bit condition (n = 2), and to 600 ms in the 2-bit condition (n = 3).
2.4. Time Course of the Study
Participants first completed the CFT, then completed the Hick task and two additional experimental tasks unrelated to the present study. The order of the three experimental tasks was balanced across participants. Each participant was tested individually.
2.5. Statistical Analyses
Analyses were conducted using the lavaan package () in the statistical Software R (version 3.2.0). Unless otherwise stated, the reported parameters were standardized.
In a first step, we computed the correlational relationship between reasoning ability (as a proxy of psychometric intelligence) and mental speed without controlling for the effect of speededness. From the items of each CFT subtest, a latent variable representing reasoning ability was extracted using (essentially tau-equivalent) one-factor models. These first-order latent variables were combined to create a second-order latent variable of reasoning ability. For this second-order latent variable, which was uncontrolled for the effect of speededness, we computed the correlation with mental speed as a latent variable extracted from RTs in the three conditions of the Hick task.
In a second step, it was probed whether the effect of speededness could be identified and represented by a latent variable in the four CFT subtests by means of bifactor models. From the four latent variables representing (first-order) reasoning ability as well as the four latent variables representing the (first-order) effect of speededness, second-order latent variables of reasoning ability and the effect of speededness were derived. Again, the correlations between mental speed as a latent variable extracted from RTs in the three Hick task conditions and (second-order) reasoning ability as well as the (second-order) effect of speededness were computed. The correlations between mental speed and reasoning ability were then compared to the respective correlations from the first step, where the effect of speededness was not controlled for.
The CFA modeling of reasoning ability and the effect of speededness need some further explanation. Variances and probability-based covariances between the binary items (1 = correct, 0 = incorrect or no response) served as input for the analyses of the CFT subtests (). This input has been shown to outperform tetrachoric correlations in a previous simulation study, especially for sample sizes smaller than n = 1000 (). For the (essentially tau-equivalent) one-factor models in the first step, one latent variable representing reasoning ability was extracted from the items of each CFT subtest with the factor loadings being fixed to the same value (“1”) assuming that all items measured reasoning in the same way.
For the bifactor models in the second step, the effect of speededness was added as a latent variable to the tau-equivalent model using the procedure suggested by (). More specifically, these authors proceeded from the assumption of normally distributed individual differences in the speededness of test taking, which should lead to an approximatively cumulative increase in omissions. When at least one omission was recorded for an item, it was assumed to be affected by speededness. In order to depict the assumed cumulative increase, the factor loadings of the latent variable reflecting the effect of speededness were fixed to follow the course of the increasing logistic function from the first item in a subtest affected by speededness (i.e., the first item for which at least one response was omitted and not followed by responses on subsequent items) to the last item (). Thus, factor loadings were fixed according to the following formula:
where i is the position of an item within all items, which were not reached by at least one participant, and j is a constant that may be selected to optimize model fit (). As in previous work with the CFT 20-R (), j was set to 1.5 for all subscales.
To bridge the gap between binary manifest and continuous latent variables as well as their distributional differences, the factor loading of each item was weighted by the standard deviation (SD) of the respective item ():
The two latent variables in the bifactor models were assumed to be independent from each other and, thus, their correlation was set to zero.
To investigate the second research question of how ECT complexity affects the relation between mental speed and reasoning ability as well as the effect of speededness, the effect of increasing task complexity on RTs had to be modeled. For this purpose, a fixed-links modeling approach was used to extract two latent variables from RTs in the Hick task (; ). To represent individual differences in RTs increasing as a function of task complexity, the first latent variable had linearly increasing factor loadings of 1, 2, and 4. The second latent variable captured individual differences in RTs, which did not vary between the Hick task conditions. Therefore, the factor loadings of this latent variable were fixed to “1”. The correlation between the two latent variables was set to “0”. After having separated basic and complexity-related mental speed for the Hick task, correlations between these latent variables and reasoning ability as well as the effect of speededness were computed.
The confirmatory factor models were evaluated as good (or acceptable) when the root mean square error of approximation (RMSEA) was below 0.08 () and the Standardized Root Mean Square Residual (SRMR) below 0.10 (). The comparative fit index (CFI) was not used for the analyses on the CFT, as it is non-informative when the investigated model is compared to a too-well-fitting baseline model (). () suggested a RMSEA below .158 as being indicative of a too-well-fitting baseline model, which was the case with all four baseline models of the CFT (Series: RMSEA = .083; Classification: RMSEA = .074; Matrices: RMSEA = .094; Topologies: RMSEA = .110). To compare models, the Akaike Information Criterion (AIC) was used. A lower AIC indicates a better model with respect to both model fit and parsimony.
3. Results
Descriptive statistics of all measures are presented in Table 1. All variables can be assumed to be approximately normally distributed as skewness was below 2 and kurtosis below 7 (; ). The sum of correct responses in the four CFT scores was transferred to age-stratified IQ scores in accordance with the manual. With M = 110, the mean IQ of the present sample was higher than the population mean of 100, but the standard deviation (SD = 14) was close to the standard deviation of 15 in the population.
Table 1.
Mean (M), Standard Deviation (SD), Minimum (Min), Maximum (Max), Skewness, and Kurtosis of the number of correct responses in the four CFT subtests as well as of the reaction times in the three conditions of the Hick task and the Flanker task.
The manipulation of task complexity in the Hick task was successful as indicated by the systematic increase in the mean RT across the three Hick task conditions, F(1.55,309) = 953.11, p < .001, η2 = .83. The degrees of freedom were adjusted due to violations of sphericity (). Bonferroni-corrected t-tests revealed that all three task conditions differed significantly from each other, all ts(199) > 13, p < .001.
3.1. The Relationship between Mental Speed and Reasoning Ability When Controlled and When Not Controlled for the Effect of Speededness
To answer the first research question, we investigated the relationship between mental speed and reasoning ability uncontrolled for the effect of speededness. Four latent variables were extracted from the items of the four subtests of the CFT by means of tau-equivalent models, which were aggregated in terms of a second-order latent variable to represent reasoning ability uncontrolled for the effect of speededness (see Figure 1). The first three items of the Series subtest, the first item of Classification, and the first four items of Matrices were excluded from the analyses because they were solved by at least 98.5% of the participants; thus, their variances were too restricted. This model described the CFT data well, χ2(1120) = 1584.96, RMSEA = .046, SRMR = .084, AIC = 7161.42.
Figure 1.
Measurement model of the four CFT subtests without controlling for the effect of speededness. For each subtest, one latent variable with constant factor loadings was extracted. From those four inductive reasoning latent variables, a second-order latent variable was extracted to represent inductive reasoning across the four CFT subtests.
Additionally, a latent variable from RTs in the three conditions of the Hick task was extracted to represent mental speed. The fit of this model is not reported as one latent variable extracted from three conditions of a task has no degrees of freedom. The latent variable was added to the measurement model of the CFT uncontrolled for the effect of speededness. In this model (Model 1), reasoning ability correlated negatively and significantly with the latent variable derived from RTs in the Hick task, r = −.42, p = .001, as presented in Panel A of Figure 2. The model showed an acceptable model fit, χ2(1263) = 1776.50, RMSEA = .045, SRMR = .083, AIC = 7648.61.
Figure 2.
Latent structural equation models assessing the relationship (1) between mental speed and reasoning ability uncontrolled for complexity and speededness (A), (2) between mental speed, reasoning ability controlled for speededness, and speededness (B), and (3) between basic and complexity-related mental speed, reasoning ability controlled for speededness, and speededness (C). * p < .05, ** p < .01, *** p < .001.
To examine how the correlation between reasoning ability and mental speed changes when the reasoning measure is controlled for the effect of speededness, a second measurement model was built as an alternative description of the CFT data. This model considered the effect of speededness in the four CFT subtests. Thus, from the responses on each of the four subtests, two latent variables were extracted by means of bifactor models. For the first latent variable, the factor loadings were fixed to 1 while, for the second latent variable, factor loadings were fixed in an increasing manner from the first item affected by speededness to the last item (see Section 2.5). The four first-order latent variables representing subtest-specific reasoning ability were combined to a second-order latent variable (overall reasoning ability). Analogously, the four first-order latent variables representing the effect of speededness in the four subtests were combined to create a second-order latent variable (overall speededness effect). This second measurement model is presented in Figure 3. It described the data well, χ2(1112) = 1462.47, RMSEA = .04, SRMR = .077, AIC = 7054.94, and better than the first measurement model of the CFT as can be seen from the lower AIC. It should be noted that the variances of the second-order latent variables were significant on the 1% level.
Figure 3.
Measurement model of the four CFT 20-R subtests. For each subtest, one latent variable with constant factor loadings (and a fading-out effect at the end) represents reasoning and one latent variable with logistically increasing factor loadings represents the effect of speededness. From both the four inductive reasoning and the four speededness latent variables, second-order latent variables were extracted to represent inductive reasoning and the effect of speededness across the four CFT 20-R subtests. Note: S = Series; C = Classifications; M = Matrices; T = Topologies.
As for Model 1 in the previous section, we added the latent variable of mental speed extracted from the RTs in the Hick task to the second CFT measurement model. The resulting model (Model 2) showed an acceptable model fit, χ2(1254) = 1654.31, RMSEA = .040, SRMR = .076, AIC = 7544.43. As illustrated in Panel B of Figure 2, reasoning ability still correlated significantly with mental speed in the Hick task, r = −.32, p < .01, and the magnitude of the association was similar to that of the model without the speededness factor (r = −.42). In addition, the correlation between mental speed in the Hick task and the overall effect of speededness in Model 2 was statistically significant, r = −.25, p < .05.
3.2. The Effect of Increasing Task Complexity on the Relationship between Mental Speed and Reasoning Ability Controlled for the Effect of Speededness
The second research question was about the effect of increasing complexity of the ECT on the relationship between mental speed and reasoning ability when the measure of reasoning ability was controlled for the effect of speededness. To answer this question, we isolated complexity-related aspects of mental speed in RTs of the Hick task from basic aspects of mental speed by means of the above-described fixed-links model. This measurement model of RTs fit the data well, χ2(1) = 0.793, RMSEA = .001, SRMR = .026, AIC = 506.64. The latent variance of both latent variables was significantly different from zero (Hick basic speed ϕ = 0.105, p < .001; Hick complex speed ϕ = 0.012, p < .001).
The measurement model of the Hick task was combined with the measurement model of the CFT data, where the effect of speededness was controlled for. The resulting structural equation model yielded an adequate fit, χ2(1253) = 1649.88, RMSEA = .04, SRMR = .076, AIC = 7541.99. The structural core of this model is shown in Panel C of Figure 2. Unstandardized and standardized factor loadings are available in the supplementary material. Reasoning ability controlled for the effect of speededness was not related to basic aspects of mental speed, r = −.07, p = .59, but to the complexity-related aspects of mental speed in the Hick task, r = −.48, p < .01. The effect of speededness, on the contrary, was related to basic aspects of mental speed, r = −.29, p <.05, but not to its complexity-related aspects, r = −.02, p = .89, (see Figure 2C).
4. Discussion
The aims of the present study were, first, to investigate the influence of the effect of speededness in a reasoning test (CFT) on the relationship between reasoning ability and mental speed and, second, to evaluate the relevance of ECT (Hick task) complexity in this context. Our expectations were twofold: First, we expected that the association between reasoning ability and mental speed would become weaker when speededness due to the time limitation in the CFT was controlled for. Second, we expected that, when speededness in the CFT was controlled for, the association between reasoning ability and mental speed would be higher for complexity-related aspects of mental speed than for basic mental speed.
Replicating previous studies, CFT scores and RTs in the Hick task were substantially correlated when ignoring the effect of speededness in the CFT and the different levels of complexity in the Hick task (r = −.42). However, when speededness caused by the time limitation in the CFT subtests was statistically separated from reasoning ability using bifactor modeling (, ; ), the relationship between reasoning ability and mental speed decreased only slightly (to r = −.32), with the effects in both models being of medium size. These results speak against our first hypothesis but are in line with () who reported that the correlational relationship between intelligence and mental speed did not substantially change depending on whether intelligence was assessed under time-limited or time-unlimited conditions. Our result differs from () who compared the correlation between mental speed and reasoning ability in conditions of a time-unlimited and a time-limited administration of reasoning tests. In their study, the relationship between reasoning ability and mental speed was significantly weaker when the assessment of reasoning ability was time-unlimited compared to the time-limited conditions. However, in both conditions of ’s () study, the correlation between reasoning ability and mental speed was significant, which corresponds to our finding that reasoning ability and mental speed are related even when speededness is controlled for. Thus, the general association between mental speed and reasoning ability (or intelligence) reported in the literature cannot be attributed to the effects of speededness alone. However, as the effect of speededness in the present study was related to mental speed (in almost the same strength as reasoning ability), failing to control for it might easily lead to an overestimation of the relationship between mental speed and reasoning ability or intelligence as also emphasized by ().
Regarding the second hypothesis, reasoning ability (controlled for speededness) was unrelated to basic aspects of mental speed (r = −.07), but showed a robust correlation with complexity-related mental speed (r = −.48). Our hypothesis was, thus, supported. The separation of complexity-related from basic aspects of speed of information processing sheds further light on the question of how mental speed and reasoning ability are related when the effect of speededness is controlled for. Specifically, reasoning ability (controlled for speededness) appears to be associated only to those mental speed aspects that are required for handling the increasing demands on decision making in the Hick task. This result resembles ’s () finding that time-unlimited intelligence tests correlated most strongly with RTs in the most complex ECTs (as defined on the basis of mean RTs). The effect of speededness, however, was only related to basic and not to complexity-related aspects of mental speed in the present study. These basic aspects of mental speed probably reflect processes of sensory encoding and motor execution as well as individual differences in alertness, fatigue, or motivation (; ) that impact performance in both the Hick task and the CFT.
Thus, the answer to the question of whether a time limitation in the administration of an intelligence test impacts the relationship between intelligence and mental speed probably depends on the complexity of the ECT to measure mental speed. In the case of a very simple, undemanding ECT, a time limitation in the reasoning test probably leads to an overestimation of the mental speed–reasoning ability relationship. RTs from more complex and cognitively demanding ECTs, on the contrary, correlate with reasoning ability even more strongly when no time limitation is used during the administration of the reasoning test or when the effect of speededness is controlled for.
A potential explanation for this differential relationship between basic and complexity-related aspects of mental speed, on the one hand, and reasoning ability and the effect of speededness, on the other hand, has been proposed by (). According to their “two-functions view”, mental speed tasks and reasoning tasks can be organized along a complexity dimension with mental speed (and low working memory demands) at the one pole and reasoning (with high working memory and low mental speed demands) at the other pole (). The correlation between two cognitive measures is determined by how close they are to each other on the complexity dimension. Reasoning ability controlled for speededness and basic mental speed controlled for complex mental speed are not expected to correlate as they are at the extreme poles of the complexity dimension. This is what we found in the present study; basic aspects of mental speed in the Hick task were not related to reasoning ability controlled for speededness. According to the “two-functions view”, however, a correlation between mental speed and reasoning is expected in two cases: first, when speededness is introduced into the reasoning measure, because in this case, reasoning also contains aspects of basic speed, and second, when the task demands in the ECT are increased and RTs not only represent basic but more complex mental speed. Both these assumptions of ’s () two-functions view were empirically confirmed in the present study. First, the effect of speededness was related to basic mental speed extracted from the Hick task and, second, complexity-related aspects of mental speed extracted from the Hick task were related to reasoning ability controlled for the effect of speededness.
Taken together, the well-established relationship between mental speed and reasoning ability might be influenced by time limitations of tests assessing reasoning ability. The present study demonstrates that this influence can be controlled for by means of bifactor CFA models. Mental speed is related to both reasoning ability and the effect of speededness. However, the effect of speededness in the reasoning measure is related to basic aspects of mental speed while reasoning is substantially related to complexity-related aspects of mental speed.
Future studies are needed to examine whether time limitations and resulting speededness effects also affect the magnitude of the association between mental speed and other tests of reasoning ability as well as broader assessments of psychometric intelligence (e.g., ). Our findings may also be relevant when choosing and interpreting reasoning or intelligence tests in applied contexts. Specifically, the predictive validity of reasoning scores for outcomes such as academic performance might change when the effects of speededness are controlled for. Implications of the present research for everyday-life correlates of intelligence, however, need to be examined explicitly.
It should be noted that the latent variable representing the effect of speededness is probably not a pure measure of this effect. Item processing does not only change from the earlier to the later items due to time limitations but also due to other processes. For example, individual differences in learning the rules underlying the items develop from item to item (; ) and effects of increasing fatigue are also plausible (). Individual differences in managing proactive interference () have been described to influence performance on Raven’s Advanced Progressive Matrices () and can be expected to increase from earlier to later items. A systematic change during test completion has also been reported for processing strategies (). All of these processes might have contributed to the latent variable, which we call the effect of speededness. The logistic function underlying the course of factor loadings was used to depict this effect as purely as possible, but it seems unlikely that this procedure completely excludes other processes from this latent variable. However, regardless of the purity of the latent variable representing the effect of speededness, we assume that the influence of speededness on the latent reasoning variable could be successfully controlled for.
Regarding the decomposition of individual differences in mental speed measures, it would be interesting to compare the fixed-links modeling approach of the present study with other contemporary methods, such as the diffusion model (), ex-Gaussian distribution (), or the linear ballistic accumulator model ().
Certainly, the present findings should be replicated and extended by using other ECTs or Hick task conditions with higher complexity than that in the present study, in which the most complex condition (2-bit Hick task) was still quite simple. Finally, our findings are based on a sample with a relatively high mean level of reasoning (see above). It would be interesting to replicate and generalize the present findings in samples with a broader range of intellectual abilities.
Supplementary Materials
The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/jintelligence11050089/s1, Table S1: Unstandardized and standardized factor loadings of the model with basic and complexity-related mental speed, reasoning ability controlled for speededness, and speededness. Data, Script.
Author Contributions
Conceptualization, N.B., K.S. and S.J.T.; methodology, N.B. and S.J.T.; formal analysis, N.B.; writing—original draft, N.B.; writing—review and editing, K.S. and S.J.T.; visualization, N.B.; supervision, S.J.T. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Institutional Review Board Statement
The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the ethical review committee of the Faculty of Human Sciences at the University of Bern (2013-8-504570, approved August 2013).
Informed Consent Statement
Informed consent was obtained from all subjects involved in the study.
Data Availability Statement
The data presented in this study are available in Supplementary Material.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Anderson, Mike, and Jeff Nelson. 2005. Individual differences and cognitive models of the mind: Using the differentiation hypothesis to distinguish general and specific cognitive processes. In Measuring the Mind: Speed, Control, and Age. Edited by Duncan John, Phillips Louise and McLeod Peter. Oxford: Oxford University Press, pp. 89–114. [Google Scholar]
- Borter, Natalie, Stefan J. Troche, and Thomas H. Rammsayer. 2018. Speed-and accuracy-related measures of an intelligence test are differentially predicted by the speed and accuracy measures of a cognitive task. Intelligence 71: 1–7. [Google Scholar] [CrossRef]
- Borter, Natalie, Annik E. Völke, and Stefan J. Troche. 2020. The development of inductive reasoning under consideration of the effect due to test speededness. Psychological Test and Assessment Modeling 62: 344–58. [Google Scholar]
- Brown, Scott D., and Andrew Heathcote. 2008. The simplest complete model of choice response time: Linear ballistic accumulation. Cognitive Psychology 57: 153–78. [Google Scholar] [CrossRef] [PubMed]
- Bunting, Michael. 2006. Proactive interference and item similarity in working memory. Journal of Experimental Psychology: Learning, Memory, and Cognition 32: 183–96. [Google Scholar] [CrossRef]
- Byrne, Barbara M. 2010. Structural Equation Modeling with AMOS: Basic Concepts, Applications, and Programming. New York: Routledge. [Google Scholar]
- Carroll, John B. 1993. Human Cognitive Abilities. Cambridge: Cambridge University Press. [Google Scholar]
- Cattell, Raymond B. 1963. Theory of fluid and crystallized intelligence: A critical experiment. Journal of Educational Psychology 54: 1–22. [Google Scholar] [CrossRef]
- Danthiir, Vanessa, Richard D. Roberts, Ralf Schulze, and Oliver Wilhelm. 2005. Mental speed. In Handbook of Understanding and Measuring Intelligence. London and New Dehli: Sage, pp. 27–46. [Google Scholar]
- Doebler, Philipp, and Barbara Scheffler. 2016. The relationship of choice reaction time variability and intelligence: A meta-analysis. Learning and Individual Differences 52: 157–66. [Google Scholar] [CrossRef]
- Estrada, Eduardo, Francisco J. Román, Francisco J. Abad, and Roberto Colom. 2017. Separating power and speed components of standardized intelligence measures. Intelligence 6: 159–68. [Google Scholar] [CrossRef]
- Gonthier, Corentin, and Jean-Luc Roulin. 2020. Intraindividual strategy shifts in Raven’s matrices, and their dependence on working memory capacity and need for cognition. Journal of Experimental Psychology: General 149: 564–79. [Google Scholar] [CrossRef]
- Greenhouse, Samuel W., and Seymour Geisser. 1959. On methods in the analysis of profile data. Psychometrika 24: 95–112. [Google Scholar] [CrossRef]
- Hair, Joseph F., William C. Black, Barry J. Babin, and Rolph E. Anderson. 2010. Multivariate Data Analysis, 7th ed. New York: Pearson Educational International. [Google Scholar]
- Heitz, Richard P., Nash Unsworth, and Randall W. Engle. 2005. Working memory capacity, attention control, and fluid intelligence. In Handbook of Understanding and Measuring Intelligence. Newbury Park: Sage Publications, Inc., pp. 61–77. [Google Scholar]
- Jensen, Arthur R. 1998. The g Factor: The Science of Mental Ability. Westport: Praeger. [Google Scholar]
- Jensen, Arthur R. 2006. Clocking the Mind: Mental Chronometry and Individual Differences. Amsterdam: Elsevier. [Google Scholar]
- Johnson, Wendy, Jan te Nijenhuis, and Thomas J. Bouchard. 2008. Still just 1 g: Consistent results from five test batteries. Intelligence 36: 81–95. [Google Scholar] [CrossRef]
- Kan, Kees-Jan, Rogier A. Kievit, Conor Dolan, and Han van der Maas. 2011. On the interpretation of the CHC factor Gc. Intelligence 39: 292–302. [Google Scholar] [CrossRef]
- Kenny, David A. 2015. Measuring Model Fit. Available online: https://davidakenny.net/cm/fit.htm (accessed on 11 August 2022).
- Lozano, José H., and Javier Revuelta. 2020. Investigating operation-specific learning effects in the Raven’s Advanced Progressive Matrices: A linear logistic test modeling approach. Intelligence 82: 101468. [Google Scholar] [CrossRef]
- Lu, Ying, and Stephen G. Sireci. 2007. Validity issues in test speededness. Educational Measurement: Issues and Practice 26: 29–37. [Google Scholar] [CrossRef]
- May, Cynthia P., Lynn Hasher, and Michael J. Kane. 1999. The role of interference in memory span. Memory and Cognition 27: 759–67. [Google Scholar] [CrossRef] [PubMed]
- McGrew, Kevin S. 2009. CHC theory and the human cognitive abilities project: Standing on the shoulders of the giants of psychometric intelligence research. Intelligence 37: 1–10. [Google Scholar] [CrossRef]
- Neubauer, Aljoscha C., Christa Bauer, and Gerald Höller. 1992. Intelligence, attention, motivation and speed-accuracy trade-off in the Hick paradigm. Personality and Individual Differences 13: 1325–32. [Google Scholar] [CrossRef]
- Pahud, Olivier, Thomas H. Rammsayer, and Stefan J. Troche. 2018. Elucidating the functional relationship between speed of information processing and speed-, capacity-, and memory-related aspects of psychometric intelligence. Advances in Cognitive Psychology 13: 3–13. [Google Scholar] [CrossRef]
- Rammsayer, Thomas H., Olivier Pahud, and Stefan J. Troche. 2017. Decomposing the functional relationship between speed of information processing in the Hick paradigm and mental ability: A fixed-links modeling approach. Personality and Individual Differences 118: 17–21. [Google Scholar] [CrossRef]
- Ren, Xuezhu, Karl Schweizer, and Fen Xu. 2013. The sources of the relationship between sustained attention and reasoning. Intelligence 41: 51–58. [Google Scholar] [CrossRef]
- Ren, Xuezhu, Tengfei Wang, Michael Altmeyer, and Karl Schweizer. 2014. A learning-based account of fluid intelligence from the perspective of the position effect. Learning and Individual Differences 31: 30–35. [Google Scholar] [CrossRef]
- Ren, Xuezhu, Tengfei Wang, Sumin Sun, Mi Deng, and Karl Schweizer. 2018. Speeded testing in the assessment of intelligence gives rise to a speed factor. Intelligence 66: 64–71. [Google Scholar] [CrossRef]
- Rosseel, Yves. 2012. Lavaan: An R package for structural equation modeling. Journal of Statistical Software 48: 1–36. [Google Scholar] [CrossRef]
- Schermelleh-Engel, Karin, Helfried Moosbrugger, and Hans Müller. 2003. Evaluating the fit of structural equation models: Tests of significance and descriptive goodness-of-fit measures. Methods of Psychological Research Online 8: 23–74. [Google Scholar]
- Schmitz, Florian, and Oliver Wilhelm. 2016. Modeling mental speed: Decomposing response time distributions in elementary cognitive tasks and correlations with working memory capacity and fluid intelligence. Journal of Intelligence 4: 13. [Google Scholar] [CrossRef]
- Schmitz, Florian, and Oliver Wilhelm. 2019. Mene Mene Tekel Upharsin: Clerical Speed and Elementary Cognitive Speed are Different by Virtue of Test Mode Only. Journal of Intelligence 7: 16. [Google Scholar] [CrossRef]
- Schneider, Walter, Amy Eschman, and Anthony Zuccolotto. 2002. E-Prime User’s Guide. Pittsburgh: Psychology Software Tools Inc. [Google Scholar]
- Schweizer, Karl. 2012. The position effect in reasoning items considered from the CFA perspective. The International Journal of Educational and Psychological Assessment 11: 44–58. [Google Scholar]
- Schweizer, Karl. 2013. A threshold-free approach to the study of the structure of binary data. International Journal of Statistics and Probability 2: 67–76. [Google Scholar] [CrossRef]
- Schweizer, Karl, Xuezhu Ren, and Tengfei Wang. 2015. A comparison of confirmatory factor analysis of binary data on the basis of tetrachoric correlations and of probability-based covariances: A simulation study. In Quantitative Psychology Research. Edited by Roger E. Millsap, Daniel M. Bolt, L. Andries van der Ark and Wen-Chung Wang. Berlin/Heidelberg: Springer, pp. 273–92. [Google Scholar]
- Schweizer, Karl, Siegbert Reiss Xuezhu Ren, Tengfei Wang, and Stefan J. Troche. 2019a. Speed effect analysis using the CFA framework. Frontiers in Psychology 10: 239. [Google Scholar] [CrossRef]
- Schweizer, Karl, Siegbert Reiss, and Stefan Troche. 2019b. Does the effect of a time limit for testing impair structural investigations by means of confirmatory factor models? Educational and Psychological Measurement 79: 40–64. [Google Scholar] [CrossRef]
- Schweizer, Karl, Stefan Troche, Thomas Rammsayer, and Florian Zeller. 2021. Inductive reasoning and its underlying structure: Support for difficulty and item position effects. Advances in Cognitive Psychology 17: 274–83. [Google Scholar] [CrossRef]
- Sheppard, Leah D., and Philip A. Vernon. 2008. Intelligence and speed of information-processing: A review of 50 years of research. Personality and Individual Differences 44: 535–51. [Google Scholar] [CrossRef]
- Stankov, Lazar, and Richard D. Roberts. 1997. Mental speed is not the “basic” process of intelligence. Personality and Individual Differences 22: 69–84. [Google Scholar] [CrossRef]
- Troche, Stefan J., Philipp Thomas, Duje Tadin, and Thomas H. Rammsayer. 2018. On the relationship between spatial suppression, speed of information processing, and psychometric intelligence. Intelligence 67: 11–18. [Google Scholar] [CrossRef]
- Vernon, Philip A., and Lida Kantor. 1986. Reaction time correlations with intelligence test scores obtained under either timed or untimed conditions. Intelligence 10: 315–30. [Google Scholar] [CrossRef]
- Vernon, Philip A., Sue Nador, and Lida Kantor. 1985. Reaction times and speed-of-processing: Their relationship to timed and untimed measures of intelligence. Intelligence 9: 357–74. [Google Scholar] [CrossRef]
- Weiß, Rudolf H. 2006. CFT 20-R—Grundintelligenztest Skala 2. Göttingen: Hogrefe. [Google Scholar]
- Wilhelm, Oliver, and Ralf Schulze. 2002. The relation of speeded and unspeeded reasoning with mental speed. Intelligence 30: 537–54. [Google Scholar] [CrossRef]
- Zeller, Florian, Siegbert Reiß, and Karl Schweizer. 2018. Does speededness in collecting reasoning data lead to a speed factor? European Journal of Psychological Assessment 36: 96–104. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).