1. Introduction
The structure of cognitive test batteries, such as the WAIS-IV or the WJ-IV are typically modeled with factor analytic procedures (
Dombrowski et al. 2017) or covariance structural modeling (
Benson et al. 2010). These methods account for the covariance between test scores with latent variables. Network models have been proposed as solutions to the problem of modeling psychological constructs such as the symptoms of psychopathology (
Cramer et al. 2010) and more recently, cognitive test scores (
Kan et al. 2019;
Schmank et al. 2019). The network models proposed by these investigators do not make use of latent variables, but rather model the structure of test batteries in terms of manifest variables (i.e., observed test scores).
Psychometric network models represent correlation matrices as a graph in which each variable is a node and each correlation an edge. Both
Kan et al. (
2019) and
Schmank et al. (
2019) used reduced partial correlation matrices of WAIS-IV test scores. The rational for the use of partial correlations is that one can rule out the relationship between any pair of variables as being due to other variables in the analysis and thus more readily infer causation (
Epskamp and Fried 2018). These partial correlation matrices were further reduced in the sense that “spurious or false positives” edges were removed by use of model selection procedures in the qgraph package from
R (i.e., they created sparse correlation matrices with many values set to zero).
Both
Kan et al. (
2019) and
Schmank et al. (
2019) compared model fit indices for network models with those for traditional latent variable models and found the former to provide a better fit.
Kan et al. (
2019) suggested that their results provided support for a mutualism model (
van der Maas et al. 2006). According to
Kan et al. (
2019), mutualism assumes that reciprocal causation occurs among cognitive abilities during development. The positive manifold (i.e., the positive correlation between cognitive test scores) is due to these interactions rather than
g.
Schmank et al. (
2019) suggested that network models are compatible with process overlap theory (
Kovacs and Conway 2016). Process overlap theory holds that the positive manifold is due to multiple (mostly generalist executive) processes that determine performance on any specific test.
One problem in attempting to compare network models with traditional latent variable models of cognitive test batteries is that the two deal with different aspects of test correlations. The network models discussed used partial correlation matrices while latent variable models used uncorrected correlation matrices. It is possible to use uncorrected correlation matrix for network analysis, as shown in Table 2 of
Kan et al. (
2019), but this model is saturated (i.e., there are no degrees of freedom remaining after estimating model weights). As such, this model cannot be compared to models with residual degrees of freedom. Indeed, the network model is basically an alternative representation of the correlation matrix and the saturated model accounts for (i.e., is identical to) all of the covariance in the matrix. This is another reason that lesser edges need to be eliminated if comparisons with structural equation models are to be made.
From the description of the methods by
Kan et al. (
2019) and
Schmank et al. (
2019), it is not clear what type of correlations were used to compare network and latent variable models. However, from the R code provided by
Schmank et al. (
2019), it is clear that the fit of network models to partial correlations were compared to the fit of latent variable models to uncorrected correlation matrices. It is also apparent from this code and tabled data in these articles that the authors used the sparse correlation matrices only to identify the correlations to be included in the model and subsequently, estimated their magnitude to derive fit indices for comparison with latent variable models.
The present study was concerned with comparing the performance of a network model of the WAIS-IV standardization samples with several latent variable models. Both types of models were evaluated for their ability to fit both partial correlations and uncorrected correlations. In order to have a meaningful comparison of models, it is necessary that they be evaluated using identical data (i.e., correlation matrices). Otherwise, it is not possible to rule out the possibility that differences in model fit are due to the fact that disparate correlation matrices are compared.
3. Results
A summary of fit indices for models evaluating partial correlation matrices is shown in
Table 1. The table shows
χ2 values for the training set and validation sets. As can be seen in
Table 1, the EBIClasso network produced the smallest values of
χ2 in all cases, essentially replicating the effects seen by
Kan et al. (
2019) and
Schmank et al. (
2019). Values for several fit indices (RMSEA, CFI and Akaike) presented in
Table 1 also support this conclusion.
A summary of model fit indices for models fit to uncorrected correlation matrices is shown in
Table 2. As can be seen in
Table 2, the EBIClasso model provided the poorest fit of any of the models evaluated. The best fitting model in this series is the Penta-factor model.
Comparison of the baseline χ2 values for the partial and uncorrected correlation matrices provides insight into the markedly divergent performance of the EBIClasso model. In each case, the baseline model for partial correlations is less than 22% of that for the uncorrected correlation matrix. The partial correlations are the result of the removal of a large amount of covariance that is common to many of the WAIS-IV subtests. The sparse matrix produced by the EBIClasso model does not account for much of this covariance.
4. Discussion
This study shows that a network model provides very good fit to matrices of partial correlations between the WAIS-IV subtests but very poor fit to matrixes of uncorrected correlations. The main issue then, is whether modeling partial correlations or uncorrected correlations is most appropriate.
Epskamp and Fried (
2018) discuss several reasons for favoring partial correlations in the analysis of networks. These include elimination of multicollinearity, identification of causal pathways, and identification of clusters in the network that correspond to latent variables. They further describe regularization of partial correlation matrices as a method to eliminate spurious edges. It should be noted that displaying regularized partial correlations as a network is simply a manner of visualizing the same data that could be examined in a table. As a result, what
Epskamp and Fried (
2018) say about correlation networks applies to correlation matrices in general.
Partial correlation networks by definition rule out multicollinearity. Whether or not this is a desirable property of a correlation matrix depends on one’s perspective.
Epskamp and Fried (
2018) wish to infer causality links in a network so that multicollinearity represents an interpretive problem. However, typical covariance structural modeling actually relies on multicollinearity to identify latent variables. The related point of inferring causation is more complex. The use of partial correlations eliminates the possibility that scores on a given variable are caused by other variables included in the partial correlation matrix. It is still possible that factors not included in the analysis are causal. Indeed, a common cause that is not directly measured is assumed to be the case in typical latent variable modeling. Other authors have pointed out that there are additional features of causal relationships, such as effects that are lagged in time, that require analyses of time series (
Granger 1969).
Conceptual problems arise from assuming that there are causal relationships between the observed scores on test batteries such as the WAIS-IV. This approach essentially requires that the test scores correspond in a one-to-one manner with the psychological traits that have causal relations. However, it seems unlikely that performance on a WAIS scale such as digit span directly causes performance on a WAIS scale such as arithmetic. At best, one would need to assume, following classical test theory, that the “true score” on a narrow trait exclusively indexed by digit span was causally related to the true score on a narrow trait exclusively related to arithmetic. However, such an approach would necessitate a proliferation of psychological traits associated with each of the various scales appearing on the multitude of cognitive test batteries that are currently in existence. The alternative is to assume that individual scales reflect the action of common latent traits. This argument is similar to that made by
McFarland and Malta (
2010) concerning the network analysis of symptoms in psychopathology. In this sense, the scores on WAIS sub-tests are analogous to the symptoms of psychopathology (i.e., they are symptoms of mental abilities). As discussed by
Epskamp et al. (
2018), it is certainly possible to have a network model of latent variables, but this is quite different from a network model with the nodes being scores on individual tests (i.e., manifest variables).
From a practical point of view, the fact that partial correlation matrices eliminate a major portion of the covariance between WAIS-IV test scales renders them inadequate for clinical interpretation. A model of WAIS partial correlations accounts for only a small fraction of the covariance in test scores, the majority of which is not unique to only two tests. Whether this common variance is best modeled with a single latent variable (i.e., g) and multiple group-level factors is a separate issue. Clinical interpretation should be based on models that account for a large portion of the actual variance in test scores. The same can be said for theoretical concerns.
Another issue concerns the role of regularization in model selection. The EBIClasso procedure provides a means of generating a sparse correlation matrix (i.e., a matrix where a portion of the elements are zero). Without some form of regularization, the model would essentially just be an alternative to the original partial correlation matrix, accounting for all of the covariance and having no residual covariance.
Epskamp and Fried (
2018, p. 620) state that “Even when two variables are conditionally independent, we still obtain nonzero (although typically small) partial correlations that will be represented as very weak edges in the network. These edges are called
spurious or
false positives”. This view seems to be based on the logic of null hypothesis significance testing. However, the authors of the graphical lasso used in the EBIClasso procedure state that it serves to increase sparisity and do not mention statistical significance or spurious values (
Friedman et al. 2008). In discussing the common features of diverse regularization procedures,
Bickel and Li (
2006) suggest that they seek to solve the inherent stability problem when estimating a large number of model parameters. Regularization facilitates both the generalization of models to novel data as well as model interpretation due to greater simplicity. These considerations align with the methodology of predictive modeling rather than hypothesis testing (
Shmueli 2010). Sparse solutions have traditionally been obtained by heuristic procedures such as stepwise regression or elimination of eigenvectors with eigenvalues less than one in principal components analysis. More recently, regularization by methods such as parameter shrinkage by the Lasso used in the EBIClasso procedure have been advocated as alternatives.
As noted earlier,
Kan et al. (
2019) suggested that their results provided support for a mutualism model (
van der Maas et al. 2006), while
Schmank et al. (
2019) suggested that network models are compatible with process overlap theory (
Kovacs and Conway 2016). From a latent variable perspective, the correlate factors model from the WAIS-IV manual and
Table 1 and
Table 2 of the present study might represent an appropriate form for the mutualism model. Likewise, the penta-factor model (
McFarland 2017) could serve as a useful form for a latent variable model of process overlap theory. For example, the general form of the penta-factor model in the present study is consistent with
Kovacs and Conway’s (
2016) conception of multiple overlapping domain-general executive cognitive processes being required for the performance of any given test item.