1. Introduction
According to the German animal protection act [
1] §1, no person may, without good cause, inflict pain, suffering, injury or lasting harm on any animal. Death of an animal is one form of harm, and, accordingly, §17 of the animal protection act states that killing an animal without reasonable cause is a criminal offense. Killing of animals in animal experiments (§7 of the animal protection act) in accordance with EU directive 2010/63 [
2] must adhere to the 3R principle (replace, reduce, refine) [
3], of which the second R stands for reduce, that is, limiting the number of animals used to the required minimum. Upon planning an animal experiment, one must provide a power calculation to demonstrate that the animal proposal does only comprise the minimum possible number of animals.
To investigate disease processes in animals, to date, interventions such as pharmacological treatments or surgeries, or genetic manipulations, are necessary which, for example, inhibit or activate signal transduction pathways to find out which effect this has on a pathophysiological process. For this purpose, treatment and untreated control groups are required or genetically manipulated and wildtype control animals. Additionally, a certain group size is necessary to obtain statistically significant results.
Against this background, it is suggested that complex correlation analyses of several data sets obtained from one single animal by multimodal measurements can be performed using a second-generation statistical method. This would allow one to conclude on the relationships between these data sets. Ideally, by this, it is possible to, for example, find out which signal transduction pathways or metabolic pathways play a role in the development or course of a disease, or its therapy.
In the future, by using second-generation statistical methods, reduction in animal numbers by minimizing the number of intervention groups can be obtained on one hand by using pilot studies with small numbers of animals, in which multimodal analyses with subsequent complex statistical analyses are performed to identify important paths which should be further investigated. On the other hand, more information can be gained from a single animal which might serve to reduce intervention groups, as interconnections which before had to be modulated pharmacologically or by surgeries can be calculated from the underlying correlation analyses. Complex statistical analyses of animal experimental data by second-generation statistical methods therefore are a potentially important contribution to “reduce” in line with the 3R principle.
Second-generation statistical methods are often used in the social sciences, as in marketing or psychology. In the veterinary medicine field, they are used in planning the breeding of farm animals [
4], epidemiology, and veterinary care [
5]. Only one animal experimental study exists, an ecotoxicology study in fish, in which the second-generation method of structural equation modeling (SEM) is used to determine which key events cause the death of animals [
6].
That study is an interesting example on the use of PSEM (piecewise structural equation modeling), which is an advanced form of SEM, on experimental animal data and on how animal numbers in toxicology studies can be saved by using complex statistical analyses. The authors investigate an adverse outcome pathway network consisting of several linear networks. From their data, they calculated which of the linear networks contribute most to adverse outcomes, and if any upstream key events could predict the adverse outcomes. Their adverse outcome pathways consist of a cascade of causal toxicology events, starting with a molecular-initiating event that extends over multiple key events—the events that occur in complex biological organisms—and ending with an adverse outcome. The study yielded reliable complex adverse outcome pathway networks despite the small number of experimental data and showed that it was possible to identify the most important adverse outcome pathways, as well as the predictive potential of key events. Most results were similar between PSEM and the Bayesian approach also used in this study, which in contrast to PSEM, did not have sufficient sensitivity. The authors suggest that using complex statistical methods serves to reduce animal tests, as the possibility to conclude on causalities between events in complex networks allows us to obtain more information from new analysis methods like in vitro high-throughput screening or high-content omics [
6].
Second-generation statistical methods might also be applied in the analysis of experiments on laboratory animals, to collect more information from one animal. This would allow us to reduce the number of intervention groups, as necessary information can already be calculated, instead of by analyzing differences between treated and control animals. In addition, this approach allows a more targeted planning of interventions, if only the most promising interventions are performed by prior identification of important links in pathophysiological processes.
Aim of this article is to, after an introduction to second-generation statistical methods, perform a theoretical transfer of this statistical technique to animal experiments by discussing how theoretical concepts like composites and factors can be applied on experimental data and disease processes. The second step towards application is to give an example for a second-generation analysis on a small data set using PLS and to explain how this helps to reduce animal numbers.
2. Second-Generation Statistical Models and Their Potential Applications in the Analysis of Animal Experiments
For the calculation of second-generation statistical analyses, R [
7] (version 4.3.3) is used, which is an open-source software background for statistical applications, which by an active community is continuously improved. This is possible by development of libraries by users, which allow multiple applications. Upon upcoming problems, it is possible to obtain support, and identified flaws are corrected timely. One problem for inexperienced users is that they must learn a programming language, as the software has no simple user-friendly point-and-click surface.
Statistical methods can be subdivided into first- and second-generation techniques [
8]. First-generation techniques are, for example, mean difference tests such as
t-tests and ANOVA, as well as correlation and regression analyses. These methods provide limited opportunities to work with models. They cannot directly investigate the causal relationships between three or more constructs but must reduce these to the calculation of relationships between pairs. That implies that they investigate the plausibility of a single cause–effect relationship and must proceed stepwise for more complex relationships. For the investigation of paths, path analyses can be used.
Second-generation techniques in contrast, allow comprehensive and flexible modeling for the calculation of causal relationships. One of these techniques is covariance-based structural equation modeling (CB-SEM), which in recent decades, has been widely applied. Other techniques are partial least squares (PLS) path modeling [
9], or newer techniques, which try to circumvent some of the problems of the older techniques, such as that of strict requirements on the data sets [
10]. Second-generation techniques can investigate a complex collection of hypotheses of a causal theory simultaneously (please see
Figure 1), making much less mistakes than first-generation techniques; therefore, complex models must be preferred before first-generation methods [
11].
Covariance-based structural equation models [
12] were especially widely applied in the social sciences in recent years. Requirements on the data are, for example, a large sample size and normal distribution. In addition, the model must be specified correctly [
9], which includes the scientist suggesting the correct causal direction (temporal priority) and the absence of confounders [
8]. It is difficult to find data sets which fulfill these requirements [
9]. Therefore, as an alternative, the partial least squares technique is often applied, which is supposed to not pose the same strict requirements on the data sets. However, in the literature, there is a vivid discussion regarding this technique, as no scientifically sound explanation is given as to why these requirements are supposed to not apply when using PLS. In addition, upon using PLS, other errors, such as confounders, arise [
10]. However, modern PLS-SEM is a well-established technique [
13]. Newer applications which develop techniques like covariance-based SEM further tackle the problem of small sample sizes and missing normality distribution, via correction of the Chi
2 statistics, multivariate transformation, modified testing statistics, robust estimators [
10], or confidence intervals with Monte Carlo simulation, and correction for confounders.
Therefore, these techniques should be suited for the statistical analysis of data acquired in animal experiments.
The relationships in second-generation or in simple first-generation analyses are presented as paths. They represent cause–effect relationships, which are investigated in a hypothesis test in which the correlation between two variables is identified. The complex second-generation analyses contain paths for every hypothesis in a theory, which allows a comprehensive testing of multilevel theoretical relationships [
8]. These paths can identify both direct and indirect effects. Graphically, paths are often displayed as arrows.
The next figure (
Figure 2) is an example of a first-generation analysis. In this case, an analysis, for example, of the effect of surgical intervention on the development of a disease can be assessed, which is a simple cause–effect relationship.
The example in
Figure 3 is a more complex path analysis. There is no simple cause–effect relationship, but moderators and mediators are involved in the relationship between the independent and the dependent variable. On the left, an interaction is displayed, in which the moderator experiences the effect of the combination of two different treatments being different from the sum of the isolated effects of the two treatments. For example, if the isolated treatments increase a symptom, their combination could result in a reduction in the symptom or its over proportional increase.
In case of a mediation, as displayed on the right side of
Figure 3, treatment does not only have a direct effect, but also, via the effect of a mediator, has an indirect effect on the development of a symptom. An example from an experimental animal study is a high-fat diet, which in addition to causing diabetes mellitus, also negatively affects heart function, via effects on the aortic valve itself which influence the sequelae of an aortic valve surgery. By aggravating the aortic valve stenosis, it further reduces cardiac performance.
Figure 4 is an example of a simple second-generation statistical analysis. The squares represent observed variables or indicators; the variables can be measured. The circles represent the latent variables, that is, theoretical assumptions about the underlying mechanisms, of which the effect can be measured via the observed variables. Endogenous variables are latent variables, which depend on the other latent variables—represented by arrow heads pointing at those latent variables—while exogenous variables are not affected by other latent variables, that is, they are independent [
9]. This even becomes more complex as single latent variables can technically act as independent and dependent variables for different parts of the model. The model, because of these components, can be subdivided into an inner and an outer model. The inner model or structure model describes the relationships between the independent and dependent latent variables, while the outer or measurement model describes the relationships between the latent variables and their observed variables [
9].
Measurement models can be differentiated into reflective and formative. The reflective models assume that the indicators, that is, the observed variables, are an effect of the latent variables. In the graphical visualization, this is represented by arrows from the latent variables to the indicators. Changes in the latent construct cause changes in the indicators. Regularly, the common factor model is used for reflective models [
8,
14,
15].
The formative model in contrast assumes that the indicators define the latent construct, that is, that the construct is a function of its indicators. The arrows point from the observed to the latent variables. For operationalization of these concepts, composite models are usually used [
8,
14,
15].
Animal experiments, in which the effects of treatments on the development of a disease are described, are suggested to be best described by reflective models as an event in the organism that is measured by different markers, such as blood parameters, which are altered if the process in the organism changes (please see
Figure 5).
3. Factor Analysis, Composites, and Their Potential Application to Animal Experimental Data
This section deals with the possibility of transferring the theoretical models behind the second-generation statistical techniques to animal experimental data. They are the basis for setting up the calculations in R, as, for example, the paths have different directions in different models.
Jöreskog [
16] invented the common factor model as a factor-based measurement model, which investigates if empirical measurements of a concept are in line with the assumed theoretical nature of this concept [
17]. In that model, researchers aim to understand the behavior of non-observable, conceptual variables, which from a scientific perspective represents real units and possessing direct and indirect causal consequences in the real world. From an empirical perspective, they are terms for observable empirical regularities, while only the observable phenomena are considered as real. Operationalists in contrast consider them as applications of a defined quantitative method for data collected after a specific protocol [
14]. The common factor model assumes that the observed indicators are a manifestation of the underlying concept, which is assumed as their common cause [
17]. In recent years, the operationalization of the theoretical concept and of the common factor has increasingly merged, so that the terms are hardly distinguishable [
17].
Wold [
18] invented a composite-based alternative to this approach. The composites are used for the investigation of constructed artifacts and abstract concepts, which serve a certain cause, and of which the indicators do not necessarily have a common cause [
15,
17]. They initially served the purpose of dimension reduction; therefore, they were conceptualized such that they as efficiently as possible captured the most important characteristics of the data [
17].
For a long time, in covariance structure analysis, abstract concepts were operationalized by latent variables (LVs) and called LV modeling [
19]. The indicators in these models are the cause of the LVs, supplemented by an error term, which represents missing causes. Each change in a formative LV is mediated by a change in at least one indicator or in the error term [
20].
In the initial composite model, the constructed concept is displayed as an emergent variable (named composite construct, aggregated construct, or formative construct). The emergent variables are defined by their indicators, that is, they are a linear combination or summary of observed variables, instead of causing them, as in the common factor model [
15,
19]. The emergent variables can be embedded in structure models, as LVs are in SEM. They are estimated by different methods amongst others by PLS or via main component analysis, linear discrimination analysis (LDA), or generalized canonical correlation analysis (GCCA) [
19].
Rigdon [
21] developed these concepts further: He suggested interpreting factors (common factors) or composites, which are the statistical models, as approaches for conceptual variables and theoretical models which cannot be observed. In this way, he explains the relationship between the concepts and the associated observed variables [
14].
For the models which are used in animal research, the question arises, under the assumption of the measurement model applied, whether the aim is to obtain results which reflect the real situation as good as possible.
In line with the theory of Jöreskog [
16], factors exist in reality but cannot be measured directly. An example from economics is the attitude towards a brand. Composites in line with the theory of Wold [
18], in contrast, are artifacts, which are created by humans, and therefore do not exist in nature. However, they can be meaningful, for example, in economy, in the form of stock indices as the NASDAQ composite. In the theory of Henseler [
17], in this relationship, the term “therapy” is used.
Therefore, some questions arise:
- (1)
Do factors exist in (veterinary) medicine? If so, a composite cannot be a solution, even if the causes of the factor are of interest. In this case, the MIMIC (multiple indicators, multiple causes) model of Jöreskog [
22] would be of interest.
- (2)
Do composites exist in (veterinary) medicine? This is not known; however, in this relation, the term “therapy” from the theory of Henseler [
17] is probably accurate.
These considerations are probably philosophical science.
If it is assumed that the disease of interest is a factor, and the symptoms are modeled as a consequence of the disease, this is a reflective measurement model. If it is of interest which treatment has which influence on the factor (the disease), this is equivalent to the MIMIC model of Jöreskog [
22] with multiple indicators (symptoms) and multiple causes (treatments).
5. Discussion
This article gives an introduction to how second-generation statistical methods can be used on experimental animal data sets. After introducing these methods, it discusses how theoretical concepts of second-generation statistical methods can be transferred to experimental data and disease processes, as a necessary basis for setting up a calculation in R. Lastly, an example of how to calculate a complex statistical analysis on an experimental data set is given, due to small n-numbers using PLS. The next section of this article mainly focusses on the criticism of PLS and on alternatives to this technique, as, to date, this second-generation technique is widely used on small data sets and other data which do not meet the strict requirements necessary to calculate classical SEM analyses, like what is often found in animal experimental data. The appropriateness of this, however, is put into question by the community.
Second-generation statistical methods are regularly applied in social sciences; however, they have not been regularly used in animal experiments. They are suited for modeling complex causal relationships, in which they are superior to simpler methods, investigating simple cause–effect relationships between variables; therefore, they must work step by step. By developing new techniques, it might be possible to work with data sets which only comprise small sample numbers and which are not normally distributed. This is often the case with animal experimental results and is even asked for due to the ethical necessity to reduce animal numbers.
As newer techniques that are suited for the analysis of causal relationships in small sample sizes exist [
27,
28,
29], application of second-generation statistical methods due to several reasons is highly relevant for animal experimental research. On one hand, from several data sets acquired in one animal, causal relationships can be calculated, without having to use additional intervention groups or genetically modified animals. This, besides the animals included in the analysis, also comprises surplus animals from breeding. Besides reduction, this serves as refinement by limiting the burden the experimental animals must endure due to experimental manipulations. At the least, these advanced statistical calculations allow a more straightforward experimental approach, by promising signal transduction pathways can be identified by statistical calculation and the subsequent experiments limited to these pathways.
Limitations
An important limitation of second-generation statistical methods is that they test hypothesized causal models derived from theory. However, they cannot calculate causal relationships from observed data alone. Instead, randomized interventions (such as pharmacological or surgical interventions or genetic manipulations) are designed to specifically infer causality. Complex statistical analysis can be applied to data from within an experimental group to model the complex, multi-factorial relationships between measured variables like food and blood parameters, as well as functional values like heart function in echocardiography but cannot replace the comparison between experimental groups against control groups. However, in times in which authorities are ever more restricting access to animal numbers and intervention groups, besides the ethical aspect, restricting the groups in which interventions are performed to a minimum by obtaining as much information on the processes of interest by complex statistical analyses is of great importance.