Next Article in Journal
Economic and Environmental Analysis of Solar Thermal and Seasonal Thermal Energy Storage Based on a Renewable Energy Conversion System for Greenhouses
Next Article in Special Issue
Physics-Informed Neural Network Solution of Point Kinetics Equations for a Nuclear Reactor Digital Twin
Previous Article in Journal
Evaluation of Thermal-Physical Properties of Novel Multicomponent Molten Nitrate Salts for Heat Transfer and Storage
Previous Article in Special Issue
Development of a Trajectory Period Folding Method for Burnup Calculations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Overview of Arbitrarily High-Order Adjoint Sensitivity and Uncertainty Quantification Methodology for Large-Scale Systems

by
Dan Gabriel Cacuci
Center for Nuclear Science and Energy, University of South Carolina, Columbia, SC 29208, USA
Energies 2022, 15(18), 6590; https://doi.org/10.3390/en15186590
Submission received: 8 August 2022 / Revised: 23 August 2022 / Accepted: 5 September 2022 / Published: 8 September 2022
(This article belongs to the Special Issue Advances in Modelling for Nuclear Science and Engineering)

Abstract

:
This work reviews from a unified viewpoint the concepts underlying the “nth-Order Comprehensive Adjoint Sensitivity Analysis Methodology for Response-Coupled Forward/Adjoint Linear Systems” (nth-CASAM-L) and the “nth-Order Comprehensive Adjoint Sensitivity Analysis Methodology for Nonlinear Systems” (nth-CASAM-N) methodologies. The practical application of the nth-CASAM-L methodology is illustrated for an OECD/NEA reactor physics benchmark, while the practical application of the nth-CASAM-N methodology is illustrated for a nonlinear model of reactor dynamics that exhibits periodic and chaotic oscillations. As illustrated both by the general theory and by the examples reviewed in this work, both the nth-CASAM-L and nth-CASAM-N methodologies overcome the curse of dimensionality in sensitivity analysis. The availability of efficiently and exactly computed sensitivities of arbitrarily high order can lead to major advances in all areas that need such high-order sensitivities, including data assimilation, model calibration, uncertainty reduction, and predictive modeling.

1. Introduction

The computational model of a physical system comprises the following fundamental conceptual components: (a) a well-posed system of equations that relate the system’s independent variables and parameters to the system’s state (i.e., dependent variables); (b) nominal or mean values for the system parameters, along with information regarding the probability distributions, moments thereof, and inequality and/or equality constraints that define the range of variations in the system’s parameters; and (c) one or several quantities, customarily referred to as model/system responses (or objective functions, or indices of performance), which are computed using the mathematical model. The functional derivatives of model responses with respect to model parameters are called “sensitivities” (of responses with respect to model parameters). The computation of higher-order sensitivities by conventional methods is subject to the “curse of dimensionality”, a term coined by Belmann [1] to describe phenomena in which the number of computations increases exponentially in the respective phase space. In the particular case of sensitivity analysis, the number of large-scale computations increases exponentially in the parameter phase space as the order of sensitivities increases if conventional methods (e.g., finite differences) are used: thus, if a model has TP parameters, the computation of first-order sensitivities requires at least 2TP large-scale computations; the computation of second-order sensitivities requires at least (3TP)2 computations, and so on; and the computation of the nth-order sensitivities requires at least (nTP)n large-scale computations.
A fundamental breakthrough in the sensitivity analysis of problems of interest to nuclear reactor physics, which are governed by the linear Boltzmann transport equation, was achieved by Wigner [2], who pioneered the use of the adjoint Boltzmann transport equation to compute first-order perturbations (and, hence, first-order sensitivities) in the reactor’s multiplication factor arising from perturbations in the model parameters (nuclear cross sections). Wigner’s method became widespread in the nuclear engineering community and was considered to be adequate for predicting the effects of small parameter changes on the model responses, as well as for performing [3,4] a first-order uncertainty quantification of the standard deviation induced in a model response due to uncertainties (standard deviations and correlations) affecting the model parameters (nuclear cross sections). This situation changed after Goldstein [5] reported studies of neutron transmission through a sodium shield, which showed that even for modest variations of 5% in the cross section at the 297 keV sodium minimum, the first-order sensitivities of responses measured by detectors differed by 20–50% from their correct/actual values. These results indicated that the nonlinear dependence of a detector’s response on the sodium cross section in deep-penetration radiation and neutron shielding problems cannot be predicted by first-order sensitivities, despite the fact that the physics of neutron/gamma transport through materials is governed by the linear Boltzmann equation. These findings have motivated Greenspan et al. [6] to propose an extension of Wigner’s method [2], so as to enable the computation of second-order sensitivities of detector responses to neutron cross sections for linear neutron transport problems. Greenspan et al. [6] concluded that their methodology “appears to provide satisfactory accuracies for most non-linear problems associated with cross section changes encountered in practice”. This conclusion, while correct for the particular “neutron-transmission through a sodium shield problem” under investigation, is actually incorrect in general since, as is shown in the remainder of this work, third- and fourth-order sensitivities of responses with respect to cross sections are even more important than the second-order sensitivities in reactor physics problems. Greenspan et al. [6] further stated that: “The use of higher (than second) order sensitivity theories like the one recently proposed by Gandini [7]—is not likely to be desirable—the added accuracy provided, say, by a third order formulation is not likely to justify the additional computational effort required for its application”. Again, this statement by Greenspan et al. [6] is correct regarding the methodology proposed by Gandini [7], which is not only cumbersome and computationally inefficient but is also limited to formal time-dependent linear neutron transport and/or diffusion problems. However, the sweepingly general statement by Greenspan et al. [6] that the “use of higher (than second) order sensitivity theories…is not likely to be desirable” is proven to be incorrect in the remainder of this work by showing that the third- and fourth-order sensitivities of responses with respect to cross sections are even more important than the second-order sensitivities in reactor physics problems.
Wigner’s methodology and its extensions by Greenspan et al. [6] and Gandini [7] were developed within the specific context of the linear neutron transport equation and are not applicable to nonlinear models, since nonlinear operators do not admit adjoint operators. The rigorous first-order adjoint sensitivity analysis methodology for generic large-scale nonlinear (as opposed to linearized) systems involving generic operator responses has been conceived and developed by Cacuci [8,9], who is also credited (see, e.g., [10,11]) for having introduced these principles to the earth, atmospheric, and other sciences. Cacuci’s [8,9] first-order adjoint sensitivity analysis methodology for nonlinear systems is just as efficient for computing first-order sensitivities as Wigner’s method is for linear systems, requiring a single large-scale (adjoint) computation for obtaining all of the first-order sensitivities, regardless of the number of model parameters. This is in contradistinction with the number of (at least) “2P computations” which would be required by conventional methods (e.g., finite differences). Thus, the methodologies of Wigner [2] and Cacuci [8,9] provided the “first-order cracking of the curse of dimensionality” in sensitivity and uncertainty analysis. Cacuci’s methodology [8,9] has been applied to many (too numerous to cite here) research fields since its publication. Second-order deterministic analyses of specific problems in diverse fields, e.g., structural analysis [12,13], mechanical system [14], atmospheric science [15], linear circuit analysis, and optimization [16], have also been published. General methodologies for the efficient and exact computation of second-order response sensitivities to model responses were proposed by Cacuci for response-coupled forward and adjoint linear systems [17] and for nonlinear systems [18].
The unparalleled efficiency of the second-order adjoint sensitivity analysis methodology for linear systems conceived by Cacuci [17] was demonstrated in [19,20,21,22,23,24] by applying this methodology to compute exactly the 21,976 first-order sensitivities and 482,944,576 s-order sensitivities for an OECD/NEA reactor physics benchmark [25], which is representative of a large-scale system that involves many parameters. The results obtained in [19,20,21,22,23,24] indicated that many second-order sensitivities were much larger than the largest first-order ones, which implies that the consideration of only the first-order sensitivities is insufficient for making credible predictions regarding the expected values and uncertainties (variances, covariances, skewness) of calculated and predicted/adjusted responses. At the very least, the second-order sensitivities must also be computed in order to enable the quantitative assessment of their impact on the predicted model responses.
The findings reported in [19,20,21,22,23,24] motivated the investigation [26,27] of the largest third-order sensitivities, many of which were found to be even larger than the second-order ones. This finding, in turn, has motivated the development [28] of the mathematical framework for determining and computing the fourth-order sensitivities for the OECD/NEA benchmark, many of which were found [29,30] to be larger than the third-order ones. The need for computing the third- and fourth-order sensitivities has been underscored in [31,32]. The efficiency of Cacuci’s new methodology for computing exactly third-order and fourth-order sensitivities while overcoming the curse of dimensionality has been highlighted in [33,34,35].
This sequence of findings has motivated the conception by Cacuci [36] of the “nth-Order Comprehensive Adjoint Sensitivity Analysis Methodology for Response-Coupled Forward/Adjoint Linear Systems” (abbreviated as “nth-CASAM-L”). The nth-CASAM-L enables the efficient computation of exactly determined expressions of arbitrarily high-order sensitivities of a generic system response—which can depend on both the forward and adjoint state functions− with respect to all of the parameters that characterize the physical system. The qualifier “comprehensive” is employed in order to highlight that the model parameters considered within the framework of the nth-CASAM-L include the system’s uncertain boundaries and internal interfaces in the phase space. The nth-CASAM-L mathematical framework was developed specifically for linear systems because the most important model responses produced by such systems are various Lagrangian functionals, which depend simultaneously on both the forward and adjoint state functions governing the respective linear system. Included among such functionals are the Raleigh quotient for computing eigenvalues and/or separation constants when solving linear partial differential equations, and the Schwinger and Rousopoulos functionals (see, e.g., [37,38,39]), which play fundamental roles in optimization and control procedures, derivation of numerical methods for solving equations (differential, integral, integro-differential), and sensitivity analysis. Responses that depend on both the forward and adjoint state functions underlying the respective model can occur only for linear systems, because responses in nonlinear systems can only depend on the system’s forward state functions (since nonlinear operators do not admit adjoint operators). Consequently, the sensitivity analysis of responses that simultaneously involve both forward and adjoint state functions makes it necessary to treat linear models/systems in their own right, rather than treating them as particular cases of nonlinear systems. The nth-CASAM-L is formulated in linearly increasing higher-dimensional Hilbert spaces (as opposed to exponentially increasing parameter-dimensional spaces), thus overcoming the curse of dimensionality in enabling the most efficient computation of exactly determined expressions of arbitrarily high-order sensitivities of generic nonlinear system responses with respect to model parameters, uncertain boundaries, and internal interfaces in the model’s phase space.
In parallel with the aforementioned developments, Cacuci [40,41] has extended his original work [8,9] on nonlinear systems, providing the mathematical framework for deriving and computing efficiently the exact expressions of sensitivities with respect to imprecisely known (i.e., uncertain) model parameters, model boundaries, and/or internal interfaces, up to and including fifth order. Subsequently, Cacuci [42] has generalized the 5th-CASAM-N methodology [41] to enable the exact and efficient computation of sensitivities of arbitrarily high order with respect to model parameters, uncertain boundaries, and internal interfaces in the model’s phase space. This novel general methodology [42], called the nth-Order Comprehensive Adjoint Sensitivity Analysis Methodology for Nonlinear Systems (nth-CASAM-N), is also formulated (just like the nth-CASAM-L) in linearly increasing higher-dimensional Hilbert spaces (as opposed to exponentially increasing parameter-dimensional spaces), thus overcoming the curse of dimensionality in sensitivity analysis of nonlinear systems.
This work is structured as follows: Section 2 highlights the fundamental role of sensitivities in defining the correlations between responses and parameters. If the negative impact of the curse of dimensionality is not alleviated when computing sensitivities, this negative impact is transmitted to the computation of the aforementioned correlations. Section 3.1 illustrates the need for computing high-order sensitivities by reviewing the results obtained in [16,17,18,19,20,21,22,23,24,25,26,27,28,29,30] for an OECD/NEA reactor physics benchmark [25], which is representative of linear models/systems. These results underscore the efficiency of the nth-CASAM-L in overcoming the curse of dimensionality while producing the respective results. Section 3.2 reviews the general mathematical framework underlying the nth-CASAM-L [36].
Section 4.1 reviews the pioneering sensitivity and uncertainty analysis results presented in [43,44] for the reduced-order model developed by March-Leuba, Cacuci, and Perez [45] for studying the nonlinear dynamics displayed by boiling water reactors (BWR). This reduced-order model comprises point neutron kinetics equations coupled nonlinearly to thermal-hydraulics equations that describe the time evolution of the fuel temperature and coolant density in the recirculation loop, which is treated as a single path of fluid of variable cross-sectional area but with constant mass flow, with the coolant entering the core at saturation temperature. This model shows that the major state variables, including power, reactivity, void fraction, and temperature, undergo limit cycle oscillations under high-power/low-flow operating conditions, and the amplitudes of these oscillations are very sensitive to the reactor’s operating conditions. When increasing the heat transfer from the reactor to the coolant, this model indicates that these limit cycle oscillations can become unstable, undergoing period-doubling bifurcations in the time evolutions of the state variables (power, delayed neutron precursors, reactivity, coolant void fraction, fuel temperature) from periodic to aperiodic (chaotic) oscillations. Thus, this nonlinear model [45] illustrates generically the most extreme forms of nonlinear phenomena (from limit cycles to chaos) that can be displayed by nonlinear physical systems. The results of the sensitivity and uncertainty analyses presented in Section 4.1 are obtained by applying the first-order particularization of the nth-CASAM-N [42], the general form of which is reviewed in Section 4.2. The work of March-Leuba, Cacuci, and Perez [45] has inspired several related models of BWR-dynamics [46,47,48,49,50,51,52,53,54,55].
Section 5 reviews the concepts underlying the nth-CASAM-L [36,56] and nth-CASAM-N [42] methodologies from a unified viewpoint, indicating under which circumstances the nth-CASAM-L could be viewed as a particular case of the nth-CASAM-N methodology. As the paradigm examples presented in Section 3.1 and Section 4.1 illustrate, using finite differences for computing high-order sensitivities of model responses to model parameters becomes not only computationally unfeasible for large-scale systems (because of the curse of dimensionality), but finding the optimal step size to minimize the error between the finite difference result and the exact result is practically impossible to achieve unless one knows beforehand what the exact result is—which would be possible only by using the CASAM-L and/or nth-CASAM-N methodologies.

2. The Fundamental Role of High-Order Sensitivities in Defining Correlations among Model Responses and Parameters

The mathematical model of a physical system comprises independent variables (e.g., space, time, etc.), dependent variables (aka “state functions”; e.g., temperature, mass, momentum, etc.), and various parameters (appearing in correlations, coordinates of physical boundaries, etc.), which are all interrelated by equations that usually represent conservation laws. Without loss of generality, the model parameters can be considered to be real scalar quantities, having known nominal (or mean) values and, possibly, known higher-order moments or cumulants (i.e., variance/covariances, skewness, kurtosis), which are usually determined from experimental data and/or processes external to the physical system under consideration. The model parameters usually stem from processes that are external to the system under consideration and are seldom, if ever, known precisely. These imprecisely known model parameters are denoted as α 1 ,…, α T P , where the subscript “TP” indicates “Total (number of) Parameters”. It is convenient to consider that these parameters are components of a “vector of parameters”, denoted as α ( α 1 , , α T P ) T P , where T P denotes the TP-dimensional subset of the set of real scalars. These model parameters are considered to include imprecisely known geometrical parameters that characterize the physical system’s boundaries in the phase space of the model’s independent variables. Thus, α T P is a TP-dimensional column vector and its components are considered to include all of the uncertain model parameters, including those that may enter into defining the system’s boundary in the phase space of independent variables. For subsequent developments, it is convenient to denote matrices and vectors using capital and lower case bold letters, respectively. The symbol “ ” will be used to denote “is defined as” or “is by definition equal to”. Transposition will be indicated by a dagger ( ) superscript.
The result of interest computing using a mathematical/computational model of a physical system is, in principle, an implicit function of the system’s parameters α ( α 1 , , α T P ) T P . In practice, the model parameters are not known exactly, even though they are not bona fide random quantities. For practical purposes, however, these model parameters are considered to be variates that obey a multivariate probability distribution function, denoted as p α ( α ) , which is seldom known, particularly for large-scale systems involving many parameters. Considering that the multivariate distribution p α ( α ) is formally defined on a domain D α , the various moments of p α ( α ) can be formally defined in a standard manner by using the following notation
u ( α ) α D α u ( α ) p α ( α ) d α
where u ( α ) is a continuous function of the parameters α . Thus, the moments of p α ( α ) are defined as follows:
  • The expected (or mean) value of a model parameter α j , denoted as α j 0 , is defined as follows:
    α j 0 α j α D α α j p α ( α ) d α ,         j = 1 , , T P .
    The expected values α j 0 are considered to be the components of the following vector of mean (expected) values:
    α 0 ( α 1 0 , , α T P 0 )
  • The covariance, cov ( α j 1 , α j 2 ) , of two parameters, α j 1 and α j 2 , is defined as follows:
    cov ( α j 1 , α j 2 ) ( α j 1 α j 1 0 ) ( α j 2 α j 2 0 ) α ,         j 1 , j 2 = 1 , , T P .
    The variance, var ( α j 1 ) , of a parameter α j 1 , is defined as follows:
    var ( α j 1 ) ( α j 1 α j 1 0 ) 2 α ,           j 1 = 1 , , T P .
    The standard deviation of a parameter α j 1 , which is denoted σ j 1 , is defined as follows: σ j var ( α j ) . The correlation, ρ j 1 j 2 , between two parameters α j 1 and α j 2 , is defined as follows:
    ρ j 1 j 2 cov ( α j 1 , α j 2 ) / ( σ j 1 σ j 2 ) ;           j 1 , j 2 = 1 , , T P .
  • The third-order correlation, t j 1 j 2 j 3 , among three parameters α j 1 , α j 2 and α j 3 , is defined as follows:
    t j 1 j 2 j 3 ( α j 1 α j 1 0 ) ( α j 2 α j 2 0 ) ( α j 3 α j 3 0 ) α / ( σ j 1 σ j 2 σ j 3 ) ;     j 1 , j 2 , j 3 = 1 , , T P .
  • The fourth-order correlation q j 1 j 2 j 3 j 4 among four parameters, is defined as follows, for j 1 , j 2 , j 3 , j 4 = 1 ,   ,   T P :
    q j 1 j 2 j 3 j 3 ( α j 1 α j 1 0 ) ( α j 2 α j 2 0 ) ( α j 3 α j 3 0 ) ( α j 4 α j 4 0 ) α σ j 1 σ j 2 σ j 3 σ j 4 .
    In the important particular case when the parameters are normally distributed, the following relation holds:
    q j 1 j 2 j 3 j 3 = ρ j 1 j 2 ρ j 3 j 4 + ρ j 1 j 3 ρ j 2 j 4 + ρ j 1 j 4 ρ j 2 j 3 .
The higher-order correlations among parameters are defined following the same pattern as displayed in Equations (6)–(8). In practice, the parameter correlations are estimated using experimental information about the parameters.
The result of interest computed using the model of the physical system under consideration is customarily called a system/model “response” and is denoted as R k ( α ) . Being an implicit function of the system’s parameters, the response is considered to admit a formal Taylor series of around a point α 0 ( α 1 0 , , α T P 0 ) , which is usually taken to be the vector of expected (or nominal) parameter values, having the following well-known formal expression:
R k ( α ) = R k ( 0 ) ( α 0 ) + n = 1 1 n ! j 1 = 1 T P j 2 = 1 T P j n 1 = 1 T P j n = 1 T P { n R k ( α ) α j 1 α j 2 α j n 1 α j n } α = α 0                   × [ ( α j 1 α j 1 0 ) ( α j 2 α j 2 0 ) ( α j n 1 α j 4 n 1 0 ) ( α j n α j n 0 ) ] .
The radius/domain of convergence of the series in Equation (10) must be established before using it for any practical application. The functional derivatives of the model response R k ( α ) with respect to the model’s parameters are customarily called the “sensitivities” (of various orders) of the model’s response with respect to the parameters. As is well known, and as indicated by Equation (10), the Taylor series of a function of T P variables (e.g., R k ( α ) ) comprises T P first-order derivatives, T P ( T P + 1 ) / 2 distinct second-order derivatives, T P ( T P + 1 ) ( T P + 2 ) / 6 distinct third-order derivatives, and so on. The computation by conventional methods of the n t h -order sensitivities of a response with respect to the T P parameters would require at least O ( T P n ) large-scale computations. This exponential increase—with the order of response sensitivities—in the number of large-scale computations needed to determine higher-order sensitivities is the manifestation of the “curse of dimensionality in sensitivity analysis”, by analogy to expression coined by Belmann [1] to express the difficulty of using “brute-force” grid search when optimizing a function with many input variables.
It follows from the Taylor series expansion presented in Equation (10) that, in principle, the response and parameters are correlated quantities. The correlations of various orders between a model’s response(s) and parameters are defined as follows.
  • The expectation (value), E ( R k ) , of a response R k ( α ) is defined as follows:
    E ( R k ) R k ( α ) α D α R k ( α ) p α ( α ) d α
  • The covariance between two responses R k and R is defined as follows:
    c o v ( R k ,   R ) [ R k ( α ) E ( R k ) ] [ R ( α ) E ( R ) ] α .
  • The triple correlations (or third-order moment of the distribution of responses), denoted as μ 3 ( R k , R , R m ) , among three responses R k ( α ) , R ( α ) , and R m ( α ) , are defined as follows:
    μ 3 ( R k , R , R m ) [ R k ( α ) E ( R k ) ] [ R ( α ) E ( R ) ] [ R m ( α ) E ( R m ) ] α
  • The quadruple correlations (or fourth-order moment of the distribution of responses), denoted as μ 4 ( R k , R , R m , R n ) , among four responses R k ( α ) , R ( α ) , R m ( α ) , and R n ( α ) , are defined as follows:
    μ 4 ( R k , R , R m , R n ) [ R k ( α ) E ( R k ) ] [ R ( α ) E ( R ) ]                                 × [ R m ( α ) E ( R m ) ] [ R n ( α ) E ( R n ) ] α .
Expressions for the response moments defined in Equations (11)−(14) were first obtained by Tuckey [57], who provided expressions up to fifth order in the parameter standard deviations. Cacuci [56] has generalized Tuckey’s work, providing corresponding expressions up to and including the sixth-order terms in parameter standard deviations. Cacuci [56] has also noticed that the existence of the Taylor-series shown in Equation (10) implies not only correlations among model responses but also correlations among model responses and model parameters, as defined below in Equations (15)−(20), and has provided the corresponding expressions up to and including the sixth-order terms in parameter standard deviations.
e.
The covariance between computed responses and a parameter is defined as follows:
cov ( α i , R k ) ( α i α i 0 ) [ R k ( α ) E ( R k ) ] α .
f.
The triple-correlations among one parameter, α i , and two responses, R k ( α ) and R ( α ) , are defined as follows:
μ 3 ( α i , R k , R ) ( α i α i 0 ) [ R k ( α ) E ( R k ) ] [ R ( α ) E ( R ) ] α .
g.
The triple correlations among two parameters  α i , α j and one response  R k ( α ) are defined as follows:
μ 3 ( α i , α j , R k ) ( α i α i 0 ) ( α j α j 0 ) [ R k ( α ) E ( R k ) ] α .
h.
The quadruple correlations denoted as μ 4 ( α i , R j , R k , R ) , among one parameter, α i , and three responses R j ( α ) , R k ( α ) , and R ( α ) , are defined as follows:
μ 4 ( α i , R j , R k , R ) ( α i α i 0 ) [ R j ( α ) E ( R j ) ] [ R k ( α ) E ( R k ) ] [ R ( α ) E ( R ) ] α .
i.
The quadruple correlations, denoted as μ 4 ( α i , α j , R k , R ) , among two parameters  α i , α j and two responses  R k ( α ) , R ( α ) , are defined as follows:
μ 4 ( α i , α j , R k , R ) ( α i α i 0 ) ( α j α j 0 ) [ R k ( α ) E ( R k ) ] [ R ( α ) E ( R ) ] α .
j.
The quadruple correlations, denoted as μ 4 ( α i , α j , α k , R ) , among three parameters α i , α j , α k and one response, R ( α ) , are defined as follows:
μ 4 ( α i , α j , α k , R ) ( α i α i 0 ) ( α j α j 0 ) ( α k α k 0 ) [ R ( α ) E ( R ) ] α .

3. High-Order Sensitivity and Uncertainty Analysis of Linear Models/Systems: Illustrative Example and General Theory

Section 3.1 illustrates the need for computing high-order sensitivities by reviewing the results obtained in [16,17,18,19,20,21,22,23,24,25,26,27,28,29,30] for an OECD/NEA reactor physics benchmark [25], comprising a subcritical polyethylene-reflected plutonium metal sphere, which is called the “PERP” (polyethylene-reflected plutonium) benchmark. The PERP benchmark is modeled by using the deterministic neutron transport Boltzmann equation, which is solved numerically (after discretization in the energy, spatial, and angular independent variables) using the software package PARTISN [58]. The numerical model of the PERP benchmark comprises 21,976 uncertain parameters, as follows: 180 group-averaged total microscopic cross sections, 21,600 group-averaged scattering microscopic cross sections, 120 fission process parameters, 60 fission spectrum parameters, 10 parameters describing the experiment’s nuclear sources, and 6 isotopic number densities. In view of the fact that the PERP benchmark is modeled by the integro-differential Boltzmann equation in six independent variables (the solving of which is representative of “large-scale” computations), and since this model involves 21,976 uncertain parameters (which are representative of “many parameters”), the PERP benchmark can be considered to be representative of linear models/systems for the purpose of computing high-order sensitivities.
The results reviewed in Section 3.1 were obtained by applying the nth-CASAM-L, for n = 1, 2, 3, 4, underscoring the efficiency of the nth-CASAM-L in overcoming the curse of dimensionality while producing the respective results. Section 3.2 reviews the general mathematical framework underlying the nth-CASAM-L [36].

3.1. Illustrating the Need for High-Order Sensitivity Analysis and Uncertainty Quantification: A Paradigm Reactor Physics Benchmark

The PERP benchmark is a spherical subcritical nuclear system driven by a source of spontaneous fission neutrons. The PERP benchmark comprises an inner sphere (designated as “material 1” and assigned to “zone 1”) which is surrounded by a spherical shell (designated as “material 2” and assigned to “zone 2”). The inner sphere of the PERP benchmark contains α-phase plutonium which acts as the source of particles; the radius is r 1 = 3.794 cm. This inner sphere is surrounded by a spherical shell reflector made of polyethylene with a thickness of 3.81 cm; the radius of the outer shell containing polyethylene is r s = 7.604 cm. Table 1 specifies the constitutive materials of the PERP benchmark.
The neutron flux distribution within the PERP benchmark has been computed by using the deterministic software package PARTISN [58], which solves the following multigroup approximation of the transport equation for the group fluxes, φ g ( r , Ω ) :
B g ( α ) φ g ( r , Ω ) = Q g ( α ; r ) ,           g = 1 , , G ,
φ g ( r s , Ω ) = 0 , r s S b , Ω · n < 0 ,         g = 1 , , G ,
where:
B g ( α ) φ g ( r , Ω ) Ω · φ g ( r , Ω ) + Σ t g ( r )   φ g ( r , Ω ) g = 1 G 4 π Σ s g g ( r , Ω · Ω )   φ g ( r , Ω ) d Ω   χ g ( r ) g = 1 G 4 π ( ν Σ ) f g ( r )   φ g ( r , Ω ) d Ω     ,
Q g ( α ; r ) i = 1 N f λ i N i , 1 F i S F ν i S F 1 I 0 E g + 1 E g d E   e E / a i sinh b i E ,   g = 1 , , G ;
I 0 π a i 3 b i 2 e a i b i 4 .
The PARTISN [58] computations used the SOURCE4C code [59] and the MENDF71X library [60] which comprise 618 group cross sections. These cross sections were collapsed to G = 30 energy groups, with group boundaries, E g , as indicated in Figure 1. The MENDF71X library [60] uses ENDF/B-VII.1 nuclear data [61]. The group boundaries, E g , are user-defined and are therefore considered to be perfectly well-known parameters.
The fundamental quantities of interest (i.e., “system responses”) for subcritical benchmarks (such as the PERP benchmark) are single-counting rates, double-counting rates, the leakage multiplication, and the total leakage. The total leakage from the outer surface of the PERP sphere was considered to be the paradigm response of interest for illustrating the need for high-order sensitivity analysis, because the total leakage does not depend on the detector configuration (as opposed to count rates) and is therefore more meaningful physically than the count rates.
The total neutron leakage from the PERP sphere, denoted as L ( α ) , depends implicitly (through the neutron flux) on all model parameters and is defined as follows:
L ( α ) V d V 0 d E Ω · n > 0 d Ω   Ω · n δ ( r r s )   φ ( r , E , Ω ) = S b d S g = 1 G Ω · n > 0 d Ω   Ω · n   δ ( r r s ) φ g ( r , Ω ) .
The numerical value of the PERP’s benchmark total leakage, computed by using Equation (26), is 1.7648 × 10 6 neutrons/sec. Figure 1 depicts the histogram plot of the leakage for each energy group for the PERP benchmark.

3.1.1. Overview of Results for the First, Second, Third, and Fourth-Order Sensitivities

The computations of the sensitivities, of all orders, were performed deliberately using a widely accessible desktop computer, namely a DELL AMD FX-8350 with an 8-core processor. A single adjoint (large-scale) computation is needed in order to obtain all 21,976 sensitives of the first-level adjoint function, which is subsequently used to compute all of the first-order sensitivities using inexpensive quadrature formulas. The CPU time for a typical PARTISN forward neutron transport computation with an angular quadrature of S32 is ca. 45 s. On the other hand, the CPU time for a typical adjoint neutron transport computation using the neutron transport solver PARTISN with an angular quadrature of S32 is ca. 24 s. The CPU time for computing the integrals over the adjoint function is ca. 0.004 s.
Thus, the exact computation of the 21,976 first-order sensitivities using the first-level adjoint function is (24 + 21,976 × 0.004) s = 112 s. By comparison, the approximate computation of the 21,976 first-order sensitivities using a two-point finite difference formula would require 21,976 × 2 × 45 s = 550 h. Most of the first-order relative sensitivities have been found to be negligibly small, having absolute values less than 0.1. The most important (i.e., largest) sensitivities of the leakage response are with respect to the group-averaged total microscopic cross sections, followed by the sensitivities of the leakage response with respect to the isotopic number densities. Only 16 first-order relative sensitivities have absolute values greater than 1.0; most of these involve the total cross sections of 1H (hydrogen) and 239Pu (plutonium-239).
After having computed all of the first-order sensitivities and ranked their importance in the order of the magnitudes of the corresponding first-order relative sensitivities, it is imperative to compute the second-order sensitivities in order to quantify and compare their impact to the impact of the first-order sensitivities. The theoretical framework for computing efficiently and exactly second-order sensitivities has therefore been developed and applied to the PERP benchmark 2nd-CASAM-L, and the computational results obtained for all of the second-order sensitivities of the PERP model’s leakage response with respect to the benchmark’s parameters listed in Table 2 were reported in [19,20,21,22,23,24]. The computation of these second-order sensitivities was performed in the order of priority indicated by the values of the first-order sensitivities. Hence, the highest priority was given to computing the total of 32,400 = 180 × 180 second-order sensitivities of the PERP model’s leakage response to the model’s group-averaged total microscopic cross sections. It was thus found [19] that 720 of these elements have relative sensitivities greater than 1.0, and many of the second-order sensitivities are much larger than the corresponding first-order ones. The largest second-order sensitivities involve the total cross sections of 239Pu and 1H. The overall largest element is the unmixed second-order relative sensitivity S ( 2 ) ( σ t , 6 30 , σ t , 6 30 ) = 429.6 , which occurs in the lowest-energy group for 1H.
The next largest sensitivities of the PERP leakage response were found to be [19,20,21,22,23,24] the mixed second-order sensitivities involving the isotopic number density of 239Pu and the microscopic total, scattering, or fission cross sections for the 12th or 30th energy groups of 239Pu or 1H, respectively. On the other hand, the numerical results obtained in [19,20,21,22,23,24] indicated that the second-order sensitivities of the PERP leakage response with respect to the remaining parameters (including scattering cross sections, fission cross sections, average number of neutrons per fission, and source parameters) are much smaller than the corresponding first-order ones, so that the effects of uncertainties in these parameters on the expected value, variance, and skewness of the response distribution were negligible by comparison to the corresponding effects stemming from uncertainties in the total cross sections.
Neglecting the second-order sensitivities would cause an erroneous reporting of the response’s expected value and also a very large non-conservative error by under-reporting of the response variance. For example, if the parameters were uncorrelated and had a uniform standard deviation of 10%, neglecting second-order (and higher) sensitivities would cause a non-conservative error by under-reporting of the response variance by a factor of 947%. If the cross sections were fully correlated, neglecting the second-order sensitivities would cause an error as large as 2000% in the expected value of the leakage response, and up to 6000% in the variance of the leakage response. In all cases, neglecting the second-order sensitivities would erroneously predict a Gaussian distribution in parameter space (for the PERP leakage response) centered around the computed value of the leakage response. In reality, the second-order sensitivities cause the leakage distribution in parameter space to be skewed towards positive values relative to the expected value, which, in turn, is significantly shifted to much larger positive values than the computed leakage value.
The above-mentioned results for the second-order sensitivities indicated that the third-order sensitivities of the PERP leakage response with respect to the total cross sections (which were the most important parameters affecting the PERP leakage response) needed to be computed in order to assess their impact on the response and compare this impact to the corresponding impact of the second-order sensitivities. Recall that the PERP leakage response admits 180 first-order sensitivities and 32,400 s-order sensitivities (of which 16,290 are distinct from one another) with respect to the total cross sections. It follows that there are 5,832,000 third-order sensitivities (of which 988,260 are distinct from each other) of the leakage response (just) with respect to the total cross sections. The results for the third-order sensitivities of the PERP leakage response and their effects on the response’s expected value, variance, and skewness were reported in [26,27]. These results indicated that many third-order sensitivities were considerably larger than the corresponding second-order ones, which underscored the need to compute the fourth-order sensitivities corresponding to the largest third-order ones. There are 1,049,760,000 fourth-order sensitivities (of which 45,212,895 are distinct from each other) of the leakage response with respect to the total cross sections. The general methodology for computing fourth-order sensitivities exactly and efficiently while overcoming the curse of dimensionality has been developed by Cacuci [28] and applied to the PERP benchmark. It was found [29,30] that the microscopic total cross sections of isotopes 1H and 239Pu are the most important parameters affecting the PERP benchmark’s leakage response, since they are involved in all of the large values (i.e., greater than 1.0) of the first-, second-, third- and fourth-order unmixed relative sensitivities. In particular, the largest unmixed fourth-order sensitivity is S ( 4 ) ( σ t , 6 g = 30 , σ t , 6 g = 30 , σ t , 6 g = 30 , σ t , 6 g = 30 ) = 2.720 × 10 6 with respect to σ t , 6 30 for the 30th energy group of isotope 1H. The 30th energy group comprises thermalized neutrons in the energy interval from 1.39 × 10−4 eV to 0.152 eV. This sensitivity is about 90 times larger than the corresponding largest third-order relative sensitivity, about 6350 times larger than the corresponding largest second-order sensitivity, and about 291,000 times larger than the corresponding largest first-order relative sensitivity. The largest mixed fourth-order relative sensitivity is S ( 4 ) ( σ t , 6 30 , σ t , 6 30 , σ t , 6 30 , σ t , 5 30 ) = 2.279 × 10 5 , which involves the 30th energy group of the microscopic total cross sections for isotopes #6 (1H) and 5 (12C). Thus, among the 180 microscopic total cross sections, the parameter σ t , 6 g = 30 , i.e., the total cross sections of isotope #6 (1H) in the lowest energy group (30th), is the single most important parameter affecting the PERP benchmark’s leakage response, as it has the largest impact on the various response moments. The comparative impact of the sensitivities of various orders, from first order to fourth order, are summarized in Section 3.1.2 below.

3.1.2. Overview of the Impacts of the First-, Second-, Third-, and Fourth-Order Sensitivities on Uncertainties Induced in the PERP’s Leakage Response

The effects of the first-, second-, third-, and fourth-order sensitivities on the expected value of the PERP benchmark’s leakage response are presented in Table 3, considering that the total cross sections are normally distributed (indicated by the superscript “N”) and uncorrelated (indicated by the superscript “U”). As shown in Table 3, the effects of the second-order and fourth-order sensitivities through [ E ( L ) ] t ( 2 , U , N ) 2.5 % × [ E ( L ) ] t ( U , N ) and [ E ( L ) ] t ( 4 , U , N ) 0.3 % × [ E ( L ) ] t ( U , N ) to the expected response value [ E ( L ) ] t ( U , N ) are negligibly small when considering a small relative standard deviation of 1% for each of the uncorrelated microscopic total cross sections of the isotopes included in the PERP benchmark. Considering a moderate (5%) relative standard deviation for each of the microscopic total cross sections, the results presented in Table 3 show that [ E ( L ) ] t ( 2 , U , N ) 65 % × L ( α 0 ) 17 % × [ E ( L ) ] t ( U , N ) indicate that the contributions from the second-order sensitivities to the expected response amount to ca. 65% of the computed leakage value L ( α 0 ) , and contribute ca. 17% to the expected value [ E ( L ) ] t ( U , N ) of the leakage response. Furthermore, the fourth-order sensitivities contribute around 56% to the expected value [ E ( L ) ] t ( U , N ) , since [ E ( L ) ] t ( 4 , U , N ) 213 % × L ( α 0 ) 56 % × [ E ( L ) ] t ( U , N ) . Hence, if the computed value, L ( α 0 ) , is considered to be the actual expected value of the leakage response, neglecting the fourth-order sensitivities would produce an error of ca. 210% for 5% relative standard deviations for uncorrelated total cross sections.
The rightmost column in Table 3 presents the results when considering a relative standard deviation of 10% for each of the uncorrelated microscopic total cross sections. These results indicate that [ E ( L ) ] t ( 2 , U , N ) 260 % × L ( α 0 ) 7 % × [ E ( L ) ] t ( U , N ) , which implies that contributions of the second-order term are about 2.6 times larger than the computed leakage value L ( α 0 ) , and contribute around 7% to the expected value [ E ( L ) ] t ( U , N ) . For the fourth-order term, the results in Table 3 show that [ E ( L ) ] t ( 4 , U , N ) 3414 % × L ( α 0 ) 90 % × [ E ( L ) ] t ( U , N ) , indicating that the contributions from the fourth-order sensitivities to the expected response are around 34 times larger than the computed leakage value L ( α 0 ) and contribute about 90% to the expected value [ E ( L ) ] t ( U , N ) . Thus, for 10% relative standard deviations for uncorrelated total cross sections, the computed value L ( α 0 ) would be ca. 3400% in error by comparison to the actual expected value of the leakage if the fourth-order sensitivities were neglected.
In summary, the impact of the fourth-order sensitivities on the expected value of the leakage response varies with the value of the standard deviation of the uncorrelated microscopic total cross sections. Generally, the larger the standard deviations of the microscopic total cross sections, the higher the impact of the fourth-order sensitivities will be on the expected value. For a small relative standard deviation of 1% for the parameters under consideration, the impact of the fourth-order sensitivities on the expected response value [ E ( L ) ] t ( U , N ) is smaller than the impact of the lower-order sensitivities. However, for a moderate relative standard deviation of 5%, the contributions from the fourth-order sensitivities are around 56% of the expected value. When the relative standard deviation is increased to 10%, the contributions from the fourth-order sensitivities to the expected value increase to nearly 90%. Notably, for the “RSD = 10%” case, neglecting the fourth-order sensitivities would cause a large error (ca. 3400%) if the computed value L ( α 0 ) were considered to be the actual expected value of the leakage response.
The effects of the first-, second-, third-, and fourth-order sensitivities on the variance of the leakage response are illustrated by the results presented in Table 4. These effects are obtained when considering parameter relative standard deviations of 1%, 5%, and 10%, respectively, and considering that the total cross sections are normally distributed and uncorrelated. For uniform relative standard deviations of 1% for the uncorrelated microscopic total cross sections, the results presented in Table 4 indicate that [ var   ( L ) ] t ( 1 , U , N ) 70 % × [ var   ( L ) ] t ( U , N ) ,   [ var   ( L ) ] t ( 2 , U , N ) 6 % × [ var   ( L ) ] t ( U , N ) , [ var   ( L ) ] t ( 3 , U , N ) 20 % × [ var   ( L ) ] t ( U , N ) , and [ var   ( L ) ] t ( 4 , U , N ) 4 % × [ var   ( L ) ] t ( U , N ) . These results indicate that, for very small relative standard deviations (e.g., 1%), the contributions from the first-order sensitivities to the response variance are significantly larger (ca. 70%) than those from higher order sensitivities.
For uniform relative standard deviations of 5% for the uncorrelated microscopic total cross sections, the results presented in Table 4 indicate that S D ( 4 ) > S D ( 3 ) S D ( 2 ) > S D ( 1 ) . Thus, the contributions from the third- and fourth-order sensitivities to the response variance are remarkably larger than those from the first- and second-order ones: [ var   ( L ) ] t ( 1 , U , N ) 2 % × [ var   ( L ) ] t ( U , N ) ,   [ var   ( L ) ] t ( 2 , U , N ) 3 % × [ var   ( L ) ] t ( U , N ) , [ var   ( L ) ] t ( 3 , U , N ) 43 % × [ var   ( L ) ] t ( U , N ) , and [ var   ( L ) ] t ( 4 , U , N ) 52 % × [ var   ( L ) ] t ( U , N ) . Hence, neglecting the fourth-order sensitivities would cause a significant error in quantifying the standard deviation of the leakage response for the PERP benchmark.
The results presented in the last column of Table 4 indicate that, for relative standard deviations of 10% for the uncorrelated microscopic total cross sections, [ var   ( L ) ] t ( 1 , U , N ) 0.1 % × [ var   ( L ) ] t ( U , N ) ,   [ var   ( L ) ] t ( 2 , U , N ) 0.9 % × [ var   ( L ) ] t ( U , N ) , [ var   ( L ) ] t ( 3 , U , N ) 40 % × [ var   ( L ) ] t ( U , N ) , and [ var   ( L ) ] t ( 4 , U , N ) 59 % × [ var   ( L ) ] t ( U , N ) . Thus, the contributions from the third- and fourth-order sensitivities amount to ca. 99% of the total contribution to the response variance, while the contributions stemming from the first- and second-order sensitivities are negligibly small by comparison. The diverging trend displayed by the standard deviations of the leakage response as the order of sensitivity increases, namely, S D ( 4 ) > S D ( 3 ) S D ( 2 ) > S D ( 1 ) , for the “10%” case is similar to the “5% case” above, but with much larger amplitudes.
The effects of the first-, second-, third-, and fourth-order sensitivities on the standard deviations of the leakage response when considering 10% relative standard deviations for the uncorrelated microscopic total cross sections are illustrated in Figure 2, which presents plots for L ( α 0 ) ± S D ( 1 ) , [ E ( L ) ] t ( U , N ) ± S D ( 1 ) , [ E ( L ) ] t ( U , N ) ±   S D ( 2 ) , [ E ( L ) ] t ( U , N ) ±   S D ( 3 ) , and [ E ( L ) ] t ( U , N ) ±   S D ( 4 ) . Very importantly, the negative value of L ( α 0 ) S D ( 1 ) shown in Figure 2 indicates that considering the computed value L ( α 0 ) to be the expected value of the leakage response would produce unphysical results for 10% relative standard deviations for the microscopic total cross sections.
In summary, the impact of the sensitivities of various orders on the standard deviation of the leakage response is as follows: (i) for small (1%) relative standard deviations for the microscopic total cross sections, the uncertainty of the leakage response arising solely from the first-order sensitivities are significantly larger than the uncertainties arising solely from the second-, third-, and fourth-order sensitivities, respectively, but following an oscillating pattern: S D ( 1 ) S D ( 3 ) > S D ( 2 ) > S D ( 4 ) ; (ii) when considering moderate (5%) and larger (10%) relative standard deviations for the microscopic total cross sections, the standard deviation of the leakage response increases substantially as successively higher-order sensitivities are considered, leading to the following sequence of inequalities, S D ( 4 ) > S D ( 3 ) S D ( 2 ) > S D ( 1 ) , which indicates the need to verify the convergence/divergence of the Taylor series which underlies the expression of the response’s standard deviation.

3.1.3. Effects of the Fourth-Order Sensitivities on the Third-Order Response Moment and Skewness

The results obtained for the third-order response moment and skewness when considering relative standard deviations of 1%, 5%, and 10%, respectively, for the microscopic total cross sections (considered to be normally distributed and uncorrelated) are summarized in Table 5.
Considering 1% relative standard deviations for the uncorrelated microscopic total cross sections, the results presented in Table 5 indicate that [ μ 3   ( L ) ] t ( 2 , U , N ) 53 % × [ μ 3   ( L ) ] t ( U , N ) , [ μ 3   ( L ) ] t ( 3 , U , N ) 31 % × [ μ 3   ( L ) ] t ( U , N ) , and [ μ 3   ( L ) ] t ( 4 , U , N ) 16 % × [ μ 3   ( L ) ] t ( U , N ) . Thus, for small (1%) relative standard deviations, the contributions to the third-order response moment [ μ 3   ( L ) ] t ( U , N ) stemming from the second-order sensitivities are the largest (e.g., around 53% in this case), followed by the contributions stemming from the third-order sensitivities, while the contributions stemming from the fourth-order sensitivities are the smallest.
Considering 5% relative standard deviations for the uncorrelated microscopic total cross sections, the results in Table 5 show that [ μ 3   ( L ) ] t ( 2 , U , N ) 10 % × [ μ 3   ( L ) ] t ( U , N ) , [ μ 3   ( L ) ] t ( 3 , U , N ) 60 % × [ μ 3   ( L ) ] t ( U , N ) , and [ μ 3   ( L ) ] t ( 4 , U , N ) 30 % × [ μ 3   ( L ) ] t ( U , N ) . In this case, the contributions from the third-order sensitivities are the largest (e.g., around 60%), followed by the contributions from the fourth-order sensitivities, which contribute about 30%; the smallest contributions stem from the second-order sensitivities.
Considering 10% relative standard deviations for the uncorrelated microscopic total cross sections, the results in Table 5 indicate that [ μ 3   ( L ) ] t ( 2 , U , N ) 8 % × [ μ 3   ( L ) ] t ( U , N ) , [ μ 3   ( L ) ] t ( 3 , U , N ) 61 % × [ μ 3   ( L ) ] t ( U , N ) , and [ μ 3   ( L ) ] t ( 4 , U , N ) 31 % × [ μ 3   ( L ) ] t ( U , N ) . These results display the same trends as the results for the “RSD = 5%” case, but the magnitudes of the respective contributions are significantly larger by comparison to the corresponding results for the “RSD = 5%” case.
For the response skewness, [ γ 1 ( L ) ] t ( U , N ) , the results shown in in the last row of Table 5 indicate that all the second-, third-, and fourth-order sensitivities produce a positive response skewness, which causes the leakage response distribution to be skewed towards the positive direction from its expected value [ E ( L ) ] t ( U , N ) . The results shown in the last row of Table 5 also indicate that as the relative standard deviation of the uncorrelated microscopic total cross sections increases from 1% to 10%, the value of the skewness decreases, thus causing the leakage response distribution to become increasingly more symmetrical about the mean value [ E ( L ) ] t ( U , N ) . Neglecting the fourth-order sensitivities would cause a significant error in the skewness. For example, for the case “RSD = 5%”, if the fourth-order sensitivities were neglected, the contributions from the first-, second-, and third-order sensitivities to the skewness would have the value [ γ 1 ( L ) ] t ( U , N ) = 0.109 , which would be 135% smaller than the more accurate result obtained by including the contributions of the fourth-order sensitivities. The fourth-order sensitivities produce a positive response skewness, causing the leakage response distribution to be skewed towards the positive direction from its expected value. The impact of the fourth-order sensitivities on the skewness of the leakage response also changes with the value of the standard deviation of the microscopic total cross sections: larger standard deviations (for the parameters) tend to decrease the value of the skewness, causing the leakage response distribution to become more symmetrical about the mean value   [ E ( L ) ] t ( U , N ) .
A “first-order sensitivity and uncertainty quantification” will always produce an erroneous third moment (and, hence, skewness) of the predicted response distribution, unless the unknown response distribution happens to be symmetrical. At least second-order sensitivities must be used in order to estimate the third-order moment (and, hence, the skewness) of the response distribution. With pronounced skewness, standard statistical inference procedures such as constructing a confidence interval for the mean (expectation) of a computed/predicted model response will be not only incorrect, in the sense that the true coverage level will differ from the nominal (e.g., 95%) level, but the error probabilities will be unequal on each side of the predicted mean. Thus, the truncation of Taylor expansion of the response (as a function of parameters) depends both on the magnitudes of the response sensitivities to parameters and the parameter uncertainties involved: if the uncertainties are small, then a fourth-order expansion suffices, in most cases, for obtaining relatively accurate results. In any case, the truncation error of a convergent Taylor series can be quantified a priori. If the parameter uncertainties are large, the Taylor series may diverge, so one would need to reduce such uncertainties, e.g., by performing additional measurements of the respective parameters. Of course, if the parameter uncertainties are large, all statistical methods are doomed to produce unreliable results for large-scale, realistic problems involving many uncertain parameters.
Obtaining the exact and complete set of first-order sensitivities of responses to model parameters is therefore of paramount importance for any analysis of a computational model. The second-order sensitivities contribute the leading correction terms to the response’s expected value, causing it to differ from the response’s computed value. The second-order sensitivities also contribute to the response variances and covariances. If the parameters follow a normal (Gaussian) multivariate distribution, the second-order sensitivities contribute the leading terms to the response’s third-order moment. In particular, the skewness of a single response, R , is customarily denoted as γ 1 ( R ) , and is defined as follows: γ 1 ( R ) = μ 3 ( R ) / [ var ( R ) ] 3 2 , where μ 3 ( R ) denotes the third central moment of the response distribution. Thus, neglecting the second-order response sensitivities to normally distributed parameters would nullify the third-order response correlations and hence would nullify the skewness of a response. Hence, a “first-order sensitivity and uncertainty quantification” will always produce a zero-valued skewness of the predicted response distribution, which would be erroneous unless the unknown response distribution happens to be symmetrical. At least second-order sensitivities must be used in order to estimate the third-order moment and, hence, the skewness of the response distribution. The skewness provides a quantitative measure of the asymmetries in the respective distribution, indicating the direction and relative magnitude of a distribution’s deviation from the normal distribution. With pronounced skewness, standard statistical inference procedures such as constructing a confidence interval for the mean (expectation) of a computed/predicted model response will be incorrect, in the sense that the true coverage level will differ from the nominal (e.g., 95%) level. Furthermore, the error probabilities will be unequal on each side of the predicted mean.

3.2. Overview of the nth-Order Comprehensive Adjoint Sensitivity Analysis Methodology for Response-Coupled Forward and Adjoint Linear Systems (nth-CASAM-L)

The nth-Order Comprehensive Adjoint Sensitivity Analysis Methodology for Response-Coupled Forward and Adjoint Linear Systems (nth-CASAM-L), which was originally presented by Cacuci [36], is briefly summarized in this subsection. As was mentioned in Section 3.1, the mathematical/computational model of a physical system comprises model parameters which are imprecisely known. In addition to the TP model parameters α ( α 1 , , α T P ) described in Section 3.1, the mathematical model of a physical system is considered to comprise T I independent variables which will be denoted as x i ,     i = 1 ,   ,   T I , and are considered to be the components of a T I -dimensional column vector denoted as x ( x 1 , , x T I ) T I , where the sub/superscript “ T I ” denotes the “Total (number of) Independent variables” and where the components x i ,     i = 1 ,   ,   T I , can be considered, without loss of generality, to be real quantities. The vector x T I of independent variables is considered to be defined on a phase-space domain, denoted as Ω ( α ) , Ω ( α ) { λ i ( α ) x i ω i ( α ) ;     i = 1 ,   ,   T I } , the boundaries of which may depend on (some of) the model parameters α . The lower boundary point of an independent variable is denoted as λ i ( α ) (e.g., the inner radius of a sphere or cylinder, the lower range of an energy variable, the initial time value, etc.), while the corresponding upper boundary point is denoted as ω i ( α ) (e.g., the outer radius of a sphere or cylinder, the upper range of an energy variable, the final time value, etc.). The boundary of Ω ( α ) is denoted as Ω ( α ) { λ i ( α )   ω i ( α ) ,     i = 1 , , T I } ; it comprises the set of all of the endpoints λ i ( α ) ,     ω i ( α ) , i = 1 ,   ,   T I .
A linear physical system is generally modeled by a system of coupled linear operator equations which can be generally represented in block-matrix form, as follows:
L ( x ; α ) φ ( x ) = q φ ( x ; α ) ,       x Ω ( α )
where φ ( x ) [ φ 1 ( x ) , , φ T D ( x ) ] is a T D -dimensional column vector of dependent variables and where the sub/superscript “ T D ” denotes the “Total (number of) Dependent variables.” The functions φ i ( x ) ,     i = 1 , , T D , denote the system’s “dependent variables” (also called “state functions”). The components L i j ( x ; α ) ,   i , j = 1 , , T D , of the T D × T D -dimensional matrix L ( x ; α )   [ L i j ( x ; α )   ] are operators that act linearly on the dependent variables φ j ( x ) and also depend on the uncertain model parameters α . The components of the T D -dimensional column vector q φ ( x ; α )   [ q φ , 1 ( x ; α ) , , q φ , T D ( x ; α ) ] represent inhomogeneous source terms which are considered to depend (nonlinearly) on the model parameters. Since the sources on the right side of Equation (27) may comprise distributions, all equalities that involve differential equations are considered to hold in the weak (i.e., “distributional”) sense.
Since the complete mathematical model is considered to be linear in φ ( x ) , the boundary and/or initial conditions needed to define the domain of L ( x ; α )   ; when L ( x ; α )   contains differential operators, it must also be linear in φ ( x ) and can be represented as follows:
B φ ( x ; α )   φ ( x ) - c φ ( x ; α ) = 0 ,         x Ω x ( α ) .
In (28), the components B i j ( x ; α ) ;     i = 1 , , N B ;     j = 1 , , T D , of the N B × T D -dimensional matrix operator B φ ( x ; α ) act linearly on φ ( x ) and nonlinearly on α . The total number of boundary and initial conditions is denoted as N B . The operator c φ ( x ; α ) [ c φ , 1 ( x ; α ) , , c φ , N B ( x ; α ) ] is a N B -dimensional vector comprising inhomogeneous boundary source terms that act nonlinearly on α . The subscript “ φ ” in Equation (28) indicates boundary conditions associated with the forward state function φ ( x ) .
Physical problems modeled by linear systems and/or operators are naturally defined in Hilbert spaces. The dependent variables φ i ( x ) ,     i = 1 , , T D , for the physical system represented by Equations (27) and (28) are considered to be square-integrable functions of the independent variables, belonging to a Hilbert space which is denoted as H 0 (where the subscript “zero” denotes “zeroth-level“ or “original”). The Hilbert space H 0 is considered to be endowed with the following inner product, denoted as φ ( x ) , ψ ( x ) 0 , between two elements φ ( x ) H 0 and ψ ( x ) H 0 :
φ ( x ) , ψ ( x ) 0 Ω x φ ( x ) · ψ ( x ) d x i = 1 T I λ i ( α ) ω i ( α ) φ ( x ) · ψ ( x )   d x                           = i = 1 T D λ 1 ( α ) ω 1 ( α ) λ i ( α ) ω i ( α ) λ T I ( α ) ω T I ( α ) φ i ( x ) ψ i ( x )   d x 1     d x 2   d x i d x T I .
The product notation i = 1 T I λ i ( α ) ω i ( α ) [     ]     d x i in Equation (29) denotes the respective multiple integrals. The linear operator L ( x ; α )   admits an adjoint operator, which is denoted as L * ( x ; α ) and which is defined through the following relation for a vector ψ ( x ) H 0 :
ψ ( x ) ,     L ( x ; α )     φ ( x ) 0 = L * ( x ; α )     ψ ( x ) ,   φ ( x ) 0
In (30), the formal adjoint operator L * ( x ; α )   is the T D × T D matrix comprising elements L j i * ( x ; α )   obtained by transposing the formal adjoints of the forward operators L i j ( x ; α )   , i.e.,
L * ( x ; α )   [ L j i * ( x ; α )   ] ,       i , j = 1 , , T D ,
Hence, the system adjoint to the linear system represented by (27) and (28) can generally be represented as follows:
L * ( x ; α )   ψ ( x ) = q ψ ( x ; α ) ,       x Ω x ( α )   ,
B ψ ( x ; α )   ψ ( x )   - c ψ ( x ; α )   = 0 ,       x Ω x ( α )   .
When L ( x ; α )   comprises differential operators, the operations (e.g., integration by parts) that implement the transition from the left side to the right side of Equation (30) give rise to boundary terms which are collectively called the “bilinear concomitant.” The domain of L * ( x ; α )   is determined by selecting adjoint boundary and/or initial conditions so as to ensure that the bilinear concomitant vanishes when the selected adjoint boundary conditions are implemented together with the forward boundary conditions given in Equation (28). The adjoint boundary conditions thus selected are represented in operator form by Equation (33), where the subscript “ ψ ” indicates adjoint boundary and/or initial conditions associated with the adjoint state function ψ ( x ) . In most practical situations, the Hilbert space H 0 is self-dual. Solving Equations (27) and (28) at the nominal parameter values α 0 yields the nominal value φ 0 ( x ) of the forward state function, while solving Equations (32) and (33) at α 0 yields the nominal value ψ 0 ( x ) of the adjoint state function.
The relationship shown in Equation (30), which is the basis for defining the adjoint operator, also provides the following fundamental “reciprocity-like” relation between the sources of the forward and the adjoint equations, cf. Equations (27) and (32), respectively:
ψ ( x ) ,     q φ ( x ; α ) 0 = q ψ ( x ; α ) ,   φ ( x ) 0
The functional on the right side of Equation (34) represents a “detector response”, i.e., a reaction rate between the particles and the medium represented by q ψ ( x ; α ) , which is equivalent to the “number of counts” of particles incident on a detector of particles that “measures” the particle flux φ ( x ) . In view of the relation provided in Equation (34), the source term q ψ ( x ; α )     [ q ψ , 1 ( x ; α ) , , q ψ , T D ( x ; α ) ] in the adjoint Equation (32) is usually associated with the “result of interest” to be measured and/or computed, which is customarily called the system’s “response.” In particular, if q ψ ( x ; α ) = δ ( x x d ) , then q ψ ( x ; α ) ,   φ ( x ) 0 = φ ( x d ) , which means that, in such a case, the right side of (34) provides the value of the dependent variable (particle flux, temperature, velocity, etc.) at the point in the phase space where the respective measurement is performed.
Since the linear operators that underly linear systems admit adjoint operators, important model responses of linear systems can be functions of both the forward and the adjoint state functions. Examples of such responses are the Schwinger [37] and Roussopoulos [38] functionals, as well as the Raleigh quotient [39] for computing eigenvalues of linear systems. Such responses cannot be encountered in nonlinear problems since nonlinear operators do not admit adjoint operators. In this sense, linear systems differ fundamentally from nonlinear systems and cannot always be considered to be particular cases of nonlinear systems but need to be treated comprehensively in their own right. Thus, the fundamental form of a response computed using a linear model is an operator which acts nonlinearly on the model’s forward and adjoint state functions, as well as on the imprecisely known model parameters, directly and indirectly through the state functions, having the following generic form:
R [ φ ( x ) , ψ ( x ) ; α ] i = 1 T I λ i ( α ) ω i ( α ) S   [ φ ( x ) , ψ ( x ) ; α ] d x i ,
where S   [ φ ( x ) , ψ ( x ) ; α ] denotes a suitably Gateaux (G)-differential function of the indicated arguments. In general, S   [ φ ( x ) , ψ ( x ) ; α ] is nonlinear in φ ( x ) , ψ ( x ) , and α , and the components of α are considered to also include parameters that may specifically appear only in the definition of the response under consideration (but which might not appear in the definition of the model). Thus, the (physical) “system” is defined and understood to comprise both the system’s computational model and the system’s response. Due to the response’s implicit dependence on the model parameters through the forward and/or adjoint state functions, the response of a linear system is a nonlinear function of the model’s parameters, except perhaps in particularly trivial situations. The nominal value of the response, R [ φ 0 ( x ) , ψ 0 ( x ) ; α 0 ] , is determined by using the nominal parameter values α 0 , the nominal value φ 0 ( x ) of the forward state function, and the nominal value ψ 0 ( x ) of the adjoint state function.
Since the model parameters α i are imprecisely known quantities, their actual values (which are not known in practice) will differ from their nominal values (which are known in practice) by quantities denoted as δ α i α i α i 0 ,     i = 1 , , T P . The variations δ α ( δ α 1 , , δ α T P ) will cause corresponding variations δ φ ( δ φ 1 ,   ,   δ φ T D ) ,   δ φ i φ i φ i 0 ,   δ ψ ( δ ψ 1 , , δ ψ T D ) ,   δ ψ i ψ i ψ i 0 ,     i = 1 , , T D , in the forward and adjoint state functions, respectively. In turn, all of these variations will cause a response variation R ( φ 0 + δ φ ; ψ 0 + δ ψ ; α 0 + δ α ) around the nominal value R [ φ 0 ( x ) , ψ 0 ( x ) ; α 0 ] . The first-order sensitivities of the response with respect to the model parameters are contained within the first-order Gateaux (G)-variation, denoted as δ R ( φ 0 , ψ 0 , α 0 ; δ φ , δ ψ , δ α ) , of the response R ( e ) for arbitrary variations δ φ ( x ) , δ ψ ( x ) , δ α in the model parameters and state functions, in a neighborhood [ φ 0 ( x ) + ε δ φ ( x ) , ψ 0 ( x ) + ε δ ψ ( x ) ; α 0 + ε δ α ] around ( φ 0 , ψ 0 , α 0 ) . By definition, the first-order Gateaux (G)-variation δ R ( φ 0 , ψ 0 , α 0 ; δ φ , δ ψ , δ α ) is obtained as follows:
δ R ( φ 0 , ψ 0 , α 0 ; δ φ , δ ψ , δ α ) { d d ε R [ φ 0 ( x ) + ε δ φ ( x ) , ψ 0 ( x ) + ε δ ψ ( x ) ; α 0 + ε δ α ] } ε = 0         { δ R ( φ 0 , ψ 0 , α 0 ; δ φ , δ ψ ) } d i r + { δ R ( φ 0 , ψ 0 , α 0 ; δ α ) } i n d ,
where the “direct-effect” term { δ R ( φ 0 , ψ 0 , α 0 ; δ α ) } d i r depends only on the parameter variation δ α and is defined as follows:
{ δ R ( φ 0 , ψ 0 , α 0 ; δ α ) } d i r { α k = 1 T I λ k ( α ) ω k ( α ) d x k S   [ φ ( x ) , ψ ( x ) ; α ] } α 0 δ α = k = 1 T I { κ k ( α ) ω k ( α ) d x k S   [ φ ( x ) , ψ ( x ) ; α ] α δ α } α 0 + k = 1 T I m = 1 , k j T I { λ m ( α ) ω m ( α ) d x m S   [ φ ( , ω k , , x N x ) , ψ ( , ω k , ) ; α ] ω k ( α ) α } α 0 δ α   k = 1 T I m = 1 , k j T I { λ m ( α ) ω m ( α ) d x m S   [ φ ( , λ k , ) , ψ ( , λ k , ) ; α ] λ k ( α ) α } α 0 δ α ,
and where the “indirect effect” term { δ R ( φ 0 , ψ 0 , α 0 ; δ φ , δ ψ ) } i n d depends only on the variations δ φ and δ ψ in the state functions, and is defined as follows:
{ δ R ( φ 0 , ψ 0 , α 0 ; δ φ , δ ψ ) } i n d                               k = 1 T I { λ k ( α ) ω k ( α ) d x k [ S   [ φ ( x ) , ψ ( x ) ; α ] φ δ φ + S   [ φ ( x ) , ψ ( x ) ; α ] ψ δ ψ ] } α 0 .
The notation { } α 0 has been used in Equations (37) and (38) to indicate that the quantity within the brackets is to be evaluated at the nominal values of the parameters and state functions. While the “direct effect” term { δ R ( φ 0 , ψ 0 , α 0 ; δ α ) } d i r defined in Equation (37), which depends only on the parameter variations δ α , can be computed immediately, the “indirect effect” term { δ R ( φ 0 , ψ 0 , α 0 ; δ φ , δ ψ ) } i n d defined in Equation (38) can be computed only after having determined the values of the variations δ φ ( x ) and δ ψ ( x ) . The variations δ φ ( x ) and δ ψ ( x ) are the solutions of the system of equations obtained by taking the G-differentials of Equations (27), (28), (32), and (33), which can be written in the following matrix-vector form:
{ V ( 1 ) ( α ) δ u ( 1 ) ( x ) } α 0 = { q ( 1 ) ( u ( 1 ) ; α ; δ α ) } α 0 ,           x Ω x ,
{ b v ( 1 ) ( u ( 1 ) ; α ; δ u ( 1 )   ; δ α ) } α 0 = 0 ,         x Ω x ( α 0 ) ,
where
V ( 1 ) ( α ) ( L ( α ) 0 0 L * ( α ) ) ;             u ( 1 ) ( x ) ( φ ( x ) ψ ( x ) ) ;           δ u ( 1 ) ( x ) ( δ φ ( x ) δ ψ ( x ) ) ;     q ( 1 ) ( u ( 1 ) ; α ; δ α )   ( q 1 ( 1 ) ( φ ; α ; δ α ) q 2 ( 1 ) ( ψ ; α ; δ α ) ) ;       b v ( 1 ) ( u ( 1 ) ;   δ u ( 1 )   ; α ; δ α )   ( b 1 ( 1 ) ( φ ; α ; δ α ) b 2 ( 1 ) ( ψ ; α ; δ α ) ) .
{ q 1 ( 1 ) ( φ , α ; δ α )   } α 0 { [ q φ ( α ) L ( α ) φ ] α   } α 0 δ α ,
{ q 2 ( 1 ) ( ψ , α ; δ α )   } α 0 { [ q ψ ( α ) L * ( α ) ψ ( x ) ] α   } α 0 δ α ,
The matrices q φ ( α ) / α and [ L ( α ) φ ] / α , which appear on the right side of Equation (39), are defined as follows:
q φ ( α ) α ( · · q φ , i α j · · )   ,     [ L ( α ) φ ] α ( · · [ k = 1 T D L i , k ( α ) φ k ] α j · · )   .
The system of equations comprising Equations (39) and (40) is called the “1st-Level Variational Sensitivity System” (1st-LVSS) and its solution, δ u ( 1 ) ( x ) , is called the “1st-Level variational sensitivity function”, as indicated by the superscript “(1)”. Since the source term, q ( 1 ) ( u ( 1 ) ; α ; δ α ) , is a function of δ α , it follows that the 1st-LVSS would need to be solved T P times in order to obtain the vector δ u ( 1 ) ( x ) which would correspond to each parameter variation δ α i , j 1 = 1 , , T P . The need for computing the functions δ φ ( x ) and δ ψ ( x ) can be avoided by applying the First-Order Comprehensive Adjoint Sensitivity Analysis Methodology (1st-CASAM-L), which enables the elimination of δ φ ( x ) and δ ψ ( x ) from the expression of the indirect effect term { δ R ( φ 0 , ψ 0 , α 0 ; δ φ , δ ψ ) } i n d defined in Equation (38). This elimination is achieved by implementing the following sequence of steps:
  • Introduce a Hilbert space, denoted as H 1 , comprising square-integrable function vector-valued elements of the form η ( 1 ) ( x ) [ η 1 ( 1 ) ( x ) , η 2 ( 1 ) ( x ) ] , η i ( 1 ) ( x ) [ η i , 1 ( 1 ) ( x ) , , η i , j ( 1 ) ( x ) , , η i , T D ( 1 ) ( x ) ] , i = 1 , 2 . The inner product between two elements, η ( 1 ) ( x ) H 1 , ξ ( 1 ) ( x ) H 1 , is denoted as η ( 1 ) ( x ) , ξ ( 1 ) ( x ) 1 and is defined as follows:
    η ( 1 ) ( x ) , ξ ( 1 ) ( x ) 1 i = 1 2 η i ( 1 ) ( x ) , ξ i ( 1 ) ( x ) 0
  • In the Hilbert H 1 , form the inner product of Equation (39) with a vector-valued function a ( 1 ) ( x ) [ a 1 ( 1 ) ( x ) , a 2 ( 1 ) ( x ) ] H 1 to obtain:
    { a ( 1 ) ( x ) , V ( 1 ) ( α ) δ u ( 1 ) ( x ) 1 } α 0 = { a ( 1 ) ( x ) , q ( 1 ) ( φ ; ψ ; α ; δ α )   1 } α 0
  • Using the definition of the adjoint operator in the Hilbert space H 1 , recast the left side of Equation (46) as follows:
    { a ( 1 ) ( x ) , V ( 1 ) ( α ) δ u ( 1 ) ( x ) 1 } α 0 = { δ u ( 1 ) ( x ) , A ( 1 ) ( α ) a ( 1 ) ( x )   1 } α 0 + { P ( 1 ) [ δ u ( 1 ) ( x ) ; a ( 1 ) ( x ) ; α ; δ α ] } α 0 ,
    where { P ( 1 ) [ δ u ( 1 ) ( x ) ; a ( 1 ) ( x ) ; α ; δ α ] } α 0 denotes the bilinear concomitant defined on the phase-space boundary x Ω x ( α 0 ) , and where A ( 1 ) ( α ) is the operator formally adjoint to V ( 1 ) ( α ) , defined as follows:
    A ( 1 ) ( α ) [ V ( 1 ) ( α ) ] * = ( L * ( α ) 0 0 L ( α ) )
  • Require the first term on right side of Equation (47) to represent the indirect effect term defined in Equation (38), to obtain the following relation:
    A ( 1 ) ( α ) a ( 1 ) ( x )   = s ( 1 ) (   u ( 1 ) ( x ) ; α ) ,
    where
    s ( 1 ) [ u ( 1 ) ( x ) ; α ] ( [ s 1 ( 1 ) (   u ( 1 ) ; α ) ] [ s 2 ( 1 ) (   u ( 1 ) ; α ) ] ) ( [ S (   u ( 1 ) ; α ) / φ ] [ S (   u ( 1 ) ; α ) / ψ ] )     .
  • Implement the boundary conditions given in Equation (40) into Equation (47) and eliminate the remaining unknown boundary values of the functions δ φ and δ ψ from the expression of the bilinear concomitant { P ( 1 ) [ δ u ( 1 ) ( x ) ; a ( 1 ) ( x ) ; α ] } α 0 by selecting appropriate boundary conditions for the function a ( 1 ) ( x ) [ a 1 ( 1 ) ( x ) , a 2 ( 1 ) ( x ) ] , to ensure that Equation (49) is well-posed while being independent of unknown values of δ φ , δ ψ , and δ α . The boundary conditions thus chosen for the function a ( 1 ) ( x ) [ a 1 ( 1 ) ( x ) , a 2 ( 1 ) ( x ) ] can be represented in operator form as follows:
    { b A ( 1 ) [   u ( 1 ) ( x ) ; a ( 1 ) ( x ) ; α ; ] } α 0 = 0 ,       x Ω x ( α 0 ) .
  • The selection of the boundary conditions for the adjoint function a ( 1 ) ( x ) [ a 1 ( 1 ) ( x ) , a 2 ( 1 ) ( x ) ] represented by Equation (51) eliminates the appearance of the unknown values of δ u ( 1 ) ( x ) in the bilinear concomitant { P ( 1 ) [ δ u ( 1 ) ( x ) ; a ( 1 ) ( x ) ; α ; δ α ] } α 0 and reduces it to a residual quantity which will be denoted as { P ^ ( 1 ) [   u ( 1 ) ( x ) ; a ( 1 ) ( x ) ; α ; δ α ] } α 0 and which contains boundary terms involving only known values of φ ( x ) , ψ ( x ) , a ( 1 ) ( x ) , α , and δ α . This residual quantity does not automatically vanish.
  • The system of equations comprising Equation (49) together with the boundary conditions represented in Equation (51) constitute the First-Level Adjoint Sensitivity System (1st-LASS). The solution a ( 1 ) ( x ) [ a 1 ( 1 ) ( x ) , a 2 ( 1 ) ( x ) ] of the 1st-LASS is called the First-level adjoint sensitivity function. The 1st-LASS is called “first-level” (as opposed to “first-order”) because it does not contain any differential or functional derivatives, but its solution, a ( 1 ) ( x ) , is used below to compute the first-order sensitivities of the response with respect to the model parameters.
  • It follows from (46) and (47) that the following relation holds:
    { a ( 1 ) ( x ) , q ( 1 ) (   u ( 1 ) ; α ; δ α )   1 } α 0 = { δ u ( 1 ) ( x ) , A ( 1 ) ( α ) a ( 1 ) ( x )   1 } α 0                               + { P ^ ( 1 ) [   u ( 1 ) ( x ) ; a ( 1 ) ( x ) ; α ; δ α ] } α 0 .
  • Recalling from (49) that the first term on the right side of (52) is the indirect effect term { δ R ( φ 0 , ψ 0 , α 0 ; δ φ , δ ψ ) } i n d , it follows from (52) that the indirect effect term can be expressed in terms of the first-level adjoint sensitivity function a ( 1 ) ( x ) [ a 1 ( 1 ) ( x ) , a 2 ( 1 ) ( x ) ] as follows:
    { δ R ( φ 0 , ψ 0 , α 0 ; δ φ , δ ψ ) } i n d = { a ( 1 ) ( x ) , q ( 1 ) (   u ( 1 ) ; α ; δ α )   1 } α 0   { P ^ ( 1 ) [ u ( 1 ) ; a ( 1 ) ; α ; δ α ] } α 0 { δ R ( u ( 1 ) ; a ( 1 ) ; α ; δ α ) } i n d .
As indicated by the identity relation in Equation (53), the variations δ φ and δ ψ have been eliminated from the original expression of the indirect effect term, which now instead depends on the first-level adjoint sensitivity function a ( 1 ) ( x ) [ a 1 ( 1 ) ( x ) , a 2 ( 1 ) ( x ) ] . Very importantly, the 1st-LASS is independent of parameter variations δ α . Hence, the 1st-LASS needs to be solved only once (as opposed to the 1st-LVSS or the 1st-OFSS, which would need to be solved anew for each parameter variation) to determine the first-level adjoint sensitivity function a ( 1 ) ( x ) [ a 1 ( 1 ) ( x ) , a 2 ( 1 ) ( x ) ] . Subsequently, the indirect effect term { δ R ( u ( 1 ) ; a ( 1 ) ; α ; δ α ) } i n d is computed efficiently and exactly by simply performing the integrations over the adjoint function a ( 1 ) ( x ) [ a 1 ( 1 ) ( x ) , a 2 ( 1 ) ( x ) ] , as indicated by Equation (53).
Adding the expression of the direct effect term defined in Equation (37) to the expression of the indirect effect term obtained in Equation (53) yields the following expression for the total first-order G-differential { δ R [ u ( 1 ) ( x ) , α ; δ φ , δ ψ , δ α ] } α 0 of the response R [ φ ( x ) , ψ ( x ) ; α ] computed at the nominal values for the models’ parameter and state functions:
{ δ R [ u ( 1 ) ( x ) , α ; δ φ , δ ψ , δ α ] } α 0 = { k = 1 T I λ k ( α ) ω k ( α ) d x k S   [ φ ( x ) , ψ ( x ) ; α ] α δ α } α 0 + k = 1 T I m = 1 , k j T I { λ m ( α ) ω m ( α ) d x m S   [ φ ( , ω k , , x N x ) , ψ ( , ω k , ) ; α ] ω k ( α ) α } α 0 δ α   k = 1 T I m = 1 , k j T I { λ m ( α ) ω m ( α ) d x m S   [ φ ( , λ k , ) , ψ ( , λ k , ) ; α ] λ k ( α ) α } α 0 δ α , + { i = 1 2 a i ( 1 ) ( x ) , q i ( 1 ) ( u ( 1 ) ; α ; δ α ) 0 } α 0 { P ^ ( 1 ) [ u ( 1 ) ; a ( 1 ) ; α ; δ α ] } α 0 { δ R [ u ( 1 ) ( x ) ; a ( 1 ) ( x ) ; α ; δ α ] } α 0 .
The expression obtained in Equation (54) no longer depends on the variations δ φ and δ ψ but instead depends on the first-level adjoint sensitivity function a ( 1 ) ( x ) [ a 1 ( 1 ) ( x ) , a 2 ( 1 ) ( x ) ] . In particular, the expression in Equation (54) reveals that the sensitivities of the response R [ φ ( x ) , ψ ( x ) ; α ] to parameters that characterize the system’s boundary and/or internal interfaces can arise both from the direct effect and indirect effect terms. It also follows from Equation (54) that the quantity δ R ( φ 0 , ψ 0 ; a ( 1 ) ; α 0 ; δ α ) has become a total differential in the parameter variation δ α so it can be expressed in the following form:
{ δ R [ u ( 1 ) ( x ) ; a ( 1 ) ( x ) ; α ; δ α ] } α 0 = j 1 = 1 T P { R ( 1 ) [ j 1 ; u ( 1 ) ( x ) ; a ( 1 ) ( x ) ; α ] } α 0 δ α j 1 ,
where the quantities R ( 1 ) [ j 1 ; u ( 1 ) ( x ) ; a ( 1 ) ( x ) ; α ] R [ φ ( x ) , ψ ( x ) ; α ] / α j 1 represent the first-order sensitivities (i.e., partial functional derivatives) of the response with respect to the model parameter α j 1 , evaluated at the nominal values of the state functions and model parameters. Furthermore, each of the first-order sensitivities R ( 1 ) [ j 1 ; u ( 1 ) ( x ) ; a ( 1 ) ( x ) ; α ] can be represented in integral form as follows:
R ( 1 ) [ j 1 ; u ( 1 ) ( x ) ; a ( 1 ) ( x ) ; α ] k = 1 T I λ k ( α ) ω k ( α ) d x k S ( 1 )   [ j 1 ; u ( 1 ) ( x ) ; a ( 1 ) ( x ) ; α ] ,     j 1 = 1 , , T P .
In particular, if the residual bilinear concomitant does not vanish, the functions S ( 1 )   [ j 1 ; φ ( x ) , ψ ( x ) ; a ( 1 ) ( x ) ; α ] would contain suitably defined Dirac delta functionals for expressing the respective non-zero boundary terms as volume integrals over the phase space of the independent variables. Dirac delta functionals would also be needed to represent, within S ( 1 )   [ j 1 ; φ ( x ) , ψ ( x ) ; a ( 1 ) ( x ) ; α ] , the terms containing the derivatives of the boundary endpoints with respect to the model and/or response parameters.
Cacuci [36] has obtained the general expression of the arbitrarily high-order (nth-order, where n is an arbitrary integer) sensitivities of the response R [ φ ( x ) , ψ ( x ) ; α ] to model parameters by treating each sensitivity as a response, applying the principles outlined in items (1)−(9) to successively higher-order sensitivities and, ultimately, using mathematical induction to demonstrate that general expression thus obtained for the nth-order sensitivity is correct. Each of the (n − 1)th-order sensitivities ( n 1 ) R [ φ ( x ) , ψ ( x ) ; α ] / α j 1 α j n 1 R ( n 1 ) [ j n 1 , , j 1 ; u ( n 1 ) ( j n 1 , , j 1 ; x ) ; a ( n 1 ) ( j n 2 , , j 1 ; x ) ; α ] of the response R [ φ ( x ) , ψ ( x ) ; α ] with respect to the parameters α j 1 α j n 1 , where j 1 , j 2 , , j n 1 = 1 , , T P , are used in the role of a “model response”, for which the first-order G-differential is obtained by applying the principles outlined in items (1)−(9), thereby ultimately obtaining the general expression of the nth-order sensitivities of the response R [ φ ( x ) , ψ ( x ) ; α ] . The details of the nth-CASAM-L methodology are provided by Cacuci [36]. The end results produced by the nth-CASAM-L methodology enable the exact and efficient computation—overcoming the curse of dimensionality in sensitivity analysis—of the nth-order partial sensitivities of the response R [ φ ( x ) , ψ ( x ) ; α ] with respect to the model parameters, using the expression provided below:
{ δ R ( n 1 ) ( j n 1 , , j 1 ; u ( n ) ; α ; δ u ( n ) ; δ α ) } α 0 { d d ε R ( n 1 ) [ u ( n ) ( x ) + ε δ u ( n ) ; α 0 + ε δ α ] } ε = 0 = { P ^ ( n ) ( a ( n ) ; u ( n ) ; α ; δ α ) } α 0 + { α i = 1 T I λ i ( α ) ω i ( α ) d x i S ( n 1 ) ( j n 1 , , j 1 ; u ( n ) ; α ) } α 0 δ α   + { a ( n ) ( j n 1 , , j 1 ; x ) , q V ( n ) [ j n 2 , , j 1 ; u ( n ) ( j n 2 , , j 1 ; x ) ; α ; δ α ]   n } α 0 j n = 1 T P { R ( n ) ( j n , , j 1 ; u ( n ) ; a ( n ) ; α ) } ( α 0 ) δ α j n = j n = 1 T P { n R [ φ ( x ) , ψ ( x ) ; α ] / α j 1 α j n } ( α 0 ) δ α j n .
The various quantities which appear in Equation (57) are defined as follows:
(1)
The quantity R ( n ) ( j n , , j 1 ; u ( n ) ; a ( n ) ; α ) n R [ φ ( x ) , ψ ( x ) ; α ] / α j 1 α j n denotes the nth-order partial sensitivity of the response R [ φ ( x ) , ψ ( x ) ; α ] with respect to the model parameters, evaluated at the nominal values of the parameters and state functions. When the symmetries among the various partial derivatives/sensitivities are considered, the respective indices have ranges as follows: j 1 = 1 , , T P ; j 2 = 1 , , j 1 ; …; j n = 1 , , j n 1 . In integral form, the nth-order partial sensitivity of the response can be written as follows:
n R [ φ ( x ) , ψ ( x ) ; α ] / α j 1 α j n R ( n ) [ j n , , j 1 ; u ( n ) ( j n 2 , , j 1 ; x ) ; a ( n ) ( j n 1 , , j 1 ; x ) ; α ] i = 1 T I λ i ( α ) ω i ( α ) d x i S ( n ) [ j n , , j 1 ; u ( n ) ( j n 2 , , j 1 ; x ) ; a ( n ) ( j n 1 , , j 1 ; x ) ; α ] ,
(2)
The nth-level state u ( n ) ( j n 2 , , j 1 ; x ) , which appears in the arguments of R ( n ) ( j n , , j 1 ; u ( n ) ; a ( n ) ; α ) , is defined as follows:
u ( n ) ( j n 2 , , j 1 ; x ) [ φ ( x ) ; ψ ( x ) ; a 1 ( 1 ) ( x ) , a 2 ( 1 ) ( x ) ; a 1 ( 2 ) ( j 1 ; x ) , a 2 ( 2 ) ( j 1 ; x ) , a 3 ( 2 ) ( j 1 ; x ) , a 4 ( 2 ) ( j 1 ; x ) , , a 1 ( n 1 ) ( j n 2 , , j 1 ) , , a k ( n 1 ) ( j n 2 , , j 1 ) , , a 2 n 1 ( n 1 ) ( j n 2 , , j 1 ) ]
(3)
The inner product which appears in Equation (57) is defined in a Hilbert space denoted as H n , comprising 2 n -dimensional block vectors η ( n ) ( x ) H n having the following structure: η ( n ) ( x ) [ , η k ( n ) ( x ) , ] , η k ( n ) ( x ) [ η k , 1 ( n ) ( x ) , , η k , T D ( n ) ( x ) ] , k = 1 , , 2 n . The inner product between two elements, η ( n ) ( x ) H n and ξ ( n ) ( x )   H n , of the Hilbert space H n + 1 , is denoted as η ( n ) ( x ) , ξ ( n ) ( x ) n and is defined as follows:
η ( n ) ( x ) , ξ ( n ) ( x ) n k = 1 2 n η k ( n ) ( x ) , ξ k ( n ) ( x ) 0 .
(4)
The adjoint operator [ A ( n ) ] 2 n × 2 n which appears on the left side of Equation (62) is constructed by using the inner product defined in Equation (60), via the following relation:
{ a ( n ) ( j n 1 , , j 1 ; x ) , [ V ( n ) ] δ u ( n ) n } α 0 = { a ( n ) ( j n 1 , , j 1 ; x ) , q V ( n ) [ j n 2 j 1 ; u ( n ) ( j n 2 , , j 1 ; x ) ; α ; δ α ] n } α 0 = { δ u ( n ) ,     A ( n ) a ( n ) ( j n 2 , , j 1 ; x )   n } α 0 + { P ( n ) [ δ u ( n ) ; a ( n ) ; α ; δ α ] } α 0 ,
(5)
The nth-level adjoint sensitivity function a ( n ) ( j n 1 , , j 1 ; x ) [ , a k ( n ) ( j n 1 , , j 1 ; x ) , ] H n ;     k = 1 , , 2 n , which appears in the arguments of R ( n ) ( j n , , j 1 ; u ( n ) ; a ( n ) ; α ) , is the solution of the following nth-level adjoint sensitivity system nth-LASS, for j 1 = 1 , , T P ; j 2 = 1 , , j 1 ; … j n 1 = 1 , , j n 2 :
[ A ( n ) ] 2 n × 2 n a ( n ) ( j n 1 , , j 1 ; x )   = s A ( n ) ( j n 1 , , j 1 ; α )   ,
{ b A ( n ) [ a ( n ) ( j n 1 , , j 1 ; x ) ; u ( n ) ( j n 2 , , j 1 ; x ) ; α ] } α 0 = 0 ,   x Ω x ( α 0 ) ,
where the vector s A ( n ) ( j n 1 , , j 1 ; α ) [ , s k ( n ) ( j n 1 , , j 1 ; α ) , ] , k = 1 , , 2 n comprises 2 n components defined for each j 1 = 1 , , T P ; j 2 = 1 , , j 1 ; …; j n 1 = 1 , , j n 2 , as follows:
s 1 ( n ) ( j n 1 , , j 1 ; α ) S ( n 1 ) ( j n 1 , , j 1 ; u ( n ) ; α ) φ ;
s 2 ( n ) ( j n 1 , , j 1 ; α ) S ( n 1 ) ( j n 1 , , j 1 ; u ( n ) ; α ) ψ ;
s 2 + k ( n ) ( j n 1 , , j 1 ; α ) S ( n 1 ) ( j n 1 , , j 1 ; u ( n ) ; α ) a k ( 1 ) ;       k = 1 , 2 ;
s 2 2 + k ( n ) ( j n 1 , , j 1 ; α )   S ( n 1 ) ( j n 1 , , j 1 ; u ( n ) ; α ) a k ( 2 ) ( j 1 ) ;       k = 1 , , 2 2 ;
s 2 3 + k ( n ) ( j n 1 , , j 1 ; α )   S ( n 1 ) ( j n 1 , , j 1 ; u ( n ) ; α ) a k ( 3 ) ( j 2 ; j 1 ) ;       k = 1 , , 2 3   ;
s 2 n 1 + k ( n ) ( j n 1 , , j 1 ; α )   S ( n 1 ) ( j n 1 , , j 1 ; u ( n ) ; α ) a k ( n 1 ) ( j n 2 , , j 1 ) ;       k = 1 , , 2 n 1 .
(6)
The quantity { P ^ ( n ) ( a ( n ) ; u ( n ) ; α ; δ α ) } α 0 denotes the residual bilinear concomitant defined on the phase-space boundary x Ω x ( α 0 ) , evaluated at the nominal values of the model parameter and respective state functions.
[ A ( n ) ] 2 n × 2 n [ V ( n ) ] 2 n × 2 n * = ( [ V ( n 1 ) ] 2 n 1 × 2 n 1 [ 0 ] 2 n 1 × 2 n 1 [ V 21 ( n ) ] 2 n 1 × 2 n 1 [ V 22 ( n 1 ) ] 2 n 1 × 2 n 1 ) * ;
V 21 ( n ) { [ A ( n 1 ) ] 2 n 1 × 2 n 1 a ( n 1 ) ( j n 2 , , j 1 ; x ) } u ( n 1 ) { s A ( n 1 ) [ j n 2 , , j 1 ; u ( n 1 ) ( j n 3 , , j 1 ; x ) ; α ] } u ( n 1 ) ;       V 22 ( n 1 ) A ( n 1 ) ;
q V ( n ) [ j n 2 , , j 1 ; u ( n ) ; α ; δ α ] ( q V ( n 2 ) [ j n 3 , , j 1 ; α ; δ α ] p ( n 1 ) [ j n 2 , , j 1 ; u ( n ) ; α ; δ α ] ) [ q k ( n ) ( j n 2 , , j 1 ; u ( n ) ; α ; δ α ) ] ,       k = 1 , , 2 n ;
p ( n 1 ) ( j n 2 , , j 1 ; u ( n ) ; α ; δ α )     { s A ( n 1 ) [ j n 2 , , j 1 ; u ( n 1 ) ( j n 3 , , j 1 ; x ) ; α ] A ( n 1 ) a ( n 1 ) } α δ α .

4. High-Order Sensitivity and Uncertainty Analysis of Nonlinear Models/Systems: Illustrative Example and General Theory

It is currently well known that a boiling water reactor (BWR) can undergo time-dependent power–amplitude oscillations, particularly when operated under high-power and low-flow conditions, due to the nonlinear coupling, through the void reactivity, between the neutronics and thermal-hydraulic processes. Such power oscillations can occur globally, when the reactor’s power oscillates in phase across the entire reactor core, or regionally, when part (e.g., half, for the so-called “first mode”) of the reactor core oscillates out of phase with respect to the other part (half), while the average power remains essentially constant across the reactor. In a pioneering sequence of works, March-Leuba, Cacuci, and Perez [45,46,47] proposed a phenomenological reduced-order model for analyzing qualitatively the dynamic behavior of BWRs, predicting global limit cycle oscillations in BWRs four years before they actually occurred [62] in an operating power reactor in the USA, on 9 March 1988, in the LaSalle County-2 BWR in Seneca, Illinois. The reduced-order model developed by March-Leuba, Cacuci, and Perez [45] comprises point neutron kinetics equations coupled nonlinearly to thermal-hydraulics equations that describe the time evolution of the fuel temperature and coolant density in the recirculation loop, which is treated as a single path of fluid of variable cross-sectional area but with constant mass flow, with the coolant entering the core at saturation temperature. This model showed that the major state variables, including power, reactivity, void fraction, and temperature, undergo limit cycle oscillations under high-power/low-flow operating conditions, and the amplitudes of these oscillations are very sensitive to the reactor’s operating conditions. This model also indicated that these limit cycle oscillations can become unstable—when increasing the heat transfer from the reactor to the coolant—and can undergo period-doubling bifurcations causing oscillatory (from periodic to aperiodic) time evolutions of the state variables (power, delayed neutron precursors, reactivity, coolant void fraction, and fuel temperature). The reduced-order model introduced by March-Leuba, Cacuci, and Perez [45] has generated great interest in the scientific community, inspiring investigations of the stability limits for in-phase and out-of-phase oscillation modes and the nature of the underlying bifurcations in the various state functions, as exemplified by [48,49,50,51,52,53,54,55].
Section 4.1 reviews the pioneering sensitivity and uncertainty analysis results presented in [43,44] for the reduced-order model of BWR dynamics developed by March-Leuba, Cacuci, and Perez [45]. This nonlinear model [45] illustrates generically the most extreme forms of nonlinear phenomena (from limit cycles to chaos) that can be displayed by nonlinear physical systems, so it is considered to be a representative paradigm for illustrating the application of the nth-CASAM-N methodology [42] for computing efficiently and exactly the sensitivities of responses to parameters for nonlinear systems. For completeness, the general mathematical formalism underlying the nth-CASAM-N [42] is reviewed in Section 4.2.

4.1. Illustrative Sensitivity and Uncertainty Analysis of a Reduced-Order Model of BWR-Dynamics

The reduced-order model developed by March-Leuba, Cacuci, and Perez [45] describes the fundamental processes underlying the dynamic behavior of a BWR when the reactor is perturbed from its originally critical steady-state condition, by insertion of reactivity. This reduced-order (lumped-parameter) model comprises differential equations that describe the following phenomena:
(a)
A point reactor representation of neutron kinetics, including the neutron precursors:
d n ( t ) d t = n ( t ) ρ ( t ) β Λ + λ c ( t ) + ρ ( t ) Λ
d c ( t ) d t = n ( t ) β Λ λ c ( t )
(b)
A “one-node” lumped-parameter representation of the heat transfer process in the fuel:
d T ( t ) d t = Q [ n ( t ) + H ( t ) Δ ] a 3 T ( t )
(c)
A “two-nodes” lumped-parameter representation of the channel thermal hydraulics, accounting for the reactivity feedback:
d 2 γ ( t ) d t 2 + a 2 d γ ( t ) d t + a 1 γ ( t ) = m k 0 T ( t )
ρ ( t ) = Γ 1 γ ( t ) + Γ 2 T ( t )
The initial conditions at t = 0 for Equations (74)−(77) are as follows:
n ( 0 ) = n 0 ,
c ( 0 ) = c 0 ,
T ( 0 ) = T 0 ,
γ ( 0 ) = γ 0 .
The state variables in Equations (74)−(78) are defined as follows:
(i)
n ( t ) [ N ( t ) N 0 ] / N 0 denotes the normalized excess neutron population, where N ( t ) denotes the actual time-dependent neutron population and N 0 denotes the neutron population of the initially critical reactor.
(ii)
c ( t ) [ C ( t ) C 0 ] / C 0 denotes the normalized excess population of delayed neutron precursors, where C ( t ) denotes the actual time-dependent population of delayed neutron precursors and C 0 denotes the population of delayed neutron precursors of the initially critical reactor.
(iii)
T ( t ) denotes the excess fuel temperature, defined as the departure from the temperature of the initially critical reactor.
(iv)
H ( t ) denotes the Heaviside functional: H ( t ) = 0 ,     f o r           t < 0 and H ( t ) = 1 ,     f o r           t > 0 .
(v)
γ ( t ) denotes the relative excess coolant density, defined as the departure from the coolant density of the initially critical reactor.
(vi)
ρ ( t ) denotes the excess reactivity, defined as the departure from the reactivity of the initially critical reactor.
The nominal values of the model parameters which appear in Equations (74)−(78) are reported in Table 6. Since the qualifier “excess” signifies that the above quantities are defined as departures from the critical reactor configuration, the nominal values of the initial conditions shown in Equations (79)−(82) are all zero at t = 0 .
The reduced-order model defined by Equations (74)−(82) has two equilibrium points in the phase space, one stable (denoted as “point A”) and the other one unstable (denoted as “point B”), at the following phase space locations:
  • Stable equilibrium (A), which corresponds to the critical reactor configuration, which is defined as follows:
n A = Δ ;     c A = β Δ λ Λ ;     T A = γ A = ρ A = 0 .
b.
Unstable equilibrium (B):
n B = 1 ;     c B = β λ Λ ;     T B = Q ( Δ 1 ) a 3 ;   γ B = k Q ( Δ 1 ) ;     ρ B = ( Γ 1 k a 3 + Γ 2 ) Q ( Δ 1 ) a 3 .
As discussed by March-Leuba, Cacuci, and Perez [45], when the feedback gain m in Equation (77) is increased, thereby increasing the heat transfer from the reactor to the coolant, the equilibrium point A becomes unstable after the stability threshold is crossed. The phase-space trajectories of the lumped-parameter model oscillate between the two unstable equilibrium points until a stable limit cycle sets in. Increasing the value of the feedback gain m beyond the stability of the first limit cycle will cause this cycle to bifurcate, through a period-doubling bifurcation, into a period-two cycle. Continuing to increase the feedback gain m leads to a cascade of period-doubling bifurcations which lead to aperiodic behavior, as illustrated in Figure 3 by the bifurcation map for the excess temperature T ( t ) , which is obtained by plotting the extrema of T ( t ) as a function of the feedback gain m . The values of the bifurcation variables are provided in Table 7. The bifurcation map presented in Figure 3 for values of k ranging from k = 1.6 · k 0 to k = 3.6 · k 0 (i.e., 1.6 m 3.6 ) highlight the following Regions in the phase space:
  • Region 1: Stable Region, before the first-order bifurcation, m · k 0 < k b i f 1 , in which the state functions oscillate with decreasing amplitudes converging toward steady-state values.
  • Region 2: Unstable Region between the first-order bifurcation and the second-order bifurcations, k b i f 1 < m · k 0 < k b i f 2 , in which the state functions oscillate indefinitely between two unstable attractors.
  • Region 3: Unstable Region between the second-order bifurcations and the third-order bifurcations, k b i f 2 < m · k 0 < k b i f 3 , in which the state functions oscillate indefinitely between four unstable attractors.
  • Region 4: Unstable Region between the third-order bifurcations and the fourth-order bifurcations, k b i f 3 < m · k 0 < k b i f 4 , in which the state functions oscillate indefinitely between eight unstable attractors.
  • Region 5: Chaotic Region stemming from the cascade of period-doubling pitchfork bifurcations produced when the feedback gain m is increased past a critical value (called accumulation point), m c < m · k 0 , where m c = 2.933995 , in which the state functions oscillate indefinitely between infinitely many unstable attractors.
The state functions are attracted in the stable Region 1 to the stable equilibrium attractor A, having the following coordinates in the phase space:
n A = 0.1 ;     c A = 175 ;     T A = γ A = ρ A = 0 .
The largest first-order sensitivity displayed by T ( t ) is with respect to the initial condition γ 0 for the excess coolant density γ ( t ) , as depicted in Figure 4, which also presents, for comparison, the sensitivities of T ( t ) to the other initial conditions. As depicted in Figure 4, the sensitivities oscillate with very large amplitudes immediately after the onset of the transient, which is initiated by perturbing the initially critical reactor in the steady state, and these amplitudes decay exponentially in time. The features depicted in Figure 4 are transmitted to the first-order total absolute standard deviation, denoted as σ , induced in T ( t ) by the sensitivities and parameter standard deviations, as shown in Figure 5.
The first-order total absolute standard deviation, denoted as σ , for T ( t ) is obtained by using the well-known formula σ = i = 1 16 ( S i σ i ) 2 , where S i n ( t ) / α i   denotes the sensitivity of T ( t ) to the parameter α i , and where σ i denotes the standard deviation for the respective parameter. The time evolution of the nominal value of the state function T ( t ) is depicted in blue in Figure 5 along with the time evolution of the functions labeled “ T ( t ) ± σ ” and depicted in yellow, where σ denotes the total absolute standard deviation for T ( t ) . Immediately after the onset of the transient effected by perturbing the initially critical reactor in the steady state, the standard deviation for T ( t ) is very large, reaching values that more than 10 times larger than the state function T ( t ) itself. In time, the standard deviation decays exponentially, reaching values comparable to the state function T ( t ) after about 140 s from the start of the transient. Very importantly, Figure 5 indicates that predictions of the excess temperature of T ( t ) during the first 120 s after the BWR model is perturbed from its initial critical steady state would be unreliable. Furthermore, it is possible that that the model could be “pushed” into one of the oscillatory Regions by the large amplitudes reached by the standard deviation of T ( t ) immediately after the onset of the perturbation.
The largest first-order sensitivity displayed by T ( t ) in the oscillatory Region 2 is also with respect to the initial condition γ 0 ; the time evolution of this sensitivity is depicted in Figure 6, together with the sensitivities of T ( t ) to the other initial conditions. As depicted in Figure 6, the amplitudes of these sensitivities are about two orders of magnitude larger than the corresponding first-order sensitivities. In contradistinction with the decaying amplitudes of the oscillations in Region 1, the oscillation amplitudes of the sensitivities in Region 2 stabilize to large values asymptotically in time. The features depicted in Figure 6 are transmitted to the first-order total absolute standard deviation induced in T ( t ) , as shown in Figure 7, which depicts (in blue) the time evolution of the nominal value of the state function T ( t ) along with the time evolution (in yellow) of the functions T ( t ) ± σ . As indicated in Figure 7, the standard deviation increases rapidly after the onset of the transient (by perturbing the initially critical reactor in the steady state), oscillating with rapidly increasing amplitudes, reaching in 40 s after the start of the transient values that are 400K larger than the amplitude of the excess temperature, T ( t ) , which oscillates around the initial value T 0 = 0 K . It is evident from Figure 7 that the standard deviation of T ( t ) renders the prediction of the evolution of this state function highly unreliable after less than 10 s from the initiation of the transient.
The evolution of the standard deviation of T ( t ) in the oscillatory Region 3 is depicted in Figure 8, which indicates that a standard deviation as small as 5% in the model’s parameters and initial conditions will be magnified in the oscillatory regions so much as to preclude any accurate prediction of the actual values of the excess temperature T ( t ) . This conclusion is also evidenced by the results presented in Table 8, which presents the largest positive and negative values attained by the oscillation amplitude of T ( t ) in the aperiodic Region 5, for 6 windows of 25 s each, covering the transient period from 0 to 150 s.

4.2. Overview of the nth-Order Comprehensive Sensitivity Analysis Methodology for Nonlinear Systems (nth-CASAM-N)

Using the same notation for the model’s independent variables and parameters as for the linear system considered in Section 3, the mathematical modeling of a nonlinear physical system can be represented in operator form as follows:
N [ u ( x ) , α ] = Q ( x , α )   ,       x Ω x ( α ) ,
The quantities which appear in Equation (86) are defined as follows:
  • u ( x ) [ u 1 ( x ) , , u T D ( x ) ] is a T D -dimensional column vector of dependent variables; the abbreviation “ T D ” denotes “total number of dependent variables.” The functions u i ( x ) ,     i = 1 ,   ,   T D denote the system’s “dependent variables” (also called “state functions”); u ( x ) E u , where E u is a normed linear space over the scalar field F of real numbers.
  • N [ u ( x ) ; α ] [ N 1 ( u ; α ) , , N T D ( u ; α ) ] denotes a T D -dimensional column vector. The components N i ( u ; α ) ,     i = 1 , , T D are operators (including differential, difference, integral, distributions, and/or finite or infinite matrices) acting (usually) nonlinearly on the dependent variables u ( x ) , the independent variables x , and the model parameters α . The mapping N ( u ; α ) is defined on the combined domains of the model’s parameters and state functions, i.e., N :   D E E Q , where D = D u D α , D u E u , D α E α , E = E u E α .
  • Q ( x , α )   [ q 1 ( x ; α ) , , q T D ( x ; α ) ] is a T D -dimensional column vector which represents inhomogeneous source terms, which usually depend nonlinearly on the uncertain parameters α . The vector Q ( x , α )   is defined on a normed linear space denoted as E Q , i.e., Q E Q .
Boundary and/or initial conditions must also be provided if differential operators appear in Equation (86). In operator form, these boundaries and/or initial conditions are represented as follows:
B [ u ( x ) ; α ;   x ] C ( x , α ) = 0 ,       x Ω x ( α ) ,
Solving Equations (86) and (87) at the nominal parameter values α 0 [ α 1 0 , , α i 0 , , α T P 0 ] provides the “nominal solution” u 0 ( x ) , i.e., the vectors u 0 ( x ) and α 0 , satisfies the following equations:
N [ u 0 ( x ) ; α 0 ] = Q ( x , α 0 )   ,       x Ω x ,
B [ u 0 ( x ) ; α 0 ;     x ] - C ( x , α 0 ) = 0 ,       x Ω x ( α 0 ) .
The model response is a functional considered to be representable in the following generic integral form:
R [ u ( x ) ; α ] λ 1 ( α ) ω 1 ( α ) λ T I ( α ) ω T I ( α ) S [ u ( x ) ; α ; x ] d x 1 d x T I ,
where S [ u ( x ) ; α ] is a suitably differentiable nonlinear function of u ( x ) and of α . It is important to note that the components of α include not only parameters that appear in the equations defining the computational model per se, i.e., in Equations (86) and (87), but also include parameters that specifically occur only in the definition of the response under consideration. It is also important to note that the system’s definition domain, Ω ( α ) , in phase space is considered to be imprecisely known, subject to uncertainties in the components of the vector of model parameters α . Therefore, the system domain’s boundary, Ω ( α ) , as well as the model response R [ u ( x ) ; α ] , will be affected by the boundary uncertainties that affect the endpoints λ i ( α ) ,     ω i ( α ) ,     i = 1 , , T I . Such boundary uncertainties stem most often from manufacturing uncertainties.
Variation δ α in the model and boundary parameters will cause corresponding variations v ( 1 ) ( x ) [ δ u 1 ( x ) , , δ u T D ( x ) ] around the nominal solution u 0 ( x ) in the forward state functions. The variations δ α and v ( 1 ) ( x ) will, in turn, induce variations in the system’s response. The first-order G-differential δ R ( e 0 ; h ) { δ R [ u ( x ) ; α ; v ( 1 ) ( x ) ; δ α ] } α 0 is obtained, by definition, as follows:
{ δ R [ u ( x ) ; α ; v ( 1 ) ( x ) ; δ α ] } α 0 { d d ε λ 1 ( α 0 + ε δ α ) ω 1 ( α 0 + ε δ α ) λ T I ( α 0 + ε δ α ) ω T I ( α 0 + ε δ α ) S ( u 0 + ε v ( 1 ) , α 0 + ε δ α ; x ) d x 1 d x T I } { δ R [ u ( x ) ; α ; δ α ] } d i r + { δ R [ u ( x ) ; α ; v ( 1 ) ( x ) ] } i n d .
In Equation (91), the direct effect term { δ R [ u ( x ) ; α ; δ α ] } d i r comprises only dependencies on δ α and is defined as follows:
{ δ R [ u ( x ) ; α ; δ α ] } d i r { R ( u ; α ) α } α 0 δ α   j 1 = 1 T P { R ( 1 ) [ j 1 ; u ( x ) ; α ] } d i r δ α j 1 ,
where
{ R ( 1 ) [ j 1 ; u ( x ) ; α ] } d i r { λ 1 ( α ) ω 1 ( α ) λ T I ( α ) ω T I ( α ) S ( u ; α ; α ) α j 1 d x 1 d x T I } α 0 + j = 1 T I { ω j ( α ) α j 1 λ 1 ω 1 d x 1 λ j 1 ω j 1 λ j + 1 ω j + 1 λ T I ω T I d x T I S [ u ( x 1     , . , ω j ( α ) , . , x N x ) ; α ] } α 0   j = 1 T I { λ j ( α ) α j 1 λ 1 ω 1 d x 1 λ j 1 ω j 1 λ j + 1 ω j + 1 λ T I ω T I d x T I S [ u ( x 1     , . , λ j ( α ) , . , x N x ) ; α ] } α 0 .
On the other hand, the quantity { δ R [ u ( x ) ; α ; v ( 1 ) ( x ) ] } i n d in Equation (91) comprises only variations in the state functions and is therefore called the indirect effect term, having the following expression:
{ δ R [ u ( x ) ; α ; v ( 1 ) ( x ) ] } i n d   λ 1 ( α 0 ) ω 1 ( α 0 ) d x 1 λ T I ( α 0 ) ω T I ( α 0 ) d x T I {   S ( u ; α ; x ) u v ( 1 ) ( x ) } α 0 ; [   ] u v ( 1 ) ( x ) i = 1 T D [   ] u i ( x ) δ u i ( x ) .
The indirect effect term induces variations in the response through the variations in the state functions, which are, in turn, caused by the parameter variations through the equations underlying the model. Evidently, the indirect effect term can be quantified only after having determined the variation v ( 1 ) ( x ) in terms of the variation δ α . The indirect effect term induces variations in the response through the variations in the state functions, which are, in turn, caused by the parameter variations through the equations underlying the model. Evidently, the indirect effect term can be quantified only after having determined the variation v ( 1 ) ( x ) in terms of the variation δ α .
The first-order relationship between the vectors v ( 1 ) ( x ) and δ α is determined by solving the equations obtained by applying the definition of the G-differential to Equations (86) and (87), which yields the following equations:
{ d d ε N [ u 0 + ε v ( 1 ) ( x ) ; α 0 + ε δ α ] } ε = 0 = { d d ε Q ( x , α 0 + ε δ α ) } ε = 0 ,
{ d d ε B [ u 0 + ε v ( 1 ) ( x ) ; α 0 + ε δ α ] } ε = 0 { d d ε C ( x , α 0 + ε δ α ) } ε = 0 = 0 .
Carrying out the differentiations with respect to ε in Equations (95) and (96) and setting ε = 0 in the resulting expressions yields the following equations:
{ N ( 1 ) ( u ; α ) v ( 1 ) ( x ) } α 0 = { q V ( 1 ) ( u ; α ; δ α ) } α 0 ,   x Ω x ,
{ b V ( 1 ) ( u ; α ; v ( 1 )   ; δ α ) } α 0 = 0 ,   x Ω x ( α 0 ) .
In Equations (97) and (98), the superscript “(1)” indicates “1st-Level” and the various quantities which appear in these equations are defined as follows:
N ( 1 ) ( u ; α ) { N (   u ; α ) u } ( N 1 u 1 N 1 u T D N T D u 1 N T D u T D )   ;
q V ( 1 ) ( u ; α ; δ α ) [ Q ( α ) N (   u ; α ) ] α   δ α   j 1 = 1 T P s V ( 1 ) ( j 1 ; u ; α ) δ α j 1 ;
s V ( 1 ) ( j 1 ; u ; α ) [ Q ( α ) N (   u ; α ) ] α j 1 ;
{ b V ( 1 ) ( u ; α ; v ( 1 ) ; δ α ) } α 0 { B ( u ; α ) u } α 0 v ( 1 ) + { [ B ( u ; α ) C ( α ) ] α } α 0   δ α .
The system comprising Equations (97) and (98) is called the “1st-Level Variational Sensitivity System” (1st-LVSS). In order to determine the solutions of the 1st-LVSS that would correspond to every parameter variation δ α j 1 , j 1 = 1 ,   ,   T P , the 1st-LVSS would need to be solved T P times, with distinct right sides for each δ α j 1 , thus requiring T P large-scale computations.
The need for computing the vector v ( 1 ) ( x ) is eliminated by applying the 1st-CASAM-N to express the indirect effect term defined in Equation (94) in terms of the solutions of the “1st-Level Adjoint Sensitivity System” (1st-LASS), the construction of which requires the introduction of adjoint operators. The 1st-CASAM-N represents a generalization of the pioneering works by Cacuci (1981a, b) which had conceived the “Adjoint Sensitivity Analysis Methodology for Nonlinear Systems” but in which the domain boundary was considered to have been perfectly well-known, free of uncertainties. The qualifier “comprehensive” is meant to indicate that that the 1st-CASAM-N considers that the internal and external boundaries Ω x ( α ) of the phase-space domain depend on the uncertain model parameters α and are thereby imprecisely known, subject to uncertainties. The construction of the “1st-Level Adjoint Sensitivity System” (1st-LASS) is accomplished by introducing a real Hilbert space denoted as H 1 ( Ω x ) , endowed with an inner product of two vectors u ( a ) ( x ) H 1 and u ( b ) ( x ) H 1 denoted as u ( a ) , u ( b ) 1 and defined as follows:
u ( a ) , u ( b ) 1   { λ 1 ( α ) ω 1 ( α ) λ T I ( α ) ω T I ( α ) [ u ( a ) ( x ) u ( b ) ( x ) ] d x 1 d x T I } α 0 ,