Next Article in Journal
Thermal Shock and Synergistic Plasma and Heat Load Testing of Powder Injection Molding Tungsten-Based Alloys
Previous Article in Journal
Deterministic Data Assimilation in Thermal-Hydraulic Analysis: Application to Natural Circulation Loops
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The First- and Second-Order Features Adjoint Sensitivity Analysis Methodologies for Neural Integro-Differential Equations of Volterra Type: Mathematical Framework and Illustrative Application to a Nonlinear Heat Conduction Model

by
Dan Gabriel Cacuci
Department of Mechanical Engineering, University of South Carolina, Columbia, SC 29208, USA
J. Nucl. Eng. 2025, 6(3), 24; https://doi.org/10.3390/jne6030024
Submission received: 14 May 2025 / Revised: 20 June 2025 / Accepted: 27 June 2025 / Published: 4 July 2025

Abstract

This work presents the mathematical frameworks of the “First-Order Features Adjoint Sensitivity Analysis Methodology for Neural Integro-Differential Equations of Volterra-Type” (1st-FASAM-NIDE-V) and the “Second-Order Features Adjoint Sensitivity Analysis Methodology for Neural Integro-Differential Equations of Volterra-Type” (2nd-FASAM-NIDE-V). It is shown that the 1st-FASAM-NIDE-V methodology enables the efficient computation of exactly-determined first-order sensitivities of the decoder response with respect to the optimized NIDE-V parameters, requiring a single “large-scale” computation for solving the 1st-Level Adjoint Sensitivity System (1st-LASS), regardless of the number of weights/parameters underlying the NIE-net. The 2nd-FASAM-NIDE-V methodology enables the computation, with unparalleled efficiency, of the second-order sensitivities of decoder responses with respect to the optimized/trained weights involved in the NIDE-V’s decoder, hidden layers, and encoder, requiring only as many “large-scale” computations as there are non-zero first-order sensitivities with respect to the feature functions. These characteristics of the 1st-FASAM-NIDE-V and 2nd-FASAM-NIDE-V are illustrated by considering a nonlinear heat conduction model that admits analytical solutions, enabling the exact verification of the expressions obtained for the first- and second-order sensitivities of NIDE-V decoder responses with respect to the model’s functions of parameters (weights) that characterize the heat conduction model.

1. Introduction

The capabilities of traditional neural nets [1,2,3] have been extended significantly by the introduction of Neural Ordinary Differential Equations (NODE), formulated by Chen et al. [4], which have enabled the use of deep learning for modeling discretely sampled dynamical systems (new [5,6,7,8,9,10,11,12]). Although NODE provide a bridge between modern deep learning and traditional numerical modelling, they are limited to describing systems that are instantaneous, each time-step being determined locally in time, without contributions from the state of the system at other times. On the other hand, integral equations (IE) are suitable for modeling global “long-distance” spatio-temporal relations and can be solved efficiently using IE solvers that sample the domain of integration continuously, as exemplified in [13,14,15,16]. Two important families of IEs are the Volterra and the Fredholm equations. In a Volterra IE, the interval of integration grows linearly during the system’s dynamics. On the other hand, the interval of integration is fixed in a Fredholm IE during the dynamic-history of the system, but at any given time instance within this interval, the system depends on the past, present, and future states of the system. For IEs of the Volterra and Fredholm types, Zappala et al. [17] have presented a general mathematical framework for Neural Integral Equations (NIE) and Attentional Neural Integral Equations (ANIE), which can be used to infer the spatio-temporal relations that generated the data, thus enabling the continuous learning of non-local dynamics with arbitrary time resolution. In particular, the ANIE interprets the self-attention mechanism as the Nystrom method for approximating integrals [18], which enables efficient integration over higher dimensions. Generalizing this work, Zappala et al. [19] have also developed a deep learning method called Neural Integro-Differential Equation (NIDE), which “learns” an integro-differential equation (IDE) whose solution is obtained using standard IDE-solvers [20,21] and approximates data sampled from given non-local dynamics. The motivation for using NIDE stems from the need to model systems that present spatio-temporal relations that transcend local modeling, as illustrated by the pioneering works of Volterra on population dynamics [22]. Combining the properties of differential and integral equations, NIDEs also present properties that are unique to their non-local behavior, with applications in various applied sciences, including physics, engineering, and computational biology [23,24,25,26,27].
The neural net is trained/optimized to reproduce the underlying physical system as closely as possible, by minimizing a “loss functional” that aims at representing the discrepancy between a “reference solution” and the output produced by the respective net’s decoder. However, the physical system modeled by a neural net comprises parameters that stem from measurements and/or computations that are subject to uncertainties. Therefore, even in the idealized/unrealistic case when a neural net would perfectly model the system under consideration, the unavoidable uncertainties inherent in the physical system’s parameters would propagate to the subsequent results of interest (customarily called “responses”) produced by the net’s decoder. The quantification of the uncertainties in the decoder’s responses, which are various functionals of the net’s hidden variables and parameters rather than some “loss functional,” requires the computation of the sensitivities of the decoder’s response with respect to the optimized weights/parameters comprised within the neural net.
Cacuci [28] has developed the “Nth-Order Comprehensive Adjoint Sensitivity Analysis Methodology for Nonlinear Systems (Nth-CASAM-N)”, which enables the efficient computation of the exact expressions of arbitrarily high-order sensitivities of model responses with respect to the model’s parameters. However, physical systems and, therefore, their neural net counterparts comprise not only scalar-valued weights/parameters but often also comprise scalar-valued functions (e.g., correlations, material properties, etc.) of the model’s scalar parameters. It is convenient to refer to such scalar-valued functions as “features of primary model parameters.” Cacuci [29] has recently generalized his Nth- CASAM methodology, introducing the “Nth-Order Features Adjoint Sensitivity Analysis Methodology for Nonlinear Systems (Nth-FASAM-N),” which enables the computation, with unparalleled efficiency, of the exact expressions of arbitrarily high-order sensitivities of model responses with respect to the model’s “features.” Subsequently, the sensitivities of the responses with respect to the primary model parameters are determined, analytically and trivially, by applying the “chain-rule” to the expressions obtained for the response sensitivities with respect to the model’s features/functions of parameters.
Based on the general framework of the Nth-FASAM-N methodology [29], Cacuci has developed specific sensitivity analysis methodologies for NODE-nets, as follows: the “First-Order Features Adjoint Sensitivity Analysis Methodology for Neural Ordinary Differential Equations (1st-FASAM-NODE)” [30] and the “Second-Order Features Adjoint Sensitivity Analysis Methodology for Neural Ordinary Differential Equations (2nd-FASAM-NODE)” [31]. The 1st-FASAM-NODE and the 2nd-FASAM-NODE are pioneering sensitivity analysis methodologies that enable the computation, with unparalleled efficiency, of exactly-determined first-order and, respectively, second-order sensitivities of decoder response with respect to the optimized/trained weights involved in the NODE’s decoder, hidden layers, and encoder.
By applying the general concepts underlying the Nth-FASAM-N methodology [29], Cacuci [32] has recently developed the “Second-Order Features Adjoint Sensitivity Analysis Methodology for Neural Integral Equations of Fredholm-Type (2nd-FASAM-NIE-F)”, which encompasses the “First-Order Features Adjoint Sensitivity Analysis Methodology for Neural Integral Equations of Fredholm-Type (1st-FASAM-NIE-F). Cacuci [33] has also developed the “Second-Order Features Adjoint Sensitivity Analysis Methodology for Neural Integral Equations of Volterra-Type (2nd-FASAM-NIE-V)”, which encompasses the “First-Order Features Adjoint Sensitivity Analysis Methodology for Neural Integral Equations of Volterra-Type (1st-FASAM-NIE-V).” The 1st-FASAM-NIE-F and 1st-FASAM-NIE-V methodologies, respectively, enable the computation, with unparalleled efficiency, of exactly-determined first-order sensitivities of decoder response with respect to the NIE-parameters, requiring a single “large-scale” computation for solving the 1st-Level Adjoint Sensitivity System (1st-LASS), regardless of the number of weights/parameters underlying the NIE-net. The 2nd-FASAM-NIE-F and 2nd-FASAM-NIE-V methodologies, respectively, enable the efficient computation of exactly-determined second-order sensitivities of decoder response with respect to the NIE-parameters, requiring only as many “large-scale” computations as there are first-order sensitivities with respect to the feature functions.
Generalizing the 2nd-FASAM-NIE-F methodology, Cacuci [34] has recently developed the “First- and Second-Order Features Adjoint Sensitivity Analysis Methodologies for Fredholm-Type Neural Integro-Differential Equations” (1st-FASAM-NIDE-F and 2nd-FASAM-NIDE-F), respectively, for computing, with unparalleled efficiency, the exact first- and second-order sensitivities, respectively, of decoder responses to model parameters in optimized NIDE-F networks.
This work presents the “First- and Second-Order Features Adjoint Sensitivity Analysis Methodologies for Neural Integro-Differential Equations of Volterra-Type” (1st- FASAM-NIDE-V and 2nd-FASAM-NIDE-V, respectively), thus generalizing the 2nd-FASAM-NIE-V methodology [33]. The 1st-FASAM-NIDE-V and, respectively, the 2nd-FASAM-NIDE-V enable, for the first time, the most efficient computation of the exact expressions of the first- and, respectively, second-order sensitivities of NIDE-V decoder-responses with respect to the optimized network’s feature functions and weights/parameters of NIDE-V neural nets.
This work is structured as follows: Section 2 presents the “First-Order Features Adjoint Sensitivity Analysis Methodology for Neural Integro-Differential Equations of Volterra-Type” (1st-FASAM-NIDE-V). It is shown that the 1st-FASAM-NDIE-V methodology enables the computation, with unparalleled efficiency, of exactly-determined first-order sensitivities of decoder response with respect to the NODE-parameters, requiring a single “large-scale” computation for solving the 1st-Level Adjoint Sensitivity System (1st-LASS), regardless of the number of weights/parameters underlying the NIDE-V net. When “feature functions of parameters “can be identified within the NIDE-V structure, the number of quadratures for computing the first-order sensitivities is smaller than the number of quadratures needed for computing the first-order decoder-response sensitivities directly with respect to the parameters, since the latter can be computed analytically and exactly by using the first-order sensitivities with respect to the feature functions.
Section 3 illustrates the application of the 1st-FASAM-NIDE-V methodology to a paradigm heat conduction model, which has been chosen for the following reasons: (i) the model is nonlinear but admits an equivalent linear model that is related to the original model through the Kirchhoff transformation; (ii) the model can be represented either as a traditional NODE net or a NIDE-V net; (iii) the model admits explicit closed-form solutions for all quantities of interest, including the functions representing the hidden/latent neural networks, the decoder response, and the sensitivities of the decoder’s response to the optimal weights/parameters; (iv) the model enables an exact comparison between the application of the 2nd-FASAM-NIDE-V methodology developed in Section 2 of this work and the application of the 2nd-FASAM-NODE methodology presented in [30,31]. It will be shown that the algebraic computations are considerably simpler in the Kirchhoff-transformed framework (in which the paradigm model can be exactly transformed from a nonlinear into a linear one) than in the conventional “temperature-framework,” in which the model’s nonlinearities cannot be circumvented.
Section 4 presents the mathematical framework of the “Second-Order Features Adjoint Sensitivity Analysis Methodology for Neural Integral Equations of Volterra Type” (2nd-FASAM-NIDE-V), which computes the second-order sensitivities most efficiently by conceptually using their basic definitions as being the “first-order sensitivities of the first-order sensitivities.” It is shown that the 2nd-FASAM-NIDE-V yields the exact expressions of the 2nd-order sensitivities and computes them most efficiently by needing only as many “large-scale” computations as there are non-zero 1st-order sensitivities with respect to the net’s feature functions. It is also shown that the mixed 2nd-order sensitivities are computed twice, using distinct 2nd-level adjoint functions, which provides the most inexpensive path for the stringent mutual verification of the accuracy of the computations performed when solving the respective 2nd-LASS for determining the respective 2nd-level adjoint functions.
Section 5 presents the illustrative application of the 2nd-FASAM-NIDE-V methodology to compute second-order sensitivities of the paradigm heat conduction model. The algebraic computations/manipulations are minimized by considering the heat conduction model in the Kirchhoff-transformed framework. All the general features of the 2nd-FASAM-NIDE-V methodology are highlighted using this model.
The discussion presented in Section 6 summarizes and concludes this work, noting that (i) the 1st-FASAM-NIDE-V methodology enables the computation, with unparalleled efficiency, of exactly-determined first-order sensitivities of decoder response with respect to the NODE-parameters, requiring a single “large-scale” computation for solving the 1st-Level Adjoint Sensitivity System (1st-LASS), regardless of the number of weights/parameters underlying the NIDE-V net; and (ii) the 2nd-FASAM-NIDE-V methodology enables the computation (with unparalleled efficiency) of exactly-determined second-order sensitivities of decoder response with respect to the model’s feature functions and parameters, requiring at most as many “large-scale” computations as there are non-zero first-order sensitivities with respect to the feature functions. Future work will examine the theoretical underpinnings and feasibility of adapting algorithms of “backpropagation-type” for computing high-order sensitivities of decoder response with respect to the feature functions of model parameters, aiming at maximizing the efficiency and accuracy of computing sensitivities of orders higher than first-order.

2. First-Order Features Adjoint Sensitivity Analysis Methodology for Neural Integro-Differential Equations of Volterra-Type (1st-FASAM-NIDE-V)

The major families of IDEs are the Fredholm and the Volterra equations. In a NIDE-F, the interval of integration is fixed during the dynamic-history of the system, but at any given time instance within this interval, the system depends on the past, present, and future states of the system. In contradistinction, in a NIDE-V, the interval of integration grows linearly during the system’s dynamics. Otherwise, the generic mathematical structures of the NIDE-V and the NIDE-F are similar. Therefore, the notation that will be used for developing the 1st-FASAM-NIDE-V sensitivity analysis methodology to be presented in this work will closely resemble the notation used by Cacuci [34] in his development of the 2nd-FASAM-NIDE-F methodology. Using similar notation will also facilitate the comparison between the 2nd-FASAM-NIDE-F methodology [34] and the 2nd-FASAM-NIDE-V methodology to be developed in this work.
The expression of the network of nonlinear neural integro-differential equations of Volterra-type (NIDE-V) considered in this work generalizes the NIE-V net introduced in [33] and is represented in component form by the following system of Nth-order integro-differential equations:
n = 1 N c i , n h t ; f θ ; t d n h i t d t n = g i h t ; f θ + j = 1 T L φ i , j f θ ; t t 0 t d τ ψ j h τ ; f θ ; τ ;   i = 1 , , T H .
The boundary conditions for the functions h i t and their time-derivatives associated with the encoder of the NIDE-V net represented by Equation (1) are typically imposed at the “initial time” t = t 0 and can be represented generically in operator form as follows:
B j h t ; f θ ; t = 0 ;   t = t 0   ;   j = 1 , , B C .
The quantities appearing in Equations (1) and (2) are defined as follows:
(i)
The real-valued scalar quantities t and τ , t 0 t , τ t f are time-like independent variables that parameterize the dynamics of the hidden/latent neuron units. Customarily, the variable t is called the “global time” while the variable τ is called the “local time”, both being defined on the domain Ω t t 0 , t f , where the initial time-value is denoted as t 0 while the stopping (“final”) time-value is denoted as t f . Thus, the dynamics modeled by Equation (1) involves both instantaneous/local and non-local information.
(ii)
The components of the T H -dimensional vector-valued function h t h 1 t , , h T H t represent the hidden/latent neural networks; T H denotes the total number of components of h t . In this work, the symbol “ ” will be used to signify “is defined as” or, equivalently, “is by definition equal to.” The various vectors will be considered to be column vectors. The dagger “ ” symbol will be used to denote “transposition.”
(iii)
The components of the column-vector θ θ 1 , , θ T W represent the “primary parameters,” namely scalar learnable adjustable parameters/weights, in all the latent neural nets, including the encoders(s) and decoder(s). The total number of adjustable parameters/weights is denoted as “ T W .” The components f i θ , i = 1 , , T F , of the vector-valued function f θ f 1 θ , , f T F θ represent the “feature/functions of the primary model parameters feature.” The quantity T F denotes the total number of such feature functions comprised in the NIDE-V. In general, the components f i θ are nonlinear functions of θ . The total number of feature functions must necessarily be smaller than the total number of primary parameters (weights), i.e., T F < T W . When the NIDE-F comprises only primary parameters, it is considered that f i θ θ i for all i = 1 , , T W T F .
(iv)
The functions ψ j h τ ; f θ ; τ model the dynamics of the neurons in a latent space where the local time integration occurs, while the functions φ i , j f θ ; t map the local space back to the original data space. The functions g i h t ; f θ model additional dynamics in the original data space. In general, these functions are nonlinear in their arguments.
(v)
The functions c i , n h t ; f θ ; t are coefficient-functions associated with the order n = 1 , , N of the time-derivatives d n h i t / d t n of the functions h i t . The functions c i , n h t ; f θ ; t may depend nonlinearly on the functions h t and f θ .
(vi)
The operators B j h t ; f θ ; t , j = 1 , , B C , represent boundary conditions imposed at t = t 0 and/or at t = t f on the functions h i t and on their time-derivatives; the quantity “BC” denotes the “total number of boundary conditions.”
The NIDE-V net is “trained” by minimizing a user-chosen loss functional representing the discrepancy between a reference solution (“target data”) and the output produced by the NIDE-V decoder. The “training” process produces “optimal” values for the primary parameters θ θ 1 , , θ T W . These optimal values become the nominal values to be used for subsequent applications/computations using the NIDE-V net. Optimal/nominal values for parameters and all other quantities will be denoted by using the superscript “zero.” Thus, the nominal parameter values will be denoted as θ 0 θ 1 0 , , θ T W 0 . Using these optimal/nominal parameter values to evaluate the NIDE-V net yields the optimal/nominal solution h 0 t , x h 1 0 t , , h T H 0 t , which will formally satisfy the following form of Equation (1):
n = 1 N c i , n h 0 t ; f θ 0 ; t d n h i 0 t d t n = g i h 0 t ; f θ 0 + j = 1 T L φ i , j f θ 0 ; t t 0 t d τ ψ j h 0 τ ; f θ 0 ; τ ;   i = 1 , , T H ;
subject to the following optimized/trained boundary conditions:
B j h 0 t ; f θ 0 ; t = 0 ; a t t = t 0   a n d / o r t = t f ; j = 1 , , B C .
After the NIDE-F net is optimized to reproduce the underlying physical system as closely as possible, the subsequent responses of interest are no longer “loss functionals” but become specific functionals of the NIDE-F’s “decoder” output. Such responses of interest can be generally represented by the functional R h ; f θ defined below:
R h ; f θ = t 0 t f D h t ; f θ ; t d t .
The function D h t ; f θ ; t models the decoder. The scalar-valued quantity R h ; f θ is a functional of h t , x and f θ , and represents the NIDE-V’s decoder-response. At the optimal/nominal parameter values, the decoder response is formally represented using the superscript “zero” as follows:
R h 0 ; f θ 0 = t 0 t f D h 0 t ; f θ 0 ; t d t .
The physical system modeled by the NIDE-V net comprises parameters that stem from measurements and/or computations. Consequently, even in an ideal situation when the NIDE-V net perfectly models the underlying physical system, the NIDE-V’s optimal weights/parameters are unavoidably afflicted by uncertainties stemming from the parameters underlying the physical system. Therefore, the known optimal/nominal values θ 0 of the primary model parameters (“weights”) characterizing the NIDE-V net will differ from the true but unknown values θ of the respective weights by variations denoted as δ θ θ θ 0 . The variations δ θ θ θ 0 will induce corresponding variations, δ f f θ f 0 , f 0 f θ 0 , in the feature functions. The variations δ θ and δ f will induce, through Equation (1), variations v 1 t v 1 1 t , , v T H 1 t δ h 1 t , , δ h T H t around the nominal/optimal functions h 0 t . In turn, the variations δ f f θ f 0 and v 1 t ; x will induce variations δ R h 0 ; F 0 ; v 1 ; δ f ; t in the NIDE-V decoder’s response.
The individual uncertainties afflicting the optimal parameters contribute to the total uncertainty in the decoder response R h ; f θ . These individual contributions are quantified by the sensitivities (i.e., functional derivatives) of the NIDE-V decoder-response with respect to the optimized NIDE-V parameters. The “First-Order Features Adjoint Sensitivity Analysis Methodology for Neural Integro-Differential Equations of Volterra-Type (1st-FASAM-NIDE-V)” aims at obtaining the exact expressions of the first-order sensitivities of the decoder’s response with respect to the feature function and the primary model parameters, followed by the most efficient computation of these sensitivities. The 1st-FASAM-NIDE-V will be established by applying the same principles as those underlying the general 1st-FASAM-N [28] methodology.
The fundamental concept for defining the first-order sensitivity of a vector-valued operator N x acting on a vector x with respect to variations δ x in a neighborhood around nominal values denoted as x 0 has been shown in 1981 by Cacuci [35] to be provided by the 1st-order Gateaux- (G-) variation δ N x 0 ; δ x of N x , which is defined as follows:
δ N x 0 ; δ x d d ε N x 0 + ε δ x ε = 0 lim ε 0 N x 0 + ε δ x N x 0 / ε ,
for a scalar ε and for arbitrary vectors δ x in a neighborhood x 0 + ε δ x around x 0 . When the G-variation δ N x 0 ; δ x is linear in the variation δ x , it is called a “G-differential”, which can be written in the form δ N x 0 ; δ x = N / x x 0 δ x , where N / x x 0 denotes the first-order G-derivative of N x with respect to x , evaluated at x 0 .
Applying the definition provided in Equation (7) to Equation (5) yields the following expression for the first-order G-variation δ R h 0 ; f 0 ; v 1 ; δ f of the response R h ; f θ :
δ R h 0 ; f 0 ; v 1 ; δ f = d d ε t 0 t f D h 0 t + ε v 1 t ; f 0 + ε δ f ; t d t ε = 0 = δ R h 0 ; f 0 ; δ f d i r + δ R h 0 ; f 0 ; v 1 i n d ,
where the “direct effect term” arises directly from variations δ f and is defined as follows:
δ R h 0 ; f 0 ; δ f d i r i = 1 T F t 0 t f d t D h t ; f θ ; t f i δ f i h 0 ; f 0 ,
and where the “indirect effect term” arises indirectly, through the variations v 1 t v 1 1 t , , v T H 1 t δ h 1 t , , δ h T H t in the hidden state functions h t , and is defined as follows:
δ R h 0 ; f 0 ; v 1 i n d i = 1 T H t 0 t f d τ D h t ; f θ ; t h i t v i 1 t h 0 ; f 0 .
The direct-effect term can be quantified using the nominal values h 0 ; f 0 ; the indirect-effect term can be quantified only after having determined the variations v 1 t , which are caused by the variations δ f through the NIDE-V net defined in Equation (1).

2.1. Nth-Order Neural Integral Equations of Volterra-Type (Nth-NIDE-V)

The first-order relationship between the variations v 1 t and δ f is obtained from the first-order G-variations of Equations (1) and (2). It is convenient to rewrite Equation (1) in the following (more compact) form, for i = 1 , , T H :
N i h t ; f θ ; t = g i h t ; f θ + j = 1 T L φ i , j f θ ; t t 0 t d τ ψ j h τ ; f θ ; τ ,
where:
N i h t ; f θ ; t n = 1 N c i , n h t ; f θ ; t d n h i t d t n ;   i = 1 , , T H .
The first-order G-variations of Equations (1) and (2), respectively, are obtained, by definition, as follows:
d N i h 0 t + ε v 1 t ; f θ 0 + ε δ f ; t d ε ε = 0 = d g i h 0 t + ε v 1 t ; f θ 0 + ε δ f d ε ε = 0 + d d ε j = 1 T L φ i , j f θ 0 + ε δ f ; t t 0 t d τ ψ j h 0 τ + ε v 1 τ ; f θ 0 + ε δ f ; τ ε = 0 .
d d ε B j h 0 t + ε v 1 t ; f θ 0 + ε δ f ; t ε = 0 = 0 ; t = t 0 a n d / o r t =   t f ; j = 1 , , B C .
Carrying out the operations indicated in Equation (13) yields the following NIDE-V system for the function v 1 t v 1 1 t , , v T H 1 t δ h 1 t , , δ h T H t :
k = 1 T H N i h ; f θ ; t h k t v k 1 t j = 1 T L φ i , j f θ ; t t 0 t d τ k = 1 T H ψ j h ; f θ ; τ h k τ v k 1 τ h 0 , f 0 k = 1 T H g i h t ; f θ ; t h k t v k 1 t h 0 , f 0 = k = 1 T F q i , k 1 h ; f δ f k , i = 1 , , T H ;
k = 1 T H B j h ; f θ ; t h k t v k 1 t h 0 , f 0 k = 1 T F B j h ; f θ ; t f k δ f k h 0 , f 0 = 0 , a t t = t 0 ;   t = t f ; j = 1 , , B C ;
where:
q i , k 1 h ; f N i h ; f ; t f k + g i h ; f ; t f k + j = 1 T L φ i , j f ; t f k t 0 t d τ ψ j h τ ; f θ ; τ + j = 1 T L φ i , j f ; t t 0 t d τ ψ j h τ ; f θ ; τ f k ; i = 1 , , T H ; k = 1 , , T F .
k = 1 T H N i h ; f θ ; t h k t v k 1 t n = 1 N d n h i t d t n k = 1 T H c i , n h ; f θ ; t h k t v k 1 t + n = 1 N c i , n h ; f θ ; t d n v i 1 t d t n ;   i = 1 , , T H ;
N i h t ; f θ ; t f k n = 1 N c i , n f θ ; t f k d n h i t d t n ; i = 1 , , T H ; k = 1 , , T F .
The NIDE-V system represented by Equations (15) and (16) is called [28] the “1st-Level Variational Sensitivity system” (1st-LVSS) and its solution, v 1 t v 1 1 t , , v T H 1 t δ h 1 t , , δ h T H t , is called [28] “1st-level variational function.” It is important to note that the 1st-LVSS is linear in the variational function v 1 t , although it remains nonlinear in h t , in general. Note also that the 1st-LVSS would need to be solved anew for each variation δ f j , j = 1 , , T F , which is prohibitively expensive computationally if T F is a large number.
The 1st-LVSS can formally be written in operator form as follows:
L h ; f v 1 = Q 1 h ; f δ f ,
where Q 1 h ; f denotes the T H × T F rectangular matrix having as elements the quantities q i , k 1 h ; f ; t , i = 1 , , T H ; k = 1 , , T F , and where the components of the vector-valued operator L h ; f v 1 L 1 h ; f v 1 , , L T H h ; f v 1       are defined as follows:
L i h ; f v 1 k = 1 T H N i h ; f θ ; t h k t v k 1 t k = 1 T H g i h t ; f θ ; t h k t v k 1 t j = 1 T L φ i , j f θ ; t t 0 t d τ ψ j h ; f θ ; τ h k τ v k 1 τ ; i = 1 , , T H .
The need for repeatedly solving the 1st-LVSS can be avoided if the variational function v 1 t could be eliminated from appearing in the expression of the indirect-effect term defined in Equation (10). This goal can be achieved by expressing the right-side of Equation (10) in terms of the solutions of the “1st-Level Adjoint Sensitivity System” (1st-LASS), to be constructed next. The construction of this 1st-LASS is performed in a Hilbert space comprising elements of the same form as v 1 t H 1 Ω t , defined on the domain Ω t t t 0 , t f . This Hilbert space is considered to be endowed with an inner product denoted as χ ( 1 ) t , η ( 1 ) t 1 , where χ ( 1 ) t χ 1 1 t , , χ T H 1 t H 1 Ω t and η ( 1 ) t η 1 1 t , , η T H 1 t H 1 Ω t , and defined as follows:
χ ( 1 ) t , η ( 1 ) t 1 t 0 t f χ ( 1 ) t · η ( 1 ) t d t t 0 t f χ ( 1 ) t η ( 1 ) t d t = j = 1 T H t 0 t f χ i 1 t η i 1 t d t .
The next step is to construct the inner product of Equation (15) with a vector a 1 t a 1 1 t , , a T H 1 t H 1 Ω t , where the superscript “(1)” indicates “1st-Level”, to obtain the following relationship:
i = 1 T H t 0 t f a i 1 t d t k = 1 T H N i h ; f θ ; t h k t v k 1 t k = 1 T H g i h t ; f θ ; t h k t v k 1 t j = 1 T L φ i , j f θ ; t t 0 t d τ k = 1 T H ψ j h ; f θ ; τ h k τ v k 1 τ = i = 1 T H t 0 t f a i 1 t d t k = 1 T F q i , k 1 h ; f ; t δ f k .
The terms appearing in Equation (23) are to be computed at the nominal values h 0 ; f 0 but the superscript “zero” has been omitted for simplicity. The relation in Equation (23) can formally be written in inner-product form as follows:
a 1 t , L h ; f v 1 1 = a 1 t , Q 1 h ; f δ f 1 .
Using the definition of the adjoint operator in H 1 Ω t , the term on the left-side of Equation (20) is integrated by parts and the order of summations is reversed to obtain the following relation:
a 1 t , L h ; f v 1 1 = v 1 t , L * h ; f a 1 1 + P h ; f ; v 1 ; a 1 ,
where the vector-valued operator L * h ; f ; a 1 is the formal adjoint of the vector-valued operator L h ; f ; v 1 and where P h ; f ; v 1 ; a 1 represents the scalar-valued bilinear concomitant evaluated on the boundary t = t 0 and t = t f of Ω t .
It follows from Equations (20) and (25) that the following relation holds:
v 1 t , L * h ; f a 1 1 = a 1 t , Q 1 h ; f δ f 1 P h ; f ; v 1 ; a 1 .
The term on the left-side of Equation (26) is now required to represent the indirect effect term defined in Equation (10) by imposing the following relation:
L * h ; f a 1 = D h t ; f θ ; t h t .
Using Equations (26) and (27) in Equation (10) yields the following expression for the indirect effect term:
δ R h 0 ; F 0 ; v 1 i n d = a 1 t , Q 1 h ; f δ f 1 P h ; f ; v 1 ; a 1 h 0 ; f 0 .
The boundary conditions accompanying Equation (27) for the function a 1 t are now chosen at the time values t = t f and/or t = t 0 so as to eliminate all unknown values of the 1st-level variational function v 1 t from the bilinear concomitant P h ; f ; v 1 ; a 1 , which remain after implementing the initial conditions provided in Equation (2). These boundary conditions for the function a 1 t can be represented in operator form as follows:
B j * h t ; a 1 t ; f θ ; t = 0 ;   a t   t = t f a n d / o r t = t 0 ; j = 1 , , B C .
The Volterra-like NIDE-V system represented by Equations (27) and (29) will be called the “1st-Level Adjoint Sensitivity System” and the solution, a ( 1 ) t , will be called the “1st-level adjoint sensitivity function.” The 1st-LASS is solved using the nominal/optimal values for the parameters and for the function h t , but this fact has not been explicitly indicated in order to simplify the notation. Notably, the 1st-LASS is independent of any parameter variations so it needs to be solved just once to obtain the 1st-level adjoint sensitivity function a ( 1 ) t The 1st-LASS is linear in a ( 1 ) t but is, in general, nonlinear in h t .
Adding the expression of the indirect-effect term obtained in Equation (28) with the expression of the direct-effect term shown in Equation (9) yields the following expression for the total G-differential δ R h 0 ; f 0 ; v 1 ; δ f :
δ R h ; f ; a 1 ; δ f = a 1 t , Q 1 h ; f δ f 1 + i = 1 T F δ f i t 0 t f D h t ; f θ ; t f i d t P ^ h ; f ; δ f ,
where the quantity P ^ h ; f ; δ f denotes the known non-zero boundary terms that might remain after using, in Equation (28), the boundary conditions provided in Equations (16) and (29). The first-order sensitivities of the decoder response with respect to the components f i θ of the feature function f θ are determined by identifying, in Equation (30), the quantities that multiply the variations δ f i . Consequently, Equation (30) can be represented formally as follows:
δ R h ; f ; a 1 ; δ f = i = 1 T F δ f i R i 1 h ; f ; a 1 ; R i 1 h ; f ; a 1 t 0 t f D i 1 h ; f ; a 1 ; t d t ,
where R i 1 h ; f ; a 1 denotes the first-order sensitivities, as indicated by the superscript “(1)”, of the decoder response with respect to the components f i θ of the feature function f θ . In view of Equation (30), these first-order sensitivities are functionals that can always be represented in the integral form shown in Equation (31).
The sensitivities with respect to the primary model parameters can be obtained by using the result shown in Equation (31) together with the “chain rule” of differentiating compound functions, as follows:
R θ j = i = 1 T F R f i f i θ j , j = 1 , , T W .
When there are only model parameters (i.e., there are no feature functions of model parameters), then f i θ θ i for all i = 1 , , T F T W , and the expression obtained in Equation (31) directly yields the first-order sensitivities R / θ j for all j = 1 , , T W . In this case, all the sensitivities R / θ j , for all j = 1 , , T W would be obtained by computing integrals using quadrature formulas. In contradistinction, when features of parameters can be established, only T F   T F < T W integrals would need to be computed (using quadrature formulas) to obtain the R / F j , j = 1 , , T F ; the sensitivities with respect to the model parameters would subsequently be obtained analytically using the chain-rule provided in Equation (32).
In the Section 2.2, below, the general 1st-FASAM-NIDE-V methodology will be particularized for the second-order (n = 2) 2nd-NIDE-V net to illustrate explicitly the form of the bilinear concomitant and the ensuing adjoint boundary conditions. Notably, the 2nd-NIDE-V system subsumes the first-order 1st-NIDE-V as a particular case.

2.2. Second-Order Neural Integral Equations of Volterra-Type (2nd-NIDE-V)

This subsection presents the development of the 1st-FASAM-NIDE-V methodology for the particular case of the Second-Order Neural Integral Equations of Volterra-Type (2nd-NIDE-V). This development will present the explicit expressions of the corresponding bilinear concomitant and ensuing adjoint boundary conditions. The representation of the second-order n = 2 neural integral equations of Volterra-type (2nd-NIDE-V) is provided below, for i = 1 , , T H :
c i , 1 h t ; f θ d h i t d t + c i , 2 h t ; f θ d 2 h i t d t 2 = g i h t ; f θ + j = 1 T L φ i , j f θ ; t t 0 t ψ j h τ ; f θ ; τ d τ ; i = 1 , , T H .
There are several combinations of boundary conditions that can be provided for the function h i t and/or for its first-derivative d h i t / d t , i = 1 , , T H , either at t = t 0 (encoder) or at t = t f (decoder), or a combination thereof. For illustrative purposes, consider that the boundary conditions are as follows:
h i t 0 = e i ;   h i t f = d i ;   i = 1 , , T H .
The 1st-LVSS is obtained by taking the G-variations of Equations (33) and (34) to obtain the following system, comprising the forms taken on for n = 2 by Equations (15) and (16), for the variational function v 1 t v 1 1 t , , v T H 1 t δ h 1 t , , δ h T H t :
c i , 1 h t ; f θ h 0 , f 0 d v i 1 t d t + d h i t d t k = 1 T H c i , 1 h t ; f θ h k t v k 1 t h 0 , f 0 + c i , 2 h t ; f θ h 0 , f 0 d 2 v i 1 d t 2 + d 2 h i t d t 2 k = 1 T H c i , 2 h t ; f θ h k t v k 1 t h 0 , f 0 j = 1 T L φ i , j f θ ; t t 0 t d τ k = 1 T H ψ j h ; f θ ; τ h k τ v k 1 τ h 0 , f 0 k = 1 T H g i h t ; f θ ; t h k t v k 1 t h 0 , f 0 = k = 1 T F q i , k 1 h ; f h 0 , f 0 δ f k , i = 1 , , T H ;
v i 1 t 0 = δ e i ; v i 1 t f = δ d i ; i = 1 , , T H ;
where the following definition was used for q i k 1 h ; f , i = 1 , , T H ;     k = 1 , , T F :
q i k 1 h ; f d h i t d t c i , 1 h t ; f θ f k d 2 h i t d t 2 c i , 2 h t ; f θ f k + g i h ; f ; t f k + j = 1 T L φ i , j f ; t f k t 0 t ψ j h τ ; f θ ; τ d τ + j = 1 T L φ i , j f ; t t 0 t d τ ψ j h τ ; f θ ; τ f k .
The 1st-LASS is constructed by forming the inner product of Equation (35) with a vector a 1 t a 1 1 t , , a T H 1 t H 1 Ω t to obtain the following relationship:
i = 1 T H t 0 t f a i 1 t d t c i , 1 h t ; f d v i 1 t d t + d h i t d t k = 1 T H c i , 1 h t ; f h k t v k 1 t + c i , 2 h t ; f d 2 v i 1 t d t 2 + d 2 h i t d t 2 k = 1 T H c i , 2 h t ; f h k t v k 1 t j = 1 T L φ i , j f ; t t 0 t d τ k = 1 T H ψ j h τ ; f ; τ h k τ v k 1 τ k = 1 T H g i h t ; f ; t h k t v k 1 t = i = 1 T H t 0 t f a i 1 t d t k = 1 T F q i , k 1 h ; f ; t δ f k .
Examining the structure of the left-side of Equation (38) reveals that the bilinear concomitant will arise from the integration by parts of the first and third terms the on the left-side of Equation (38), as follows:
i = 1 T H t 0 t f a i 1 t c i , 1 h ; f d v i 1 t d t d t + i = 1 T H t 0 t f a i 1 t c i , 2 h ; f d 2 v i 1 t d t 2 d t = P h ; f ; v 1 ; a 1 i = 1 T H t 0 t f v i 1 t d d t a i 1 t c i .1 h ; f d t + i = 1 T H t 0 t f v i 1 t d 2 d t 2 a i 1 t c i .2 h ; f d t ,
where the bilinear concomitant P h ; f ; v 1 ; a 1 has the following expression:
P h ; f ; v 1 ; a 1 i = 1 T H a i 1 t f c i , 1 h t f ; f v i 1 t f a i 1 t 0 c i , 1 h t 0 ; f v i 1 t 0 + i = 1 T H a i 1 t f c i , 2 h t f ; f d v i 1 t f d t a i 1 t 0 c i , 2 h t 0 ; f d v i 1 t 0 d t i = 1 T H v i 1 t f d d t a i 1 t c i , 2 h ; f t = t f + i = 1 T H v i 1 t 0 d d t a i 1 t c i , 2 h ; f t = t 0 .
The second term on the left-side of Equation (38) will be recast in its “adjoint form” by reversing the order of summations so as to transform the inner product involving the function a ( 1 ) t into an inner product involving the function v 1 t , as follows:
i = 1 T H t 0 t f a i 1 t d t d h i t d t k = 1 T H c i , 1 h ; f h k t v k 1 t = i = 1 T H t 0 t f v i 1 t c i , 1 h ; f h i t k = 1 T H a k 1 t d h k t d t d t .
The fourth term on the left-side of Equation (38) will be recast in its “adjoint form” by reversing the order of summations so as to transform the inner product involving the function a ( 1 ) t into an inner product involving the function v 1 t , just as was done for obtaining the relation in Equation (41), to obtain the following relation:
i = 1 T H t 0 t f a i 1 t d t d 2 h i t d t 2 k = 1 T H c i , 2 h ; f h k t v k 1 t = i = 1 T H t 0 t f v i 1 t c i , 2 h ; f h i t k = 1 T H a k 1 t d 2 h k t d t 2 d t .
The fifth term on the left-side of Equation (38) is now recast in its “adjoint form” by “integrating by parts” and reversing the order of summations and integrations so as to transform the inner product involving the function a ( 1 ) t into an inner product involving the function v 1 t , as follows:
i = 1 T H t 0 t f d t a i 1 t j = 1 T L φ i , j f ; t t 0 t k = 1 T H ψ j h ; f ; τ h k τ v k 1 τ d τ = i = 1 T H t 0 t d t a i 1 t j = 1 T L φ i , j f ; t t 0 t k = 1 T H ψ j h ; f ; τ h k τ v k 1 τ d τ t = t f i = 1 T H t 0 t d t a i 1 t j = 1 T L φ i , j f ; t t 0 t k = 1 T H ψ j h ; f ; τ h k τ v k 1 τ d τ t = t 0 i = 1 T H t 0 t f d t v i 1 t j = 1 T L ψ j h ; f ; t h i t t 0 t k = 1 T H a k 1 τ φ k , j f ; τ d τ = i = 1 T H t 0 t f d t v i 1 t j = 1 T L ψ j h ; f ; t h i t t t f k = 1 T H a k 1 τ φ k , j f ; τ d τ
The sixth term on the left-side of Equation (38) will be recast in its “adjoint form” by reversing the order of summations and integrations so as to transform the inner product involving the function a ( 1 ) t into an inner product involving the function v 1 t , as follows:
i = 1 T H t 0 t f a i 1 t d t k = 1 T H g i h ; f ; t h k t v k 1 t = i = 1 T H t 0 t f v i t k = 1 T H a k 1 t g k h t ; f ; t h i t .
Collecting the results obtained in Equations (39)–(44) yields the following expression for the left-side of Equation (38):
i = 1 T H t 0 t f a i 1 t d t c i , 1 h ; f d v i 1 t d t + d h i t d t k = 1 T H c i , 1 h ; f h k t v k 1 t + c i , 2 h ; f d 2 v i 1 t d t 2 + d 2 h i t d t 2 k = 1 T H c i , 2 h ; f h k t v k 1 t j = 1 T L φ i , j f ; t t 0 t f d τ k = 1 T H ψ j h ; f ; τ h k τ v k 1 τ k = 1 T H g i h ; f ; t h k t v k 1 t = P h ; f ; v 1 ; a 1 + i = 1 T H t 0 t f v i 1 t d d t a i 1 t c i , 1 h ; f d t + i = 1 T H t 0 t f v i 1 t d 2 d t 2 a i 1 t c i , 2 h ; f d t + i = 1 T H t 0 t f v i 1 t c i , 1 h ; f h i t k = 1 T H a k 1 t d h k t d t d t + i = 1 T H t 0 t f v i 1 t c i , 2 h ; f h i t k = 1 T H a k 1 t d 2 h k t d t 2 d t i = 1 T H t 0 t f v i t k = 1 T H a k 1 t g k h ; f ; t h i t i = 1 T H t 0 t f d τ v i 1 τ j = 1 T L ψ j h ; f ; τ h i τ t t f k = 1 T H a k 1 t φ k , j f ; t d t
Using Equation (39) and rearranging the terms on the right-side of Equation (45) yields the following relation:
i = 1 T H t 0 t f a i 1 t d t k = 1 T F q i , k 1 h ; f ; t δ F k P h ; f ; v 1 ; a 1 = i = 1 T H t 0 t f v i t d t d d t a i 1 t c i , 1 h ; f + d 2 d t 2 a i 1 t c i , 2 h ; f + c i , 1 h ; f h i t k = 1 T H a k 1 t d h k t d t + c i , 2 h ; f h i t k = 1 T H a k 1 t d 2 h k t d t 2 d t k = 1 T H a k 1 t g k h ; f ; t h i t j = 1 T L ψ j h ; f ; t h i t t t f k = 1 T H a k 1 τ φ k , j f ; τ d τ .
The term on the right-side of Equation (46) is now required to represent the “indirect-effect” term defined in Equation (10), which is achieved by requiring the components of the function a 1 t a 1 1 t , , a T H 1 t to satisfy the following 1st-LASS:
d d t a i 1 t c i , 1 h ; f + d 2 d t 2 a i 1 t c i , 2 h ; f + c i , 1 h ; f h i t k = 1 T H a k 1 t d h k t d t + c i , 2 h ; f h i t k = 1 T H a k 1 t d 2 h k t d t 2 d t k = 1 T H a k 1 t g k h ; f ; t h i t j = 1 T L ψ j h ; f ; t h i t t t f k = 1 T H a k 1 τ φ k , j f ; τ d τ = D h ; f ; t h i t .
Recalling from Equation (36) that the values v i 1 t 0 and v i 1 t f , i = 1 , , T H , are known, it follows that the unknown values involving the function v i t in the bilinear concomitant P h ; f ; v 1 ; a 1 defined in Equation (40) are eliminated by imposing the following conditions on the components of the 1st-level adjoint sensitivity function a 1 t a 1 1 t , , a T H 1 t :
a i 1 t 0 = 0 ; a i 1 t f = 0 ; i = 1 , , T H .
It follows from Equations (45)–(48) that the indirect-effect term defined in Equation (10) has the following expression in terms of the 1st-level adjoint sensitivity function a ( 1 ) t :
δ R h ; f ; a ( 1 ) i n d = t 0 t f a i 1 t d t k = 1 T F q i k 1 h ; f ; t δ f k P ^ h ; f ; v 1 ; a 1 ,
where the boundary quantity P ^ h ; f ; v 1 ; a 1 contains the remaining terms after having implemented the known boundary conditions given in Equations (36) and (48), and has the following explicit expression:
P ^ h ; f ; v 1 ; a 1 i = 1 T H δ d i c i , 2 h ; f d a i 1 t d t t = t f + i = 1 T H δ e i c i , 2 h ; f d a i 1 t d t t = t 0 .
Using the results obtained in Equations (9), (37), (49), and (50) in Equation (8) yields the following expression for the G-variation δ R h 0 ; f 0 ; v 1 ; δ f , which is seen to be linear in the variations δ d i , δ e i ( i = 1 , , T H ) and δ f j ( j = 1 , , T F ):
δ R h t ; f θ ; a ( 1 ) t ; δ f j = 1 T F t 0 t f d t D h ; f ; t F j δ f j + i = 1 T H t 0 t f a i 1 t d t j = 1 T F q j k 1 h ; f ; t δ f j + i = 1 T H δ d i c i , 2 h ; f d a i 1 t d t t = t f i = 1 T H δ e i c i , 2 h ; f d a i 1 t d t a i 1 t 0 c i , 1 h ; f t = t 0 j = 1 T F R f j δ f j + i = 1 T H R d i δ d i + i = 1 T H R e i δ e i .
The expression in Equation (51) is to be satisfied at the nominal/optimal values for the respective model parameters, but this fact has not been indicated explicitly in order to simplify the notation.
It also follows from Equations (51) and (37) that the sensitivities R / f j of the response R h ; f θ with respect to the components f j θ of the feature function f θ have the following expressions:
R h ; f θ f j = t 0 t f D h t ; f θ ; t f j d t i = 1 T H t 0 t f a i 1 t d h i t d t c i , 1 h t ; f θ f j d t i = 1 T H t 0 t f a i 1 t d 2 h i t d t 2 c i , 2 h t ; f θ f j d t + i = 1 T H t 0 t f a i 1 t g i h ; f ; t f j d t + i = 1 T H t 0 t f a i 1 t k = 1 T L φ i , k f ; t f j t 0 t f ψ k h τ ; f θ ; τ d τ + i = 1 T H t 0 t f a i 1 t k = 1 T L φ i , k f ; t d t t 0 t f d τ ψ k h τ ; f θ ; τ f j ; j = 1 , , T F .
Identifying in Equation (51) the expressions that multiply the variations δ e i yields the following expressions for the decoder response sensitivities with respect to the encoder’s initial-time conditions:
R e i = c i , 2 h ; f d a i 1 t d t + a i 1 t 0 c i , 1 h ; f t = t 0 ; i = 1 , , T H .
Identifying in Equation (51) the expressions that multiply the variations δ d i yields the following expressions for the decoder response sensitivities with respect to the final-time conditions:
R d i = c i , 2 h ; f d a i 1 t d t t = t f ; i = 1 , , T H .
As a particular case, the 2nd-NIDE-V system considered in Equations (33) and (34) encompasses the 1st-NIDE-V below, which is obtained by setting c i , 2 h ; f 0 and d i 0 , i = 1 , , T H , in these equations:
c i , 1 h t ; F θ d h i t d t = g i h t ; F θ + j = 1 T L φ i , j F θ ; t t 0 t d τ ψ j h τ ; F θ ; τ .
h i t 0 = e i ; i = 1 , , T H ,
The first-order sensitivities of the decoder with respect to the feature functions and initial conditions for the above 1st-NIDE-V are obtained by setting c i , 2 h ; f 0 and d i 0 , i = 1 , , T H , in Equations (52)–(54), to obtain the following expressions:
R h ; f θ f j = t 0 t f D h t ; f θ ; t f j d t i = 1 T H t 0 t f a i 1 t d h i t d t c i , 1 h t ; f θ f j d t + i = 1 T H t 0 t f a i 1 t g i h ; f ; t f j d t + i = 1 T H t 0 t f a i 1 t k = 1 T L φ i , k f ; t f j t 0 t f ψ k h τ ; f θ ; τ d τ + i = 1 T H t 0 t f a i 1 t k = 1 T L φ i , k f ; t d t t 0 t f d τ ψ k h τ ; f θ ; τ f j ; j = 1 , , T F .
R e i = a i 1 t 0 c i , 1 h ; f t = t 0 ; i = 1 , , T H .
It is apparent from Equation (58) that, for the 1st-NIDE-V defined by Equations (55) and (56), the sensitivities R / e i are proportional to the values of the respective component a i 1 t 0 of the 1st-level adjoint function evaluated at the initial-time t = t 0 . This relation provides an independent mechanism for verifying the correctness of solving the 1st-LASS from t = t f to t = t 0 (backwards in time) since the sensitivities R / e i can be computed independently of the 1st-LASS using finite differences of appropriately high-order in conjunction with known variations δ e i and the correspondingly induced variations in the decoder response. Special attention needs to be devoted, however, to ensure that the respective finite-difference formula is accurate, which may need several trials with different values chosen for the variation δ e i .

Effects of Alternative Boundary Conditions for the Forward Functions h i t

If the boundary conditions imposed on the forward functions h i t and/or the first-derivatives d h i t / d t , i = 1 , , T H , differ from the illustrative ones selected in Equation (34) then the corresponding boundary conditions for the 1st-level adjoint function a 1 t a 1 1 t , , a T H 1 t would also differ from the ones shown in Equation (48), as would be expected. The components of a 1 t a 1 1 t , , a T H 1 t would consequently have different values and so all of the first-order sensitivities R / F j would have values different from those computed using Equation (52), even though the formal mathematical expressions of the respective sensitivities would remain unchanged. Of course, the sensitivities R / e i and R / d i would have expressions that would differ from those in Equations (53) and (54), respectively, if the boundary conditions in Equation (34), and consequently those in Equation (48), were different. This is because the residual bilinear concomitant P ^ h ; F ; v 1 ; a 1 would have a different expression from that shown in Equation (50).
The above considerations will be illustrated next by replacing the boundary conditions considered in Equation (34) by the following typical “initial conditions” for the encoder of the NIDE-V net defined in Equation (33):
h i t 0 = e i ;   d h i t d t t = t 0 = ξ i ;   i = 1 , , T H .
Note that the same symbols, e.g., h i t , v i 1 t , a i 1 t , etc., will be used in this “Subsubsection” for the various functions, even though the values of these functions will differ from the corresponding ones in Section 2.2, since the “initial encoder conditions” considered in Equation (59) differ from the “initial/final encoder/decoder conditions” considered in Equation (34).
The G-differentiated expressions of the above “initial conditions”, which are the counterparts of the expressions provided in Equation (36), are as follows:
v i 1 t 0 = δ e i ; d v i 1 t d t t = t 0 = δ ξ i ; i = 1 , , T H .
For the above boundary conditions, the 1st-LVSS for the function v 1 t v 1 1 t , , v T H 1 t δ h 1 t , , δ h T H t comprises Equations (35) and (60). The 1st-LASS corresponding to the 1st-LVSS comprising Equations (35) and (60) is obtained by following the same procedure as detailed in Section 2.2. The main distinction from the case considered in Section 2.2 will be the expression of the residual bilinear concomitant P ^ h ; f ; v 1 ; a 1 . Thus, applying the same sequence of steps as in Section 2.2 to Equations (35) and (60), but omitting the details for brevity, yields the following expressions for the 1st-order sensitivities of the decoder response:
R h ; f θ f j = t 0 t f D h t ; f θ ; t f j d t i = 1 T H t 0 t f a i 1 t d h i t d t c i , 1 h t ; f θ f j d t i = 1 T H t 0 t f a i 1 t d 2 h i t d t 2 c i , 2 h t ; f θ f j d t + i = 1 T H t 0 t f a i 1 t g i h ; f ; t f j d t + i = 1 T H t 0 t f a i 1 t k = 1 T L φ i , k f ; t f j t 0 t f ψ k h τ ; f θ ; τ d τ + i = 1 T H t 0 t f a i 1 t k = 1 T L φ i , k f ; t d t t 0 t f d τ ψ k h τ ; f θ ; τ f j ; j = 1 , , T F .
R e i = a i 1 t 0 c i , 1 h ; f d d t a i 1 t c i , 2 h ; f t = t 0 ; i = 1 , , T H .
R ξ i = a i 1 t 0 c i , 2 h t 0 ; f ; i = 1 , , T H .
The 1st-level adjoint function that appears in Equations (61)–(63) is the solution of a 1st-LASS comprising a system of NIDE-V equations that formally resembles the system presented in Equation (47) but having boundary conditions that differ from those in Equation (48), as follows:
d d t a i 1 t c i , 1 h ; f + d 2 d t 2 a i 1 t c i , 2 h ; f + c i , 1 h ; f h i t k = 1 T H a k 1 t d h k t d t + c i , 2 h ; f h i t k = 1 T H a k 1 t d 2 h k t d t 2 d t k = 1 T H a k 1 t g k h ; f ; t h i t j = 1 T L ψ j h ; f ; t h i t t t f k = 1 T H a k 1 τ φ k , j f ; τ d τ = D h ; f ; t h i t .
a i 1 t f = 0 ; d d t a i 1 t c i , 2 h ; f t = t f = 0 ; i = 1 , , T H .

3. Illustrative Application of the 1st-FASAM-NIDE-V Methodology to a Heat Conduction Model

The application of the 1st-FASAM-NIDE-V methodology will be illustrated in this section by considering a paradigm model of nonlinear heat conduction through a one-dimensional slab of thickness l , made of material having a temperature dependent conductivity k T , insulated at one side at a temperature T 0 . The slab is heated internally by a nonlinear temperature-dependent heat source that is proportional to the power within the slab. There are several reasons for choosing this illustrative paradigm heat conduction model, including:
(i)
Conduction heat transfer is a physical process occurring ubiquitously in energy engineering, being the subject of many textbooks [36,37,38,39,40,41].
(ii)
The illustrative model is nonlinear but admits an equivalent linear model that is related to the original model through the Kirchhoff transformation [40].
(iii)
The illustrative model can be represented either as a traditional NODE system or a NIDE-V system.
(iv)
The illustrative model admits explicit closed-form solutions for all quantities of interest, including the functions representing the hidden/latent neural networks, the decoder response, and the sensitivities of the decoder’s response to the optimal weights/parameters.
(v)
The illustrative model enables an exact comparison between the application of the 2nd-FASAM-NIDE-V methodology developed in Section 2 and the equivalent application of the 2nd-FASAM-NODE methodology developed in [30,31].
The traditional representation of the heat conduction process within the slab described above is by means of the following nonlinear second-order NODE-net:
d d x k T d T x d x + Q T x + β 2 T 2 x = 0 ; 0 < x < l ; k T k 0 1 + β T x ;
T 0 = T 0 ; d T x d x x = 0 = 0 .
The quantities T 0 , k 0 , β , and Q in Equation (66) are imprecisely-known model parameters/weights subject to experimental uncertainties; these parameters are considered to be components of the “vector of model parameters” denoted as θ T 0 , k 0 , β , Q . The optimal/nominal values of these parameters are considered to be known and are denoted using a superscript “zero”, as follows: T 0 0 , k 0 0 , β 0 , Q 0 , with θ 0 T 0 0 , k 0 0 , β 0 , Q 0 .
Alternatively, the above heat conduction model can be represented by the following non-linear NIDE-V system:
1 + β T x d T x d x + b 2 θ 0 x T y + β 2 T 2 y d y = 0 ; b 2 θ Q     / k 0   ; 0 < x < l ;
T 0 = T 0 .
The decoder’s response of interest is chosen to be the “power per unit distance” (also called the “linear power”) developed within the slab, which depends nonlinearly on the temperature and is defined as follows:
R = 0 l k T d T x d x d x .
Heat conduction models involving temperature-dependent conductivities, k T , are typically solved by using the Kirchhoff transformation defined below:
U T = 1 k 0 0 T x k τ d τ = T x + β 2 T 2 x .
The NODE-form of the heat conduction model defined by Equations (66) and (67) has the following expression in the Kirchhoff-transformed framework:
d 2 U x d x 2 + b 2 θ U x = 0 ; 0 < x < l ;
U 0 = U 0 θ T 0 + β 2 T 0 2 ; d U x d x x = 0 = 0 .
On the other hand, the NIDE-V form of the heat conduction model defined by Equations (68) and (69) has the following form in the Kirchhoff-transformed framework:
d U x d x + b 2 θ 0 x U y d y = 0 ; 0 < x < l ;
U 0 = U 0 θ .
The expression of the decoder response in the Kirchhoff-transformed framework is obtained by using Equation (71) in Equation (70), which yields:
R = k 0 0 l d U x d x d x = k 0 U l U 0 = k 0 0 l U x δ x l d x k 0 U 0 θ .
Solving either the NODE or the NIDE-V forms of the illustrative heat conduction model, in either the Kirchhoff-transformed framework or the temperature-framework, yields the following exact closed-form expressions, respectively:
U x = U 0 θ cos x b θ ;
T x = 1 + 2 β U x 1 / β = 1 + 2 β U 0 θ cos x b θ 1 / β .
Using Equation (76), the result obtained in Equation (77), or, alternatively, using in Equation (70) the result obtained Equation (78), yields the following closed-form expression for the decoder response:
R = k 0 U 0 θ cos l b θ 1 = k 0 T 0 + β 2 T 0 2 cos l b θ 1 .

3.1. Application of the 1st-FASAM-NIDE-V to Compute 1st-Order Sensitivities of the NIDE-V Representation of the Illustrative Heat Conduction Model

The 1st-FASAM-NIDE-V methodology will be applied in this subsection to the paradigm heat conduction model in order to illustrate the path to obtaining the exact expressions of the first-order sensitivities of the decoder response with respect to the model parameters and subsequently compute these sensitivities most efficiently, requiring a single large-scale computation for solving the respective 1st-Level Adjoint Sensitivity System (1st-LASS). The illustrative mathematical derivations will be performed in the temperature-framework formulation (in Section 3.1.1) and also in the Kirchhoff-transformed framework formulation (in Section 3.1.2) in order to illustrate the advantages of using linearizing transformations, such as Kirchhoff’s transformation, whenever possible.

3.1.1. Representation in the NIDE-V Temperature-Framework

The first-order sensitivities of the decoder response defined in Equation (70) are obtained by determining the first-order G-differential of the respective expression, for variations δ T and δ θ around the nominal values θ 0 T 0 0 , k 0 0 , β 0 , Q 0 , which is by definition obtained as follows:
δ R T 0 , θ 0 ; δ T , δ θ d d ε k 0 0 + ε δ k 0 0 l 1 + β 0 + ε δ β T 0 + ε δ T × d T 0 x + ε δ T x d x d x ε = 0 = δ k 0 0 l 1 + β T x d T x d x θ 0 + δ β k 0 0 l T x d T x d x d x θ 0 + k 0 β 0 l d d x T x δ T x d x θ 0 = δ R T 0 , θ 0 ; δ θ d i r + δ R T 0 , θ 0 ; δ T i n d ,
where the direct-effect term δ R T 0 , θ 0 ; δ θ d i r and indirect-effect term δ R T 0 , θ 0 ; δ T i n d , respectively, are defined below:
δ R T 0 , θ 0 ; δ θ d i r δ k 0 0 l 1 + β T x d T x d x θ 0 + δ β k 0 0 l T x d T x d x d x θ 0 δ T 0 T 0 k 0 β θ 0 ;
δ R T 0 , θ 0 ; δ T i n d k 0 β T l δ T l θ 0 = k 0 β T l 0 l δ T x δ x l d x θ 0 .
The direct-effect term can be computed at this time, but the indirect-effect term can be computed after having determined the variational function δ T x . This variational function is the solution of the 1st-LVSS obtained by G-differentiating Equations (68) and (69). Applying the definition of the G-differential to these equations yields the following relations for the variational function δ T x :
d d ε 1 + β 0 + ε δ β T 0 + ε δ T d T 0 + ε δ T d x ε = 0 + d d ε b 0 + ε δ b 2 0 x T 0 + ε δ T + β 0 + ε δ β 2 T 0 + ε δ T 2 d y ε = 0 = 0 ;          
d d ε T 0 x + ε δ T x x = 0 ε = 0 = δ T 0 .
Carrying out the operations indicated in Equations (83) and (84) yields the following NIDE-V system representing the 1st-LVSS for the variational function δ T x :
L 1 T ; θ δ T x β δ T x d T x d x + T x d d x δ T x + b 2 0 x 1 + β T y δ T y d y = δ β T x d T x d x δ β b 2 2 0 x T 2 y d y 2 b δ b 0 x T y + β 2 T 2 y d y ;  
δ T 0 = δ T 0 .
The above 1st-LVSS for the variational function δ T x is to be satisfied at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation. It is important to note that the operator L 1 T ; θ δ T x acts linearly on δ T x . The 1st-LVSS would need to be solved repeatedly to obtain the variational function δ T x for every parameter variation of interest. Repeatedly solving the 1st-LVSS can be avoided by expressing the indirect effect term defined in Equation (82) in terms of a “1st-level adjoint sensitivity function” that will be the solution of a 1st-LASS to be constructed next by following the general procedure outlined in Section 2. The inner product appropriate for constructing the 1st-LASS corresponding to the 1st-LVSS defined by Equations (85) and (86) for the single-component variational function δ T x takes on the following particular form of Equation (22):
χ ( 1 ) x , η ( 1 ) x 1 0 l χ ( 1 ) x η ( 1 ) x d x .
The construction of the 1st-LASS will be accomplished by implementing the following sequence of steps:
  • Use Equation (87) to construct the inner product of Equation (85) with a yet specified function η 1 x to obtain the following relation:
    0 l η 1 x L 1 T ; θ δ T x = 0 l η 1 x β δ T x d T x d x + β T x d d x δ T x d x + b 2 0 l η 1 x 0 x 1 + β T y δ T y d y d x = q 1 T ; η 1 ; θ ; δ θ ,
    where:
    q 1 T ; η 1 ; θ ; δ θ δ β 0 l η 1 x T x d T x d x d x δ β b 2 2 0 l η 1 x d x 0 x T 2 y d y 2 b δ b 0 l η 1 x 0 x T y + β 2 T 2 y d y d x .
  • Transfer the operations on δ T x to operations (differentiation and integration) on η 1 x in Equation (88) by constructing the formal adjoint operator L 1 * T ; θ satisfying the relation
    0 l η 1 x L 1 T ; θ δ T x d x = 0 l δ T x L 1 * T ; θ η 1 x d x + P 1 T ; θ ; δ T ; η 1 ,
    where P 1 T ; θ ; δ T ; η 1 denotes the bilinear concomitant comprising boundary terms. The relationship shown in Equation (90) is derived by integrating by parts the second and third terms in the middle expression of Equation (88) to obtain the following relations:
    β 0 l η 1 x T x d d x δ T x d x = β η 1 l T l δ T l β η 1 0 T 0 δ T 0 β 0 l δ T x d d x η 1 x T x d x .
    b 2 0 l η 1 x d x 0 x 1 + β T y δ T y d y = b 2 0 x η 1 y d y × 0 x 1 + β T y δ T y d y x = 0 x = l b 2 0 l 1 + β T x δ T x d x 0 x η 1 y d y = b 2 0 l δ T x 1 + β T x d x x l η 1 y d y .
  • Insert the results obtained in Equations (91) and (92) into the left-side Equation (88) and collect the terms multiplying the function δ T x to obtain the relation below:
    0 l δ T x L 1 * T ; θ η 1 x d x = 0 l δ T x β η 1 x d T x d x β d d x η 1 x T x + b 2 1 + β T x x l η 1 y d y d x = q 1 T ; η 1 ; θ ; δ θ β η 1 l T l δ T l + β η 1 0 T 0 δ T 0 ,
  • Require the term on the left-side of Equation (93) to represent the indirect-effect term defined in Equation (82), by requiring that the function η 1 x satisfy the following NIDE-V net:
    L 1 * T ; θ η 1 x = β η 1 x d T x d x β d d x η 1 x T x + b 2 1 + β T x x l η 1 y d y = k 0 β T l δ x l .
  • Eliminate the unknown boundary term in Equation (93) by imposing the following boundary condition of the function η 1 x :
    η 1 l = 0 .
    The NIDE-V system comprising Equations (94) and (95) constitutes the 1st-Level Adjoint Sensitivity System (1st-LASS) for the 1st-level adjoint sensitivity function η 1 x .
  • Use Equations (86), (94), and (95) in Equation (93) to obtain the following expression for the indirect-effect term:
    δ R T 0 , θ 0 ; δ T i n d = q 1 T ; η 1 ; θ ; δ θ + δ T 0 β η 1 0 T 0 .
  • Adding the expression obtained in Equation (96) with the expression for the direct-effect term defined in Equation (81) and using the definition of the quantity q 1 T ; η 1 ; θ ; δ θ provided in Equation (89) yields the following expression for the G-differential δ R T 0 , θ 0 ; δ T , δ θ :
    δ R T 0 , θ 0 ; δ T , δ θ = δ k 0 0 l 1 + β T x d T x d x + δ β k 0 0 l T x d T x d x d x δ T 0 T 0 k 0 β + δ T 0 β η 1 0 T 0 δ β 0 l η 1 x T x d T x d x d x δ β b 2 2 0 l η 1 x d x 0 x T 2 y d y 2 b δ b 0 l η 1 x d x 0 x T y + β 2 T 2 y d y .
    The above expression is to be satisfied at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation.
  • From the definition of the “feature-function” b 2 θ Q     / k 0 provided in Equation (68), it follows that:
    δ b θ = δ Q 2 k 0 Q δ k 0 2 k 0 Q k 0 .
  • Inserting the expression provided in Equation (98) into Equation (97) and collecting the terms that multiply the respective parameter variations yields the following expressions of the first-order sensitivities of the decoder response with respect to the model parameters:
    R T 0 = T 0 k 0 β + β η 1 0 T 0 ;
    R β = k 0 0 l T x d T x d x d x 0 l η 1 x T x d T x d x d x Q 2 k 0 0 l η 1 x d x 0 x T 2 y d y ;
    R k 0 = Q k 0 2 0 l η 1 x d x 0 x T y + β 2 T 2 y d y
    R Q = 1 k 0 0 l η 1 x d x 0 x T y + β 2 T 2 y d y .
The expressions presented in Equations (99)–(102) can be evaluated after having obtained the 1st-level adjoint sensitivity function η 1 x by solving, a single time, the 1st-LASS represented by the NIDE-V system comprising Equations (94) and (95). These computations will not be performed explicitly at this stage because the closed-form expressions of the sensitivities of the decoder response with respect to the model parameters will be obtained by much simpler algebraic manipulations, in the NIDE-V Kirchhoff-transformed framework, as will be presented in Section 3.1.2, below.

3.1.2. Representation in the NIDE-V Kirchhoff-Transformed Framework

In the Kirchhoff-transformed framework, the 1st-order sensitivities of the decoder response defined in Equation (76) are obtained by determining the first-order G-differential, δ R U 0 , θ 0 ; δ U , δ θ , of the respective expression, for variations δ U , δ θ around the nominal values θ 0 T 0 0 , k 0 0 , β 0 , Q 0 , which is by definition obtained as follows:
δ R U 0 , θ 0 ; δ U , δ θ d d ε k 0 0 + ε δ k 0 0 l U 0 x + ε δ U x δ x l d x d d ε k 0 0 + ε δ k 0 U 0 0 + ε δ U 0 ε = 0 = δ R U 0 , θ 0 ; δ θ d i r + δ R U 0 , θ 0 ; δ U i n d ,
where the direct-effect term δ R U 0 , θ 0 ; δ θ d i r and indirect-effect term δ R U 0 , θ 0 ; δ U i n d are defined below:
δ R U 0 , θ 0 ; δ θ d i r δ k 0 U l k 0 δ U 0 U 0 δ k 0 θ 0 ,
δ R U 0 , θ 0 ; δ U i n d k 0 0 l δ U x δ x l d x θ 0 .
The direct-effect term can be computed at this time, but the indirect-effect term can only be computed after having determined the variational function δ U x . This variational function is the solution of the 1st-LVSS obtained by G-differentiating the NIDE-V form of the heat conduction model, namely Equations (74) and (75). Applying the definition of the G-differential to these equations yields the following NIDE-V form for the 1st-LVSS to be satisfied by the variational function δ U x :
d d x δ U x + b 2 θ 0 x δ U y d y = 2 b δ b 0 x U y d y ; 0 < x < l ;
δ U 0 = δ U 0 θ = 1 + β T 0 δ T 0 + T 0 2 2 δ β
The above 1st-LVSS for the variational function δ U x is to be satisfied at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation. The 1st-LVSS would need to be solved repeatedly to obtain the variational function δ U x for every parameter variation of interest. Repeatedly solving the 1st-LVSS can be avoided by expressing the indirect effect term defined in Equation (105) in terms of an adjoint function to be constructed by following the general procedure outlined in Section 2, as follows:
  • Use Equation (87) to construct the inner product of Equation (106) with a yet specified function φ 1 x to obtain the following relation:
    0 l φ 1 x d d x δ U x + b 2 θ 0 x δ U y d y d x = 2 δ b b θ 0 l φ 1 x 0 x U y d y d x .
  • Integrate by parts the terms on the left-side of Equation (108) to obtain the following relation:
    0 l φ 1 x d d x δ U x + b 2 θ 0 x δ U y d y d x = φ 1 l δ U l φ 1 0 δ U 0 + 0 l δ U x d φ 1 x d x + b 2 θ x l φ 1 y d y d x . .
  • Require the last term on the right-side of Equation (109) to represent the indirect-effect term defined in Equation (105), by requiring the function φ 1 x to satisfy the following relation:
    d φ 1 x d x + b 2 θ x l φ 1 y d y = k 0 δ x l .
  • In Equation (109), use the boundary condition provided in Equation (107) and eliminate the unknown remaining boundary term by imposing the following boundary condition of the function φ 1 x :
    φ 1 l = 0 .
    The NIDE-V net comprising Equations (110) and (111) constitutes the 1st-Level Adjoint Sensitivity System (1st-LASS) for the 1st-level adjoint function φ 1 x . Note that this 1st-LASS is linear in φ 1 x and is independent of parameter variations, so it needs to be solved only once (as opposed to many times, like in the case of the 1st-LVSS) to obtain the function φ 1 x .
  • Use Equations (107), (110), and (111) in Equation (105) to obtain the following expression for the indirect-effect term:
    δ R U 0 , θ 0 ; δ U i n d = δ U 0 φ 1 0 2 δ b b θ 0 l φ 1 x 0 x U y d y d x .
  • Adding the expression obtained in Equation (112) with the expression for the direct-effect term defined in Equation (104) yields the following expression for the G-differential δ R U 0 , θ 0 ; δ U , δ θ :
    δ R U 0 , θ 0 ; δ U , δ θ = δ k 0 U l k 0 δ U 0 U 0 δ k 0 + δ U 0 φ 1 0 2 δ b b θ 0 l φ 1 x 0 x U y d y d x .
    The above expression is to be satisfied at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation.
  • Inserting into Equation (113) the result obtained in Equation (98) and identifying in Equation (113) the expressions that multiply the respective parameter variations yields the following expressions for the first-order sensitivities of the decoder response:
    R U 0 = φ 1 0 k 0 ;
    R Q = 1 k 0 0 l φ 1 x 0 x U y d y d x ;
    R k 0 = Q k 0 2 0 l φ 1 x 0 x U y d y d x + U l U 0 .
The sensitivities of the decoder response with respect to the model parameters T 0 and β are obtained by using the “chain rule” of differentiation of the expression provided in Equation (114) in conjunction with the definition of the feature function U 0 θ provided in Equation (73) to obtain the following expressions:
R T 0 = R U 0 U 0 T 0 = φ 1 0 k 0 1 + β T 0 ;
R β = R U 0 U 0 β = φ 1 0 k 0 T 0 2 2 ;
The sensitivities obtained in Equations (114)–(118) can be computed after having determined the 1st-level adjoint sensitivity function φ 1 x . Solving the 1st-LASS comprising Equations (110) and (111) yields the following expression for the 1st-level adjoint sensitivity function φ 1 x :
φ 1 x = k 0 1 H x l cos l x b θ .
Using in Equations (114)–(118) the expression for φ 1 x obtained in Equation (119) and the expression for U x obtained in Equation (77) yields the following explicit expressions for the first-order sensitivities of the decoder response:
R U 0 = k 0 cos l b θ 1 ;
R Q = U 0 l 2 k 0 Q sin l b θ ;
R k 0 = U 0 l 2 Q k 0 sin l b θ + U 0 cos b θ l 1 .
R T 0 = R U 0 U 0 T 0 = k 0 cos l b θ 1 1 + β T 0 ;
R β = R U 0 U 0 β = k 0 cos l b θ 1 T 0 2 2 ;
Comparing the derivations presented in this subsection (which yielded the explicit exact expressions shown in Equations (120)–(124) for all of the first-order sensitivities of the decoder response with respect to the model parameters) with the derivations presented in Section 3.1.2 indicates that the derivations/computations in the Kirchoff-transformed framework (in which the heat conduction equation and the decoder response become linear functions of the dependent variable U x ) are considerably simpler than in the original temperature-framework, in which both the heat conduction equation and the decoder response are nonlinear functions of the dependent variable T x .

3.2. Application of the 1st-FASAM-NODE to Compute 1st-Order Sensitivities of the NODE Representation of the Illustrative Heat Conduction Model

The 1st-FASAM-NODE methodology [30] will be applied, in this subsection, to the paradigm heat conduction model in order to illustrate the path to obtaining the exact expressions of the first-order sensitivities of the decoder response with respect to the model parameters and to subsequently compute these sensitivities most efficiently, requiring a single large-scale computation for solving the corresponding 1st-Level Adjoint Sensitivity System (1st-LASS). The illustrative mathematical derivations will be performed in the temperature-framework (in Section 3.2.1) and also in the Kirchhoff-transformed framework (in Section 3.2.2) in order to illustrate the advantages of using linearizing transformations, such as Kirchhoff’s transformation, whenever possible. Finally, the application of the 1st-FASAM-NODE methodology will be compared to the application of the 1st-FASAM-NIDE-V methodology, which was presented in Section 3.1

3.2.1. Representation in the NODE Temperature- Framework

In the NODE temperature- framework, the first-order G-differential of the response is as defined in Equation (80) but the variational function δ T x is the solution of the 1st-LVSS obtained by G-differentiating the NODE-form provided in Equations (66) and (67). Applying the definition of the G-differential to these equations yields the following 1st-LVSS for the variational function δ T x :
L 2 T ; θ δ T x d d x β d T x d x δ T x + 1 + β T x d d x δ T x + Q 1 + β T x δ T x = δ k 0 d d x 1 + β T x d T x d x δ β k 0 d d x T x d T x d x δ Q T x + β 2 T 2 x δ β Q 2 T 2 x ;
δ T 0 = δ T 0 ; d d x δ T x x = 0 = 0 .
The above 1st-LVSS for the variational function δ T x is to be satisfied at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation. It is important to note that the operator L 2 T ; θ δ T x acts linearly on δ T x . The 1st-LVSS would need to be solved repeatedly to obtain the variational function δ T x for every parameter variation of interest. Repeatedly solving the 1st-LVSS can be avoided by expressing the indirect effect term defined in Equation (82) in terms of a “1st-level adjoint sensitivity function” that will be the solution of a 1st-LASS to be constructed next by following the general procedure outlined in Section 2. The construction of this 1st-LASS will be accomplished by implementing the following sequence of steps:
  • Use Equation (87) to construct the inner product of Equation (125) with a yet specified function χ 1 x to obtain the following relation:
    0 l χ 1 x L 2 T ; θ δ T x = 0 l χ 1 x d d x β d T x d x δ T x + 1 + β T x d d x δ T x d x + 0 l χ 1 x Q 1 + β T x δ T x d x = p 1 T ; η 1 ; θ ; δ θ ,
    where:
    p 1 T ; η 1 ; θ ; δ θ δ k 0 0 l d x χ 1 x d d x 1 + β T x d T x d x δ β k 0 0 l d x χ 1 x d d x T x d T x d x δ Q 0 l d x χ 1 x T x + β 2 T 2 x δ β 0 l d x χ 1 x Q 2 T 2 x .
  • Transfer the differentiations on δ T x in Equation (127) to differentiations on χ 1 x by constructing the formal adjoint operator L 2 * T ; θ satisfying the relation
    0 l χ 1 x L 2 T ; θ δ T x d x = 0 l δ T x L 2 * T ; θ χ 1 x d x + P 2 T ; θ ; δ T ; χ 1 ,
    where P 2 T ; θ ; δ T ; χ 1 denotes the bilinear concomitant comprising boundary terms. The relationship shown in Equation (129) is derived by integrating by parts the first and the second terms on the left-side of Equation (127) so as to transfer the differentiations on δ T x to differentiations on χ 1 x , to obtain the following relations:
    0 l χ 1 x d d x β d T x d x δ T x d x = χ 1 x β d T x d x δ T x x = l χ 1 x β d T x d x δ T x x = 0 β 0 l δ T x d T x d x d χ 1 x d x d x ;
    0 l χ 1 x d d x 1 + β T x d d x δ T x d x = χ 1 x 1 + β T x d d x δ T x x = l χ 1 x 1 + β T x d d x δ T x x = 0 1 + β T x d χ 1 x d x δ T x x = l + 1 + β T x d χ 1 x d x δ T x x = 0 + 0 l δ T x d d x 1 + β T x d χ 1 x d x d x .
  • Insert the results obtained in Equations (130) and (131) into the left-side Equation (127) and collect the terms that multiply the function δ T x to obtain the relation below:
    0 l δ T x L 2 * T ; θ χ 1 d x = p 1 T ; χ 1 ; θ ; δ θ P 2 T ; χ 1 ; θ ; δ θ ,
    where the operator L 2 * T ; θ has the expression shown below:
    L 2 * T ; θ χ 1 β d T x d x d χ 1 x d x + d d x 1 + β T x d χ 1 x d x .
    and where the bilinear concomitant denoted as P 2 T ; θ ; δ T ; χ 1 in Equation (132) is defined as follows:
    P 2 T ; θ ; δ T ; χ 1 χ 1 x β d T x d x δ T x x = l χ 1 x β d T x d x δ T x x = 0 + χ 1 x 1 + β T x d d x δ T x x = l χ 1 x 1 + β T x d d x δ T x x = 0 1 + β T x d χ 1 x d x δ T x x = l + 1 + β T x d χ 1 x d x δ T x x = 0 .
  • Require the term on the left-side of Equation (132) to represent the indirect-effect term defined in Equation (82), by requiring the function χ 1 x to satisfy the following NIDE-V system:
    L 2 * T ; θ χ 1 = k 0 β T l δ x l .
  • Eliminate the unknown boundary terms in Equation (132) by imposing the following boundary condition of the function χ 1 x :
    χ 1 l = 0 ; d χ 1 x d x x = l = 0 .
    The NIDE-V system comprising Equations (135) and (136) constitutes the 1st-Level Adjoint Sensitivity System (1st-LASS) for the 1st-level adjoint sensitivity function χ 1 x .
  • Use Equations (67), (126), and (134)–(136) in Equation (132) to obtain the following expression for the indirect-effect term:
    δ R T 0 , θ 0 ; δ T i n d = p 1 T ; η 1 ; θ ; δ θ δ T 0 1 + β T 0 d χ 1 x d x x = 0 .
  • Adding the expression obtained in Equation (137) with the expression for the direct-effect term defined in Equation (81) and using the definition of the quantity provided in Equation (128) yields the following expression for the G-differential δ R T 0 , θ 0 ; δ T , δ θ :
    δ R T 0 , θ 0 ; δ T , δ θ = δ k 0 0 l 1 + β T x d T x d x + δ β k 0 0 l T x d T x d x d x δ T 0 T 0 k 0 β δ k 0 0 l d x χ 1 x d d x 1 + β T x d T x d x δ β k 0 0 l d x χ 1 x d d x T x d T x d x δ Q 0 l d x χ 1 x T x + β 2 T 2 x δ β 0 l d x χ 1 x Q 2 T 2 x δ T 0 1 + β T 0 d χ 1 x d x x = 0 .
    The above expression is to be satisfied at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation.
  • Inserting the expression provided in Equation (98) into Equation (138) and collecting the terms that multiply the respective parameter variations yields the following expressions for the first-order sensitivities of the decoder response with respect to the model parameters:
    R T 0 = T 0 k 0 β 1 + β T 0 d χ 1 x d x x = 0 ;
    R β = k 0 0 l T x d T x d x d x k 0 0 l d x χ 1 x d d x T x d T x d x 0 l d x χ 1 x Q 2 T 2 x ;
    R k 0 = 0 l 1 + β T x d T x d x 0 l d x χ 1 x d d x 1 + β T x d T x d x ;
    R Q = 0 l d x χ 1 x T x + β 2 T 2 x .
The expressions presented in Equations (139)–(142) are to be evaluated at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation. These expressions can be evaluated after having obtained the 1st-level adjoint sensitivity function χ 1 x by solving, a single time, the 1st-LASS represented by the NIDE-V net comprising Equations (135) and (136). These computations will not be performed explicitly at this stage because the closed-form expressions of the sensitivities of the decoder response with respect to the model parameters will be obtained by much simpler algebraic manipulations in the NODE Kirchhoff-transformed framework, as will be presented in Section 3.2.2, below.

3.2.2. Representation in the NODE Kirchhoff-Transformed Framework

In the NODE Kirchhoff-transformed framework, the first-order G-differential of the response is as defined in Equations (103) but the variational function δ U x is the solution of the 1st-LVSS obtained by G-differentiating the NODE-system provided in Equations (72) and (73). Applying the definition of the G-differential to these equations yields the following 1st-LVSS for the variational function δ U x :
d 2 d x 2 δ U x + b 2 θ δ U x = 2 δ b b θ U x ; 0 < x < l ;
δ U 0 = δ U 0 ; d d x δ U x x = 0 = 0 .
The above 1st-LVSS for the variational function δ U x is to be satisfied at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation. Since the variational function δ U x depends on parameter variations, its repeated computation by solving the 1st-LVSS for every parameter variation of interest can be avoided by expressing the indirect effect term defined in Equation (105) in terms of an adjoint function to be constructed by following the general procedure outlined in Section 2, as follows:
  • Use Equation (87) to construct the inner product of Equation (143) with a yet specified function ψ 1 x to obtain the following relation:
    0 l ψ 1 x d 2 d x 2 δ U x + b 2 θ δ U x d x = 2 δ b b θ 0 l ψ 1 x U x d x .
  • Integrate by parts twice the first term on the left-side of Equation (145) to obtain the following relation:
    0 l ψ 1 x d 2 d x 2 δ U x + b 2 δ U x d x = ψ 1 l d d x δ U x x = l ψ 1 0 d d x δ U x x = 0 δ U l d ψ 1 x d x x = l + δ U 0 d ψ 1 x d x x = 0 + 0 l δ U x d 2 ψ 1 x d x 2 + b 2 ψ 1 x d x .
  • Require the last term on the right-side of Equation (146) to represent the indirect-effect term defined in Equation (105) by requiring that the function ψ 1 x satisfy the following relation:
    d 2 ψ 1 x d x 2 + b 2 θ ψ 1 x = k 0 δ x l .
  • Use in Equation (146) the boundary conditions provided in Equation (144) and eliminate the unknown remaining boundary terms by imposing the following boundary conditions of the function ψ 1 x :
    ψ 1 l = 0 ; d ψ 1 x d x x = l = 0 .
    The NODE-type system comprising Equations (147) and (148) constitutes the 1st-Level Adjoint Sensitivity System (1st-LASS) for the 1st-level adjoint sensitivity function ψ 1 x .
  • Use Equations (144), (145), (147), and (148) in Equation (146) to obtain the following expression for the indirect-effect term:
    δ R T 0 , θ 0 ; δ T i n d = 2 b δ b 0 l ψ 1 x U x d x δ U 0 d ψ 1 x d x x = 0 .
  • Add the expression obtained in Equation (149) with the expression for the direct-effect term defined in Equation (104) to obtain the following expression for the G-differential:
    δ R T 0 , θ 0 ; δ T , δ θ = 2 b δ b 0 l ψ 1 x U x d x δ U 0 d ψ 1 x d x x = 0 + δ k 0 U l U 0 k 0 δ U 0 θ 0 .
  • Inserting into Equation (150) the result obtained in Equation (98) and identifying in Equation (150) the expressions that multiply the respective parameter variations yields the following expressions for the first-order sensitivities of the decoder response:
    R U 0 = k 0 d ψ 1 x d x x = 0 ;
    R Q = b θ k 0 Q 0 l ψ 1 x U x d x ;
    R k 0 = b θ k 0 Q k 0 0 l ψ 1 x U x d x + U l U 0 .
The expressions presented in Equations (151)–(153) are to be evaluated at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation. These expressions can be evaluated after having obtained the 1st-level adjoint sensitivity function ψ 1 x by solving, a single time, the 1st-LASS represented by the NODE-net comprising Equations (147) and (148). Solving this 1st-LASS yields the following closed-form expression for the 1st-level adjoint sensitivity function ψ 1 x :
ψ 1 x = k 0 b θ 1 H x l sin l x b θ .
Using in Equations (151)–(153) the expression for ψ 1 x obtained in Equation (154) and the expression for U x obtained in Equation (77) yields the following expressions for the first-order sensitivities of the decoder response:
R U 0 = k 0 cos l b θ 1 ;
R Q = U 0 l 2 k 0 Q sin l b θ ;
R k 0 = U 0 l 2 Q k 0 sin l b θ + U 0 cos b θ l 1 .

3.3. Comparative Discussion: Application of the 1st-FASAM-NIDE-V Versus the 1st-FASAM-NODE

The developments presented in Section 3.1 and Section 3.2 indicate the following conclusions regarding the comparison of applying the 1st-FASAM-NIDE-V versus the 1st-FASAM-NODE for the nonlinear paradigm heat conduction model:
(i)
The illustrative nonlinear heat conduction model can be equivalently framed either as a nonlinear NODE-net in the original temperature-framework or as a linear NIDE-V net in the Kirchhoff-transformed framework. The solutions for the corresponding dependent variables, i.e., T x in the temperature-framework and U x in the Kirchhoff-transformed framework, are equivalent to each other, as expected.
(ii)
The 1st-Level Variational Sensitivity System (1st-LVSS) in the Kirchhoff-transformed framework is also equivalent to the 1st-LVSS in the temperature-framework.
(iii)
On the other hand, the 1st-Level Adjoint Sensitivity System (1st-LASS) in the Kirchhoff-transformed framework is not equivalent to the 1st-LASS in the temperature-framework. Consequently, their respective solutions (1st-level adjoint functions) are not equivalent to each other. For example, comparing the solution φ 1 x provided in Equation (119) for the 1st-LASS in the NIDE-V Kirchhoff-transformed framework (cf., Equations (110) and (111)) with the solution ψ 1 x provided in Equation (154) for the 1st-LASS in the NODE Kirchhoff-transformed framework (cf., Equations (147) and (148)) reveals the following relations:
d ψ 1 x d x = φ 1 x .
(iv)
The mathematical derivations/computations in the Kirchhoff-transformed framework, in which the model is linear in the Kirchhoff-transformed dependent variable U x , are simpler (as would be intuitively expected) than the mathematical derivations/computations in the temperature-framework, in which the model is nonlinear in dependent variable T x .

4. Second-Order Features Adjoint Sensitivity Analysis Methodology for Neural Integral Equations of Volterra Type (2nd-FASAM-NIDE-V)

The second-order sensitivities of the response R h ; f θ defined in Equation (5) are obtained by computing the first-order sensitivities of the expression R i 1 h ; f ; a 1 defined in Equation (31), which represents the ith-first-order sensitivity of the response R h ; f θ with respect to the component f i θ of the feature function f θ f 1 θ , , f T F θ . In other words, the second-order sensitivities of R h ; f θ will be computed by conceptually using their basic definitions as being the “first-order sensitivities of the first-order sensitivities” starting from the G-differential of R i 1 h ; f ; a 1 , which is obtained by definition as follows:
f o r i = 1 , , T F : δ R i 1 h 0 ; f 0 ; a 1 , 0 ; v 1 ; δ a 1 ; δ f d d ε t 0 t f D i 1 h 0 t + ε v 1 t ; f 0 + ε δ f ; a 1 , 0 + ε δ a 1 ; t d t ε = 0 = δ R i 1 h 0 ; f 0 ; a 1 , 0 ; δ f d i r + R i 1 h 0 ; f 0 ; a 1 , 0 ; v 1 ; δ a 1 i n d ,
where the “direct effect term” arises directly from variations δ f and is defined as follows:
δ R i 1 h 0 ; f 0 ; a 1 , 0 ; δ f d i r j = 1 T F t 0 t f d t D i 1 h t ; a 1 ; f θ ; t f j δ f j h 0 ; a 1 , 0 ; f 0 ,
and where the “indirect effect term” arises indirectly, through the variations v 1 t v 1 1 t , , v T H 1 t and δ a 1 δ a 1 t , , δ a T H t in the forward and 1st-level adjoint sensitivity functions h t and a 1 t , respectively, and is defined below:
δ R i 1 h 0 ; f 0 ; v 1 ; δ a 1 i n d t 0 t f d τ D i 1 h t ; a 1 ; f θ ; t h v 1 t h 0 ; a 1 , 0 ; f 0 + t 0 t f d τ D i 1 h t ; a 1 ; f θ ; t a 1 δ a 1 t h 0 ; a 1 , 0 ; f 0 .
The direct-effect term δ R i 1 h 0 ; f 0 ; a 1 , 0 ; δ f d i r can be computed immediately while the indirect-effect term δ R i 1 h 0 ; f 0 ; v 1 ; δ a 1 i n d can be computed only after having determined the vectors v 1 x and δ a 1 x . The vector v 1 x is the solution of the 1st-LVSS defined by Equations (16) and (20). On the other hand, the vector δ a 1 x is the solution of the G-differentiated 1st-LASS comprising Equations (27) and (29), which has the formal expression provided below:
V 21 2 v 1 t + V 22 2 δ a 1 t θ 0 = Q 2 2 h ; f ; a 1 δ f θ 0 ,
B j * h ; a 1 ; f ; t h v 1 t + B j * h ; a 1 ; f ; t a 1 δ a 1 t + B j * h ; a 1 ; f ; t f δ f θ 0 = 0 , a t t = t f   a n d / o r t = t 0 ; j = 1 , , B C .
where:
V 21 2 L * h ; f a 1 h 2 D h t ; f θ ; t h h ;
V 22 2 u ; α L * h ; f a 1 a 1 ;
Q 2 2 h ; f ; a 1 δ f 2 D h t ; f θ ; t f h t δ f L * h ; f a 1 f δ f .
Note that Equations (162) and (163) are coupled to the 1st-LVSS through the function v 1 t . Therefore, the functions v 1 t and δ a 1 t are determined by solving the “2nd-Level Variational Sensitivity System (2nd-LVSS)” obtained by concatenating the 1st-LVSS to Equations (162) and (163). This 2nd-LVSS can be written formally in operator form as follows:
V 2 v 2 t = Q V 2 h ; f ; a 1 ; δ f ,
B V 2 h ; a 1 ; f ; v 1 ; δ a 1 ; δ f = 0 , a t t = t f   a n d / o r t = t 0 ; j = 1 , , B C . .
where the superscript “(2)” indicates “2nd-level”, where Equation (168) represents the concatenation of all boundary conditions comprised in Equations (16) and (163), and where the matrices and vectors appearing in Equation (167) are defined as follows:
V 2 L 0 V 21 2 V 22 2 ; v 2 t v 1 t δ a 1 t .
Q V 2 h ; f ; a 1 ; δ f Q 1 h ; f δ f Q 2 2 h ; f ; a 1 δ f ;
The relations in Equations (167) and (168) are to be satisfied at the nominal functions and parameter values, θ 0 , but this fact has not been indicated explicitly in order to simplify the notation.
Solving the 2nd-LVSS requires T F 2 large-scale computations, which are unrealistic to perform for large-scale systems comprising many parameters. The need for solving the 2nd-LVSS is alleviated by deriving an alternative expression for the indirect-effect term defined in Equation (161), in which the function v 2 t is replaced by a 2nd-level adjoint function that will not depend on the variations in the model parameter and state functions. This 2nd-level adjoint function will satisfy a 2nd-Level Adjoint Sensitivity System (2nd-LASS), which will be constructed by using the 2nd-LVSS as starting point and following the same principles as underlying the 2nd-FASAM-N [28], in a Hilbert space that will be denoted as H 2 Ω t , Ω t t t 0 , t f , and which comprises as elements block-vectors of the same form as v 2 t . Thus, a generic vector in H 2 Ω t , denoted as ψ ( 2 ) t ψ 1 ( 2 ) t , ψ 2 ( 2 ) t H 2 Ω t , comprises two components of the form ψ 1 ( 2 ) t ψ 1 , 1 2 t , , ψ 1 , T H 2 t H 1 Ω t , ψ 2 ( 2 ) t ψ 2 , 1 2 t , , ψ 2 , T H 2 t H 1 Ω t , each of which are T H -dimensional column vectors; hence, ψ ( 2 ) t is a 2 × T H -dimensional column vector. The inner product of two vectors ψ ( 2 ) t ψ 1 ( 2 ) t , ψ 2 ( 2 ) t H 2 Ω t and φ ( 2 ) t φ 1 ( 2 ) t , φ 2 ( 2 ) t H 2 Ω t in the Hilbert space H 2 Ω t will be denoted as ψ ( 2 ) t , φ ( 2 ) t 2 and defined below:
ψ ( 2 ) t , φ ( 2 ) t 2 i = 1 2 ψ i ( 2 ) t , φ i ( 2 ) t 1 = i = 1 2 j = 1 T H t 0 t f ψ i , j 2 t φ i , j 2 t d t .
Using the definition of the inner product defined in Equation (171), construct the inner product of Equation (167) with a vector a ( 2 ) t a 1 ( 2 ) t , a 2 ( 2 ) t H 2 Ω t , with a 1 ( 2 ) t a 1 , 1 2 t , , a 1 , T H 2 t H 1 Ω t , a 2 ( 2 ) t a 2 , 1 2 t , , a 2 , T H 2 t H 1 Ω t , to obtain the following relation:
a ( 2 ) t , V 2 v 2 t 2 = a ( 2 ) t , Q V 2 h ; f ; a 1 ; δ f 2
The inner product on the left-side of Equation (172) is now further transformed by using the definition of the adjoint operator to obtain the following relation:
a ( 2 ) t , V 2 v 2 t 2 = v 2 t , A 2 a ( 2 ) t 2 + P 2 v 2 ; a ( 2 ) t ; f ,
where P 2 v 2 ; a ( 2 ) t ; f denotes the corresponding bilinear concomitant on the domain’s boundary, evaluated at the nominal values for the parameters and respective state functions and where the adjoint operator A 2 is defined as follows:
A 2 V 2 * = L * V 21 2 * 0 V 22 2 * .
The definition domain of the adjoint (matrix-valued) operator A 2 is specified by requiring the function a ( 2 ) t a 1 ( 2 ) t , a 2 ( 2 ) t to satisfy adjoint boundary/initial conditions that will be formally denoted in operator form as follows:
B A 2 v 2 t ; a ( 2 ) t ; f = 0 , t Ω t .
The 2nd-level adjoint boundary/initial conditions represented by Equation (175) are determined by requiring that: (a) they must be independent of unknown values of v 2 t , and (b) the substitution of the boundary and/or initial conditions represented by Equations (168) and (175) into the expression of P 2 v 2 ; a ( 2 ) t ; f must cause all terms containing unknown values of v 2 t to vanish.
Implementing the 2nd-level (forward and adjoint) boundary/initial conditions, namely Equations (168) and (175), into Equation (173) will transform the later into the following form:
v 2 t , A 2 a ( 2 ) t 2 = a ( 2 ) t , V 2 v 2 t 2 P ^ 2 v 2 ; a ( 2 ) t ; f ; δ f ,
where P ^ 2 v 2 ; a ( 2 ) t ; f ; δ f denotes residual boundary terms that may not have vanished after having implemented the boundary/initial conditions represented by Equations (168) and (175). The right-side of Equation (172) is now used to replace the vector V 2 v 2 t in the first term on the right-side of Equation (176), thereby obtaining the following relation:
v 2 t , A 2 a ( 2 ) t 2 = a ( 2 ) t , Q V 2 h ; f ; a 1 ; δ f 2 P ^ 2 v 2 ; a ( 2 ) t ; f ; δ f .
The definition of the 2nd-level adjoint function a ( 2 ) t a 1 ( 2 ) t , a 2 ( 2 ) t is now completed by requiring that the left-side of Equation (177) be the same as the “indirect-effect term” δ R i 1 h 0 ; f 0 ; v 1 ; δ a 1 i n d defined by Equation (161), for each index i = 1 , , T F . Recall that the index i = 1 , , T F corresponds to the “ith” first-order sensitivity R i 1 h ; f ; a 1 . Hence, there will be a total of T F 2nd-level adjoint sensitivity functions of the form a ( 2 ) i ; t a 1 ( 2 ) i ; t , a 2 ( 2 ) i ; t , i = 1 , , T F , each such adjoint function corresponding to a specific (i-dependent) indirect-effect term. The left-side of Equation (177) will be identical to the right-side of Equation (161) by requiring that the following relation be satisfied by the 2nd-level adjoint functions (block-vectors) a ( 2 ) i ; t a 1 ( 2 ) i ; t , a 2 ( 2 ) i ; t , for each i = 1 , , T F :
A 2 a ( 2 ) i ; t = D i 1 h t ; a 1 ; f θ ; t / u D i 1 h t ; a 1 ; f θ ; t / a 1 ; i = 1 , , T F .
The boundary conditions to be satisfied by each of the 2nd-level adjoint functions a ( 2 ) i ; t a 1 ( 2 ) i ; t , a 2 ( 2 ) i ; t are as represented by Equation (175), namely:
B A 2 v 2 t ; a ( 2 ) i ; t ; f = 0 , t Ω t .
The system of equations represented by Equations (178) and (179) will be called the 2nd-Level Adjoint Sensitivity System (2nd-LASS); its solution, a ( 2 ) i ; t a 1 ( 2 ) i ; t , a 2 ( 2 ) i ; t , will be called the 2nd-level adjoint function. The 2nd-LASS is independent of parameter variations. Furthermore, the 2 × T H 2 -dimensional matrix A 2 is independent of the index i = 1 , , T F . Only the source-term on the right-side of Equation (178) depends on the index i = 1 , , T F . Therefore, the same solver can be used to invert the matrix A 2 in order to solve numerically the 2nd-LASS that corresponds to each first-order sensitivity R i 1 h ; f ; a 1 , i = 1 , , T F . Computationally, it would be efficient to store, if possible, the inverse matrix A 2 1 , in order to multiply directly this inverse matrix with the source term corresponding to each index i = 1 , , T F , in order to obtain the corresponding 2nd-level adjoint function a ( 2 ) i ; t a 1 ( 2 ) i ; t , a 2 ( 2 ) i ; t .
Using Equations (176)–(179) in Equation (161) yields the following expression for the indirect-effect term δ R i 1 h 0 ; f 0 ; v 1 ; δ a 1 i n d , for each i = 1 , , T F :
δ R i 1 h 0 ; f 0 ; v 1 ; δ a 1 i n d = a ( 2 ) i ; t , Q V 2 h ; f ; a 1 ; δ f 2 P ^ 2 v 2 ; a ( 2 ) i ; t ; f ; δ f .
Adding the expression obtained in Equation (180) for the indirect-effect term together with the expression for the direct-effect term defined in Equation (160) yields the following expression for the total differential defined by Equation (159), for i = 1 , , T F :
δ R i 1 h 0 ; f 0 ; a 1 , 0 ; v 1 ; δ a 1 ; δ f = t 0 t f d t D i 1 h t ; a 1 ; f θ ; t f θ δ f θ + a ( 2 ) i ; t , Q V 2 h ; f ; a 1 ; δ f 2 P ^ 2 v 2 ; a ( 2 ) i ; t ; f ; δ f j = 1 T F 2 R u x ; α f j f i δ f j .
The second-order sensitivities 2 R / f j f i , i , j = 1 , , T F are obtained by identifying in Equation (181) the expressions that multiply the variations δ f j , which yields the following expressions after detailing the second-term on the right side of Equation (181):
2 R u x ; α f j f i = t 0 t f d t D i 1 h t ; a 1 ; f θ ; t f j P ^ 2 v 2 ; a ( 2 ) i ; t ; f ; δ f f j + m = 1 T H t 0 t f a 1 , m 2 t q m , j 1 h ; f d t + t 0 t f a 2 ( 2 ) t · 2 D h t ; f θ ; t f j h t L * h ; f a 1 f j d t .
As Equations (178) and (179) indicate, solving the 2nd-LASS once provides the 2nd-level adjoint function a ( 2 ) i ; t for each index i = 1 , , T P . Thus, the exact computation of all of the partial second-order sensitivities, 2 R α , u / f j f i   i , j = 1 , , T P , requires solving the 2nd-LASS at most T F -times, amounting to at most T F “large-scale” computations for, rather than at least O T W 2 large-scale computations as would be required by forward methods. It is important to note that if the 2nd-LASS is solved T F -times, the 2nd-order mixed sensitivities 2 R α , u / f j f i will be computed twice, in two different ways, in terms of two distinct 2nd-level adjoint functions. Consequently, the symmetry property 2 R α , u / f j f i = 2 R α , u / f i f j inherent to the second-order sensitivities provides an intrinsic (numerical) verification that the respective components of the 2nd-level adjoint function a ( 2 ) i ; t , for two distinct values of the index “i”, are computed accurately.
Since the adjoint matrix A 2 is block-diagonal, solving the 2nd-LASS is equivalent to solving two 1st-LASS, with two different source terms. Thus, the “solvers” and the computer program used for solving the 1st-LASS can also be used for solving the 2nd-LASS. The 2nd-LASS was designated as the “second-level” rather than the “second-order” adjoint sensitivity system, since the 2nd-LASS does not involve any explicit 2nd-order G-derivatives of the operators underlying the original system but involves the inversion of the same operators that needed to be inverted for solving the 1st-LASS. The block-triangular structure of the 2nd-LASS enables full flexibility for prioritizing the computation of the 2nd-order sensitivities. The computation of the 2nd-order sensitivities would logically be prioritized based on the relative magnitudes of the 1st-order sensitivities: the largest relative 1st-order response sensitivity should have the highest priority for computing the corresponding 2nd-order mixed sensitivities; then, the second largest relative 1st-order response sensitivity should be considered next, and so on. The unimportant 2nd-order sensitivities can be deliberately neglected while knowing the error incurred by neglecting them. Computing 2nd-order sensitivities that correspond to vanishing 1st-order sensitivities may also be of interest, since vanishing 1st-order sensitivities may indicate critical points of the response in the phase-space of model parameters.

5. Illustrative Application of the 2nd-FASAM-NIDE-V Methodology to Compute Second-Order Sensitivities of the Heat Conduction Model

The application of the 2nd-FASAM-NIDE-V methodology will be illustrated in this section by continuing the analysis of the heat conduction model considered in Section 3. As highlighted while developing the general 2nd-FASAM-NIDE-V methodology in Section 4, the second-order sensitivities are conceptually determined by considering them to be the “first-order sensitivities of the first-order sensitivities.” To underscore the essential concepts and steps underlying the application of the 2nd-FASAM-NIDE-V methodology, the heat conduction model will be considered in the Kirchhoff-transformed framework, since the algebraic manipulation in this framework are considerably less involved than in the temperature-framework.

5.1. Second-Order Sensitivities of the Heat Conduction Model in the NIDE-V Kirchhoff-Transformed Framework

The expressions of the first-order sensitivities to be considered as “responses that give rise to the second-order sensitivities of the decoder response with respect to the feature functions and model parameters” are presented in Equations (115)–(118).

5.1.1. Determining the Second-Order Sensitivities Stemming from R / T 0

The second-order sensitivities stemming from the first-order sensitivity R / T 0 are comprised in the G-differential of the expression provided in Equation (117), which yields, by definition, the following relation:
δ R T 0 1 + β T 0 d d ε 0 l φ 1 , 0 x + ε δ φ 1 x δ x d x ε = 0 + φ 1 0 δ β T 0 + β δ T 0 δ k 0 δ k 0 T 0 0 β 0 δ T 0 k 0 0 β 0 δ β k 0 0 T 0 0 δ R T 0 d i r + δ R T 0 i n d ,
where the direct-effect term δ R / T 0 d i r and the indirect-effect term δ R / T 0 i n d , respectively, are defined below:
δ R / T 0 d i r φ 1 0 δ β T 0 + β δ T 0 δ k 0 δ k 0 T 0 β δ T 0 k 0 β δ β k 0 T 0 ;
δ R / T 0 i n d 1 + β T 0 0 l δ φ 1 x δ x d x .
The above expressions are to be evaluated at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation.
The direct-effect term δ R / T 0 d i r can be computed immediately, while the indirect-effect term δ R / T 0 i n d can be computed only after having determined the function δ φ 1 t , which is the solution of the G-differentiated 1st-LASS comprising Equations (110) and (111). Applying the definition of the G-differential to these equations yields the following NIDE-V system:
d d x δ φ 1 x + b 2 θ x l δ φ 1 y d y = δ k 0 δ x l 2 δ b b θ x l φ 1 y d y ;
δ φ 1 l = 0 .
The NIDE-V system comprising Equations (186) and (187) constitutes the 2nd-LVSS for the 2nd-level variational function δ φ 1 x , which depends on parameter variations. Therefore, the 2nd-LVSS would need to be solved repeatedly to determine δ φ 1 x for all possible parameter variations of interest. The need for solving the 2nd-LVSS is alleviated by deriving an alternative expression for the indirect-effect term defined in Equation (185), in which the function δ φ 1 x is replaced by a 2nd-level adjoint function that is independent of variations in the model parameter and state functions. This 2nd-level adjoint function will satisfy a 2nd-Level Adjoint Sensitivity System (2nd-LASS), which will be constructed by using the 2nd-LVSS as starting point and following the same principles outlined in Section 4. Since the 2nd-LVSS involves a one-component dependent variable (namely: δ φ 1 x ), the corresponding 2nd-level adjoint function will also be a one-component function, which will be denoted a ( 2 ) 1 ; x , using the notation introduced in Section 4. The superscript “(2)” in a ( 2 ) 1 ; x indicates “2nd-level” while the argument “1” indicates that this 2nd-level adjoint function corresponds to the sensitivity R / T 0 , which is the “first” first-order sensitivity to be considered. The Hilbert space appropriate for constructing the 2nd-LASS for the function a ( 2 ) 1 ; x is endowed with the inner product defined in Equation (87). The construction of this 2nd-LASS proceeds as follows:
  • Use Equation (87) to construct the inner product of Equation (186) with a yet unspecified function a ( 2 ) 1 ; x to obtain the following relation:
    0 l a ( 2 ) 1 ; x d d x δ φ 1 x + b 2 θ x l δ φ 1 y d y d x = 0 l a ( 2 ) 1 ; x δ k 0 δ x l 2 δ b b x l φ 1 y d y d x .
  • Integrate by parts the terms on the left-side of Equation (188) to obtain the following relation:
    0 l a ( 2 ) 1 ; x d d x δ φ 1 x + b 2 θ x l δ φ 1 y d y d x = a ( 2 ) 1 ; l δ φ 1 l + a ( 2 ) 1 ; 0 δ φ 1 0 + 0 l δ φ 1 x d d x a ( 2 ) 1 ; x + b 2 θ 0 x a ( 2 ) 1 ; y d y d x .
  • Require the last term on the right-side of Equation (189) to represent the indirect-effect term defined in Equation (185), by requiring the function a ( 2 ) 1 ; x to satisfy the following relation:
    d a ( 2 ) 1 ; x d x + b 2 θ 0 x a ( 2 ) 1 ; y d y = 1 + β T 0 δ x .
  • Use in Equation (189) the boundary condition provided in Equation (187) and eliminate the unknown remaining boundary term by imposing the following boundary condition of the function a ( 2 ) 1 ; x :
    a ( 2 ) 1 ; 0 = 0 .
    The NIDE-V net comprising Equations (190) and (191) constitutes the 2nd-Level Adjoint Sensitivity System (2nd-LASS) for the 2nd-level adjoint function a ( 2 ) 1 ; x . Note that this 2nd-LASS is linear in a ( 2 ) 1 ; x and is independent of parameter variations, so it needs to be solved only once to obtain the function a ( 2 ) 1 ; x .
  • Use Equations (187), (190), and (191) in Equation (189) to obtain the following expression for the indirect-effect term defined in Equation (185):
    δ R / T 0 i n d = 0 l a ( 2 ) 1 ; x δ k 0 δ x l 2 δ b b θ x l φ 1 y d y d x .
  • Adding the expression obtained in Equation (192) with the expression for the direct-effect term defined in Equation (184) yields the following expression for the G-differential δ R / T 0 :
    δ R / T 0 = φ 1 0 δ β T 0 + β δ T 0 δ k 0 δ k 0 T 0 β δ T 0 k 0 β δ β k 0 T 0 + 0 l a ( 2 ) 1 ; x δ k 0 δ x l 2 δ b b θ x l φ 1 y d y d x .
    The above expression is to be satisfied at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation.
  • Inserting into Equation (193) the result obtained in Equation (98) and identifying in Equation (193) the expressions that multiply the respective parameter variations yields the expressions for the following second-order sensitivities of the decoder response:
    2 R T 0 T 0 = β φ 1 0 k 0 = β cos l b θ 1 ;
    2 R β T 0 = T 0 φ 1 0 k 0 = T 0 cos l b θ 1 ;
    2 R Q T 0 = 1 k 0 0 l a ( 2 ) 1 ; x d x x l φ 1 y d y ;
    2 R k 0 T 0 = 1 T 0 β + a ( 2 ) 1 ; l + Q k 0 2 0 l a ( 2 ) 1 ; x x l φ 1 y d y d x .
The sensitivities obtained in Equations (196) and (197) can be computed after having determined the 1st-level adjoint sensitivity function a ( 2 ) 1 ; x . Solving the 2nd-LASS comprising Equations (190) and (191) yields the following expression for the 2nd-level adjoint sensitivity function a ( 2 ) 1 ; x :
a ( 2 ) 1 ; x = 1 + β T 0 H x cos x b θ .
Inserting into Equation (196) and Equation (197), respectively, the expressions obtained in Equations (198) and (119) and performing the respective integrations yields the following results:
2 R Q T 0 = 1 + β T 0 l 2 k 0 Q sin l b θ
2 R k 0 T 0 = 1 + β T 0 cos l b θ 1 + l 2 Q k 0 sin l b θ .

5.1.2. Determining the Second-Order Sensitivities Stemming from R / β

The second-order sensitivities stemming from the first-order sensitivity R / β are comprised in the G-differential of the expression provided in Equation (118), which yields, by definition, the following relation:
δ R β T 0 0 2 2 d d ε 0 l φ 1 , 0 x + ε δ φ 1 x δ x d x ε = 0 + φ 1 0 T 0 0 δ T 0 δ k 0 T 0 0 2 2 δ T 0 k 0 0 T 0 0 δ R β d i r + δ R β i n d ,
where the direct-effect term δ R / β d i r and the indirect-effect term δ R / β i n d , respectively, are defined below:
δ R / β d i r φ 1 0 T 0 δ T 0 δ k 0 T 0 2 2 δ T 0 k 0 T 0 ;
δ R / β i n d T 0 2 2 0 l δ φ 1 x δ x d x .
The above expressions are to be evaluated at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation. The direct-effect term δ R / β d i r can be computed immediately while the indirect-effect term δ R / β i n d can be computed only after having determined the function δ φ 1 x , which is the solution of the 2nd-LVSS represented by Equations (186) and (187). As before, the need for solving this 2nd-LVSS repeatedly, for every parameter variation of interest, is alleviated by deriving an alternative expression for the indirect-effect term defined in Equation (203), in which the function δ φ 1 x is replaced by a 2nd-level adjoint function that is independent of variations in the model parameter and state functions. This 2nd-level adjoint function will also be a one-component function, which will be denoted a ( 2 ) 2 ; x , using the notation introduced in Section 4. The superscript “(2)” in a ( 2 ) 2 ; x indicates “2nd-level” while the argument “2” indicates that this 2nd-level adjoint function corresponds to the sensitivity R / β , which is the “second” first-order sensitivity to be considered. The Hilbert space appropriate for constructing the 2nd-LASS for the function a ( 2 ) 2 ; x is also endowed with the inner product defined in Equation (87). The construction of this 2nd-LASS proceeds along the same steps as outlined in Section 5.1.1, above, as follows:
  • Use Equation (87) to construct the inner product of Equation (186) with a yet unspecified function a ( 2 ) 2 ; x to obtain the following relation:
    0 l a ( 2 ) 2 ; x d d x δ φ 1 x + b 2 θ x l δ φ 1 y d y d x = 0 l a ( 2 ) 2 ; x δ k 0 δ x l 2 δ b b x l φ 1 y d y d x .
  • Integrate by parts the terms on the left-side of Equation (204) to obtain the following relation:
    0 l a ( 2 ) 2 ; x d d x δ φ 1 x + b 2 θ x l δ φ 1 y d y d x = a ( 2 ) 2 ; l δ φ 1 l + a ( 2 ) 2 ; 0 δ φ 1 0 + 0 l δ φ 1 x d d x a ( 2 ) 2 ; x + b 2 θ 0 x a ( 2 ) 2 ; y d y d x .
  • Require the last term on the right-side of Equation (205) to represent the indirect-effect term defined in Equation (203), by requiring the function a ( 2 ) 2 ; x to satisfy the following relation:
    d a ( 2 ) 1 ; x d x + b 2 θ 0 x a ( 2 ) 1 ; y d y = T 0 2 2 δ x .
  • Use in Equation (205) the boundary condition provided in Equation (187) and eliminate the unknown remaining boundary term by imposing the following boundary condition of the function a ( 2 ) 2 ; x :
    a ( 2 ) 2 ; 0 = 0 .
    The NIDE-V net comprising Equations (206) and (207) constitutes the 2nd-Level Adjoint Sensitivity System (2nd-LASS) for the 2nd-level adjoint function a ( 2 ) 2 ; x . Note that this 2nd-LASS is linear in a ( 2 ) 2 ; x and is independent of parameter variations, so it needs to be solved only once to obtain the function a ( 2 ) 2 ; x .
  • Use Equations (187), (206), and (207) in Equation (205) to obtain the following expression for the indirect-effect term defined in Equation (203):
    δ R / β i n d = 0 l a ( 2 ) 2 ; x δ k 0 δ x l 2 δ b b θ x l φ 1 y d y d x .
  • Adding the expression obtained in Equation (208) with the expression for the direct-effect term defined in Equation (202) yields the following expression for the G-differential δ R / β :
    δ R / β = φ 1 0 T 0 δ T 0 δ k 0 T 0 2 2 δ T 0 k 0 T 0 + 0 l a ( 2 ) 2 ; x δ k 0 δ x l 2 δ b b θ x l φ 1 y d y d x .
    The above expression is to be satisfied at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation.
  • Inserting into Equation (209) the result obtained in Equation (98) and identifying in the resulting form of Equation (209) the expressions that multiply the respective parameter variations yields the expressions for the following second-order sensitivities of the decoder response:
    2 R T β 0 = T 0 φ 1 0 k 0 = T 0 cos l b θ 1 ;
    2 R β β = 0 ;
    2 R Q β = 1 k 0 0 l a ( 2 ) 2 ; x d x x l φ 1 y d y ;
    2 R k 0 β = T 0 2 2 + a ( 2 ) 2 ; l + Q k 0 2 0 l a ( 2 ) 2 ; x x l φ 1 y d y d x .
The sensitivities obtained in Equations (196) and (197) can be computed after having determined the 2nd-level adjoint sensitivity function a ( 2 ) 1 ; x . Solving the 2nd-LASS comprising Equations (190) and (191) yields the following expression for the 2nd-level adjoint sensitivity function a ( 2 ) 1 ; x :
a ( 2 ) 2 ; x = T 0 2 2 H x cos x b θ .
Inserting into Equation (212) and (213), respectively, the expressions obtained in Equations (214) and (119), and performing the respective integrations yields the following results:
2 R Q β = l T 0 2 4 k 0 Q sin l b θ ;
2 R k 0 β = T 0 2 2 cos l b θ 1 l 2 Q k 0 sin l b θ .

5.1.3. Determining the Second-Order Sensitivities Stemming from R / Q

The second-order sensitivities stemming from the first-order sensitivity R / Q are comprised in the G-differential of the expression provided in Equation (115), which yields, by definition, the following relation:
δ R Q = d d ε 1 k 0 0 + ε δ k 0 0 l d x φ 1 , 0 x + ε δ φ 1 x 0 x U 0 y + ε δ U y d y ε = 0 = δ R Q d i r + δ R Q i n d ,
where the direct-effect term δ R / Q d i r and the indirect-effect term δ R / Q i n d , respectively, are defined below:
δ R / Q d i r δ k 0 k 0 2 0 l φ 1 x 0 x U y d y d x
δ R / Q i n d 1 k 0 0 l δ φ 1 x 0 x U y d y d x 1 k 0 0 l φ 1 x 0 x δ U y d y d x .
The above expressions are to be evaluated at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation. The direct-effect term δ R / Q d i r can be computed immediately while the indirect-effect term δ R / Q i n d can be computed only after having determined the variational functions δ φ 1 x and δ U x . The variational function δ φ 1 x is the solution of the 2nd-LVSS represented by Equations (186) and (187). On the other hand, the function δ U x is the solution of the Equations (106) and (107). Altogether, these equations constitute the 2nd-LVSS for the two-component 2nd-level variational function v 2 x δ U x , δ φ 1 x , cf. Equation (169). As before, the need for solving this 2nd-LVSS repeatedly, for every parameter variation of interest, is circumvented by deriving an alternative expression for the indirect-effect term defined in Equation (203), in which the function v 2 x δ U x , δ φ 1 x is replaced by a 2nd-level adjoint function that is independent of variations in the model parameter and state functions. This 2nd-level adjoint function will be a two-component function, which will be as denoted a ( 2 ) 3 ; x a 1 ( 2 ) 3 ; x , a 2 ( 2 ) 3 ; x , using the notation introduced in Section 4. The superscript “(2)” in a ( 2 ) 3 ; x indicates “2nd-level” while the argument “3” indicates that this 2nd-level adjoint function corresponds to the sensitivity R / Q , which is the “third” first-order sensitivity to be considered. The Hilbert space appropriate for constructing the 2nd-LASS for the two-component function a ( 2 ) 3 ; x is endowed with the particular form (two-component) of the inner product defined in Equation (171). The construction of this 2nd-LASS proceeds along the same steps as outlined in Section 4, as follows:
  • Use Equation (171) to construct the inner product of Equations (106) and (186) with a yet unspecified function a ( 2 ) 3 ; x to obtain the following relation:
    0 l a 1 ( 2 ) 3 ; x d d x δ U x + b 2 θ 0 x δ U y d y d x + 0 l a 2 ( 2 ) 3 ; x d d x δ φ 1 x + b 2 θ x l δ φ 1 y d y d x = 2 b δ b 0 l a 2 ( 2 ) 3 ; x 0 x U y d y d x + 0 l a 1 ( 2 ) 3 ; x δ k 0 δ x l 2 δ b b x l φ 1 y d y d x .
  • Integrate by parts the terms on the left-side of Equation (220) to obtain the following relation:
    0 l a 1 ( 2 ) 3 ; x d d x δ U x + b 2 θ 0 x δ U y d y d x + 0 l a 2 ( 2 ) 3 ; x d d x δ φ 1 x + b 2 θ x l δ φ 1 y d y d x = a 1 ( 2 ) 3 ; l δ U l a 1 ( 2 ) 3 ; 0 δ U 0 a 2 ( 2 ) 3 ; l δ φ 1 l + a 2 ( 2 ) 3 ; 0 δ φ 1 0 + 0 l δ U x d a 1 ( 2 ) 3 ; x d x + b 2 θ x l a 1 ( 2 ) 3 ; y d y d x . + 0 l δ φ 1 x d a 2 ( 2 ) 3 ; x d x + b 2 θ 0 x a 2 ( 2 ) 3 ; y d y d x .
  • The last two terms on the right-side of Equation (221) will be required to represent the indirect-effect term δ R / Q i n d defined in Equation (219). For this purpose, the second expression in the definition of the indirect-effect term δ R / Q i n d needs to be recast in the following form:
    0 l d x φ 1 x 0 x δ U y d y = 0 x δ U y d y 0 x φ 1 y d y 0 l 0 l δ U x d x 0 x φ 1 y d y = 0 l δ U x d x x l φ 1 y d y .
    Replacing the result obtained in Equation (222) into Equation (219) yields the following expression indirect-effect term δ R / Q i n d :
    δ R / Q i n d 1 k 0 0 l δ φ 1 x 0 x U y d y d x 1 k 0 0 l δ U y d y x l φ 1 y d y .
  • Express the indirect-effect term δ R / Q i n d in Equation (223) in terms of the 2nd-level adjoint sensitivity function a ( 2 ) 3 ; x a 1 ( 2 ) 3 ; x , a 2 ( 2 ) 3 ; x by requiring this function to satisfy the following relations:
    d a 1 ( 2 ) 3 ; x d x + b 2 θ x l a 1 ( 2 ) 3 ; y d y = 1 k 0 x l φ 1 y d y = 1 b sin b l x ;
    d a 2 ( 2 ) 3 ; x d x + b 2 θ 0 x a 2 ( 2 ) 3 ; y d y = 1 k 0 0 x U y d y = U 0 k 0 b sin b x .
  • Use in Equation (221) the boundary conditions provided in Equations (107) and (187), and eliminate the unknown remaining boundary terms by imposing the following boundary conditions of the function a ( 2 ) 3 ; x a 1 ( 2 ) 3 ; x , a 2 ( 2 ) 3 ; x :
    a 1 ( 2 ) 3 ; l = 0 ; a 2 ( 2 ) 3 ; 0 = 0 .
    The NIDE-V net comprising Equations (224) –(226) constitutes the 2nd-Level Adjoint Sensitivity System (2nd-LASS) for the 2nd-level adjoint function a ( 2 ) 3 ; x . Note that this 2nd-LASS is linear in a ( 2 ) 3 ; x and is independent of parameter variations, so it needs to be solved only once to obtain the function a ( 2 ) 3 ; x .
  • Use in Equation (220) the equations underlying the 2nd-LVSS for the function v 2 x δ U x , δ φ 1 x together with the equations underlying the 2nd-LASS for the function a ( 2 ) 3 ; x to obtain the following expression for the indirect-effect term defined in Equation (223):
    δ R / Q i n d = a 1 ( 2 ) 3 ; 0 δ U 0 2 δ b b θ 0 l a 2 ( 2 ) 3 ; x 0 x U y d y d x + 0 l a 1 ( 2 ) 3 ; x δ k 0 δ x l 2 δ b b θ x l φ 1 y d y d x .
  • Adding the expression obtained in Equation (227) with the expression for the direct-effect term defined in Equation (218) yields the following expression for the G-differential δ R / U 0 :
    δ R Q = δ k 0 k 0 2 0 l φ 1 x 0 x U y d y d x + a 1 ( 2 ) 3 ; 0 δ U 0 + δ k 0 a 1 ( 2 ) 3 ; l 2 δ b b θ 0 l a 2 ( 2 ) 3 ; x 0 x U y d y + x l φ 1 y d y d x .
    The above expression is to be satisfied at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation.
  • Inserting into Equation (228) the result obtained in Equation (98) and identifying in the resulting form of Equation (228) the expressions that multiply the respective parameter variations yields the expressions for the following second-order sensitivities of the decoder response:
    2 R T 0 Q = 1 + β T 0 a 1 ( 2 ) 3 ; 0 ;
    2 R β Q = T 0 2 2 a 1 ( 2 ) 3 ; 0 ;
    2 R Q Q = 1 k 0 0 l a 2 ( 2 ) 3 ; x 0 x U y d y d x 1 k 0 0 l a 1 ( 2 ) 3 ; x x l φ 1 y d y d x ;
    2 R k 0 Q = 1 k 0 2 0 l φ 1 x 0 x U y d y d x + a 1 ( 2 ) 3 ; l + Q k 0 2 0 l a 2 ( 2 ) 3 ; x 0 x U y d y d x . + Q k 0 2 0 l a 1 ( 2 ) 3 ; x x l φ 1 y d y d x .
The sensitivities obtained in Equations (229)–(232) can be computed after having determined the 2nd-level adjoint sensitivity function a ( 2 ) 3 ; x . In this particular case, the equations underlying the 2nd-LASS, comprising Equations (224)–(226), for the two components of a ( 2 ) 3 ; x are decoupled and can therefore be solved independently of each other. Solving Equations (224) and (226) yields the following expression for the component a 1 ( 2 ) 3 ; x of the 2nd-level adjoint sensitivity function a ( 2 ) 3 ; x :
a 1 ( 2 ) 3 ; x = x l 2 b θ sin l x b θ = x l 2 k 0 Q sin l x Q k 0 ;
Solving Equations (225) and (226) yields the following expression for the component a 2 ( 2 ) 3 ; x of the 2nd-level adjoint sensitivity function a ( 2 ) 3 ; x :
a 2 ( 2 ) 3 ; x = U 0 2 k 0 b θ x sin x b θ = U 0 2 Q k 0 x sin x Q k 0 .
Inserting into Equations (229)–(232) the expressions of obtained in Equations (233) and (234) together with the previously obtained expressions for the functions U x and φ 1 x , and performing the respective integrations yields the following closed-form expressions for the second-order sensitivities stemming from the first-order sensitivity R / Q :
2 R T 0 Q = 1 + β T 0 l 2 k 0 Q sin l b θ ;
2 R β Q = T 0 2 l 4 k 0 Q sin l b θ ;
2 R Q Q = U 0 l 4 Q k 0 Q sin l b θ l cos l b θ ;
2 R k 0 Q = U 0 l 4 l k 0 cos l b θ 1 Q k 0 sin l b θ .

5.1.4. Determining the Second-Order Sensitivities Stemming from R / k 0

The second-order sensitivities stemming from the first-order sensitivity R / k 0 are comprised in the G-differential of the expression provided in Equation (116), which yields, by definition, the following relation:
δ R k 0 = d d ε Q 0 + ε δ Q k 0 0 + ε δ k 0 2 0 l d x φ 1 , 0 x + ε δ φ 1 x 0 x U 0 y + ε δ U y d y ε = 0   + 0 l U x δ x l d x δ U 0 = δ R k 0 d i r + δ R k 0 i n d ,
where the direct-effect term δ R / k 0 d i r and the indirect-effect term δ R / k 0 i n d , respectively, are defined below:
δ R / k 0 d i r δ Q k 0 2 2 Q δ k 0 k 0 3 0 l φ 1 x 0 x U y d y d x δ U 0 ;
δ R / k 0 i n d Q k 0 2 0 l δ φ 1 x 0 x U y d y d x + Q k 0 2 0 l φ 1 x 0 x δ U y d y d x + 0 l δ U x δ x l d x .
The above expressions are to be evaluated at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation. The direct-effect term δ R / k 0 d i r can be computed immediately while the indirect-effect term δ R / k 0 i n d can be computed only after having determined the two-component variational function v 2 x δ U x , δ φ 1 x , which is the solution of the 2nd-LVSS represented by Equations (106), (107), (186), and (187). As before, the need for solving this 2nd-LVSS repeatedly, for every parameter variation of interest, is circumvented by deriving an alternative expression for the indirect-effect term defined in Equation (241), in which the function v 2 x δ U x , δ φ 1 x is replaced by a 2nd-level adjoint function that is independent of variations in the model parameter and state functions. This 2nd-level adjoint function will be a two-component function, which will be as denoted a ( 2 ) 4 ; x a 1 ( 2 ) 4 ; x , a 2 ( 2 ) 4 ; x , using the notation introduced in Section 4. The superscript “(2)” in a ( 2 ) 4 ; x indicates “2nd-level” while the argument “4” indicates that this 2nd-level adjoint function corresponds to the sensitivity R / k 0 , which is the “fourth” first-order sensitivity to be considered. The Hilbert space appropriate for constructing the 2nd-LASS for the two-component function a ( 2 ) 4 ; x is endowed with the particular form (two-component) of the inner product defined in Equation (171). The construction of this 2nd-LASS proceeds along the same steps as outlined in Section 5.1.3, as follows:
  • Use Equation (171) to construct the inner product of Equations (106) and (186) with a yet unspecified function a ( 2 ) 4 ; x to obtain the following relation:
    0 l a 1 ( 2 ) 4 ; x d d x δ U x + b 2 θ 0 x δ U y d y d x + 0 l a 2 ( 2 ) 4 ; x d d x δ φ 1 x + b 2 θ x l δ φ 1 y d y d x = 2 b δ b 0 l a 2 ( 2 ) 4 ; x 0 x U y d y d x + 0 l a 1 ( 2 ) 4 ; x δ k 0 δ x l 2 δ b b x l φ 1 y d y d x .
  • Integrate by parts the terms on the left-side of Equation (242) to obtain the following relation:
    0 l a 1 ( 2 ) 4 ; x d d x δ U x + b 2 θ 0 x δ U y d y d x + 0 l a 2 ( 2 ) 4 ; x d d x δ φ 1 x + b 2 θ x l δ φ 1 y d y d x = a 1 ( 2 ) 4 ; l δ U l a 1 ( 2 ) 4 ; 0 δ U 0 a 2 ( 2 ) 4 ; l δ φ 1 l + a 2 ( 2 ) 4 ; 0 δ φ 1 0 + 0 l δ U x d a 1 ( 2 ) 4 ; x d x + b 2 θ x l a 1 ( 2 ) 4 ; y d y d x . + 0 l δ φ 1 x d a 2 ( 2 ) 4 ; x d x + b 2 θ 0 x a 2 ( 2 ) 4 ; y d y d x .
  • The last two terms on the right-side of Equation (243) will be required to represent the indirect-effect term δ R / k 0 i n d defined in Equation (241). For this purpose, the second expression in the definition of the indirect-effect term δ R / k 0 i n d needs to be recast into the form presented in Equation (222). Replacing this form into Equation (241) yields the following expression indirect-effect term δ R / Q i n d :
    δ R / k 0 i n d Q k 0 2 0 l δ φ 1 x 0 x U y d y d x + Q k 0 2 0 l δ U x d x x l φ 1 y d y + 0 l δ U x δ x l d x .
  • Express the indirect-effect term δ R / k 0 i n d in Equation (244) in terms of the 2nd-level adjoint sensitivity function a ( 2 ) 4 ; x a 1 ( 2 ) 4 ; x , a 2 ( 2 ) 4 ; x by requiring this function to satisfy the following relations:
    d a 1 ( 2 ) 4 ; x d x + b 2 θ x l a 1 ( 2 ) 4 ; y d y = Q k 0 2 x l φ 1 y d y + δ x l ;
    d a 2 ( 2 ) 4 ; x d x + b 2 θ 0 x a 2 ( 2 ) 4 ; y d y = Q k 0 2 0 x U y d y .
  • Use in Equation (243) the boundary conditions provided in Equations (107) and (187), and eliminate the unknown remaining boundary terms by imposing the following boundary conditions of the function a ( 2 ) 4 ; x a 1 ( 2 ) 4 ; x , a 2 ( 2 ) 4 ; x :
    a 1 ( 2 ) 4 ; l = 0 ; a 2 ( 2 ) 4 ; 0 = 0 .
    The NIDE-V system comprising Equations (245)–(247) constitutes the 2nd-Level Adjoint Sensitivity System (2nd-LASS) for the 2nd-level adjoint function a ( 2 ) 4 ; x . This 2nd-LASS is linear in a ( 2 ) 4 ; x and is independent of parameter variations, so it needs to be solved only once to obtain the function a ( 2 ) 4 ; x .
  • Use in Equation (243) the equations underlying the 2nd-LVSS for the function v 2 x δ U x , δ φ 1 x together with the equations underlying the 2nd-LASS for the function a ( 2 ) 4 ; x to obtain the following expression for the indirect-effect term defined in Equation (244):
    δ R / k 0 i n d = a 1 ( 2 ) 4 ; 0 δ U 0 2 b δ b 0 l a 2 ( 2 ) 4 ; x 0 x U y d y d x + 0 l a 1 ( 2 ) 4 ; x δ k 0 δ x l 2 δ b b θ x l φ 1 y d y d x .
  • Adding the expression obtained in Equation (248) with the expression for the direct-effect term defined in Equation (240) yields the following expression for the G-differential δ R / U 0 :
    δ R k 0 = δ Q k 0 2 2 Q δ k 0 k 0 3 0 l φ 1 x 0 x U y d y d x δ U 0 + a 1 ( 2 ) 4 ; 0 δ U 0 2 δ b b θ 0 l a 2 ( 2 ) 4 ; x 0 x U y d y + x l φ 1 y d y d x + δ k 0 a 1 ( 2 ) 4 ; l .
    The above expression is to be satisfied at the nominal parameter values θ 0 but this fact has not been indicated explicitly (using the superscript “zero”) in order to simplify the notation.
  • Inserting into Equation (249) the result obtained in Equation (98) and identifying in the resulting form of Equation (249) the expressions that multiply the respective parameter variations yields the expressions for the following second-order sensitivities of the decoder response:
    2 R T 0 k 0 = δ U 0 + a 1 ( 2 ) 4 ; 0 δ U 0 = 1 + β T 0 a 1 ( 2 ) 4 ; 0 1 ;
    2 R β k 0 = T 0 2 2 a 1 ( 2 ) 4 ; 0 1 ;
    2 R Q k 0 = 1 k 0 2 0 l φ 1 x 0 x U y d y d x 1 k 0 0 l a 2 ( 2 ) 4 ; x 0 x U y d y + x l φ 1 y d y d x ;
    2 R k 0 k 0 = 2 Q k 0 3 0 l φ 1 x 0 x U y d y d x + a 1 ( 2 ) 4 ; l + Q k 0 2 0 l a 2 ( 2 ) 4 ; x 0 x U y d y + x l φ 1 y d y d x .
The sensitivities obtained in Equations (250)–(253) can be computed after having determined the 2nd-level adjoint sensitivity function a ( 2 ) 4 ; x . In this particular case, the equations underlying the 2nd-LASS, comprising Equations (245)–(247), for the two components of a ( 2 ) 4 ; x are decoupled and can therefore be solved independently of each other. Solving Equations (245) and (247) yields the following expression for the component a 1 ( 2 ) 4 ; x of the 2nd-level adjoint sensitivity function a ( 2 ) 4 ; x :
a 1 ( 2 ) 4 ; x = Q 2 k 0 b θ x l sin x l b θ + 1 H x l cos l x b θ .
Solving Equations (246) and (247) yields the following expression for the component a 2 ( 2 ) 4 ; x of the 2nd-level adjoint sensitivity function a ( 2 ) 4 ; x :
a 2 ( 2 ) 4 ; x = Q U 0 θ 2 b θ k 0 2 x sin x b θ .
Inserting into Equations (250)–(253) the expressions of obtained in Equations (254) and (255) together with the previously obtained expressions for the functions U x and φ 1 x , and performing the respective integrations yields the following closed-form expressions for the second-order sensitivities stemming from the first-order sensitivity R / k 0 :
2 R T 0 k 0 = 1 + β T 0 l 2 Q k 0 sin l b θ + cos l b θ 1 ;
2 R β k 0 = T 0 2 2 l 2 Q k 0 sin l b θ + cos l b θ 1 ;
2 R Q k 0 = l U 0 θ 4 l k 0 cos l b θ 1 Q k 0 sin l b θ ;
2 R k 0 k 0 = l U 0 θ 4 k 0 Q k 0 sin l b θ Q l k 0 cos l b θ .

5.2. Applying the 2nd-FASAM-NIDE-V Methodology to Compute Second-Order Sensitivities: Highlights

When applying the 2nd-FASAM-NIDE-V methodology, the “large-scale” computations occur when solving the 2nd-LASS to determine the respective 2nd-level adjoint function, since solving the 2nd-LASS involves inversion of differential and integral operators. After having obtained the 2nd-level adjoint function, the computation of the actual sensitivities is comparatively trivial, since it involves computing integrals by using quadrature formulas. As shown in Equation (178), the 2nd-LASS is a lower-triangular system of block-matrices for a 2-component 2nd-level adjoint function of the form a ( 2 ) i ; t a 1 ( 2 ) i ; t , a 2 ( 2 ) i ; t , the components of which are obtained by solving coupled subsystems of equations. Occasionally, however, the 2nd-LASS may be block-diagonal, so the components of the respective 2nd-level adjoint function are decoupled and can be determined independently of each other. This is the case for the 2nd-LASS developed for the 2nd-level adjoint function a ( 2 ) 3 ; x a 1 ( 2 ) 3 ; x , a 2 ( 2 ) 3 ; x , in Section 5.1.3, and for the 2nd-level adjoint function a ( 2 ) 4 ; x a 1 ( 2 ) 4 ; x , a 2 ( 2 ) 4 ; x in Section 5.1.4, respectively. Occasionally, the 2nd-LASS may be “degenerate,” involving a 2nd-level adjoint function comprising a single-component, as has been the case for the functions a ( 2 ) 1 ; x and, respectively, a ( 2 ) 2 ; x , in Section 5.1.1 and Section 5.1.2, respectively.
When applying the 2nd-FASAM-NIDE-V methodology, the 2nd-order unmixed sensitivities are obtained just once, using a distinct 2nd-level adjoint function for each unmixed 2nd-order sensitivity, as shown in Equation (194), Equation (211), Equation (231), and Equation (253), respectively. On the other hand, the 2nd-order mixed sensitivities are obtained twice, using distinct expressions involving distinct 2nd-level adjoint functions. For example, the sensitivity 2 R / Q T 0 is obtained using Equation (196) and/or Equation (229); the sensitivity 2 R / k 0 T 0 is obtained using Equation (197) and/or Equation (250), the sensitivity 2 R / Q k 0 is obtained using Equation (232) and/or Equation (252); and so on. These mixed 2nd-order sensitivities are obtained as inexpensively as practicable, involving just quadrature formulas, since the needed 2nd-level adjoint functions are already available. Thus, computing the mixed 2nd-order sensitivities twice, using distinct 2nd-level adjoint functions, provides the most inexpensive path for the stringent mutual verification of the accuracy of the computations performed when solving the respective 2nd-LASS for determining the respective 2nd-level adjoint functions.

6. Discussion and Conclusions

This work commenced by introducing the First-Order Features Adjoint Sensitivity Analysis Methodology for Neural Integro-Differential Equations of Volterra-Type (1st-FASAM-NIDE-V) for computing most efficiently the exact expressions of the first-order sensitivities of NIDE-V decoder-responses with respect to the optimized NIDE-V weights/parameters. The computation of the first-order sensitivities of the decoder response requires a single “large-scale” computation for solving the 1st-Level Adjoint Sensitivity System (1st-LASS), regardless of the number of weights/parameters underlying the NIE-net. As the comparison presented Section 3.3 has indicated, it is useful to investigate the advantages/disadvantages of applying the 1st-FASAM-NIDE-V versus the 1st-FASAM-NODE, when the system under consideration admits alternative representations as either an ordinary differential system or as an integro-differential Volterra system.
This work has also presented the 2nd-FASAM-NIDE-V, which builds on the 1st-FASAM-NIDE-V by considering the first-order sensitivities as “system responses” and applying the general principles underlying the Nth-CASAM [28] to determine the second-order sensitivities by determining the first-order sensitivities of these “system responses” (thus determining the “first-order sensitivities of the first-order sensitivities”). It has been shown that the 2nd-FASAM-NIDE-V methodology yields the exact expressions of all of the 2nd-order sensitivities and computes them with unparalleled efficiency, needing only as many “large-scale” computations as there are non-zero 1st-order sensitivities. These “large-scale” computations occur when solving the 2nd-LASS that corresponds to the respective non-zero 1st-order sensitivity considered as a “system response.” The unparalleled efficiency of the 2nd-FASAM-NIDE-V methodology stems from the fact that it scales at most linearly with the number of features functions as opposed to scaling exponentially with the number of model parameters, as do all the other (deterministic or statistical) methods. It has been shown that the mixed 2nd-order sensitivities are computed twice, using distinct 2nd-level adjoint functions, which provides the most inexpensive path for the stringent mutual verification of the accuracy of the computations performed when solving the respective 2nd-LASS for determining the respective 2nd-level adjoint functions.
This work has also shown that the existence of a linear analog of a nonlinear system enables significant simplifications and computational savings by performing the sensitivity analysis on the linear analog rather than directly on the nonlinear model. Albeit relatively rare, such instances are worth exploiting.
The mathematical framework of the 2nd-FASAM-NIDE-V can be generalized for computing arbitrarily-high order sensitivities of decoder response with respect to the NIDE-V feature functions and/or model parameters by applying the principles underlying the Nth-FASAM [29], which entails determining the next-order sensitivities as the “first-order sensitivities of the current sensitivities considered as system responses.” Thus, the 3rd-order sensitivities are obtained as the “first-order sensitivities” of the 2nd-order sensitivities considered as “system responses,” and so on. Future work will examine the theoretical underpinnings and feasibility of adapting algorithms of “backpropagation-type” for computing high-order sensitivities of decoder response with respect to feature functions of model parameters, rather than directly with respect to the model’s weights/parameters, in order to maximize the efficiency and accuracy of computing sensitivities of order higher than first-order.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author(s).

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Bishop, C.M. Neural Networks for Pattern Recognition; Clarendon Press: Oxford, UK, 1995. [Google Scholar]
  2. Bishop, C.M. Pattern Recognition and Machine Learning; Springer Science + Business Media: New York, NY, USA, 2006. [Google Scholar]
  3. Bishop, C.M.; Bishop, H. Deep Learning: Foundations and Concepts; Springer Nature: Cham, Switzerland, 2024. [Google Scholar]
  4. Chen, R.T.Q.; Rubanova, Y.; Bettencourt, J.; Duvenaud, D.K. Neural ordinary differential equations. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: New York, NY, USA, 2018; Volume 31, pp. 6571–6583. [Google Scholar] [CrossRef]
  5. Lu, Y.; Zhong, A.; Li, Q.; Dong, B. Beyond finite layer neural networks: Bridging deep architectures and numerical differential equations. In Proceedings of the International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 3276–3285. [Google Scholar]
  6. Ruthotto, L.; Haber, E. Deep neural networks motivated by partial differential equations. J. Math. Imaging Vis. 2018, 62, 352–364. [Google Scholar] [CrossRef]
  7. Grathwohl, W.; Chen, R.T.Q.; Bettencourt, J.; Sutskever, I.; Duvenaud, D. Ffjord: Free-form continuous dynamics for scalable reversible generative models. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  8. Dupont, E.; Doucet, A.; The, Y.W. Augmented neural odes. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Volume 32, pp. 14–15. [Google Scholar]
  9. Zhong, Y.D.; Dey, B.; Chakraborty, A. Symplectic ODE-net: Learning Hamiltonian dynamics with control. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 30 April 2020. [Google Scholar]
  10. Kidger, P.; Morrill, J.; Foster, J.; Lyons, T. Neural controlled differential equations for irregular time series. In Proceedings of the Advances in Neural Information Processing Systems, Virtual, 6–12 December 2020; Volume 33, pp. 6696–6707. [Google Scholar]
  11. Morrill, J.; Salvi, C.; Kidger, P.; Foster, J. Neural rough differential equations for long time series. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 18–24 July 2021; pp. 7829–7838. [Google Scholar]
  12. Kidger, P. On Neural Differential Equations. arXiv 2022, arXiv:2202.02435. [Google Scholar]
  13. Rokhlin, V. Rapid solution of integral equations of classical potential theory. J. Comput. Phys. 1985, 60, 187–207. [Google Scholar] [CrossRef]
  14. Rokhlin, V. Rapid solution of integral equations of scattering theory in two dimensions. J. Comput. Phys. 1990, 86, 414–439. [Google Scholar] [CrossRef]
  15. Greengard, L.; Kropinski, M.C. An integral equation approach to the incompressible Navier-Stokes equations in two dimensions. SIAM J. Sci. Comput. 1998, 20, 318–336. [Google Scholar] [CrossRef]
  16. Effati, S.; Buzhabadi, R. A neural network approach for solving Fredholm integral equations of the second kind. Neural Comput. Appl. 2012, 21, 843–852. [Google Scholar] [CrossRef]
  17. Xiong, Y.; Zeng, Z.; Chakraborty, R.; Tan, M.; Fung, G.; Li, Y.; Singh, V. Nystromformer: A nystrom-based algorithm for approximating self-attention. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 2–9 February 2021; Volume 35, p. 14138. [Google Scholar]
  18. Zappala, E.; de Oliveira Fonseca, A.H.; Caro, J.O.; van Dijk, D. Neural Integral Equations. arXiv 2023, arXiv:2209.15190v4. [Google Scholar]
  19. Zappala, E.; de Oliveira Fonseca, A.H.; Moberly, A.H.; Higley, J.M.; Abdallah, C.; Cardin, J.; Van Dijk, D. Neural Integro-Differential Equations. arXiv 2022, arXiv:2206.14282v1. [Google Scholar] [CrossRef]
  20. Gelmi, C.A.; Jorquera, H. IDSOLVER: A general-purpose solver for nth-order integrodifferential equations. Comput. Phys. Commun. 2014, 185, 392–397. [Google Scholar] [CrossRef]
  21. Karpel, J.T. IDESolver: A general purpose integro-differential equation solver. J. Open-Source Softw. 2018, 3, 542. [Google Scholar] [CrossRef]
  22. Volterra, V. Theory of functionals and of integral and integro-differential equations. Bull. Amer. Math. Soc. 1932, 38, 623. [Google Scholar]
  23. Lakshmikantham, V. Theory of Integro-Differential Equations; CRC Press: Boca Raton, FL, USA, 1995; Volume 1. [Google Scholar]
  24. Caffarelli, L.; Silvestre, L. Regularity theory for fully nonlinear integro-differential equations. Commun. Pure Appl. Math. A J. Issued by Courant Inst. Math. Sci. 2009, 62, 597–638. [Google Scholar] [CrossRef]
  25. Grigoriev, Y.N.; Ibragimov, N.H.; Kovalev, V.F.; Meleschko, S.V. Symmetries of integro-differential equations: With applications in mechanics and plasma physics. In Lecture Notes in Physics; Springer Science + Business Media: Dordrecht, The Netherlands, 2010; Volume 806. [Google Scholar]
  26. Wilson, H.R.; Cowan, J.D. Excitatory and inhibitory interactions in localized populations of model neurons. Biophys. J. 1972, 12, 1–24. [Google Scholar] [CrossRef]
  27. Medlock, J.; Kot, M. Spreading disease: Integro-differential equations old and new. Math. Biosci. 2003, 184, 201–222. [Google Scholar] [CrossRef] [PubMed]
  28. Cacuci, D.G. The nth-Order Comprehensive Adjoint Sensitivity Analysis Methodology (nth-CASAM): Overcoming the Curse of Dimensionality in Sensitivity and Uncertainty Analysis; Springer Nature: Switzerland, Cham, 2023; Volumes I and III. [Google Scholar] [CrossRef]
  29. Cacuci, D.G. Introducing the nth-Order Features Adjoint Sensitivity Analysis Methodology for Nonlinear Systems (nth-FASAM-N): I. Mathematical Framework. Am. J. Comput. Math. 2024, 14, 11–42. [Google Scholar] [CrossRef]
  30. Cacuci, D.G. First-Order Comprehensive Adjoint Sensitivity Analysis Methodology for Neural Ordinary Differential Equations: Mathematical Framework and Illustrative Application to the Nordheim–Fuchs Reactor Safety Model. J. Nucl. Eng. 2024, 5, 347–372. [Google Scholar] [CrossRef]
  31. Cacuci, D.G. Introducing the Second-Order Features Adjoint Sensitivity Analysis Methodology for Neural Ordinary Differential Equations. I: Mathematical Framework. Processes 2024, 12, 2660. [Google Scholar] [CrossRef]
  32. Cacuci, D.G. Introducing the Second-Order Features Adjoint Sensitivity Analysis Methodology for Fredholm-Type Neural Integral Equations. Mathematics 2025, 13, 14. [Google Scholar] [CrossRef]
  33. Cacuci, D.G. Introducing the Second-Order Features Adjoint Sensitivity Analysis Methodology for Neural Integral Equations of Volterra-Type: Mathematical Methodology and Illustrative Application to Nuclear Engineering. J. Nucl. Eng. 2025, 6, 8. [Google Scholar] [CrossRef]
  34. Cacuci, D.G. The First- and Second-Order Features Adjoint Sensitivity Analysis Methodologies for Fredholm-Type Neural Integro-Differential Equations: Mathematical Framework and Illustrative Application to a Heat Transfer Model. Processes 2025, 2025042086. [Google Scholar] [CrossRef]
  35. Cacuci, D.G. Sensitivity Theory for Nonlinear Systems: I. Nonlinear Functional Analysis Approach. J. Math. Phys. 1981, 22, 2794–2802. [Google Scholar] [CrossRef]
  36. Schneider, P.J. Conduction Heat Transfer; Addison-Wesley Publishing, Co.: Reading, MA, USA, 1955. [Google Scholar]
  37. Carslaw, H.S.; Jaeger, J.C. Conduction of Heat in Solids; Clarendon Press: London, UK, 1959. [Google Scholar]
  38. Arpaci, V.S. Conduction Heat Transfer; Addison-Wesley Publishing, Co.: Reading, MA, USA, 1966. [Google Scholar]
  39. Ozisik, M.N. Boundary Value Problems of Heat Conduction; International Textbook Company: Scranton, PA, USA, 1968. [Google Scholar]
  40. Ozisik, M.N. Heat Conduction; John Wiley & Sons: New York, NY, USA, 1980. [Google Scholar]
  41. Todreas, N.E.; Kazimi, M.S. Nuclear Systems I: Thermal Hydraulic Fundamentals; Taylor & Francis: Bristol, PA, USA, 1993. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cacuci, D.G. The First- and Second-Order Features Adjoint Sensitivity Analysis Methodologies for Neural Integro-Differential Equations of Volterra Type: Mathematical Framework and Illustrative Application to a Nonlinear Heat Conduction Model. J. Nucl. Eng. 2025, 6, 24. https://doi.org/10.3390/jne6030024

AMA Style

Cacuci DG. The First- and Second-Order Features Adjoint Sensitivity Analysis Methodologies for Neural Integro-Differential Equations of Volterra Type: Mathematical Framework and Illustrative Application to a Nonlinear Heat Conduction Model. Journal of Nuclear Engineering. 2025; 6(3):24. https://doi.org/10.3390/jne6030024

Chicago/Turabian Style

Cacuci, Dan Gabriel. 2025. "The First- and Second-Order Features Adjoint Sensitivity Analysis Methodologies for Neural Integro-Differential Equations of Volterra Type: Mathematical Framework and Illustrative Application to a Nonlinear Heat Conduction Model" Journal of Nuclear Engineering 6, no. 3: 24. https://doi.org/10.3390/jne6030024

APA Style

Cacuci, D. G. (2025). The First- and Second-Order Features Adjoint Sensitivity Analysis Methodologies for Neural Integro-Differential Equations of Volterra Type: Mathematical Framework and Illustrative Application to a Nonlinear Heat Conduction Model. Journal of Nuclear Engineering, 6(3), 24. https://doi.org/10.3390/jne6030024

Article Metrics

Back to TopTop