Development of Deterministic Artiﬁcial Intelligence for Unmanned Underwater Vehicles (UUV)

: The major premise of deterministic artiﬁcial intelligence (D.A.I.) is to assert deterministic self-awareness statements based in either the physics of the underlying problem or system identiﬁcation to establish governing di ﬀ erential equations. The key distinction between D.A.I. and ubiquitous stochastic methods for artiﬁcial intelligence is the adoption of ﬁrst principles whenever able (in every instance available). One beneﬁt of applying artiﬁcial intelligence principles over ubiquitous methods is the ease of the approach once the re-parameterization is derived, as done here. While the method is deterministic, researchers need only understand linear regression to understand the optimality of both self-awareness and learning. The approach necessitates full (autonomous) expression of a desired trajectory. Inspired by the exponential solution of ordinary di ﬀ erential equations and Euler’s expression of exponential solutions in terms of sinusoidal functions, desired trajectories will be formulated using such functions. Deterministic self-awareness statements, using the autonomous expression of desired trajectories with buoyancy control neglected, are asserted to control underwater vehicles in ideal cases only, while application to real-world deleterious e ﬀ ects is reserved for future study due to the length of this manuscript. In totality, the proposed methodology automates control and learning merely necessitating very simple user inputs, namely desired initial and ﬁnal states and desired initial and ﬁnal time, while tuning is eliminated completely.


Introduction
Artificial intelligence is most often expressed in stochastic algorithms that often have no knowledge whatsoever of the underlying problem being learned (a considerable strength of the methods). The field of non-stochastic or deterministic artificial intelligence breaks from this notion by first asserting the nature of the underlying problem using a self-awareness statement that permits the controlled item to have knowledge of itself, and this assertion allows the controlled item to learn in reaction to the environment. Thus, an unmanned vehicle with a deterministic self-awareness and learning can respond to significant damage that removes substantial vehicle parts or instead increases the vehicle's math models via inelastic collisions (e.g., bird strikes on aircraft, or robotic capture for spacecraft and underwater vehicles).
It is sometimes said that science fiction can be a precursor to science-fact, and such it is with artificial intelligence [1] vis-à-vis KarelČapek's Rossum's Universal Robots and Frankenstein by Mary Shelley [2]. Ethical issues of artificial intelligence were first grappled with in such fictional works [1]. As ancient mathematicians and philosophers studied reasoning formally, computation as a theory emerged, embodied in the Church-Turing thesis suggesting that any mathematical action could be represented digitally by combinations of zeros and ones [3], where "intelligence" was defined as the ability for these mathematical actions being indistinguishable from human responses [4]. In 1943, Pitts and McCullouch formalized Turing's design, whose artificial neurons instantiated artificial intelligence [5]. In 1956, Dartmouth College hosted a summer workshop [6] attended by A.I. pioneers IBM's Arthur Samuel, MIT's Marvin Minsky and John McCarthy, and Carnegie-Mellon's Herbert Simon and Allen Newell [2]. The result was computer programs that learned checkers strategies [2], and within four years [7], these programs were better than average humans [8]. Logical Theorems were proved in the computers' ability to solve algebra word problems and speaking English [2]. Artificial intelligence research subsequently launched in earnest following considerable monetary support by the U.S. defense department [9], but was not limited to the United States [10]. An original Dartmouth summer workshop attendee from MIT, Marvin Minsky, was so enthusiastic to state in writing, "within a generation . . . the problem of creating 'artificial intelligence' will substantially be solved" [9,11,12], clearly failing to acknowledge some very difficult challenges.
Sir James Lighthill's critique [13] of the early 1970s exemplifies the admission of the lost potential realized by the de-funding of artificial intelligence research by Britain and American governments, demarking a time-period of artificial intelligence research drought [5,9,10,14,15]. Seeking to develop artificially intelligent systems mimicking human experts' knowledge and analytical skills, a billion-dollar commercial sector substantially launched in the 1980s [1,2,9,[14][15][16][17]. Research by academia was shortly salvaged by re-initiation of government support to compete for honor in the face of the support by Japan [1,2,9,16], but a longer drought immediately followed in the late 1980s [1,9,16,17]. By the 1990s, other research [18] presented logic, rule, object and agent-based architectures, along with example programs written in LISP and PROLOG. Accompanying continual computation evolution often referred to as "Moore's Law," artificial intelligence was afterwards adopted for statistics and mathematics, data mining, economics, medicine, and logistics [2,16,17,19], leading to the adoption of standards [1,2] An IBM company artificial intelligence system easily beat two of the best players at the TV show Jeopardy in 2011 [20], and by 2012, so-called deep-learning methods became the most accurate instantiations [21]. Currently, artificial intelligence algorithms are used every day in smart phones and video game consoles such as Xbox [22,23], and even the abstract strategy board game Go [24][25][26][27][28]. Shortly thereafter, the search-engine giant Google had, by 2015, nearly 3000 artificial intelligence software projects, while the maturity of artificial intelligence was documented in Bloomberg vis-à-vis low image processing error rates attributable to cost reduction of neural networks accompanying the significant expansion of infrastructure for cloud computing and proliferation of data sets and tools for research [29]. Image processing error improvements led the social media company Facebook to develop a system to describe images to the blind, while Microsoft enhanced their Skype telecommunication software with artificial intelligent language translation abilities [29]. Twenty percent of companies surveyed in 2017 used artificial intelligence in some form [30,31] amidst increased government funding in China [32,33]; however, it must be elaborated that in a modern continuation of the original exaggeration in the 1960s and 1970s of the abilities of artificial intelligence, current reporting has also been established as exaggerated [34][35][36].
In order to statistically maximize success likelihood, artificial intelligent systems continually analyze their environment [2,15,37,38], while induced or explicit goals are expressed by figures of merit often referred to as a utility function, which can be quite complex or quite simple. Utility functions in deterministic systems are sometimes referred to as cost functions, where the cost is pre-selected and subsequently maximized or minimized to produce the deterministic form of the best behavior similarly to how animals evolved to innately possess goals such as finding food [39]. Some systems, such as nearest-neighbor are non-goal systems framed as systems whose "goal" is to successfully accomplish its narrow classification task [40].
The goal pursued in the proposed deterministic artificial intelligence approach will be the elimination of motion trajectory tracking error (and do so in an optimal manner), and therefore, the approach will seek to exactly track unknown or arbitrary user commands, and learning will be performed on truly unknowable parameters that are typically assumed to be reasonable constants in classical linear, time-invariant (LTI) methods. The physical system will be an autonomously piloted (without human intervention) underwater robotic vehicle that has already been well-studied by classical methods [41]. Improvement over the most recent investigation of classical methods [42] was proposed to and funded by the U.S. Office of Naval Research, where the winning proposal [43] contained two parts: deterministic artificial intelligence methods and non-propulsive maneuvering by momentum exchange stemming from a recent lineage [44][45][46][47][48][49][50][51]. While the former is the subject of this manuscript, the latter is neglected here (reserved for a future manuscript).
While most artificial intelligence methods are stochastic, emerging deterministic artificial intelligence (not to be confused with deterministic algorithms [52] or deterministic environments [53]) ignores articulated uncertainty, preferring instead to determine the analytic relationship that minimizes uncertainty (e.g., so-called two-norm optimization among others). Therefore, a key challenge is re-parameterization of the underlying problem into a form that permits such optimization to minimize variance and/or uncertainty.
Thus, a key novel contribution of this manuscript is re-parameterization, in addition to the illustration of optimality and validation of basic performance. Autonomous trajectory generation accepting arbitrary user commands and outputting statements of desired full state trajectories will be adopted and elaborated from the very recent literature [54], while parameter modification and disturbance formulation will similarly adopt the most current methods [55]. Taking these new developments for granted, the remaining task addressed in this manuscript will be accomplished by asserting physics-based [56][57][58] mathematical deterministic self-awareness statements, while learning will be developed using simple methods (stemming from a heritage in adaptive systems) [59-79] and also new optimal methods embodying the aforementioned reparameterization seeking 2-norm error minimization learning. [59-61] describe use of mathematic models are references, where adaption rules are invoked to make the system behave like the chosen model. [61][62][63][64][65][66] articulate methods of identifying system models from data of performance for various systems, rather than invocation of preferred reference models. References [67][68][69][70][71][72][73] utilize data-derived models in the various adaptive schemes, while [74] substantiates the first evolution from simple adaptive systems with the innovation of deterministic self-awareness applied to the forced van der Pol equation. References [75][76][77][78] illustrate the utilization of the system models towards optimization and prediction. Lastly, [79] is the first book devoted to the application of the methods developed here applied to space systems, while this manuscript applies the methods to underwater vehicles. Care is taken to cite the original source of the inspiring first principle relationships for deterministic self-awareness statements in addition to the references of subsequent expansion and eventual application to deterministic self-awareness.
Why Use the Proposed Approach on a UUV?
One instance to consider using deterministic artificial intelligence (D.A.I.) is where the users strictly need autonomous operations, since the method is developed to make the Unmanned Underwater Vehicles (UUVs) depicted in Figure 1 self-aware with learning mechanisms that permit the UUV to autonomously operate (a goal shared with common robotics and controls methods), but also understand its performance in terms of its mathematical model leading to an ability to self-correct, but also to self-diagnose problems and notify land-based or ship-based infrastructure details of the vehicle's dynamic models.
Another reason to consider using D.A.I. for UUVs lies in the elimination of the necessity for backgrounds in robotics and/or controls, substituting instead with a prerequisite understanding of (1) dynamics and (2) regression. In model predictive control [75][76][77], the dynamics are used to predict system behavior, which is compared to the behavior of a desired (simpler) dynamic model, and the difference between the two is used to formulate a control that seeks to drive the actual dynamic behavior to mimic the behavior of the desired dynamic. Reference [76] also highlights a very realistic issue: actuator saturation and under actuation (driven by hardware selection). This motivates the final Another reason to consider using D.A.I. for UUVs lies in the elimination of the necessity for backgrounds in robotics and/or controls, substituting instead with a prerequisite understanding of (1) dynamics and (2) regression. In model predictive control [75][76][77], the dynamics are used to predict system behavior, which is compared to the behavior of a desired (simpler) dynamic model, and the difference between the two is used to formulate a control that seeks to drive the actual dynamic behavior to mimic the behavior of the desired dynamic. Reference [76] also highlights a very realistic issue: actuator saturation and under actuation (driven by hardware selection). This motivates the final sections of this manuscript: an implementable operational procedure that starts with a selection of actuators, and uses their limits to determine the desired maneuver times for the autonomous trajectory generation scheme, which feeds the D.A.I. calculations.
With D.A.I, the dynamics comprise the self-awareness statements (albeit necessitating the articulation of desired trajectories). This assertion of the dynamics stems from the physics-based control methods [56][57][58], which was proven to eliminate lag in references [58,67]. Such assertion is also shared with the field of nonlinear adaptive control [69][70][71][72][73], while the feedback is typically of the classical type (utilizing proportional, derivative, and sometimes integral actions). The use of feedback in the form of learning presented here neglects these classical forms, separating the method from the nonlinear adaptive control. Other forms of adaptive control and system identification (e.g., autoregressive moving average parameterization, among others) do not parameterize the problem in terms of the dynamics, as presented in [62][63][64][65][66]. One key development illustrating the powerful idea of asserting the dynamics as self-awareness statements comes in a research by Cooper et al. [74], who eliminated feedback altogether in the control of the forced van der Pol equation. Other approaches emphasizing the use of the dynamics include real-time optimal control as compared in [75] to several other approaches, including classical feedforward plus feedback, open-loop optimal control and also predictive control. Offline nonlinear optimization [78] also emphasizes use of the dynamics, but adjoins the dynamics and constraints with a cost function to numerically find trajectories and controls that minimize the combined adjoined cost function. This use of the dynamics is effective but relatively unwieldy compared to the analytic approach developed here.

Materials and Methods
Deterministic artificial intelligence (D.A.I.) requires a self-awareness (mathematical) statement derived from the governing physics followed by some type of learning (either stochastic or deterministic). The deterministic self-awareness statement requires the creation of a desired state With D.A.I, the dynamics comprise the self-awareness statements (albeit necessitating the articulation of desired trajectories). This assertion of the dynamics stems from the physics-based control methods [56][57][58], which was proven to eliminate lag in references [58,67]. Such assertion is also shared with the field of nonlinear adaptive control [69][70][71][72][73], while the feedback is typically of the classical type (utilizing proportional, derivative, and sometimes integral actions). The use of feedback in the form of learning presented here neglects these classical forms, separating the method from the nonlinear adaptive control. Other forms of adaptive control and system identification (e.g., auto-regressive moving average parameterization, among others) do not parameterize the problem in terms of the dynamics, as presented in [62][63][64][65][66]. One key development illustrating the powerful idea of asserting the dynamics as self-awareness statements comes in a research by Cooper et al. [74], who eliminated feedback altogether in the control of the forced van der Pol equation. Other approaches emphasizing the use of the dynamics include real-time optimal control as compared in [75] to several other approaches, including classical feedforward plus feedback, open-loop optimal control and also predictive control. Offline nonlinear optimization [78] also emphasizes use of the dynamics, but adjoins the dynamics and constraints with a cost function to numerically find trajectories and controls that minimize the combined adjoined cost function. This use of the dynamics is effective but relatively unwieldy compared to the analytic approach developed here.

Materials and Methods
Deterministic artificial intelligence (D.A.I.) requires a self-awareness (mathematical) statement derived from the governing physics followed by some type of learning (either stochastic or deterministic). The deterministic self-awareness statement requires the creation of a desired state trajectory, and rather than demand offline work by a designing engineer, the first section of this section of the manuscript introduces one potential online autonomous (desired) trajectory generation scheme. Afterwards, the deterministic self-awareness statement will be introduced neglecting standard kinematic representations [80] in favor or recent expressions for UUVs [42], followed by two approaches to learning: simple learning, and optimal learning. Each component of D.A.I. will be elaborated in sufficient detail to allow the reader to replicate the published results and use the methods on systems not elaborated in this document. Graphic images of systems coded in MATLAB/SIMULINK (depicted in Figure 2) will aid the reader in creating their own validating simulations. For the sake of easier reading, variable definitions are provided in each section where the reader will be introduced to the variables, while a summary table of all variable definitions is provided in the Appendix A.
section of the manuscript introduces one potential online autonomous (desired) trajectory generation scheme. Afterwards, the deterministic self-awareness statement will be introduced neglecting standard kinematic representations [80] in favor or recent expressions for UUVs [42], followed by two approaches to learning: simple learning, and optimal learning. Each component of D.A.I. will be elaborated in sufficient detail to allow the reader to replicate the published results and use the methods on systems not elaborated in this document. Graphic images of systems coded in MATLAB/SIMULINK (depicted in Figure 2) will aid the reader in creating their own validating simulations. For the sake of easier reading, variable definitions are provided in each section where the reader will be introduced to the variables, while a summary table of all variable definitions is provided in the appendix.

Deterministic Artificial Intelligence Self-Awareness Statement
In a two-step process, first impose the governing equations of motion expressing the dominant physics of the problem as the deterministic artificial intelligence self-awareness statement. The second step will be to assert these governing equations using "desired trajectories." Rearranging the equations allows expression in the so-called state-space form ubiquitously associated with modern control methods, a very well-articulated topic [81][82][83][84][85]. Subsequently (in the next section of this manuscript), the self-awareness statement(s) will be reparametrized to isolate variables that are to be learned (in this instance: center of gravity, masses, and mass moments of inertia); first using simple learning algorithms, then using optimal learning methods. Items that possess mass behave in accordance with several first principles, among them Hamilton's Principle [86], the conservation of energy or of momentum [87][88][89][90][91] and angular momentum [92][93][94][95], Lagrange's equations [96], Kane's Method [97][98][99][100], and Newton's Law [101] together with Euler's equations [102][103][104][105][106][107], which may be invoked in accordance with Chasle's theorem [108], while invocation motivates the assertion of the dynamic equations as deterministic self-awareness statements. The dynamics of unmanned underwater vehicles, per the defining components of dynamics comprises kinetics in Equations (1)-(6) and kinematics, and these are expressed in Equations (7)- (11). As equations are developed small "mini tables" of acronyms are provide in Tables 1-3 to aid readability without continually flipping back and forth between pages, while a complete "summary table" of acronyms is provided in the appendix in Table A1. Table 1. Definitions of variables in this part of the manuscript (while a consolidated list is provided in the Appendix).

Mass
Position of center of mass in meters Mass moment of inertia with respect to a vertical axis that passes through the vehicle's geometric center (amidships) , ̇ Lateral (sway) velocity and rate , ̇ Heading angle and rate , , , ̇ Cartesian position coordinates and derivatives

Deterministic Artificial Intelligence Self-Awareness Statement
In a two-step process, first impose the governing equations of motion expressing the dominant physics of the problem as the deterministic artificial intelligence self-awareness statement. The second step will be to assert these governing equations using "desired trajectories." Rearranging the equations allows expression in the so-called state-space form ubiquitously associated with modern control methods, a very well-articulated topic [81][82][83][84][85]. Subsequently (in the next section of this manuscript), the self-awareness statement(s) will be reparametrized to isolate variables that are to be learned (in this instance: center of gravity, masses, and mass moments of inertia); first using simple learning algorithms, then using optimal learning methods. Items that possess mass behave in accordance with several first principles, among them Hamilton's Principle [86], the conservation of energy or of momentum [87][88][89][90][91] and angular momentum [92][93][94][95], Lagrange's equations [96], Kane's Method [97][98][99][100], and Newton's Law [101] together with Euler's equations [102][103][104][105][106][107], which may be invoked in accordance with Chasle's theorem [108], while invocation motivates the assertion of the dynamic equations as deterministic self-awareness statements. The dynamics of unmanned underwater vehicles, per the defining components of dynamics comprises kinetics in Equations (1)-(6) and kinematics, and these are expressed in Equations (7)- (11). As equations are developed small "mini tables" of acronyms are provide in Tables 1-3 to aid readability without continually flipping back and forth between pages, while a complete "summary table" of acronyms is provided in the Appendix A in Table A1. ψ Heading angle and rate x, y,

Nondimensional Variable Definition
Sway force coefficients: coefficients describing sway forces from resolved lift, drag, and fluid inertia along body lateral axis. These occur in response to individual (or multiple) velocity, acceleration and plane surface components, as indicated by the corresponding subscripts Arbitrarily labeled constants used to simplify expressions Variable found ubiquitously throughout Equations (1)- (52), but this reduced list aids readers follow proximal Equations (1)-(21) [41].
νdt. Using these facts, desired states may be calculated using Equations (10) and (11) where subscripted d are added to indicate desired states: Equations (1)-(11) merely contain the vehicle dynamics (kinetics and kinematics), neglecting deleterious real-world factors such as disturbances, noise, unmodeled dynamics, and especially mismodeled dynamics. In Section 3, the vehicle dynamics are used to codify self-awareness statements and re-parameterized to substantiate learning. Sequel treatment of deleterious effects will follow exactly the same process, highlighting the generic appeal of the methodology. For example, hydrodynamic forces and moments are well-modeled phenomenon of physics. Mathematical expression of hydrodynamics (and other deleterious effects) will be asserted as self-awareness, and then re-parameterized for learning in the sequel.
In accordance with physics-based control methods, Equations (7) and (8) in particular, will be used to formulate deterministic self-awareness statements that will be asserted to form the control algorithm of the unmanned underwater vehicles. The equations will be re-parameterized, not around the motion states, but instead the motion states are assumed to be known (via sensors and inertial navigation such as state observers, Kalman filter, etc.), while the unknown/unknowable states will be mass properties.
The key procedure is to simply re-write the equations of motion isolating the mass properties, add a "d" subscript to the motion states (necessitating articulation of a desired state trajectory), and lastly, add a "ˆ" superscript to the properties, indicating these quantities will be learned subsequently. Section 3 will use Equations (7)- (11) to formulate optimal self-awareness statements where motion states will be replaced with desired motion states to be provided by autonomously generated trajectories in accordance with the methods presented in [55].

Autonomous Trajectory Generation
The goal is to have an intelligent system accept a desired endstate from the user and use that to autonomously created the entire maneuver trajectory without assistance. This maneuver trajectory will be used subsequently to formulate the deterministic self-awareness statement. One approach to autonomous trajectory generation is to impose a structure. Inspired by the nature of the exponential function as a solution-form for differential equations of motion, this structure will be imposed here. The simple motion ordinary differential equation . z = Az can be assumed to have an exponential solution. The solution for z(t) may be differentiated and substituted back into the original motion equation to solve for the constants, A and λ. Additionally, recall that Euler's Formula may be use to express the exponential as a sum of sines and cosines as seen in Equations (12) and (13), where the initial condition is assumed quiescent, eliminating the cosine: If a nominal sine curve constitutes a position trajectory, and the first and second derivatives, respectively, constitute the velocity and acceleration trajectory per Equation (13): A brief quiescent period ∆t quiescant (for trouble-shooting) is preferred, followed by a maneuver from the initial position (amplitude, A 0 ) to a commanded position (amplitude, A) in a specified maneuver time, ∆t maneuver , subsequently followed by regulation at the new position, as depicted in Figure 3, where overshoot, oscillation, and settling are all undesirable traits. The following paragraph illustrates a systematic procedure to modify the nominal sine equation to achieve these desires while also ensuring smooth initiation of the maneuver (to avoid exciting un-modeled flexible vibration modes).
A brief quiescent period ∆ (for trouble-shooting) is preferred, followed by a maneuver from the initial position (amplitude, ) to a commanded position (amplitude, ) in a specified maneuver time, ∆ , subsequently followed by regulation at the new position, as depicted in Figure 3, where overshoot, oscillation, and settling are all undesirable traits. The following paragraph illustrates a systematic procedure to modify the nominal sine equation to achieve these desires while also ensuring smooth initiation of the maneuver (to avoid exciting un-modeled flexible vibration modes).
(a) Piecewise-continuous sinusoidal maneuver trajectory. An initial quiescent period is followed by a designed maneuver, subsequently followed by regulation at the new commanded position.
(b) Nominal sine curve x = Asin(ωt) used to reveal the relationship between points on the curve's time (fraction of the curves total period) and phase angle (fraction of 2π).

Figure 3.
Creation of a piecewise-continuous sinusoidal maneuver trajectory autonomously generated to guarantee a specified maneuver time (to be provided by offline optimization).
A nominal sine curve = is depicted in Figure 3b represented by the thick-dashed line. Note that it starts abruptly at time = 0, while a smooth initiation is preferred to permit future expansion to flexible multi-body equations of motion (while rigid body dynamics are assumed here). The low point occurring at time = is desired to be placed at the designated maneuver time ( = 5 here, assuming a five second quiescent period).
1. Choose the maneuver time: ∆ = 2 is used here illustratively. Express the maneuver time as a portion (half) of the total sinusoidal period, , as depicted in figure (3b).
-The result is Equation (14): Important side comment: ∆t is provided by the user, thus, this time period can be optimized (often represented as t * ) to meet any number of cost functions, J and constraint equations.
2. Phase shift the curve to place the smooth low-point from = 3 to the desired maneuver start time, following the quiescent period at = 5.
-The result is Equation (15) plotted in Figure 4a: Figure 3. Creation of a piecewise-continuous sinusoidal maneuver trajectory autonomously generated to guarantee a specified maneuver time (to be provided by offline optimization).
A nominal sine curve z = Asin(ωt) is depicted in Figure 3b represented by the thick-dashed line. Note that it starts abruptly at time t = 0, while a smooth initiation is preferred to permit future expansion to flexible multi-body equations of motion (while rigid body dynamics are assumed here). The low point occurring at time t = 3T 4 is desired to be placed at the designated maneuver time (t = 5 here, assuming a five second quiescent period).

1.
Choose the maneuver time: ∆t maneuver = 2 is used here illustratively. Express the maneuver time as a portion (half) of the total sinusoidal period, T, as depicted in Figure 3b.
-The result is Equation (14): Important side comment: ∆t maneuver is provided by the user, thus, this time period can be optimized (often represented as t*) to meet any number of cost functions, J and constraint equations.

2.
Phase shift the curve to place the smooth low-point from t = 3 to the desired maneuver start time, following the quiescent period at t quiescant = 5. - The result is Equation (15) plotted in Figure 4a: 3.
Compress the amplitude for the desired final change in amplitude to equate to the top-to-bottom total of curve.
-The result is Equation (16) plotted in Figure 4b:

4.
Amplitude-shift the curve up for smooth initiation at arbitrary starting position used here by adding A 0 .
-The result is Equation (17) plotted in Figure 4c:

5.
Craft a piecewise continuous trajectory such that amplitude is zero until the termination of t quiescant . Follow the sinusoidal trajectory during the maneuver time indicated by ∆t maneuver , and then hold the final amplitude afterwards. - The result is Equation (18) plotted in Figure 4d: 6.
∆t quiescant is often omitted operationally, while ∆t maneuver is defined in Section 3 of this manuscript for any available control authority. Equations (19)-(21) are used to autonomously generate sway velocity and yaw rate trajectories for any initial and final state, any maneuver start time, and any desired total maneuver time (equivalently time at desired end state).

Topologies and Implementation in SIMULINK
Section 2.1 of this manuscript introduced the notion of asserting deterministic self-awareness statements invoking the mathematics of physics, and the description revealed the necessity of analytic expressions for trajectories that permit autonomous generation, and one such method is articulated in Section 2.2. Next, Section 2.3 displays topologies using the SIMULINK program where equation numbers were used as labels, and those equation numbers beg further explanation. Section 3 contains detailed development (from first principles) of the proposed method of deterministic artificial intelligence, and along the way, equations are labeled consistent with the presentations of topologies in Section 2. The first principles are the work of such famous scientists as Newton, Euler, and Chasle and some modern expressions are cited including Kane, Goldstein, Wie. The principles are presented as factually self-evident without much articulation, instead cited for independent pursuit by the reader.

Results
Stemming from the earlier described first principles [97][98][99][100][101][102][103][104][105][106][107][108], whose classical usage is described in [86][87][88][89][90][91][92][93][94][95][96], we accept these principles and adopt them as deterministic self-awareness statements. Next, we illustrate the parameterization of the statements in standard state-variable form [81][82][83][84][85]. We illustrate simple learning techniques that are honestly indistinguishable from non-linear adaptive control techniques [68][69][70][71][72][73] mentioned as "the lineage" of the subsequent technique presented: optimal learning using the state-variable formulation to highlight the pseudo-inverse optimal solution as the learning relationship. Nonlinear-adaptive techniques still require tuning, while the optimal learning relationship proposed here negates this requirement. The assertion of self-awareness statements is validated with several maneuvers, while the validating simulations with optimal learning are expressed in the presentation of the empirically derived relationship between maximum available actuation versus achievable minimum maneuver time (using both assertion of self-awareness statements and optimal learning). The motivation for this is presented at the end of the section: a nominal implementation procedure for deterministic artificial intelligence, whose first step is to establish minimum achievable maneuver time as a function of available actuators. The implementation methods bring Section 3 to a close, while Section 4 includes an operational implementation checklist.

Table 3. Definitions of variables in this part of the manuscript (while a consolidated list is provided in the Appendix A).
Variable Definition a 1 , a 2 , a 3 , a 4 Variables in state-variable formulation ("state space") of equations of motion associated with motion states Variables in state-variable formulation ("state space") of equations of motion associated with controls δ * s Deterministic error-optimal stern rudder displacement commands δ * b Deterministic error-optimal bow rudder displacement commandŝ m Learned vehicle masŝ mx G Learned product of vehicle mass and location of center of masŝ x G Learned location of center of masŝ I Z ,Î Z0 Learned mass moment of inertia and initial value K m1 Control gain for mass simple-learning K mxG1 Control gain for learning product of mass and location of center of mass K I1 , K I2 Control gain for learning mass moment of inertia Variables (combinations of motion states) used to reparametrize problem into optimal learning form Superscript "ˆ" indicates quantities estimated by learning or feedback in Equations (1)- (52), but this reduced list aids readers to follow proximal Equations (22)-(52).

Articulate Optimal Deterministic Self-Awareness Statement
Equations (22)-(25) articulate the equations of motion from Equations (7)-(9) expressed using desired trajectories indicated by subscript "d". This comprises the first step of applying the deterministic artificial intelligence technique. Learning will be described in Sections 3.3 and 3.4 after first reformulating the dynamics in Section 3.2.
Note: Equations (22) through (25) do not yet have learned estimates (which are indicated by superscriptˆ). These equations define the deterministic self-awareness statements to be asserted and subsequently learning will be applied to these statements as elaborated in Sections 3.3 and 3.4. ¢ i ∀i ∩ (1 . . . 8) and ¢ u 1 ∀i ∩ (1 . . . 2) defined in Equations (1)-(8) contain m, I z , mx g , and these values will be assumed known for use in Equations (22)-(25).

Formulate Optimal Deterministic Self-Awareness Statement in MIMO State Space Form
This section derives the full-form state-variable representation of the optimal deterministic self-awareness statement, where δ s = −δ b rudder constraint equation is not enforced. Rudders are free to be designed separately in a "many-in, many-out" (MIMO) formulation. Isolate . υ in Equation (24), and then isolate . r in Equation (25): .
Parameterize in a general form .
x = Ax to reveal optimal rudder commands based on the deterministic self-awareness statement: Note: Equations (36) and (37) do not yet have learned estimates (which are indicated by a "hat" superscriptˆ). ¢ i ∀i ∩ (1 . . . 8) and ¢ u 1 ∀i ∩ (1 . . . 2) defined in Equations (1)-(8) contain m, I z , mx g , and these values will be assumed known for use in Equations (36) and (37). These equations define the deterministic self-awareness statements to be asserted and subsequently learning will be applied to these statements as elaborated in Sections 3.3 and 3.4.

Validating Simulations
This section displays many simulations of various maneuvers and scenarios to validate a functioning simulation and the ability of the proposed method of deterministic artificial intelligence to control the unmanned underwater. Figures 5-9 illustrate general topologies with inputs, mappings, and outputs. The mappings are labeled with equation numbers elaborated in Sections 2 and 3 of this manuscript, permitting readers to create their own simulations using this manuscript as a guide. For operational implementation, the exact SIMULINK program may be used to command and control laboratory hardware by replacing the "Phoenix vehicle dynamics" block with electrical connectivity to the lab hardware.  (52) and (48) execute optimal learning. Equation (37) is used to find optimal rudder commands and within that SIMULINK block is also the assertion of self-awareness statements.  . Topology (using actual SIMULINK model) of optimal rudder commands whose block has embedded the assertion of self-awareness; simple learning (using proportional + derivative components); and optimal learning (vis a vis the two-norm error optimal pseudoinverse solution). Figure 6 displays sub-systems of Figure 5, specifically the second subsystem from the left labeled "Optimal Rudder Commands." The subsystem displayed in Figure 6 accepts the desired motion states and the actual motion states as inputs and outputs the two-norm optimal rudder commands using Equation (37), whose constants are learned by simple methods in Equations (42)-(44), or alternatively two-norm optimal learning in Equation (52) with Equation (48) and analytic expression used for solving the location of the center of gravity, which appears nonlinearly coupled with the vehicle mass.   Figure 6. Simple learning uses Equation (42) to learning the vehicle mass, which is used in Equation  (52) and (48) execute optimal learning. Equation (37) is used to find optimal rudder commands and within that SIMULINK block is also the assertion of self-awareness statements. Figure 5 displays the SIMULINK simulation of the equations developed in Section 3 of this manuscript, and was used to produce the validating simulations presented in Section 3. The simulation was also used to develop the coding procedure and operational implementation procedure presented in Section 4. These individual blocks displayed in Figure 5 are expanded in Figures 6-9. Compare Figure 5 to Figure 2 to reveal assumptions of known or subsequently executable components: specific actuators, sensors, state observers, filters, and specific disturbances (withheld for future research seeking the limits of the robustness of the proposed techniques).  (52) and (48) execute optimal learning. Equation (37) is used to find optimal rudder commands and within that SIMULINK block is also the assertion of self-awareness statements.  . Topology (using actual SIMULINK model) of optimal rudder commands whose block has embedded the assertion of self-awareness; simple learning (using proportional + derivative components); and optimal learning (vis a vis the two-norm error optimal pseudoinverse solution). Figure 6 displays sub-systems of Figure 5, specifically the second subsystem from the left labeled "Optimal Rudder Commands." The subsystem displayed in Figure 6 accepts the desired motion states and the actual motion states as inputs and outputs the two-norm optimal rudder commands using Equation (37), whose constants are learned by simple methods in Equations (42)-(44), or alternatively two-norm optimal learning in Equation (52) with Equation (48) and analytic expression used for solving the location of the center of gravity, which appears nonlinearly coupled with the vehicle mass.   Figure 6. Simple learning uses Equation (42) to learning the vehicle mass, which is used in Equation Figure 6. Topology (using actual SIMULINK model) of optimal rudder commands whose block has embedded the assertion of self-awareness; simple learning (using proportional + derivative components); and optimal learning (vis a vis the two-norm error optimal pseudoinverse solution). Figure 6 displays sub-systems of Figure 5, specifically the second subsystem from the left labeled "Optimal Rudder Commands." The subsystem displayed in Figure 6 accepts the desired motion states and the actual motion states as inputs and outputs the two-norm optimal rudder commands using Equation (37), whose constants are learned by simple methods in Equations (42)-(44), or alternatively two-norm optimal learning in Equation (52) with Equation (48) and analytic expression used for solving the location of the center of gravity, which appears nonlinearly coupled with the vehicle mass.  (52) and (48) execute optimal learning. Equation (37) is used to find optimal rudder commands and within that SIMULINK block is also the assertion of self-awareness statements. Figure 5 displays the SIMULINK simulation of the equations developed in Sections 3 of this manuscript, and was used to produce the validating simulations presented in Sections 3. The simulation was also used to develop the coding procedure and operational implementation procedure presented in Section 4. These individual blocks displayed in Figure 5 are expanded in Figures 6-9. Compare Figure 5 to Figure 2 to reveal assumptions of known or subsequently executable components: specific actuators, sensors, state observers, filters, and specific disturbances (withheld for future research seeking the limits of the robustness of the proposed techniques). Figure 6. Topology (using actual SIMULINK model) of optimal rudder commands whose block has embedded the assertion of self-awareness; simple learning (using proportional + derivative components); and optimal learning (vis a vis the two-norm error optimal pseudoinverse solution). Figure 6 displays sub-systems of Figure 5, specifically the second subsystem from the left labeled "Optimal Rudder Commands." The subsystem displayed in Figure 6 accepts the desired motion states and the actual motion states as inputs and outputs the two-norm optimal rudder commands using Equation (37), whose constants are learned by simple methods in Equations (42)-(44), or alternatively two-norm optimal learning in Equation (52) with Equation (48) and analytic expression used for solving the location of the center of gravity, which appears nonlinearly coupled with the vehicle mass.   Figure 6. Simple learning uses Equation (42) to learning the vehicle mass, which is used in Equation    (1) and (4) expressed with desired rather than actual trajectories. Used here as actual rather than desired trajectories, no "d" subscript is present.  Recall that no tuning was required and nonetheless motion states are controlled through several types of maneuvers in Figures 10 and 11. Figure 11 also displays the nature of rudder motions where a single rudder is locked-down (either intentionally or by damage received while underway).
(a) Validating maneuver: tight turn. x(t) on the abscissa and y(t) on the ordinant.    Figure 6.
(a) Equation (36): note that this is a state space representation of the dynamic equations of motion in Equations (1) and (4) expressed with desired rather than actual trajectories. Used here as actual rather than desired trajectories, no "d" subscript is present.  Recall that no tuning was required and nonetheless motion states are controlled through several types of maneuvers in Figures 10 and 11. Figure 11 also displays the nature of rudder motions where a single rudder is locked-down (either intentionally or by damage received while underway).
(a) Validating maneuver: tight turn. x(t) on the abscissa and y(t) on the ordinant.   Recall that no tuning was required and nonetheless motion states are controlled through several types of maneuvers in Figures 10 and 11. Figure 11 also displays the nature of rudder motions where a single rudder is locked-down (either intentionally or by damage received while underway).  (43) to learn the time-varying location of the vehicle center of gravity, while Equation (44) is used to learn the mass moments of inertia. Figure 8. Optimal learning topology manifest as a SIMULINK simulation program representing the lower left subsystem in Figure 6.
(a) Equation (36): note that this is a state space representation of the dynamic equations of motion in Equations (1) and (4) expressed with desired rather than actual trajectories. Used here as actual rather than desired trajectories, no "d" subscript is present.  Recall that no tuning was required and nonetheless motion states are controlled through several types of maneuvers in Figures 10 and 11. Figure 11 also displays the nature of rudder motions where a single rudder is locked-down (either intentionally or by damage received while underway).
(a) Validating maneuver: tight turn. x(t) on the abscissa and y(t) on the ordinant.    Following this brief demonstration of utility of self-awareness, the next section introduces two paradigms for learning: simple and optimal to allow self-awareness to learn a time-varying selfawareness. Following the technical developments of Section 3.3, the reader will have definitions for all the equations in the topology of Figures 5-9 and will be in a position to attempt their own algorithm development. With that motivation, Section 3.5 of the manuscript will list a procedure to create software to control unmanned underwater vehicles, but then afterwards illustrate the use of the procedure. The self-awareness statement, together with optimal learning, will be shown to reveal the required actuator authority necessary to accomplish an unspecified maneuver duration.

Deterministic Artificial Intelligence Simple-Learning
Start by defining variables and states , , . Then, rewrite the equations of motion isolating the states to be learned: Use classical terms that are proportional to the error and error rate (derivative). The reader familiar with classical proportional, derivative (PD) control [80][81][82] will recognize the approach, but note that we are not formulating a feedback control signal using a PD controller; instead, we are learning the presumed unknown parameters in the deterministic self-awareness statement with proportional and derivative components, and then using that self-awareness statement to formulate the control. Let: Following this brief demonstration of utility of self-awareness, the next section introduces two paradigms for learning: simple and optimal to allow self-awareness to learn a time-varying self-awareness. Following the technical developments of Section 3.3, the reader will have definitions for all the equations in the topology of Figures 5-9 and will be in a position to attempt their own algorithm development. With that motivation, Section 3.5 of the manuscript will list a procedure to create software to control unmanned underwater vehicles, but then afterwards illustrate the use of the procedure. The self-awareness statement, together with optimal learning, will be shown to reveal the required actuator authority necessary to accomplish an unspecified maneuver duration.

Deterministic Artificial Intelligence Simple-Learning
Start by defining variables and states m, I z , mx g . Then, rewrite the equations of motion isolating the states to be learned: .
Use classical terms that are proportional to the error and error rate (derivative). The reader familiar with classical proportional, derivative (PD) control [80][81][82] will recognize the approach, but note that we are not formulating a feedback control signal using a PD controller; instead, we are learning the presumed unknown parameters in the deterministic self-awareness statement with proportional and derivative components, and then using that self-awareness statement to formulate the control. Let: It is presumed here that any/all states are knowable by using sensors, inertial navigation, simultaneous location and mapping (SLAM), state observers, Kalman filters, and such. Thus, in Figure 5, full state knowledge is assumed in the feedback signal used for learning.

Deterministic Artificial Intelligence Optimal Learning
Recall from Equations (2) and (5) the constants starting with ¢ i ∀i = 1, 2, . . . 8, ending with a i ∀i = 1, 2, 3, 4 and b i ∀i = 1, 2, 3, 4 require knowledge of mass,m and mass moment of inertia,Î Z . Section 3.3 describes a simple learning approach to discover values ofm andÎ Z . This section seeks to express an optimal learning methodology, eliminating the need for tuning. Solve Equations (1) and (4) for the product mx G . Solving Equation (1) for the product of mass and location of the center of gravity, mx G produces Equation (45): then substitute the result of Equation (45) into Equation (4) to produce Equation (46), a new version of the second equation of the deterministic self-awareness statement, which may be formulated to parameterize with m and I z as the states resulting in Equation (46): Solve Equation (4) for the product of mass and location of the center of gravity, mx G : then substitute the result of Equation (48) into Equation (1) to produce Equation (49), a new version of the first equation of the deterministic self-awareness statement, which may be formulated to parameterize with m and I z as the states: .
Expressing Equations (47) and (50) together in state-variable form (i.e., "state space") yields Equation (51), which may be inverted to solve for the optimal learned values for mass and mass moment of inertia in Equation (52), where the arbitrarily labeled constants are calculated in Equations (47) and (50) using desired states: The location of the center of gravity may then be located by solving Equation (45) with all components known, except x G . This estimate of the location of the center of gravity may then be located by solving Equation (45) with all components known, except x G together with Equation (52) provide the optimal learned values that should replace m, x G and I z in the deterministic self-awareness statements embodied in Equation (37) (2) and (5).

Optimize maneuver Time for the Allowable Maximum Non-Dimensional Force for a Representative Maneuver
One approach to designing maximum-performance maneuvers is to design the command to fit the amount of available control force (illustrated in Table 4 corresponding to Figure 12a). Iterating maneuver times yield performance curves (Figure 12a), allowing engineers to understand how quickly a maneuver can be demanded, while Figure 12b reveals the force required to accomplish the maneuver depicted in Figure 12a, among others. These two key figures produce a design procedure where the available rudder force yields the minimum time maneuver that can be produced by that vehicle configuration. The resulting maneuver duration may be substituted into Equations (19)-(21) in place of ∆t maneuver , producing full-state autonomously generated trajectories. Especially since deterministic artificial intelligence (in its error-optimal learning instantiation) requires no tuning, this single decision is the only thing required to formulate all other constituent analytic relationships elaborated in this manuscript.

Procedures to implement deterministic artificial intelligence as proposed
Section 3.5 revealed an operational implementation procedure with the revelations of Figure  11a,b. This section succinctly describes a nominal procedure for use by operators of unmanned underwater vehicles to wish to apply the proposed deterministic artificial intelligence methods.

Figure 12.
Assertion of deterministic self-awareness with optimal learning: Maneuver duration options and corresponding required maximum force for representative maneuver. For any given control authority provided by available motors, the selection of force in (b) yields ∆t maneuver for Equations (23)-(25).

5.
Use the parameters learned in step 4 to update the constants and constituent constants, repeating step 3 optimal rudder commands

Follow-On Research
This manuscript described the development of the proposed techniques from first principles, articulated optimality, and illustrated step-by-step procedures for algorithm development and operational implementation. Future research includes critical analysis intended to reveal limitations of the technique. Algorithm discretization time step will be iterated to reveal the potential existence of a minimum processor speed required to run the algorithm. Simulated damage will be imposed on the vehicle to ascertain the efficacy of the method's robustness. Parameter variation will be imposed to investigate the algorithms sensitivity and bridge to real-world implementation in lab and open ocean experiments. Disturbances (e.g., currents and wave action) will be added to reveal the method's dynamic stiffness. It is noteworthy that any instance of disturbance where a physics-based mathematical description is available, those descriptions can be used to formulate deterministic self-awareness statements to augment those already presented in this manuscript. In essence, the vehicle can become aware of its disturbance environment as part of its self-awareness. The rigid body can know that it is an unmanned underwater vehicle in a specifically disturbed environment. Consideration will be given in future research to a reverse design procedure that permits the algorithm to tell the operator which actuator to purchase and install to accomplish some known maneuver driven by a trajectory tracking task given weigh-point guidance (not addressed here). An additional interesting procedure includes underactuated maneuvers that are nonetheless demanded by operational imperatives.
Lastly, experimentation should include comparison to existing solutions, particularly those described in the introduction of this manuscript (i.e., typical stochastic A.I. methods and other analytic methods articulated in Section 1). Experimentation is contingent upon further research funding.

Conflicts of Interest:
The author declares no conflict of interest. No personal circumstances or interest may be perceived as inappropriately influencing the representation or interpretation of reported research results. The funding sponsors had absolutely no role in the design of the study; in the collection, analyses, or interpretation of data; or in the writing of the manuscript, but did require publication or presentation of the results. ψ

Appendix A
Heading angle and rate x, y Cartesian position coordinates r Turning rate (yaw) δ s , δ * s Deflection of stern rudder and its optimal variant δ b , δ * b Deflection of bow rudder and its optimal variant Sway force coefficients: coefficients describing sway forces from resolved lift, drag, and fluid inertia along body lateral axis. These occur in response to individual (or multiple) velocity, acceleration, and plane surface components, as indicated by the corresponding subscripts N δs , N δ b ,N r , N . r ,N ν , N . ν Yaw moment coefficients  Learned mass moment of inertia and initial value K m1 Control gain for mass simple-learning K mxG1 Control gain for learning product of mass and location of center of mass K I1 , K I2 Control gain for learning mass moment of inertia Variables (combinations of motion states) used to reparametrize problem into optimal learning form