Abstract
In the theory of gene networks, the mathematical apparatus that uses dynamical systems is fruitfully used. The same is true for the theory of neural networks. In both cases, the purpose of the simulation is to study the properties of phase space, as well as the types and the properties of attractors. The paper compares both models, notes their similarities and considers a number of illustrative examples. A local analysis is carried out in the vicinity of critical points and the necessary formulas are derived.
MSC:
34C60; 34D45; 92B20
1. Introduction
In this article, we study Neural Networks, called also Artificial Neural Networks (ANN), and their mathematical models, using ordinary differential equations. The motivation for the study of ANNs came from attempts to understand the principles and organization of the human brain. Understanding came that human brains work differently from digital computers. Their effectiveness comes from high complexity, nonlinear modes of regulation, and parallelism of actions. The elements of the human brain were called neurons.
These elements still perform calculations faster than the fastest digital computers. The human brain is able to perceive information about the environment in the form of images and, moreover, it can process the received information needed for interaction with the environment.
At birth, the human brain has a ready structure for learning which, in familiar terms, is understood as experience. So, the neural network is designed to model the way in which the human brain solves usual problems and performs a particular task. A particular interest in ANN stems from the fact that an important group of neural networks is needed to solve a problem computations through the process of learning. So, following [1], an ANN can generally be imagined as a parallel distributed processor, consisting of separate units, which is able to analyze experimental data and prepare them for use.
Many natural processes involve networks of elements that affect each other following a general pattern of conditions and the updating rules for any elements. Both genomic networks and neuronal networks are of this kind. In mathematical models of networks of both types, the regulatory effect of one element to the outputs of other elements is defined by a weight matrix. Therefore, the models describing the evolution of these networks have a lot in common. But, there are also differences. This paper compares models using systems of ordinary differential equations. To distinguish between these systems, we use the designations GRN system and ANN system. At the same time, we realize that the term ANN system has too general a meaning. An ANN system in the established sense is understood as a network that operates according to certain rules and is focused on performing certain tasks. At the same time, the networks undergo training and thus improve their qualities. This article looks at neural networks from a different point of view. We are interested in the behavior of systems of both types for different forms of interaction of elements. The structure of both systems assumes the presence of attractors that determine future states. The description and comparison of possible attractors for the systems of both types is our result.
ANNs are made up of many interconnected elements. Weighted signals from different elements are received by a separate element and processed. A positive signal is understood as an excitatory connection, while negative one means an inhibitory connection. The received signals are linearly summed and modified by a nonlinear sigmoidal function which is called an activation one. The activation function controls the amplitude of an output. “Each neuron has a sigmoid transfer function, and a continuous positive and bounded output activity that evolves according to weighted sums of the activities in the networks. Neural networks with arbitrary connections are often called recurrent networks” [2]. The dynamics of the continuous time recurrent neural network with n units, can be described by the system of ordinary differential equations (ODE) ([3])
where is the internal state of the i-th unit, is the time constant for the i-th unit, are connection weights, is the input to the i-th unit, and is the response function of the i-th unit. Usually, f is taken as a sigmoidal function. There are particular response functions that are non-negative. For instance, functions were used in [4]. More general cases can be modeled by the system using the function which takes values in the open interval If the recurrent neural networks without input are considered, the system
can be considered.
Applications of Artificial Neural Networks are multiple. They can be used in different fields. These fields can be categorized as function approximations, including time series prediction and modeling; pattern and sequence recognition, novelty detection and sequential decision making; and data processing, including filtering and clustering. For applications in Machine Learning (ML), Deep Learning and related problems, consult the review article [5]. For neuroscience applications and their relation to ML, and machine learning using biologically realistic models of neurons to carry out the computation, consider the review [6]. The problems of pattern recognition by ANNs, including applications in manufacturing industries, were studied and analyzed in the review paper [7]. In the paper [8], the ANN approach is applied for the study of a genetic system.
In this article, we mainly study properties of the mathematical model of a three-dimensional ANN, but part of our results will refer to two-dimensional or, more generally, to n-dimensional networks. In particular, we provide information on the types of possible attractors, and their birth and evolution under changes in multiple parameters. The asymptotic properties of the system are important for prediction of future states. This, in turn, can provide instruments for control and management of the modeling network. We use analytical tools for the study of the phase space and its elements. A set of formulas is obtained for the local analysis near equilibria. The necessary data for the analysis were collected by conducting computational experiments and constructing several examples. A broader study involves examining the model and interpreting the findings for the actual process being modeled. Examples of this approach are the works [9,10].
Let us describe the structure of the paper. The Problem formulation section provides the necessary material for the study. The Preliminary results section describes some basic properties of the main systems of ordinary differential equations. It deals also with technical details concerning nullclines, critical points, local analysis by linearization, and some special cases. The next two sections concern some particular but important cases. The systems possessing critical points of the type focus, and systems exhibiting the inhibition-activation behavior, are treated. Both types of systems can have periodic solutions, and that means that cyclic processes can occur in the modeled network. The system of the special triangular structure is analyzed in Section 6. It is convenient for analysis and the main conclusions can be transferred to systems of arbitrary dimensions. The process of birth of stable periodic trajectories from stable critical points of the type focus is considered in Section 7. The mechanism of the Andronov–Hopf bifurcation is illustrated for two-dimensional and three-dimensional neuronal systems. As a by-product, an example of a 3D system that has three limit cycles is constructed. Some suggestions on the management of neuronal systems are provided in Section 7. The possibility of effectively changing the properties of the system, and therefore to partially controlling the network in question, is emphasized. The last section summarizes the results obtained so far, and outlines further studies in this direction.
2. Problem Formulation
The mathematical model using ordinary differential equations, is
The same system can be written as ([11])
since
The elements of this 3D network are called neurons. The connections between them are synapses (or nerves). There is an algorithm that describes how the impulses are propagated through the network. In the above model, this algorithm is encoded by the matrix
Each neuron accepts signals from others and produces a single output. The extent to which the input of neuron i is driven by the output of neuron j is characterized by its output and synaptic weight The dynamic evolution leads to attractors of the system (4), and it was experimentally observed in neural systems. In theoretical modeling, the emphasis is put on the attractors of a system. We wish to study them for System (4).
Similar systems arise in the theory of genetic regulatory networks. The difference is that the nonlinearity is represented by a positive valued sigmoidal functions. One of such systems is
Notice that System (3), and therefore also System (4), can be obtained from System (6), where by two arithmetic operations, namely multiplication of the nonlinearity in (6) by 2 and subtracting 1. This changes the range of values in (3) to
Systems of the form (6) were studied before by many authors. The interested reader may consult the works ([12,13,14,15,16,17,18,19,20]). Similar systems appear in the theory of telecommunication networks ([21]).
In this article, we study the different dynamic regimes for System (4) which can be observed under various conditions. In particular, we first speak about critical points in System (4) and evaluate the number of them. Then, we focus on periodic regimes, study their attractiveness for other trajectories. This can be performed, under some restrictions, for systems of relatively high dimensionality. Also, the evidences of chaotic behavior are presented.
3. Preliminary Results
This section contains the description of basic properties of systems under consideration, and provides information about nullclines, critical points, and their role in the study.
3.1. Invariant Set
Consider the 3D system (4).
Proposition 1.
System (4) has an invariant set
Proof.
By inspection of the vector field generated by System (4) on the opposite faces of the three-dimensional cube Notice, that the value range for the function tanh z is □
3.2. Nullclines
The nullclines for the system are defined by the relations
Example 1.
Consider the system with the matrix
and .
3.3. Critical Points
The critical points, which are also called the equilibria, can be obtained from System (4). Geometrically, they are the cross points of the nullclines. The nullclines are defined by the relations
Proposition 2.
All critical points are in the invariant set.
The nullclines are located in the sets , respectively, and these sets intersect by the invariant set only.
Proposition 3.
At least one critical point exists.
The invariant set may be considered as a topological ball. Since the vector field on the border is directed inward, is mapped into itself continuously. The continuous contraction mapping to has a fixed point. Any fixed point is a solution of the system (7).
Remark 1.
The number of critical points may be greater, up to 27, but finite.
Remark 2.
Both assertions 2 and 3 are valid for the n-dimensional case also.
Example 2.
Consider System (4) with the matrix
and . There is one critical point .
Example 3.
Consider example of multiple critical points and the system (4) with the matrix
and .
There are three critical points , and .
3.4. Linearization at a Critical Point
Let be a critical point. The linearization around it is given by the system
where
One has
and the characteristic equation for is
where
3.5. Regulatory Matrices With Zero Diagonal Elements
Set The regulatory matrix is
and the system of differential equations takes the form
Let be a critical point. The respective linearized system around it is
where given in (13) to (15), are computed assuming that the regulatory matrix is (18). The characteristic equation for takes the form
where
Equation (21) has the form
Recall the Cardano formulas for Equation (24). This equation has complex roots if
is positive. The complex roots can be obtained as
where
are real cubic roots satisfying The real root of Equation (24) is
Example 4.
Consider System (19) with the matrix
and .
There is a single critical point The characteristic numbers obtained by the linearization process are .
4. Focus Type Critical Points
Consider again Equation (21). In our notation,
Suppose that Let be a critical point in question. The associated characteristic numbers are
where
are the real values of cubic roots, and Q is given by (28). We will call such a critical point 3D-focus. It is unstable if the real part is positive. We arrive at the following assertion.
Proposition 4.
Let be a critical point of the system (19). Suppose that
Then, and this critical point is a 3D-focus.
Corollary 1.
Suppose the condition holds for a critical point. Then, this point is a 3D-focus.
Proof.
The relation (31) is fulfilled if □
Proposition 5.
Suppose is a critical point of type focus of the system (19). This point is an unstable focus if the condition holds.
Example 5.
Consider System (19) with the matrix
and .
Figure 5.
The nullclines for System (19) (—red, —green, —blue).
The system has three critical points: , and at , and . The characteristic numbers are given in Table 1.
Table 1.
The characteristic numbers .
5. Inhibition-Activation
Consider the system
where are negative, are positive.
Let the regulatory matrix be
and There is a single critical point. Introduce
The range of values of is the interval The linearized system is
One can obtain the matrix
and the characteristic equation
The roots of the characteristic equation are
Summing up, we arrive at the following assertion.
Proposition 6.
A critical point of System (33) under the above conditions is 3D-focus; that is, the following is true: there is 2D-subspace with a stable focus and attraction in the remaining dimension.
6. The Case of Triangular Regulatory Matrix
We consider the special case of the regulatory matrix being in triangular form,
Since the presentation for the general case differs little from the three-dimensional case, let us consider the n-dimensional variant. The system of differential equations takes the form
where Suppose that the coefficients take values in the interval
6.1. Critical Points
The critical points of System (43) can be determined from
Since the right sides in (44) are less than unity in modulus, all critical points locate in Due to sigmoidal character of the function the last equation in (44) may have one , two or three roots.
Proposition 7.
There are, at most, three values for in System (44).
Proposition 8.
At most, critical points are possible in System (43).
Proof.
The last equation in (44) may have, at most, three roots, due to the S-shape of the graph to a sigmoidal function on the right side. Consequently, the penultimate equation in (44) may have, at most, roots . In total, there are nine roots. Proceeding in this way, we obtain, at most, roots for the very first equation in (44), and therefore, at most critical points for System (43). Hence, the proof. □
6.2. Linearized System
The linearized system is
where
The values of are positive and not greater than unity. The characteristic values for a critical point are to be obtained from
and
Evidently,
Therefore, the characteristic values for any critical point are real, and the following assertion follows.
Proposition 9.
The triangular system (43) cannot have critical points of type focus.
7. Systems with Stable Periodic Solutions: Andronov–Hopf Type Bifurcations
7.1. 2D Case
We first study the second-order system
where and is the parameter. Choose a k small enough that a unique critical point is a stable focus. Then, increase k until the stable focus turns to unstable one. Then, the limit cycle emerges, surrounding the critical point. This is called Andronov–Hopf bifurcation for 2D systems.
Example 6.
Consider System (52) with the matrix
and .
There is one critical point: the stable focus. If the parameter k increases, the stable focus turns to an unstable one. Then, the limit cycle emerges, surrounding the critical point.
Example 7.
Consider System (52) with the matrix
and .
7.2. 3D Case
Consider now the 3D system with the matrix
where are as in 2D system (52). The second nullcline is defined by the relation
Choose the parameters so that Equation (56) has three roots. Then, the second nullcline is a union of three parallel planes.
8. Control and Management of ANN
First, a citation from [22]: “Models of ANN are specified by three basic entities: models of the neurons themselves–that is, the node characteristics; models of synaptic interconnections and structures–that is, net topology and weights; and training or learning rules—that is, the method of adjusting the weights or the way the network interprets the information it receives”.
In this section, we discuss the problem of changing the behavior of the trajectories of System (4). This may be interpreted as partial control over the system. The system has as parameters the coefficients the values and in the linear part. Properties of the system may be changed by varying any of mentioned.
We would like demonstrate how a system of the form (4) can be modified so that trajectories start to tend to some of indicated attractor. For this, consider the system (4), which has as attractors three limit cycles. This can be performed via three operations: (1) put the entries of the 2D regulatory matrix, which corresponds to 2D system with the limit cycle L, to the four corners of a 3D matrix A; (2) choose the middle element of the 3D matrix A so, that the equation with respect to has exactly three roots (3) set the four remaining values of to zero. Set also to unity. After finishing these preparations, the second nullcline will be three parallel planes , going through Each of these planes will contain the limit cycle. Two side limit cycles will attract trajectories from their neighborhoods. The middle limit cycle will attract only trajectories, lying in the plane
Now, let us solve the problem of control. Let the limit cycle at be conditionally “bad”. The problem is to change the system so that all trajectories in are attracted to the limit cycle which, at the beginning of the process, was in the plane Problems of this kind may arise often. In the paper [20], a similar problem was treated mathematically for genetic networks.
Solution: Change so that the equation has now the unique root near The second nullcline is now the plane, passing near This operation is possible, since the graph of is sigmoidal, and changing means shifting the original plane in both directions. After that, only one attractor (limit cycle) remains. The problem is solved.
In neuronal systems, the parameters express the threshold of a response function f ([4]). In genetic networks, stands for the influence of external input on gene which modulates the gene’s sensitivity of response ([23]). The technique of changing the parameters and thus shifting the nullclines was applied in the work [24] for building the partial control over model of genetic network.
9. Conclusions
Modeling of genetic and neural networks, using dynamical systems, is effective in both cases. The advantage of this approach, compared with other models, is the possibility of following the evolution of modeled networks. Both systems have invariant sets trapping the trajectories. As a consequence, the attracting sets exist. The structure and properties of attractors are important for the prediction of future states of networks. Both systems must have critical points. These points may be attractive (stable) or repelling. The limit cycles are possible in both cases. The attractors, exhibiting sensitivity to the initial data, are possible for three-dimensional GRN and ANN systems. Systems with specific structures can have predictable properties. For instance, the triangular systems cannot have critical points of the focus type. In contrast, the inhibition-activation systems typically have critical points of this type, and can suffer bifurcations of Andronov–Hopf type. Partial control and management are possible for GRN and ANN systems. In particular, some realistically large-sized GRN systems allow for control and management by changing the adjustable parameters. This problem is relevant to modern medicine.
Author Contributions
Writing—review & editing, D.O. and F.S. The authors contributed equally to the creation of this article. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Data Availability Statement
No data created in this article.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Haykin, S. Neural networks. In A Comprehensive Foundation, 2nd ed.; Prentice Hall: Hoboken, NJ, USA, 1999. [Google Scholar]
- Chapeau-Blondeau, F.; Chauvet, G. Stable, Oscillatory, and Chaotic Regimes in the Dynamics of Small Neural Networks with Delay. Neural Netw. 1992, 5, 735–743. [Google Scholar]
- Funahashi, K.; Nakamura, Y. Approximation of dynamical systems by continuous time recurrent neural networks. Neural Netw. 1993, 6, 801–806. [Google Scholar]
- Das, A.; Roy, A.B.; Das, P. Chaos in a three dimensional neural network. Appl. Math. Model. 2000, 24, 511–522. [Google Scholar] [CrossRef]
- Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of deep learning: Concepts, CNN architectures, challenges, applications, future directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef] [PubMed]
- Yamazaki, K.; Vo-Ho, V.K.; Darshan, B.; Le, N. Spiking Neural Networks and Their Applications: A Review. Brain Sci. 2022, 12, 63. [Google Scholar] [CrossRef]
- Abiodun, O.I.; Jantan, A.; Omolara, A.E.; Dada, K.V.; Umar, A.M.; Linus, O.U.; Arshad, H.; Kazaure, A.A.; Gana, U.; Kiru, M.U. Comprehensive Review of Artificial Neural Network Applications to Pattern Recognition. IEEE Access 2019, 7, 158820–158846. [Google Scholar] [CrossRef]
- Vohradský, J. Neural network model of gene expression. Faseb J. 2001, 15, 846–854. [Google Scholar] [CrossRef]
- Kraynyukova, N.; Tchumatchenko, T. Stabilized supralinear network can give rise to bistable, oscillatory, and persistent activity. Proc. Natl. Acad. Sci. USA 2018, 115, 3464–3469. [Google Scholar]
- Ahmadian, Y.; Miller, K.D. What is the dynamical regime of cerebral cortex? Neuron 2021, 109, 3373–3391. [Google Scholar] [CrossRef]
- Sprott, J.C. Elegant Chaos; World Scientific: Singapore, 2010. [Google Scholar]
- Wilson, H.R.; Cowan, J.D. Excitatory and inhibitory interactions in localized populations of model neurons. Biophys. J. 1972, 12, 1–24. [Google Scholar]
- Furusawa, C.; Kaneko, K. A generic mechanism for adaptive growth rate regulation. PLoS Comput. Biol. 2008, 4, e3. [Google Scholar] [CrossRef]
- Alakwaa, F.M. Modeling of Gene Regulatory Networks: A Literature Review. J. Comput. Syst. Biol. 2014, 1, 1. [Google Scholar] [CrossRef]
- Brokan, E.; Sadyrbaev, F. On a differential system arising in the network control theory. Nonlinear Anal. Model. Control. 2016, 21, 687–701. [Google Scholar] [CrossRef]
- Schlitt, T. Approaches to Modeling Gene Regulatory Networks: A Gentle Introduction. In Silico Systems Biology; Methods in Molecular Biology (Methods and Protocols); Humana: Totowa, NJ, USA, 2013; Volume 1021, pp. 13–35. [Google Scholar] [CrossRef]
- Jong, H.D. Modeling and Simulation of Genetic Regulatory Systems: A Literature Review. J. Comput Biol. 2002, 9, 67–103. [Google Scholar] [CrossRef]
- Ogorelova, D.; Sadyrbaev, F.; Sengileyev, V. Control in Inhibitory Genetic Regulatory Network Models. Contemp. Math. 2020, 1, 421–428. [Google Scholar] [CrossRef]
- Sadyrbaev, F.; Ogorelova, D.; Samuilik, I. A nullclines approach to the study of 2D artificial network. Contemp. Math. 2019, 1, 1–11. [Google Scholar] [CrossRef]
- Wang, L.Z.; Su, R.Q.; Huang, Z.G.; Wang, X.; Wang, W.X.; Grebogi, C.; Lai, Y.C. A geometrical approach to control and controllability of nonlinear dynamical networks. Nat. Commun. 2016, 7, 11323. [Google Scholar] [CrossRef]
- Koizumi, Y.; Miyamura, T.; Arakawa, S.I.; Oki, E.; Shiomoto, K.; Murata, M. Adaptive Virtual Network Topology Control Based on Attractor Selection. J. Light. Technol. 2010, 28, 1720–1731. [Google Scholar] [CrossRef]
- Vemuri, V. Artificial Neural Networks in Control Applications. Adv. Comput. 1993, 36, 203–254. [Google Scholar]
- Kozlovska, O.; Samuilik, I. Quasi-periodic solutions for a three-dimensional system in gene regulatory network. WSEAS Trans. Syst. 2023, 22, 727–733. [Google Scholar] [CrossRef]
- Ogorelova, D.; Sadyrbaev, F.; Samuilik, I. On Targeted Control over Trajectories on Dynamical Systems Arising in Models of Complex Networks. Mathematics 2023, 11, 2206. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).