Abstract
Dynamics of neural fields are tools used in neurosciences to understand the activities generated by large ensembles of neurons. They are also used in networks analysis and neuroinformatics in particular to model a continuum of neural networks. They are mathematical models that describe the average behavior of these congregations of neurons, which are often in large numbers, even in small cortexes of the brain. Therefore, change of average activity (potential, connectivity, firing rate, etc.) are described using systems of partial different equations. In their continuous or discrete forms, these systems have a rich array of properties, among which is the existence of nontrivial stationary solutions. In this paper, we propose an estimator for nontrivial solutions of dynamical neural fields with a single layer. The estimator is shown to be consistent and a computational algorithm is proposed to help carry out implementation. An illustrations of this consistency is given based on different inputs functions, different kernels, and different pulse emission rate functions.
1. Introduction
It is known that any small piece of human or animal cortex contains a vast number of neurons. Therefore, a continuum approach in modeling these large ensembles makes sense and was pioneered by the work of Beurle []. This work was designed to accommodate only excitable networks of neurons and was subsequently generalized by Wilson and Cowan [] to include inhibitory neurons as well. Amari [] considered ensembles of neurons when studying pattern formation. Since then, there have been applications and extensions of his work in several directions and the birth of the field of dynamic field theory as byproduct. These extensions have for instance enabled analyses of electroencephalograms [], short-term memory [], visual hallucinations [,], and most recently robotics using dynamics neural fields. Applications to robotics has proven very effective, as shown, for instance, by the works of Bicho, Mallet, and Schöner [], Erlhangen and Bicho [], Erlhangen and Schoner [], and Bicho, Louro, and Erlhagen []. The authors of the latter provided studies in which robots to humans interactions were implemented based on information from Dynamic Neural Fields (DNF). The theoretical aspects started in [,] and are summarized below.
Let be a manifold. In the presence of neurons located a position at time t arranged on L layers, the average potential function is often used to understand the continuous field on the kth layer. is the average membrane potential of the neurons located at position at time t of the kth layer. When , can also be understood as the synaptic input or activation at time t of a neuron at position or direction . It satisfies the Amari equation (see []), which is given as
where is the intensity of the connection between a neuron located at position on the kth layer with a neuron a position y on the lth layer and is the pulse emission rate (or activity) at time t of the neuron located at position on the kth layer. G is often chosen as a monotone increasing function. represents the intensity of the external stimulus at time t arriving on the neuron at position on the kth layer, see Figure 1 below.

Figure 1.
Illustration of the DNF for: single layer (a); and multiple layers (b).
DNFs have also branched out to dynamical systems; for instance, in [], the authors studied a heterogeneous aspect of DNF and found existence of attractors and saddle nodes for solutions of (1). The existence of solutions of DNF is based on fixed point theory using Hammerstein integral equation (see []) such as in []. Now, based on recent developments in recurrent neural networks (RNN), Equation (1) can be discretized using nearly exact discretization schemes (see []) to give rise to discrete dynamical neural fields as
where represents the state of the membrane potential on the neuron at position at time on the lth layer, (where ) is a time scale parameter, is a parameter depending on the time scale and the size of the manifold , are heterogeneous weights representing the connectivity between a neuron at position on the kth layer with a neuron at position on the lth layer, and is the intensity of the external stimulus arriving at the neuron at position at time on the kth layer. We observe that (2) represents a discrete dynamical system. To study the stability analysis of the discrete dynamical system (2), one needs to first find the stationary solutions given as and evaluate the derivative of the map F at these stationary solutions. This is a difficult if not impossible task if we do not know how to estimate the stationary solutions for the DNF. This is one of the main motivating factors behind the current paper.
Moreover, from Elman [], Willams and Zipser [], and, most recently, Durstewitz [], this equation is also a RNN. Therefore, the tools of discrete dynamical systems can be applied non only to single-layer DNFs but also to multiple-layers DNFs, where conditions for stability are well-known.
Another interesting aspect of DNFs is that, if we restrict to the unit circle where , then solutions may exist in the complex unit disk ; with the absence of external stimulus, such solutions would also be solutions of a Dirichlet problem associated with Equation (1) (see []). Indeed, suppose that and let for and , for some complex-valued function G. Consider the Poisson kernel on is defined as , where . From the theory of complex analysis (see []), consider then the complex single-layer Amari equation
Suppose F is a smooth function on or F is a distribution (in a functional sense) on . If a nontrivial stationary solution for this equation exists, then it satisfies the equation
An obvious corollary is that, if the complex function has a fixed point, then a nontrivial solution for the complex Amari Equation (3) without stimulus () when it exists, is a harmonic function in , in that the Laplacian operator applied to V is identically zero. Therefore, as a harmonic function, such a solution may be written as a power series , where the coefficient are to be determined. This would be an interesting aspect of this non trivial solutions of DNF worth investigating akin to the Lotka–Volterra expansion proposed in [].
Most analyses of DNFs focus on their applications and theoretical properties. However, given that kernels often used in practice are either Gaussian, Laplacian, or tangent hyperbolic kernels, and the function G is monotone increasing, there are avenues to also study statistical properties of the DNFs, albeit in specific situations. Indeed, the aforementioned kernels can be thought of as density functions of a random variable Y so that the integrand in Equation (1) can be viewed as the average of the random variable over the manifold . With that understanding at hand, our goal is to use this new statistical paradigm to propose a consistent estimator for nontrivial solutions of DNF. The remainder of the paper is organized as follows. In Section 2, we state the necessary definitions and the main result. In Section 3, we propose a computation algorithm for the implementation of the estimator. In Section 4, we state the technical considerations to be used in the implementation, by proposing other functions G beyond the usual sigmoid function. In Section 5, we perform the Monte Carlo simulations based on different kernel functions. We make concluding remarks in Section 6.
2. Main Results
We observe that non-stationary solutions of (1) are given as
Henceforth, for simplicity sake, given that, at up to a kernel, the stationary solutions would have the same form, we consider a dynamic neural field with a single layer, so that (1) becomes
Let V be a stationary solution of the integro-differential Equation (5). According to Hammerstein [], such a nontrivial solution exists if is symmetric positive definite and G satisfies , where are positive constants. We know that V is defined over the domain as
Definition 1.
The indicator function of the set A is defined as
We recall the definition of a consistent estimator.
Definition 2.
Let random sample from distribution with parameter θ. Let be an estimator of θ for some function h. Then, is said to a consistent estimator for θ if for any real number
Theorem 1.
For a given , suppose Y is a random variable supported on Ω with probability density function distribution . Suppose G is the cumulative distribution function of some random variable U supported on . Then, given positive integers n and m, define
where for and , and are random points from U and Y, respectively.
Then, for , we have that
Proof.
From the Markov inequality, we know that, given , ; therefore, it is enough to prove that, given ,
We have that
Put
Since is the empirical distribution of the random variable U, we have that
and since is density function, we have
We observe that by definition , therefore
Given , by continuity of G, there exists such that
It follows that for any
To finish, we note that
Therefore,
This concludes the proof that is a consistent estimator for V. □
Remark 1.
(1) We observe that depends on the knowledge of , which is not known in general. However, we observe that given in Equation (7) has minimum and a maximum of because has minimum 0 and maximum 1. This also means that although we may not know exactly the value of , we can estimate it to be between and . We can therefore select between and for and the estimate of the nontrivial solution will exist in a small interval of length (bandwidth) . Hence, if the domain Ω is very dense in points ’s, then the nontrivial solution will be a small perturbation of the initial external input, and, if the domain is sparse, then perturbation will be greater. Henceforth, for simplicity sake, we assume that . (2) Another observation is that S depends on the position of the neuron ξ in a single-layer system; however, in a multiple-layer system, it would be reasonable to think of S as depending on both the layer k and the position of the neuron on the layer, that is, .
3. Computational Algorithm
In this section, we have the setting of Theorem 1. From Remark 1, we can use the following algorithm to estimate :
- Step 1: Select positive integers n and m.Here, the experimenter should choose the values of relative to how much computational capabilities ones has, knowing that very large values can lead to a significant slowdown of convergence.
- Step 2: Select from the distribution .Knowing that is a known probability distribution (Gaussian, Laplace, or tangent hyperbolic, see the section below), this should be achievable with relative ease from any software.
- Step 3: Select from the distribution of of U associated with G.As in the previous step, sampling from a known probability distribution should be achievable. However, if G is not given as bounded function between 0, and 1, we can still truncate it adequately to obtain a probability distribution (see Section 4.1).
- Step 4: For , select from a uniform distribution .This step assumes that we have an external stimulus S arriving on the neuron at position given as a function of .
- Step 5: For given , evaluate .In this final step, one can choose different values of to plot the estimator in the space .
We use Step 4 only to evaluate . From the above algorithm, it is clear that the activation function of the neuron at position is the sum of the average of activations of neurons at position and the external stimulus arriving at . Thus, essentially, the function is a perturbation of the function by the quantity that depends on and possibly of parameters of the distribution of random variables U and Y. These parameters play the role of smoothing to compensate from the noise created by small values of (see Section 5).
4. Technical Considerations
In this section, we discuss the choices for the pulse emission rate function G and the connection intensity .
4.1. Pulse Emission Rate Function
We note that Amari considered the dynamic of neural fields with pulse emission function G defined as the sigmoid function. However, the equation still has nontrivial stationary solution even if G is not the sigmoid. In fact, there is a large class of nonlinear functions G for which this is true (see Figure 2 and Table 1 below). For example, the following functions, often used in for training algorithm in artificial neural networks, have been adequately truncated for our purposes. Here, is a positive real number.

Figure 2.
Five different activation functions.

Table 1.
A list of potential pulse emission rate functions one can consider in applications.
Remark 2.
- (1) We observe that the choice of the sigmoid activation function is widely preferred in the literature for its bounded nature, without condition.
- (2) Another reason is the fact that it is also suitable when thes are binary, that is, they may take the value 0 or 1, where 0 represents a non-active neuron at time n and 1 represents an active neuron at time n. In this case,would represent the probability that there is an activity on neuron at positionat time.
- (3) A third reason, which is important in our situation, is that it has an inverse that can be written in close form, unlike many other activation functions sometime used in the artificial neural networks (see, e.g., []) making it easy to generate random numbers from. The other functions would require the use of numerical inversion methods such as the bisection method, the secant method, or the Newton–Raphson method, all of which are computationally intensive (see, e.g., Chapter 4 in []).
4.2. Connection Intensity Function
There are various connection intensities functions (or kernel) that one can choose from. These include the Gaussian kernel introduced above. One could also consider the Laplacian kernel defined as or the hyperbolic tangent kernel , see Figure 3 below for an illustration.

Figure 3.
Illustration of Gaussian, Laplacian, and hyperbolic tangent kernels.
5. Simulations
In each of the simulations below, we used the function as if it were the true solution, just to evaluate for . Using the algorithm above, we then compare the estimates of obtained using our estimator V with V unknown (red curves) and V known (blue curves), using various kernel functions and various external stimulus functions . In all simulations below, we selected . We used Gaussian, Laplacian, and hyperbolic tangent kernels with a sigmoid function G. The value of was set as 1 for the Gaussian and Laplacian kernels.
5.1. Simulation 1: Constant External Stimulus
In this simulation, we illustrate the algorithm above by selection a constant intensity of external stimulus arriving at at point . To check if the algorithm is correct, we select , with true function (see Figure 4, Figure 5 and Figure 6).

Figure 4.
The dotted line represents the input , which from above is similar to the . The kernel is a Gaussian. The maximum estimator (red) has the same patterns as the external stimulus and the estimator (blue) takes the form of the true sine function.

Figure 5.
The dotted line still represents the input , which from above is similar to the . The kernel is Laplacian. The maximum estimator (red) has the same patterns as the external stimulus and the estimator (blue) takes the form of the true sine function, but noisier at low and progressively smoother at high values of .

Figure 6.
The dotted line represents the input , which from above is similar to the . The kernel is a Hyperbolic tangent. The maximum estimator (red) has the same patterns as the external stimulus and the estimator (blue) takes the form of the initial external input, unable to replicate the form of the true since function, even at high values of .
5.2. Simulation 2: Logarithm External Stimulus
In this simulation, we illustrate the algorithm above by selection a constant intensity of external stimulus arriving at at point . To check if the algorithm is correct, we select , with true function (see Figure 7, Figure 8 and Figure 9).

Figure 7.
The dotted line represents the input , which from above is similar to the . The kernel is Gaussian. The maximum estimator (red) has the same patterns as the external stimulus and the estimator (blue) is a distorted version of the original since function, with distortion that is increased around 0, which is caused by the presence of that is in the neighborhood of 0 dominating the sine function that we know is close to 0 in a small neighborhood of 0. However, as we get farther from zero, the influence of external stimulus wanes and the estimator starts to take the shape of the true since function.

Figure 8.
In this case, we use a Laplacian Kernel and we observe a similar pattern as above. However, the estimator is much noisier. There is also a noticeable phase difference between the estimations from a Gaussian kernel and a laplacian kernel.

Figure 9.
In this case, we used hyperbolic tangent kernel and clearly the sine pattern of the true function is never recovered. This suggests that the external input is overwhelming the noise, even after a close look within the interval [5.0,5.1].
5.3. Simulation 3: Exponentially Decaying External Stimulus
In this simulation, we illustrate the algorithm above by selection an intensity of external stimulus arriving at at point as , with true function (see Figure 10, Figure 11 and Figure 12).

Figure 10.
Clearly, with this external inputs, the situation is different from the above cases. In a small neighborhood of 0, we still have the external stimulus dominating; the noise and the estimates (blue) remains between and . However, as we move farther away from 0, takes the shape of but it oscillates between and . This is explained but the fact that is traded with due to the periodic nature of the latter. The estimator on the other hand reproduces the expected pattern is this where a Gaussian kernel is used.

Figure 11.
In this case, the kernel is Laplacian and the observations are the same as above. However, we observe much more noise in the estimator, together with a phase shift.

Figure 12.
In this case, the expected pattern is not reproduced due to the external stimulus dominating and the estimator serving as noise.
5.4. Simulation 4: Mexican Hat True Function
In [], the authors used a Gaussian kernel to obtain an estimate of the nontrivial solution that had the form of a Mexican hat function in the space . Our method differs from theirs in two aspects: First, they assumed that where which implies that . This would restrict us to only independent random variables X and Y with marginals and . We do not make such an assumption because there are many kernels (bivariate functions) that cannot be factored as the product of two marginals. Second, we do not make the assumption that has power series about certain state , so that . This assumption would obviously fail for the Heaviside and Ramp functions. The main reason for the difference is that they were interested in second-order synaptic dynamics, which is not the case here. In this section, we show that our method still yields a comparable estimate even without these assumptions. Indeed, in [], the authors showed that the true solution obtained with a Gaussian kernel in the space has the form of a Mexican hat function. In this simulation, we use our estimator to verify this fact, that is, we set our external stimulus as a Gaussian distribution with mean zero and standard deviation 0.03 as in their case and compare the estimates obtained from the use of a Gaussian, Laplacian, and hyperbolic tangent kernels (see Figure 13).

Figure 13.
The curves are estimates of using the proposed estimator for different kernels, Gaussian (blue), Laplacian (red), and hyperbolic tangent (green) with a mean zero Gaussian external stimulus with standard deviation 0.03. Clearly, all kernels yield the expected pattern: (a) the sigmoid function ; (b) the Heaviside function ; (c) the Ramp function with ; and (d) the hyperbolic tangent inverse function .
5.5. Discussion
(1) The conclusion we can draw from the first simulation is that Gaussian and Laplacian kernels both fare well when the external input is constant. The latter produces noisier outputs at low resolution values and becomes smother at high resolution values .
(2) The major takeaway from the second and third simulations is that the external stimulus can have a significant effect on the estimator, especially near boundary points where there is a significant frequency change between the true value and the external stimulus. In reality, in practical applications, the true value is not known; therefore, a careful choice on the external input is needed if one would like to obtain accurate estimations.
(3) The fourth simulation shows that estimates obtained using a Gaussian kernel are smoother. As for the pulse rate function G, the sigmoid function fares much better for all three kernels used in comparison to other functions.
(4) Ultimately, the point of the first three simulations is the hope to extend this type of estimator beyond solutions of DNF. In fact, Equation (1) can be considered a linear Boltzmann equation with stochastic kernel if, instead of thinking of as the position of a neuron, we think of it as the velocity of a particle and as the velocity distribution overtime. In this case, nontrivial solutions are now solutions of the nonlinear Markov operator where , which by the Hille–Yosida theorem exist. The proposed estimator provides in a sense another way of thinking about this problem using successive approximations (see []).
(5) A drawback of the estimator is that it depends on a great first guest of the solutions at points . However, if enough of these points are selected, one stands a great chance of obtaining a good approximation.
(6) An advantage of this estimator is that, locally, it is a great point estimator of the value of for a given , and given . As mentioned in Section 3, although we may not know the values of needed to evaluate , one way to go around the issue is to select them uniformly from a small interval of length .
(7) We observe that the computational aspect of our algorithm depends on Monte Carlo Simulations. This is not the only way of efficiently achieve this. One may also choose to achieve this using methods such as sparse grids and Bayesian Monte Carlo with appropriate priors on the parameters (see, e.g., [,,,,]).
(8) One other possible use of this estimator is that it can help initialize a RNN algorithm or help find the phase space diagram in a discrete dynamical system with two different layers.
6. Conclusions
In this paper, we propose an estimator for nontrivial solutions of dynamic neural fields. The proposed estimator is shown to be consistent. Moreover, the proposed estimator exists within a small interval depending on the number of points selected in the domain where these nontrivial solutions are defined.
The choice of the kernel, as in previous studies, is shown to be crucial in determining the accuracy of the estimates.
We also show that Gaussian kernels provide the best balance among accuracy, smoothness, and the number of points used. In the space domain, the estimates obtained are visually similar to those obtained using, for instance, Lotka–Volterra series, as in []. The proposed estimator has the advantage that it is simple to implement and may serve as initial guess or initialization when for example using a recurrent neural networks to find nontrivial solutions in the time and space domain. This is particularly important in robotics and to a certain extent in neuroinformatics because it could potentially help with accuracy of movements of robots.
In this paper, we also show how the DNF can be extended to functional and complex analysis, which could further extend theoretical properties of DNFs using techniques from these areas. The proposed estimator in this paper can be used to initialize a discrete dynamical system associated with the DNF.
The present work could be useful for new insights into the connection between DNF and dynamical systems and overall contribute to the literature in these areas and in computational neuroscience.
Funding
This research received no external funding.
Data Availability Statement
The data presented in this study are available on request from the corresponding author. The data are not publicly available due to their randomly generated nature from known distributions.
Acknowledgments
The author thank Corine M. Kwessi for the administrative and technical support that helped conduct this study to its completion. The author would also like to thank the referees for careful reading and useful comments that helped to improve the paper.
Conflicts of Interest
The author declares no conflict of interest.
References
- Beurle, R.L. Properties of a mass of cells capable of regenerating pulses. Philos. Trans. R. Soc. Lond. B 1956, 240, 55–94. [Google Scholar]
- Wilson, H.R.; Cowan, J.D. Excitatory and inhibitory interactions in localized populations ofmodel neurons. Biophys. J. 1972, 12, 1–24. [Google Scholar] [CrossRef]
- Amari, S.I. Dynamics of pattern formation in lateral-inhibition type neural fields. Biol. Cybern. 1977, 27, 77–87. [Google Scholar] [CrossRef] [PubMed]
- Nunez, P.L.N.; Srinivasan, R. Electric Fields of the Brain: The Neurophysics of EEG, 2nd ed.; Oxford University Press: Oxford, UK, 2006. [Google Scholar]
- Camperi, M.; Wang, X.J. A model of visuospatial short-term memory in prefrontal cortex: Recurrent network and cellular bistability. J. Comp. Neurosci. 1998, 4, 383–405. [Google Scholar] [CrossRef] [PubMed]
- Ermentrout, G.B.; Cowan, J.D. A mathematical theory of visual hallucination patterns. Biol. Cybern. 1979, 34, 137–150. [Google Scholar] [CrossRef] [PubMed]
- Tass, P. Cortical pattern formation during visual hallucinations. J. Biol. Phys. 1995, 21, 177–210. [Google Scholar] [CrossRef]
- Bicho, E.; Mallet, P.; Schöner, G. Target representation on an autonomous vehicle with low-levelsensors. Int. J. Robot. Res. 2000, 19, 424–447. [Google Scholar] [CrossRef]
- Erlhangen, W.; Bicho, E. The dynamics neural field approach to cognitive robotics. J. Neural Eng. 2006, 3, R36–R54. [Google Scholar] [CrossRef] [PubMed]
- Erlhangen, W.; Schöner, G. Dynamic field theory of movement preparation. Psychol. Rev. 2001, 109, 545–572. [Google Scholar] [CrossRef] [PubMed]
- Bicho, E.; Louro, L.; Erlhangen, W. Integrating verbal and non-verbal communication in adynamic neural field for human-robot interaction. Front. Neurorobot. 2010, 4, 1–13. [Google Scholar]
- Beim, P.G.; Hutt, A. Attractor and saddle node dynamics in heterogeneous neural fields. EPJ Nonlinear Biomed. Phys. EDP Sci. 2014, 2. [Google Scholar] [CrossRef]
- Hammerstein, A. Nichtlineare Integralgleichungen nebst Anwendungen. Acta Math. 1930, 54, 117–176. [Google Scholar] [CrossRef]
- Djitte, N.; Sene, M. An Iterative Algorithm for Approximating Solutions of Hammerstein IntegralEquations. Numer. Funct. Anal. Optim. 2013, 34, 1299–1316. [Google Scholar] [CrossRef]
- Kwessi, E.; Elaydi, S.; Dennis, B.; Livadiotis, G. Nearly exact discretization of single species population models. Nat. Resour. Model. 2018. [Google Scholar] [CrossRef]
- Elman, J.L. Finding Structure in time. Cogn. Sci. 1990, 14, 179–211. [Google Scholar] [CrossRef]
- Williams, R.J.; Zipser, D. A learning algorithm for continually running fully recurrent neural networks. Neural Comput. 1990, 1, 256–263. [Google Scholar] [CrossRef]
- Durstewitz, D. Advanced Data Analysis in Neuroscience; Bernstein Series in Computational Neuroscience; Springer: Cham, Switzerland, 2017. [Google Scholar]
- Green, R.E.; Krantz, S.G. Function Theory of One Complex Variable; Pure and Applied Mathematics (New York); John Wiley & Sons, Inc.: New York, NY, USA, 1997. [Google Scholar]
- Rudin, W. Real and Complex Analysis; McGraw-Hill: New York, NY, USA, 1987. [Google Scholar]
- Kwessi, E.; Edwards, L. Artificial neural networks with a signed-rank objective function and applications. Commun. Stat. Simul. Comput. 2020. [Google Scholar] [CrossRef]
- Devroye, L. Complexity questions in non-uniform random variate generation. In Proceedings of COMPSTAT’2010; Physica-Verlag/Springer: Heidelberg, Germany, 2010; pp. 3–18. [Google Scholar]
- Lasota, A.; Mackey, M.C. Chaos, Fractals, and Noise, 2nd ed.; Applied Mathematical Sciences; Springer: New York, NY, USA, 1994; Volume 97. [Google Scholar]
- Rasmussen, C.E.; Ghahramani, Z. Bayesian Monte Carlo. In Proceedings of the 15th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 9–14 December 2002; pp. 505–512. [Google Scholar]
- Deisenroth, M.P.; Huber, M.F.; Henebeck, U.D. Analytic Moment-based Gaussian Process Filtering. In Proceedings of the 26th International Conference on Machine Learning (ICML), Montreal, QC, Canada, 14–18 June 2009. [Google Scholar]
- Gerstner, T.; Griebel, M. Numerical integration using sparse grids. Numer. Algorithms 1998, 18, 209. [Google Scholar] [CrossRef]
- Xu, Z.; Liao, Q. Gaussian Process Based Expected Information Gain Computation for Bayesian Optimal Design. Entropy 2020, 22, 258. [Google Scholar] [CrossRef] [PubMed]
- Movaghar, M.; Mohammadzadeh, S. Bayesian Monte Carlo approach for developing stochastic railway track degradation model using expert-based priors. Struct. Infrastruct. Eng. 2020, 1–22. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).