Next Article in Journal
Applications of Entropy in Data Analysis and Machine Learning: A Review
Previous Article in Journal
Shannon Entropy Analysis of a Nuclear Fuel Pin Under Deep Burnup
Previous Article in Special Issue
Probabilistic PARAFAC2
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ornstein–Uhlenbeck Adaptation as a Mechanism for Learning in Brains and Machines

by
Jesús García Fernández
,
Nasir Ahmad
and
Marcel van Gerven
*
Department of Machine Learning and Neural Computing, Donders Institute for Brain, Cognition and Behaviour, Radboud University, 6500HB Nijmegen, The Netherlands
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(12), 1125; https://doi.org/10.3390/e26121125
Submission received: 24 October 2024 / Revised: 19 December 2024 / Accepted: 20 December 2024 / Published: 22 December 2024

Abstract

:
Learning is a fundamental property of intelligent systems, observed across biological organisms and engineered systems. While modern intelligent systems typically rely on gradient descent for learning, the need for exact gradients and complex information flow makes its implementation in biological and neuromorphic systems challenging. This has motivated the exploration of alternative learning mechanisms that can operate locally and do not rely on exact gradients. In this work, we introduce a novel approach that leverages noise in the parameters of the system and global reinforcement signals. Using an Ornstein–Uhlenbeck process with adaptive dynamics, our method balances exploration and exploitation during learning, driven by deviations from error predictions, akin to reward prediction error. Operating in continuous time, Ornstein–Uhlenbeck adaptation (OUA) is proposed as a general mechanism for learning in dynamic, time-evolving environments. We validate our approach across a range of different tasks, including supervised learning and reinforcement learning in feedforward and recurrent systems. Additionally, we demonstrate that it can perform meta-learning, adjusting hyper-parameters autonomously. Our results indicate that OUA provides a promising alternative to traditional gradient-based methods, with potential applications in neuromorphic computing. It also hints at a possible mechanism for noise-driven learning in the brain, where stochastic neurotransmitter release may guide synaptic adjustments.

1. Introduction

One of the main properties of any intelligent system is that it has the capacity to learn. This holds for biological systems, ranging from bacteria and fungi to plants and animals [1,2,3,4], as well as for engineered systems designed by artificial intelligence (AI) researchers [5,6,7]. Modern intelligent systems, such as those used in machine learning, typically rely on gradient descent for learning by minimizing error gradients [8,9,10]. While gradient-based methods have driven significant advances in AI [6], their reliance on exact gradients, centralized updates, and complex information pathways limits their applicability in biological and neuromorphic systems.
In contrast, biological learning likely relies on different mechanisms, as organisms often lack the exact gradient information and centralized control that gradient descent requires [11,12]. Neuromorphic computing, inspired by these principles, aims to replicate the distributed, energy-efficient learning of biological systems [13,14]. However, integrating traditional gradient-based methods into neuromorphic hardware has proven challenging, highlighting a critical gap: the need for gradient-free learning mechanisms that exclusively rely on operations that are local in space and time [15,16].
To address this, alternative learning principles to gradient descent have been proposed for both rate-based [17,18,19,20,21] and spike-based models [22,23,24,25]. A class of methods that leverages inherent noise present in biological systems to facilitate learning is perturbation-based methods [26,27,28], which adjust the system’s parameters based on noise effects and global reinforcement signals, offering gradient-free, local learning suitable for biological or neuromorphic systems. Specifically, these methods inject random fluctuations (noise) into the system and evaluate the impact of this noise on performance through a reinforcement signal. Then, they adjust the system’s parameters to increase or decrease alignment with the output of the system featuring noise, depending on the feedback provided by the reinforcement signal. Within this framework, node perturbation methods [28,29,30,31,32,33,34] inject noise into the nodes, while weight perturbation methods [35,36,37,38,39] inject noise into the parameters. Particularly, node perturbation has been shown to approximate gradient descent updates on average [28], but it relies on a more stochastic exploration of the loss landscape rather than consistently following the steepest descent direction, as in gradient-based methods. Nevertheless, both approaches require generating two outputs per input—one noisy and one noise-free—to accurately measure the impact of the noise on the system, and access to the noise process, making them impractical in many real-world or biological scenarios. Reward-modulated Hebbian learning (RMHL) [40,41], a bio-inspired alternative, overcomes some of these limitations as it does not need a noiseless system nor access to the noise process to learn. Yet, RMHL often struggles to solve practical tasks efficiently, limiting their applicability.
In this work, we propose Ornstein–Uhlenbeck Adaptation (OUA), a novel learning mechanism that extends the strengths of perturbation-based methods while addressing some of their limitations. Like RMHL methods, it does not require a noiseless system or direct access to the noise process to learn. Differently, OUA introduces a mean-reverting Ornstein–Uhlenbeck (OU) process [42,43] to inject noise directly into system parameters, adapting their mean based on a global modulatory reinforcement signal. This signal is derived from deviations in error predictions, resembling a reward prediction error (RPE) thought to play a role in biological learning [44]. Unlike traditional methods, OUA operates in continuous time using differential equations, making it particularly suited for dynamic, time-evolving environments [45]. While RMHL was developed to model biological learning dynamics, OUA’s online nature and ability to adapt continually offer advantages in practical applications. Additionally, the simple and local nature of the proposed mechanism makes it well-suited for implementation on neuromorphic hardware.
We validate our approach across various experiments, including feedforward and recurrent systems, covering input–output mappings, control tasks, and a real-world weather forecasting task, all within a continuous-time framework. Additionally, we demonstrate that the method can be extended to a meta-learning setting by learning the system hyper-parameters. Finally, we discuss the implications of OUA, as our experiments demonstrate its efficiency and versatility, positioning it as a promising approach for both neuromorphic computing and understanding biological learning mechanisms.

2. Methods

2.1. Inference

Consider an inference problem, where the goal is to map inputs x ( t ) R m to outputs y ( t ) R k for t R + . In a continual learning setting, this problem can be formulated as a stochastic state space model
d z ( t ) = f θ ( z ( t ) , x ( t ) ) d t + d ζ ( t )
y ( t ) = g θ ( z ( t ) , x ( t ) ) + ϵ ( t )
Here, θ R n are (learnable) parameters, z ( t ) R d is a latent process, f θ ( · ) and g θ ( · ) are nonlinear functions parameterized by θ , ζ ( t ) is process noise, and  ϵ ( t ) is observation noise. We assume that we integrate the process from an initial time t 0 up to a time horizon T. We also refer to Equation (1) as a (latent) neural stochastic differential equation [46], where ‘neural’ refers to the use of learnable parameters θ . The dependency structure of the variables in OUA are depicted in Figure 1.

2.2. Reward Prediction

To enable learning, we assume the existence of a global scalar reward signal r ( t ) , providing instantaneous feedback on the efficacy of the system’s output y ( t ) . The goal is to adapt the parameters θ to maximize the cumulative reward
G ( t ) = t 0 t r ( τ ) d τ
as t T . This can be expressed as an update equation G ( t ) = r ( t ) d t with initial state G 0 = G ( t 0 ) = 0 . We also refer to the expected cumulative reward as the return. To facilitate learning, the system maintains a moving average of the reward [47] according to
d r ¯ ( t ) = ρ ( r ( t ) r ¯ ( t ) ) d t .
This is equivalent to applying a low-pass filter to the reward with time-constant 1 / ρ . We also refer to the difference δ r ( t ) = r ( t ) r ¯ ( t ) as the reward prediction error, which can be interpreted as a global dopaminergic neuromodulatory signal, essential for learning in biological systems [44].

2.3. Learning

Often, learning is viewed as separate from inference. Here, in contrast, we cast learning and inference as processes that co-evolve over time by making the parameters part of the system dynamics. That is, in Equations (1) and (2), we assume that θ ( t ) evolves over time in parallel to the other variables. In this sense, the only distinction between learning and inference is that the former is assumed to evolve at a slower time scale compared to the latter.
The question remains of how to set up learning dynamics such that the parameters adapt towards a more desirable state. To this end, we define learning as a stochastic process evolving forward in time. Specifically, let us assume that parameter dynamics are given by an Ornstein–Uhlenbeck process
d θ ( t ) = λ ( μ ( t ) θ ( t ) ) d t + Σ d W ( t )
with μ ( t ) the mean parameter, λ the rate parameter, Σ = diag ( σ 1 , , σ n ) the diffusion matrix, and  W ( t ) = ( W 1 ( t ) , , W n ( t ) ) a stochastic process, which we take here to be a multivariate Wiener process. The Wiener process introduces stochastic perturbations, characterized by normally distributed increments with zero mean and variance proportional to dt, providing a mathematical model of random noise.
The OU process balances two key forces: (i) stability through mean-reversion via the term λ ( μ ( t ) θ ( t ) ) and (ii) exploration through stochastic noise, as the term Σ d W ( t ) introduces randomness, allowing the system to explore the parameter space. Together, these terms embody the classical exploration–exploitation dilemma at the level of individual parameters. Exploration is driven by stochastic perturbations, which allow the parameters to sample a broad range of values. Exploitation, on the other hand, is guided by the mean-reverting force that nudges the parameters toward favorable regions identified by the current estimate of μ ( t ) .
If we were to run Equation (5) in isolation, the parameters would simply fluctuate around the mean μ ( t ) , with no directed learning. To enable adaptation, we define the mean parameter dynamics using an ordinary differential equation:
d μ ( t ) = η δ r ( t ) ( θ ( t ) μ ( t ) ) d t
where η is the learning rate, and  δ r ( t ) is the RPE, acting as a global modulatory reinforcement signal. This reinforcement mechanism dynamically adjusts μ ( t ) , shifting the focus of exploration towards regions associated with higher reward. The term ( θ ( t ) μ ( t ) ) ensures that the adaptation of μ reflects the influence of recent stochastic updates to θ . The interplay between the stochastic term and the mean-reversion term results in probabilistic convergence behavior. As  μ is refined through updates driven by δ r , the system converges to parameter values θ ( T ) N ( μ ( T ) , C ) with mean μ ( T ) and stationary covariance C = Σ Σ 2 λ .
Since Equation (5) defines learning dynamics in terms of an Ornstein–Uhlenbeck process, we refer to our proposed learning mechanism as Ornstein–Uhlenbeck adaptation. OUA combines stochastic exploration with adaptive exploitation, making it particularly well-suited for continual learning in dynamic, time-evolving environments, such as those encountered in neuromorphic systems.

2.4. Experimental Validation

To test OUA as a learning mechanism, we designed experiments across several distinct scenarios. First, we analyzed the learning dynamics using a single-parameter model to gain insight into fundamental behavior. Subsequently, we examined recurrent and multi-parameter models to explore interactions among parameters and assess scalability. We then applied OUA to a real-world weather prediction task, forecasting temperature 24 h ahead based on current measurements of temperature, humidity, wind speed, wind direction (expressed as sine and cosine components), and atmospheric pressure. The dataset used in this task contains hourly recordings for Szeged, Hungary, collected between 2006 and 2016. Data can be obtained from https://www.kaggle.com/datasets/budincsevity/szeged-weather/ (accessed on the 20 December 2024). Outliers were removed using linear interpolation, and data were either standardized or whitened prior to further processing. To further assess OUA, we tackled a control problem known as the stochastic double integrator (SDI), where the objective was to maintain a particle’s position and velocity near zero despite the effects of Brownian motion. We refer to this control task as the stochastic double integrator (SDI) problem [48]. Finally, we extended our investigation to meta-learning, testing OUA’s ability to adapt hyper-parameters dynamically.
To implement OUA, we rely on numerical integration. The learning process is governed by a coupled system of stochastic and ordinary differential equations, defined in Equations (1), (2), and (4)–(6). Specifically, Equations (1) and (2) describe inference, while Equations (4)–(6) capture the learning dynamics. To integrate the system from the initial time t 0 to the time horizon T, we used an Euler–Heun solver implemented in the Python Diffrax package [49]. In each experiment, the step size for numerical integration was set to Δ t = 0.05 . To interpolate inputs for the weather prediction task across the time window of interest, we used cubic Hermite splines with backward differences [50]. To ensure reproducibility, all scripts needed to replicate the results presented in this study are available via https://github.com/artcogsys/OUA (accessed on the 20 December 2024).

3. Results

In the following, we demonstrate OUA-based learning in increasingly complex systems. Here, we suppress the time index t from our notation to reduce clutter.

3.1. Learning a Single-Parameter Model

To analyze learning dynamics, we begin with a non-linear model containing a single learnable parameter θ . We assume that
y = g θ ( x ) = tanh ( θ x )
with input x and output y, and no latent state z is used. The input is given by a sinusoidal signal x ( t ) = sin ( 0.1 t ) and the reward is given by r = ( y y * ) 2 . The target output is given by y * = tanh ( θ * x ) , which is generated by a ground-truth parameter θ * = 1 . Thus, this setup focuses on supervised learning for a non-linear input–output mapping. The learning dynamics for the single-parameter model are described by the following stochastic differential equations:
d θ = λ ( μ θ ) d t + σ d W
d μ = η δ r ( θ μ ) d t
where r ¯ is the expected reward, δ r = r r ¯ is the RPE and W is a standard Wiener process. Here, θ represents the model parameter and μ represents its mean.
Figure 2 illustrates the learning dynamics of this model, simulated over 15 trials using different random seeds. Figure 2a depicts the target output vs. the model output for the different random seeds. In Figure 2b,c, we observe the trajectories of θ and μ as they converge towards values that allow the model to approximate the target output. Due to the stochastic nature of the dynamics, convergence exhibits probabilistic behavior, leading to variability in θ (and thus μ ) around their optimal values. Nonetheless, OUA ensures the model’s output closely follows the target, even in the presence of continuous noise.
Figure 2d shows how the RPE δ r tends to zero over time. Despite achieving convergence, δ r remains noisy due to the intrinsic stochasticity in the parameters. Figure 2e displays the cumulative reward G over time. The dashed line represents the cumulative reward for an untrained model ( θ = θ 0 ). The results demonstrate that learning significantly improves accumulated reward, with variability in the cumulative reward arising from differences in the time it takes for θ to converge in individual trials.
Analyzing the sensitivity of parameter convergence to hyper-parameter choices provides valuable insights. Figure 3 shows how the choice of hyper-parameters influences the final obtained cumulative reward for the same task (input–output mapping learning). For all hyper-parameters, we see a clear peak in G, except for ρ since the true average reward is equal to the initial estimate of r ¯ 0 = 0 . Even for high noise levels σ , we still observe effective learning.

3.2. Learning a Multi-Parameter Model

Having demonstrated the feasibility of learning with a single parameter, we now investigate whether effective learning extends to cases involving multiple parameters θ = ( θ 1 , , θ n ) . For this, we assume a model given by
y = g θ ( x ) = tanh ( θ x )
where x = ( x 1 , , x n ) is the input vector and y is the scalar output. The input x is composed of multiple sine waves, defined as x i ( t ) = sin ( i , 0.1 , t + ( i 1 ) 2 π / n ) for 1 i n . The reward is given by r = ( y y * ) 2 with target output y * = tanh ( ( θ * ) x ) , generated by ground-truth parameters θ * = ( 0.3 , 1.1 , 0.0 , 0.3 , 1.5 , 0.4 ) . Finally, learning dynamics are given by
d θ = λ ( μ θ ) d t + d Σ W
d μ = η δ r ( θ μ ) d t
with Σ = σ I and W = ( W 1 , , W n ) a multivariate standard Wiener process.
Figure 4 shows that effective learning can still be achieved when having multiple parameters. Figure 4a depicts the target output vs. the model output. In Figure 4b,c, we observe the trajectories of θ and μ as they converge towards values that allow the model to approximate the target output. Figure 4d shows how the RPE δ r tends to zero over time. As in the previous models, δ r remains noisy after convergence due to the intrinsic stochasticity in the parameters. Figure 4e displays the cumulative reward (G) over time. The dashed line represents the return for an untrained model.
Note that we may also choose to use the final mean μ ( T ) as the parameters estimated after learning. While this diverges from the continual learning setting, it is of importance when deploying trained systems in real-world applications. The orange line in Figure 4d shows that this indeed provides optimal performance.

3.3. Weather Prediction Task

We now apply our approach to a real-world weather prediction task. The input vector x consists of several weather features: current temperature, humidity, wind speed, sine of the wind direction angle, cosine of the wind direction angle, and another humidity measurement. The model aims to predict the temperature, denoted as y, 24 h ahead. The setup follows the structure used in the multiple-parameter analysis, with the main difference being that y * now represents the target temperature 24 h ahead. The predicted temperature is given by y = θ x . To validate test performance, a separate segment of the weather dataset was used. Additionally, motivated by recent work that shows that input decorrelation improves learning efficiency [51,52], we tested the model both with and without applying ZCA decorrelation [53] to the input features.
Figure 5a–c shows the learning dynamics for this task. Figure 5d shows that, during training, the model learns to improve the cumulative reward. Figure 5e shows a scatter plot comparing the true versus predicted 24-h ahead temperature before training (gray) and after training (orange). The final mean values μ ( T ) were used as the parameters θ ( t ) when performing inference on separate test data. Figure 5f shows these final mean values, indicating that the current temperature mostly determines the prediction outcome. Figure 5d–f show results over ZCA-decorrelated data. As summarized in Table 1, the model achieves accurate predictions, with performance comparable to stochastic gradient descent (SGD), which was included as a baseline in this experiment. The OUA results were generated using the mean parameters as the model parameters, i.e., θ = μ .

3.4. Learning in Recurrent Systems

While the previous analysis explored learning a static input–output mapping, we now examine the learning dynamics of a non-linear recurrent system, given by
d z = ( f ( θ 1 z + θ 2 x ) z ) d t y = θ 3 z
where x is the input, z is the latent state, and y is the output. This model can be interpreted as a continuous-time recurrent neural network (CTRNN) or a latent ordinary differential equation (ODE). The reward is given by r = ( y y * ) 2 , where y * is the target output generated using fixed target parameters θ * = ( θ 1 , θ 2 , θ 3 ) = ( 0.3 , 0.7 , 1.0 ) . This setup requires fast dynamics in z and slow dynamics in θ , creating a challenging learning scenario. We demonstrate OUA’s capability to learn the parameters of a recurrent model by training it on a 1D input–output mapping.
Figure 6 illustrates the learning dynamics of this three-parameter recurrent system, which includes connections from the input to the latent state, recurrent dynamics within the latent state, and connections from the latent state to the output. Figure 6a depicts the target output vs. the model output for the different random seeds. In Figure 6b,c, we observe the trajectories of θ and μ as they converge towards values that allow the model to approximate the target output. We can observe the same probabilistic convergence behavior described in the non-recurrent model. Figure 2d shows how the RPE δ r tends to zero over time. As in the non-recurrent model, δ r remains noisy after convergence due to the intrinsic stochasticity in the parameters. Figure 2e displays the cumulative reward G over time. The dashed line represents the return for an untrained model.

3.5. Learning to Control a Stochastic Double Integrator

Next, we explore the task of controlling a stochastic double integrator (SDI) [48], which presents a more complex learning environment compared to the supervised learning tasks considered earlier. In this setup, the agent’s actions influence the controlled system, which in turn affects the agent’s observations. This feedback loop can lead to potentially unstable dynamics if not handled properly.
Let s = ( s 1 , s 2 ) represent the state vector, where s 1 and s 2 denote the position and velocity of a particle moving in one dimension. The SDI system is described by the following state-space equations:
d s = 0 1 0 γ s + 0 1 y t + 0 α d W
x = s + β ϵ
where ϵ N ( 0 , I ) . Here, α represents process noise, β represents observation noise, and γ represents a friction term. Note that we again employ an OU process to model the stochastic dynamics of the velocity. The reward is given by a (negative) quadratic cost r = 0.5 | | s | | 2 0.5 y 2 , penalizing deviations of the state s from the set-point, where both the position and velocity are equal to zero, while also penalizing large values of the control y. The agent is defined by y = θ x , with learning dynamics given by Equations (8) and (9), as before. Note that the output y of the agent is the control input to the SDI, whereas the SDI generates the observations x to the agent. Hence, both the agent and the environment are modeled as coupled dynamical systems.
Figure 7 shows that OUA can learn to control a stochastic double integrator, demonstrating effective learning in this more challenging control setting. Both the parameters for the position and the velocity converge to negative values. This is indeed optimal since it induces accelerations that move the particle’s position and velocity to zero.
Figure 7 illustrates the learning dynamics of the stochastic double integrator control task. As shown in Figure 7a–c, the agent learns to adjust the parameters θ and their means μ during the learning process. Figure 7d shows that the return improves over time, with the model achieving a better performance compared to the baseline, where θ is fixed to its initial values θ 0 .
The learning process allows the agent to effectively control the particle, as seen in Figure 7e,f. The dashed lines represent the behavior of the particle without learning, showing a significant deviation of the position from zero. The learned controller, however, successfully adjusts the velocity to drive the position to zero, demonstrating that the agent has effectively learned to stabilize the system. Thus, OUA can successfully learn to control the stochastic double integrator, even in the presence of observation and process noise.

3.6. Meta-Learning

In the previous analyses, we estimated the parameters θ and μ while keeping the hyper-parameters fixed. However, we can also choose to learn the hyper-parameters using the same mechanism. To illustrate this, we introduce a learnable diffusion coefficient σ for the single-parameter model. Rather than fixing σ , we allow it to adjust, which in turn modulates the exploration versus exploitation trade-off in the learning process. This can be viewed as a form of meta-learning, where the adjustment of σ influences the dynamics of learning based on the problem at hand [54].
To implement meta-learning, we define additional dynamics for σ , given by
d σ = λ σ ( μ σ σ ) d t + ρ d W
d μ σ = η σ δ r ( σ μ σ ) d t
where W is a standard Wiener process, analogous to the equations for θ and μ in previous sections (Equations (8) and (9)).
To verify that the meta-learning process effectively identifies the optimal value for the chosen hyper-parameter, in this case, σ , we test the model in a volatile environment. For this, we tackled a simple input–output mapping task, where the target parameter θ * , used to generate the target output, switches from +1 to −1 during learning.
Figure 8 presents the learning dynamics of σ under these volatile conditions. As seen in Figure 8a, the system converges faster to the optimal θ when meta-learning is employed, compared to using a fixed σ . Additionally, Figure 8 demonstrates how σ adapts by increasing its value when θ changes, promoting exploration, and decreasing when the target remains stable, favoring exploitation. The faster convergence is also reflected by the larger return in Figure 8f. Figure 8c,d show that σ and μ σ rapidly adapt to a more suitable higher value of σ , encouraging fast exploration. At a later stage, we observe adaptation to a value of σ below the initial σ 0 value, encouraging exploitation.

4. Discussion

In this paper, we introduced Ornstein–Uhlenbeck adaptation as a learning mechanism for naturally and artificially intelligent systems. Our results show that learning can emerge purely from simulating the dynamics of parameters governed by an Ornstein–Uhlenbeck process, where mean values are updated based on a global reward prediction error. We showed that OUA effectively learns in supervised and reinforcement learning settings, including single- and multi-parameter models, recurrent systems, and meta-learning tasks. Notably, meta-learning via OUA allows the model to balance exploration and exploitation [55,56] automatically. We hypothesize that, in this setting, the remaining stochastic drift of θ around its mean approximates the posterior distribution of θ , similar to Bayesian methods [57].
OUA offers significant advantages for machine learning as it offers a gradient-free learning framework. Unlike backpropagation, no exact gradients, backward passes, or non-local information beyond access to a global reward signal are needed. OUA is also highly parallelizable since multiple agents may simultaneously explore the parameter space using one global parameter mean μ , which may have implications for federated learning [58]. Our experiments highlight the importance of hyper-parameter selection for learning stability (Figure 3), suggesting that meta-learning or black-box optimization techniques, such as Bayesian optimization [59,60] or evolution strategies [61], could further enhance OUA’s performance.
It should be noted that OUA is inherently a stochastic algorithm since noise fluctuations drive learning. This means that, in the limit, instead of converging towards exact parameter values, individual parameter values θ fluctuate about their mean μ with variance σ 2 2 λ . This is also reflected by induced RPE fluctuations around zero. We consider this behavior a feature rather than a bug since it allows the agent to adapt to changing circumstances by continuously probing for better parameter values. This can be seen in Figure 8, where the system responds to volatile target parameters. As also shown in this figure, the meta-learning formulation further allows the system to reduce parameter variance in stable situations. Suppose one does want to enforce fixed parameter values, as shown in Figure 4 and Figure 5, one may choose to set the parameters θ to their estimated mean values μ , or enforce convergence by increasing the rate parameter λ (or decreasing the noise variance σ 2 ) over time, similar to the use of learning rate schedulers in conventional neural network training.
A key area for future work is scaling OUA to more complex tasks. This could involve replacing Wiener processes with other noise processes, such as compound Poisson processes, to selectively update subsets of parameters, potentially reducing weight entanglement. Additionally, as demonstrated in Figure 5, decorrelating the input data via ZCA whitening accelerates convergence [51,52]. Such decorrelation strategies can generalize across deep and recurrent networks [33,34] and even be implemented locally as part of the forward dynamics [52]. OUA is also well-suited for training deep networks, which can be interpreted in our framework as recurrent networks with block-diagonal structures.
OUA might be relevant for biological learning as well since it suggests potential mechanisms for noise-based learning in the brain [62,63]. Here, we draw a connection with the stochastic nature of neurotransmitter release [64,65], which introduces variability in synaptic transmission, akin to the exploratory noise in OUA’s parameter updates. Hypothetically, this stochasticity, paired with global reward signals such as dopaminergic RP [44,66], could guide synaptic weight adjustments dynamically, as modeled by the mean-reverting updates in OUA. Whether or not such noise-based mechanisms are at play in biological learning remains an open question and requires experimental validation. OUA also shares conceptual similarities with reward-modulated Hebbian learning, where the update of the synaptic weight θ i j is a function of the pre- and post-synaptic activity (specifically, the activity deviation from their mean activity) and the RPEs [40]. Formally, while RMHL provides a biologically plausible interpretation of node perturbation [28,30,31,32,33,34], OUA offers a biologically plausible interpretation of weight perturbation [37,38,39]. It is important to note that, while our framework is inspired by biological processes, it operates at a conceptual level rather than by simulating specific neurophysiological mechanisms. Importantly, the conceptual parallels we draw to brain processes serve to provide a high-level understanding of how noise and local reward signals might facilitate efficient learning in biological systems. That is, we do not aim to model detailed neurophysiological processes but to leverage these principles in a computational framework that remains biologically inspired.
Our work is of particular relevance for neuromorphic computing and other unconventional (non-von-Neumann-style) architectures [19,67,68,69,70,71,72], where learning and inference emerge from the physical dynamics of the system [15,73]. These approaches pave the way for sustainable, energy-efficient intelligent systems. Unlike conventional AI, where learning and inference are treated as distinct processes, our method integrates them seamlessly. It relies solely on the online adaptation of parameter values through drift and diffusion terms, which can be directly implemented in physical systems. OUA’s reliance on local, continuous-time parameter updates makes it a natural fit for spiking neuromorphic systems [74], eliminating the need for differentiability or separate forward and backward stages for inference and learning.
Looking ahead, we envision future AI systems where intelligence emerges purely from running a system’s equations of motion forward in time to maximize efficiency and effectiveness. OUA exemplifies how such physical learning machines could be realized through local operations, leveraging the mean-reverting dynamics of the Ornstein–Uhlenbeck process. Future work will focus on theoretical extensions, scaling OUA to high-dimensional tasks, addressing challenges like delayed rewards and catastrophic forgetting, and implementing OUA on neuromorphic and unconventional computing platforms.

Author Contributions

Conceptualization, M.v.G.; methodology, M.v.G. and N.A.; software, M.v.G. and J.G.F.; validation, M.v.G. and J.G.F.; writing—original draft preparation, M.v.G., N.A. and J.G.F.; writing—review and editing, M.v.G., N.A. and J.G.F.; visualization, M.v.G. and J.G.F.; supervision, M.v.G. and N.A.; project administration, M.v.G.; and funding acquisition, M.v.G. All of the authors have read and agreed to the published version of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This publication is part of the DBI2 project (024.005.022, Gravitation), which is financed by the Dutch Ministry of Education (OCW) via the Dutch Research Council (NWO).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The weather data used in this study are openly available in Kaggle at https://www.kaggle.com/datasets/budincsevity/szeged-weather/ (accessed on 20 December 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Fernando, C.T.; Liekens, A.M.; Bingle, L.E.; Beck, C.; Lenser, T.; Stekel, D.J.; Rowe, J.E. Molecular circuits for associative learning in single-celled organisms. J. R. Soc. Interface 2009, 6, 463–469. [Google Scholar] [CrossRef]
  2. Gagliano, M.; Vyazovskiy, V.V.; Borbély, A.A.; Grimonprez, M.; Depczynski, M. Learning by association in plants. Sci. Rep. 2016, 6, 38427. [Google Scholar] [CrossRef]
  3. Money, N.P. Hyphal and mycelial consciousness: The concept of the fungal mind. Fungal Biol. 2021, 125, 257–259. [Google Scholar] [CrossRef]
  4. Sasakura, H.; Mori, I. Behavioral plasticity, learning, and memory in C. Elegans. Curr. Opin. Neurobiol. 2013, 23, 92–99. [Google Scholar] [CrossRef] [PubMed]
  5. Sutton, R.S. Reinforcement Learning: An Introduction; A Bradford Book; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  6. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  7. Brunton, S.L.; Kutz, J.N. Data-Driven Science and Engineering: Machine Learning, Dynamical Systems, and Control; Cambridge University Press: Cambridge, UK, 2022. [Google Scholar]
  8. Linnainmaa, S. The Representation of the Cumulative Rounding Error of an Algorithm as a Taylor Expansion of the Local Rounding Errors. Master’s Thesis, University of Helsinki, Helsinki, Finland, 1970; pp. 6–7. (In Finnish). [Google Scholar]
  9. Werbos, P. Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences. Ph.D. Thesis, Harvard University, Cambridge, MA, USA, 1974. [Google Scholar]
  10. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning Internal Representations by Error Propagation; Technical Report; California Univ San Diego La Jolla Inst for Cognitive Science: La Jolla, CA, USA, 1985. [Google Scholar]
  11. Lillicrap, T.P.; Santoro, A.; Marris, L.; Akerman, C.J.; Hinton, G. Backpropagation and the brain. Nat. Rev. Neurosci. 2020, 21, 335–346. [Google Scholar] [CrossRef]
  12. Whittington, J.C.; Bogacz, R. Theories of error back-propagation in the brain. Trends Cogn. Sci. 2019, 23, 235–250. [Google Scholar] [PubMed]
  13. Mead, C. Neuromorphic Electronic Systems. Proc. IEEE 1990, 78, 1629–1636. [Google Scholar] [CrossRef]
  14. Modha, D.S.; Ananthanarayanan, R.; Esser, S.K.; Ndirango, A.; Sherbondy, A.J.; Singh, R. Cognitive computing. Commun. ACM 2011, 54, 62–71. [Google Scholar] [CrossRef]
  15. Jaeger, H.; Noheda, B.; van der Wiel, W.G. Toward a formal theory for computing machines made out of whatever physics offers. Nat. Commun. 2023, 14, 4911. [Google Scholar] [CrossRef]
  16. Davies, M.; Wild, A.; Orchard, G.; Sandamirskaya, Y.; Guerra, G.A.F.; Joshi, P.; Plank, P.; Risbud, S.R. Advancing neuromorphic computing with loihi: A survey of results and outlook. Proc. IEEE 2021, 109, 911–934. [Google Scholar] [CrossRef]
  17. Oja, E. Simplified neuron model as a principal component analyzer. J. Math. Biol. 1982, 15, 267–273. [Google Scholar] [CrossRef]
  18. Bienenstock, E.L.; Cooper, L.N.; Munro, P.W. Theory for the development of neuron selectivity: Orientation specificity and binocular interaction in visual cortex. J. Neurosci. 1982, 2, 32–48. [Google Scholar] [CrossRef] [PubMed]
  19. Scellier, B.; Bengio, Y. Equilibrium propagation: Bridging the gap between energy-based models and backpropagation. Front. Comput. Neurosci. 2017, 11, 24. [Google Scholar] [CrossRef] [PubMed]
  20. Bengio, Y. How auto-encoders could provide credit assignment in deep networks via target propagation. arXiv 2014, arXiv:1407.7906. [Google Scholar]
  21. Whittington, J.C.; Bogacz, R. An approximation of the error backpropagation algorithm in a predictive coding network with local Hebbian synaptic plasticity. Neural Comput. 2017, 29, 1229–1262. [Google Scholar] [CrossRef]
  22. Markram, H.; Lübke, J.; Frotscher, M.; Sakmann, B. Regulation of synaptic efficacy by coincidence of postsynaptic APs and EPSPs. Science 1997, 275, 213–215. [Google Scholar] [CrossRef] [PubMed]
  23. Bellec, G.; Scherr, F.; Subramoney, A.; Hajek, E.; Salaj, D.; Legenstein, R.; Maass, W. A solution to the learning dilemma for recurrent networks of spiking neurons. Nat. Commun. 2020, 11, 3625. [Google Scholar] [CrossRef]
  24. Haider, P.; Ellenberger, B.; Kriener, L.; Jordan, J.; Senn, W.; Petrovici, M.A. Latent equilibrium: A unified learning theory for arbitrarily fast computation with arbitrarily slow neurons. Adv. Neural Inf. Process. Syst. 2021, 34, 17839–17851. [Google Scholar]
  25. Payeur, A.; Guerguiev, J.; Zenke, F.; Richards, B.A.; Naud, R. Burst-dependent synaptic plasticity can coordinate learning in hierarchical circuits. Nat. Neurosci. 2021, 24, 1010–1019. [Google Scholar] [CrossRef] [PubMed]
  26. Spall, J.C. Multivariate stochastic approximation using a simultaneous perturbation gradient approximation. IEEE Trans. Autom. Control 1992, 37, 332–341. [Google Scholar] [CrossRef]
  27. Widrow, B.; Lehr, M.A. 30 years of adaptive neural networks: Perceptron, madaline, and backpropagation. Proc. IEEE 1990, 78, 1415–1442. [Google Scholar] [CrossRef]
  28. Werfel, J.; Xie, X.; Seung, H. Learning curves for stochastic gradient descent in linear feedforward networks. Adv. Neural Inf. Process. Syst. 2003, 16, 1197–1204. [Google Scholar] [CrossRef] [PubMed]
  29. Flower, B.; Jabri, M. Summed weight neuron perturbation: An O(n) improvement over weight perturbation. Adv. Neural Inf. Process. Syst. 1992, 5, 212–219. [Google Scholar]
  30. Hiratani, N.; Mehta, Y.; Lillicrap, T.; Latham, P.E. On the stability and scalability of node perturbation learning. Adv. Neural Inf. Process. Syst. 2022, 35, 31929–31941. [Google Scholar]
  31. Williams, R.J. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Mach. Learn. 1992, 8, 229–256. [Google Scholar] [CrossRef]
  32. Fiete, I.R.; Seung, H.S. Gradient learning in spiking neural networks by dynamic perturbation of conductances. Phys. Rev. Lett. 2006, 97, 048104. [Google Scholar] [CrossRef]
  33. Dalm, S.; Offergeld, J.; Ahmad, N.; van Gerven, M. Efficient deep learning with decorrelated backpropagation. arXiv 2024, arXiv:2405.02385. [Google Scholar]
  34. Fernández, J.G.; Keemink, S.; van Gerven, M. Gradient-free training of recurrent neural networks using random perturbations. Front. Neurosci. 2024, 18, 1439155. [Google Scholar] [CrossRef]
  35. Kirk, D.B.; Kerns, D.; Fleischer, K.; Barr, A. Analog VLSI implementation of multi-dimensional gradient descent. Adv. Neural Inf. Process. Syst. 1992, 5, 789–796. [Google Scholar]
  36. Lippe, D.; Alspector, J. A study of parallel perturbative gradient descent. Adv. Neural Inf. Process. Syst. 1994, 7, 803–810. [Google Scholar]
  37. Züge, P.; Klos, C.; Memmesheimer, R.M. Weight versus node perturbation learning in temporally extended tasks: Weight perturbation often performs similarly or better. Phys. Rev. X 2023, 13, 021006. [Google Scholar] [CrossRef]
  38. Cauwenberghs, G. A fast stochastic error-descent algorithm for supervised learning and optimization. Adv. Neural Inf. Process. Syst. 1992, 5, 244–251. [Google Scholar]
  39. Dembo, A.; Kailath, T. Model-free distributed learning. IEEE Trans. Neural Netw. 1990, 1, 58–70. [Google Scholar] [CrossRef] [PubMed]
  40. Legenstein, R.; Chase, S.M.; Schwartz, A.B.; Maass, W. A reward-modulated hebbian learning rule can explain experimentally observed network reorganization in a brain control task. J. Neurosci. 2010, 30, 8400–8410. [Google Scholar] [CrossRef]
  41. Miconi, T. Biologically plausible learning in recurrent neural networks reproduces neural dynamics observed during cognitive tasks. eLife 2017, 6, e20899. [Google Scholar] [CrossRef] [PubMed]
  42. Uhlenbeck, G.E.; Ornstein, L.S. On the theory of the Brownian motion. Phys. Rev. 1930, 36, 823–841. [Google Scholar] [CrossRef]
  43. Doob, J.L. The Brownian movement and stochastic equations. Ann. Math. 1942, 43, 351–369. [Google Scholar] [CrossRef]
  44. Schultz, W.; Dayan, P.; Montague, P.R. A neural substrate of prediction and reward. Science 1997, 275, 1593–1599. [Google Scholar]
  45. Kudithipudi, D.; Aguilar-Simon, M.; Babb, J.; Bazhenov, M.; Blackiston, D.; Bongard, J.; Brna, A.P.; Raja, S.C.; Cheney, N.; Clune, J.; et al. Biological underpinnings for lifelong learning machines. Nat. Mach. Intell. 2022, 4, 196–210. [Google Scholar] [CrossRef]
  46. Tzen, B.; Raginsky, M. Neural stochastic differential equations: Deep latent Gaussian models in the diffusion limit. arXiv 2019, arXiv:1905.09883. [Google Scholar]
  47. Wan, Y.; Naik, A.; Sutton, R.S. Learning and planning in average-reward Markov decision processes. In Proceedings of the International Conference on Machine Learning, PMLR, 18–24 July 2021; pp. 10653–10662. [Google Scholar]
  48. Rao, V.G.; Bernstein, D.S. Naïve control of the double integrator. IEEE Control Syst. Mag. 2001, 21, 86–97. [Google Scholar]
  49. Kidger, P. On Neural Differential Equations. Ph.D. Thesis, University of Oxford, Oxford, UK, 2021. [Google Scholar]
  50. Morrill, J.; Kidger, P.; Yang, L.; Lyons, T. Neural controlled differential equations for online prediction tasks. arXiv 2021, arXiv:2106.11028. [Google Scholar]
  51. Ahmad, N.; Schrader, E.; van Gerven, M. Constrained parameter inference as a principle for learning. arXiv 2022, arXiv:2203.13203. [Google Scholar]
  52. Ahmad, N. Correlations are ruining your gradient descent. arXiv 2024, arXiv:2407.10780. [Google Scholar]
  53. Bell, A.J.; Sejnowski, T.J. The “independent components” of natural scenes are edge filters. Vis. Res. 1997, 37, 3327–3338. [Google Scholar] [CrossRef]
  54. Schmidhuber, J. Evolutionary Principles in Self-Referential Learning. Ph.D. Thesis, Technische Universität München, Munich, Germany, 1987. [Google Scholar]
  55. Feldbaum, A.A. Dual control theory, I. Avtomat. Telemekh. 1960, 21, 1240–1249. [Google Scholar]
  56. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction, 2nd ed.; The MIT Press: Cambridge, MA, USA, 2015. [Google Scholar]
  57. Xu, W.; Chen, R.T.Q.; Li, X.; Duvenaud, D. Infinitely deep Bayesian neural networks with stochastic differential equations. arXiv 2022, arXiv:2102.06559v4. [Google Scholar]
  58. McMahan, H.B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics (AISTATS), PMLR, Fort Lauderdale, FL, USA, 20–22 April 2017; Volume 54, pp. 1273–1282. [Google Scholar]
  59. Snoek, J.; Larochelle, H.; Adams, R.P. Practical Bayesian optimization of machine learning algorithms. Adv. Neural Inf. Process. Syst. 2012, 25, 2951–2959. [Google Scholar]
  60. Shahriari, B.; Swersky, K.; Wang, Z.; Adams, R.P.; De Freitas, N. Taking the human out of the loop: A review of Bayesian optimization. Proc. IEEE 2015, 104, 148–175. [Google Scholar] [CrossRef]
  61. Beyer, H.G.; Schwefel, H.P. Evolution strategies—A comprehensive introduction. Nat. Comput. 2002, 1, 3–52. [Google Scholar] [CrossRef]
  62. Faisal, A.A.; Selen, L.P.J.; Wolpert, D.M. Noise in the nervous system. Nat. Rev.. Neurosci. 2008, 9, 292–303. [Google Scholar] [CrossRef] [PubMed]
  63. Engel, T.A.; Chaisangmongkon, W.; Freedman, D.J.; Wang, X.J. Choice-correlated activity fluctuations underlie learning of neuronal category representation. Nat. Commun. 2015, 6, 6454. [Google Scholar] [CrossRef] [PubMed]
  64. Branco, T.; Staras, K. The probability of neurotransmitter release: Variability and feedback control at single synapses. Nat. Rev. Neurosci. 2009, 10, 373–383. [Google Scholar]
  65. Rolls, E.T.; Deco, G. The Noisy Brain: Stochastic Dynamics as a Principle of Brain Function; Oxford University Press: Oxford, UK, 2010. [Google Scholar]
  66. Schultz, W. Dopamine reward prediction error coding. Dialogues Clin. Neurosci. 2016, 18, 23–32. [Google Scholar] [CrossRef] [PubMed]
  67. Markovic, D.; Mizrahi, A.; Querlioz, D.; Grollier, J. Physics for neuromorphic computing. Nat. Rev. Phys. 2020, 2, 499–510. [Google Scholar] [CrossRef]
  68. Stern, M.; Hexner, D.; Rocks, J.W.; Liu, A.J. Supervised learning in physical networks: From machine learning to learning machines. Phys. Rev. X 2021, 11, 021045. [Google Scholar] [CrossRef]
  69. Nakajima, M.; Inoue, K.; Tanaka, K.; Kuniyoshi, Y.; Hashimoto, T.; Nakajima, K. Physical deep learning with biologically inspired training method: Gradient-free approach for physical hardware. Nat. Commun. 2022, 13, 7847. [Google Scholar] [CrossRef]
  70. López-Pastor, V.; Marquardt, F. Self-learning machines based on Hamiltonian echo backpropagation. Phys. Rev. X 2023, 13, 031020. [Google Scholar] [CrossRef]
  71. Momeni, A.; Rahmani, B.; Malléjac, M.; del Hougne, P.; Fleury, R. Backpropagation-free training of deep physical neural networks. Science 2023, 382, 1297–1303. [Google Scholar] [CrossRef] [PubMed]
  72. Van Doremaele, E.R.W.; Stevens, T.; Ringeling, S.; Spolaor, S.; Fattori, M.; van de Burgt, Y. Hardware implementation of backpropagation using progressive gradient descent for in situ training of multilayer neural networks. Sci. Adv. 2024, 10, 8999. [Google Scholar] [CrossRef] [PubMed]
  73. Kaspar, C.; Ravoo, B.J.; van der Wiel, W.G.; Wegner, S.V.; Pernice, W.H. The rise of intelligent matter. Nature 2021, 594, 345–355. [Google Scholar] [CrossRef] [PubMed]
  74. Shahsavari, M.; Thomas, D.; van Gerven, M.A.J.; Brown, A.; Luk, W. Advancements in spiking neural network communication and synchronization techniques for event-driven neuromorphic systems. Array 2023, 20, 100323. [Google Scholar] [CrossRef]
Figure 1. Dependency structure of the variables that together determine Ornstein–Uhlenbeck adaptation (hyper-parameters not shown). Variables r ¯ , μ , and θ (green) are related to learning, whereas variables z and y (blue) are related to inference. The average reward estimate r ¯ depends on rewards r (red) that indirectly depend on the outputs y generated by the model. The output itself depends on input x (black).
Figure 1. Dependency structure of the variables that together determine Ornstein–Uhlenbeck adaptation (hyper-parameters not shown). Variables r ¯ , μ , and θ (green) are related to learning, whereas variables z and y (blue) are related to inference. The average reward estimate r ¯ depends on rewards r (red) that indirectly depend on the outputs y generated by the model. The output itself depends on input x (black).
Entropy 26 01125 g001
Figure 2. Dynamics of a single-parameter model across 15 random seeds (each color represents a different seed) with ρ = λ = η = 1 and σ = 0.3 . Initial conditions are r ¯ 0 = 1 , θ 0 = 0 and μ 0 = 0 . The target output is generated by a ground-truth parameter θ * = 1 . (a) Target output vs. model output (b) Evolution of θ over time. (c) Evolution of μ over time. (d) RPE δ r over time, shown on a logarithmic axis to better visualize initial convergence. The dotted line denotes zero reward prediction error. (e) Cumulative reward G over time, showing improvement with learning compared to the untrained model (dashed line).
Figure 2. Dynamics of a single-parameter model across 15 random seeds (each color represents a different seed) with ρ = λ = η = 1 and σ = 0.3 . Initial conditions are r ¯ 0 = 1 , θ 0 = 0 and μ 0 = 0 . The target output is generated by a ground-truth parameter θ * = 1 . (a) Target output vs. model output (b) Evolution of θ over time. (c) Evolution of μ over time. (d) RPE δ r over time, shown on a logarithmic axis to better visualize initial convergence. The dotted line denotes zero reward prediction error. (e) Cumulative reward G over time, showing improvement with learning compared to the untrained model (dashed line).
Entropy 26 01125 g002
Figure 3. Sensitivity of the final cumulative reward G ( T ) to model hyper-parameters for the input–output mapping task. Results are averaged over 15 runs using different random seeds. The shaded area represents variability across runs, showing stability across a wide range of hyper-parameter settings. Vertical lines show the chosen hyper-parameter values, and horizontal lines show the return without learning. (a) Impact of λ . (b) Impact of σ . (c) Impact of ρ . (d) Impact of η .
Figure 3. Sensitivity of the final cumulative reward G ( T ) to model hyper-parameters for the input–output mapping task. Results are averaged over 15 runs using different random seeds. The shaded area represents variability across runs, showing stability across a wide range of hyper-parameter settings. Vertical lines show the chosen hyper-parameter values, and horizontal lines show the return without learning. (a) Impact of λ . (b) Impact of σ . (c) Impact of ρ . (d) Impact of η .
Entropy 26 01125 g003
Figure 4. Learning dynamics in a multi-parameter model with ρ = λ = η = 1 and σ = 0.2 . Initial conditions are r ¯ 0 = 1 and θ 0 = μ 0 = 0 . The ground-truth parameters used to generate the target output are θ * = ( 0.3 , 1.1 , 0.0 , 0.3 , 1.5 , 0.4 ) . (a) Target output vs. model output. (b) Evolution of θ = ( θ 1 , , θ 6 ) with individual parameter values denoted by different colors. (c) Evolution of μ = ( μ 1 , , μ 6 ) with individual mean values denoted by different colors. (d) RPE δ r , shown on a logarithmic axis to better visualize initial convergence. The dotted line denotes zero prediction error. (e) Cumulative reward G, showing improvement with learning compared to the untrained model (dashed line). The blue line indicates the cumulative reward during parameter learning. The dashed line denotes the return when θ is fixed to the initial value θ 0 . The orange line indicates the return obtained when we fix parameters to the final mean parameters θ ( t ) = μ ( T ) .
Figure 4. Learning dynamics in a multi-parameter model with ρ = λ = η = 1 and σ = 0.2 . Initial conditions are r ¯ 0 = 1 and θ 0 = μ 0 = 0 . The ground-truth parameters used to generate the target output are θ * = ( 0.3 , 1.1 , 0.0 , 0.3 , 1.5 , 0.4 ) . (a) Target output vs. model output. (b) Evolution of θ = ( θ 1 , , θ 6 ) with individual parameter values denoted by different colors. (c) Evolution of μ = ( μ 1 , , μ 6 ) with individual mean values denoted by different colors. (d) RPE δ r , shown on a logarithmic axis to better visualize initial convergence. The dotted line denotes zero prediction error. (e) Cumulative reward G, showing improvement with learning compared to the untrained model (dashed line). The blue line indicates the cumulative reward during parameter learning. The dashed line denotes the return when θ is fixed to the initial value θ 0 . The orange line indicates the return obtained when we fix parameters to the final mean parameters θ ( t ) = μ ( T ) .
Entropy 26 01125 g004
Figure 5. Dynamics of the weather prediction model whose parameters θ denote the different weather features. The goal is to learn to predict the temperature 24 h ahead. Hyper-parameters: ρ = λ = 1 , σ = 0.05 , and η = 0.1 . Initial conditions: r ¯ = 0 , θ 0 N ( 0 , 10 6 I ) and μ 0 = θ 0 . (a) Evolution of θ = ( θ 1 , , θ 6 ) with individual parameter values denoted by different colors. (b) Evolution of μ = ( μ 1 , , μ 6 ) with individual mean values denoted by different colors. (c) RPE δ r , shown on a logarithmic axis to better visualize initial convergence. The dotted line denotes zero prediction error. (d) Cumulative reward G using ZCA-decorrelated data, showing improvement with learning compared to the untrained model (dashed line) and comparable performance to stochastic gradient descent (green line). The blue line indicates the return during parameter learning. The dashed line denotes the return when θ are fixed to their initial values θ 0 . The orange line indicates the return obtained when we fix parameters to the final mean parameters θ ( t ) = μ ( T ) . (e) Predicted vs. true 24 h-ahead temperature on ZCA-decorrelated test data, using final μ ( T ) and initial θ 0 . (f) Final parameter values for each regressor using ZCA-decorrelated data.
Figure 5. Dynamics of the weather prediction model whose parameters θ denote the different weather features. The goal is to learn to predict the temperature 24 h ahead. Hyper-parameters: ρ = λ = 1 , σ = 0.05 , and η = 0.1 . Initial conditions: r ¯ = 0 , θ 0 N ( 0 , 10 6 I ) and μ 0 = θ 0 . (a) Evolution of θ = ( θ 1 , , θ 6 ) with individual parameter values denoted by different colors. (b) Evolution of μ = ( μ 1 , , μ 6 ) with individual mean values denoted by different colors. (c) RPE δ r , shown on a logarithmic axis to better visualize initial convergence. The dotted line denotes zero prediction error. (d) Cumulative reward G using ZCA-decorrelated data, showing improvement with learning compared to the untrained model (dashed line) and comparable performance to stochastic gradient descent (green line). The blue line indicates the return during parameter learning. The dashed line denotes the return when θ are fixed to their initial values θ 0 . The orange line indicates the return obtained when we fix parameters to the final mean parameters θ ( t ) = μ ( T ) . (e) Predicted vs. true 24 h-ahead temperature on ZCA-decorrelated test data, using final μ ( T ) and initial θ 0 . (f) Final parameter values for each regressor using ZCA-decorrelated data.
Entropy 26 01125 g005
Figure 6. Learning dynamics in a recurrent model with ρ = λ = 1 , η = 50 , and σ = 0.2 . Initial conditions are r ¯ 0 = 0.1 and θ 0 = μ 0 = ( 0.2 , 0.1 , 0.5 ) . The ground-truth parameters used to generate the target output are θ * = ( 0.3 , 0.7 , 1.0 ) . (a) Target output vs. model output. (b) Evolution of θ = ( θ 1 , θ 2 , θ 3 ) in blue, orange and green. (c) Evolution of μ = ( μ 1 , μ 2 , μ 3 ) in blue, orange and green. (d) RPE δ r , shown on a logarithmic axis to better visualize initial convergence. The dotted line denotes zero reward prediction error. (e) Cumulative reward G, showing improvement with learning compared to the untrained model (dashed line). The blue line indicates the return during parameter learning. The dashed line denotes the return when θ are fixed to their initial values θ 0 . The orange line indicates the return obtained when we fix parameters to the final mean parameters θ ( t ) = μ ( T ) .
Figure 6. Learning dynamics in a recurrent model with ρ = λ = 1 , η = 50 , and σ = 0.2 . Initial conditions are r ¯ 0 = 0.1 and θ 0 = μ 0 = ( 0.2 , 0.1 , 0.5 ) . The ground-truth parameters used to generate the target output are θ * = ( 0.3 , 0.7 , 1.0 ) . (a) Target output vs. model output. (b) Evolution of θ = ( θ 1 , θ 2 , θ 3 ) in blue, orange and green. (c) Evolution of μ = ( μ 1 , μ 2 , μ 3 ) in blue, orange and green. (d) RPE δ r , shown on a logarithmic axis to better visualize initial convergence. The dotted line denotes zero reward prediction error. (e) Cumulative reward G, showing improvement with learning compared to the untrained model (dashed line). The blue line indicates the return during parameter learning. The dashed line denotes the return when θ are fixed to their initial values θ 0 . The orange line indicates the return obtained when we fix parameters to the final mean parameters θ ( t ) = μ ( T ) .
Entropy 26 01125 g006
Figure 7. Learning to control a stochastic double integrator. Hyper-parameters: ρ = 2 , λ = 1 , σ = 0.02 , and η = 50 , γ = 0.01 , α = β = 0.005 . Initial conditions are set to zero for all variables. (a) Evolution of θ = ( θ 1 , θ 2 ) . (b) Evolution of μ = ( μ 1 , μ 2 ) . (c) Reward prediction error δ r . (d) Cumulative reward G, where higher is better. Blue line: return during learning; dashed line: return with fixed θ 0 . (e) Particle position after learning (blue) with observation noise (light blue); dashed line: position changes without learning. (f) Particle velocity after learning (orange) with observation noise (light orange); dashed line: velocity fluctuations without learning.
Figure 7. Learning to control a stochastic double integrator. Hyper-parameters: ρ = 2 , λ = 1 , σ = 0.02 , and η = 50 , γ = 0.01 , α = β = 0.005 . Initial conditions are set to zero for all variables. (a) Evolution of θ = ( θ 1 , θ 2 ) . (b) Evolution of μ = ( μ 1 , μ 2 ) . (c) Reward prediction error δ r . (d) Cumulative reward G, where higher is better. Blue line: return during learning; dashed line: return with fixed θ 0 . (e) Particle position after learning (blue) with observation noise (light blue); dashed line: position changes without learning. (f) Particle velocity after learning (orange) with observation noise (light orange); dashed line: velocity fluctuations without learning.
Entropy 26 01125 g007
Figure 8. Results when using a learnable σ (blue line) compared to a constant σ = σ 0 (black line). Parameters are set to ρ = λ = η = 1 , λ σ = 2 , and η σ = 3.0 for the model with learnable σ . The initial conditions are given by r ¯ = θ 0 = μ 0 = 0 and σ 0 = μ 0 σ = 0.15 . (a) Dynamics of θ . (b) Dynamics of μ . (c) Dynamics of σ . (d) Dynamics of μ σ . (e) Dynamics of δ r . (f) Dynamics of G.
Figure 8. Results when using a learnable σ (blue line) compared to a constant σ = σ 0 (black line). Parameters are set to ρ = λ = η = 1 , λ σ = 2 , and η σ = 3.0 for the model with learnable σ . The initial conditions are given by r ¯ = θ 0 = μ 0 = 0 and σ 0 = μ 0 σ = 0.15 . (a) Dynamics of θ . (b) Dynamics of μ . (c) Dynamics of σ . (d) Dynamics of μ σ . (e) Dynamics of δ r . (f) Dynamics of G.
Entropy 26 01125 g008
Table 1. Test performance comparison between SGD and OUA on the weather task. Included metrics are MSE and Pearson correlation using both ZCA-decorrelated data and original data.
Table 1. Test performance comparison between SGD and OUA on the weather task. Included metrics are MSE and Pearson correlation using both ZCA-decorrelated data and original data.
ZCA DataOriginal Data
MSEPearson Corr.MSEPearson Corr.
SGD0.210.8710.210.874
OUA0.220.8710.220.871
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

García Fernández, J.; Ahmad, N.; van Gerven, M. Ornstein–Uhlenbeck Adaptation as a Mechanism for Learning in Brains and Machines. Entropy 2024, 26, 1125. https://doi.org/10.3390/e26121125

AMA Style

García Fernández J, Ahmad N, van Gerven M. Ornstein–Uhlenbeck Adaptation as a Mechanism for Learning in Brains and Machines. Entropy. 2024; 26(12):1125. https://doi.org/10.3390/e26121125

Chicago/Turabian Style

García Fernández, Jesús, Nasir Ahmad, and Marcel van Gerven. 2024. "Ornstein–Uhlenbeck Adaptation as a Mechanism for Learning in Brains and Machines" Entropy 26, no. 12: 1125. https://doi.org/10.3390/e26121125

APA Style

García Fernández, J., Ahmad, N., & van Gerven, M. (2024). Ornstein–Uhlenbeck Adaptation as a Mechanism for Learning in Brains and Machines. Entropy, 26(12), 1125. https://doi.org/10.3390/e26121125

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop