Next Article in Journal
MKF-NET: KAN-Enhanced Vision Transformer for Remote Sensing Image Segmentation
Previous Article in Journal
Lightweight Detection Method of Wheelset Tread Defects Based on Improved YOLOv7
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

From Predictive Coding to EBPM: A Novel DIME Integrative Model for Recognition and Cognition

by
Ionel Cristian Vladu
1,*,
Nicu George Bîzdoacă
2,
Ionica Pirici
3 and
Bogdan Cătălin
4
1
Department of Electromechanics, Environment and Applied Informatics, Faculty of Electrical Engineering, University of Craiova, 200440 Craiova, Romania
2
Department of Mechatronics and Robotics, Faculty of Automation, Computers and Electronics, University of Craiova, 200440 Craiova, Romania
3
Department of Human Anatomy, University of Medicine and Pharmacy of Craiova, 2000349 Craiova, Romania
4
Experimental Research Centre for Normal and Pathological Aging, University of Medicine and Pharmacy of Craiova, 2000349 Craiova, Romania
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(20), 10904; https://doi.org/10.3390/app152010904
Submission received: 10 September 2025 / Revised: 2 October 2025 / Accepted: 3 October 2025 / Published: 10 October 2025
(This article belongs to the Section Robotics and Automation)

Abstract

Featured Application

The DIME model can be applied to the development of adaptive robotic systems that combine ultra-rapid recognition of familiar objects with the ability to operate in unfamiliar or ambiguous environments, thereby optimizing both reaction speed and behavioral robustness.

Abstract

Predictive Coding (PC) frameworks claim to model recognition via prediction–error loops, but they often lack explicit biological implementation of fast familiar recognition and impose latency that limits real-time robotic control. We begin with Experience-Based Pattern Matching (EBPM), a biologically grounded mechanism inspired by neural engram reactivation, enabling near-instantaneous recognition of familiar stimuli without iterative inference. Building upon this, we propose Dynamic Integrative Matching and Encoding (DIME), a hybrid system that relies on EBPM under familiar and low-uncertainty conditions and dynamically engages PC when confronted with novelty or high uncertainty. We evaluate EBPM, PC, and DIME across multiple image datasets (MNIST, Fashion-MNIST, CIFAR-10) and on a robotic obstacle-course simulation. Results from multi-seed experiments with ablation and complexity analyses show that EBPM achieves minimal latency (e.g., ~0.03 ms/ex in MNIST, ~0.026 ms/step in robotics) but poor performance in novel or noisy cases; PC exhibits robustness at a high cost; DIME delivers strong trade-offs—boosted accuracy in familiar clean situations (+4–5% over EBPM on CIFAR-10), while cutting PC invocations by ~50% relative to pure PC. Our contributions: (i) formalizing EBPM as a neurocomputational algorithm built from biologically plausible principles, (ii) developing DIME as a dynamic EBPM–PC integrator, (iii) providing ablation and complexity analyses illuminating component roles, and (iv) offering empirical validation in both perceptual and embodied robotic scenarios—paving the way for low-latency recognition systems.

1. Introduction

Over the past three decades, the Predictive Coding (PC) paradigm has become one of the most influential theoretical frameworks in cognitive neuroscience. It posits that perception arises from a hierarchical process of top-down prediction generation, which is continuously compared with incoming sensory data, with discrepancies propagated as prediction errors [1,2]. The PC model has successfully accounted for phenomena such as directed attention, multimodal integration, and the adaptive flexibility of the cortex, being closely linked to Bayesian mathematical principles and the “free energy principle” hypothesis [3].
Nevertheless, this framework presents both conceptual and practical limitations. First, PC entails a high computational and energetic cost, as it requires the continuous generation of internal predictions even in familiar contexts. Second, it assumes that perception is essentially a hypothesis-testing process, which does not accurately capture the rapid and reflexive responses observed at both neuronal and behavioral levels. Third, a substantial body of neurophysiological evidence suggests that recognition in familiar situations occurs through the direct reactivation of stored engrams, without an intermediate stage of generative simulation [4,5].
To address these limitations, we propose a new model—Experience-Based Pattern Matching (EBPM)—developed in this study (Figure 1). EBPM conceptualizes perception and action as the result of the direct activation of multimodal engrams (visual, auditory, motor, emotional), previously formed through experience. Each engram consists of elementary engram units, such as visual edges, motion fragments, or interoceptive cues, which can be reused and combined into complex networks. In this framework, novel stimuli initially activate these units, which may trigger a full familiar engram or partially activate several engrams. If none surpasses an activation threshold, attentional processes are engaged, and new connections are formed—a process equivalent to biological supervised learning.
The EBPM model presents significant advantages over PC (Figure 2) in familiar contexts: faster recognition, reduced energetic cost, and a natural alignment with neurophysiological data (Hebbian plasticity, lateral inhibition, direct memory reactivation). However, like PC, EBPM also has limitations: low robustness to noisy data and reduced flexibility when confronted with entirely novel stimuli.
Building on the observation that neither PC nor EBPM can fully account for all cognitive phenomena, this paper proposes a novel hybrid model—Dynamic Integrative Matching and Encoding (DIME). DIME integrates the two paradigms: it employs EBPM’s direct matching in familiar and rapid situations but can recruit PC’s generative mechanisms when the context is ambiguous or incomplete. A dynamic control mechanism determines which pathway is dominant, ensuring both speed and efficiency, as well as flexibility and robustness. The general architecture of the DIME model is illustrated in Figure 3, which depicts the integration of EBPM and PC processing streams through a central control module.
The aim of this study is twofold: (i) to present the original EBPM model, together with its neurophysiological and algorithmic foundations and (ii) to define the DIME architecture as an integrative solution between EBPM and PC. This work has an interdisciplinary character, with implications in cognitive neuroscience, artificial intelligence, and robotics. In addition, it proposes a methodological framework for validation through neuroimaging (fMRI, EEG/MEG), artificial neural network simulations, and robotic experiments, thereby providing both theoretical and applied relevance.
Other recognition systems incorporate pattern-matching expert frameworks [6] or hybrid architectures [7] but lack biologically grounded engram reactivation as in EBPM.
In addition to Predictive Coding, alternative frameworks have been proposed to explain recognition and cognition. Prior work on attractor-based memory models and Predictive Coding has argued for complementary roles of fast pattern completion and iterative inference. We build on this view by integrating both into a single decision process. However, these models typically focus either on abstract pattern completion without multimodality or on purely Bayesian formulations without explicit engram reactivation. EBPM therefore differs in scope: it formalizes multimodal engrams and introduces a novelty-driven supervised-like learning process, providing a distinct alternative to both attractor and predictive theories.
See Section 2.3.6 for a method-by-method comparison between EBPM and attractor models, CLS, HTM, SDM, memory-augmented NNs, episodic control, metric-learning, and PC/Active Inference.
While modern deep architectures such as convolutional networks and Transformers achieve state-of-the-art accuracy on benchmarks like CIFAR-10, their objectives differ from ours. EBPM/DIME is not designed to compete with high-capacity deep nets in terms of maximum accuracy but rather to provide a biologically grounded framework that explicitly balances recognition latency and robustness. In this sense, our approach is complementary rather than competitive, emphasizing interpretability and real-time feasibility.
Notation: Throughout this manuscript, we adopt the following notation. Input data are denoted by x , with C classes in total. The embedding produced by EBPM is z = f θ x , and class prototypes are P c R d . Prototype similarity is s c x = cos z , P c . The scalars T E and T P are temperature parameters for EBPM and PC, respectively. In PC, K denotes the number of iterative inference steps. Familiarity is denoted as F ^ 0,1 , and uncertainty is denoted as   U 0,1 . The gain parameter of the controller is γ . τ F denotes a time constant for the temporal integration of familiarity, while Λ is a threshold shift used for calibration. Finally, α 0,1 is the dynamic weighting factor combining EBPM and PC in the DIME model.

2. Materials and Methods

2.1. The Theoretical Framework of the EBPM Model

Similarity, logits, and softmax for EBPM
s c x = z · P c z P c ,   l c E B P M x = T E · s c x ,       p E B P M x = s o f t m a x l E B P M x
where z = f θ x is the embedding of input, x ,   P c R d is the prototype of class c ,   T E is a temperature scaling factor, and p E P B M x denotes the class probability distribution estimated by EBPM.
Familiarity EBPM
F r a w x = max c Familiar s c x ,       F ^ x = c l i p F r a w x Λ + 1 2 , 0,1
where F r a w x is the maximum prototype similarity among familiar classes, Λ is a calibration shift parameter, and F ^ x 0,1 is the normalized familiarity estimate used later by the DIME controller.
The Experience-Based Pattern Matching (EBPM) model, proposed in this paper, describes perception and action as the result of the direct activation of multimodal engrams. Each engram integrates the following:
  • Sensory components (e.g., visual edges and colors, auditory timbres, tactile cues);
  • Motor components (articular configurations, force vectors);
  • Interoceptive and emotional components (visceral state, affective value);
  • Cognitive components (associations, abstract rules).
  • These engrams are constructed through experience and consolidated by Hebbian plasticity and spike-timing-dependent plasticity (STDP) mechanisms [4,8]. They are composed of elementary sub-engrams (feature primitives), such as edge orientations or elementary movements, which can be reused in multiple combinations.
The EBPM Processing Stream:
1.
Stimulus activation—sensory input activates elementary engram units (sub-engrams) such as edges, textures, tones, and motor or interoceptive signals.
2.
Engram co-activation—shared sub-engrams may partially activate multiple parent engrams in parallel.
3.
Scoring and aggregation—each parent engram accumulates an activation score based on the convergence of its sub-engrams.
4.
Winner selection—if a single engram surpasses the recognition threshold, it is fully activated and determines the output; if several engrams are partially active, the strongest one prevails through lateral inhibition.
5.
Novelty detection—if all activation scores remain below the threshold, the system enters a novelty state: no complete engram is recognized.
6.
Attention engagement—attention is mobilized to increase neuronal gain and plasticity, strengthening the binding of co-active sub-engrams into a proto-engram.
7.
Consolidation vs. fading—if the stimulus/event is repeated or behaviorally relevant, the proto-engram is consolidated into a stable engram; if not, the weak connections gradually fade.
Formally, an engram unit (EU) can be defined as a feature vector E U i = s i , m , e i , c i , where s i denotes sensory primitives (edges, colors, timbres), m i denotes motor fragments (joint configurations, force vectors), e i captures interoceptive or affective values, and c i encodes cognitive associations or rules. A full engram is a sparse set E U i , , E U n , stabilized by Hebbian and spike-timing-dependent plasticity (STDP). This formalization differentiates EBPM from generic episodic traces or attractor states by explicitly integrating multimodal dimensions.
Thus, EBPM accounts for rapid recognition in familiar contexts and provides a robust architecture for the storage of multimodal experiences.
This process is illustrated in Figure 1 and is consistent with the principles of Hebbian learning and sparse distributed coding [4,5]. It shows how sensory input activates atomic sub-engrams, which can either fully trigger a stored engram or partially activate multiple candidates. The winner-takes-all mechanism determines the output, while novelty detection engages supervised-like learning.
It is important to emphasize that in EBPM, even novel or weakly familiar stimuli always activate sub-engrams (e.g., edges, textures, tones, and tactile or interoceptive states). These activations may remain below the recognition threshold, being unable to complete any stable engram. In such cases, attention is mobilized, not only to enhance further sensory sampling but also to increase neuronal gain and plasticity, creating temporary proto-engrams by binding co-active sub-engrams according to the Hebbian rule (‘fire together, wire together’). If the stimulus or event is repeated or behaviorally relevant, these proto-engrams undergo consolidation into stable engrams; if the event does not recur, the weak connections gradually fade. Thus, attention serves a dual role: intensifying the analysis of unfamiliar input while simultaneously facilitating the structural formation of new memory programs for future recognition.

2.2. The Theoretical Framework of Predictive Coding (PC)

Iterative update in PC
z k + 1 = z k + η x x ^ k · W
where z k is the latent state at inference step k, x ^ k is the top-down prediction at that step, x is the input, W are the synaptic weights, and η is a learning rate parameter. The number of inference steps is denoted by K.
Final probabilities in PC
p P C x = s o f t m a x l K x
where l K are the logits after K inference steps, and p P C x denotes the final class probability distribution estimated by the PC model.
Predictive Coding (PC) is an established framework according to which the cortex continuously generates top-down predictions, compares them with actual sensory data, and propagates the discrepancies as prediction errors [1,2]. PC has three fundamental elements:
  • Internal generative models (hierarchical representations of the environment);
  • Prediction errors (differences between input and predictions);
  • Adaptive updating (learning through error minimization).
PC provides a good account of the brain’s flexibility in novel environments, its robustness to noise, and its capacity for cross-modal integration. However, its high energetic cost and long latencies make it less suitable for rapid responses.
This architecture is schematically represented in Figure 2, where predictions are transmitted top-down, while errors propagate bottom-up for adjustment. It illustrates the generative top-down predictions, the comparison with incoming sensory data, and the propagation of prediction errors upwards in the cortical hierarchy [1,3].

2.3. Comparison of EBPM and PC

2.3.1. Theoretical Foundation

Over the past two decades, Predictive Coding (PC) has become the dominant framework for explaining perception and cognition. PC posits that the nervous system constructs hierarchical generative models that continuously predict sensory input and update these models through the minimization of prediction errors.
In contrast, the EBPM model proposed here argues that perception and action rely on the direct activation of pre-existing multimodal engrams, without requiring the active generation of predictions. EBPM is grounded in the principle that recognition is determined by the similarity between input and stored patterns; only when the input is insufficient to activate an existing engram do attentional processes and supervised learning come into play.
Thus, the fundamental difference is that PC operates through predictive reconstruction, whereas EBPM operates through direct matching (Table 1).

2.3.2. Processing Dynamics in the Two Models

  • Predictive Coding (PC)
  • The brain generates hierarchical top-down predictions.
  • Sensory input is compared with predictions.
  • Prediction errors are transmitted upward and correct the internal model.
  • Perception arises as the result of minimizing the difference between prediction and input.
  • EBPM
  • The input activates elementary engram units.
  • The elementary engram units simultaneously activate multiple engrams.
  • Through lateral inhibition, the most active engram prevails.
  • If no winner exceeds the threshold, attention is triggered, and a new engram is formed.
  • Perception is the complete reactivation of a pre-existing network, not a reconstruction process.
In EBPM, novelty does not mean absence of activation, but rather fragmentary activation: elementary sub-engrams (edges, textures, sounds, interoceptive states, etc.) respond, yet no parent engram reaches the recognition threshold. This state triggers attention mechanisms, which increase neuronal gain and synaptic plasticity in the relevant circuits. According to the principle, ‘fire together, wire together’, co-active sub-engrams are linked into a proto-engram. If the stimulus/event is repeated (or remains behaviorally relevant), these linkages are consolidated into a stable engram; if not, they remain weak and fade. Thus, attention not only collects additional data but also guides and accelerates the very process of inscribing new engrams.

2.3.3. Comparative Advantages and Limitations

The Predictive Coding (PC) framework describes perception as an iterative process of error minimization between sensory inputs and top-down expectations. In this context, the network dynamically adjusts its internal generative model to predict sensory outcomes, while error units signal mismatches between the predicted and received stimuli. The hierarchical feedback mechanism allows the system to progressively refine its estimates, achieving perceptual stability and adaptive generalization. Table 2 summarizes the structural and computational characteristics of the PC model, highlighting the functional correspondence between biological prediction–error loops and their artificial implementation within the proposed comparative framework.

2.3.4. Divergent Experimental Predictions

The Dynamic Integrative Matching and Encoding (DIME) model extends the EBPM and PC principles by incorporating adaptive control and context-dependent integration between recognition and prediction layers. The system dynamically modulates the balance between bottom-up sensory matching and top-down predictive adjustments, optimizing performance under both familiar and novel conditions. This hybrid mechanism reflects the flexible coupling of biological recognition and executive systems observed in neural circuits. Table 3 presents the comparative architecture and control parameters of the DIME model, emphasizing its hybrid integrative dynamics and real-time adaptive behavior.

2.3.5. Interpretive Analysis

  • EBPM better explains rapid recognition, complex reflexes, and sudden cue-based recall.
  • PC is superior in situations with incomplete or ambiguous data due to its generative reconstruction capability.
  • The two models are therefore complementary, each having its own set of advantages and limitations.
This observation sets the stage for an integrative model (DIME), which combines the speed and efficiency of EBPM with the flexibility of PC.

2.3.6. EBPM vs. Related Theories—Method-by-Method Comparison

Context: EBPM (direct reactivation of multimodal engrams) is defined in Section 2.1 and contrasted with Predictive Coding (PC) in Section 2.2 and Section 2.3; DIME integrates them in Section 2.4. Below, we align EBPM against adjacent frameworks, highlighting (i) what is identical, (ii) what each prior method adds, and (iii) what EBPM adds (Table 4).
The detailed step-by-step implementation of the EBPM process is summarized in Algorithm 1, which illustrates the hierarchical activation, competition, and adaptive plasticity sequence that leads to recognition and novelty gating.
Algorithm 1. EBPM (step-by-step) with prior-art mapping.
1.
Extract low-level primitives (edges, textures, timbre, phonemes).
↳ Links to sparse coding and columnar representations [16,23].
2.
Activate candidate multimodal Engram Units (EUs) from those primitives.
Cell assemblies/neural assemblies concepts [4].
3.
Spread + compete among candidate EUs with lateral inhibition/WTA.
↳ Cortical microcircuit motifs; biased competition in attention [24,25].
4.
Select the winner (fully activated engram) → recognition/action trigger.
Attractor/pattern-completion dynamics [9,11,12].
5.
Familiarity/novelty test (no clear winner → novelty branch).
Salience/attention control; locus-coeruleus NE gain [21,25,26].
6.
Rapid plasticity (Hebb/STDP-like) to form/adjust new or refined engrams.
↳ Hebbian/STDP literature [8].
7.
Bind multimodal facets (sensory–motor–affective–cognitive) within the engram; optional hippocampal indexing for episode-level links.
Indexing theory and process-memory hierarchy [11,27].
8.
Real-time reuse: for familiar inputs, jump straight to Steps 3–4 (no generative loop).
↳ Distinct from PC/Active Inference, which iterate prediction-error loops [1,2,3].

2.3.7. Distinctive Contributions of EBPM (Beyond Attractors/CLS/HTM/PC)

While Section 2.3.6 aligns EBPM with adjacent frameworks, here, we summarize the contributions that are unique or substantially sharpened by EBPM:
  • Multimodal Engram as the Computational Unit: EBPM defines an engram as a compositional binding of sensory, motor, interoceptive/affective, and cognitive facets. This goes beyond classic attractor networks (binary pattern vectors) and metric-learning embeddings by specifying the unit of recall and action as a multimodal assembly.
  • Novelty-First Decision with Explicit Abstention: By design, EBPM yields near-zero classification on truly novel inputs. Partial activation of sub-engrams is treated as a reliable novelty flag—not as a misclassification. This sharp separation of recognition versus novelty is absent from PC’s iterative reconstruction loop and from generic prototype schemes.
  • Attention-Gated Proto-Engram Formation: EBPM assigns a concrete role to attention and neuromodulatory gain (e.g., LC-NE) in binding co-active sub-engrams into a proto-engram during novelty, followed by consolidation. This specifies “when and how new representations form” more explicitly than attractors/CLS/HTM.
  • Real-Time Reuse without Iterative Inference: For familiar inputs, EBPM supports one-step reactivation of stored engrams (winner selection via lateral inhibition), avoiding costly predictive iterations. This explains ultra-fast recognition and procedural triggering in overlearned contexts.
  • Integration Readiness via DIME: EBPM is not an isolated alternative: the DIME controller provides a formal arbitration (α) that privileges EBPM under high familiarity and shifts toward PC under uncertainty, yielding a principled hybrid.

2.4. The DIME Integrative Model

To overcome the limitations of both paradigms, we propose DIME (Dynamic Integrative Matching and Encoding)—a model that combines EBPM’s direct matching with PC’s generative prediction.
Principles:
  • Dual-path processing: sensory input is processed simultaneously through direct engram matching (EBPM) and top-down prediction (PC).
  • Dynamic controller: a central mechanism evaluates the context (level of familiarity, uncertainty, noise) and selects the dominant pathway.
  • Adaptive flexibility: in familiar environments → EBPM becomes dominant; in ambiguous environments → PC is engaged.
Algorithmically, DIME can be formulated as follows:
p x = α x · p E B P M x + 1 α x · p P C x
where p E B P M x and p P C x   denote the probability distributions from EBPM and PC, respectively, 0,1 is the dynamic weighting factor defined by Equation (6)
The adaptive controller α(t) is defined as
α x = σ γ · F ^ x U x
where σ · is the logistic sigmoid, γ is the controller gain, F ^ x is the normalized familiarity, and U x is the normalized uncertainty.
The general architecture of the DIME model is presented in Figure 3, which illustrates the integration of the EBPM and PC processing streams through a central control module [24,28]. Diagram of the dual-pathway system: EBPM (direct engram reactivation) vs. PC (generative prediction), with a dynamic controller weighting both outputs depending on familiarity and uncertainty.

2.5. Methodology of Experimental Validation

Testable Predictions and Analysis Plan

We highlight a set of testable predictions that distinguish EBPM from PC and can be empirically evaluated:
  • H1 (EEG/MEG latency): EBPM-dominant trials (familiar stimuli) yield shorter recognition latencies (≈200–300 ms) compared to PC-dominant trials (≈350–450 ms). We base the ≈200–300 ms range on recent EEG evidence indicating that recognition memory signals emerge from around 200 ms post-stimulus across diverse stimulus types [29].
  • H2 (fMRI representational similarity): EBPM produces higher RSA between learning and recognition patterns in hippocampal/cortical regions; PC shows anticipatory activation in higher areas under uncertainty.
  • H3 (Cue-based recall completeness): Partial cues reinstantiate full engram patterns under EBPM, while PC engages reconstructive dynamics without full reinstatement.
  • H4 (Behavioral/robotic efficiency): Under stable familiarity, EBPM leads to lower inference time and energy cost than PC; under novelty/noise, DIME shifts toward PC to preserve robustness.
  • H5 (NE-gain modulation): Pupil-indexed or physiological proxies of LC-NE correlate with novelty detection and proto-engram formation during EBPM-dominant novelty trials.
These hypotheses are structured to allow empirical falsification and ensure that EBPM/DIME is not merely a descriptive framework but a model with measurable consequences.
To test the EBPM, PC, and DIME hypotheses, we propose a multi-level approach:
  • Neuroimaging:
    • fMRI (representational similarity analysis, RSA) → measures the overlap of patterns between learning and recognition.
    • EEG/MEG → measures recognition latencies (EBPM ≈ shorter; PC ≈ longer).
  • Computational simulations:
    • Artificial neural networks inspired by EBPM vs. PC.
    • Comparison of inference times and computational cost.
  • Robotic experiments:
    • Implementation in visuomotor control.
    • Measurement of reaction times and energy consumption in navigation tasks.
This methodology enables the integrated testing of theoretical hypotheses and the assessment of the practical applicability of the DIME model.
The methodological structure is illustrated in Figure 4: (i) neuroimaging establishes the cortical dynamics associated with EBPM vs. PC; (ii) simulations on artificial neural networks (ANNs) compare execution times and noise robustness for the two models; and (iii) robotic tests assess practical applicability in real-time visuomotor control tasks. Such a combined approach is recommended in the computational neuroscience literature to triangulate and validate the proposed mechanisms [30]. Overview of the three validation layers: (i) neuroimaging (fMRI, RSA, EEG/MEG latency), (ii) ANN simulations, and (iii) robotic implementations.

2.6. ANN Simulation Protocol

Datasets: MNIST, Fashion-MNIST, and CIFAR-10. Familiar classes: 0–7; novel classes: 8–9. Inputs are converted to grayscale and resized to 28 × 28 for MNIST/F-MNIST; CIFAR-10 uses standard 32 × 32 RGB but is projected to the model’s input space as in our code.
Noise: We evaluate four scenarios per dataset: Familiar-Clean, Familiar-Noisy, Novel-Clean, Novel-Noisy. For the Noisy conditions, we add zero-mean Gaussian noise with standard deviation σ relative to pixel values in [0,1], followed by clamping to [0,1]. MNIST/Fashion-MNIST: σ = 0.30. CIFAR-10: σ = 0.15. No additional corruptions are used.
EBPM: A prototype-based encoder maps inputs to a normalized embedding; class logits are cosine similarities to learned class prototypes (familiar classes).
PC (Predictive Coding). An autoencoder with a classifier at the latent layer; inference uses K = 10 gradient steps with step size η = 0.1 to reduce the reconstruction error before classification.
DIME: We use two integration strategies:
(i)
DIME-Prob (probability-level fusion): p = α·p_EBPM + (1 − α)·p_PC, with α driven by a familiarity score calibrated on familiar validation data. On CIFAR-10, we tune {gain = 4.0, thr_shift = 0.75, T_EBPM = 6.0, T_PC = 1.0}.
(ii)
DIME-Lazy (runtime-aware gating, used in robotics): δ = mean(|I_t − I_{t − 1}|). If δ ≤ 0.004 → EBPM-only; if δ ≥ 0.015 → PC-only; otherwise, fuse with fixed α = 0.65.
Metrics: Top-1 accuracy and latency (ms/ex.) measured on the same GPU (Tesla T4; PyTorch 2.8.0 + cu126). We provide all raw TXT/CSV outputs in the Supplementary package (see Data and Code Availability).
Unless otherwise noted, results are reported for a single deterministic seed (SEED = 42) with CuDNN determinism enabled; numbers are point estimates per test set. Multi-seed statistics are left to future work.
Full architecture specifications (layers, activations, optimizer, learning rate schedules, training epochs) are provided in Supplementary Section S1, together with configuration files (*.yaml).
All simulations and analyses were performed using Python 3.10 and PyTorch 2.8.0 + CUDA 12.6 (NVIDIA Corporation, Santa Clara, CA, USA) on a Tesla T4 GPU (Google Colab Pro+ environment). Data preprocessing and visualization were conducted with NumPy 1.26, Matplotlib 3.9, and Pandas 2.2. The robotic experiments were implemented in PyBullet 3.2.5 (Boston Dynamics AI Institute, Waltham, MA, USA). The complete source code and configuration files (*.yaml) are provided in the Supplementary Materials.

2.7. Intuitive Derivation of the Controller

The DIME controller adaptively weighs EBPM and PC based on two intuitive signals: familiarity (F) and uncertainty (U). Familiarity increases when an input strongly matches stored engrams, while uncertainty reflects the degree of mismatch or noise. The control variable α\alphaα can therefore be expressed as a logistic gating function α = σ γ F U , where γ is a sensitivity parameter and σ is the logistic function. For example, in a highly familiar and low-uncertainty case F U ,   α 1 and EBPM dominates. Conversely, when uncertainty is high relative to familiarity, α 1 and PC iterations are prioritized. This formulation provides an intuitive justification for the control equations and links them directly to neurocognitive interpretations of recognition under familiarity vs. novelty.

3. Results

3.1. Preface—Scope of the Simulations (Proof-of-Concept)

We emphasize that the simulations reported in Section 3 are minimal proof-of-concept instantiations designed to demonstrate that the EBPM/DIME framework can be implemented and exercised end-to-end. The core contribution of this paper is the conceptual and architectural proposal, not exhaustive benchmarking. Accordingly, MNIST/Fashion-MNIST/CIFAR-10 and the PyBullet obstacle course serve to illustrate feasibility and the predicted trade-offs: speed and efficiency for EBPM on familiar inputs, robustness for PC under noise/novelty, and adaptive arbitration for DIME. Comprehensive large-scale evaluations and richer robotic tasks are deferred to future work (see Section 4.5).
This section evaluates EBPM, PC, and DIME on MNIST, Fashion-MNIST, CIFAR-10, and a PyBullet obstacle course. We report accuracy, latency (ms/sample or ms/step), and stability across five random seeds. Results are organized to (i) quantify the contribution of each component via ablations (controller ON/OFF; PC iterations K ∈ {0,5,10,20}), (ii) assess statistical significance with paired tests and effect sizes, and (iii) analyze computational complexity through runtime and number of PC invocations. Together, these experiments test the central claim: EBPM enables very low-latency familiar recognition, PC provides robustness under uncertainty, and DIME dynamically balances the two for real-time feasibility.

3.2. ANN Simulations on Familiar vs. Novel Splits

We evaluated EBPM (prototype matching), PC (predictive autoencoder with iterative latent inference, K = 10, η = 0.1), and DIME (fusion of EBPM and PC) on MNIST, Fashion-MNIST, and CIFAR-10 under four scenarios: Familiar-Clean, Familiar-Noisy (Gaussian σ as in the code), Novel-Clean, and Novel-Noisy. Familiar classes were 0–7; novel were 8–9.
Noise parameterization: For completeness, the Noisy conditions use zero-mean Gaussian noise with σ relative to [0,1] pixel values, followed by clamping: MNIST/Fashion-MNIST σ = 0.30; CIFAR-10 σ = 0.15.
  • Metrics: Top-1 accuracy, latency (ms/ex.), and for DIME, the mean controller weight is α ¯ where applicable.
  • These results support the complementarity… (see Section 2.6).
MNIST: EBPM is near-optimal on familiar inputs but, by design, does not classify novel digits (≈0%). This reflects its novelty-flagging mechanism: sub-engrams are activated but do not complete a full engram, signaling unfamiliarity rather than misclassifying. PC is slower but generalizes better to novelty, while DIME adapts by following EBPM on familiar inputs and engaging PC under uncertainty.
Key numbers from our runs:
  • EBPM—Familiar-Clean: 98.63%, 0.100 ms/ex; Familiar-Noisy: 97.13%, 0.132 ms/ex; Novel: 0.0%.
  • PC—Familiar-Clean: 96.11%, 0.281 ms/ex; Familiar-Noisy: 56.28%, 0.177 ms/ex; Novel-Clean: 94.40%, 0.154 ms/ex; Novel-Noisy: 37.27%, 0.188 ms/ex.
  • DIME (adaptive)—Familiar-Clean: 98.63%, 0.247 ms/ex, α ¯ = 0.993; Familiar-Noisy: 96.93%, 0.188 ms/ex, α ¯ = 0.974; Novel≈0% (as EBPM dominates unless fused).
  • DIME-Prob (probability-level fusion)—Familiar-Clean: 96.47%, 0.264 ms; Familiar-Noisy: 58.68%, 0.208 ms; Novel-Clean: 93.65%, 0.197 ms; Novel-Noisy: 36.16%, 0.239 ms.
Fashion-MNIST: Trends are similar to MNIST, with a larger familiar/noise gap.
  • EBPM—Familiar-Clean: 82.56%, 0.101 ms; Familiar-Noisy: 52.35%, 0.139 ms; EBPM reports ≈0% accuracy on novel inputs, consistent with its design as a novelty detector: sub-engrams activate but do not yield a full engram. This produces an explicit novelty signal, rather than a forced classification.
  • PC—Familiar-Clean: 83.14%, 0.182 ms; Familiar-Noisy: 50.45%, 0.238 ms; Novel-Clean: 96.50%, 0.160 ms; Novel-Noisy: 97.00%, 0.203 ms.
  • DIME (adaptive)—Familiar-Clean: 84.14%, 0.133 ms, α ¯ = 0.536; Familiar-Noisy: 51.31%, 0.170 ms, α ¯ = 0.221; Novel-Clean: 81.10%, 0.165 ms, α ¯ = 0.506; Novel-Noisy: 97.10%, 0.213 ms, α ¯ = 0.443.
CIFAR-10:
  • EBPM—Familiar-Clean: 70.42%, 0.198 ms; Familiar-Noisy: 42.67%, 0.192 ms; Novel: 0%.
  • PC—Familiar-Clean: 65.99%, 0.477 ms; Familiar-Noisy: 27.15%, 0.478 ms; Novel-Clean: 83.10%, 0.536 ms; Novel-Noisy: 50.95%, 0.548 ms.
  • DIME-Prob (best tuned)—Familiar-Clean: 68.31%, 0.529 ms; Familiar-Noisy: 30.38%, 0.535 ms; Novel-Clean: 62.50%, 0.609 ms; Novel-Noisy: 38.05%, 0.675 ms.
Remarks: On Fashion-MNIST, PC shows a slight reversal—Novel-Noisy (97.00%) marginally exceeds Novel-Clean (96.50%). With σ = 0.30 Gaussian noise (clamped to [0,1]), this acts as a mild regularizer that attenuates high-frequency artifacts for some classes, yielding a small improvement within the expected variance band. The effect is not observed on MNIST or CIFAR-10 and does not change our overall conclusion (EBPM excels on familiar; PC on novel; DIME trades off adaptively) (Table 5).
These results support the complementarity of EBPM and PC and the usefulness of DIME as a context-adaptive integration. In particular, EBPM’s absence of classification on novel inputs should be read as a feature rather than a flaw: it provides a reliable novelty flag that complements PC’s generative reconstruction. This complementarity underlies the adaptive trade-offs realized in DIME.
Figure 5, Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11 and Figure 12 illustrate the experimental results across multiple datasets (MNIST, Fashion-MNIST, and CIFAR-10), highlighting the comparative performance and latency of EBPM, PC, and DIME models under both familiar and novel conditions.
Additional views are provided in Appendix A, Figure A1 and Figure A2.

3.3. Comparative Results and Simulations

To evaluate the potential of the DIME model, we conducted a comparison across five critical dimensions: speed, energy efficiency, robustness to noise, flexibility, and biological plausibility.
  • EBPM: Performs well in speed and efficiency but is limited in adaptation and robustness.
  • PC: Flexible and robust but entails high costs.
  • DIME: Integrates the advantages of both, providing an optimal balance between speed, efficiency, and flexibility.
For a conceptual comparison, see Appendix A, Figure A6; the main text reports only empirical results.

3.4. General Information Flow in DIME

DIME organizes perception along two complementary paths (Figure 13): (1) a fast EBPM path that maps inputs to a normalized embedding and matches against class prototypes, and (2) a slower PC path that iteratively refines a latent code to improve reconstruction and classification. A controller α ∈ [0,1] balances the two outputs at inference time—favoring EBPM in familiar, low-noise settings and shifting toward PC as uncertainty or novelty increases. This flow underlies the empirical trade-offs reported in Section 3.1 and the runtime behavior in Section 3.4 [27,31].

3.5. Virtual Robotics (Obstacle Course): Runtime-Aware DIME-Lazy (α Fixed)

We evaluated EBPM, PC, DIME, and a runtime-aware variant, DIME-Lazy, in a PyBullet obstacle course with four static boxes and an overhead camera. The controller was kinematic (velocity) following a fixed waypoint slalom; thus, task success is identical across methods, and any runtime differences isolate perception cost. The camera frames were down-sampled to 28 × 28 grayscale (see Figure 14 and Figure 15). DIME-Lazy uses a delta-frame gate: for each step, we compute δ = m e a n I t I t 1 . If δ ≤ 0.004 ⇒ EBPM-only; if δ ≥ 0.015 ⇒ PC-only; otherwise, we fuse with a fixed α = 0.65: p = α p_EBPM + (1 − α) p_PC.
Waypoints: (0.45, 0.45) → (0.85, −0.45) → (1.25, 0.40) → (1.60, −0.30) → (1.80, 0.00); controller v_max = 1.2 m/s; WP_RADIUS = 0.55 m. Each condition ran for three episodes. Metrics: ms/step, success rate, collisions (contact frames), and #PC calls.
Key findings: With noise = 0.0, DIME-Lazy matches EBPM’s decisions (0 PC calls) and achieves the lowest cost (10.50 ms/step), ~−42% vs. PC (18.04 ms) and ~−32% vs. EBPM (15.43 ms), at 100% success. With noise = 0.15, DIME-Lazy escalates toward PC (≈380 PC calls/381) and its cost matches PC (19.17 vs. 19.13 ms) (see Table 6), again at 100% success. Collisions are equal across methods (mean 113), confirming that runtime comparisons are not confounded by control (see Figure 16 and Figure 17).
Additional views are provided in Appendix A, Figure A3 and Figure A4.

3.6. Ablation Studies

To validate the necessity of each DIME component, we performed ablations by disabling the adaptive controller (DIME_ctrlOFF) and by varying the number of Predictive Coding steps (K ∈ {0,5,10,20}). Controller ablations directly test whether dynamic switching improves performance beyond a fixed EBPM–PC mixture, while step ablations quantify the trade-off between robustness and latency in Predictive Coding (see Figure 18, and Table 7).
These results show that EBPM achieves the lowest latency but suffers under noise or novelty. Increasing the number of PC iterations improves robustness but with large latency costs. The DIME controller consistently matches EBPM speed in familiar settings while reducing reliance on expensive PC iterations, confirming the benefit of dynamic switching.

3.7. Statistical Significance

To ensure that observed performance differences are not due to random variation, we conducted paired statistical tests across five random seeds. We report p-values, effect sizes (Cohen’s d), and 95% confidence intervals (CIs) (see Table 8).
These significance tests confirm that the observed trends are statistically robust. On CIFAR-10 Familiar-Clean, DIME significantly outperforms EBPM (p = 0.000127, d = −4.79) and reduces reliance on PC (p = 0.000807, d = 4.07). In noisy or novel scenarios, PC maintains an advantage, but at the cost of high latency. These results validate the role of the DIME controller in balancing speed and robustness.

3.8. Complexity Analysis

For real-time robotic deployment, computational complexity is as important as accuracy. We therefore report average latency, the number of Predictive Coding (PC) iterations invoked, and the success rate in the obstacle-course environment. Results are averaged over five runs (see Table 9).
EBPM inference requires only a dot product with prototypes, resulting in extremely low latency (~0.026 ms/step) but no capacity to adapt under novel inputs. PC achieves robustness but requires thousands of iterative updates (~5 k calls per episode), making it impractical for real-time use. DIME dynamically reduces PC calls by about 50% relative to pure PC (≈2.4 k vs. ≈5.2 k per episode), while maintaining a high success rate. This analysis confirms that DIME provides a feasible trade-off for resource-constrained robotic scenarios.

3.9. Conclusions of Results

Taken together, these results validate the complementary contributions of EBPM, PC, and the DIME controller. EBPM alone provides extremely low-latency recognition of familiar stimuli but fails under novelty or noise. PC ensures robustness under uncertainty, at the cost of iterative inference and high latency. DIME dynamically combines the two, achieving accuracy improvements of up to +5% on CIFAR-10 while reducing PC usage by ~50% compared to pure PC. Ablation studies confirm the necessity of the adaptive controller, statistical tests demonstrate significance across multiple datasets, and complexity analysis supports feasibility in resource-constrained robotic scenarios. These findings establish DIME as a biologically plausible and practically applicable cognitive architecture for low-latency recognition and control.

4. Discussion

4.1. Advantages of EBPM

The EBPM model explains the rapid and energy-efficient reactions observed in familiar contexts. In contrast to PC, where the cortex must continuously generate internal predictions [2], EBPM assumes that a sensory stimulus directly activates the relevant engram. This mechanism reduces latencies and energy cost, in agreement with neurophysiological data showing that neuronal assemblies reactivate almost instantaneously when familiar stimuli are encountered [4].
Another advantage lies in procedural robustness. Complex but overlearned actions—such as the use of a tool—can be initiated through the direct reactivation of a multimodal engram, without the need for predictive simulations. This aspect correlates with the observations of Hawkins and Ahmad [5], who showed that neurons possess thousands of synapses precisely to facilitate the rapid reactivation of memorized sequences.
The comparison in Section 2.3.6 shows that EBPM covers familiar inputs with minimal compute, while generative or external-memory approaches excel on novelty; DIME reconciles the two.
Nevertheless, EBPM recognition is contingent on the prior existence of well-consolidated engrams. In situations where no such engrams are available, EBPM cannot operate effectively, which limits its explanatory power in novel or ambiguous contexts.

4.2. Limitations of EBPM

A frequently raised concern is that EBPM, by design, yields near-zero accuracy on entirely novel inputs. However, this should not be interpreted as a failure but rather as a functional distinction: EBPM explicitly separates the processes of recognition and novelty detection.
In EBPM, unfamiliar stimuli still activate sub-engrams (elementary sensory, motor, or interoceptive primitives). Although none of the stored engrams reach full activation, the fragmentary sub-engram activity defines a novelty state. This state automatically triggers attentional mechanisms that increase neuronal gain and plasticity, promoting the binding of co-active sub-engrams into a proto-engram. With repetition or behavioral relevance, this proto-engram consolidates into a stable engram; without reinforcement, the weak associations fade [32].
Thus, the “0% accuracy” in novel conditions is not a bug but an intentional feature: EBPM does not force a misclassification under uncertainty. Instead, it signals novelty and reallocates resources toward exploration and learning. From a biological standpoint, this aligns with evidence that attention and neuromodulatory systems (e.g., locus coeruleus–noradrenaline) are engaged precisely when established assemblies fail to reactivate. From an engineering perspective, this provides an explicit novelty-flagging mechanism that can be exploited in hybrid systems such as DIME, where PC takes over under unfamiliar or noisy conditions.
In summary, EBPM’s limitation is also its defining contribution: while it cannot generalize to novel inputs on its own, it provides a biologically grounded mechanism for the rapid recognition of the familiar and explicit detection of the unfamiliar, thereby laying the groundwork for adaptive integration with generative models.

4.3. Advantages of PC

Predictive Coding (PC) compensates for these limitations through its flexibility and robustness. Because the cortex generates top-down predictions, it can reconstruct incomplete or noisy inputs. This explains, for example, why we are able to recognize an object even when it is partially occluded [33]. Moreover, PC is supported by strong neuroimaging evidence demonstrating the active presence of anticipatory signals in higher-order cortical areas [34].
Importantly, PC also remains valuable for explaining how new engrams can be formed under conditions of partial or conflicting input. Its recursive comparison mechanism provides a plausible account of how the brain integrates incomplete information and adapts to unforeseen circumstances.

4.4. Limitations of PC

However, PC entails a high energetic cost and longer latencies. The continuous process of generating and verifying predictions is inefficient in familiar situations, where direct recognition would be sufficient. This observation is consistent with the arguments put forward by Kanai et al. [3], who highlighted that maintaining a complete generative model is redundant in stable and well-known contexts.
A more fundamental critique of PC concerns the very notion of “prediction” as an active top-down mechanism. From a neurobiological perspective, there is no evidence for explicit reconstructions of sensory input at lower levels. What actually exists are engrams, i.e., networks consolidated through experience and synaptic plasticity, according to the principle “neurons that fire together wire together”. When incoming stimuli sufficiently reactivate these networks (for example, edges + red color + cylindrical shape → cup), recognition arises directly from this reactivation. There is no top-down “drawing” of the cup into the visual cortex; there are only pre-trained networks that fire automatically when the input matches. If, however, the object is missing, chipped, or altered (e.g., covered with flour), the activation of the engram is partial or inconsistent, which triggers attention, learning, and the formation of new connections. In this interpretation, the top-down flow does not generate explicit predictions but functions as a modulation of excitability in lower-level networks, sensitizing them according to context and goals.
This role of top-down signaling becomes particularly evident in imagination and internal simulation. When a person “falls into thought” or mentally visualizes an object, associative and hippocampal networks reactivate engrams and transmit activity top-down to sensory cortices, so that one can “see” the cup or “hear” a melody in the mind without any external input. In parallel, there is an active inhibition of real sensory inputs, which explains why people become less attentive to the environment during deep thought. Everyday examples confirm this inhibitory control: when an ophthalmologist blows air into the eye, but the voluntary system suppresses the blink reflex; or when a nurse administers an injection and, despite the painful stimulus, the prefrontal cortex modulates motor pathways to block the withdrawal reflex. Similar effects are seen when athletes suppress avoidance reflexes to perform risky movements, or when musicians ignore sudden noises to remain focused. These examples illustrate that top-down influence is primarily about gain control and inhibition, not about generating full sensory predictions.
Moreover, PC also fails to account for fast reflexes and highly automated responses, which bypass any generative comparison process. Reflexes such as blinking to corneal stimulation or withdrawing from a painful stimulus are executed through subcortical or spinal circuits within tens of milliseconds, long before any top-down “prediction” could be generated or verified. Similarly, well-learned sensorimotor routines (e.g., skilled grasping or locomotion patterns) unfold primarily through engram reactivation in cortico-basal ganglia and cerebellar loops, with minimal involvement of predictive feedback. This further highlights that PC cannot serve as a universal principle of recognition and behavior, but at most as an auxiliary attentional mechanism in situations of novelty, conflict, or noise.
Taken together, these considerations suggest that PC is not the primary engine of recognition but rather a set of attentional and executive sub-mechanisms. EBPM provides a more direct explanation for rapid and low-cost recognition through engram reactivation, while PC becomes relevant mainly when engram activation is insufficient, inconsistent, or requires inhibition of automatic responses.
A further limitation of PC lies in the absence of a clear neurobiological specification of how predictions are generated. In most formulations, the “prediction” is treated as an abstract top-down signal, with little explanation of the concrete neuronal mechanisms that produce it. Empirical evidence suggests that top-down pathways modulate the excitability of lower-level networks rather than reconstructing full sensory input, leaving the precise nature of the predictive signal underspecified. In contrast, EBPM explicitly defines recognition as the reactivation of consolidated engrams, including the conditions under which activation is complete, partial, or inconsistent, and how attention and learning intervene in the latter cases. Within the DIME framework, this provides a mechanistic clarity absent from PC, while still allowing PC-like processes to be reinterpreted as attentional subroutines engaged in novelty and ambiguity.

4.5. Threats to Validity and Limitations

Internal validity: Our familiar/novel split (0–7 vs. 8–9) is standard but creates a sharp novelty boundary; different splits could change absolute numbers. Familiarity calibration (μ,σ) for DIME was estimated on familiar validation images; miscalibration would shift α and the EBPM↔PC balance.
Construct validity: Accuracy and ms/step capture complementary aspects (task performance vs. compute cost) but do not measure uncertainty calibration or open-set risk explicitly.
External validity: CIFAR-10 and MNIST/F-MNIST are limited in complexity; generalization to natural, long-tail distributions requires larger models/datasets. The PyBullet course uses a fixed waypoint controller; success equality across methods isolates perception cost but does not evaluate closed-loop policy learning.
Conclusion validity: Some robotics metrics average over three episodes; while stable, they may under-estimate rare failure modes. We therefore report means and provide raw CSV files for replication.
Future work: We will evaluate on stronger open-set/OOD benchmarks, vary the novelty boundary, increase episode counts, and explore learned gating for DIME-Lazy beyond fixed α and thresholds.
Appendix A captions: Figure A7—Hypothesized trends (non-empirical): Hypothetical plots used during design; retained for transparency. Not used to support claims.
Parallels with the Predictive Coding trajectory: It is important to emphasize that the EBPM/DIME framework should be understood as a theoretical synthesis, not as an isolated speculation. Its components—engram reactivation, Hebbian/STDP, lateral inhibition, attentional gain, and LC–NE novelty signaling—are all individually supported by empirical evidence in both animal and human studies. The novelty of EBPM lies in integrating these mechanisms into a unified account of recognition and novelty detection. In this respect, EBPM’s current status is comparable to that of Predictive Coding (PC) at the time of its initial formulation: PC was accepted as a dominant paradigm long before all of its assumptions were empirically confirmed, and to this day, some of its central claims (e.g., explicit generative reconstructions in primary cortices, “negative spike” signals) remain debated. EBPM/DIME thus follows a similar trajectory: built on solid empirical fragments, it provides a coherent alternative framework whose integrative claims now require systematic validation through the experimental program outlined in Section 2.5.

4.6. DIME as an Integrative Solution

Given these complementary characteristics, the proposed model—DIME (Dynamic Integrative Matching and Encoding)—offers a balanced solution. By combining the two paradigms, DIME enables both rapid recognition through EBPM in familiar situations and generative inference through PC in ambiguous conditions. The dynamic controller determines the dominant pathway depending on the level of familiarity and uncertainty. This type of architecture reflects the recent view of the brain as a hybrid system that combines fast associations with slower but more flexible inferences [21].
To evaluate the advantages and limitations of each model, we synthesized the main characteristics in Table 10. This comparison shows that EBPM achieves peak performance in speed and energy efficiency but is less adaptable. PC excels in flexibility and noise robustness yet entails higher computational and energetic costs. DIME successfully integrates the advantages of both paradigms, offering a better balance between performance, adaptability, and biological plausibility. This suggests that such integration not only reflects the functioning of the brain more faithfully but also provides a promising direction for the design of efficient AI and robotic systems.
A comparative table-style figure or infographic:
  • EBPM = fast, energy-efficient, but less flexible.
  • PC = robust, adaptive, but computationally costly.
  • DIME = hybrid integration, balancing both.
Table 10 summarizes the strengths and limitations of EBPM, PC, and DIME. EBPM excels in speed and energy efficiency, PC in flexibility and noise robustness, and DIME provides an integrative balance. Such comparative tables are common in neurocomputational model reviews [22,35].
The ANN results (Table 5a–c, Figure 5, Figure 6, Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11) empirically validate these claims: EBPM excels in speed and efficiency, PC in robustness, and DIME dynamically balances both. The α controller consistently shifted toward EBPM on familiar inputs and toward PC under noise and novelty, in agreement with the theoretical model.

4.7. Interdisciplinary Implications

  • Neuroscience: EBPM provides an alternative explanation for the phenomenon of memory replay, while DIME may serve as a more faithful framework for understanding the interaction between memory and prediction.
  • Artificial Intelligence: EBPM-inspired architectures are more energy-efficient, and their integration with PC-type modules can enhance noise robustness.
  • Robotics: DIME can be implemented in autonomous systems that require rapid reactions in familiar contexts but also adaptability in novel situations, similar to the approaches tested [21,22].
In artificial intelligence, EBPM also resonates with memory-augmented neural architectures such as Neural Turing Machines [17], Differentiable Neural Computers [18], and episodic control models in reinforcement learning [19]. Unlike these approaches, which are primarily engineered solutions, EBPM emphasizes biological grounding through multimodal engram reactivation. DIME can therefore be interpreted as a bridge between biologically plausible engram-based recall and machine-oriented predictive architectures.
In future work, we plan to integrate methods from Expert Systems (e.g., [6]) and hybrid recognition frameworks (e.g., [7]) within our DIME architecture to further strengthen interpretability and multimodal integration.
A critical next step will be extending EBPM/DIME beyond image recognition to multimodal engrams. For example, combining spoken digits with visual digits or pairing motor action sequences with visual object recognition would provide cross-modal evidence for engram reactivation. Such multimodal validation is essential for establishing the generality of the model in realistic cognitive scenarios.

5. Conclusions

This paper has introduced and compared two fundamental paradigms of neural processing: Predictive Coding (PC), an established framework in neuroscience [1,2], and the original Experience-Based Pattern Matching (EBPM) model, proposed here for the first time. The comparative analysis highlighted that EBPM provides clear advantages in familiar contexts: faster recognition, increased energy efficiency, and better consistency with phenomena of direct memory reactivation (multimodal engrams) [4,32]. In contrast, PC remains superior in ambiguous situations or with incomplete inputs, due to its anticipatory reconstruction capability [33].
Nevertheless, neither of the two models fully accounts for all observed cognitive phenomena. For this reason, we proposed a new integrative framework—DIME (Dynamic Integrative Matching and Encoding)—which combines EBPM’s direct matching with PC’s generative prediction. DIME incorporates a dynamic controller that determines the dominant processing pathway depending on familiarity, uncertainty, and goals. Theoretical results and preliminary simulations suggest that DIME provides an optimal balance between speed, efficiency, and robustness, making it both more biologically plausible and more practically useful than the individual models.
This contribution has interdisciplinary implications:
  • In neuroscience, it provides a more comprehensive explanation of the interaction between memory and prediction;
  • In artificial intelligence, it suggests hybrid architectures that are more energy-efficient;
  • In robotics, it paves the way for autonomous systems that can combine rapid recognition with adaptability to unfamiliar environments.
The present ANN simulations provide the first concrete validation of the DIME model, supporting its potential for integration into both neuroscience frameworks and robotic control architectures.
This study has several limitations that define directions for future research. First, EBPM was evaluated primarily on image classification tasks; extending the model to multimodal engrams, including speech, motor actions, and cross-modal datasets, remains an important step toward demonstrating generality. Second, comparisons focused on PC and classical models, whereas benchmarking against modern deep learning architectures (CNNs, Transformers) is necessary to further position EBPM/DIME within the landscape of contemporary AI. Third, robotic evaluation was restricted to a virtual obstacle course; future work should investigate real-time deployment on physical humanoid or assistive robotic platforms, under hardware and energy constraints. Addressing these limitations will provide stronger evidence for EBPM as a biologically grounded and practically applicable cognitive model.
In conclusion, EBPM represents an original contribution to cognitive modeling, while DIME provides a hybrid framework with the potential to redefine how we interpret neural computation and apply it in intelligent systems.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/app152010904/s1, Supplementary File: Supplementary.zip, containing all supplementary figures, tables, configuration files, raw TXT/CSV outputs, and robotic simulation scripts and results corresponding to the experiments described in the main manuscript.

Author Contributions

Conceptualization, I.C.V.; methodology, I.C.V. and N.G.B.; software, I.C.V.; validation, I.C.V., I.P. and B.C.; formal analysis, I.C.V. and B.C.; investigation, I.C.V. and I.P.; resources, I.C.V. and N.G.B.; data curation, I.C.V.; writing—original draft preparation, I.C.V.; writing—review and editing, I.C.V., I.P., and B.C.; visualization, I.C.V.; supervision, I.C.V. and N.G.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

All scripts, configuration files, and raw outputs for ANN (MNIST, Fashion-MNIST, CIFAR-10) and PyBullet experiments are provided as a supplementary package and anonymized ZIP provided to reviewers; public links added after acceptance. The figures in the main text correspond to image files listed in MANIFEST.md under Supplementary/figures/, and the raw numeric results are provided as CSV/TXT under Supplementary/results_csv/ and Supplementary/results_txt/. For exact reproduction we include environment versions (Python/PyTorch/CUDA), random seeds, and runnable commands in README.md. An immutable archive will be deposited on Zenodo at acceptance, and the DOI will be added to this section. A blind, anonymous link to the supplementary ZIP is provided via the submission system; the public GitHub/Zenodo links will be added upon acceptance. The supplementary repository includes all scripts used to produce the results (see experiments/mnist_all.py, fashion_mnist_all.py, cifar10_all.py) and a minimal robotics scaffold (robotics/pybullet_setup_demo.py), together with exact seeds, default hyper-parameters, and ready-to-run commands. Raw TXT/CSV outputs and plotting utilities are provided alongside.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Figure A1. EBPM vs. PC (Latency). Latency comparison restricted to EBPM vs. PC (K = 10, η = 0.1).
Figure A1. EBPM vs. PC (Latency). Latency comparison restricted to EBPM vs. PC (K = 10, η = 0.1).
Applsci 15 10904 g0a1
Figure A2. EBPM vs. PC (Accuracy). Accuracy comparison restricted to EBPM vs. PC.
Figure A2. EBPM vs. PC (Accuracy). Accuracy comparison restricted to EBPM vs. PC.
Applsci 15 10904 g0a2
Figure A3. Legacy ms/step bar chart. Historic bar chart of ms/step on an older configuration. Superseded by the v3c_alpha results.
Figure A3. Legacy ms/step bar chart. Historic bar chart of ms/step on an older configuration. Superseded by the v3c_alpha results.
Applsci 15 10904 g0a3
Figure A4. Legacy DIME-Lazy ms/step. Older view of DIME-Lazy perception cost; thresholds and controller differ from the final setup.
Figure A4. Legacy DIME-Lazy ms/step. Older view of DIME-Lazy perception cost; thresholds and controller differ from the final setup.
Applsci 15 10904 g0a4
Figure A5. Legacy obstacles ms/step. Earlier ms/step summary on the obstacle scenario. Kept as a historical reference; the main text reports v3c_alpha.
Figure A5. Legacy obstacles ms/step. Earlier ms/step summary on the obstacle scenario. Kept as a historical reference; the main text reports v3c_alpha.
Applsci 15 10904 g0a5
Figure A6. Conceptual radar comparison. Conceptual (non-empirical) illustration of expected tendencies for EBPM, PC, and DIME across speed, robustness to noise/novelty, and compute cost. Provided for intuition only; not derived from measured data.
Figure A6. Conceptual radar comparison. Conceptual (non-empirical) illustration of expected tendencies for EBPM, PC, and DIME across speed, robustness to noise/novelty, and compute cost. Provided for intuition only; not derived from measured data.
Applsci 15 10904 g0a6
Figure A7. (a). Predicted EEG/MEG recognition latency (ms) with 95% CI: EBPM ≈ 250 ± 40, PC ≈ 400 ± 60. (b) Predicted fMRI RSA (r) with 95% CI: EBPM ≈ 0.75 ± 0.05, PC ≈ 0.60 ± 0.07.
Figure A7. (a). Predicted EEG/MEG recognition latency (ms) with 95% CI: EBPM ≈ 250 ± 40, PC ≈ 400 ± 60. (b) Predicted fMRI RSA (r) with 95% CI: EBPM ≈ 0.75 ± 0.05, PC ≈ 0.60 ± 0.07.
Applsci 15 10904 g0a7aApplsci 15 10904 g0a7b
Figure A8. (a). Predicted ANN execution time (ms) with 95% CI: EBPM ≈ 80 ± 10, PC ≈ 110 ± 12. (b). Predicted robotic energy consumption (arb. units) with 95% CI: EBPM ≈ 70 ± 8, PC ≈ 95 ± 6.
Figure A8. (a). Predicted ANN execution time (ms) with 95% CI: EBPM ≈ 80 ± 10, PC ≈ 110 ± 12. (b). Predicted robotic energy consumption (arb. units) with 95% CI: EBPM ≈ 70 ± 8, PC ≈ 95 ± 6.
Applsci 15 10904 g0a8aApplsci 15 10904 g0a8b

Appendix B

Table A1. Rdata.
Table A1. Rdata.
File (Relative Path)What It Contains
results_csv/pybullet_obstacles_metrics_v3c_alpha.csvv3c_alpha (α = 0.65): ms/step mean, steps, success, collisions, avg PC calls per mode and noise
results_csv/pybullet_obstacles_metrics_v3c.csvv3c (dyn gating, no fixed α): same fields
results_csv/pybullet_obstacles_metrics_v3b.csvv3b: success and collisions breakdown
results_csv/pybullet_perception_metrics.csvdemo perception ms/step EBPM vs. PC vs. DIME

References

  1. Friston, K. The free-energy principle: A unified brain theory? Nat. Rev. Neurosci. 2010, 11, 127–138. [Google Scholar] [CrossRef]
  2. Rao, R.P.N.; Ballard, D.H. Predictive coding in the visual cortex: A functional interpretation of some extra-classical receptive-field effects. Nat. Neurosci. 1999, 2, 79–87. [Google Scholar] [CrossRef]
  3. Kanai, R.; Komura, Y.; Shipp, S.; Friston, K. Cerebral hierarchies: Predictive processing, precision and the pulvinar. Philos. Trans. R. Soc. B Biol. Sci. 2015, 370, 20140169. [Google Scholar] [CrossRef]
  4. Buzsáki, G. Neural Syntax: Cell Assemblies, Synapsembles, and Readers. Neuron 2010, 68, 362–385. [Google Scholar] [CrossRef]
  5. Hawkins, J.; Ahmad, S. Why Neurons Have Thousands of Synapses, a Theory of Sequence Memory in Neocortex. Front. Neural Circuits 2016, 10, 23. [Google Scholar] [CrossRef]
  6. Xiang, Q.; Wang, X.; Song, Y.; Lei, L. ISONet: Reforming 1DCNN for aero-engine system inter-shaft bearing fault diagnosis via input spatial over-parameterization. Expert Syst. Appl. 2025, 277, 127248. [Google Scholar] [CrossRef]
  7. Xiang, Q.; Wang, X.; Lei, L.; Song, Y. Dynamic bound adaptive gradient methods with belief in observed gradients. Pattern Recognit. 2025, 168, 111819. [Google Scholar] [CrossRef]
  8. Markram, H.; Gerstner, W.; Sjöström, P.J. Spike-Timing-Dependent Plasticity: A Comprehensive Overview. Front. Synaptic Neurosci. 2012, 4, 2. [Google Scholar] [CrossRef] [PubMed]
  9. Hopfield, J.J. Neural networks and physical systems with emergent collective computational abilities. Proc. Natl. Acad. Sci. USA 1982, 79, 2554–2558. [Google Scholar] [CrossRef] [PubMed]
  10. McClelland, J.L.; McNaughton, B.L.; O’Reilly, R.C. Why there are complementary learning systems in the hippocampus and neocortex: Insights from the successes and failures of connectionist models of learning and memory. Psychol. Rev. 1995, 102, 419–457. [Google Scholar] [CrossRef]
  11. Teyler, T.J.; DiScenna, P. The hippocampal memory indexing theory. Behav. Neurosci. 1986, 100, 147–154. [Google Scholar] [CrossRef]
  12. Rolls, E.T. The mechanisms for pattern completion and pattern separation in the hippocampus. Front. Syst. Neurosci. 2013, 7, 74. [Google Scholar] [CrossRef] [PubMed]
  13. Nosofsky, R.M. Attention, similarity, and the identification–categorization relationship. J. Exp. Psychol. Gen. 1986, 115, 39–57. [Google Scholar] [CrossRef] [PubMed]
  14. Logan, G.D. Toward an instance theory of automatization. Psychol. Rev. 1988, 95, 492–527. [Google Scholar] [CrossRef]
  15. Biederman, I. Recognition-by-components: A theory of human image understanding. Psychol. Rev. 1987, 94, 115–147. [Google Scholar] [CrossRef] [PubMed]
  16. Kanerva, P. Sparse Distributed Memory; MIT Press: Cambridge, MA, USA, 1988. [Google Scholar]
  17. Graves, A.; Wayne, G.; Danihelka, I. Neural Turing Machines. arXiv 2014, arXiv:1410.5401. [Google Scholar] [CrossRef]
  18. Graves, A.; Wayne, G.; Reynolds, M.; Harley, T.; Danihelka, I.; Grabska-Barwińska, A.; Colmenarejo, S.G.; Grefenstette, E.; Ramalho, T.; Agapiou, J.; et al. Hybrid computing using a neural network with dynamic external memory. Nature 2016, 538, 471–476. [Google Scholar] [CrossRef]
  19. Blundell, C.; Uria, B.; Pritzel, A.; Li, Y.; Ruderman, A.; Leibo, J.Z.; Rae, J.; Wierstra, D.; Hassabis, D. Model-Free Episodic Control. arXiv 2016, arXiv:1606.04460. [Google Scholar] [CrossRef]
  20. Snell, J.; Swersky, K.; Zemel, R.S. Prototypical Networks for Few-shot Learning. arXiv 2017, arXiv:1703.05175. [Google Scholar] [CrossRef]
  21. Clark, A. Whatever next? Predictive brains, situated agents, and the future of cognitive science. Behav. Brain Sci. 2013, 36, 181–204. [Google Scholar] [CrossRef]
  22. Bastos, A.M.; Usrey, W.M.; Adams, R.A.; Mangun, G.R.; Fries, P.; Friston, K.J. Canonical Microcircuits for Predictive Coding. Neuron 2012, 76, 695–711. [Google Scholar] [CrossRef]
  23. Olshausen, B.A.; Field, D.J. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature 1996, 381, 607–609. [Google Scholar] [CrossRef]
  24. Bengio, Y.; Courville, A.; Vincent, P. Representation Learning: A Review and New Perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef]
  25. Itti, L.; Koch, C. Computational modelling of visual attention. Nat. Rev. Neurosci. 2001, 2, 194–203. [Google Scholar] [CrossRef]
  26. Aston-Jones, G.; Cohen, J.D. AN INTEGRATIVE THEORY OF LOCUS COERULEUS-NOREPINEPHRINE FUNCTION: Adaptive Gain and Optimal Performance. Annu. Rev. Neurosci. 2005, 28, 403–450. [Google Scholar] [CrossRef]
  27. Dehaene, S.; Meyniel, F.; Wacongne, C.; Wang, L.; Pallier, C. The Neural Representation of Sequences: From Transition Probabilities to Algebraic Patterns and Linguistic Trees. Neuron 2015, 88, 2–19. [Google Scholar] [CrossRef]
  28. Douglas, R.J.; Martin, K.A.C. Recurrent neuronal circuits in the neocortex. Curr. Biol. 2007, 17, R496–R500. [Google Scholar] [CrossRef] [PubMed]
  29. Ambrus, G.G. Shared neural codes of recognition memory. Sci. Rep. 2024, 14, 15846. [Google Scholar] [CrossRef]
  30. Gerstner, W.; Kistler, W.M.; Naud, R.; Paninski, L. Neuronal Dynamics: From Single Neurons to Networks and Models of Cognition, 1st ed.; Cambridge University Press: Cambridge, UK, 2014. [Google Scholar] [CrossRef]
  31. Evans, J.S.B.T.; Stanovich, K.E. Dual-Process Theories of Higher Cognition: Advancing the Debate. Perspect. Psychol. Sci. 2013, 8, 223–241. [Google Scholar] [CrossRef]
  32. Hasson, U.; Chen, J.; Honey, C.J. Hierarchical process memory: Memory as an integral component of information processing. Trends Cogn. Sci. 2015, 19, 304–313. [Google Scholar] [CrossRef]
  33. Summerfield, C.; De Lange, F.P. Expectation in perceptual decision making: Neural and computational mechanisms. Nat. Rev. Neurosci. 2014, 15, 745–756. [Google Scholar] [CrossRef] [PubMed]
  34. Kok, P.; Jehee, J.F.M.; de Lange, F.P. Less Is More: Expectation Sharpens Representations in the Primary Visual Cortex. Neuron 2012, 75, 265–270. [Google Scholar] [CrossRef] [PubMed]
  35. Kiebel, S.J.; Daunizeau, J.; Friston, K.J. A Hierarchy of Time-Scales and the Brain. PLoS Comput. Biol. 2008, 4, e1000209. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Schematic representation of the EBPM framework.
Figure 1. Schematic representation of the EBPM framework.
Applsci 15 10904 g001
Figure 2. Predictive Coding (PC) architecture.
Figure 2. Predictive Coding (PC) architecture.
Applsci 15 10904 g002
Figure 3. DIME (Dynamic Integrative Matching and Encoding) model.
Figure 3. DIME (Dynamic Integrative Matching and Encoding) model.
Applsci 15 10904 g003
Figure 4. Proposed experimental validation methodology.
Figure 4. Proposed experimental validation methodology.
Applsci 15 10904 g004
Figure 5. MNIST: Latency—Per-scenario inference latency (ms/ex.) on MNIST for EBPM, PC (K = 10, η = 0.1), and DIME across Familiar-Clean/Noisy and Novel-Clean/Noisy. EBPM is fastest on familiar; DIME approaches PC when uncertainty grows.
Figure 5. MNIST: Latency—Per-scenario inference latency (ms/ex.) on MNIST for EBPM, PC (K = 10, η = 0.1), and DIME across Familiar-Clean/Noisy and Novel-Clean/Noisy. EBPM is fastest on familiar; DIME approaches PC when uncertainty grows.
Applsci 15 10904 g005
Figure 6. MNIST: Accuracy—Top-1 accuracy on MNIST for the same four scenarios. EBPM excels on familiar but fails by design on novel (8–9); PC generalizes better to novelty; DIME balances both.
Figure 6. MNIST: Accuracy—Top-1 accuracy on MNIST for the same four scenarios. EBPM excels on familiar but fails by design on novel (8–9); PC generalizes better to novelty; DIME balances both.
Applsci 15 10904 g006
Figure 7. Fashion-MNIST: Latency—Per-scenario latency (ms/ex.) on Fashion-MNIST for EBPM, PC (K = 10, η = 0.1), and DIME. Relative ordering mirrors MNIST; PC is slower due to iterative inference.
Figure 7. Fashion-MNIST: Latency—Per-scenario latency (ms/ex.) on Fashion-MNIST for EBPM, PC (K = 10, η = 0.1), and DIME. Relative ordering mirrors MNIST; PC is slower due to iterative inference.
Applsci 15 10904 g007
Figure 8. Fashion-MNIST: Accuracy—Top-1 accuracy on Fashion-MNIST across the four scenarios. PC dominates on novel; EBPM is stronger on familiar; DIME interpolates context-dependently.
Figure 8. Fashion-MNIST: Accuracy—Top-1 accuracy on Fashion-MNIST across the four scenarios. PC dominates on novel; EBPM is stronger on familiar; DIME interpolates context-dependently.
Applsci 15 10904 g008
Figure 9. CIFAR-10: Latency—Per-scenario latency on CIFAR-10. Absolute times are higher than MNIST/F-MNIST; the relative ordering of EBPM < DIME ≈ PC persists.
Figure 9. CIFAR-10: Latency—Per-scenario latency on CIFAR-10. Absolute times are higher than MNIST/F-MNIST; the relative ordering of EBPM < DIME ≈ PC persists.
Applsci 15 10904 g009
Figure 10. CIFAR-10: Accuracy—Top-1 accuracy on CIFAR-10 for the four scenarios. PC leads on novel; EBPM degrades on noise/novel; tuned DIME-Prob trades off between the two.
Figure 10. CIFAR-10: Accuracy—Top-1 accuracy on CIFAR-10 for the four scenarios. PC leads on novel; EBPM degrades on noise/novel; tuned DIME-Prob trades off between the two.
Applsci 15 10904 g010
Figure 11. MNIST: Fusion ablation—Ablation of fusion strategies: probability- vs. logit-level, familiarity calibration on/off, and iterations K. Probability-level fusion with calibration yields the best familiar/novel trade-off at comparable latency.
Figure 11. MNIST: Fusion ablation—Ablation of fusion strategies: probability- vs. logit-level, familiarity calibration on/off, and iterations K. Probability-level fusion with calibration yields the best familiar/novel trade-off at comparable latency.
Applsci 15 10904 g011
Figure 12. MNIST: Mean α. Mean fusion weight α ¯ per scenario showing how DIME shifts between EBPM and PC as uncertainty changes.
Figure 12. MNIST: Mean α. Mean fusion weight α ¯ per scenario showing how DIME shifts between EBPM and PC as uncertainty changes.
Applsci 15 10904 g012
Figure 13. General information flow in the DIME model.
Figure 13. General information flow in the DIME model.
Applsci 15 10904 g013
Figure 14. Obstacle course (snapshot). Top-down orthographic view of the obstacle course. R = robot, T = target, O1–O4 = obstacles; frames from this camera are down-sampled to 28 × 28 grayscale.
Figure 14. Obstacle course (snapshot). Top-down orthographic view of the obstacle course. R = robot, T = target, O1–O4 = obstacles; frames from this camera are down-sampled to 28 × 28 grayscale.
Applsci 15 10904 g014
Figure 15. Obstacle course (schematic). Clean 2D schematic with waypoints used by the kinematic velocity controller (v_max = 1.2 m/s, WP_RADIUS = 0.55 m). R = robot, T = target.
Figure 15. Obstacle course (schematic). Clean 2D schematic with waypoints used by the kinematic velocity controller (v_max = 1.2 m/s, WP_RADIUS = 0.55 m). R = robot, T = target.
Applsci 15 10904 g015
Figure 16. Runtime cost vs. noise (v3c_alpha, α = 0.65). Mean perception time per step (3 episodes) for EBPM, PC, DIME, and DIME-Lazy, under noise ∈ {0.0, 0.15}. DIME-Lazy equals EBPM on stable frames and escalates to PC under higher dynamics, preserving 100% success.
Figure 16. Runtime cost vs. noise (v3c_alpha, α = 0.65). Mean perception time per step (3 episodes) for EBPM, PC, DIME, and DIME-Lazy, under noise ∈ {0.0, 0.15}. DIME-Lazy equals EBPM on stable frames and escalates to PC under higher dynamics, preserving 100% success.
Applsci 15 10904 g016
Figure 17. Runtime cost and PC calls (two-panel summary). Top: ms/step (mean of 3 episodes). Bottom: average #PC calls. Runtime-aware gating reduces cost on stable scenes and increases it when dynamics/noise rise, with unchanged task success.
Figure 17. Runtime cost and PC calls (two-panel summary). Top: ms/step (mean of 3 episodes). Bottom: average #PC calls. Runtime-aware gating reduces cost on stable scenes and increases it when dynamics/noise rise, with unchanged task success.
Applsci 15 10904 g017
Figure 18. Ablation study on MNIST: Latency (ms/sample) as a function of Predictive Coding iterations K, comparing PC baselines and DIME controller variants. Error bars denote ± 1 standard deviation over five seeds.
Figure 18. Ablation study on MNIST: Latency (ms/sample) as a function of Predictive Coding iterations K, comparing PC baselines and DIME controller variants. Error bars denote ± 1 standard deviation over five seeds.
Applsci 15 10904 g018
Table 1. Extended comparison between Predictive Coding (PC) and Experience-Based Pattern Matching (EBPM), with explicit treatment of novelty, attention, and new engram formation.
Table 1. Extended comparison between Predictive Coding (PC) and Experience-Based Pattern Matching (EBPM), with explicit treatment of novelty, attention, and new engram formation.
AspectPredictive Coding (PC)Engram-Based Program Model (EBPM)
Basic principlePredictions compared with input; error is propagated.Recognition through engrams composed of reusable sub-engrams; full activation only for familiar stimuli.
Partial activation-Partial/similar stimuli → shared sub-engrams are activated; multiple engrams become partially active.
Selection-Lateral inhibition: the engram with the most complete activation prevails.
Novelty/unfamiliarityUnclear how new representations are detected and formed.If all engrams remain below the threshold: sub-engrams are active, but no parent engram is completed → novelty state. Attention is engaged.
Formation of new engrams (role of attention)Does not specify a concrete biological mechanism.Automatic (“fire together, wire together”): in the novelty state, attention increases neuronal gain and plasticity; co-active sub-engrams are bound into a proto-engram. Without repetition, the linkage fades/does not finalize. With repetition, it consolidates into a stable engram.
Signaling absencePostulates “omission error” (biologically problematic).No “negative spike.” Silence at the parent-engram level + fragmentary activation of sub-engrams defines novelty; attention manages exploration and learning.
Neuronal logicRequires an internal comparator/explicit error signals.Pure spike/silence + lateral inhibition + co-activation-dependent plasticity; attention sets the learning threshold.
Table 2. (a) Conceptual differences between EBPM and PC. (b) Conceptual differences extended with related models.
Table 2. (a) Conceptual differences between EBPM and PC. (b) Conceptual differences extended with related models.
(a)
DomainEBPMPC Prediction
Central MechanismDirect activation of multimodal engramsContinuous generation and updating of predictions
The Role of PredictionNon-existent; only goal-driven sensitizationFundamental for processing
Recognition SpeedInstantaneous for familiar stimuliSlower, requires iterative computation
Energy EfficiencyHigh (without generative simulation)Lower (increased metabolic cost)
Attention ControlGoal-driven (sensitization of relevant networks)Through the magnitude of prediction errors
Biological PlausibilitySupported by studies on engrams and inhibitionSupported by cortical predictive microcircuits
Typical Application DomainsRapid recognition, episodic memories, reflexesPerception in ambiguous or noisy conditions
(b)
DomainEBPMPCSimilar existing theories
Central MechanismDirect multimodal engram activationGenerative Predictive CodingHopfield attractors, episodic replay
Recognition SpeedFast, one-step matchIterative, slowerFast (pattern completion)
Energy EfficiencyHigh, no generative loopLow, high metabolic demandVariable
FlexibilityLimited to stored engramsHigh, reconstructiveActive inference
Biological PlausibilitySupported by engram studiesSupported by cortical microcircuitsBoth partially supported
Table 3. Comparative predictions: EBPM vs. PC.
Table 3. Comparative predictions: EBPM vs. PC.
DomainEBPM PredictionPC PredictionTestable Through
fMRI (similarity analysis)Recognition = reactivation of storage networksAnticipatory activation in higher cortical areasfMRI + RSA
EEG/MEG (latency)Rapid recognition of familiar patternsLonger latencies due to generative processingEEG/MEG
Cue-based recallA partial cue reactivates the entire engramThe cue triggers generative reconstructionBehavioral and neurophysiological studies
Robotic implementationFaster execution of learned proceduresBetter adaptive reconstruction with incomplete dataAI/Robotics benchmarks
Table 4. Side-by-side comparison (operational summary).
Table 4. Side-by-side comparison (operational summary).
Method/FrameworkWhat the Method DoesIdentical to EBPMWhat It Adds vs. EBPMWhat EBPM Adds
Hopfield/Attractor Networks [9]Recurrent dynamics converge into stored patterns (content-addressable recall from partial cues).Pattern completion from partial input; competition among states.Explicit energy formalism; well-characterized attractor dynamics.Explicit multimodality (sensory–motor–affective–cognitive) and engram units beyond binary patterns.
CLS—Complementary Learning Systems (hippocampus fast; neocortex slow) [10]Episodic rapid learning + slow semantic consolidation.Separation of “fast vs. slow” memory routes.System-level transfer between stores.EBPM is operational in real-time via direct matching; no mandatory inter-system consolidation for recognition.
Hippocampal indexing: pattern completion/separation [11,12]Hippocampus index episodes; CA3 supports completion, and DG supports separation.Fast episodic reactivation from indices; completion logic.Fine neuro-anatomical mapping (CA3/DG) with predictions.EBPM generalizes to neocortex and to multimodal action-oriented engrams, not only episodic replays.
Exemplar/Instance-Based (GCM; Instance Theory) [13,14]Decisions by similarity to stored instances; latencies fall with instance accrual.Direct “match to memory” with similarity score.Quantitative psychophysics of categorization and automatization.EBPM includes motor/interoceptive bindings and neuronal lateral inhibition/WTA selection.
Recognition-by-Components (RBC) [15]Object recognition via geons and relations.Sub-engrams akin to primitive features.Explicit 3D structural geometry.EBPM is not vision-limited; it covers sequences/skills and affective valence.
HTM/Sequence Memory (cortical columns) [5]Sparse codes, sequence learning, local predictions.Reuse of sparse distributed codes; sequence handling.Continuous prediction emphasis.EBPM does not require continuous generative prediction; for familiar inputs, it yields lower latency/energy via direct recall.
Sparse Distributed Memory (SDM) [16]Approximate addressing in high-D sparse spaces; noise-tolerant recall.Proximity-based matching in feature space.Abstract memory addressing theory.EBPM specifies engram units and novelty/attention gating with biological microcircuit motifs.
Memory-Augmented NNs (NTM/DNC) [17,18]Neural controllers read/write an external memory differentiably.Key-based recall from stored traces.General algorithmic read/write; task-universal controllers.EBPM is bio-plausible (engram units; STDP-like plasticity) without opaque external memory.
Episodic Control in RL [19]Policies exploit episodic memory for rapid action.Fast reuse of familiar episodes.RL-specific credit assignment and returns.EBPM operates for perception and action with multimodal bindings.
Prototypical/Metric-Learning (few-shot) [20]Classify by distance to learned prototypes in embedding space.Prototype-like matching ≈ EBPM encoder + similarity.Episodic training protocols; few-shot guarantees.EBPM grounds the embedding in neuro-plausible engrams + inhibition, not just vector spaces.
Predictive Coding/Active Inference [1,2,3,21,22]Top-down generative predictions + iterative error minimization.Shares attentional modulation and context use.Robust reconstruction under missing data via generative loops.EBPM delivers instant responses on familiar inputs (no iterative inference) with lower compute/energy; DIME then falls back to PC for novelty.
Table 5. (a). MNIST (single-seed, SEED = 42). (b) Fashion-MNIST. (c) CIFAR-10.
Table 5. (a). MNIST (single-seed, SEED = 42). (b) Fashion-MNIST. (c) CIFAR-10.
ModelFamiliar-Clean
(Acc/Ms)
Familiar-Noisy (Acc/Ms)Novel-Clean
(Acc/Ms)
Novel-Noisy
(Acc/Ms)
α ¯
(a)
EBPM98.63%/0.10097.13%/0.1320.00% */0.1260.00% */0.161-
PC96.11%/0.28156.28%/0.17794.40%/0.15437.27%/0.188-
DIME98.63%/0.24796.93%/0.1880.00%/0.1350.00%/0.2890.99/0.97/–/–
DIME-Prob96.47%/0.26458.68%/0.20893.65%/0.19736.16%/0.2390.61/0.38/0.41/0.23
(b)
EBPM82.56%/0.10152.35%/0.1390.00%/0.1470.00%/0.174-
PC83.14%/0.18250.45%/0.23896.50%/0.16097.00%/0.203-
DIME84.14%/0.13351.31%/0.17081.10%/0.16597.10%/0.2130.54/0.22/0.51/0.44
(c)
EBPM70.42%/0.19842.67%/0.1920.00%/0.2010.00%/0.242
PC65.99%/0.47727.15%/0.47883.10%/0.53650.95%/0.548
DIME-Prob **68.31%/0.52930.38%/0.53562.50%/0.60938.05%/0.675
* for EBPM, novel = 0% indicates explicit novelty detection (no engram completed), not misclassification. Sub-engrams are activated but recognition is withheld until learning occurs. ** best tuned fusion (gain = 4.0, thr_shift = 0.75, temp_ebpm = 6.0, temp_pc = 1.0).
Table 6. PyBullet obstacle course—means over 3 episodes (v3c_alpha, α = 0.65).
Table 6. PyBullet obstacle course—means over 3 episodes (v3c_alpha, α = 0.65).
ModeNoiseMs/Step
(Mean)
Steps
(Mean)
SuccessCollisions
(Mean)
EBPM0.0015.433811.00113
PC0.0018.043811.00113
DIME0.0018.443811.00113
DIME-Lazy0.0010.503811.00113
EBPM0.1512.553811.00113
PC0.1519.133811.00113
DIME0.1519.523811.00113
DIME-Lazy0.1519.173811.00113
Table 7. Ablation study results on MNIST (Familiar-Clean). Accuracy and latency are reported as mean ± standard deviation over five runs.
Table 7. Ablation study results on MNIST (Familiar-Clean). Accuracy and latency are reported as mean ± standard deviation over five runs.
ModelAccuracy (Mean ± Std)Latency (Ms/Sample ± Std)
EBPM97.78 ± 0.120.0031 ± 0.0009
PC (K = 0)96.97 ± 0.230.0034 ± 0.0009
PC (K = 5)96.97 ± 0.230.0461 ± 0.0038
PC (K = 10)96.97 ± 0.230.0694 ± 0.0108
PC (K = 20)96.97 ± 0.230.1082 ± 0.0024
DIME_ctrlOFF (α = 0.65)97.83 ± 0.080.0693 ± 0.0107
DIME_ctrlON97.81 ± 0.090.0648 ± 0.0011
Table 8. Paired statistical tests (t-tests) on CIFAR-10 across five seeds. Reported values include p-value, Cohen’s d effect size, and 95% confidence intervals (CIs). Significant results (p < 0.05) indicate that DIME significantly outperforms EBPM in familiar settings and reduces reliance on PC under uncertainty.
Table 8. Paired statistical tests (t-tests) on CIFAR-10 across five seeds. Reported values include p-value, Cohen’s d effect size, and 95% confidence intervals (CIs). Significant results (p < 0.05) indicate that DIME significantly outperforms EBPM in familiar settings and reduces reliance on PC under uncertainty.
ScenarioComparison Pairp-ValueCohen’s dCI LowCI Highn
Familiar-CleanEBPM vs. DIME_ctrlON0.000127−4.79−6.54−3.275
Familiar-CleanDIME_ctrlON vs. PC_K100.0008074.072.594.865
Familiar-NoisyEBPM vs. PC_K100.004774−2.54−13.97−4.795
Novel-CleanEBPM vs. PC_K100.000022−10.23−66.36−51.995
Novel-CleanDIME_ctrlON vs. PC_K100.000016−11.12−60.48−48.325
Table 9. Complexity analysis in PyBullet obstacle-course control. EBPM is fastest but fails under novelty; PC achieves robustness at the cost of high complexity; DIME balances the two by invoking PC only when necessary.
Table 9. Complexity analysis in PyBullet obstacle-course control. EBPM is fastest but fails under novelty; PC achieves robustness at the cost of high complexity; DIME balances the two by invoking PC only when necessary.
ModeLatency (Ms/Step, Mean ± Std)PC Calls (Mean ± Std)Success Rate
EBPM0.026 ± 0.0120.0 ± 0.01.0
PC4.01 ± 1.315214 ± 14291.0
DIME1.97 ± 1.082420 ± 12321.0
Table 10. Strengths and limitations of EBPM, PC, and DIME.
Table 10. Strengths and limitations of EBPM, PC, and DIME.
AspectEBPMPCDIME
SpeedHighMediumHigh
Energy EfficiencyHighLowMedium–High
Noise RobustnessMedium–LowHighHigh
FlexibilityLowHighHigh
Biological PlausibilityHighMediumHigh
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Vladu, I.C.; Bîzdoacă, N.G.; Pirici, I.; Cătălin, B. From Predictive Coding to EBPM: A Novel DIME Integrative Model for Recognition and Cognition. Appl. Sci. 2025, 15, 10904. https://doi.org/10.3390/app152010904

AMA Style

Vladu IC, Bîzdoacă NG, Pirici I, Cătălin B. From Predictive Coding to EBPM: A Novel DIME Integrative Model for Recognition and Cognition. Applied Sciences. 2025; 15(20):10904. https://doi.org/10.3390/app152010904

Chicago/Turabian Style

Vladu, Ionel Cristian, Nicu George Bîzdoacă, Ionica Pirici, and Bogdan Cătălin. 2025. "From Predictive Coding to EBPM: A Novel DIME Integrative Model for Recognition and Cognition" Applied Sciences 15, no. 20: 10904. https://doi.org/10.3390/app152010904

APA Style

Vladu, I. C., Bîzdoacă, N. G., Pirici, I., & Cătălin, B. (2025). From Predictive Coding to EBPM: A Novel DIME Integrative Model for Recognition and Cognition. Applied Sciences, 15(20), 10904. https://doi.org/10.3390/app152010904

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop