Next Article in Journal
Results from Cryo-PoF Project: Power over Fiber at Cryogenic Temperature for Fundamental and Applied Physics
Next Article in Special Issue
Scattering-Based Machine Learning Algorithms for Momentum Estimation in Muon Tomography
Previous Article in Journal / Special Issue
Information Field Theory for Two Applications in Astroparticle Physics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unsupervised Particle Tracking with Neuromorphic Computing

1
Dipartimento di Fisica e Astronomia, Università di Padova, Via F. Marzolo 8, 35131 Padova, Italy
2
Dipartimento di Fisica, Università di Bologna, Via Irnerio, 40126 Bologna, Italy
3
Department of Computer Science, Electrical and Space Engineering, Luleå University of Technology, 971 87 Luleå, Sweden
4
Istituto Nazionale di Fisica Nucleare, Sezione di Padova, Via F. Marzolo 8, 35131 Padova, Italy
5
Department of Physics, Central University of Tamil Nadu, Thiruvarur 610 001, India
*
Authors to whom correspondence should be addressed.
Universal Scientific Education and Research Network.
Particles 2025, 8(2), 40; https://doi.org/10.3390/particles8020040
Submission received: 1 February 2025 / Revised: 28 February 2025 / Accepted: 1 April 2025 / Published: 7 April 2025

Abstract

:
We study the application of a neural network architecture for identifying charged particle trajectories via unsupervised learning of delays and synaptic weights using a spike-time-dependent plasticity rule. In the considered model, the neurons receive time-encoded information on the position of particle hits in a tracking detector for a particle collider, modeled according to the geometry of the Compact Muon Solenoid Phase-2 detector. We show how a spiking neural network is capable of successfully identifying in a completely unsupervised way the signal left by charged particles in the presence of conspicuous noise from accidental or combinatorial hits, opening the way to applications of neuromorphic computing to particle tracking. The presented results motivate further studies investigating neuromorphic computing as a potential solution for real-time, low-power particle tracking in future high-energy physics experiments.

1. Introduction

The aspiration of enhancing the scientific potential of High Energy Physics (HEP) experiments leads to an extremely large data volume [1]. Conventional computing solutions alone struggle with the demand for online identification and reconstruction of particle signals [2,3]. Another significant challenge lies in the temporal realm. Given that particles close to light speed traverse 3 c m in just 100 picoseconds, the exploitation of time patterns of detected signals requires sensitivity to sub-nanosecond time intervals. Current detectors neglect the temporal information generated by the passage of particles through sensitive material, while it could, in principle, help extract further information. The integration of fast online dimensional reduction and spatio-temporal pattern recognition using parallel analog-digital neuromorphic computing architectures at the detector end may allow us to overcome these limitations.
Throughout this paper, we use natural units where = c = 1 , so that transverse momentum ( p T ), energy, and mass are all expressed in units of GeV.

1.1. Neuromorphic Computing

As digital computing technologies approach their physical and architectural limits, alternative mixed-signal processing methods that exploit intrinsic physical properties of materials are investigated [4], a concept sometimes referred to as ‘in-materia’ and ‘analog in-memory’ computing. A prominent example is neuromorphic computing and engineering [5], where sensors, processors, and cybernetic system architectures are developed using biological brains as inspiration. Brains process massively parallel event-based representations of uncertain information, providing new insights into how spatio-temporal detector signal patterns can be efficiently encoded and compressed.
Neuromorphic solutions imitate mixed-signal circuits and architectural motifs inspired by the brain to improve efficiency, robustness, and learning [3,6]. Brains demonstrate orders-of-magnitude more energy-efficient learning and intelligence solutions and show that relatively slow, stochastic, and plastic circuits can promote robust processing of unstructured and noisy avalanches of sensor information. For example, the FPGA-based neuromorphic supercomputer DeepSouth, presently under construction, aims to reach above 200 trillion synaptic operations per second at 40 k W power, while a human brain with a comparable number of neurosynaptic operations (and higher complexity) requires about 20 W [7].
Although the time constants of biological neurons are matched to the long, behaviorally relevant timescales of milliseconds or more, the general pattern-learning capacity of such neuro synaptic circuits can be generalized to other environments. By implementing circuits that mimic useful aspects of the neuro synaptic mixed-signal dynamics in, e.g., nanoscale semiconductors and photonic devices, neuromorphic circuits with sub-nanosecond time constants can be realized [8,9,10]. This opens new opportunities for developing efficient event-triggered information sampling and spiking neural network processing solutions [6,11] for high-energy physics detector readouts.

1.2. Spiking Neural Networks

Spiking Neural Networks (SNNs) are used to model biological neurons using differential equations describing neurosynaptic dynamics at some spatial and temporal approximation level [12]. Unlike conventional artificial neural networks (ANNs), which use continuous values to represent the activation of neurons, SNNs are based on discrete spikes that neurons generate in response to incoming stimuli. These spikes occur at precise points in time, adding a temporal dimension to neural processing that enhances the model’s capacity to asynchronously process time-dependent information efficiently. This spike-based approach makes SNNs particularly well suited for event-based spatio-temporal processing and neuromorphic computing, where neurosynaptic dynamics and asynchronous parallel processing are efficiently implemented using specialized circuits and materials.
The modeling of spiking neurons within SNNs typically involves integrating and firing action potentials represented as spikes, with each neuron having a membrane potential that varies in response to synaptic inputs; see Figure 1. When this potential surpasses a threshold or the dynamical system reaches an unstable fixed point, the neuron emits one or several spikes that are transmitted to connected neurons with some delays. Various mathematical models, such as the Leaky Integrate-and-Fire (LIF) model and the (adaptive) exponential integrate-and-fire model, are commonly used to describe this process. These models vary in complexity but share the core concept that the timing of incoming signals drives neuronal activity.
In neuromorphic computing, SNNs are used to model neural networks and brain-like algorithms in a biologically plausible way, taking advantage of the temporal coding of spikes. This makes them suitable for applications in spatio-temporal pattern recognition, sensor fusion, and real-time signal processing tasks that are often central to particle detection and analysis. By leveraging SNNs, neuromorphic systems can process high-dimensional events with complex asynchronous temporal structures.
In the context of particle detectors, SNNs offer significant potential for sparse encoding and processing of detector events. Their ability to process time-series data, such as signals from detectors that change over time or are influenced by various factors, could lead to more efficient, robust, and low-power systems for real-time dimension reduction, triggering, and event classification. As the field of neuromorphic computing continues to evolve, SNNs can have a pivotal role in revolutionizing how particle detection systems process information, making them an area of active research within both the neuromorphic and particle physics communities.

1.3. Neuromorphic Computing in High-Energy Physics

Particle physics has a long history of synergy with computer science developments. Some of the applications required to study subnuclear matter and interactions, in fact, often require access to extensive computing resources, innovative algorithms, and specialized computing hardware. Given this history, it is no surprise that in recent years, particle physics experiments have extensively adopted deep learning technologies, integrating them into their data analysis workflows [14]. In parallel, a number of efforts have been directed toward the integration of neural networks and other machine-learning models in online data acquisition [15]. Quantum Computing (QC) developments are also being followed closely by the HEP community, with a view to offering specific use cases where QC may provide successful encoding and implementation of the computing tasks [16].
A similar trajectory is now starting to become apparent for the Neuromorphic Computing (NC) paradigm. While the specific points of strength of NC technologies over traditional digital computing are not necessarily aligned with the most pressing demands of particle physics experiments in design or commissioning, there are situations where NC can provide advantageous alternative solutions that open a view into entirely different design concepts. In this study, we focus on the Compact Muon Solenoid (CMS) Phase-2 detector at the High-Luminosity Large Hadron Collider (HL-LHC) as a case study to demonstrate the potential of neuromorphic computing in HEP. The challenging conditions present in the detector, with an average of 200 simultaneous proton–proton collisions per LHC bunch-crossing (also referred to as pileup), present a significant computational challenge for real-time track reconstruction. By leveraging SNNs, we aim to address these challenges while maintaining high efficiency and low fake rates.
The structure of this article is as follows. In Section 2 we summarize how particle trajectories are identified and measured in the CMS tracking system, which is the use case on which we focus our attention in our study. In Section 3, we describe the SNN model we employ to demonstrate how a neuromorphic system may identify patterns of hits left by charged particles in silicon sensors in an unsupervised way. We discuss the data samples we generated for our study in Section 4. In Section 5 we describe the tuning of hyperparameters of the SNN model, performed with a two-staged approach also employing a genetic algorithm. We detail our results in Section 6 and offer some concluding remarks in Section 7.

2. Track Reconstruction with the CMS Phase-2 Experiment at HL-LHC

The Phase-2 CMS detector incorporates significant upgrades to its tracking system to meet the challenges of the High-Luminosity LHC (HL-LHC) [17]. Figure 2 illustrates the layout of the silicon sensors within a single sector of the Phase-2 CMS tracker. Charged particle trajectories are reconstructed using a sophisticated iterative algorithm, which processes detector hits and ensures robust track reconstruction under the challenging conditions of the HL-LHC:
  • Seeding: Track seeds are formed using hits from the detector, focusing on high-efficiency seed generation even in high-density environments.
  • Trajectory building: Using the Kalman filter, the algorithm extends the seed through the detector layers, accounting for multiple scattering and energy loss.
  • Fitting: A final fit refines the trajectory, providing precise momentum, charge, and vertex information.
Figure 2. Layout of the silicon sensors in one sector of the Phase-2 CMS tracker. The Inner Tracker (green and yellow) is made by pixel sensors, while the Outer Tracker (red and blue) is made by modules built using both macro-pixel and strip sensors. This is an evolved layout from the TDR, available at [18].
Figure 2. Layout of the silicon sensors in one sector of the Phase-2 CMS tracker. The Inner Tracker (green and yellow) is made by pixel sensors, while the Outer Tracker (red and blue) is made by modules built using both macro-pixel and strip sensors. This is an evolved layout from the TDR, available at [18].
Particles 08 00040 g002
This iterative approach prioritizes high-momentum and prompt tracks early and handles low-momentum tracks and complex patterns in subsequent iterations, improving efficiency and fake rejection. The following key performance metrics emphasize the capabilities of the track reconstruction process and its impact on the overall experimental performance [17].
  • Track efficiency: The track reconstruction algorithm achieves exceptionally high efficiency (defined as the fraction of charged particles in a given momentum range that are identified by the algorithm). For charged particles with transverse momentum p T > 1 GeV , it exceeds 99% in the central region (pseudorapidity | η | < 1.5 ), and it still remains above 95% in the forward region ( 1.5 < | η | < 4.0 ).
  • Transverse momentum resolution: Phase-2 track reconstruction achieves a resolution of transverse momentum (defined as the relative uncertainty in transverse momentum, δ p T / p T ) that is better than 1–2% for high- p T tracks in the central region ( | η | < 1.5 ), which is critical for high-precision measurements of particle momenta. Although the resolution slightly degrades in the forward region due to the increased material and multiple scattering effects, it remains within acceptable bounds for reliable track reconstruction across the entire detector.
  • Fake track rate: the fraction of reconstructed tracks that do not correspond to real particle trajectories, often arising from noise, overlapping hits, or mis-reconstruction, is expected to be 0.5% to 1% for tracks with p T > 1 GeV .
Efficient computing timing for track reconstruction in the CMS Phase-2 detector is crucial for handling the high-luminosity, high-pileup conditions expected at the HL-LHC. Key performance optimizations include multithreading, GPU acceleration, and potentially machine learning algorithms to handle the increased event complexity while maintaining high precision and efficiency.

3. Spiking Neural Network Model for Particle Tracking

3.1. SNN Architecture

The model employed in this work builds upon the SNN architecture proposed by Masquelier et al. [19], specifically designed to perform spike-timing-based learning via Spike-Timing-Dependent Plasticity (STDP). This architecture integrates LIF neurons to achieve the recognition of complex spatio-temporal patterns in a noisy environment facilitated through an unsupervised learning process.
In neuroscience, afferent axons are nerves that transmit information from sensory receptors to the central nervous system [20]. By extension, in this paper, we use the term ‘afferent’ to refer to the channels that carry input spikes from the CMS tracker to the SNN.
Unlike the model proposed by Masquelier et al., the architecture employed in this work consists of two primary layers, L 0 and L 1 , which are densely connected to afferents serving as channels for input signals. Each afferent corresponds to a specific input source, which introduces spikes into the network, simulating a diverse array of sensory inputs. Each of the N L 0 N L 1 neurons of the layer L 0 L 1 is characterized by the activation threshold T 0 T 1 . Furthermore, each synapse j linked to an afferent has a synaptic delay d j , which are additional degrees of freedom that our model exploits with a novel learning algorithm presented in Section 3.4. Similar unsupervised learning rules for synaptic delays have been studied in other works, such as [21,22,23]. A simplified scheme of the network architecture employed for this study is shown in Figure 3.

3.2. Initialization of the Synaptic Weights and Delays

The initial synaptic weights are drawn from a Gaussian distribution with mean μ = 1 and standard deviation σ = 2 / N afferents = 2 / 10 . Then, they are normalized so that their sum is unitary. Synaptic delays are initialized to random values within the range [ d max / 2 Δ , d max / 2 + Δ ] where Δ is a hyperparameter set to ensure sufficient temporal spread and d max is the maximum value allowed for the synaptic delay. These initial values were chosen to provide a diverse starting point for the unsupervised learning of spatio-temporal patterns.

3.3. Numerical Simulation of the Neuron Potentials

The post-synaptic potential (EPSP) produced by an incoming spike at time t j is computed as:
ϵ ( t t j ) = K · exp t t j τ m exp t t j τ s · θ ( t t j )
Here, τ m and τ s represent the membrane and synaptic time constants, respectively, and θ ( · ) is a unit step function. K is a scaling constant such that the peak of the potential is set to 1: max t ϵ ( t ) = ϵ t max = 1 . This potential change captures the dynamics of how an individual synapse contributes to the membrane potential of a neuron upon receiving an input spike. After reaching a certain threshold T, the neuron fires. Upon firing at time t i , the neuron experiences a reset in its membrane potential, described by:
η ( t t i ) = T · K 1 · exp t t i τ m K 2 · exp t t i τ m exp t t i τ s · θ ( t t i )
with K 1 and K 2 acting as constants that dictate the post-firing behavior. Inhibitory Post-synaptic Potentials (IPSPs) are incorporated to introduce competition between neurons within the network. When a neuron fires at time t k , the resulting inhibitory potential it sends to its neighbors is expressed as
μ ( t t k ) = α · T · ϵ ( K μ ( t t k ) )
where α is a coefficient representing the strength of inhibition, and K μ adjusts its temporal extension. This competitive interaction among neurons promotes selective firing, preventing over-activation of the network. As a consequence of the IPSPs, a ‘Winner-Takes-All’ competition mechanism emerges, where neurons compete to respond to input patterns. This process drives specialization, with neurons becoming increasingly tuned to distinct features of the input streams. Thus, at a given point in time, the membrane potential of a neuron is given by
p ( t ) = η ( t t i ) + j | ( t j + d j ) > t i w j · ϵ ( t ( t j + d j ) ) + k | t k > t i μ ( t t k )
with w j and d j representing the synaptic weight and delay. This defines the model of a LIF neuron. Figure 4 shows a simulation of the time evolution of the membrane potential of 3 different neurons. The main advantage of the model is that the calculations on potentials are performed only when a neuron receives or emits spikes, making it extremely efficient compared with solving a differential equation that describes potential dynamics.

3.4. Modified STDP for Unsupervised Synaptic Delay Learning

STDP is a learning mechanism inspired by biological processes that modify synaptic weights according to the time difference between incoming spikes and neuronal activation [24,25]. Within SNNs, STDP provides a framework for unsupervised learning. The Hebbian STDP model in [19] is formulated as follows:
Definition 1
(Hebbian STDP rule for Synaptic Weights). 
Δ w j = a + · exp t j t i τ + if t j t i Synaptic Long - Term Potentiation a · exp t j t i τ if t j > t i Synaptic Long - Term Depression
where t j denotes the pre-synaptic spike arrival time, t i is the neuron activation time, a + and a define the maximum weight update, and τ + and τ are constants that control the duration of the time window in which spikes lead to either synaptic potentiation or depression.
According to this rule, if an input spike arrives prior to neuronal activation, the synaptic weight increases, reinforcing the causality relation; conversely, if it arrives afterward, the weight decreases.
This rule alone has proven insufficient for the purposes of this work, as it fails to lead neurons to specialize in recognizing distinct patterns (see Section 5.2). To address this issue, we propose a modified version of this rule (see Figure 5) specifically for learning synaptic delays instead by incorporating the following:
  • Delay potentiation (DLTP) If a pre-synaptic spike occurs before the neuron’s activation, the delay is increased, effectively delaying the pre-synaptic spike.
  • Delay depression (DLTD) If a pre-synaptic spike occurs after the neuron’s activation, the delay is decreased, advancing the pre-synaptic spike.
Figure 5. Model of spike-time-dependent plasticity of connection delays implementing a modified Hebbian learning concept. The x-axis denotes the time difference, Δ t = t j t i + t max , and the y-axis, Δ d , is the corresponding update of the delay. The parameters defining this unsupervised learning rule were found by the genetic algorithm.
Figure 5. Model of spike-time-dependent plasticity of connection delays implementing a modified Hebbian learning concept. The x-axis denotes the time difference, Δ t = t j t i + t max , and the y-axis, Δ d , is the corresponding update of the delay. The parameters defining this unsupervised learning rule were found by the genetic algorithm.
Particles 08 00040 g005
Through this mechanism, neurons can adjust connection delays (heretofore addressed as “synaptic delays”) to minimize the temporal pattern length, increasing the activation likelihood. Analytically, this rule is defined as follows:
Definition 2
(STDP rule for synaptic delays). 
Δ d j = d + exp t j t i + t max τ d + exp t j t i + t max τ d + if t j t i t max DLTP , d exp t i t j t max τ d exp t i t j t max τ d if t j > t i t max DLTD , 0 if synapse j is linking two neurons
where t j denotes the pre-synaptic spike arrival time, t i represents the neuron’s activation time, t max indicates the EPSP signal’s peak time, d + and d are the learning rates, and τ d + , τ d + , τ d , and τ d are the time constants for potentiation and depression, respectively. The rule updates the delay of synapses linking afferents to neurons.
As a regularization method, the sum of the synaptic delays of each neuron remains unchanged throughout the process, with a renormalization step conducted after each update. Furthermore, delays are clipped in a definite range d j [ 0 , d max ] .

3.5. Spatio-Temporal Information Encoding of Detector Hits

The challenge of encoding information involves determining how to map the three-dimensional coordinates ( r , ϕ , η ) of hits recorded by the tracker into input spikes ( a , t ) received at a specific time t from a specific afferent a. Given our objective to distinguish particles from noise and quantify their transverse momenta, we initially concentrate on the radial coordinate and azimuthal angle, postponing the integration of the third dimension for future investigations. This approach effectively considers events as projected onto the transverse plane. Furthermore, our current focus is limited to events involving only the Inner Tracker and Outer Tracker barrel, deferring any inclusion of the endcaps for later expansions. The geometry corresponds to what is displayed in Figure 6. Specifically, we assign to each tracking layer a unique afferent and scan the tracker counterclockwise beginning from the positive half of the x-axis. Define the angular reading speed as ω = 2 π · f , with f = 40 MHz representing the LHC event rate. The j t h hit ( r j , ϕ j ) is thus encoded into a spike ( a j , t j ) where a j corresponds to the afferent related to its tracking layer and t j = ϕ j + 2 π · θ ( ϕ j ) ω represents the arrival time. Consequently, we establish a mapping r j a j and ϕ j t j that remains monotonic and continuous for signal tracks, except at the discontinuity transitioning between ϕ = 2 π rad and ϕ = 0 .
The implemented signal encoding method faces challenges with edge cases near the ϕ = 2 π rad to ϕ = 0 transition, where trajectories overstep the boundary of the encoding region. To resolve this, the encoding framework is expanded to include an auxiliary region ϕ [ 0 , δ ] with δ = 0.7 rad . This covers the maximum anticipated angular discrepancy between the closest and farthest hits for the lowest momentum track ( 1 GeV ), ensuring continuous signal representation and preserving monotonicity, albeit introducing minimal redundancy.
To validate the proposed SNN model, a carefully simulated dataset is essential. This dataset must capture the spatio-temporal patterns of charged particle trajectories while incorporating realistic noise and detector geometry. As illustrated in Figure 7, the dataset exhibits distinct clusters of signal spikes amid background noise, effectively preserving the event-based nature of the encoding. In the following section, we describe the dataset preparation process and its role in training and evaluating the SNN. In the following section, we describe the dataset preparation process and its role in training and evaluating the SNN.

4. Simulated Dataset for Training and Validation

The dataset employed for training the SNN was derived from Monte Carlo simulations based on the CMS Phase-2 detector geometry. Half of the events contain just random noise, while the other half were generated assuming the production of a single particle (either a muon or an antimuon) without pile-up conditions. The kinematic properties of the particles were uniformly distributed in azimuthal angle ϕ [ π , π ] and pseudorapidity η [ 1 , 1 ] , focusing on the barrel region while excluding the endcap disks and the tilted detector modules in the barrel. Three transverse momentum classes were considered, corresponding to p T { 1 , 3 , 10 } GeV .
The spatial density of hits ρ ( x ) is modeled as ρ ( x ) F ( x ) · G ( x ) , where F ( x ) represents the structural features of the tracker (vanishing in regions without active sensors) and G ( x ) accounts for the trajectory geometry. Given that the primary events were uniformly distributed in ϕ and η , the factor G ( x ) is proportional to 1 / r . The generation of the background hits for network training and testing has been performed following a custom generation process, extracting random signal hits through an inversion-based sampling algorithm:
  • Hits from the Monte Carlo simulations were collected into a reference set.
  • Each hit was assigned a weight equal to 1.
  • A cumulative distribution function P ( i ) = j i r j was computed for the reference set.
  • A random value x was drawn uniformly in [ 0 , P max ] , where P max is the maximum value of P ( i ) .
  • The corresponding hit index i hit was determined by inverting P ( i ) .
  • Steps 4 and 5 were repeated until the desired number of noise hits was obtained.
This approach maintains the spatial consistency of the noise distribution with respect to the detector geometry. The number of hits of noise in an event is supposed to follow a Poisson distribution with a given average N bg . Figure 8a illustrates a representative event where N bg = 100 noise clusters generated using this method are overlaid onto the primary signal. In contrast, Figure 8b shows the event with 300 background noise clusters. This dataset preparation strategy ensures a balanced and realistic input for network training, preserving both signal fidelity and background complexity. Figure 9, Figure 10 and Figure 11 show the distributions of the generated datasets.
The next step involves training and optimizing the SNN to handle high-noise environments effectively. This requires fine-tuning hyperparameters such as synaptic delays and learning rates to maximize efficiency and selectivity while minimizing fake rates. The following section describes our approach to hyperparameter optimization using genetic algorithms.

5. Methodology—Hyperparameter Search and Optimization

In this section, we present the definitions of the utilities used to measure the performance of the SNN (Section 5.1), discuss the problem of the specialization of the neurons (Section 5.2), and explain how we tackled hyperparameter tuning via a genetic algorithm (Section 5.3 and Section 5.4).

5.1. Utilities Definition

The index n represents the N neurons neurons, while c denotes the index corresponding to the N classes particle classes, using a notation where c = 0 refers to events containing only background hits. Alternatively, to specify a particular class, the notation c = ( q , p T ) is used, where q { 1 , + 1 } represents the charge state, and p T { 1 , 3 , 10 } GeV corresponds to the transverse momentum. The symbol ϵ serves as the index over the N ϵ events, while ϵ c specifically denotes events that include a track from a given class c. To evaluate the functionality of the network and monitor its learning progression, the following definitions are introduced.
Definition 3
(Neuron Activation Indicator Function). This function is used to identify events in which the neuron n has been activated at least once. A neuron is considered “activated at least once” during an event if its membrane potential exceeds the firing threshold T 0 or T 1 within the duration of the event.
1 n ( ϵ ) = 1 if the neuron n has activated at least once during the event ϵ 0 otherwise
Definition 4
(Network Activation Indicator Function).
1 ( ϵ ) = 1 if at least one neuron in the network has activated during event ϵ 0 otherwise
Definition 5
(Acceptance per neuron per class). Measures the fraction of events of class c in which the neuron n has been activated at least once.
A n , c = ϵ c 1 n ( ϵ c ) ϵ c 1
Definition 6
(Fake rate per neuron). Measures the fraction of events containing just hits of background in which the neuron n has been activated at least once.
F n = A n , 0 = ϵ 0 1 n ( ϵ 0 ) ϵ 0 1
Definition 7
(Aggregate acceptance per class). Measures the fraction of events of class c in which at least one neuron in the network has been activated.
A c = ϵ c 1 ( ϵ c ) ϵ c 1
Definition 8
(Aggregate fake rate). Measures the fraction of events containing just hits of background in which at least one neuron in the network has activated.
F = A 0 = ϵ 0 1 ( ϵ 0 ) ϵ 0 1
Definition 9
(Selectivity of the network). Selectivity quantifies the ability of the network to discriminate patterns and is derived from mutual information by comparing the distribution of activations across neurons and particle classes:
S = n , c P n , c · log 2 P n , c + δ P c · P n
with P n , c = ϵ c 1 n ( ϵ c ) N ϵ , P n = c = 1 N classes P n , c , P c = n = 1 N neurons P n , c and δ 1 to avoid numerical instabilities.

5.2. The Problem of Specializing Neurons

Initially, given the similarity of the task, the network was operated using the optimal parameters found in [19], with all synaptic delays set to zero and weight learning enabled only. This configuration resulted in high acceptance rates (greater than 90%) and low fake rates (below 5%). For aggregated results, see Table 1. However, with this configuration, the neurons did not achieve meaningful specialization, as shown in Figure 12. Certain neurons, such as neuron 8, exhibited excessive responsiveness across all classes, while others remained underutilized, highlighting considerable redundancy and inefficiency in the network’s behavior.
Addressing this issue required increasing the network’s complexity, which proved to be the key solution. The introduction of synaptic delays and a delay-learning mechanism allowed neurons to fine-tune their activation times based on incoming spikes. As a result, the network achieved more precise alignment with specific spatio-temporal patterns, ultimately fostering meaningful specialization. This development is thoroughly analyzed and presented in Section 6.1.

5.3. Hyperparameter Tuning and Genetic Algorithm

To improve the model’s ability to recognize particle trajectories, we introduced synaptic delays, which required precise hyperparameter tuning. Due to the high dimensionality of the parameter space, we employed a genetic algorithm (GA) to efficiently optimize network configuration.
The optimization process focused on a set of key hyperparameters governing synaptic and neuronal dynamics. These include temporal scaling factors, reset potential adjustments, membrane potential thresholds, and learning rates for synaptic delay modulation. The genetic algorithm was designed to efficiently navigate this high-dimensional space by iteratively improving parameter sets based on performance criteria. Table 2 summarizes the main hyperparameters considered in the optimization.
The optimization of these parameters is performed using NSGA-II (Non-dominated Sorting Genetic Algorithm II), a state-of-the-art multi-objective evolutionary algorithm [26]. Implemented via the pyGAD library [27], NSGA-II is particularly suited for problems involving multiple conflicting objectives, such as maximizing network efficiency while minimizing the fake rate and maximizing selectivity. The algorithm employs a random mutation strategy and a single-point crossover mechanism to generate new candidate solutions, with parent selection based on non-dominated sorting to maintain population diversity across generations. Each iteration evaluates the fitness of candidate solutions by simulating the SNN and analyzing key performance metrics. The process continues until convergence towards an optimal set of hyperparameters is achieved, ensuring a robust and efficient network configuration.

5.4. Optimization Workflow for Spiking Neural Networks in Noisy Environments

To achieve an SNN capable of classifying particles with varying charges and momenta in dense and noisy environments, an optimization workflow divided into different stages was employed. This progressive strategy systematically exposes the network and the accompanying GA to increasingly complex tasks, ensuring robustness and generalization. The workflow is outlined as follows:
  • Stage 1: Simplified Problem with Minimal Background
    -
    Background noise is controlled, with the average number of background hits set to N bg = 100 .
    -
    The GA optimizes the network for a reduced classification task focused on identifying negative muons with transverse momenta p T { 1 , 3 , 10 } GeV , resulting in a three-class classification problem.
    -
    Delay learning, a mechanism for modulating synaptic delays, is employed to enhance selectivity and minimize false positives.
    -
    The hyperparameter search space for this stage includes all the ones defined in Table 2.
  • Stage 2: Incorporating Antimuon Classification
    -
    The optimal network configuration from Stage 1 is selected as a baseline. The most active and specialized neurons are preserved, while their delay properties are mirrored and adjusted to initialize new neurons. Initial individuals in the GA are mutations of this configuration.
    -
    Antimuon events are introduced into the training and validation datasets, expanding the classification task.
    -
    The hyperparameter search space is refined by fixing the values of K 1 , K 2 , τ m , τ s , τ d , τ d + , τ d , τ d + , and constraining the ranges of T 0 , T 1 , K μ , α .
    -
    Delay learning rates ( d ± ) are reduced to allow finer convergence during training.
  • Stage 3: Evaluating Network Robustness Across Noise Levels
    -
    The optimal network from Stage 2 undergoes additional refinement, where neurons with high activation and specialization are retained, and less significant neurons are pruned.
    -
    The network is tested at varying noise levels to assess robustness. Based on these results, the network undergoes further retraining at elevated noise levels during the next stage.
  • Stage 4: Fine-tuning at High Noise Levels Results for this stage are presented in Section 6.1.
    -
    The optimized network from Stage 3 is selected as the basis for this stage. Initial individuals in the GA are mutations of this configuration.
    -
    Background noise is increased, with the average number of background hits raised to N bg = 300 .
    -
    The hyperparameter search space is limited to T 0 , T 1 , K μ , α , τ m , τ s , with narrower parameter ranges to focus on fine adjustments.
    -
    The delay in learning rates ( d ± ) is further reduced to achieve incremental improvements in accuracy and robustness.

6. Results

6.1. Single Track Pattern Recognition Under High Levels of Noise

Before presenting the final results under high noise conditions, we first highlight key findings of Stages 1 and 3, which shaped our approach to improving the optimization process. Figure 13 illustrates the evolution of the synaptic delays during Stage 1 for three neurons specializing in the recognition of muons with distinct transverse momenta ( p T = 1 , 3, and 10 GeV ). By processing a total of 20,000 training events, the delays transitioned from random initial values to stable configurations in an unsupervised manner, effectively aligning the network with the corresponding spatio-temporal patterns. Figure 14 illustrate the evolution of acceptance, selectivity, and fake rate as functions of the number of training events processed by an SNN initialized with the same hyperparameters as the best-performing individual from Stage 1. The plots indicate that after 20,000 training events, both acceptance and selectivity are already near their maximum values, while extended training could further reduce fake rates exponentially. This trend is further supported by Figure 15a, which demonstrates how three L0 neurons progressively adjust their firing rates toward the expected value for a perfectly specialized neuron under these conditions (red dashed line) while the other firing rates are lower by an order of magnitude. Notably, these three neurons correspond to the most specialized units within the analyzed network.
In Stage 3, the robustness of the network was tested by training it on events with N bg = 100 noise hits and then evaluating its performance under varying noise levels. The outcomes, summarized in Figure 16, reveal a general decline in neuron acceptances and an exponential increase in fake rates for N bg test 200 noise hits. These results underscore that network performance is closely related to noise levels in the training environment. In particular, specific time constants that regulate the build-up of neuron potentials, such as τ m and τ s , are crucial in determining the robustness of the network at different levels of noise.
For the reason mentioned above, the SNN was re-optimized and assessed under high-noise conditions by simulating an average of 300 background clusters per event ( N bg = 300 ). This noise level represents a more challenging and more realistic environment for the running conditions in the HL-LHC. In fact, studies conducted on the central region of the CMS Phase-1 detector for the t t ¯ process at a pileup level of 60 estimates that, on average, each reconstructed signal particle generates one cluster for every eight clusters originating from noise or non-reconstructed particles [17]. Extrapolating these results to the barrel region of the Phase-2 tracker, we conservatively estimate that each reconstructed signal particle may correspond to as many as 80 spurious clusters not associated with the main interaction. Consequently, simulating 300 background clusters per event provides a meaningful and interesting test scenario for evaluating the network’s capacity to identify true particle trajectories amidst significant combinatorial noise. Figure 8b provides an example of a transverse plane projection of an event with this level of noise. A complete list of the final hyperparameters is available in Table A1. The aggregate performance of the network in terms of signal acceptance (A), fake rate (F), and selectivity (S) is summarized in Table 3. Key findings include the following:
  • High acceptance: the SNN achieved greater than 98% acceptance for all transverse momentum classes, with peak values reaching 100% for particles with p T = 10 GeV .
  • Low fake rate: The network effectively suppressed noise, achieving a fake rate of approximately 3.0 ± 0.2 % . This aggregate fake rate is higher than the individual fake rates of neurons (<1%) because it accounts for the global response of the entire network: even if only a single neuron activates in a background-only event, the event is considered a false positive.
  • Strong selectivity: a selectivity score of 3.48 indicates robust discrimination between signal and noise patterns.
Table 3. Aggregate acceptance and fake rate of the network for different patterns and average noise level. Each test dataset contains N ev test = 25,000 events.
Table 3. Aggregate acceptance and fake rate of the network for different patterns and average noise level. Each test dataset contains N ev test = 25,000 events.
N bg test A 1 [%] A + 1 [%] A 3 [%] A + 3 [%] A 10 [%] A + 10 [%]F [%]S
300 98.2 ± 0.3 98.2 ± 0.3 99.90 ± 0.08 99.2 ± 0.2 100.00 ± 0.05 100.00 ± 0.05 3.0 ± 0.2 3.48
Figure 17 presents the final acceptance of individual neurons across the different particle classes. The results demonstrate the remarkable specialization of neurons in recognizing specific patterns associated with distinct particle trajectories, even under high-noise conditions. The analysis highlights the performance of the SNN in classifying particle trajectories, even in high-noise conditions. Each neuron exhibits a distinct preference for a specific particle class, with at least one neuron achieving an acceptance rate exceeding 90% for every class. This specialization indicates that the network effectively learns to recognize characteristic spatio-temporal patterns associated with different transverse momentum p T values and charge states. Furthermore, the network demonstrates a low level of confusion across classes, with overlap in neuronal responses remaining below 5% in most cases. This ability to maintain clear distinctions between particle classes underscores the network’s effectiveness in accurately discriminating between different trajectories. Additionally, the fake rate per neuron, as illustrated in the first row of the heatmap, is consistently low, on the order of 10 3 , which highlights the robustness of the network in suppressing background noise even under challenging conditions.
The network’s reliance on synaptic delays to temporally align spike-based inputs with neuronal firing played a critical role in achieving these results. Figure 18 illustrates the final synaptic delay configurations between neurons, revealing distinct temporal specializations for different classes of particles. In particular, neurons that specialize in lower p T patterns exhibit a wider spread in the magnitudes of synaptic delays compared with those that specialize in higher p T patterns. Figure 19 illustrates the spatio-temporal spike patterns under high noise conditions and the activation of different neurons within the network. It is particularly noticeable that neurons activate in a time window close to the arrival of spikes associated with signal hits, whereas they do not activate in periods primarily occupied by spikes associated with noise.
While the single-track results demonstrate the feasibility of neuromorphic computing for real-time track reconstruction, high-energy physics experiments require the reconstruction of multiple overlapping particle trajectories per event. The next section proposes a preliminary study on how the proposed SNN model scales to multi-track events.

6.2. Searching for Multiple Tracks

Although the network described in Section 6.1 was specifically trained for the reconstruction of single tracks, no fundamental limitations were identified that would prevent its extension to multiparticle tracking within the same event. Figure 20 presents an example of spatio-temporal spike patterns for events containing 10 tracks, demonstrating that, even in this more complex scenario, neurons activate according to the input tracks. This behavior arises from the fact that the SNN processes events as a time series, primarily relying on local information near the tracks. As a result, the primary factor affecting performance is the angular separation between tracks. When two tracks are too close in ϕ , the following challenges may arise:
  • Temporal overlap: tracks appearing within the same time window may produce closely spaced spikes, making it difficult for the network to distinguish separate trajectories.
  • Refractory periods: once a neuron fires in response to one track, it enters a refractory period, which may hinder the detection of other nearby tracks.
  • Neuron competition: Lateral inhibition is a crucial mechanism that prevents multiple neurons from firing for the same pattern, allowing them to specialize. However, when two patterns occur within the same time window, this mechanism can become counterproductive, reducing the network’s ability to specialize effectively.
Figure 20. Spatio-temporal spike patterns in events containing 10 tracks. In the bottom graph, the x-axis represents encoding time, while the y-axis designates each afferent. Each dot symbolizes an incoming spike from the associated afferent. Blue dots denote spikes associated with clusters of background, whereas red dots represent spikes associated with clusters of signal. In the top graph, the x-axis represents encoding time, while the y-axis designates each neuron. The green dots represent the activation of a neuron.
Figure 20. Spatio-temporal spike patterns in events containing 10 tracks. In the bottom graph, the x-axis represents encoding time, while the y-axis designates each afferent. Each dot symbolizes an incoming spike from the associated afferent. Blue dots denote spikes associated with clusters of background, whereas red dots represent spikes associated with clusters of signal. In the top graph, the x-axis represents encoding time, while the y-axis designates each neuron. The green dots represent the activation of a neuron.
Particles 08 00040 g020
In order to analyze this relationship, we examined eight datasets that were considered for the study, each of which contained 25,000 double-track events, with a fixed angular separation between them, defined as the difference of the average azimuthal angle of the two tracks. Furthermore, we define the following metrics.
Definition 10
(Detection Efficiency). Detection efficiency is the ratio of correct activations (i.e., events when a neuron fires following the signal of a track of the class it was assigned to) to the total number of events of that class in the dataset:
ϵ = T P T P + F N
where the following is true:
  • T P (True Positives) are correctly classified events.
  • F N (False Negatives) are missed events.
Definition 11
(Misclassification Rate). The misclassification rate is the ratio of misclassified events (i.e., events when a neuron fires and that do not contain a track of the class it was assigned to) to the total number of events in which the neuron fires in the dataset:
α = F P T P + F P
where the following is true:
  • T P (True Positives) are correctly classified particle track events.
  • F P (False Positives) are events mistakenly classified.
Figure 21a examines the detection efficiency of each neuron versus the angular difference between tracks. For all particle combinations, we observed a significant loss in efficiency as azimuthal angular separation between particle pairs decreases. This tendency was confirmed by a concurrent rise in the misclassification rate, represented in Figure 21b. Figure 21c is intended as a summary of those observations, displaying the average behavior for efficiency and misclassification rates overall particle types. Those results can be ascribed to the refractory period, governed by the dynamics of the reset potential and the inhibition mechanism that prevents the simultaneous firing of neurons. In principle, this effect represents the main limitation of the network in multi-track events. We believe that it can be mitigated by carefully tuning the parameters that regulate the reset potential after neuron activation, as well as by adjusting the strength and duration of inhibition. Furthermore, we believe that an encoding of the so-far neglected transversal dimension (corresponding to particles of different rapidity generating hits at different z positions in the detector) could greatly improve the capability of the SNN to distinguish particles with similar azimuthal angles. These aspects will be further investigated in future studies.

7. Conclusions

In this work, we have considered a simplified model of the CMS Phase-2 silicon tracker to study how a neuromorphic computing readout and processing of the information may allow the unsupervised identification of particle trajectories at particle colliders in the presence of significant noise in the detector.
Our proposed approach introduces several novel concepts, including time-encoding of ionization hits and neuromorphic time-series processing, which could significantly reduce power consumption in online tracking systems. With appropriate hardware implementation, this method has the potential to enable ultra-fast identification of trigger primitives, making it a promising candidate for real-time applications in future high-energy physics experiments.
While in our model, we limited ourselves to a 2D geometry, and we mainly focused on the issue of single-track reconstruction, we identified no conceptual hindrances to the scalability of this model to three spatial dimensions, which would strongly reduce backgrounds, nor to the tracking of large numbers of particles in the same event. One challenging aspect of the produced model is, on the other hand, the need for tuning its hyperparameters in order for the system to achieve the best performance. In our study, we explored two different techniques to identify the best operating point of those parameters through an evolutionary technique and the separate learning of input signal delays to the neurons. Since the amount of training data necessary to learn patterns by the system is quite limited (a few tens of thousands of events corresponding to O   ( 1 μ s) of data taken at the LHC), hyperparameter tuning does not appear as a potential showstopper on the way to a real hardware implementation of a neuromorphic-computing-based triggering system for tracking applications in future colliders.

Author Contributions

Conceptualization: E.C., F.C., M.A., T.D., E.L., E.P., J.R., F.S. and M.T.; Data curation: E.C., F.C. and M.T.; Formal analysis: E.C., F.C., E.P. and J.R.; Funding acquisition: T.D. and F.S.; Methodology: E.C., F.C., M.A., T.D., E.L., E.P., J.R., F.S. and M.T.; Project administration: T.D. and F.S.; Resources: T.D., F.S. and M.T.; Software: E.C., F.C., T.D., E.P. and J.R.; Supervision: T.D., F.S. and M.T.; Validation: E.C., F.C., E.P. and J.R.; Visualization: E.C., F.C., E.P. and J.R.; Writing—original draft: E.C., F.C., M.A., T.D., E.P., J.R., F.S. and M.T.; Writing— review and editing: E.C., F.C., M.A., T.D., E.P., J.R., F.S. and M.T. All authors have read and agreed to the published version of the manuscript.

Funding

The work by TD and FS was partially supported by the Wallenberg AI, Autonomous Systems and Software Program (WASP) funded by the Knut and Alice Wallenberg Foundation. The work by MA and FS was partially supported by the Jubilee Fund at the Luleå University of Technology.

Data Availability Statement

The datasets presented in this article are not readily available because the data are part of an ongoing study and due to time limitations. Requests to access the datasets should be directed to the corresponding authors.

Acknowledgments

The authors thank the Tracker Group of the CMS Collaboration for the use of the geometry model of the Phase-2 detector.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Parameters

Table A1. Final hyperparameters of the spiking neural network, optimized for robust particle trajectory classification under high-noise conditions using a genetic algorithm.
Table A1. Final hyperparameters of the spiking neural network, optimized for robust particle trajectory classification under high-noise conditions using a genetic algorithm.
ParameterValueDescription
K μ 0.167Temporal Scaling Factor for Inhibitory Post-Synaptic Potential (IPSP) dynamics.
K2.27 VScaling constant for Excitatory Post-Synaptic Potential (EPSP).
K 1 3.45 VScaling constant for the Reset Potential after neuron activation.
K 2 5.00 VScaling constant for the Reset Potential after neuron activation.
d max 2.5 × 10 9 sMaximum allowable synaptic delay.
Δ 5 × 10 10 sDefines the initial spread of the synaptic delays.
T 0 1.58 VMembrane potential threshold for neuron activation in Layer 0.
T 1 0.733 VMembrane potential threshold for neuron activation in Layer 1.
α 1.31Strength coefficient for Inhibitory Post-Synaptic Potential (IPSP) dynamics.
d 1.98 × 10 13 sLearning rate for synaptic delay depression.
d + 2.24 × 10 13 sLearning rate for synaptic delay potentiation.
τ m 1.24 × 10 10 sMembrane potential decay time constant.
τ s 3.46 × 10 11 sSynaptic potential time constant for EPSP dynamics.
τ d 1.31 × 10 9 sTime constant for synaptic delay depression.
τ d 2.90 × 10 10 sAuxiliary time constant for synaptic delay depression.
τ d + 2.70 × 10 9 sTime constant for synaptic delay potentiation.
τ d + τ d + 6.15 × 10 10 sAuxiliary time constant for synaptic delay potentiation.
t m a x 6.14 × 10 11 sPeak time for Excitatory Post-Synaptic Potential (EPSP).

References

  1. CERN Yellow Reports: Monographs; High-Luminosity Large Hadron Collider (HL-LHC): Technical Design Report; CERN: Geneva, Switzerland, 2020; Volume 10. [CrossRef]
  2. Dorigo, T.; Giammanco, A.; Vischia, P.; Aehle, M.; Bawaj, M.; Boldyrev, A.; de Castro Manzano, P.; Derkach, D.; Donini, J.; Edelen, A.; et al. Toward the end-to-end optimization of particle physics instruments with differentiable programming. Rev. Phys. 2023, 10, 100085. [Google Scholar] [CrossRef]
  3. Mehonic, A.; Ielmini, D.; Roy, K.; Mutlu, O.; Kvatinsky, S.; Serrano-Gotarredona, T.; Linares-Barranco, B.; Spiga, S.; Savel’ev, S.; Balanov, A.G.; et al. Roadmap to neuromorphic computing with emerging technologies. APL Mater. 2024, 12, 109201. [Google Scholar] [CrossRef]
  4. Jaeger, H.; Noheda, B.; van der Wiel, W.G. Toward a formal theory for computing machines made out of whatever physics offers. Nat. Commun. 2023, 14, 4911. [Google Scholar] [CrossRef] [PubMed]
  5. Mead, C. How we created neuromorphic engineering. Nat. Electron. 2020, 3, 434–435. [Google Scholar] [CrossRef]
  6. Kudithipudi, D.; Schuman, C.; Vineyard, C.M.; Panditit, T.; Merkel, C.; Kubendran, R.; Aimone, J.B.; Orchard, G.; Mayr, C.; Benosman, R.; et al. Neuromorphic computing at scale. Nature 2025, 637, 801. [Google Scholar] [CrossRef] [PubMed]
  7. Home—Deep South. Available online: https://www.deepsouth.org.au (accessed on 1 January 2025).
  8. Winge, D.O.; Limpert, S.; Linke, H.; Borgström, M.T.; Webb, B.; Heinze, S.; Mikkelsen, A. Implementing an Insect Brain Computational Circuit Using III–V Nanowire Components in a Single Shared Waveguide Optical Network. ACS Photonics 2020, 7, 2787. [Google Scholar] [CrossRef] [PubMed]
  9. Wittenbecher, L.; Viñas Boström, E.; Vogelsang, J.; Lehman, S.; Dick, K.A.; Verdozzi, C.; Zigmantas, D.; Mikkelsen, A. Unraveling the Ultrafast Hot Electron Dynamics in Semiconductor Nanowires. ACS Nano 2021, 15, 1133. [Google Scholar] [CrossRef] [PubMed]
  10. Winge, D.; Borgström, M.; Lind, E.; Mikkelsen, A. Artificial nanophotonic neuron with internal memory for biologically inspired and reservoir network computing. Neuromorphic Comput. Eng. 2023, 3, 034011. [Google Scholar] [CrossRef]
  11. Nilsson, M.; Schelén, O.; Lindgren, A.; Bodin, U.; Paniagua, C.; Delsing, J.; Sandin, F. Integration of neuromorphic AI in event-driven distributed digitized systems: Concepts and research directions. Front. Neurosci. 2023, 17, 1074439. [Google Scholar] [CrossRef] [PubMed]
  12. Gerstner, W.; Kistler, W.M.; Naud, R.; Paninski, L. Neuronal Dynamics: From Single Neurons to Networks and Models of Cognition; Cambridge University Press: Cambridge, UK, 2014. [Google Scholar]
  13. Verhelst, M.; Bahai, A. Where Analog Meets Digital: Analog-to-Information Conversion and Beyond. IEEE Solid-State Circuits Mag. 2015, 7, 67–80. [Google Scholar] [CrossRef]
  14. Calafiura, P.; Rousseau, D.; Terao, K. AI For High-Energy Physics; World Scientific: Singapore, 2022. [Google Scholar]
  15. di Torino, P.; Lazarescu, M.T.; Lazarescu, M.T. FPGA-Based Deep Learning Inference Acceleration at the Edge. Ph.D. Thesis, Politecnico di Torino, Torino, Italy, 2021. [Google Scholar]
  16. Di Meglio, A.; Jansen, K.; Tavernelli, I.; Alexandrou, C.; Arunachalam, S.; Bauer, C.W.; Borras, K.; Carrazza, S.; Crippa, A.; Croft, V.; et al. Quantum Computing for High-Energy Physics: State of the Art and Challenges. PRX Quantum 2024, 5, 037001. [Google Scholar] [CrossRef]
  17. CMS Collaboration. The Phase-2 Upgrade of the CMS Tracker; CERN: Geneva, Switzerland, 2017. [Google Scholar] [CrossRef]
  18. The Tracker Group of the CMS Collaboration. Phase-2 CMS Tracker Layout Information. Available online: https://cms-tklayout.web.cern.ch/cms-tklayout/layouts/recent-layouts/OT801_IT701/index.html (accessed on 1 September 2023).
  19. Masquelier, T.; Guyonneau, R.; Thorpe, S.J. Competitive STDP-Based Spike Pattern Learning. Neural Comput. 2009, 21, 1259–1276. [Google Scholar] [CrossRef] [PubMed]
  20. Bear, M.; Connors, B.; Paradiso, M. Neuroscience: Exploring the Brain; Wolters Kluwer: Alphen aan den Rijn, The Netherlands, 2016. [Google Scholar]
  21. Hammouamri, I.; Khalfaoui-Hassani, I.; Masquelier, T. Learning Delays in Spiking Neural Networks using Dilated Convolutions with Learnable Spacings. arXiv 2023, arXiv:2306.17670. [Google Scholar]
  22. Nadafian, A.; Ganjtabesh, M. Bio-plausible Unsupervised Delay Learning for Extracting Temporal Features in Spiking Neural Networks. arXiv 2020, arXiv:2011.09380. [Google Scholar]
  23. Hazan, H.; Caby, S.; Earl, C.; Siegelmann, H.; Levin, M. Memory via Temporal Delays in weightless Spiking Neural Network. arXiv 2022, arXiv:2202.07132. [Google Scholar]
  24. Song, S.; Miller, K.D.; Abbott, L.F. Competitive Hebbian learning through spike-timing-dependent synaptic plasticity. Nat. Neurosci. 2000, 3, 919–926. [Google Scholar] [CrossRef] [PubMed]
  25. Bi, G.Q.; Poo, M.M. Synaptic modification by correlated activity: Hebb’s postulate revisited. Annu. Rev. Neurosci. 2001, 24, 139–166. [Google Scholar] [CrossRef] [PubMed]
  26. Gad, A.F. PyGAD: An Intuitive Genetic Algorithm Python Library. arXiv 2021, arXiv:2106.06158. [Google Scholar] [CrossRef]
  27. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef]
Figure 1. Spike-based encoding (a) and processing of information using a spiking neuron unit (b). Spikes are asynchronous binary events used to represent detector hits and neuron activations. By encoding information in a succession of spikes, such as their order or temporal distance, non-binary modes of information coding can be used. By leveraging physical phenomena for integration (in synapses) and processing (in dendrites and the soma of the neuron) of spike codes, the energy efficiency and latency can be improved versus ordinary logic information processing [3], in particular when the output precision is limited by noise [13].
Figure 1. Spike-based encoding (a) and processing of information using a spiking neuron unit (b). Spikes are asynchronous binary events used to represent detector hits and neuron activations. By encoding information in a succession of spikes, such as their order or temporal distance, non-binary modes of information coding can be used. By leveraging physical phenomena for integration (in synapses) and processing (in dendrites and the soma of the neuron) of spike codes, the energy efficiency and latency can be improved versus ordinary logic information processing [3], in particular when the output precision is limited by noise [13].
Particles 08 00040 g001
Figure 3. Sketch of the Network Architecture. In the final network presented in Section 6.1, there are 10 afferents, N L 0 = 6 neurons in layer L 0 and N L 1 = 6 neurons in layer L 1 .
Figure 3. Sketch of the Network Architecture. In the final network presented in Section 6.1, there are 10 afferents, N L 0 = 6 neurons in layer L 0 and N L 1 = 6 neurons in layer L 1 .
Particles 08 00040 g003
Figure 4. Comparison of membrane potential evolution in firing and non-firing events. In the top plot, neuron 10 surpasses the firing threshold at t 143.7 ns , experiencing a sharp increase in potential followed by a reset governed by the reset potential η ( t t i ) . An entire event is 25 ns . The firing neuron also induces an IPSP μ ( t t k ) in the other neurons, suppressing their activation and reinforcing the competition mechanism. In the bottom plot, neurons exhibit sub-threshold oscillations but do not fire, indicating that the input signals were insufficient to reach the activation threshold.
Figure 4. Comparison of membrane potential evolution in firing and non-firing events. In the top plot, neuron 10 surpasses the firing threshold at t 143.7 ns , experiencing a sharp increase in potential followed by a reset governed by the reset potential η ( t t i ) . An entire event is 25 ns . The firing neuron also induces an IPSP μ ( t t k ) in the other neurons, suppressing their activation and reinforcing the competition mechanism. In the bottom plot, neurons exhibit sub-threshold oscillations but do not fire, indicating that the input signals were insufficient to reach the activation threshold.
Particles 08 00040 g004
Figure 6. Layout of the silicon sensors in the barrel of the Phase-2 CMS tracker in the transverse plane [18].
Figure 6. Layout of the silicon sensors in the barrel of the Phase-2 CMS tracker in the transverse plane [18].
Particles 08 00040 g006
Figure 7. Spatio-temporal spike patterns. The x-axis represents encoding time, while the y-axis designates each afferent. Each dot symbolizes an incoming spike from the associated afferent. Blue dots denote spikes associated with clusters of background, whereas red dots represent spikes associated with clusters of signal.
Figure 7. Spatio-temporal spike patterns. The x-axis represents encoding time, while the y-axis designates each afferent. Each dot symbolizes an incoming spike from the associated afferent. Blue dots denote spikes associated with clusters of background, whereas red dots represent spikes associated with clusters of signal.
Particles 08 00040 g007
Figure 8. Comparison of transverse plane projections of events with different noise levels. The left panel (a) shows an event containing 100 noise clusters, and the right panel (b) shows an event with 300 noise clusters. Red points represent clusters associated with the main event, while blue points indicate background noise.
Figure 8. Comparison of transverse plane projections of events with different noise levels. The left panel (a) shows an event containing 100 noise clusters, and the right panel (b) shows an event with 300 noise clusters. Red points represent clusters associated with the main event, while blue points indicate background noise.
Particles 08 00040 g008
Figure 9. Distribution of the hits in R, the radial distance in the transverse plane from the nominal interaction point.
Figure 9. Distribution of the hits in R, the radial distance in the transverse plane from the nominal interaction point.
Particles 08 00040 g009
Figure 10. Distribution of η , the pseudorapidity, indicating the spread of clusters along the beam axis direction.
Figure 10. Distribution of η , the pseudorapidity, indicating the spread of clusters along the beam axis direction.
Particles 08 00040 g010
Figure 11. Distribution of ϕ , the azimuthal angle, showing the uniformity of clusters around the detector transverse plane.
Figure 11. Distribution of ϕ , the azimuthal angle, showing the uniformity of clusters around the detector transverse plane.
Particles 08 00040 g011
Figure 12. Heatmap of neuron activations across particle classes, with neurons represented on the x-axis and particle classes (characterized by charge and transverse momentum) on the y-axis. The color intensity indicates the acceptance rate or the fraction of events in which the neuron was activated for a given class. The fake rate (false positives for noise-only events) is shown in the bottom row. This visualization demonstrates the lack of specialization of the neurons for specific classes despite maintaining low false-positive rates.
Figure 12. Heatmap of neuron activations across particle classes, with neurons represented on the x-axis and particle classes (characterized by charge and transverse momentum) on the y-axis. The color intensity indicates the acceptance rate or the fraction of events in which the neuron was activated for a given class. The fake rate (false positives for noise-only events) is shown in the bottom row. This visualization demonstrates the lack of specialization of the neurons for specific classes despite maintaining low false-positive rates.
Particles 08 00040 g012
Figure 13. Evolution of synaptic delays over the course of training for neurons specializing in recognizing particles with transverse momentum values of 1 GeV (a) and 10 GeV (b). The x-axis represents the training iteration, and the y-axis shows the delay value for each afferent.
Figure 13. Evolution of synaptic delays over the course of training for neurons specializing in recognizing particles with transverse momentum values of 1 GeV (a) and 10 GeV (b). The x-axis represents the training iteration, and the y-axis shows the delay value for each afferent.
Particles 08 00040 g013
Figure 14. Comparison of acceptance, selectivity, and fake rate evolution during the learning process. (a) Evolution of acceptance and selectivity of the SNN during the learning process. The X-axis represents the number of events processed, whereas the Y-axis denotes the acceptances and selectivities of L0 and L1 neurons. The selectivities have been normalized by their maximum values for visualization purposes. (b) Evolution of the fake rates of neurons during the learning process. The X-axis represents the number of events processed, whereas the Y-axis denotes the fake rates of L0 and L1 neurons.
Figure 14. Comparison of acceptance, selectivity, and fake rate evolution during the learning process. (a) Evolution of acceptance and selectivity of the SNN during the learning process. The X-axis represents the number of events processed, whereas the Y-axis denotes the acceptances and selectivities of L0 and L1 neurons. The selectivities have been normalized by their maximum values for visualization purposes. (b) Evolution of the fake rates of neurons during the learning process. The X-axis represents the number of events processed, whereas the Y-axis denotes the fake rates of L0 and L1 neurons.
Particles 08 00040 g014
Figure 15. Comparison of spike rates, defined as the number of neuron firings per event, throughout the learning process. The X-axis represents the number of processed events. (a) The Y-axis shows the firing rate of L0 neurons. (b) The Y-axis displays the firing rate of L1 neurons on a logarithmic scale. The target spike rate represents the expected value for a perfectly specialized neuron and corresponds to r = 1 6 firings event .
Figure 15. Comparison of spike rates, defined as the number of neuron firings per event, throughout the learning process. The X-axis represents the number of processed events. (a) The Y-axis shows the firing rate of L0 neurons. (b) The Y-axis displays the firing rate of L1 neurons on a logarithmic scale. The target spike rate represents the expected value for a perfectly specialized neuron and corresponds to r = 1 6 firings event .
Particles 08 00040 g015
Figure 16. Performance of a network trained with a background level of N bg = 100 when tested at higher background levels.
Figure 16. Performance of a network trained with a background level of N bg = 100 when tested at higher background levels.
Particles 08 00040 g016
Figure 17. Heatmap of neuron activations across particle classes, with neurons represented on the x-axis and particle classes (characterized by charge and transverse momentum) on the y-axis. The color intensity indicates the acceptance rate or the fraction of events in which the neuron was activated for a given class. The fake rate (false positives for noise-only events) is shown in the bottom row. This visualization demonstrates the network’s ability to specialize neurons for specific classes while maintaining low false-positive rates.
Figure 17. Heatmap of neuron activations across particle classes, with neurons represented on the x-axis and particle classes (characterized by charge and transverse momentum) on the y-axis. The color intensity indicates the acceptance rate or the fraction of events in which the neuron was activated for a given class. The fake rate (false positives for noise-only events) is shown in the bottom row. This visualization demonstrates the network’s ability to specialize neurons for specific classes while maintaining low false-positive rates.
Particles 08 00040 g017
Figure 18. Synaptic delays of the different neurons. The x-axis represents the neuron ID, and the y-axis corresponds to the afferent ID. The color intensity indicates the synaptic delay (in nanoseconds), with brighter colors representing longer delays. This figure showcases the specialization of synaptic delays for different neurons to detect distinct spatio-temporal patterns in the input.
Figure 18. Synaptic delays of the different neurons. The x-axis represents the neuron ID, and the y-axis corresponds to the afferent ID. The color intensity indicates the synaptic delay (in nanoseconds), with brighter colors representing longer delays. This figure showcases the specialization of synaptic delays for different neurons to detect distinct spatio-temporal patterns in the input.
Particles 08 00040 g018
Figure 19. Spatio-temporal spike patterns. In the bottom graph, the x-axis represents encoding time, while the y-axis designates each afferent. Each dot symbolizes an incoming spike from the associated afferent. Blue dots denote spikes associated with clusters of background, whereas red dots represent spikes associated with clusters of signal. In the top graph, the x-axis represents encoding time, while the y-axis designates each neuron. The green dots represent the activation of a neuron.
Figure 19. Spatio-temporal spike patterns. In the bottom graph, the x-axis represents encoding time, while the y-axis designates each afferent. Each dot symbolizes an incoming spike from the associated afferent. Blue dots denote spikes associated with clusters of background, whereas red dots represent spikes associated with clusters of signal. In the top graph, the x-axis represents encoding time, while the y-axis designates each neuron. The green dots represent the activation of a neuron.
Particles 08 00040 g019
Figure 21. Study of the network’s performance as a function of the angular difference between tracks. As Δ ϕ decreases, the network’s ability to correctly classify the different tracks diminishes, indicating a loss of specialization in the learned representations. The error associated with each point is of order O ( 10 3 ) . (a) Each line represents the detection efficiency of a neuron. (b) Each line represents the misclassification rate of a neuron. (c) Average misclassification rate and detection efficiency of the SNN. (d) Example of an event containing two particles with an angular separation of Δ ϕ = 300 mrad .
Figure 21. Study of the network’s performance as a function of the angular difference between tracks. As Δ ϕ decreases, the network’s ability to correctly classify the different tracks diminishes, indicating a loss of specialization in the learned representations. The error associated with each point is of order O ( 10 3 ) . (a) Each line represents the detection efficiency of a neuron. (b) Each line represents the misclassification rate of a neuron. (c) Average misclassification rate and detection efficiency of the SNN. (d) Example of an event containing two particles with an angular separation of Δ ϕ = 300 mrad .
Particles 08 00040 g021
Table 1. Aggregate acceptance and fake rate of the networks for different patterns and average noise levels. The networks were trained and tested with the same average number of noise hits. Each test dataset contains N ev test = 10,000 events.
Table 1. Aggregate acceptance and fake rate of the networks for different patterns and average noise levels. The networks were trained and tested with the same average number of noise hits. Each test dataset contains N ev test = 10,000 events.
N bg A 1 [%] A + 1 [%] A 3 [%] A + 3 [%] A 10 [%] A + 10 [%]F [%]
50 86 ± 1 76 ± 1 98.6 ± 0.4 99.9 ± 0.2 93.5 ± 0.8 96.4 ± 0.7 2.2 ± 0.1
100 70 ± 2 61 ± 2 98.4 ± 0.4 98.9 ± 0.4 97.0 ± 0.6 97.6 ± 0.5 2.1 ± 0.1
200 64 ± 2 49.4 ± 2 93.5 ± 0.8 92 ± 1 97.5 ± 0.6 97.4 ± 0.6 3.9 ± 0.2
Table 2. Hyperparameters of the SNN, optimized for particle trajectory classification under high-noise conditions using a genetic algorithm.
Table 2. Hyperparameters of the SNN, optimized for particle trajectory classification under high-noise conditions using a genetic algorithm.
ParameterDescription
K μ , α Temporal scaling factor and strength coefficient for IPSP dynamics.
K 1 , K 2 Scaling constants for the reset potential after neuron activation.
T 0 , T 1 Membrane potential thresholds for neuron activation in Layer 0 and Layer 1.
d , d + Learning rates for synaptic delay depression and potentiation.
τ m , τ s Membrane potential decay time constant and synaptic potential time constant.
τ d , τ d Time constant and auxiliary time constant for synaptic delay depression.
τ d + , τ d + Time constant and auxiliary time constant for synaptic delay potentiation.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Coradin, E.; Cufino, F.; Awais, M.; Dorigo, T.; Lupi, E.; Porcu, E.; Raj, J.; Sandin, F.; Tosi, M. Unsupervised Particle Tracking with Neuromorphic Computing. Particles 2025, 8, 40. https://doi.org/10.3390/particles8020040

AMA Style

Coradin E, Cufino F, Awais M, Dorigo T, Lupi E, Porcu E, Raj J, Sandin F, Tosi M. Unsupervised Particle Tracking with Neuromorphic Computing. Particles. 2025; 8(2):40. https://doi.org/10.3390/particles8020040

Chicago/Turabian Style

Coradin, Emanuele, Fabio Cufino, Muhammad Awais, Tommaso Dorigo, Enrico Lupi, Eleonora Porcu, Jinu Raj, Fredrik Sandin, and Mia Tosi. 2025. "Unsupervised Particle Tracking with Neuromorphic Computing" Particles 8, no. 2: 40. https://doi.org/10.3390/particles8020040

APA Style

Coradin, E., Cufino, F., Awais, M., Dorigo, T., Lupi, E., Porcu, E., Raj, J., Sandin, F., & Tosi, M. (2025). Unsupervised Particle Tracking with Neuromorphic Computing. Particles, 8(2), 40. https://doi.org/10.3390/particles8020040

Article Metrics

Back to TopTop