1. Introduction
The rapid evolution of neuromorphic computing has led to a resurgence in the study of computational models inspired by biological neural systems, aiming to process information in a way that is both efficient and capable of handling real-world, dynamic data [
1,
2,
3]. Among these models, Liquid State Machines (LSMs) stand out as a promising framework for processing spatio-temporal data due to their inherent adaptability and efficiency in tasks requiring high temporal precision [
4,
5]. LSMs leverage a “reservoir” dynamic, using recurrently connected neurons to process information by mapping inputs into high-dimensional representations, which can then be classified or further processed using simpler learning algorithms [
6,
7]. This property makes LSMs particularly well suited for applications in neuromorphic hardware, where energy efficiency and real-time processing are paramount [
8,
9,
10].
However, one of the fundamental challenges in deploying LSMs for spatio-temporal classification is optimizing reservoir separability, which refers to the reservoir’s capacity to map inputs into distinct, separable states [
11,
12,
13]. High separability enhances the model’s ability to classify complex temporal patterns by creating distinct trajectories for different input signals within the reservoir state space [
14]. Achieving optimal separability is crucial for the effectiveness of LSMs, particularly in neuromorphic hardware where constraints on memory and computational resources demand the careful tuning of the network’s parameters and structure [
15,
16].
In recent years, various approaches have been explored to enhance reservoir separability in LSMs, such as tuning synaptic weights, adjusting connectivity patterns, and optimizing the neuron model’s dynamics [
7,
13,
17,
18]. These efforts have yielded promising improvements, yet they often rely on heuristics that may not generalize well to different tasks or hardware implementations. Therefore, the development of systematic optimization strategies for enhancing reservoir separability remains an open problem, with significant implications for the deployment of LSMs in neuromorphic hardware [
19]. Additionally, although there have been successful neuromorphic implementations of LSMs, their capability to address several problems using a single liquid and multiple task-related readout models has been scarcely exploited in both software and hardware implementations. For example, in [
9], an LSM was implemented on the SpiNNaker board solely to solve the spatio-temporal version of the well-known MNIST benchmark (N-MNIST [
20]). This research aims to leverage the capability of LSMs on a neuromorphic platform to address different problems by focusing on improving the separability property of the liquid and simplifying the task-related readout models.
Possible applications of this technology are broad and especially impactful in fields where real-time processing and low power consumption are paramount. In biomedical signal processing, for instance, it holds significant promise for the early detection of conditions such as epilepsy or arrhythmias through the analysis of EEG and ECG signals, which are inherently spatio-temporal in nature. Additionally, this approach can enhance pattern recognition and classification tasks in areas like intelligent surveillance, real-time video analysis, and audio processing, such as robust speech recognition in noisy environments. Since spiking neural networks are particularly adept at handling spatio-temporal data, this technology is well suited for these complex, dynamic signal types. Moreover, brain–computer interfaces (BCIs) could benefit from this technology, enabling more efficient and accurate control of external devices using neural signals, all while minimizing energy consumption.
This paper proposes a novel optimization approach using Particle Swarm Optimization (PSO) to improve the reservoir separability of LSMs specifically for spatio-temporal classification tasks. Our approach demonstrated significant improvements in classification accuracy, with the Mono-objective approach achieving accuracies ranging from 81.09% to 95.52% and the multi-objective approach delivering accuracies ranging from 90.23% to 98.77% for the software implementation. In addition, when implemented on the SpiNNaker neuromorphic system, the Mono-objective approach achieved accuracies ranging from 86.20% to 97.70%, with the multi-objective approach further improving accuracies, ranging from 94.42% to 99.52%. PSO was utilized to systematically evaluate and refine reservoir configurations, enhancing the discriminative power of the network and enabling it to better distinguish between complex input patterns. Our approach leverages recent advances in neural network optimization and is designed with a focus on the constraints and requirements of neuromorphic hardware platforms. Through extensive experimental validation, we demonstrated that PSO-optimized reservoir separability not only enhances classification accuracy but also allows for the more efficient and scalable deployment of LSMs in practical applications. This research contributes to the growing body of work on neuromorphic computing by addressing a core challenge in LSM design and providing a pathway toward more robust and adaptable spatio-temporal classifiers.
The rest of this paper is organized as follows: In
Section 2, we provide a detailed description of Liquid State Machines (LSMs) and explain how they are configured within our method, including the spiking neuron model used. In
Section 3, we describe the methodology applied to improve reservoir separability based on the classes present in the dataset, along with an overview of the SpiNNaker neuromorphic system and its configuration parameters.
Section 4 presents the results obtained from the implementation on SpiNNaker. Finally, conclusions are provided in
Section 5.
3. Materials and Methods
Our methodology, illustrated in
Figure 1, was designed to enhance the performance of the liquid state in Liquid State Machines (LSMs) by increasing its separability [
23], which is crucial for improving classification accuracy in multi-class data scenarios [
24]. To achieve this, we applied Particle Swarm Optimization (PSO), leveraging both single-objective and multi-objective optimization approaches to fine-tune the liquid’s separability properties. PSO was particularly well suited for this task due to its effectiveness in optimizing non-differentiable functions, which was essential given the spiking nature of neurons in LSMs. Traditional gradient-based optimization methods were not applicable in this context because the spiking behavior of neurons introduces discontinuities that prevent the calculation of gradients [
25]. By using PSO, we could circumvent these limitations and systematically improve the liquid’s ability to differentiate between different input patterns. The PSO process involved initializing a swarm of candidate solutions, each representing a potential configuration of the liquid’s parameters. These candidates were evaluated based on their performance in enhancing separability, and they iteratively adjusted their positions in the parameter space by considering their own best performance and the best performances observed by the swarm. This iterative process continued until a satisfactory level of separability was achieved. In our approach, this optimization was performed offline, allowing for a thorough search of the parameter space without the constraints of real-time processing. Once the optimal parameters were determined, the enhanced liquid state was then deployed on a neuromorphic platform, such as SpiNNaker, which is capable of emulating the spiking neuron dynamics efficiently.
3.1. Separability Metric
The reservoir’s performance was evaluated using the separability metric introduced in [
14]. This metric measured how effectively the reservoir transformed inputs into distinct, separable states, thereby enhancing the readout layer’s ability to classify these states accurately. The separability metric is defined as the ratio of the inter-class distance
to the intra-class variance
, with 1 added to the denominator to prevent division by zero, as shown in Equation (
3). This evaluation was applied to the set of state vectors
, determined by the reservoir’s firing rate over the final 50 ms of the simulation.
The inter-class distance,
, is defined as the average distance between the centers of mass for each pair of classes, as expressed in Equation (
4).
where
and
represent the centers of mass for the classes
l and
m, respectively. The total number of classes in the problem is denoted by
n. The center of mass for the
l-th class is given by the following equation:
with
representing the quantity of state vectors that correspond to the
l-th class.
On the other hand, the intra-class variance
is defined as the average variance of each cluster of state vectors, and it is calculated as follows:
In Equation (
6), the term
indicates the mean variance of the state vectors belonging to the class
l, as outlined in Equation (
7).
3.2. Optimization Methods
3.2.1. Particle Swarm Optimization
Particle Swarm Optimization (PSO) [
26,
27] is an optimization algorithm inspired by the collective behavior of swarm individuals observed in nature. A swarm consists of a group of elements known as particles, which are responsible for exploring and exploiting the
d-dimensional search space in a given optimization problem to find the best solution based on the combined experience of all particles. For this research, the value of
d could not be explicitly defined because it depended on the characteristics of the liquid. Since the liquids were initially created randomly based on the connectivity between the input layer and the liquid, as well as the internal neuron connectivity, the value of
d varied accordingly. The search process was guided primarily by two key attributes of each particle: its velocity
V and position
X. Velocity represents the direction and speed of movement from the current position to the next position within the search space during each iteration. Position corresponds to the particle’s current location in the search space.
In each iteration, k, the particles required an update, which involved obtaining two key values:
Using
and
, PSO updated the velocity through Equation (
8) and the position through Equation (
9).
where we have the following:
is the d-th value of the i-th particle’s velocity.
is the inertia weight.
are the cognitive and social acceleration coefficients.
are two uniform random values generated within the interval.
is the d-th value of the personal best position.
is the d-th value of the global best position.
is the d-th value of the i-th particle’s position.
Figure 2 presents a diagram illustrating the methodology for implementing PSO. We started by initializing the parameters
as previously described, along with
S, the number of particles, and
D, the problem dimensionality. A swarm of
S particles was generated, each with an initial position,
X, and velocity,
V. Each particle was then evaluated based on the fitness function (separability level from Equation (
3)) in its current position,
X, with this initial position serving as the initial
value.
As long as no stopping condition was met, both the velocity and position of each particle were updated, and the fitness function was evaluated. We then updated and accordingly. Ultimately, the particle with the best was obtained as the solution.
3.2.2. OMOPSO
The Original Multi-Objective PSO (OMOPSO) [
28] is an algorithm proposed to solve multi-objective optimization problems. It is a variant of Multi-Objective Particle Swarm Optimization (MOPSO) [
29], an extension of the PSO algorithm designed specifically for multi-objective optimization. When addressing complex problems with conflicting objectives, it becomes essential to generate a set of non-dominated solutions. These optimal solutions offer a balanced representation for multi-objective problems and are organized along the Pareto front, enabling the consideration of multiple optimal solutions to a given problem.
Figure 3 illustrates the methodology for implementing OMOPSO. The process begins by initializing a swarm with
S particles and selecting an initial set of leaders. An external archive (
e-archive) is generated to store these initial leaders. Crowding is applied to control the distribution of particles within the search space. The update for each particle is then based on the following steps:
Select Leader: the selection of a leader from the e-archive to guide the particle’s movement.
Flight: updating the particle’s velocity and position based on the previously selected leader.
Mutation: the implementation of mutation to introduce variability.
Evaluation: the particle’s updated position is evaluated based on the fitness function.
Update : updating if a better performance is achieved during the evaluation.
Once the update process for each particle is complete, the leaders, , are updated by transferring the non-dominated solutions to the e-archive and applying crowding again. If no stopping condition, such as the maximum number of generations, is met, the particles are updated once more. Otherwise, the e-archive containing the solutions is obtained.
3.3. Dataset
Our approach utilized four datasets, initially proposed in [
14]: Frequency Recognition with 5 classes (FR5), Pattern Recognition with 4 classes (PR4), Pattern Recognition with 8 classes (PR8), and Pattern Recognition with 12 classes (PR12). Each dataset consisted of 400 instances per class for training and 100 instances per class for testing, totaling 2000 and 500, 1600 and 400, 3200 and 800, and 4800 and 1200 instances for the FR5, PR4, PR8, and PR12 training and test sets, respectively. In the single-objective phase, the PR4, PR8, and PR12 datasets used 8 neurons each, while FR5 employed 4 neurons. However, for the multi-task phase, the FR5 dataset was modified to ensure consistent dimensionality in synaptic connectivity across all tasks, which required each problem to have the same number of channels or neurons. Therefore, 4 non-spiking channels were added to FR5 in this phase.
Figure 4 shows a sample from each dataset.
3.4. Offline Training with the NEST Framework
For the experiments, the NEST (Neural Simulation Tool) framework [
30] and PyNN [
31] were used. NEST was employed to simulate spiking neural networks, while PyNN served as an interface for implementing experiments both in NEST and on SpiNNaker. NEST is a spiking neural network simulator that captures the dynamics of neuronal action potentials, efficiently representing the heterogeneity of biological neural networks. PyNN is an interface for spiking neural networks that facilitates the use of different simulators without requiring significant modifications between them.
A total of 33 experiments were conducted for both the single-objective and multi-objective implementations, using only 3 randomly selected samples per class with replacement. These were simulated with PyNN and NEST prior to implementation on SpiNNaker. In the single-objective phase, the total number of experiments were conducted independently for each dataset. In the multi-objective phase, only 33 experiments were performed, including all the datasets. In each experiment, a total population of 25 individuals was used. This value was selected because it falls within the range suggested by the literature, which recommends using between 20 and 30 particles [
32,
33]. Additionally, we conducted experiments with a population of 50 particles [
34,
35], but no significant improvement in performance was observed. Thus, the population size of 25 was deemed optimal for the experiments, balancing performance and computational efficiency. The initial individual was defined by the synaptic weights obtained before the optimization process. Then, 24 additional individuals were generated by introducing Gaussian noise, thus completing the total population. One limitation of implementing SpiNNaker with PyNN is that synaptic weights are stored as 16-bit integers. Since the optimization process involved fine-tuning the model in each experiment, the synaptic weights were treated as fixed-point values with 16-bit precision. This approach allowed for more representative behavior on SpiNNaker compared to what was experienced in NEST. In each experiment, the goal was to improve the LSM’s performance by optimizing the synaptic weights of the input layer and reservoir through the separability metric, evaluated with the state vectors. The state vectors in this study were obtained from the firing rate of each neuron during the last 50 ms of a 100 ms simulation.
3.5. Implementation on Neuromorphic SpiNNaker Hardware
The SpiNNaker system is a multicore computing platform specifically engineered to simulate the behavior of up to a billion spiking neurons in real time [
36]. At the heart of SpiNNaker’s innovation is its unique communication infrastructure, optimized for transmitting large volumes of compact packets efficiently [
37]. Each packet represents a ‘spike’, a discrete communication unit among neurons encoding both the originating neuron and the timing of the spike. This architecture is especially suited for handling spatio-temporal datasets.
In this study, we enhanced the reservoir’s performance by implementing metaheuristic algorithms. Subsequently, the optimized synaptic weights for the input layer and reservoir were extracted and deployed onto the SpiNNaker system. These synaptic weights, encoded as 16-bit fixed-point values, are subject to range constraints determined at the compile time. This limitation affects the precision of synaptic weights, as the smallest representable value depends on the largest specified weight [
37]. For the simulation, a scale factor of 1 ms per step was used, as increasing the scale did not yield performance improvements.
Our work utilized the SpiNNaker-3 board, illustrated in
Figure 5. This board integrated 4 SpiNNaker chips, each containing 18 ARM968 processing cores with dedicated local memory, allowing each core to support one or more neurons while maintaining low power consumption. Additionally, each chip had access to 1 GB (128 MB) of SDRAM. The board connected to a host machine via an RJ45 Ethernet port, which facilitated code downloads and real-time system communication at 10 or 100 Mbit/s [
38].
4. Results
The different experiments focused on evaluating the LSM, primarily assessing the reservoir’s ability to discriminate samples from each dataset. Four datasets were considered for the LSM evaluation: FR5, PR4, PR8, and PR12. Each experiment was conducted with variations in synaptic connectivity, leading to greater heterogeneity in the reservoir and influencing each optimization process. For each experiment, we gathered information from both the initial state (before optimization) and the final state (after optimization), considering the following characteristics:
Initial Separability (InitialSP) and Final Separability (FinalSP), which represent how well the liquid states generated by the reservoir could be separated based on the classes, in one or more tasks, before and after optimization, respectively.
Initial Accuracy (InitialAcc) and Final Accuracy (FinalAcc), which indicate the model’s performance when evaluated using the test datasets in NEST.
Final Accuracy Hardware (FinalAccHW), which represents the model’s performance when simulated on SpiNNaker.
In the SpiNNaker implementations, we only evaluated FinalAcc and assess its behavior based on the results obtained in NEST.
4.1. PSO Experiments
In this section, we review the results obtained from implementing PSO in the LSM with four tasks simulated independently in NEST and the migration of the synaptic weights of interest to SpiNNaker. For the optimization process simulations, we used the Pymoo framework [
39] with the following parameters: a population size of 25, an inertia weight of 0.9 applied in each iteration for velocity updates, a cognitive acceleration factor of 2.0, a social acceleration factor of 2.0, a randomly initialized velocity, and mutation applied to the global best.
In
Table 2, we can see how
and
represent a significant improvement when considering the initial and final phases of the experiments in terms of InitialSP–FinalSP and InitialAcc–FinalAcc. This ensures a smaller
value after the optimization process for each of the tasks, guaranteeing the proper behavior of the process. Taking the median as the reference performance value, the quantitative range of difference between the NEST and SpiNNaker results with the PSO optimization method is 0.043. This indicates, in the context of accuracy classification, that there is a small difference between all the results from both implementations.
In
Table 2 and
Table 3, the primary focus is not on comparing FinalACC and FinalACCHW but rather on comparing these metrics with InitialACC. This comparison reveals a significant improvement, demonstrating the effectiveness of our approach. Additionally, it is important to consider the separability measures, which further highlight the enhanced performance of our method. This is true for both the single-objective and multi-objective problems. The goal was to ensure that FinalACC and FinalACCHW exhibited similar performance, as this would indicate that our method was effective and consistent across both software and hardware implementations. Achieving comparable results with both platforms highlights the robustness and adaptability of our approach, showcasing its potential for practical applications. Furthermore, SpiNNaker operates at around 1–5 watts per chip, while traditional systems can require 50–100 watts for similar tasks [
40], emphasizing the significant energy efficiency advantage of our hardware-based solution.
Figure 6a shows the performance improvement in the reservoir in each task compared to the results obtained in [
14], based on accuracy (Acc), with less precision in the synaptic weights. Additionally, we were able to exceed the maximum accuracy achieved in three of the tasks compared to the total experiments, as shown in
Figure 6b.
The SpiNNaker implementation is represented by FinalAccHW, where a difference is observed compared to FinalAcc, particularly in FR5, which corresponds to the task with the lowest FinalAcc. This highlights how the reservoir’s separability and an adequate firing rate influenced the performance of the implementation on SpiNNaker. If the firing rate is too high, it may not be possible to process all spikes in each time step in a real-world environment due to the inherent architecture of the hardware.
In
Figure 7, we observe the behavior of each experiment both in NEST and on SpiNNaker after the optimization process. In certain experiments, such as 4, 14, and 26, the single-objective optimization on SpiNNaker demonstrated improved accuracy compared to simulations conducted in NEST. This improvement is likely attributable to the hardware-specific advantages of SpiNNaker, a neuromorphic platform designed to emulate spiking neural networks with high efficiency. Unlike traditional simulators, SpiNNaker’s asynchronous architecture and real-time spike processing closely mirror biological neural dynamics, enabling more precise temporal interactions within a Liquid State Machine (LSM). These characteristics contribute to enhanced reservoir separability and stability, which are critical for improving classification performance in spatio-temporal tasks. The reduced latency and optimized handling of spikes on SpiNNaker further support its suitability for implementing LSMs, leading to notable gains in performance under specific experimental conditions. A similar situation occurred with the multi-objective optimization, as will be seen below.
In
Figure 8, we can observe the behavior of the liquid state based on a simulated sample from the last 50 ms in three stages, each represented by a raster plot and a heatmap: The first corresponds to Before Optimization, where we observed the liquid state behavior before optimization, simulated in NEST, where the reservoir showed a high firing rate. In After Optimization—NEST, we observed the behavior of the reservoir after the optimization process, simulated in NEST, resulting in a slightly more regularized firing pattern. Finally, After Optimization—SpiNNaker represents the behavior of the reservoir simulated on the hardware, showing a pattern very similar to the one obtained in NEST after optimization. Focusing on the heatmap, which is a representation of the liquid state, we can observe the effect of the separability-based optimization process, where the liquid’s neurons neither overreacted nor reacted similarly to any stimuli, allowing for an enhanced overall LSM performance.
4.2. OMOPSO Experiments
In this section, we review the results of implementing OMOPSO in the LSM with four tasks simultaneously simulated in NEST and the migration of the synaptic weights of interest to SpiNNaker. For the optimization simulations, we used the Jmetalpy framework (
https://github.com/jMetal/jMetalPy, accessed on 22 october 2024) with the following parameters: a swarm size of 25 particles, a mutation probability of 1/|variables|, an
factor of 0.075, a crowding distance archive of 25 leaders, and a maximum evaluation stopping criterion of 500 evaluations. These parameters were determined through empirical evidence.
In
Table 3, we observe, just like in the PSO implementation, an improvement after the optimization process considering InitialSP–FinalSP and InitialAcc–FinalAcc. This improvement is reflected in the statistics
and
, where an increasing trend can be seen between the initial and final states of the reservoir. Additionally, the value of
decreases in each task, demonstrating consistency in the reservoir’s performance. Similarly to PSO, for the case of the OMOPSO optimizer, we considered the median as the reference performance value to calculate the quantitative range of difference between the NEST and SpiNNaker results. This gave a value of 0.032, which similarly reflects that there is a small difference between all the results from both implementations, even smaller than the results obtained by using PSO.
In
Figure 9a, we can see that the reservoir performance improved in most of the implemented tasks, specifically in PR4, PR8, and PR12, compared to the results in [
14], with the exception of FR5. However, for FR5, we achieved a higher
performance than in the mono-objective case, with a lower
value. From the above, we can observe how FR5 may have been influenced by “optimization conformity” during the training process, where solutions were generated to benefit the majority of tasks. This suggests an underrepresentation of the FR5 task. In
Figure 9b, the maximum accuracy obtained from the 33 experiments across the four tasks is shown. The results demonstrate acceptable performance in most tasks, with FR5 achieving a performance consistent with the mean accuracy.
The values used for comparison were obtained from [
14] and were selected for each task based on the highest average performance,
, after evaluating them in each base reservoir. This reservoir was considered for transferring learning to the other tasks. The results from the SpiNNaker implementation are shown in FinalAccHW, where a smaller difference between FinalAcc and FinalAccHW can be observed compared to the mono-objective case, based on the model’s average performance. In
Figure 10, we can see
for each experiment using NEST and SpiNNaker, showing an improvement in the model’s performance dispersion, indicating appropriate behavior even with high spike rates. Although the final stage of the multi-objective optimization process did not result in a higher
for FR5, it is evident that this method outperformed all tasks when optimized independently, as seen in
Figure 11, allowing for improved reservoir performance and reduced training time.
In
Figure 12, we observe the liquid state behavior based on a sample considering the last 50 ms of simulation, with the corresponding heatmap. This process involved three stages: The first, labeled “Before Optimization”, was simulated with NEST, showing the initial behavior of the reservoir without implementing any strategies to improve its performance. This stage was characterized by a high spike rate. “After Optimization—NEST” represents the reservoir’s behavior after the optimization process, where a lower spike rate was observed. This is reflected in the heatmap, where some neurons reduce their spike frequency, further regularizing the reservoir’s behavior. “After Optimization—SpiNNaker” shows the behavior obtained by implementing the hardware, where we observe that some neurons further reduce their spike frequency.
Both in mono-objective and multi-objective optimizations, the SpiNNaker implementation showed a slight reduction in the spike rate, even in simulations with high spike frequency, when compared to the NEST implementation. This highlights the excellent performance of the hardware and the challenges we encountered, such as the noise generated by the hardware itself due to asynchronous core synchronization, the fine-tuning required through metaheuristic implementation, and the scalability of the weight representation in the hardware, among others. A significant limitation for not running the optimization process directly on the hardware was the amount of available memory for execution, which would become an even greater constraint for multi-objective implementations.
4.3. Statistical Significance Test
To validate our results, both in software (NEST simulator) and hardware (SpiNNaker board), we proceeded to perform individual statistical tests according to the used optimization method to verify whether the behavior of both implementations had the same quality or if there was any statistically significant difference and to determine which implementation was better [
41].
For both optimization methods, PSO and OMOPSO, the first step was to conduct Shapiro–Wilk normality tests [
42,
43] to determine the distribution of their results for each implementation and dataset. The results from each dataset provided by both implementations for each optimization methods showed that none of them followed a normal distribution, which led us to perform the Wilcoxon signed-rank non-parametric test [
41,
44]. Next, for each optimization method and implementation (NEST or SpiNNaker), the median value for each result set per benchmark dataset was calculated as a reference value to perform the Wilcoxon signed-rank test; thus, the statistical test for each optimization method included two implementations with four paired samples.
For the PSO optimization method, the results of the Wilcoxon signed-rank test were used to compare the performance of the two implementations, one in NEST and the other in SpiNNaker. The test statistic (V) was 0, with a p-value of 0.125. Since the p-value was greater than the used significance level of 0.05, we failed to reject the null hypothesis. This suggests that there was no statistically significant difference between the two implementations.
For the OMOPSO optimization methods, the results of the Wilcoxon signed-rank test were applied to compare the two implementations, one in NEST and the other in SpiNNaker. The test statistic (V) was 0, with a p-value of 0.25. Given that the p-value exceeded the significance level of 0.05, we failed to reject the null hypothesis.
From both statistical tests, we can conclude that the software and hardware implementations had similar performances with respect to the median accuracy. Thus, there was no decrease in performance when software implementations were migrated to hardware.
5. Conclusions
In this work, we proposed and validated an optimization approach based on Particle Swarm Optimization (PSO) to enhance reservoir separability in Liquid State Machines (LSMs), specifically for spatio-temporal classification tasks with neuromorphic systems. The experimental results demonstrated significant improvements in classification accuracy and computational efficiency, validating the implementation on the SpiNNaker neuromorphic platform. For example, with the Mono-objective approach, average performance improvements in benchmark problems such as FR5, PR4, PR8, and PR12 were achieved, with values of 86.2%, 97.7%, 94.32%, and 91.77%, respectively. Additionally, the multi-objective approach delivered exceptional results with 88.53%, 99.52%, 97.06%, and 94.42% in the same problems, consistently outperforming the Mono-objective approach. Although for both optimization methods there were no significant statistical differences in accuracy between the performance of the NEST simulator and the SpiNNaker board, the superiority of the neuromorphic platform became evident in implementation aspects, such as the reported processing time comparisons [
45].
A key aspect of this work is that SpiNNaker not only allows embedded implementations based on spiking neural networks but also does so with significantly lower power consumption compared to traditional systems like NEST [
45,
46]. This energy efficiency, coupled with real-time simulation capabilities, positions SpiNNaker as an ideal platform for the scalable implementation of LSMs.
Furthermore, PSO-optimized reservoir configurations not only improved sample separability but also reduced the neuron spike rate, contributing to more regular and efficient reservoir behavior, despite the technical challenges of the hardware. This approach is scalable and adaptable, allowing implementation on other neuromorphic platforms such as TruNorth [
47] and Loihi [
48], which also provide energy-efficient processing for spiking neural networks.
The robustness of our approach lies in its versatility, as the model can be directly implemented both in software and hardware, without requiring significant modifications. This dual applicability ensures that the solution is adaptable to a variety of deployment environments, from traditional computational platforms to specialized neuromorphic hardware. When implemented in software, the model can take advantage of the flexibility and scalability provided by general-purpose computing systems, allowing for fast prototyping and testing across different configurations and datasets. On the other hand, when deployed on neuromorphic hardware like SpiNNaker, TruNorth, or Loihi, the model benefits from the inherent efficiency and low power consumption of these platforms, enabling the real-time processing of spatio-temporal signals with minimal energy demands. This ability to seamlessly transition between software and hardware implementations not only enhances the practicality of the approach but also broadens its potential applications, making it a robust solution for both research and real-world use cases in fields such as biomedical signal processing, intelligent surveillance, and brain–computer interfaces.
In conclusion, this work marks an important step toward the optimization and scalability of LSMs on neuromorphic systems. The improvements in performance, energy efficiency, and accuracy highlight the potential of PSO as an effective tool for fully exploiting the capabilities of neuromorphic hardware. This approach benefits from envisioning suitable paths for overcoming current limitations and expanding the practical applications of LSMs on platforms such as SpiNNaker, TruNorth, and Loihi.