Next Article in Journal
Online Calibration Method of LiDAR and Camera Based on Fusion of Multi-Scale Cost Volume
Previous Article in Journal
Optimal Convolutional Networks for Staging and Detecting of Diabetic Retinopathy
Previous Article in Special Issue
Enhanced RF Power Conversion for Sensor Networks and Embedded Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Memristor-Based Neuromorphic System for Unsupervised Online Learning and Network Anomaly Detection on Edge Devices

1
Department of Electrical and Computer Engineering, University of Dayton, Dayton, OH 45469, USA
2
Center for Computational and Data Sciences, Independent University, Dhaka 1229, Bangladesh
*
Author to whom correspondence should be addressed.
Information 2025, 16(3), 222; https://doi.org/10.3390/info16030222
Submission received: 27 January 2025 / Revised: 1 March 2025 / Accepted: 5 March 2025 / Published: 13 March 2025
(This article belongs to the Special Issue Intelligent Information Processing for Sensors and IoT Communications)

Abstract

:
An ultralow-power, high-performance online-learning and anomaly-detection system has been developed for edge security applications. Designed to support personalized learning without relying on cloud data processing, the system employs sample-wise learning, eliminating the need for storing entire datasets for training. Built using memristor-based analog neuromorphic and in-memory computing techniques, the system integrates two unsupervised autoencoder neural networks—one utilizing optimized crossbar weights and the other performing real-time learning to detect novel intrusions. Threshold optimization and anomaly detection are achieved through a fully analog Euclidean Distance (ED) computation circuit, eliminating the need for floating-point processing units. The system demonstrates 87% anomaly-detection accuracy; achieves a performance of 16.1 GOPS—774× faster than the ASUS Tinker Board edge processor; and delivers an energy efficiency of 783 GOPS/W, consuming only 20.5 mW during anomaly detection.

Graphical Abstract

1. Introduction

Artificial Intelligence (AI) computation is undergoing a transformative shift, with a significant movement toward edge computing. By 2025, it is projected that 75% of enterprises will adopt edge-based computation to leverage its advantages in latency reduction and localized processing [1]. With billions of devices connected to the internet through various protocols, the volume of generated and transmitted data has grown exponentially. However, this proliferation of internet-connected devices has escalated concerns about system integrity and data security. Consequently, real-time learning of network packets for intrusion detection has become essential for ensuring uninterrupted and secure network operations [2].
Network intrusion detection systems are broadly categorized into two types: (i) signature-based systems and (ii) anomaly-based systems [3]. Signature-based systems detect threats by comparing incoming packets against a database of known signatures using state machine algorithms such as Snort, Sagan, and Suricata [4]. However, these systems are ineffective against unknown or zero-day attacks. In contrast, anomaly-based systems evaluate the similarity between incoming packets and previously known packets to identify potential threats. These systems employ statistical models, finite state machines, or Artificial Neural Networks (ANNs) [5]. ANNs are particularly effective due to their ability to dynamically adapt to new knowledge by adjusting synaptic parameters, making them a robust tool for anomaly detection in dynamic network environments.
ANNs are further classified based on their training methodologies, such as supervised and unsupervised learning [6]. Supervised neural networks, like Multilayer Perceptrons (MLPs) and Convolutional Neural Networks (CNNs), require labeled data and excel at classifying known patterns. However, their reliance on labeled data limits their suitability for real-time training, where data labels may not be readily available. In contrast, unsupervised neural networks do not require labeled data, making them ideal for real-time and adaptive applications.
Traditionally, training ANN architectures demands high-power computing platforms, such as GPUs in data centers, which come with substantial memory and energy requirements [7]. Transferring data to these centralized facilities for training introduces latency, memory overhead, and high power consumption, all of which are significant challenges in the edge AI paradigm. Additionally, deploying GPU-based training systems on resource-constrained edge or IoT devices is impractical.
As edge computing becomes increasingly prevalent, the security of edge devices has emerged as a critical concern. To address these challenges, there is a pressing need for AI training platforms capable of performing real-time data processing and online learning directly on local devices. These platforms enable personalized and secure applications while reducing reliance on centralized, energy-intensive facilities [8].
Low-power AI neural-network accelerators present a promising solution to these challenges. However, most commercially available AI accelerators are designed for inference tasks only, limiting their applicability for on-device training [9]. Furthermore, CMOS technology is nearing its physical and performance limits, with Von Neumann computing architectures facing bottlenecks due to Moore’s law, the memory wall, and the heat wall [10]. Non-Von Neumann computing paradigms, such as neuromorphic on-chip learning, offer a viable alternative for implementing neural-network systems with on-device training capabilities. In this paradigm, data are processed and stored in the same location, eliminating the need for extensive data movement. This approach, known as Processing-In-Memory (PIM), significantly enhances computational efficiency [10].
Among emerging technologies, memristors have garnered significant attention as a fundamental building block for neuromorphic computing systems. Memristors are nanoscale, non-volatile memory devices that enable non-Von Neumann computing and on-chip learning without off-chip memory access [10,11]. These devices, fabricated in crossbar architectures, perform multiply–accumulate operations—a dominant function in neural networks—efficiently and in parallel within the analog domain [11]. The small feature size of memristors enables the development of high-density neuromorphic systems with exceptional computational efficiency [10].
Memristors present a compelling alternative to traditional memory technologies, offering higher density, lower cost, and non-volatility compared to SRAM- and DRAM-based computing [12]. These resistive devices, characterized by their inherent memory effect, enable data storage and processing within the same physical structure, reducing latency and energy consumption. By integrating ultrahigh-density memristive memory directly into processor chips, the longstanding memory bottleneck in computing architectures can be significantly mitigated, leading to substantial improvements in system speed and energy efficiency [10,12]. Recent studies have further advanced memristor technology by addressing key challenges such as thermal stability, endurance, and energy efficiency, making them more viable for large-scale implementation [13,14]. As we evaluate the potential of memristors to drive computing beyond Moore’s law, their scalability and adaptability position them as key enablers of next-generation architectures, paving the way for novel paradigms in data-intensive and neuromorphic computing applications [10,12,13,14].
This article focuses on the design and implementation of a neural network-based intrusion and anomaly-detection system. The key contributions are as follows:
  • Development of an ultralow-power memristor-based unsupervised on-chip and online-learning system for edge security applications.
  • Resolution of circuit challenges for implementing online threshold computation to support real-time learning.
  • Design and simulation of a Euclidean Distance computation circuit, integrated into the memristor-based neuromorphic system for unsupervised online training.
The rest of the paper is organized as follows: Section 2 describes the dataset used in the experiment. Section 3 discusses the unsupervised learning approach for anomaly detection. Section 4 details the memristor-based implementation for the on-chip and online-learning system. Section 5 provides an overview of related works, while Section 6 elaborates on the memristor-based online-learning system. Section 7 explains the threshold-computing circuits, and Section 8 presents SPICE simulation results for these circuits. Section 9 showcases the results and analysis of online learning. Finally, Section 10 concludes the paper.

2. Network Dataset

The NSL-KDD dataset is a popular benchmark for validating machine-learning and neural-network models [15]. The dataset has one normal class and four malicious classes (DoS, Probe, R2L, and U2R). The dataset is separated into training and testing categories, and they contain 125,673 and 22,544 packets, respectively. A sample of normal and malicious network data packets is presented in Figure 1a. Both datasets consist of normal and malicious packages. Nevertheless, the test dataset contains 17 more malicious data types that are not present in the training dataset. The distribution of normal and malicious packets, along with their variants, is presented in Figure 1b.
These new categories remain unknown datatypes to supervised learning networks, as they were not present in the training session. However, the anomaly-detection system can handle these unknown categories as a zero-day attack by comparing the sequence with normal or known packets.
Each packet has 43 attributes with binary, alphanumeric, and nominal values. The nominal attributes are the protocol, service, flag, and attach types and occupied 2nd to 4th and 42nd positions. First, the nominal values are replaced with integers, and then the attributes are normalized to be bounded between 0 and 1. Examples of network packets before and after normalizations are presented in Figure 1a.

3. Unsupervised Learning

An unsupervised neural network learns the input feature space without knowing the identity or label of the sample. The purpose of unsupervised learning is to learn the underlying structure of the dataset and cluster that data according to the similarity measures.
The conceptual diagram of a simples vanilla autoencoder is presented in Figure 2. In Figure 2, X is the input feature space fed to the input layer. The encoder section learns to compress the sample into lower-dimensional latent feature space at the Z bottleneck layer. On the other hand, the decoder section learns to regenerate the input feature space at the output, X ´ . Once the AE is trained, it becomes a very efficient machine to regenerate the known classes and discriminate with the unknown inputs [16].
The encoder and decoder transformation of an AE can be denoted by ϕ and ψ , ϕ :   X Z and ψ : Z   X ´ . For instance, if we consider the simplest vanilla autoencoder [17] with a single hidden layer (Z) with input and output layers, as in Figure 2, the mathematical expression for the layer-wise signal propagation can be expressed as in Equations (1) and (2).
Z = f i = 1 n 1 w x i + b
X ´ = g j = 1 n 2 w T z j + b T
where b is biasing toward the neuron, f is the activation function, and n is the number of neurons in any layer. The inputs x i X and z j Z are the elements in latent space. The model is trained by backpropagation of error [17] to minimize the reconstruction error or loss function, l X , X ´ = X X ´ 2 . The same technique can be applied for layer-wise computation for autoencoders with multiple hidden layers, as presented in Figure 3. Equations (3)–(6) represent the system of equations for layer-wise implementation.
L 1 j = f i = 1 41 w 1 i , j . x i + b 1 j
L 2 k = f j = 1 90 w 2 j , k . h 1 j + b 2 k
L 3 j = f k = 1 10 w 2 k , j . h 2 k + b 3 j
L 4 i = f j = 1 90 w 1 j , i . h 3 j + b 4 i

4. Memristor Implementation

4.1. Memristor Device Model

The memristor SPICE model is governed by a generalized I–V relationship, as expressed in Equation (7), where the current, I(t), depends not only on the applied voltage, V(t), but also on the state variable, x(t), which dynamically modulates the device resistance. The variation in x(t) is dictated by two primary functions: g(V(t)) and f(x(t)).
I t = a 1 x t sinh b V t ,   V t 0 a 2 x t sinh b V t ,   V t < 0
g V t = A p   e V t e V p ,   V t > V p A n   e V t e V n ,   V t < V n 0 ,                               V n V t V p
The variation in x(t) is dictated by two primary functions: g(V(t)) and f(x(t)). The function g(V(t)), defined in Equation (8), enforces a threshold voltage requirement, ensuring that the state variable changes only when V(t) exceeds a certain positive or negative threshold (Vp and Vn respectively). This feature is crucial for accurately capturing the switching behavior of practical memristive devices.
f x = e a p x x p w p x , x p ,                                       x x p 1 ,                                                                                                     x < x p
f x = e a n x + x n 1 w n x , x n ,                                       x 1 x n 1 ,                                                                                                       x < 1 x n
Meanwhile, the function f(x(t)), presented in Equations (9) and (10), models the nonlinear motion of ions within the memristor, accounting for the increased difficulty in altering the state as x(t) approaches its boundary limits. The overall state variable dynamics are then expressed in Equation (10), where the time derivative, dx/dt, is determined by the product of g(V(t)) and f(x(t)), scaled by a coefficient, η. This comprehensive modeling framework ensures that the proposed memristor model can accurately represent devices with varying physical structures, making it suitable for simulating diverse memristive systems. Readers are referred to [18,19] for more details on the memristor model.
d x d t = η g V t f x t

4.2. Memristor Neuron Circuit

Figure 4a,b present a memristor-based neuron circuit consisting of N inputs (X) and one bias (β). The output neuron is connected to the error generation circuits (see Figure 4b). The Euclidean distance computing circuitry is implemented for online-learning experiments to compute the threshold in real time. In the neuron circuit, the synaptic element is represented by 0T2M, where there is a pair of memristor devices along the row, as presented in Figure 4 [18,20,21,22]. No transistor was used for the cell unit, which reduces the cell size and makes it about 36 times smaller than a typical SRAM cell. The synaptic weight can be negative ( σ + σ < 0 ) or positive ( σ + σ > 0 ) based on the magnitude of the conductance of memristor pairs. The synapse receives the inputs as an analog voltage pulse, multiplies with the synaptic conductance, and gives a resultant current that flows along the common column. The total current is sensed at the bottom of the neuron circuit as in Equation (12), and the output voltage from the op-amp circuits is the resultant voltage Ym (see Equation (13)). The neuron circuit creates approximated sigmoid activation (Equation (14)) in op-amp circuits, where the power rail of the op-amp is connected to VDD = 1 V and VSS = 0 V.
The sigmoid is typically used as the activation function in deep learning, which is represented as f x = 1 / ( 1 + e x ) . However, in the presented memristor-based neuromorphic system, the approximated sigmoid function in Equation (14) is used, as it is easier to generate using an amplifier circuit [20]. Figure 5 displays the typical sigmoid function, along with the approximation used in this work.
I j = N X N σ i j + σ i j + X β σ β + σ β
Y m = R f I j
y = 1 ,                                           y > 2 0.25 y + 0.5 ,               y 2 0 ,                                                 y < 2

4.3. Crossbar Training Circuit

The memristor-based crossbar circuit is trained in the full analog domain. The forward propagation is performed the same way the inference is computed in multilayered neural networks. The training samples are applied to the input nodes of the neural networks, producing vector–matrix multiplication and giving a neuronal sum at the output nodes. An error backpropagation process [23,24] is utilized in the training process.
For training the autoencoder circuits, transposable arrays are designed where the input can be applied both row-wise and column-wise. The error ( δ j ) is generated from the error generation circuit for each input sample, as shown in Figure 6. The error is backpropagated to the same crossbar circuit, but from the transposed direction, and produce a resulted sum, j δ j σ i j .
The amount is multiplied with the derivation of neuron activation to compute the layer error gradient. The weight update of the neuron is performed based on the product of activation and error gradient ( x N δ j ) of each layer. For weight update, pulses are produced with varible heights and widths. The height is modulated by the input activations and the widht is shaped by the product of learning rate and error gradient ( η δ j ), as discussed in [10]. The detailed training steps are presented in pictorial form in Figure 5, with transposable crossbar array circuits and peripheral details. The Euclidean distance computing module is necessary for determining the anomalies in the computer networks. This module is embedded in the online-learning process to calculate the threshold. Analog computations are severely affected by noise from different pieces of analog circuits, such as ADC/DAC, op-amps, and the memristor array. The noise is introduced in the circuit simulation during weight updates, as described in Equation (15), where η δ j is the weight update amount, and n j is random noise.
w n e w j = w o l d j + η δ j +   n j

5. Online Learning and Related Works

5.1. Online-Learning Memristor Neuron Circuit

Online learning, akin to continual lifelong learning [25], addresses the challenge of sequential data streams, in contrast to the batch-learning approach commonly used in deep-learning training. In online learning, the system learns from a continuous sequence of incoming data without requiring pretrained knowledge or a fixed dataset. As new samples arrive, the network updates its parameters in real time. However, a major drawback of this approach is catastrophic forgetting [26], where the model tends to forget previously learned information when new data are introduced. This phenomenon significantly reduces model performance, as older knowledge is overwritten by new data.
State-of-the-Art deep-learning algorithms have demonstrated superior performance in classification tasks, especially in large-scale data processing. However, these algorithms assume that all possible data types are included in the training set. When trained sequentially, the model experiences a decline in accuracy due to catastrophic forgetting. To mitigate this issue, an effective learning system must possess the ability to acquire new knowledge while preserving previously learned information, thus overcoming the challenge of catastrophic forgetting

5.2. Related Works on Online Learning

Real-world computational units often encounter continuous streams of information, necessitating real-time learning systems capable of handling dynamic data distributions and learning multiple tasks. Various machine-learning and deep-learning algorithms have been proposed for online-learning applications. This section highlights some relevant systems.
Continuous and lifelong learning, a key area in online learning, has been an active research frontier for decades, with numerous algorithms developed for this purpose [27]. S. A. Rebuffi et al. introduced the iCARL algorithm, a supervised lifelong online-learning approach that can identify new categories and incrementally learn classes, though it requires storing samples of all classes for fine-tuning [28]. M. Pratama et al. proposed the evolving denoising autoencoder (DEVDAN) for online incremental learning on non-stationary data streams; it adapts by generating, pruning, and learning hidden nodes on the fly [29]. S. S. Sarwar et al. presented a deep CNN (DCNN) for incremental learning, addressing catastrophic forgetting by partially sharing trained knowledge and using new branches for new classes [30]. Hierarchical temporal memory (HTM) has been applied for online learning and anomaly detection in various domains, including 3D printing and video anomaly detection [31,32]. Extreme learning machines (ELMs) and Adaptive Resonance Theory (ART) algorithms have been explored for dynamic modeling, anomaly detection, and intrusion detection [33,34,35,36,37,38,39,40].
While most of these systems are implemented in software, there are a few hardware implementations of online-learning algorithms with emerging resistive memory devices. ELM has been implemented in both spiking [41] and non-spiking [42] neuromorphic systems, and ART in FPGA [43] and memristor-based systems [37]. Autoencoders have been implemented in FPGA for unsupervised learning, such as detecting new physical phenomena in the large hadron collider [44]. However, to the best of our knowledge, no one has previously implemented unsupervised autoencoders in memristor neuromorphic systems for online learning and anomaly detection. This work presents such an implementation, detailing the supporting circuits for online threshold computation and neural-network optimization [20,21].

5.3. Online-Learning Systems

The online-learning system in this study is presented in Figure 7a, with a conceptual diagram. The system is developed as a complement to the existing rule-based IDS system. AE-1 is a pretrained autoencoder that is trained on normal packets. AE-2 is implemented to learn only on malicious packets. The malicious packets are coming from AE-1 in real time.

6. Memristor-Based Online-Learning Systems

Unsupervised autoencoders are implemented in memristor-based neuromorphic architecture for anomaly-detection applications. The system has two main parts, the pretrained section and the online-learning section. In the pretrained section, the pretrained autoencoder detects potential malicious packets and directs them to the second autoencoder, where the AE-2 learns from the malicious packets only. The flowchart of the online-learning system is presented in Figure 7b. While learning about the malicious packets, the system can identify any new potential threat to the network, known as the zero-day attack, which is yet unfamiliar or an anomaly. However, with the repeated appearance of unknown packets, they become familiar to the network and are eventually recognized as known malicious packets by the system.
For online anomaly detection, the system needs to compute Euclidean distance for every network packet coming to AE-2. The ED is computed between the input and output feature vector of the network packets to determine the threshold for detecting the known or unfamiliar malicious packets. Online learning is taking place on edge devices, so there is no high-power floating processor to compute ED. In this work, we have proposed a detailed circuitry for computing ED in real time for optimizing the network and computing the threshold.
In online learning, the system needs to update the threshold for network anomaly detection, as it continuously learns and optimizes the network parameters for detecting the newly learned packets and the unknown packets. The threshold is computed using a series of analog squaring circuits that are able to compute reconstruction errors in a single shot. The block diagram of the Euclidean distance computing block is presented in Figure 8a. The differences between the input vector and generated output vector are d1, d2, …, dn fed into the analog squaring circuit. A summing amplifier accumulates the output currents of the squaring circuits. The output of the summing amplifier is digitized through an analog-to-digital converter (ADC) and stored as the equivalent Euclidean distance in the look-up table (LUT), a peripheral memory module. After each learning cycle, the stored data are used to compute the threshold as the standard deviation in the digital system and then convert the result into an analog equivalent voltage signal for anomaly detection, as described in Figure 8b.

7. Analog Threshold Computing

The analog squaring circuits are integral to computing the training error as the Euclidean distance for online training. Each incoming sample’s distance is digitized using a low-precision ADC, stored in peripheral memory, and periodically used to update the anomaly-detection threshold. The distance computing circuit, shown in Figure 9, is divided into three main components: (A) absolute differential voltage computation, (B) voltage-to-current conversion, and (C) the squaring section.
Section A computes the absolute L1 norm (Manhattan distance) between the input and output vectors of the autoencoder circuit, providing the result at node P. Section B converts the input voltage Vp into a proportional current. Section C calculates the square of the Manhattan distance using the current input, (Ix-y).
In the online training of the NSL-KDD network packet dataset, the system requires 41 squaring circuits to efficiently perform training and anomaly detection. The system can also employ time-multiplexing techniques for enhanced functionality. All 41 squaring circuits compute the Euclidean distance simultaneously. After computation, the output current is passed through transistor M7, where the current, Iout, generates a voltage drop at the output node Q. The voltage outputs from all squaring circuits are then accumulated using a summing amplifier and stored in a look-up table (LUT). The square root of the accumulated voltage values is used to determine the actual Euclidean distance.
The circuit presented in Figure 9 is implemented in SPICE simulation with actual device parameters to obtain the realistic characteristics of the training and anomaly-detection performance. Figure 10a presents the SPICE simulation for the switching time of analog squaring circuits implemented for computing the reconstruction error of the autoencoder. The graph shows two inputs (X and Y) applied at the differential circuit in Figure 10a and the differential voltage at node P. The timing for executing this circuit operation is 400 ns, considered the switching time. The switching time is required for the transistors in Figure 10a to develop the maximum output voltage level at node Q, while the input voltage is applied to the differential circuits. To compute the reconstruction error of the NSL-KDD, network packets need 41 of these circuits; thus, total energy consumption can be computed accordingly, as shown in Equation (16).
E = 41 t 1 t 2 P · t   d t
The power consumption of the error computing circuit is presented in Figure 10b over the simulated switching time. To compute the total energy consumption, the area under the curve in Figure 9B is computed by the integration over time t2 to t1, and the computed energy consumption is 20.2 nJ for computing the reconstruction error of an NSL-KDD packet.

8. Simulation of Online Threshold-Computing Circuit

The squaring circuit is the main computing component for online learning and anomaly detection. The actual circuit is implemented in the LTSpice and simulated with realistic circuit parameters by using actual device parameters. For this implementation, the ASU 90 nm transistor model was utilized as a well-established hardware preference. For differential calculations, ideal op-amp devices are considered for simplicity. The simulation results of Figure 9 are presented in Figure 11a,b.
The output current is the resulted mirror current flowing through M7, and that is a squared value normalized by the bias current (Ibias). The simulated result is compared to the theoretical fitting relation, ( I i n 2 I b i a s ) [45], and follows the quadratic fit. The output voltage is measured at node Q (Figure 11b), and fitting parameters of the output voltage are utilized in the MATLAB R2022a simulation to estimate the threshold for anomaly detection. All the parameters for circuit simulation are presented in Table 1. The W and L are the width and length of the transistor and given in µm.

9. Results and Discussion

The proposed online-learning and anomaly-detection system is highly dependent on the threshold-computing system. This work implemented an analog threshold-computing system with a low-precision and power-efficient ADC [46] for data conversion.

9.1. Crossbar Training Analysis

The crossbar training is performed completely in the analog domain, without using any data conversion units attached with crossbar circuits. The crossbars layers are connected toe-to-toe style. The crossbar training procedure is discussed in section III(B). This section discusses the training results and performance of AE on intrusion detection. The circuit parameters for training are presented in Table 2. The realistic SPICE model was utilized for memristor device circuit simulation [18,19]. The memristor device with a higher Roff/Ron ratio was considered from practical memristor device data presented in [47].

9.2. Pretrained AE

The pretraining of AE-1 in Figure 7a is performed on-chip, and the threshold is computed in floating. For AE-based intrusion detection, V t h is used as the reference or threshold voltage in the comparator circuit to find knowns/anomalies among the incoming packets. If the amount of VED − Vm > Vth, then the system identifies the incoming packet as an unknown and labels it as 1. Otherwise, the packet is labeled as known or 0. The comparator op-amp is biased with 1 and 0 voltage supply. Figure 12a is a comparator circuit to detect the incoming packets as normal or malicious. Figure 12b presents the intrusion detection accuracy, sensitivity, and precision of memristor-based AE-1, while the baseline software accuracy is 95.86% and 93.54%, respectively, for known and unknown network packets. AEs outperform supervised models in zero-day attack identification using anomaly-based intrusion detection, as discussed in our previous work [21].

9.3. Online-Training Analysis

The training process of AE-2 (Figure 7a) operates in real time without relying on a floating-point processor. Various test datasets, including subsets of the NSL-KDD dataset, were used for online-learning experiments. For instance, in Set-1, Normal and Probe packets were sent to the system, and AE-2 was instructed to learn only the Probe packets since AE-1 identified them as potential threats. Online training of AE-2 is illustrated in Figure 13a.
When new malicious packets, such as those in Set-2, breached the rule-based IDS, AE-1 flagged and routed them to AE-2 for further learning. AE-2 identified these zero-day attacks by flagging the unknown packets, leading to a temporary increase in anomalous packets and training errors (Figure 13a). However, with subsequent learning cycles, the system adapted, reducing anomalies and training errors as the packets became familiar. Pretraining of AE-1 (Figure 7a) was performed on-chip, with the threshold Vth computed in floating point for reference in AE-based intrusion detection.
Figure 13b illustrates network packet detection at different stages of the intrusion and anomaly-detection system. While incoming packets are unlabeled, labels 0 (True Normal) and 1 (True Malicious) are used to assess detection accuracy. Pretrained AE-1, recognizing only normal packets, achieves over 93.5% accuracy in distinguishing normal (PN) and malicious (PM) packets, with the threshold marked by the red line. The lower part of Figure 13b shows AE-2’s online anomaly detection, where its threshold is computed in real time using analog–digital hybrid CMOS circuits (Figure 9 and Figure 10).
The online threshold is updated after each learning cycle, with the system storing the Euclidean distance as a digital equivalent in a LUT. While neural-network computations occur entirely in the analog domain (Figure 6), an ADC is used solely for converting the AE’s generative error to compute thresholds. Training is online, learning sequentially from individual samples without batch processing. Figure 14a compares the floating-point threshold with that obtained using ADCs of various bit widths.
Figure 14b illustrates AE-2’s online incremental learning and real-time anomaly detection. Initially, AE-2’s parameters are random, and Testset-1 (containing normal and Probe packets) is processed. AE-1 identifies and forwards malicious packets to AE-2 for learning. At first, anomalies exceed 500, as AE-2 lacks knowledge of Probe packets. Over time, AE-2 learns, increasing known packets and reducing anomalies. At learning cycle 20, DoS packets cause a spike in anomalies, but subsequent cycles show adaptation, with anomalies decreasing as AE-2 becomes familiar. This iterative process enables the system to continuously learn and detect new malicious packets.
During online learning, AE-1 and AE-2 detect intrusions and anomalies, with detection accuracy shown in Figure 15a. AE-1’s accuracy remains steady until new intrusions occur, while AE-2 starts at zero and improves over training cycles. At cycle 19, AE-1’s accuracy rises slightly as it identifies unknown attacks, while AE-2’s accuracy drops temporarily due to new anomalies. Over time, AE-2 adapts, and accuracy stabilizes until further anomalies appear.
The ADC resolution significantly impacts threshold computation in anomaly-detection systems. While Figure 14a shows minimal difference in anomaly-detection performance between floating-point and low-precision threshold computation, Figure 15b highlights a noticeable accuracy drop with a 5-bit ADC. However, using a 6-bit or 7-bit ADC provides a balanced trade-off, making them suitable options for developing online-learning and anomaly-detection systems in edge security applications.
Memristor-based analog computing systems are often affected by device variability and circuit noise from analog sources. Figure 16 demonstrates the system’s performance under varying levels of noise, showing a clear decline in accuracy as noise increases. The random noise in the system is scaled relative to the minimum conductivity, highlighting the impact of analog noise on system reliability.

9.4. System Energy, Power, and Performance Analysis

The anomaly-detection system was benchmarked against a low-power commercial edge processor, the ASUS Tinker Board, which features a Rockchip AI processor (RK3288), ARM Mali-T764 GPU, and 2 GB DDR3 RAM [48]. Using 2000 NSL-KDD packets, the Tinker Board required 1.8 s for detection, whereas the memristor-based system completed the task in just 0.0023 s—three orders of magnitude faster. The memristor system demonstrated 184,000× higher energy efficiency, achieving 16.1 GOPS and 783 GOPS/W, compared to the Tinker Board’s 20 MOPS and 4.12 MOPS/W. Table 3 compares numerical results for both systems. Other NSL-KDD dataset-based systems using floating-point processors and machine-learning algorithms typically achieve a performance ranging from kilo to mega operations per second [49,50,51].
Figure 17a,b provide a breakdown of chip area and energy consumption for anomaly-detection operations. During forward propagation, crossbar arrays and op-amps occupy the largest portion of the chip area (60%) if only one autoencoder is used. When two autoencoders are stacked, they occupy less area, allowing ADCs, DACs, and LUTs to take up 16%, and other digital circuits, including weight update and control units, to occupy 27%. Implementing ADC/DAC modules beneath the autoencoder circuits can significantly reduce chip footprint.
Energy analysis shows that threshold-computing circuits account for 62% of energy usage during forward propagation. Differential and squaring circuits consume 20.2 nJ (0.493 nJ per circuit), while ADC/DAC modules, which are only used for reconstruction error and error gradient conversion, consume minimal energy. This energy-efficient design highlights the advantages of the memristor-based system over conventional processors.

10. Conclusions

In conclusion, a memristor-based on-chip system has been developed for online anomaly detection in edge devices, showcasing its potential for energy-efficient and real-time network security applications. The system integrates analog training circuits for unsupervised neural networks with a Euclidean distance computation unit for reconstruction error calculations, enabling effective detection of zero-day attacks. Compared to commercially available edge devices, this system offers significantly lower energy consumption while achieving impressive performance metrics of 16.1 GOPS and 783 GOPS/W, consuming just 20.5 mW during anomaly detection. These results highlight the system’s viability for scalable and efficient edge security solutions.
In our future work, we plan to implement this system in physical hardware and conduct extensive experiments with various edge devices. Additionally, we aim to perform a detailed experimental analysis of the proposed system to further validate its performance and feasibility.

Author Contributions

Conceptualization, M.S.A. and T.M.T.; Methodology, M.S.A. and T.M.T.; software development, M.S.A. and R.H.; investigation M.S.A.; formal analysis M.S.A.; original draft preparation, M.S.A.; review and editing, M.S.A., C.Y., R.H. and T.M.T.; supervision, C.Y. and T.M.T.; resources C.Y. and T.M.T.; review and editing, M.S.A., C.Y., T.M.T. and R.H.; project administration, T.M.T.; funding acquisition, T.M.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Science Foundation, under the grant 1718633.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available NSL-KDD dataset was utilized for this study. Research data will be available upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Singh, R.; Gill, S.S. Edge AI: A survey. Internet Things Cyber-Phys. Syst. 2023, 3, 71–92. [Google Scholar] [CrossRef]
  2. Wang, B.; Zheng, H.; Qian, K.; Zhan, X.; Wang, J. Edge computing and AI-driven intelligent traffic monitoring and optimization. Appl. Comput. Eng. 2024, 67, 225–230. [Google Scholar] [CrossRef]
  3. Bhavsar, M.; Roy, K.; Kelly, J.; Olusola, O. Anomaly-based intrusion detection system for IoT application. Discov. Internet Things 2023, 3, 5. [Google Scholar] [CrossRef]
  4. Zarpelão, B.B.; Miani, R.S.; Kawakani, C.T.; de Alvarenga, S.C. A survey of intrusion detection in Internet of Things. J. Netw. Comput. Appl. 2017, 84, 25–37. [Google Scholar] [CrossRef]
  5. Thudumu, S.; Branch, P.; Jin, J.; Singh, J. A comprehensive survey of anomaly detection techniques for high dimensional big data. J. Big Data 2020, 7, 42. [Google Scholar] [CrossRef]
  6. Srinivas, T.; Aditya Sai, G.; Mahalaxmi, R. A comprehensive survey of techniques, applications, and challenges in deep learning: A revolution in machine learning. Int. J. Mech. Eng. 2022, 7, 286–296. [Google Scholar]
  7. Naumov, M.; Kim, J.; Mudigere, D.; Sridharan, S.; Wang, X.; Zhao, W.; Yilmaz, S.; Kim, C.; Yuen, H.; Ozdal, M.; et al. Deep learning training in facebook data centers: Design of scale-up and scale-out systems. arXiv 2020, arXiv:2003.09518. [Google Scholar]
  8. Lin, J.; Zhu, L.; Chen, W.M.; Wang, W.C.; Gan, C.; Han, S. On-device training under 256 kb memory. Adv. Neural Inf. Process. Syst. 2022, 35, 22941–22954. [Google Scholar]
  9. Luo, Y.; Peng, X.; Yu, S. MLP + NeuroSimV3. 0: Improving on-chip learning performance with device to algorithm optimizations. In Proceedings of the International Conference on Neuromorphic Systems, Knoxville, TN, USA, 23–25 July 2019; pp. 1–7. [Google Scholar]
  10. Sebastian, A.; Le Gallo, M.; Khaddam-Aljameh, R.; Eleftheriou, E. Memory devices and applications for in-memory computing. Nat. Nanotechnol. 2020, 15, 529–544. [Google Scholar] [CrossRef]
  11. Rao, M.; Tang, H.; Wu, J.; Song, W.; Zhang, M.; Yin, W.; Zhuo, Y.; Kiani, F.; Chen, B.; Jiang, X.; et al. Memristor devices denoised to achieve thousands of conductance levels. Nature 2023, 615, 823–829. [Google Scholar] [CrossRef]
  12. Zidan, M.A.; Strachan, J.P.; Lu, W.D. The future of electronics based on memristive systems. Nat. Electron. 2018, 1, 22–29. [Google Scholar] [CrossRef]
  13. Lv, Z.; Zhu, S.; Wang, Y.; Ren, Y.; Luo, M.; Wang, H.; Zhang, G.; Zhai, Y.; Zhao, S.; Zhou, Y.; et al. Development of Bio-Voltage Operated Humidity-Sensory Neurons Comprising Self-Assembled Peptide Memristors. Adv. Mater. 2024, 36, 2405145. [Google Scholar] [CrossRef] [PubMed]
  14. Lv, Z.; Jiang, M.; Liu, H.; Li, Q.; Xie, T.; Yang, J.; Wang, Y.; Zhai, Y.; Ding, G.; Zhu, S.; et al. Temperature-Resilient Polymeric Memristors for Effective Deblurring in Static and Dynamic Imaging. Adv. Funct. Mater. 2025, 2424382. [Google Scholar] [CrossRef]
  15. NSL-KDD Dataset. Available online: https://www.unb.ca/cic/datasets/nsl.html (accessed on 5 May 2020).
  16. Takiddin, A.; Ismail, M.; Zafar, U.; Serpedin, E. Deep autoencoder-based anomaly detection of electricity theft cyberattacks in smart grids. IEEE Syst. J. 2022, 16, 4106–4117. [Google Scholar] [CrossRef]
  17. Sorrenson, P.; Draxler, F.; Rousselot, A.; Hummerich, S.; Zimmerman, L.; Köthe, U. Maximum Likelihood Training of Autoencoders. arXiv 2023, arXiv:2306.01843. [Google Scholar]
  18. Yakopcic, C.; Taha, T.M.; Subramanyam, G.; Pino, R.E. Memristor SPICE model and crossbar simulation based on devices with nanosecond switching time. In Proceedings of the 2013 International Joint Conference on Neural Networks (IJCNN), Dallas, TX, USA, 4–9 August 2013; pp. 1–7. [Google Scholar]
  19. Yakopcic, C.; Taha, T.M.; Mountain, D.J.; Salter, T.; Marinella, M.J.; McLean, M. Memristor model optimization based on parameter extraction from device characterization data. IEEE Trans. Comput. Des. Integr. Circuits Syst. 2019, 39, 1084–1095. [Google Scholar] [CrossRef]
  20. Alam, M.S.; Fernando, B.R.; Jaoudi, Y.; Yakopcic, C.; Hasan, R.; Taha, T.M.; Subramanyam, G. Memristor based autoencoder for unsupervised real-time network intrusion and anomaly detection. In Proceedings of the International Conference on Neuromorphic Systems, Knoxville, TN, USA, 23–25 July 2019; pp. 1–8. [Google Scholar]
  21. Alam, M.S.; Yakopcic, C.; Taha, T.M. Unsupervised Learning of Memristor Crossbar Neuromorphic Processing Systems. U.S. Patent Application 17/384,306, 23 July 2021. [Google Scholar]
  22. Hasan, R.; Taha, T.M.; Yakopcic, C. On-chip training of memristor based deep neural networks. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 3527–3534. [Google Scholar]
  23. Krestinskaya, O.; Salama, K.N.; James, A.P. Learning in memristive neural network architectures using analog backpropagation circuits. IEEE Trans. Circuits Syst. I Regul. Pap. 2018, 66, 719–732. [Google Scholar] [CrossRef]
  24. Tsai, S.; Ambrogio, S.; Narayanan, P.; Shelby, B.; Mackin, C.; Burr, G. Analog memory-based techniques for accelerating the training of fully-connected deep neural networks. In Proceedings of the SPIE Advanced Lithography, San Jose, CA, USA, 26–27 February 2019. [Google Scholar]
  25. Hayes, T.L.; Kanan, C. Lifelong machine learning with deep streaming linear discriminant analysis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; pp. 220–221. [Google Scholar]
  26. Shao, C.; Feng, Y. Overcoming catastrophic forgetting beyond continual learning: Balanced training for neural machine translation. arXiv 2022, arXiv:2203.03910. [Google Scholar]
  27. Chaudhri, Z.; Liu, B. Lifelong Machine Learning; Springer International Publishing: Cham, Switzerland, 2022. [Google Scholar]
  28. SRebuffi, A.; Kolesnikov, A.; Sperl, G.; Lampert, C.H. iCaRL: Incremental Classifier and Representation Learning. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5533–5542. [Google Scholar]
  29. Ashfahani, A.; Pratama, M.; Lughofer, E.; Ong, Y.-S. DEVDAN: Deep evolving denoising autoencoder. Neurocomputing 2020, 390, 297–314. [Google Scholar] [CrossRef]
  30. Sarwar, S.S.; Ankit, A.; Roy, K. Incremental Learning in Deep Convolutional Neural Networks Using Partial Network Sharing. IEEE Access 2020, 8, 4615–4628. [Google Scholar] [CrossRef]
  31. Malawade, A.V.; Costa, N.D.; Muthirayan, D.; Khargonekar, P.P.; Al Faruque, M.A. Neuroscience-Inspired Algorithms for the Predictive Maintenance of Manufacturing Systems. IEEE Trans. Ind. Inform. 2021, 17, 7980–7990. [Google Scholar] [CrossRef]
  32. Monakhov, V.; Thambawita, V.; Halvorsen, P.; Riegler, M.A. GridHTM: Grid-Based Hierarchical Temporal Memory for Anomaly Detection in Videos. Sensors 2023, 23, 2087. [Google Scholar] [CrossRef] [PubMed]
  33. Faezi, S.; Yasaei, R.; Barua, A.; Al Faruque, M.A. Brain-Inspired Golden Chip Free Hardware Trojan Detection. IEEE Trans. Inf. Forensics Secur. 2021, 16, 2697–2708. [Google Scholar] [CrossRef]
  34. Wang, J.; Lu, S.; Wang, S.-H.; Zhang, Y.-D. A review on extreme learning machine. Multimed. Tools Appl. 2022, 81, 41611–41660. [Google Scholar] [CrossRef]
  35. Zhang, J.; Li, Y.; Xiao, W. Adaptive online sequential extreme learning machine for dynamic modeling. Soft Comput. 2021, 25, 2177–2189. [Google Scholar] [CrossRef]
  36. Wang, X.; Tu, S.; Zhao, W.; Shi, C. A novel energy-based online sequential extreme learning machine to detect anomalies over real-time data streams. Neural Comput. Appl. 2022, 34, 823–831. [Google Scholar] [CrossRef]
  37. Albadr, M.A.A.; Tiun, S.; Ayob, M.; Al-Dhief, F.T.; Abdali, T.-A.N.; Abbas, A.F. Extreme Learning Machine for Automatic Language Identification Utilizing Emotion Speech Data. In Proceedings of the 2021 International Conference on Electrical, Communication, and Computer Engineering (ICECCE), Kuala Lumpur, Malaysia, 12–13 June 2021; pp. 1–6. [Google Scholar]
  38. Alam, M.S.; Yakopcic, C.; Subramanyam, G.; Taha, T.M. Memristor based neuromorphic adaptive resonance theory for one-shot online learning and network intrusion detection. In Proceedings of the International Conference on Neuromorphic Systems, 2020 (ICONS 2020), Oak Ridge, TN, USA, 28–30 July 2020; pp. 1–8, Article No. 25. [Google Scholar]
  39. Jones, C.B.; Carter, C.; Thomas, Z. Intrusion Detection & Response using an Unsupervised Artificial Neural Network on a Single Board Computer for Building Control Resilience. In Proceedings of the 2018 Resilience Week (RWS), Denver, CO, USA, 20–23 August 2018; pp. 31–37. [Google Scholar]
  40. Kumar, A.; Sachdeva, N. Cyberbullying checker: Online bully content detection using Hybrid Supervised Learning. In International Conference on Intelligent Computing and Smart Communication: Proceedings of ICSC 2019; Springer: Singapore, 2020; pp. 371–382. [Google Scholar]
  41. Dong, Z.; Lai, C.S.; Zhang, Z.; Qi, D.; Gao, M.; Duan, S. Neuromorphic extreme learning machines with bimodal memristive synapses. Neurocomputing 2021, 453, 38–49. [Google Scholar] [CrossRef]
  42. Liang, H.; Cheng, H.; Wei, J.; Zhang, L.; Yang, L.; Zhao, Y.; Guo, H. Memristive neural networks: A neuromorphic paradigm for extreme learning machine. IEEE Trans. Emerg. Top. Comput. Intell. 2019, 3, 15–23. [Google Scholar] [CrossRef]
  43. Bureneva, O.I.; Prasad, M.S.; Verma, S. FPGA-based Hardware Implementation of the ART-1 classifier. In Proceedings of the 2023 XXVI International Conference on Soft Computing and Measurements (SCM), Saint Petersburg, Russia, 24–26 May 2023; pp. 171–174. [Google Scholar]
  44. Govorkova, E.; Puljak, E.; Aarrestad, T.; James, T.; Loncar, V.; Pierini, M.; Pol, A.A.; Ghielmetti, N.; Graczyk, M.; Summers, S.; et al. Autoencoders on field-programmable gate arrays for real-time, unsupervised new physics detection at 40 MHz at the Large Hadron Collider. Nat. Mach. Intell. 2022, 4, 154–161. [Google Scholar] [CrossRef]
  45. Nag, A.; Paily, R.P. Low power squaring and square root circuits using subthreshold MOS transistors. In Proceedings of the 2009 International Conference on Emerging Trends in Electronic and Photonic Devices & Systems, Varanasi, India, 22–24 December 2009; pp. 96–99. [Google Scholar]
  46. Jiang, H.; Li, W.; Huang, S.; Cosemans, S.; Catthoor, F.; Yu, S. Analog-to-Digital Converter Design Exploration for Compute-in-Memory Accelerators. IEEE Des. Test 2022, 39, 48–55. [Google Scholar] [CrossRef]
  47. Yu, S.; Wu, Y.; Wong, H.-S.P. Investigating the switching dynamics and multilevel capability of bipolar metal oxide resistive switching memory. Appl. Phys. Lett. 2011, 98, 103514. [Google Scholar] [CrossRef]
  48. ASUS. Tinkerboard. Available online: https://tinker-board.asus.com/product/tinker-board.html (accessed on 1 January 2020).
  49. Suleiman, M.F.; Issac, B. Performance Comparison of Intrusion Detection Machine Learning Classifiers on Benchmark and New Datasets. In Proceedings of the 2018 28th International Conference on Computer Theory and Applications (ICCTA), Alexandria, Egypt, 30 October–1 November 2018; pp. 19–23. [Google Scholar]
  50. Sidharth, V.; Kavitha, C.R. Network intrusion detection system using stacking and boosting ensemble methods. In Proceedings of the 2021 Third International Conference on Inventive Research in Computing Applications (ICIRCA), Coimbatore, India, 2–4 September 2021; pp. 357–363. [Google Scholar]
  51. Hamolia, V.; Melnyk, V.; Zhezhnych, P.; Shilinh, A. Intrusion detection in computer networks using latent space representation and machine learning. Int. J. Comput. 2020, 19, 442–448. [Google Scholar] [CrossRef]
Figure 1. NSL-KDD dataset: (a) normal and malicious packets in raw and normalized format; and (b) distribution of normal and malicious packets in training and test dataset.
Figure 1. NSL-KDD dataset: (a) normal and malicious packets in raw and normalized format; and (b) distribution of normal and malicious packets in training and test dataset.
Information 16 00222 g001
Figure 2. Schematic presentation of an autoencoder learning network.
Figure 2. Schematic presentation of an autoencoder learning network.
Information 16 00222 g002
Figure 3. Neural-network model of the autoencoder with the layer size.
Figure 3. Neural-network model of the autoencoder with the layer size.
Information 16 00222 g003
Figure 4. Single neuron: (a) input and hidden layers; and (b) output layer.
Figure 4. Single neuron: (a) input and hidden layers; and (b) output layer.
Information 16 00222 g004
Figure 5. Plot displaying the traditional sigmoid function, along with approximated sigmoid function.
Figure 5. Plot displaying the traditional sigmoid function, along with approximated sigmoid function.
Information 16 00222 g005
Figure 6. A crossbar circuit representing the ANN output layer with inputs and corresponding outputs, error generation circuit, and Euclidean distance computation circuitry.
Figure 6. A crossbar circuit representing the ANN output layer with inputs and corresponding outputs, error generation circuit, and Euclidean distance computation circuitry.
Information 16 00222 g006
Figure 7. (a) The online-learning and anomaly-detection system. (b) Flowchart of online-learning and anomaly-detection systems.
Figure 7. (a) The online-learning and anomaly-detection system. (b) Flowchart of online-learning and anomaly-detection systems.
Information 16 00222 g007
Figure 8. (a) Block diagram of circuitry that is implemented for online Euclidean distance computation for anomaly detection. (b) Block diagram of Euclidean distance calculation circuit.
Figure 8. (a) Block diagram of circuitry that is implemented for online Euclidean distance computation for anomaly detection. (b) Block diagram of Euclidean distance calculation circuit.
Information 16 00222 g008
Figure 9. Threshold-computing circuit: (A) gives a non-negative or absolute output voltage from the difference of input and output vector of autoencoder; (B) converts voltage into current, Ix-y; and (C) squares the incoming current and the output voltage sensed in the load resistance, RH. The SPICE simulation utilized device parameters for 90 nm technology; other parameters are R = 1 kΩ, R1 = 1.5 MΩ, and RH = 10 kΩ.
Figure 9. Threshold-computing circuit: (A) gives a non-negative or absolute output voltage from the difference of input and output vector of autoencoder; (B) converts voltage into current, Ix-y; and (C) squares the incoming current and the output voltage sensed in the load resistance, RH. The SPICE simulation utilized device parameters for 90 nm technology; other parameters are R = 1 kΩ, R1 = 1.5 MΩ, and RH = 10 kΩ.
Information 16 00222 g009
Figure 10. (a) Switching timing and input–output voltage measurement in proposed circuit (Figure 8). (b) Power consumption over time of ED computing circuit.
Figure 10. (a) Switching timing and input–output voltage measurement in proposed circuit (Figure 8). (b) Power consumption over time of ED computing circuit.
Information 16 00222 g010
Figure 11. (a) Measured current and estimated current flowing in the output squaring circuit. (b) Measured voltage and estimated voltage at the output node.
Figure 11. (a) Measured current and estimated current flowing in the output squaring circuit. (b) Measured voltage and estimated voltage at the output node.
Information 16 00222 g011
Figure 12. (a) The comparator circuit for identifying the known anomalies of incoming packets. (b) Performance score of autoencoder AE-1 for known and zero-day attacks offline.
Figure 12. (a) The comparator circuit for identifying the known anomalies of incoming packets. (b) Performance score of autoencoder AE-1 for known and zero-day attacks offline.
Information 16 00222 g012
Figure 13. (a) Online-training error. Error increases once anomalies are encountered in the system and decreases with successive learning cycles. (b) An excerpt from the incoming normal and malicious packets and the predicted normal and malicious packets.
Figure 13. (a) Online-training error. Error increases once anomalies are encountered in the system and decreases with successive learning cycles. (b) An excerpt from the incoming normal and malicious packets and the predicted normal and malicious packets.
Information 16 00222 g013
Figure 14. (a) Threshold optimization during the online-training process. (b) Anomaly detection and incremental learning on malicious packets in real time.
Figure 14. (a) Threshold optimization during the online-training process. (b) Anomaly detection and incremental learning on malicious packets in real time.
Information 16 00222 g014
Figure 15. (a) Anomaly-detection accuracy during online training. Accuracy is computed after each learning cycle. (b) Accuracy variations utilizing ED circuits compared to the floating-point threshold computing.
Figure 15. (a) Anomaly-detection accuracy during online training. Accuracy is computed after each learning cycle. (b) Accuracy variations utilizing ED circuits compared to the floating-point threshold computing.
Information 16 00222 g015
Figure 16. Experiment of online learning with noise. The noise level is scaled relative to the minimum conductivity of the memristor devices.
Figure 16. Experiment of online learning with noise. The noise level is scaled relative to the minimum conductivity of the memristor devices.
Information 16 00222 g016
Figure 17. (a) Breakdown of (a) area occupancy and (b) energy of various circuit elements in memristor-based anomaly-detection and online-learning systems.
Figure 17. (a) Breakdown of (a) area occupancy and (b) energy of various circuit elements in memristor-based anomaly-detection and online-learning systems.
Information 16 00222 g017
Table 1. Parameters for threshold-computing circuit simulation.
Table 1. Parameters for threshold-computing circuit simulation.
ParameterMagnitudeUnit
R1000Ohm
R11.50 × 106Ohm
M11.99/0.65W/L
M21.99/0.65W/L
M30.2/0.16W/L
M40.24/0.18W/L
M51.28/0.65W/L
M61.85/0.65W/L
M71.85/0.65W/L
Vbias0.4volt
Ibias210nA
RH10,000Ohm
Table 2. Parameters for memristor circuit simulation.
Table 2. Parameters for memristor circuit simulation.
ParameterMagnitudeUnit
Cycle time2 × 10−9s
Max training pulse duration5 × 10−9s
Roff10
Ron50
Ravg5.025
Wire resistance5Ω
Vmem1.3V
Pmem0.336µW
Op-amp power3µW
Max read voltage1.3V
Feature size, F45nm
Transistor size50F2
Memristor area10,000nm2
Cycle time2ns
Crossbar processing time50ns
Table 3. Hardware performance analysis.
Table 3. Hardware performance analysis.
ParametersTinker BoardMemristor System
Test sample20002000
Time (sec)1.8075632.3 × 10−3
Time/sample9.04 × 10−41.17 × 10−6
Speedup1774
Energy (joule)4.52 × 10−32.08 × 10−7
Power (W)50.0205
Performance (OPS)20 × 10616.1 × 109
Energy efficiency (OPS/W)4.12 × 1067.83 × 1011
Area (mm2) ---4.43 × 10−3
Test sample20002000
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alam, M.S.; Yakopcic, C.; Hasan, R.; Taha, T.M. Memristor-Based Neuromorphic System for Unsupervised Online Learning and Network Anomaly Detection on Edge Devices. Information 2025, 16, 222. https://doi.org/10.3390/info16030222

AMA Style

Alam MS, Yakopcic C, Hasan R, Taha TM. Memristor-Based Neuromorphic System for Unsupervised Online Learning and Network Anomaly Detection on Edge Devices. Information. 2025; 16(3):222. https://doi.org/10.3390/info16030222

Chicago/Turabian Style

Alam, Md Shahanur, Chris Yakopcic, Raqibul Hasan, and Tarek M. Taha. 2025. "Memristor-Based Neuromorphic System for Unsupervised Online Learning and Network Anomaly Detection on Edge Devices" Information 16, no. 3: 222. https://doi.org/10.3390/info16030222

APA Style

Alam, M. S., Yakopcic, C., Hasan, R., & Taha, T. M. (2025). Memristor-Based Neuromorphic System for Unsupervised Online Learning and Network Anomaly Detection on Edge Devices. Information, 16(3), 222. https://doi.org/10.3390/info16030222

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop