Next Article in Journal
Deep Learning Approaches for Quantifying Ventilation Defects in Hyperpolarized Gas Magnetic Resonance Imaging of the Lung: A Review
Next Article in Special Issue
Carpal Tunnel Syndrome Automated Diagnosis: A Motor vs. Sensory Nerve Conduction-Based Approach
Previous Article in Journal
Endoscope Capsules: The Present Situation and Future Outlooks
Previous Article in Special Issue
Interobserver Agreement in Automatic Segmentation Annotation of Prostate Magnetic Resonance Imaging
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Utilizing Deep Learning Algorithms for Signal Processing in Electrochemical Biosensors: From Data Augmentation to Detection and Quantification of Chemicals of Interest

by
Fatemeh Esmaeili
1,2,
Erica Cassie
2,3,
Hong Phan T. Nguyen
2,3,
Natalie O. V. Plank
2,3,
Charles P. Unsworth
1,2 and
Alan Wang
4,5,6,*
1
Department of Engineering Science, University of Auckland, Auckland 1010, New Zealand
2
The MacDiarmid Institute for Advanced Materials and Nanotechnology, Victoria University of Wellington, Wellington 6021, New Zealand
3
School of Chemical and Physical Sciences, Victoria University of Wellington, Wellington 6021, New Zealand
4
Auckland Bioengineering Institute, University of Auckland, Auckland 1010, New Zealand
5
Center for Medical Imaging, Faculty of Medical and Health Sciences, University of Auckland, Auckland 1010, New Zealand
6
Centre for Brain Research, University of Auckland, Auckland 1010, New Zealand
*
Author to whom correspondence should be addressed.
Bioengineering 2023, 10(12), 1348; https://doi.org/10.3390/bioengineering10121348
Submission received: 12 October 2023 / Revised: 14 November 2023 / Accepted: 21 November 2023 / Published: 23 November 2023
(This article belongs to the Special Issue Artificial Intelligence-Based Diagnostics and Biomedical Analytics)

Abstract

:
Nanomaterial-based aptasensors serve as useful instruments for detecting small biological entities. This work utilizes data gathered from three electrochemical aptamer-based sensors varying in receptors, analytes of interest, and lengths of signals. Our ultimate objective was the automatic detection and quantification of target analytes from a segment of the signal recorded by these sensors. Initially, we proposed a data augmentation method using conditional variational autoencoders to address data scarcity. Secondly, we employed recurrent-based networks for signal extrapolation, ensuring uniform signal lengths. In the third step, we developed seven deep learning classification models (GRU, unidirectional LSTM (ULSTM), bidirectional LSTM (BLSTM), ConvGRU, ConvULSTM, ConvBLSTM, and CNN) to identify and quantify specific analyte concentrations for six distinct classes, ranging from the absence of analyte to 10 μM. Finally, the second classification model was created to distinguish between abnormal and normal data segments, detect the presence or absence of analytes in the sample, and, if detected, identify the specific analyte and quantify its concentration. Evaluating the time series forecasting showed that the GRU-based network outperformed two other ULSTM and BLSTM networks. Regarding classification models, it turned out signal extrapolation was not effective in improving the classification performance. Comparing the role of the network architectures in classification performance, the result showed that hybrid networks, including both convolutional and recurrent layers and CNN networks, achieved 82% to 99% accuracy across all three datasets. Utilizing short-term Fourier transform (STFT) as the preprocessing technique improved the performance of all datasets with accuracies from 84% to 99%. These findings underscore the effectiveness of suitable data preprocessing methods in enhancing neural network performance, enabling automatic analyte identification and quantification from electrochemical aptasensor signals.

1. Introduction

Deep learning algorithms have gained attention in the context of sensor development and application. These algorithms have been effective and beneficial for addressing problems such as noise reduction, classification, object detection, anomaly detection, and process monitoring [1,2,3].
There has been a wide range of deep learning models for classification, such as convolutional neural network (CNN), generative adversarial network (GAN)-based architectures [4], and recurrent-based neural networks [5,6]. For example, Zhang [7] proposed a CNN-based network to automatically identify and quantify heavy-metal ions, which obtained with an accuracy of 99.99%. Moreover, Li [8] proposed a hybrid network, utilizing convolutional and long short-term memory layers in the network architectures, for EEG signal classification to find Parkinson’s disease. The classification accuracy for this model was 98.6%.
Moreover, deep learning-based generative models, such as variational autoencoders (VAEs), generative adversarial networks (GANs), and diffusion probabilistic models, have found extensive use in data augmentation [9]. Specifically, variational autoencoders (VAEs) and their extension, conditional variational autoencoders (CVAE), have been applied in research for anomaly detection, generating new sample data, and reconstructing input data by learning the fundamental features and structure of the training data. For instance, Liu [10] utilized VAE and CVAE structures for data augmentation in an intrusion detection system (IDS) to address data imbalances. The application of these networks significantly improved model performance metrics, specifically the Macro F1-score, by 5.32%. Moreover, CVAE proved to be effective in detecting structural anomalies in scanning transmission electron microscopy (STEM) images. By accurately reproducing input data and highlighting discrepancies in defect input data, CVAE was capable of automatically differentiating various point defect types [11].
Furthermore, deep learning networks have emerged as powerful tools for time series forecasting or predicting future data, using techniques such as CNNs, recurrent neural networks (RNNs), and temporal convolutional networks (TCNs) [12]. For example, Pirani [13] explored various RNN-based architectures for financial time series forecasting, demonstrating that models incorporating gated recurrent unit (GRU) layers outperformed other recurrent networks. Likewise, Mahjob [14] used quite similar architectures for energy consumption prediction, and the results showed that the network containing long short-term memory (LSTM) layer in its architectures outperformed other RNN-based models.
The motivation for this paper was related to our previous work [15,16]. In the former, we used LSTM-based networks for classification. We aimed to enhance the robustness, performance and speed of the automatic detection and quantification of analyte concentrations registered by sensors. Also, the results of both studies raised an issue that the time for recording the signals might have been effective in the outcomes of the deep learning models. Thus, we were inspired to employ techniques to forecast the future output of the sensors and examine our hypothesis about the signals’ lengths in the sensors’ signal analysis.
Thus, it can be briefly said that, in this study, we applied deep learning algorithms for signal processing obtained from three comparable CNT FET biosensors. Our primary objective was to automatically identify and quantify specific analytes using segments of these recorded signals. However, achieving this classification objective posed challenges that needed to be addressed. Consequently, we employed three key deep learning-based signal analysis techniques to overcome these challenges and fulfil our research objectives. These techniques were categorized into three main steps: (1) data augmentation, (2) signal extrapolation, and (3) classification models. Additionally, all calculations and deep learning algorithms were executed using MATLAB R2022b.
Figure 1 illustrates the framework of the applied deep learning methods in this research. Initially, essential data preparation steps were applied to the available datasets, including z-score scaling and STFT. Subsequently, a CVAE-based data augmentation technique was used to handle the limited data availability. In the following step, a signal extrapolation method was employed, utilizing recurrent neural networks (RNN) to generate uniform signal lengths for all the datasets.
Finally, we designed two classification models, C1 and C2, incorporating recurrent and convolutional layers. The C1 model aimed to identify and measure precise analyte levels within six distinct categories, ranging from no analyte presence to 10 μM. Each dataset’s segments were analyzed separately in this model. The C2 classification model, on the other hand, was developed to differentiate abnormal data from normal segments, identify the absence or presence of analytes in the sample, and, if present, recognize the specific substance and measure its concentration.

2. Background of Deep Learning Models

2.1. Recurrent Neural Network

Recurrent Neural Networks (RNNs) [17,18] are a specialized type of feed-forward neural network designed for processing sequential data. An RNN consists of hidden recurrent units, where the output at a specific time step is calculated based on the output from the previous time step.
In Figure 2, the movement of information within an RNN layer across three consecutive time steps is illustrated. Here, X R N , h R L , and Y R M represent the input to the RNN layer, the output of the RNN layer, and the subsequent layer after the RNN, respectively. A recurrent block comprises multiple recurrent units within the RNN layer. In this depiction, the number of hidden recurrent units, denoted as L, is equivalent to the dimension of h.
In the following, we will delve into the RNN structures utilized in this study. We will begin by exploring the basic structure of an RNN, followed by its two main extensions: GRU and LSTM.

2.1.1. Conventional Recurrent Neural Network

The vanilla RNN, also known as the conventional RNN, represents the most basic form of recurrent neural networks, as shown in Figure 3. It undergoes training utilizing the backpropagation through time technique [17]. In this model, an input sequence X t R N is fed into a recurrent unit at time step t. The output h t of the unit at the time step t is calculated using Equation (1):
h t = 0 t = 0 , ϕ ( W x X t + W h h t 1 + b ) o t h e r w i s e ,
where ϕ represents an activation function, W x , W h , and b indicate weight matrices and the bias vector, respectively.
Vanilla RNNs, while powerful dynamic systems, face challenges in training due to the vanishing gradient problem during backpropagation. This problem arises because the backpropagated gradients tend to either shrink or grow at each time step, leading to vanishing or exploding gradients over many time steps. Consequently, vanilla RNNs struggle to capture long-term dependencies. To address this issue, two main extensions of RNNs were developed: (i) LSTM, and (ii) GRU networks. These extensions integrate gating mechanisms within their recurrent units, resolving the issue of vanishing gradients and facilitating the capture of long-term dependencies.

2.1.2. Long Short-Term Memory

Long Short-Term Memory or LSTM [19], an advanced extension of RNNs, excels in learning temporal dependencies within sequential input during the training phase due to the unique structure of its recurrent units, known as LSTM cells. These cells incorporate three key components: the input gate, forget gate, and output gate, as depicted in Figure 4. These gates play a vital role in regulating the transmission of information entering and exiting the unit, enabling the LSTM network to capture temporal dependency in sequential data.
The input gate, represented as i t , and the cell candidate, denoted as c ˜ t , are in charge of modifying and managing the memory cell state, c t . The forget gate, denoted as f t , decides which details from the previous time step t 1 stored in the cell state c t 1 should be disregarded. The output gate, o t , selects the portion of the cell state that contributes to the output, h t . Equations (2)–(6) governing these processes at time step t are as follows:
i t = σ ( W i h h t 1 + W i x X t + b i ) ,
f t = σ ( W f h h t 1 + W f x X t + b f ) .
c ˜ t = t a n h ( W c h h t 1 + W c x X t + b c ) ,
c t = i t × c ˜ t + f t × c t 1 ,
o t = σ ( W o h h t 1 + W o x X t + b o ) ,
h t = o t × t a n h ( c t ) ,
where W i x , W f x , W c x , and W o x denote the input gate’s weight matrices for input, forget, cell value, and output, respectively. Similarly, W i h , W f h , W c h , and W o h represent the recurrent gate weights for input, forget, cell value, and output, respectively. In the same sequence, the corresponding bias vectors are denoted as b i , b f , b c , and b o .
In an LSTM layer, multiple LSTM units are recurrently connected. As shown in Figure 5, there are two types of LSTM layers: unidirectional LSTM (ULSTM) and bidirectional LSTM (BLSTM). These categories are based on the information flow within the layer. ULSTM processes information in one direction, moving forward in time, as illustrated in Figure 5a. Conversely, BLSTM, depicted in Figure 5b, consists of both forward and backward layers. The forward layer operates in the forward time direction, whereas the backward layer handles information processing in the reverse time direction.

2.1.3. Gated Recurrent Unit

A Gated Recurrent Unit or GRU network is an extension of RNNs and a simplified variant of LSTM, lacking an output gate [20,21,22]. In comparison to LSTM, GRU is more straightforward and faster during training. Unlike LSTM, GRU does not utilize a memory cell state to retain information. Instead, it only controls information within the unit. Unlike LSTMs, GRUs fully expose the memory in each state. The GRU architecture comprises three key components: the update gate, the reset gate, and a candidate state, as illustrated in Figure 6.
The update gate, z t , determines the proportion of the previous memory that should be retained during the training process, while the reset gate, r t , decides how much of the new input and the previous memory to combine. The following equations mathematically depict the calculation of the output h t at time step t.
z t = σ ( W z x X t + W z h h t 1 + b z ) ,
r t = σ ( W r x X t + W r h h t 1 + b r ) ,
h ˜ t = t a n h ( W h ˜ x X t + W h ˜ r ( r t h t 1 ) + b h ˜ ) ,
h t = ( 1 z t ) h t 1 + z t h ˜ t ,
where X t is the input vector, z t is the update gate, r t is the reset gate, h ˜ t is the candidate state, and σ is the sigmoid activation function. W and b refer to the weights matrices and bias vectors, respectively.

2.2. Convolutional Neural Network

Convolutional neural networks (CNN) are designed to analyze data presented in 1D, 2D, or 3D arrays, such as signals, images, or videos. CNNs possess four essential characteristics: local connections, where neurons focus on specific regions; shared weights, enabling parameter efficiency; pooling, reducing spatial dimensions; and the incorporation of multiple layers, allowing the network to extract complicated features from the data. These characteristics optimize CNNs for processing array data effectively [23]. As shown in Figure 7, the typical composition of a CNN includes fundamental layers like convolutional, pooling, and fully connected layers.
Convolutional layers process multiple input feature maps using filters to generate output feature maps. This operation involves sliding the filter across the entire input array. Each neuron calculates a weighted sum in a convolutional layer based on a subset of the prior layer’s outputs [24]. Consider X i , j as the prior layer’s output, where i and j denote the specific position in the two-dimensional grid. The weights assigned to an individual node within the present layer are denoted as w k , l , with k and l indicating the respective weight indices. As represented in Equation  (12), the calculated 2D convolution or the sum of weights, can be written as follows:
Y i , j = k l w k , l · X i , j .
It needs to be mentioned that the output size of a convolution layer, l n , by the padding p, stride s, is calculated according to Equation (13).
l n = [ l n 1 + 2 p n f s + 1 ] ,
where l n 1 and n f indicate the previous layer size and number of filters, respectively.
Following convolutional layers, the output neurons are fed into a non-linear activation function, such as the Rectified Linear Unit (ReLU) or Leaky ReLU, illustrated in Equations (14) and (15), respectively.
f ( x ) = m a x ( 0 , x ) .
f ( x ) = x x 0 , a · x x < 0 ,
where a, the slope coefficient or threshold, is a small positive constant that multiplies the negative input values, ensuring a small gradient for negative inputs, allowing some information to flow even for negative inputs.

2.3. Variational Autoencoder

A variational autoencoder (VAE) is a generative learning model developed from the Bayesian framework, as introduced in [25]. A VAE comprises a recognition model, also known as a variational approximator, denoted as q ϕ ( z | x ) and a generative model p θ ( z ) p θ ( x | z ) , where x and z represent input data and latent space, respectively. The architecture of a VAE network is depicted in Figure 8.
The objective of the VAE is to jointly learn the parameters of variational approximator ϕ and the generative model θ . This is achieved by maximizing the evidence lower bound (ELBO), defined in Equation (16):
L vae θ , ϕ ; x = D K L ( q ϕ ( z | x ) | | p θ ( z ) ) + E q ϕ ( z | x ) [ log p θ ( x | z ) ] ,
where D K L denotes the Kullback–Leibler (KL) divergence between the approximated posterior q ϕ ( z | x ) and the prior of the latent variable p θ ( z ) . The first term on the right-hand side acts as a regularizer, encouraging minimising the KL divergence between q ϕ ( z | x ) and p θ ( z ) . The second term represents an expected negative reconstruction error, maximizing the log-likelihood l o g p θ x | z with samples drawn from the approximated posterior.
To efficiently estimate this optimization, the stochastic gradient variational Bayes (SGVB) method, employing the autoencoding variational Bayes (AEVB) algorithm and a reparameterization trick [25], is used. The reparameterization trick involves reparameterizing the approximated posterior using a differentiable transformation g ϕ ϵ , x , where ϵ is a random noise variable acting as an auxiliary variable with an independent marginal p ( ϵ ) . This trick overcomes the challenge of calculating the gradient of L vae with respect to ϕ .
In practical applications, the choice of probability distributions is pivotal. Typically, a Gaussian distribution N ( μ z , Σ z ) is selected for the posterior, where μ and Σ are arbitrary deterministic functions [26]. For the prior, a standard normal distribution N ( 0 , 1 ) is commonly chosen. Regarding the likelihood distribution, a multivariate Gaussian distribution is used for continuous data, while a Bernoulli distribution is suitable for binary data [27].

2.4. Conditional Variational Autoencoder

The conditional variational autoencoder (CVAE) extends the capabilities of the VAE by incorporating external information, termed conditional data, during the generative process. Introduced in [28], the CVAE addresses a limitation inherent in VAEs. Traditional VAEs can generate data solely from latent variables, representing a specific class of sample data. In contrast, as depicted in Figure 9, a CVAE can create data using both latent variables and additional conditional input, allowing it to generate sample data for a specific class based on corresponding class labels [10]. While VAEs are primarily considered an unsupervised training framework, CVAEs operate in a semisupervised manner as their inputs include both sample data and class labels.
Equation (17) represents the objective function of the CVAE, which is similar to that of the VAE but includes additional data class label information:
L cvae θ , ϕ ; x = D K L ( q ϕ ( z | x , y ) | | p θ ( z ) ) + E q ϕ ( z | x ) [ log p θ ( x | z , y ) ] ,
where y refers to the class label.

3. Materials

This section gives a summary of the datasets used in this study, the sensing protocols for their corresponding signal registration, along with a detailed explanation of the data preparation and preprocessing techniques applied. A comprehensive understanding of the datasets, their preparation, and preprocessing is crucial for gaining insights into the subsequent classification models.

3.1. Datasets Description

A dataset of sensors, represented as D = { X 1 , , X s , , X S } , consists of a collection of signals denoted as X s , which are registered from identical sensors. Here, the subscript s is the index of each signal, while S signifies the total count of signals within this dataset.
In this study, we used and analyzed three distinct datasets: (1) 35-mer adenosine, (2) 31-mer oestradiol, and (3) 35-mer oestradiol. These datasets include multiple single-variable time-series signals that record the drain current from three distinct sensors. Specifically, the sensors were based on technology utilizing aptamer-based electrochemical carbon nanotube field-effect transistors (CNT-FETs).
Table 1 presents the available datasets and their sizes, including the total number of signals in each dataset. In order to enhance comprehension of the registered signals, a brief introduction to the sensors and the sensing protocols used to measure drain current is provided below, outlining the components of the sensors and the methodology used for signal measurement.
Table 2 represents three fundamental components of the sensors employed in collecting data to concisely compare the datasets. The datasets were primarily differentiated by two key elements of their sensors: (i) their target analytes and (ii) their aptamers employed as the bioreceptors. The other key component, which is the transducer, was CNT FET for all three sensors. Providing a comprehensive explanation of the sensor functionalization details is outside the scope of this paper. Nevertheless, comprehensive details regarding the 35-mer adenosine sensor, including specifics about receptor functionalization and transistor fabrication, are available, in [29].
In terms of sensing protocols, the procedures for the 31-mer and 35-mer oestradiol sensors shared similarities, but they differed from those employed for the adenosine sensors. Table 3 compares and summarizes the sensing procedures used in the datasets.
The signals for oestradiol and adenosine were registered at specific time intervals, with oestradiol measured every 1.081 with a standard deviation of 5 × 10 3 and adenosine every 1 s, respectively. The drain and gate voltages were held consistently at V D = 100 mV and V G = 0 V for both.
Initially, the polydimethylsiloxane (PDMS) wells of the sensors were filled with designated solutions. The adenosine sensor employed a 2 mM Tris-HCl buffer, while the oestradiol sensors utilized a solution consisting of 0.05 times PBS (phosphate-buffered saline) with 5% ethanol (EtOH).
In the case of the adenosine aptasensor, the process commenced by filling the PDMS well with 110 μM of 2 mM Tris-HCl, which was maintained for 1000 s. Following this, adenosine solution was introduced into the PDMS well every 500 s, with each injection containing progressively higher concentrations, taking into account the total adenosine concentration in the well before each addition. Through this incremental approach, the adenosine concentration in the well was raised, ranging from 100 pM to 10 μM.
Considering the oestradiol aptasensors, the procedure started by introducing an initial volume of 100 µL of 0.05 times PBS with 5% EtOH into the PDMS well, which was maintained for 300 s. During the subsequent 300 s, an additional 20 μL of 0.05 times PBS with 5% EtOH was added without altering the oestradiol concentration. Afterward, oestradiol solution was injected into the well every 300 s, taking into account the existing oestradiol concentration before each addition. Furthermore, with each increment in the oestradiol concentration, 20 µL of 0.05 times PBS with 5% EtOH was introduced into the well. This step-by-step process elevated the oestradiol concentration within the well, ranging from 1 nM to 10 μM.
Figure 10 illustrates the stereotypical original signals from the mentioned datasets. Vertical lines are employed to distinguish between various analyte concentrations (ACs) within a signal. It should be noted that the initial ACs for the 35-mer adenosine experiments may not have been the same; the initial AC in Figure 10a was 1 nM and it was 1 μM for Figure 10b. Conversely, the initial ACs for experiments involving the 31-mer and 35-mer oestradiol sensors were entirely the same, as depicted in Figure 10c and Figure 10d, respectively.
Further clarity is needed regarding the terms entire signal and segment, as they are frequently used throughout this paper. In this context, the term entire signal points out all the registered sensing responses from the start to the end of an experiment. Conversely, a segment denotes a specific portion of the entire signal that illustrates the sensor’s registered data to a particular concentration of the target analyte. For example, in Figure 10a, the data at t [ 1 , 2000 ] are considered as the entire signal, and this signal consists of three distinct segments: t [ 1 , 1000 ] for the No Analyte segment, t [ 1001 , 1500 ] for the 1 μM segment, and t [ 1501 , 2000 ] for the 10 μM segment.

3.2. Data Preparation

There were three steps in order to prepare each dataset for further analysis in this study. These steps include data normalization, segmentation and segment labelling.

3.2.1. Data Normalization

The purpose of feature scaling or data normalization was to standardize the entire signals within a specific dataset to a uniform scale. This process also aimed to prevent certain features from controlling and dominating others. We utilized Z-score scaling for normalization, employing the mean (μ) and standard deviation ( σ ) calculated from the entire signal.
To demonstrate this concept, take an entire raw signal X = [ x 1 , , x i , , x n ] of length n. Equation (18) outlines the normalization process, producing the normalized signal denoted as X N o r m , following the method described.
X N o r m = [ x 1 N o r m , , x i N o r m = x i μ σ , , x n N o r m ] .
Figure 11 displays the effect of Z-score scaling. The impact of the normalization process becomes evident when you compare Figure 11a,b. Figure 11a represents an original signal sample from DS 1, while Figure 11b shows the same signal after undergoing normalization using Equation (18).
It is important to emphasize that the entire signal underwent Z-score scaling. The primary purpose behind normalizing the signals in this manner was to ensure that all signals within each dataset were brought to a consistent scale. Additionally, it should be noted that, for the sake of simplicity, we will subsequently refer to the normalized signals as X throughout this paper instead of X N o r m .

3.2.2. Segmentation

After rescaling the signal, each signal was divided into its component segments. In this study, a segment is defined as a part of a signal with a consistent analyte concentration from beginning to end.
Table 4 displays the total count of segments corresponding to various analyte concentrations in each dataset. It is important to note that the oestradiol signal’s analyte concentration ranged from 1 nM to 10 μM. Consequently, the total count of signals and segments in each class across these datasets remained consistent. On the other hand, the recorded analyte concentration ranges differed among the adenosine signals. Moreover, due to an insufficient number of segments in the 100 pM label in DS 1, this particular class had to be omitted.
Regarding the oestradiol datasets, in the next step, we used the retime function in MATLAB R2022b for the resampling to increase the data points from 166 data points in each segment to 300 data points. The reason for resampling was to increase the data points that were essential for the signal extrapolation part.

3.2.3. Anomaly Detection

Anomaly detection is the task of detecting data samples that exhibit statistically distinctive characteristics compared to the other available data instances, regarded as normal instances [32]. Identifying abnormal data is a vital task for machine learning or deep learning scenarios, as abnormal observations hold misinformation and incorrect details and deteriorate the performance of predictive models.
With respect to the signals studied in this research, a number of factors influenced the signals and led to the production of anomalous data. These contributing factors could include complications in the receptor’s immobilization on the surface of CNT, challenges in constructing the sensing interface, transistor malfunctions, difficulties and complications with the receptor on CNT surfaces, laboratory background noise, and so on.
Figure 12 presents a comparison between normal patterns in DS 1 and its abnormal time series. Figure 12a displays the normal pattern. Figure 12b represents two instances of entire signals considered anomalies. The blue and red signals correspond to recordings from non-sensing and broken transistor sensors, respectively. Figure 12c exhibits two signals with abnormal behaviour during a specific time interval, specifically between 300 s and 750 s. In these observations, segments displaying abnormal behaviour were categorized as anomalies, while the others as normal segments. A comprehensive explanation of anomaly description can be found in [16].
The data collector classified the existing segments as standard, borderline, non-sensing signals, and malfunctioning transistors, relying on prior understanding of the sensors’ capabilities and visual data analysis. In our initial study [15], we considered marginal, no-sensing signals and broken transistors as abnormal data or contextual outliers. However, in our subsequent work [16], we introduced an autoencoder-based anomaly detection method as an additional criterion. Consequently, abnormal data in this study were identified using the results of [16] and data visualization. Table 5 displays the total number of normal segments corresponding to each analyte concentration within each dataset after the application of anomaly detection.

3.2.4. Segment Labeling

In this study, as depicted in Figure 1, we developed two distinct classification models. Given the differences between these models and the importance of understanding data classes and labels, we provide an explanation of the segment labelling approach for both models in this section. For a visual overview of the labelling, refer to Figure 10.
Table 6 provides a quick view of the labels in both models C1 and C2. In classification model C1, the objective was to detect and quantify the specific analyte concentration within six different classes, ranging from No Analyte to 10 µM. These models were developed individually for each dataset, excluding abnormal data. In contrast, classification model C2 aimed to distinguish between normal and abnormal segments. For normal segments, model C2 also detected the presence of target analytes and their respective quantities.
It is important to note that all three datasets contain both normal and abnormal data. In the C1 models, abnormal data were excluded, and only normal data were utilized for prediction models. Conversely, in the C2 model, both normal and abnormal data from all three datasets were combined. The approach for generating augmented abnormal data in the C2 model involved creating abnormal data specific to each dataset in the data augmentation process using their respective VAE. Subsequently, the augmented abnormal data from all three datasets were treated as a unified category and labeled as class 12 for the C2 classification model.

3.3. Data Preprocessing

The short-time Fourier transform (STFT), introduced by Stockwell [33], is an extension of the Fourier transform that is used for analyzing the time-varying frequency components of non-stationary signals by processing them in short time intervals [34]. This transformation has found application as a data preprocessing technique in signal processing for machine learning and deep learning models [35,36,37,38].
The STFT involves a sliding window to divide a signal into short time segments, often with overlap. Fourier transforms are computed for these short segments, and this process is applied iteratively to cover the entire signal using windowing. In continuous form, the STFT of a single-channel signal x ( t ) is expressed as Equation (19):
S T F T x ( t ) ( τ , ω ) = X ( τ , ω ) = x ( t ) ω ( t τ ) e j ω τ d t ,
where ω ( τ ) represents the analysis window function. It should be noted that in the discrete version of signal x(t), the integral sign in Equation (19) is replaced by a sum.
The parameters of the window function ω ( τ ) include the type of window, its length, and the degree of overlap. Various types of window functions are available, such as the rectangular window, Sine window, and Blackman window. Equation (20) describes the calculation of Blackman window:
ω [ n ] = a 0 a 1 c o s ( 2 π n N ) + a 2 c o s ( 4 π n N ) , a 0 = 1 α 2 , a 1 = 1 2 , a 2 = α 2 , α = 0.16 .
The squared magnitude of the STFT S is known as the spectrogram that represents the power spectral density of the function over the joint time-frequency plane, as shown in Equation (21):
S x ( t ) ( τ , ω ) = | X ( τ , ω ) | 2 .
In this work, we used the short-time Fourier transform (STFT) for data preprocessing with a Blackman window. The window parameters were set to a 128-sample Blackman window, with 64 samples of overlap between adjoining segments and a 128-point Fourier transform length.
Figure 13 illustrates the spectrogram of a typical and normal segment from the 35-mer adenosine dataset with an AC of 10 μM. Figure 13a corresponds to the first 300 s of the segment, while Figure 13b shows the spectrogram for the entire 500 s.
The resulting matrices for this transformation, applied to all datasets, were of size 128 × 3 for the initial 300 s and 128 × 6 for the full 500 s. These matrices were utilized as the primary input data for all classification models employed in our study. In our analytical assessment, we conducted a comparative evaluation to evaluate the influence of this transformation in contrast to the impact of employing data normalization as a standalone preprocessing technique.

3.4. Data Split

Traditionally, data provided for deep learning models is divided into three segments: the training set (60%), the validation set (20%), and the test set (20%) [39]. However, in this study, where we tackled three distinct deep learning tasks—data augmentation, signal extrapolation, and classification—we employed varied strategies for data partitioning.
During the data augmentation phase, 80% of the original datasets were allocated for training, and 20% for validation. The test sets comprised scaled augmented data from [15]. For the signal extrapolation and classification models, we used a split ratio of 60% for training, 20% for validation, and 20% for testing. The data utilized in these stages were the CVAE-augmented data.

4. Methods

In this section, we provide a comprehensive overview of the methodologies implemented to tackle the challenges and fulfil the research objectives of our study. Our approach includes the development of three core deep learning models: (1) data augmentation, (2) signal extrapolation, and (3) classification. These models were developed to address the specific requirements of our research tasks. Moreover, it should be noted that the entire implementation process for these deep learning algorithms was conducted using the MATLAB R2022b Deep Learning Toolbox. Moreover, it should be noted that the design of the networks in this study was based on extensive experimentation and evaluation.

4.1. Data Augmentation

The exceptional performance of deep learning predictive models relies heavily on the size and consistency of the training datasets to prevent overfitting. However, real-world datasets often suffer from scarcity and class imbalance issues. To effectively increase dataset size and enhance data quality, data augmentation plays a crucial role in the successful application of deep learning models, including time series data [40]. Specifically, generative models like VAE, CVAE, and GAN have been employed for augmenting time series datasets. These models are directly applied to capture the data’s underlying probability distribution, enabling the generation of new samples that closely imitate the original data distribution [41,42].
In this study, we employed data augmentation using CVAE networks for normal data and VAE for anomaly data to address limited dataset sizes, enabling the development of deep learning classification models with generalization ability and reduced risk of overfitting. Figure 14 illustrates the architecture of encoder and decoder networks utilized for generating normal data across all datasets.
As illustrated in Figure 14, both encoder and decoder networks shared the same architecture with minor variations in layers 1, 9, 12, 22, and 23. In the first layer, the encoder received signal data, while the decoder received random noise. Layers 9 and 12 were convolutional and deconvolutional layers for the encoder and decoder, respectively. In layer 22, the encoder employed a fully connected layer, while the decoder used a deconvolution layer. The last layer differed, with the encoder employing a sampling layer and the decoder using a scaling layer to replicate the original signal.
Figure 15 illustrates the architecture of encoder and decoder networks utilized for generating abnormal data across all datasets. Notably, AC labels for anomaly data were impractical and lacked meaningful results. Consequently, layers related to label information were omitted from the CVAE, and the corresponding networks were modified to VAE networks. As depicted in this figure, layers 1, 10, 15, and 16 exhibited differences in the encoder and decoder, which was similar to layers differentiation in the CVAE structure.
In the context of the sampling layer, let X = [ x 1 , x 2 , , x m ] T denote the input data to this layer, where m represents the length of the vector. This vector was a concatenation of the mean vector μ z and the variance vector Σ z , namely, μ z = [ x 1 , , x m / 2 ] T and Σ z = [ x ( m / 2 ) + 1 , , x m ] T . The sampling layer produced three outputs: μ z , σ z , and Z. Equations (22) and (23) describe the calculations for these outputs:
σ z = e 1 2 Σ z ,
Z = ϵ · σ z + μ z ,
where ϵ was a vector with normally distributed random numbers generated with randn function in MATLAB R2022b.
Regarding the scaling layer, Equation (24) represents the function utilized in this layer to scale and map the input vector X to the output data f s ( X ) while maintaining the same scale as the original registered data:
f s ( X ) = α · t a n h ( X ) ,
where α R was the variable α that varied according to the dataset scale. For example, for normal data across all datasets, we set α to 1.5, which aligns with the normalized drain current segments for normal data, lying within the range of [ 1.5 , 1.5 ] . Conversely, when generating anomalies in the 35-mer adenosine dataset, α is set to 5, reflecting the approximate range of abnormal data within [ 5 , 5 ] .

4.2. Signal Extrapolation

Time series forecasting using deep learning has gained prominence in academic research across diverse domains [43,44]. In the sensors industry, deep learning-based time series modelling has found application in tasks such as denoising, dimensionality reduction, anomaly detection, structural damage identification, and predicting future sensor outputs [45]. Various deep learning algorithms have been employed for forecasting time series data, including CNN, temporal convolutional network TCN, LSTM, and GRU [43,44,45].
In this section, we employed recurrent-based networks for forecasting future sensor outputs in the oestradiol datasets, a task we simply refer to as signal extrapolation in this study. Our motivation for this approach stemmed from the results of our previous work [15,16], where we raised concerns about the impact of signal length on classification and anomaly detection outcomes. In both of those studies, deep learning models performed better when using the 35-mer adenosine dataset with segments lasting 500 s, as opposed to the 31-mer and 35-mer oestradiol datasets with segments lasting 300 s. As a result, we applied recurrent-based networks to extrapolate the oestradiol datasets from 300 s to 500 s.
Figure 16 visualizes the network architecture used for signal extrapolation. These networks comprised four successive layers: a sequential input layer, a recurrent layer, a fully connected layer, and a regression layer. The primary distinction among these networks lay in their recurrent layer, which incorporated either a GRU, unidirectional LSTM, or bidirectional LSTM layer. For a more detailed comparison of the network layers and parameters, please refer to Table 7.
Remember that the input layer’s size fed into the networks matched the length of the segments and was treated as a single sequence. Similarly, the output size of the networks was a sequence with a length identical to that of the input segment. Also, the n refers to the number of hidden units in the recurrent layers. In this section, the number of hidden units, n, was set to 128. Note that the input weights, along with the recurrent weight and bias matrices, were combined to create the input weights for both GRU and LSTMs ( W x = [ W r x ; W z x ; W h ˜ x ] , W x = [ W i x ; W f x ; W c x ; W o x ] ), recurrent weights in the same order ( W h = [ W r h ; W z h ; W h ˜ h ] , W h = [ W i h ; W f h ; W c h ; W o h ] ), and bias in the same order ( b = [ b r ; b z ; b h ˜ ] , b = [ b i ; b f ; b c ; b o ] ).
The pseudocode outlined in Algorithm 1 describes the technique used to extend the length of data segments to meet a specified desired length, set at 500 s for this study. With an input dataset denoted as D and an initial data segment, X, the algorithm proceeded by iteratively predicting future sensor readings at variable time intervals, controlled by the timeStep parameter, and appending these predictions to the existing data. This process continued until the desired data segment length was achieved. Key steps included window-based data preparation, neural network training, and iterative prediction updates.
Figure 17 illustrates the application of the algorithms described on a segment with AC of 10 μM from the 35-mer adenosine dataset. In the first iteration, as visualized in Figure 17a, the predictor window covered the time interval [ 40 , 300 ] , which was used to predict the subsequent time interval [ 55 , 315 ] . Consequently, the data from the interval [ 301 , 315 ] , as the extrapolated part, was appended to the actual segment.
During the second iteration, according to Figure 17b, the predictor and predicted segments were adjusted to [ 50 , 315 ] and [ 65 , 330 ] , respectively. As a result, the data from the interval [ 316 , 330 ] , as the extrapolated part, was appended to the actual segment.
Algorithm 1 Algorithm for forecasting future sensor outputs
Input: Dataset D, Segment X = [ x 1 , , x n ] , Desired Length L
Output: Extrapolated Segment X e = [ x 1 , , x n , x n + 1 , , x L ]
X e X
l Segment Length
startPoint Start Point
timeStep 15
stepOffset 0
while  l L  do
   for all  X D  do
          predictorStart startPoint + stepOffset
          predictorEnd l timeStep
          Predictor [ x predictorStart , , x predictorEnd ]
          targetStart predictorStart + timeStep + stepOffset
          targetEnd l
          Target [ x targetStart , , x targetEnd ]
   end for
    TrainNetwork ( Predictor , Target )
    PredictFutureDataPoints X p = [ x l + 1 , , x l + timeStep ]
    X e Concatenate ( X e , X p )
    l Length ( X e )
    stepOffset stepOffset + 10
end while

4.3. Classification Models Architectures

As explained before, in this study, we have developed two distinct classification models: C1 and C2 models. The primary objective of the C1 models is to detect and quantify analyte concentrations for each dataset individually. On the other hand, the C2 model is designed to identify abnormal and normal data, detect various analytes, and quantify their respective concentrations throughout all available datasets.
Seven distinct deep learning models were employed for the classification tasks: (1) GRU, (2) ULSTM, (3) BLSTM, (4) ConvGRU, (5) ConvULSTM, (6) ConvBLSTM, (7) CNN. The architectures of these networks remained consistent across both the C1 and C2 models. The only difference in the network architectures was in their classification output layers: the C1 networks had output layers with a size of six classes, whereas the C2 models’ output layers consisted of twelve classes.
Figure 18 and Table 8 visualize and describe the three employed RNN-based networks for both classification models and compare the networks with regard to their parameters in different layers. These networks consisted of 5 consecutive layers: (1) a sequential input layer, (2) a recurrent-based, (3) a fully connected layer, (4) a Softmax layer, and (5) a classification layer. The main differences among these networks were in their second layers, with the implementation of GRU, ULSTM, and BLSTM structures within the recurrent layer.
Note that n and m refer to the hidden node numbers in the recurrent layer and the class numbers in the classification models, respectively. In this part, the hidden node numbers, n, was set to 128, and m for C1 and C2 models were 6 and 12, respectively. Furthermore, the details of the input weights ( W x ), recurrent weights ( W h ), and biases (b) have been previously explained in this section.
Moreover, Figure 19 illustrates the three proposed networks for classification, while Table 9 provides detailed descriptions and relevant information about their architectures, collectively referred to as ConvRNNs. These networks, namely ConvGRU, ConvULSTM, and ConvBLSTM, differ from the previously mentioned RNN-based networks by incorporating a 2D convolutional layer after the input layer. Some structural modifications were made to refine and fine-tune the networks.
Note that the variable l s represents the length of the segments that are fed into the networks. In our earlier models, the input layers were configured as sequential input layers. However, in the case of the ConvRNNs, we opted for an image input layer. As a result, the input size for the sequential layers was set to one, while for the image input layer, it matched the length of the segments.
The variable l c , as the output size of the convolution layers, was calculated according to Equation (13). Additionally, s f and n f denote the size and number of filters in the 2D convolutional layer, respectively, with p representing padding and s indicating the stride size for this layer. In these classification models, the mentioned variables were set as follows: l s { 300 , 500 } , s f = 5 , n f = 32 , p = 1 , s = 1 , n = 128 , and m { 6 , 12 } .
It is important to note that the variable l s represents the input length of the segments fed into the networks, which can vary between 300 s and 500 s. This variation allowed us to assess the impact of segment length on the models’ performance.
Finally, Figure 20 provides an illustration of the proposed CNN architecture used for classification, while Table 10 offers a detailed description of its layers.
The variables l c i for i = 1 , , 6 , as the output size of the convolution layers, were calculated according to Equation (13), where filters’ size s f = 5 , padding p = 1 , and stride s = 1 . Also, the number of filters were set as follows: n f 1 = 32 , n f 2 = 16 , and n f 3 = 8 .

4.4. Model Performance Evaluation

There are two approaches for the evaluation of deep learning models in this study, which are prediction and classification.

4.4.1. Prediction Model Evaluation Metrics

Regarding the evaluation of the data augmentation and signal extrapolation methods, we reconstructed their corresponding test data using the training networks and then calculated the reconstruction error. This error quantifies the dissimilarity between the original data and the reconstructed output and can be measured using statistical metrics such as mean absolute error (MAE) [46] or mean square error (MSE) [47]. In this study, we employed MSE as the metric for calculating the reconstruction error.
Consider X as an input segment represented by [ x 1 , x 2 , , x N ] , and let X ^ represent its reconstructed output given by [ x ^ 1 , x ^ 2 , , x ^ N ] . Equation (25) defines MSE as the metric for measuring reconstruction errors:
MSE = 1 N i = 1 n ( x ^ i x i ) 2 .

4.4.2. Classification Model Evaluation Metrics

The classification performance of C1 and C2 models was assessed by the overall accuracy (ACC) and the Macro F1-score (MF1). To calculate these metrics, initially, a confusion matrix was created using the classification of the test data, and subsequently, the two mentioned metrics were calculated.
The overall accuracy, as described in Equation (26), represents the proportion of truly classified elements in the test data out of the total elements. It was calculated by adding the diagonal elements of the confusion matrix and dividing the sum by the total number of elements.
ACC = Number of correct preditions The overall count of items in the test set .
The Macro F1-score serves as a valuable metric for classification models dealing with multiple classes. To comprehend how the macro F1-score is computed, it is crucial to understand its constituent elements: recall, precision, and F1-score. Precision in Equation (27), also known as positive predicted value (PPV), represents the ratio of accurately classified positive instances out of all the instances predicted positive by the model. Recall in Equation (28), denoted as the true positive rate (TPR), signifies the proportion of accurately classified positive instances out of all actual positive instances. The F1-score, outlined in Equation (29), is calculated from precision and recall. Finally, the macro F1-score, described in Equation (30), calculates the average of class-wise F1-scores concerning the model.
Precision = PPV = TP TP + FP ,
Recall = TPR = TP TP + FN ,
F 1 - score = 2 1 / precision + 1 / recall ,
Macro F 1 - score = 1 m i = 1 m { F 1 - score } i ,
In these equations, m represents the total number of classes within a specific classification model.

5. Results

5.1. Data Augmentation

Table 11 presents detailed information about the architecture of the networks and the layers’ output sizes that were utilized for data augmentation. These networks are comprised of both encoder and decoder modules [10]. The CVAE networks were employed to augment normal data, primarily due to variations in AC labels across available datasets. In contrast, the VAE networks were used for generating anomaly data, as these signals lacked different labels, making the use of AC labels impractical. Additionally, since the sizes of the segments related to the 31-mer and 35-mer oestradiol datasets, referred to as DS 2 and DS 3 were similar, their network information was merged in one column. In contrast, the information related to 35-mer adenosine as DS 1 remained separate.
Table 12 shows two types of information regarding the data augmentation with CVAE for normal data on the three available datasets: (1) the performance evaluation of the networks and (2) examples of augmented normal segments. After training the networks as described, we initially evaluated their performance using segments generated with a scaling data augmentation method, as previously detailed in our work [15]. These generated segments served as the test data, since the original datasets lacked sufficient data for testing purposes. Therefore, we first reconstructed the test data using the trained CVAE networks. The examples of reconstructed normal segments are shown in the left column of Table 12, while the middle column presents the histogram of reconstruction errors calculated using mean square error (MSE) according to [48].
Subsequently, the decoder modules of the networks were employed to generate new data. For each AC, the corresponding label was input into the feature input layer, and the desired amount of data was augmented. The examples of augmented data are visualized in the right column of Table 12. The total size of both the original and augmented data for each AC was set to 200. Likewise, the anomaly data was generated with VAEs’ decoder modules.

5.2. Signal Extrapolation

In the signal extrapolation phase, we employed the 35-mer adenosine dataset for forecasting future sensor outputs. This dataset had segments with a length of 500 s, which was longer than the segments in both oestradiol datasets, each with a length of 300 s. This longer segment length provided a suitable basis for evaluating the performance of the recurrent-based prediction models in signal extrapolation. Specifically, we used the initial 300 s as the predictor and extrapolated the subsequent 200 s.
The data used here for training and then evaluating the models were the original and augmented data with CVAE, explained in the previous section. The MSE was used as the statistical metric to calculate the prediction error, and to assess and compare the performance of the prediction models. In this part, MSE calculated the difference between the original segments and their corresponding extrapolated parts at the final 200 s.
Figure 21 displays the results of the signal extrapolation process. Figure 21a–c show histograms of the prediction error assessed by MSE for each prediction model. Additionally, Figure 21d provides an example of extrapolated outputs using the GRU network.
Similar to [13], as shown in these histograms, we observed that the GRU network outperformed the other two networks. Furthermore, as indicated in Table 7, the GRU model had fewer learnable parameters compared to ULSTM and BLSTM, resulting in shorter training times for computation than the LSTM models. Thus, we implemented the signal extrapolation procedure on the oestradiol datasets with a GRU-based prediction model. Subsequently, Figure 22 represents examples of the predicted outputs from the oestradiol datasets.

5.3. Classification Models

In this section, we delve into the comprehensive performance evaluation of the C1 and C2 models on their respective test sets. Our study focused on three key aspects: data preprocessing methods, signal extension techniques, and the comparative analysis of various deep learning-based neural networks for classification.
Regarding the data preprocessing, we compared the impact of two approaches: normalized data with z-score scaling and normalized data with the STFT method. For signal extrapolation, we examined the effect of segment lengths, specifically 300 s and 500 s, on the classification outcomes. Finally, our study culminated in a thorough comparison across these four categories, employing the explained deep learning algorithms.
The recurrent layers in the algorithms were configured with 128 hidden units when applicable. According to our previous work [15], the number of hidden units did not greatly affect the networks’s performance. Thus, in this study, we just considered one number, 128. The training process comprised 100 epochs with a minibatch size of 128, utilizing the Adam optimization algorithm. The learning rate was set to 0.002. Moreover, to enhance the robustness of the prediction models and avoid the risk of overfitting, K-fold cross-validation with k = 10 was applied. All other hyperparameters were maintained at their default values in MATLAB R2022b.

5.3.1. C1 Models

Figure 23 illustrates the performance and impact of deep learning models on the 35-mer adenosine dataset. It is important to note that in C1 models, there were six distinct classes. This configuration implies that the primary objective of the classification model was to individually identify and quantify the analyte concentration for each dataset.
Figure 24 and Figure 25 showcase the performance and impact of deep learning models on the 31-mer and 35-mer oestradiol datasets, respectively.
The choice of the data processing method significantly influenced the classification outcomes. Utilizing the STFT method as the data preprocessing technique had a noticeable effect on the employed classification models. As a result, all the networks for the three datasets yielded comparable, consistent and high-performance results, ranging between 84% to 99% in both metrics, i.e., accuracy and macro F1-score.
Comparing the deep learning networks, it is evident that ConvGRU, ConvULSTM, ConvBLSTM, and CNN models demonstrated consistent and comparable performance, surpassing RNNs across all datasets.
In the context of segment length, it can be deduced that segment length played a crucial role in recurrent networks such as GRU, ULSTM, and BLSTM models just in the 35-mer adenosine dataset. In the case of the 35-mer adenosine dataset, using the lengthier segments resulted in a notable accuracy boost, reaching up to 8% in the ULSTM and BLSTM models and as high as 15% in the GRU model. However, it did not affect networks with convolution layers in their structures.
Considering signal extension for the oestradiol datasets, it is important to note that this method did not prove effective in our developed classification models. Nevertheless, its inclusion in this study does not support our previous hypothesis regarding the significance of segment length for these two datasets in predictive modelling [15,16].

5.3.2. C2 Model

Figure 26 showcases the performance and impact of deep learning models on the C2 classification model.
The outcomes of this model closely mirrored those of the C1 models. It is evident that employing STFT as the data preprocessing method significantly enhanced the performance of classification tasks across all networks. The accuracy rates for networks utilizing STFT ranged between approximately 95% and 98%, underlining its consistent effectiveness in improving classification outcomes.
In the context of signal extrapolation, it is crucial to note that this method did not demonstrate effectiveness in our developed classification. Specifically, the results for GRU and ULSTM models using extrapolated segments were consistently lower than those with original segments, showing a decrease of at least 10%.
Upon comparing the networks, it can be deduced that models with convolution layers consistently delivered stable results, outperforming RNNs across various signal lengths and data preprocessing techniques, showcasing high performance.

6. Discussion

Aptasensors utilizing nanomaterials are valuable biosensors as they can detect minute chemicals and species. A key objective in biosensor development is identifying and measuring trace amounts of specific analytes. Deep learning techniques have become highly appealing tools for advancing biosensors and their data analysis.
Deep learning generative algorithms, like GANs or VAEs, provide a valuable solution for data augmentation, addressing the limitations of original data collection. They offer a cost-effective and time-saving alternative for analysis, crucial in biosensor development [49]. Additionally, deep learning algorithms have been extensively utilized in time series forecasting, allowing the accurate prediction of future data points [12,50]. This forecasting approach is effective in tasks such as sensor signal extrapolation and predicting future outputs. Moreover, deep learning networks serve as powerful tools for various tasks, including identification, quantification, and classification models [51,52]. These methods utilize advanced neural network structures to handle intricate data patterns, making them indispensable in both contemporary research and industrial applications.
In this work, we have successfully proposed and compared several classification models designed to automatically identify and measure specific analytes based on segments of signals captured by the aptasensors. However, addressing the challenge of inadequate dataset sizes, common in real-world scenarios, remains crucial to achieving our primary goals. Additionally, employing a signal extrapolation method was essential to ensure the uniformity and equality of all available datasets’ lengths. This step was necessary to test our earlier hypothesis, which was proposed in our previous papers and involved variations in segment lengths [15,16].
Regarding the data augmentation, we designed and employed CVAE and VAE networks to augment the normal and abnormal data, respectively. The data augmentation and network designs were applied separately for each dataset. The networks were trained by all available and original segments within their corresponding datasets. In the test phase, we used the data that were generated by the proposed scaling data augmentation in [15] as the test sets. Finally, MSE was used as the evaluation metric for calculating the reconstruction error. The potential contribution of the presented data augmentation method is that it can be reused for anomaly detection or outlier detection.
In the context of signal extrapolation, we designed and compared networks incorporating recurrent layers to generate segments with uniform lengths [53]. For training and testing the signal extrapolation algorithm, we utilized the 35-mer adenosine dataset, which featured 500-s data points and was longer than the two oestradiol datasets. This dataset allowed us to thoroughly assess the proposed algorithm. To evaluate prediction accuracy, MSE was employed, comparing real data with extrapolated data in the last 200 s. Among the three recurrent networks studied, the GRU-based network outperformed the other two LSTM-based networks [13]. Subsequently, this technique was applied to the oestradiol datasets to extend the segments from 300 s to 500 s. Furthermore, this proposed signal extrapolation can be regarded as an algorithm for time series forecasting for similar datasets.
We categorized our classification objectives into two models: C1 and C2. C1 models were tailored for individual datasets, focusing on six classes that spanned from zero analyte presence to 10 μM concentrations. Conversely, the C2 model considered all datasets, distinguishing 12 classes. These classifications included identifying normal and abnormal data, detecting analyte presence, and measuring its concentration when present.
We designed seven deep learning networks incorporating recurrent, convolutional, or hybrid layers to achieve our classification objectives. These networks were subjected to two preprocessing methods: z-score scaling and z-score combined with the STFT method. Furthermore, we examined the influence of segment length on the classification outcomes. Evaluation metrics included overall accuracy and macro F1-score, given the presence of multiple classes within each classification model.
The results indicated a significant improvement in model performance, particularly in recurrent-based networks, when utilizing the STFT method as a data preprocessing technique. However, the signal extrapolation method did not yield consistent effects on classification models. Networks combining both recurrent and convolutional layers consistently outperformed others. Interestingly, the performance of CNN networks was comparable to that of Conv-RNN networks [38].
In our analysis, we observed that the C1 models and RNN networks achieved lower accuracy levels with the 35-mer oestradiol dataset, consistent with the findings in [15]. However, when accounting for randomness, all datasets exhibited similar results in other networks.
In our future research, we plan to explore deep transfer learning techniques such as VGG-19, Inception, and ResNet-50 for our classification models. This will enable us to assess and compare their impact on the overall classification performance [54,55]. Additionally, we intend to incorporate attention mechanisms into our classification models. Attention mechanisms can improve feature extraction from input data, potentially enhancing the overall performance of our models [56]. We also can use hybrid networks [57] or CNN networks [58] for signal extrapolation to examine and compare their effects on time series forecasting.
The limitation in our study came from not having sufficient data. The first limitation was related to the signal analysis technique used. In machine learning tasks, it is typical to apply the same changes to both training and test data. However, we employed a method to scale each time series derived from its mean and standard deviation. This occurred due to the unfeasibility of calculating the mean and standard deviation using the existing statistical techniques. The second limitation was related to the lack of data for testing the CVAE networks. Thus, we used the scaling augmented data from our previous work. The third limitation was related to the data augmentation with VAE for anomaly data. There was insufficient abnormal data to evaluate the VAE networks for augmented data.

7. Conclusions

In this paper, we proposed three main techniques utilizing deep learning algorithms to examine the signals from the drain current of three comparable electrochemical sensors across the datasets named the 35-mer adenosine, and the 31-mer and 35-mer oestradiol. These signals captured the sensors’ responses as the concentrations of the target substances gradually rose from 1 nM to 10 μM. Our primary goal was to achieve automatic identification and measurement of specific analytes based on a section of the signal captured by these sensors. These three main steps to fulfil our goal can be categorized into (1) data augmentation, (2) signal extrapolation, and (3) classification models.
The CVAE-based data augmentation method proved highly effective in enhancing and strengthening the generalization abilities of classification models, while only insufficient original data was provided. Furthermore, the results indicate that this augmentation technique was particularly impactful when coupled with the STFT method as the data preprocessing technique.
We used a signal extrapolation method to generate uniform signal lengths for all the available datasets as the feed for classification models. The proposed GRU-based signal extrapolation algorithm demonstrated exceptional performance in reconstructing original data within the 35-mer adenosine dataset, which featured lengthier signals. Consequently, this algorithm was applied to extrapolate oestradiol datasets with segments lasting 300 s. Notably, while this method did not yield effective results in our classification models, its inclusion in the study was pivotal for investigating our hypothesis concerning the significance of segment length in predictive modelling. Ultimately, the findings of this study do not support our initial hypothesis.
In terms of classification models, initially, we designed seven deep learning networks with the aim of detecting and quantifying specific analyte concentrations across six distinct classes. These classes ranged from the absence of the analyte to 10 μM, with each dataset’s segments being analyzed independently. The classification performance was evaluated with two metrics: accuracy and macro F1-score. The outcomes indicated that ConvGRU, ConvULSTM, ConvBLSTM, and CNN exhibited robustness and consistently delivered high-performance results compared to recurrent networks.
In the second classification model, we employed the same set of deep learning networks for a twelve-class classification task. The objective was to differentiate abnormal data from normal segments, ascertain the absence or presence of the analytes in the sample, and, if an analyte was present, identify the specific substance and quantify its concentration. The results mirrored those of the previous model, demonstrating that ConvGRU, ConvULSTM, ConvBLSTM, and CNN exhibited stability and consistently outperformed recurrent networks.
The outstanding performance of ConvGRU, ConvULSTM, ConvBLSTM, and CNN suggests that sequential data analysis does not necessarily yield better results with recurrent networks, which are specifically designed for such data. Our findings indicate that the combination of convolutional and recurrent layers, as seen in ConvRNN networks, outperforms pure RNN architectures, even with the increase in learnable parameters. This suggests that, in the context of our datasets, the optimal classification strategy involves using both recurrent and convolutional layers for effective time series feature extraction.
Additionally, our results highlight the efficiency of CNN models, which possess fewer learnable parameters yet demonstrate computational efficiency and speed and satisfactory classification performance when compared to ConvRNN models. This underscores the versatility of CNNs in handling time series data with remarkable efficiency and accuracy. The evaluation metrics of the classification models indicate that the proposed networks are effective tools for automatic identification and measurement of analyte concentration.
Furthermore, our results indicate that the exceptional performance of these networks can be effectively enhanced by employing an appropriate data preprocessing technique, such as the STFT method, as a time-frequency feature extraction utilized in our study. This suggests that the choice of data preprocessing technique could serve as a beneficial tool for improving neural network performance, aligning with our goal of automatic identification and measurement of specific analytes.

Author Contributions

Conceptualization, F.E., N.O.V.P., C.P.U. and A.W.; methodology, F.E.; software, F.E.; validation, F.E.; formal analysis, F.E.; investigation, F.E. and A.W.; resources—computing resources and analysis tools, F.E.; resources—laboratory samples, E.C. and H.P.T.N.; resources, N.O.V.P.; data curation, E.C.; writing—original draft preparation, F.E.; writing—review and editing, F.E., N.O.V.P., C.P.U. and A.W.; visualization, F.E.; supervision, A.W.; funding acquisition, N.O.V.P. and C.P.U. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Marsden Fund, managed by the Royal Society Te Apārangi, grant number VUW1708.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study might be available on request from the corresponding author. There are restrictions on data availability due to their necessity for our future work.

Acknowledgments

The authors express gratitude to Erica Happe for her assistance. It is acknowledged that Erica Happe identified the abnormal data within the 35-mer adenosine dataset.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MLMachine Learning
DLDeep Learning
STEM  Scanning Transmission Electron Microscopy
IDSIntrusion Detection System
ACAnalyte Concentrations
CNTCarbon Nanotube
STFTShort-Term Fourier Transform
FETField-Effect Transistor
GANGenerative Adversarial Network
AEAutoencoder
VAEVariational Autoencoder
CVAEConditional Variational Autoencoder
CNNConvolution Neural Network
RNNRecurrent Neural Networks
LSTMLong Short-Term Memory
ULSTMUnidirectional Long Short-Term Memory
BLSTMBidirectional Long Short-Term Memory
GRUGated Recurrent Unit
FCFully Connected
ELBOEvidence Lower Bound
AEVBAutoencoding Variational Bayes
SGVBStochastic Gradient Variational Bayes
KLKullback–Leibler

References

  1. Al-Ashwal, N.H.; Al Soufy, K.A.; Hamza, M.E.; Swillam, M.A. Deep learning for optical sensor applications: A review. Sensors 2023, 23, 6486. [Google Scholar] [CrossRef] [PubMed]
  2. Yasir, M.; Jianhua, W.; Shanwei, L.; Sheng, H.; Mingming, X.; Hossain, M. Coupling of deep learning and remote sensing: A comprehensive systematic literature review. Int. J. Remote Sens. 2023, 44, 157–193. [Google Scholar] [CrossRef]
  3. Incel, O.D.; Bursa, S.O. On-Device Deep Learning for Mobile and Wearable Sensing Applications: A Review. IEEE Sens. J. 2023, 23, 5501–5512. [Google Scholar] [CrossRef]
  4. Thapa, A.; Horanont, T.; Neupane, B.; Aryal, J. Deep Learning for Remote Sensing Image Scene Classification: A Review and Meta-Analysis. Remote Sens. 2023, 15, 4804. [Google Scholar] [CrossRef]
  5. Pan, J.; Hu, Z.; Yin, S.; Li, M. GRU with dual attentions for sensor-based human activity recognition. Electronics 2022, 11, 1797. [Google Scholar] [CrossRef]
  6. Zhou, K.; Liu, Y. Early-stage gas identification using convolutional long short-term neural network with sensor array time series data. Sensors 2021, 21, 4826. [Google Scholar] [CrossRef] [PubMed]
  7. Zhang, J.; Chen, F.; Zou, R.; Liao, J.; Zhang, Y.; Zhu, Z.; Yan, X.; Jiang, Z.; Tan, F. A CNN-Based Method for Heavy-Metal Ion Detection. Appl. Sci. 2023, 13, 4520. [Google Scholar] [CrossRef]
  8. Li, K.; Ao, B.; Wu, X.; Wen, Q.; Ul Haq, E.; Yin, J. Parkinson’s disease detection and classification using EEG based on deep CNN-LSTM model. Biotechnol. Genet. Eng. Rev. 2023. Online ahead of print. [Google Scholar] [CrossRef]
  9. Garcea, F.; Serra, A.; Lamberti, F.; Morra, L. Data augmentation for medical imaging: A systematic literature review. Comput. Biol. Med. 2022, 152, 106391. [Google Scholar] [CrossRef]
  10. Liu, C.; Antypenko, R.; Sushko, I.; Zakharchenko, O. Intrusion detection system after data augmentation schemes based on the VAE and CVAE. IEEE Trans. Reliab. 2022, 71, 1000–1010. [Google Scholar] [CrossRef]
  11. Prifti, E.; Buban, J.P.; Thind, A.S.; Klie, R.F. Variational Convolutional Autoencoders for Anomaly Detection in Scanning Transmission Electron Microscopy. Small 2023, 19, 2205977. [Google Scholar] [CrossRef] [PubMed]
  12. Chen, Z.; Ma, M.; Li, T.; Wang, H.; Li, C. Long sequence time-series forecasting with deep learning: A survey. Inf. Fusion 2023, 97, 101819. [Google Scholar] [CrossRef]
  13. Pirani, M.; Thakkar, P.; Jivrani, P.; Bohara, M.H.; Garg, D. A comparative analysis of ARIMA, GRU, LSTM and BiLSTM on financial time series forecasting. In Proceedings of the 2022 IEEE International Conference on Distributed Computing and Electrical Circuits and Electronics (ICDCECE), Ballari, India, 23–24 April 2022; IEEE: Piscataway, NJ, USA; pp. 1–6. [Google Scholar]
  14. Mahjoub, S.; Chrifi-Alaoui, L.; Marhic, B.; Delahoche, L. Predicting Energy Consumption Using LSTM, Multi-Layer GRU and Drop-GRU Neural Networks. Sensors 2022, 22, 4062. [Google Scholar] [CrossRef] [PubMed]
  15. Esmaeili, F.; Cassie, E.; Nguyen, H.P.T.; Plank, N.O.; Unsworth, C.P.; Wang, A. Predicting analyte concentrations from electrochemical aptasensor signals using LSTM recurrent networks. Bioengineering 2022, 9, 529. [Google Scholar] [CrossRef]
  16. Esmaeili, F.; Cassie, E.; Nguyen, H.P.T.; Plank, N.O.; Unsworth, C.P.; Wang, A. Anomaly Detection for Sensor Signals Utilizing Deep Learning Autoencoder-Based Neural Networks. Bioengineering 2023, 10, 405. [Google Scholar] [CrossRef]
  17. Rumelhart, D.E.; Hinton, G.E.; Williams, R.J. Learning representations by back-propagating errors. Nature 1986, 323, 533–536. [Google Scholar] [CrossRef]
  18. Elman, J.L. Finding structure in time. Cogn. Sci. 1990, 14, 179–211. [Google Scholar] [CrossRef]
  19. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  20. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  21. Cho, K.; Van Merriënboer, B.; Bahdanau, D.; Bengio, Y. On the properties of neural machine translation: Encoder-decoder approaches. arXiv 2014, arXiv:1409.1259. [Google Scholar]
  22. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  23. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  24. Contreras-Cruz, M.A.; Novo-Torres, L.; Villarreal, D.J.; Ramirez-Paredes, J.P. Convolutional neural network and sensor fusion for obstacle classification in the context of powered prosthetic leg applications. Comput. Electr. Eng. 2023, 108, 108656. [Google Scholar] [CrossRef]
  25. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  26. Doersch, C. Tutorial on variational autoencoders. arXiv 2016, arXiv:1606.05908. [Google Scholar]
  27. Park, D.; Hoshi, Y.; Kemp, C.C. A multimodal anomaly detector for robot-assisted feeding using an lstm-based variational autoencoder. IEEE Robot. Autom. Lett. 2018, 3, 1544–1551. [Google Scholar] [CrossRef]
  28. Sohn, K.; Lee, H.; Yan, X. Learning structured output representation using deep conditional generative models. Adv. Neural Inf. Process. Syst. 2015, 28, 3483–3491. [Google Scholar]
  29. Nguyen, H.P.T.; Murugathas, T.; Plank, N.O. Comparison of duplex and quadruplex folding structure adenosine aptamers for carbon nanotube field effect transistor aptasensors. Nanomaterials 2021, 11, 2280. [Google Scholar] [CrossRef]
  30. Jauset-Rubio, M.; Botero, M.L.; Skouridou, V.; Aktas, G.B.; Svobodova, M.; Bashammakh, A.S.; El-Shahawi, M.S.; Alyoubi, A.O.; O’Sullivan, C.K. One-pot SELEX: Identification of specific aptamers against diverse steroid targets in one selection. ACS Omega 2019, 4, 20188–20196. [Google Scholar] [CrossRef]
  31. Alsager, O.A.; Kumar, S.; Zhu, B.; Travas-Sejdic, J.; McNatty, K.P.; Hodgkiss, J.M. Ultrasensitive colorimetric detection of 17β-estradiol: The effect of shortening DNA aptamer sequences. Anal. Chem. 2015, 87, 4201–4209. [Google Scholar] [CrossRef]
  32. Maleki, S.; Maleki, S.; Jennings, N.R. Unsupervised anomaly detection with LSTM autoencoders using statistical data-filtering. Appl. Soft Comput. 2021, 108, 107443. [Google Scholar] [CrossRef]
  33. Stockwell, R.G.; Mansinha, L.; Lowe, R. Localization of the complex spectrum: The S transform. IEEE Trans. Signal Process. 1996, 44, 998–1001. [Google Scholar] [CrossRef]
  34. Akan, A.; Cura, O.K. Time–frequency signal processing: Today and future. Digit. Signal Process. 2021, 119, 103216. [Google Scholar] [CrossRef]
  35. Tang, S.; Yuan, S.; Zhu, Y. Data preprocessing techniques in convolutional neural network based on fault diagnosis towards rotating machinery. IEEE Access 2020, 8, 149487–149496. [Google Scholar] [CrossRef]
  36. Jafari, M.; Shoeibi, A.; Khodatars, M.; Bagherzadeh, S.; Shalbaf, A.; García, D.L.; Gorriz, J.M.; Acharya, U.R. Emotion recognition in EEG signals using deep learning methods: A review. Comput. Biol. Med. 2023, 165, 107450. [Google Scholar] [CrossRef]
  37. Yuan, Y.; Xun, G.; Jia, K.; Zhang, A. A multi-view deep learning method for epileptic seizure detection using short-time fourier transform. In Proceedings of the 8th ACM International Conference on Bioinformatics, Computational Biology, and Health Informatics, Boston, MA, USA, 20–23 August 2017; pp. 213–222. [Google Scholar]
  38. Garcia, C.I.; Grasso, F.; Luchetta, A.; Piccirilli, M.C.; Paolucci, L.; Talluri, G. A comparison of power quality disturbance detection and classification methods using CNN, LSTM and CNN-LSTM. Appl. Sci. 2020, 10, 6755. [Google Scholar] [CrossRef]
  39. Cui, F.; Yue, Y.; Zhang, Y.; Zhang, Z.; Zhou, H.S. Advancing biosensors with machine learning. ACS Sens. 2020, 5, 3346–3364. [Google Scholar] [CrossRef]
  40. Kebaili, A.; Lapuyade-Lahorgue, J.; Ruan, S. Deep Learning Approaches for Data Augmentation in Medical Imaging: A Review. J. Imaging 2023, 9, 81. [Google Scholar] [CrossRef]
  41. Wen, Q.; Sun, L.; Yang, F.; Song, X.; Gao, J.; Wang, X.; Xu, H. Time series data augmentation for deep learning: A survey. arXiv 2020, arXiv:2002.12478. [Google Scholar]
  42. Iglesias, G.; Talavera, E.; González-Prieto, Á.; Mozo, A.; Gómez-Canaval, S. Data Augmentation techniques in time series domain: A survey and taxonomy. Neural Comput. Appl. 2023, 35, 10123–10145. [Google Scholar] [CrossRef]
  43. Lim, B.; Zohren, S. Time-series forecasting with deep learning: A survey. Philos. Trans. R. Soc. A 2021, 379, 20200209. [Google Scholar] [CrossRef] [PubMed]
  44. Torres, J.F.; Hadjout, D.; Sebaa, A.; Martínez-Álvarez, F.; Troncoso, A. Deep learning for time series forecasting: A survey. Big Data 2021, 9, 3–21. [Google Scholar] [CrossRef] [PubMed]
  45. Namuduri, S.; Narayanan, B.N.; Davuluru, V.S.P.; Burton, L.; Bhansali, S. Deep learning methods for sensor based predictive maintenance and future perspectives for electrochemical sensors. J. Electrochem. Soc. 2020, 167, 037552. [Google Scholar] [CrossRef]
  46. Marimon, X.; Traserra, S.; Jiménez, M.; Ospina, A.; Benítez, R. Detection of abnormal cardiac response patterns in cardiac tissue using deep learning. Mathematics 2022, 10, 2786. [Google Scholar] [CrossRef]
  47. Cheng, D.; Fan, Y.; Fang, S.; Wang, M.; Liu, H. ResNet-AE for Radar Signal Anomaly Detection. Sensors 2022, 22, 6249. [Google Scholar] [CrossRef]
  48. Zheng, M.; Man, J.; Wang, D.; Chen, Y.; Li, Q.; Liu, Y. Semi-supervised multivariate time series anomaly detection for wind turbines using generator SCADA data. Reliab. Eng. Syst. Saf. 2023, 235, 109235. [Google Scholar] [CrossRef]
  49. Yang, Z.; Li, Y.; Zhou, G. TS-GAN: Time-series GAN for Sensor-based Health Data Augmentation. ACM Trans. Comput. Healthc. 2023, 4, 12. [Google Scholar] [CrossRef]
  50. Morid, M.A.; Sheng, O.R.L.; Dunbar, J. Time series prediction using deep learning methods in healthcare. ACM Trans. Manag. Inf. Syst. 2023, 14, 2. [Google Scholar] [CrossRef]
  51. Ghali, R.; Akhloufi, M.A. Deep Learning Approaches for Wildland Fires Remote Sensing: Classification, Detection, and Segmentation. Remote Sens. 2023, 15, 1821. [Google Scholar] [CrossRef]
  52. Dhaka, V.S.; Kundu, N.; Rani, G.; Zumpano, E.; Vocaturo, E. Role of Internet of Things and Deep Learning Techniques in Plant Disease Detection and Classification: A Focused Review. Sensors 2023, 23, 7877. [Google Scholar] [CrossRef]
  53. Zheng, W.; Chen, G. An accurate GRU-based power time-series prediction approach with selective state updating and stochastic optimization. IEEE Trans. Cybern. 2021, 52, 13902–13914. [Google Scholar] [CrossRef] [PubMed]
  54. Jana, S.; Middya, A.I.; Roy, S. Participatory Sensing Based Urban Road Condition Classification using Transfer Learning. Mob. Netw. Appl. 2023, 1–17. [Google Scholar] [CrossRef]
  55. Neupane, B.; Horanont, T.; Aryal, J. Real-time vehicle classification and tracking using a transfer learning-improved deep learning network. Sensors 2022, 22, 3813. [Google Scholar] [CrossRef]
  56. Li, W.; Chen, H.; Liu, Q.; Liu, H.; Wang, Y.; Gui, G. Attention mechanism and depthwise separable convolution aided 3DCNN for hyperspectral remote sensing image classification. Remote Sens. 2022, 14, 2215. [Google Scholar] [CrossRef]
  57. Livieris, I.E.; Pintelas, E.; Pintelas, P. A CNN–LSTM model for gold price time-series forecasting. Neural Comput. Appl. 2020, 32, 17351–17360. [Google Scholar] [CrossRef]
  58. Silva, A.Q.B.; Gonçalves, W.N.; Matsubara, E.T. DESCINet: A hierarchical deep convolutional neural network with skip connection for long time series forecasting. Expert Syst. Appl. 2023, 228, 120246. [Google Scholar] [CrossRef]
Figure 1. Workflow of the proposed deep learning techniques in this study.
Figure 1. Workflow of the proposed deep learning techniques in this study.
Bioengineering 10 01348 g001
Figure 2. Visualization of an RNN unfolded in time, illustrating the transmission of information within an RNN layer over three consecutive time steps.
Figure 2. Visualization of an RNN unfolded in time, illustrating the transmission of information within an RNN layer over three consecutive time steps.
Bioengineering 10 01348 g002
Figure 3. Structure of a vanilla RNN unit using a hyperbolic tangent function as its activation function.
Figure 3. Structure of a vanilla RNN unit using a hyperbolic tangent function as its activation function.
Bioengineering 10 01348 g003
Figure 4. Structure of an LSTM unit or an LSTM hidden unit (⊕ and ⊙ refer to element-wise addition and multiplication, respectively).
Figure 4. Structure of an LSTM unit or an LSTM hidden unit (⊕ and ⊙ refer to element-wise addition and multiplication, respectively).
Bioengineering 10 01348 g004
Figure 5. Transmission of data within an LSTM layer: (a) a ULSTM layer with forward states, and (b) a BLSTM layer with both forward and backward states. Here, X, Y, h , and h represent the input, output, forward and backward states in the BLSTM layer, respectively.
Figure 5. Transmission of data within an LSTM layer: (a) a ULSTM layer with forward states, and (b) a BLSTM layer with both forward and backward states. Here, X, Y, h , and h represent the input, output, forward and backward states in the BLSTM layer, respectively.
Bioengineering 10 01348 g005
Figure 6. Illustration of GRU structure (⊕ and ⊙ refer to element-wise addition and multiplication, respectively).
Figure 6. Illustration of GRU structure (⊕ and ⊙ refer to element-wise addition and multiplication, respectively).
Bioengineering 10 01348 g006
Figure 7. A typical structure of CNN.
Figure 7. A typical structure of CNN.
Bioengineering 10 01348 g007
Figure 8. Ilustration of a variational autoencoder (VAE) network architecture.
Figure 8. Ilustration of a variational autoencoder (VAE) network architecture.
Bioengineering 10 01348 g008
Figure 9. Illustration of conditional variational autoencoder (CVAE) network architecture that integrates conditional information for enhanced generative modeling.
Figure 9. Illustration of conditional variational autoencoder (CVAE) network architecture that integrates conditional information for enhanced generative modeling.
Bioengineering 10 01348 g009
Figure 10. The stereotypical sensing time series registered by aptasensors include (a) the typical signal from DS 1 with an initial AC starting from 1 nM, (b) another typical signal from the same dataset with AC starting from 1 μM, (c) a sample from DS 2, and (d) a sample from DS 3. Note that AC, C1 Labels, and C2 Labels refer to the analyte concentration and the segment labels for classification models 1 and 2 in this paper, respectively.
Figure 10. The stereotypical sensing time series registered by aptasensors include (a) the typical signal from DS 1 with an initial AC starting from 1 nM, (b) another typical signal from the same dataset with AC starting from 1 μM, (c) a sample from DS 2, and (d) a sample from DS 3. Note that AC, C1 Labels, and C2 Labels refer to the analyte concentration and the segment labels for classification models 1 and 2 in this paper, respectively.
Bioengineering 10 01348 g010aBioengineering 10 01348 g010b
Figure 11. An example of data normalization for the adenosine dataset: (a) the raw signal, (b) the normalized signal according to Z-score scaling.
Figure 11. An example of data normalization for the adenosine dataset: (a) the raw signal, (b) the normalized signal according to Z-score scaling.
Bioengineering 10 01348 g011
Figure 12. These plots illustrate a comparison between normal 35-mer oestradiol signal patterns and the patterns of abnormal signals: (a) signals showing normal and standard behaviour, (b) abnormal time series (the blue and red lines represent signals from non-sensing and malfunctioning transistor sensors, respectively), (c) signals with abnormal time intervals, captured by functional sensors but exhibiting unusual patterns from 300 to 750 s.
Figure 12. These plots illustrate a comparison between normal 35-mer oestradiol signal patterns and the patterns of abnormal signals: (a) signals showing normal and standard behaviour, (b) abnormal time series (the blue and red lines represent signals from non-sensing and malfunctioning transistor sensors, respectively), (c) signals with abnormal time intervals, captured by functional sensors but exhibiting unusual patterns from 300 to 750 s.
Bioengineering 10 01348 g012
Figure 13. Illustration of the spectrogram for a typical and normal 10 μM 35-mer adenosine segment: (a) STFT of the initial 300 s, and (b) STFT of the full 500 s.
Figure 13. Illustration of the spectrogram for a typical and normal 10 μM 35-mer adenosine segment: (a) STFT of the initial 300 s, and (b) STFT of the full 500 s.
Bioengineering 10 01348 g013
Figure 14. Structure of CVAE networks for data augmentation of normal data: (a) the encoder module and (b) the decoder network.
Figure 14. Structure of CVAE networks for data augmentation of normal data: (a) the encoder module and (b) the decoder network.
Bioengineering 10 01348 g014
Figure 15. Structure of VAE networks for data augmentation of abnormal data: (a) the encoder network, and (b) the decoder module. Note that the anomaly data augmentation was essential for C2 models.
Figure 15. Structure of VAE networks for data augmentation of abnormal data: (a) the encoder network, and (b) the decoder module. Note that the anomaly data augmentation was essential for C2 models.
Bioengineering 10 01348 g015
Figure 16. A schematic of the network architecture. The primary distinction among these networks lay in their recurrent layer, which incorporated either a GRU, unidirectional LSTM, or bidirectional LSTM layer.
Figure 16. A schematic of the network architecture. The primary distinction among these networks lay in their recurrent layer, which incorporated either a GRU, unidirectional LSTM, or bidirectional LSTM layer.
Bioengineering 10 01348 g016
Figure 17. Illustration of two iterations of signal extrapolation procedure on a segment of the 35-mer adenosine dataset with AC of 10 μM: (a) first iteration, and (b) second iteration.
Figure 17. Illustration of two iterations of signal extrapolation procedure on a segment of the 35-mer adenosine dataset with AC of 10 μM: (a) first iteration, and (b) second iteration.
Bioengineering 10 01348 g017
Figure 18. A schematic of the recurrent-based network architecture for classification models. The primary distinction among these networks lay in their second layer, which included either a GRU, ULSTM, or BLSTM layer.
Figure 18. A schematic of the recurrent-based network architecture for classification models. The primary distinction among these networks lay in their second layer, which included either a GRU, ULSTM, or BLSTM layer.
Bioengineering 10 01348 g018
Figure 19. Illustration of ConvRNN networks architectures used for enhanced feature extraction in classification tasks.
Figure 19. Illustration of ConvRNN networks architectures used for enhanced feature extraction in classification tasks.
Bioengineering 10 01348 g019
Figure 20. Illustration of CNN network architecture used for classification models.
Figure 20. Illustration of CNN network architecture used for classification models.
Bioengineering 10 01348 g020
Figure 21. Performance evaluation and an example of signal extrapolation on the 35-mer adenosine dataset: histograms of prediction error for (a) GRU, (b) ULSTM, and (c) BLSTM networks, along with (d) an example of forecasting future sensor outputs with three GRU-based prediction models at AC of 10 μM.
Figure 21. Performance evaluation and an example of signal extrapolation on the 35-mer adenosine dataset: histograms of prediction error for (a) GRU, (b) ULSTM, and (c) BLSTM networks, along with (d) an example of forecasting future sensor outputs with three GRU-based prediction models at AC of 10 μM.
Bioengineering 10 01348 g021
Figure 22. Examples of signal extrapolation with GRU network on oestradiol datasets: (a) a 31-mer Oestradiol segment with AC of 1 μM, (b) a 35-mer Oestradiol segment with AC of 100 nM.
Figure 22. Examples of signal extrapolation with GRU network on oestradiol datasets: (a) a 31-mer Oestradiol segment with AC of 1 μM, (b) a 35-mer Oestradiol segment with AC of 100 nM.
Bioengineering 10 01348 g022
Figure 23. Performance metrics of C1 model on 35-mer adenosine dataset: (a) accuracy and (b) macro F1-score.
Figure 23. Performance metrics of C1 model on 35-mer adenosine dataset: (a) accuracy and (b) macro F1-score.
Bioengineering 10 01348 g023
Figure 24. Performance metrics of C1 model on 31-mer oestradiol dataset: (a) accuracy and (b) macro F1-score.
Figure 24. Performance metrics of C1 model on 31-mer oestradiol dataset: (a) accuracy and (b) macro F1-score.
Bioengineering 10 01348 g024
Figure 25. Performance metrics of C1 model on 35-mer oestradiol dataset: (a) accuracy and (b) macro F1-score.
Figure 25. Performance metrics of C1 model on 35-mer oestradiol dataset: (a) accuracy and (b) macro F1-score.
Bioengineering 10 01348 g025
Figure 26. Performance metrics of C2 model: (a) accuracy and (b) macro F1-score.
Figure 26. Performance metrics of C2 model: (a) accuracy and (b) macro F1-score.
Bioengineering 10 01348 g026
Table 1. The total count of signals in the provided datasets.
Table 1. The total count of signals in the provided datasets.
Dataset IDDataset NameDataset Size
DS 135-mer Adenosine15
DS 231-mer Oestradiol24
DS 335-mer Oestradiol24
Table 2. The datasets and the fundamental elements of the sensors responsible for recording signals in each dataset.
Table 2. The datasets and the fundamental elements of the sensors responsible for recording signals in each dataset.
Dataset NameAnalyteTransducerBioreceptorRef.
35-mer AdenosineAdenosineCNT FET 5 -NH2-AAAAAAAAAACCTGGGGGAGTATTGCGGAGGAAGG- 3 [29]
31-mer OestradiolOestradiolCNT FET 5 -GGTCCTGACCGGAGGCTGACCGGAGTGGGAA- 3 [30] 1
35-mer OestradiolOestradiolCNT FET 5 -AAGGGATGCCGTTTGGGCCCAAGTTCGGCATAGTG- 3 [31]
1 Erica S. Cassie created the aptamer for this dataset, which was a variation of the one described in the cited source. This sequence incorporated the shared segment of the best three oestradiol aptamers, with additional segments added at both ends.
Table 3. Comparison of the sensing procedures of the sensors. Note that the details regarding the oestradiol sensors are combined into a single column, as they share similar sensing procedures.
Table 3. Comparison of the sensing procedures of the sensors. Note that the details regarding the oestradiol sensors are combined into a single column, as they share similar sensing procedures.
CharacteristicsAdenosine SensorOestradil Sensors
Time interval measurement1 s1.081 s with std 5 × 10 3
Gate voltage ( V G )0 V0 V
Drain voltage ( V D )100 mV100 mV
Buffer solution2 mM Tris-HCI 0.05 × PBS with 5% EtOH
Initial step load chemical110 μM of 2 mM Tris-HCI100 μL of 0.05 × PBS 5% EtOH
Next steps load chemical-20 μL of 0.05 × PBS 5% EtOH
Initial analyte load time1000 s600 s
Time interval of analyte injection500 s300 s
Time interval of chemical injection-300 s
Variation of analyte concentration100 pM–10 μM1 nM–10 μM
Table 4. The total number of segments corresponds to the datasets and is categorized based on their analyte concentration levels. It should be noted that the 100 pM segment in DS 1 was removed due to data scarcity and will not be mentioned in this study afterwards.
Table 4. The total number of segments corresponds to the datasets and is categorized based on their analyte concentration levels. It should be noted that the 100 pM segment in DS 1 was removed due to data scarcity and will not be mentioned in this study afterwards.
Analyte ConcentrationDS 1DS 2DS 3
No Analyte152424
100 pM100
1 nM52424
10 nM72424
100 nM92424
1 μM122424
10 μM152424
Total Segments63144144
Table 5. The total size of normal segments for available datasets after applying anomaly detection.
Table 5. The total size of normal segments for available datasets after applying anomaly detection.
Analyte ConcentrationDS 1DS 2DS 3
No Analyte9185
1 nM4185
10 nM41811
100 nM41811
1 μM41811
10 μM61811
Table 6. Segment labeling for classification models C1 and C2.
Table 6. Segment labeling for classification models C1 and C2.
Analyte ConcentrationC1 LabelsC2 Labels
DS 1DS 2DS 3DS 1DS 2DS 3
No Analyte111111
1 nM222277
10 nM333388
100 nM444499
1 μM55551010
10 μM66661111
Anomaly---121212
Table 7. Layer description of the recurrent-based networks developed for signal extrapolation.
Table 7. Layer description of the recurrent-based networks developed for signal extrapolation.
Layer NumberLayer TypeHyperparametersLearnable ParametersState Parameters
1Sequential inputOutput size: 1--
2GRUInput size: 1 W x : 3 n × 1 Hidden state: n × 1
Hidden units: n W h : 3 n × n
Output size: nb: 3 n × 1
ULSTMInput size: 1 W x : 4 n × 1 Hidden state: n × 1
Hidden units: n W h : 4 n × n
Output size: nb: 4 n × 1 Cell state: n × 1
BLSTMInput size: 1 W x : 8 n × 1 Hidden state: 2 n × 1
Hidden units: n W h : 8 n × n
Output size: 2 n b: 8 n × 1 Cell state: 2 n × 1
3FC (GRU, ULSTM) 1Input size: nWeights: 1 × n -
Output size: 1Bias: 1 × 1
FC (BLSTM) 2Input size: 2 n Weights: 1 × 2 n -
Output size: 1Bias: 1 × 1
4Regression outputOutput size: 1--
1,2 A fully connected layer that succeeds GRU, ULSTM, and BLSTM layers, respectively.
Table 8. Layer description and comparison of the developed recurrent-based networks for classification models.
Table 8. Layer description and comparison of the developed recurrent-based networks for classification models.
Layer NumberLayer TypeHyperparameters (Output Size)Learnable ParametersState Parameters
1Sequential input1--
2GRUInput size: 1 W x : 3 n × 1
Hidden units: n W h : 3 n × n Hidden state: n × 1
Output size: nb: 3 n × 1
ULSTMInput size: 1 W x : 4 n × 1 Hidden state: n × 1
Hidden units: n W h : 4 n × n
Output size: nb: 4 n × 1 Cell state: n × 1
BLSTMInput size: 1 W x : 8 n × 1 Hidden state: 2 n × 1
Hidden units: n W h : 8 n × n
Output size: 2 n b: 8 n × 1 Cell state: 2 n × 1
3FC (GRU, ULSTM)mWeights: m × n -
Bias: m × 1
FC (BLSTM)mWeights: m × 2 n -
Bias: m × 1
4, 5Softmax, Classificationm--
Table 9. Architecture and layer descriptions of the proposed Conv-GRU, Conv-ULSTM, and Conv-BLSTM networks for classification models.
Table 9. Architecture and layer descriptions of the proposed Conv-GRU, Conv-ULSTM, and Conv-BLSTM networks for classification models.
Layer NumberLayer TypeHyperparameters (Output Size)Learnable ParametersState Parameters
1Image Input l s --
22D Convolution [ l c n f ] Weights: s f × n f -
Bias: 1 × n f
3Leaky ReLU, scale = 0.1 [ l c n f ] --
4Flatten k = l c × n f --
5GRUInput size: k W x : 3 n × k
Hidden units: n W h : 3 n × n Hidden state: n × 1
Output size: nb: 3 n × 1
ULSTMInput size: k W x : 4 n × k Hidden state: n × 1
Hidden units: n W h : 4 n × n
Output size: nb: 4 n × 1 Cell state: n × 1
BLSTMInput size: k W x : 8 n × k Hidden state: 2 n × 1
Hidden units: n W h : 8 n × n
Output size: 2 n b: 8 n × 1 Cell state: 2 n × 1
6Leaky ReLU, scale = 0.1GRU, ULSTM: n--
BLSTM: 2 n
7FC (GRU, ULSTM)mWeights: m × n -
Bias: m × 1
FC (BLSTM)mWeights: m × 2 n -
Bias: m × 1
8, 9Softmax, Classificationm--
Table 10. Architecture and layer descriptions of the proposed CNN network for classification models. Note that ConvBlock refers to a combination of a 2D convolution layer, a Leaky ReLU activation function, and a dropout layer placed sequentially.
Table 10. Architecture and layer descriptions of the proposed CNN network for classification models. Note that ConvBlock refers to a combination of a 2D convolution layer, a Leaky ReLU activation function, and a dropout layer placed sequentially.
Layer NumberLayer TypeHyperparameters (Output Size)Learnable Parameters
1Image Input l s -
22D Convolution [ l c 1 n f 1 ] Weights: s f × n f 1
Bias: 1 × n f 1
3Leaky ReLU, scale = 0.1 [ l c 1 n f 1 ] -
4Dropout, probability = 0.25 [ l c 1 n f 1 ] -
5, 6, 72D ConvBlock [ l c 2 n f 2 ] Weights: s f × n f 2
Bias: 1 × n f 2
8, 9, 102D ConvBlock [ l c 3 n f 3 ] Weights: s f × n f 3
Bias: 1 × n f 3
11, 12, 132D ConvBlock [ l c 4 n f 3 ] Weights: s f × n f 3
Bias: 1 × n f 3
14, 15, 162D ConvBlock [ l c 5 n f 2 ] Weights: s f × n f 2
Bias: 1 × n f 2
17, 18, 192D ConvBlock [ l c 6 n f 1 ] Weights: s f × n f 1
Bias: 1 × n f 1
20FCInput size: k = l c 6 × n f 1 Weights: m × k
Output size: mBias: m × 1
21, 22Softmax, Classificationm-
Table 11. Detailed information regarding the output sizes of the layers in the networks developed for data augmentation. Additionally, “S”, “C”, and “B” denote the spatial size, channel number, and batch size, respectively. Furthermore, CVAE and VAE represent the networks used for data augmentation of normal and anomaly data, respectively.
Table 11. Detailed information regarding the output sizes of the layers in the networks developed for data augmentation. Additionally, “S”, “C”, and “B” denote the spatial size, channel number, and batch size, respectively. Furthermore, CVAE and VAE represent the networks used for data augmentation of normal and anomaly data, respectively.
LayersLayers NameEncoderDecoder
CVAEVAEEncoder/DecoderDS 1DS 2 & DS 3DS 1DS 2 & DS 3
11Image/Feature Input500 (S) × 1 (S) × 1 (C) × 1 (B)300 (S) × 1 (S) × 1 (C) × 1 (B)32 (C) × 1 (B)32 (C) × 1 (B)
22FC500 (C) × 1 (B)300 (C) × 1 (B)4096 (C) × 1 (B)2048 (C) × 1 (B)
33Reshape500 (S) × 1 (S) × 1 (C) × 1 (B)300 (S) × 1 (S) × 1 (C) × 1 (B)64 (S) × 1 (S) × 64 (C) × 1 (B)32 (S) × 1 (S) × 64 (C) × 1 (B)
4-Feature Input1 (C) × 1 (B)1 (C) × 1 (B)1 (C) × 1 (B)1 (C) × 1 (B)
5-Embedding32 (C) × 1 (B)32 (C) × 1 (B)32 (C) × 1 (B)32 (C) × 1 (B)
6-FC500 (C) × 1 (B)300 (C) × 1 (B)64 (C) × 1 (B)32 (C) × 1 (B)
7-Reshape500 (S) × 1 (S) × 1 (C) × 1 (B)300 (S) × 1 (S) × 2 (C) × 1 (B)64 (S) × 1 (S) × 1 (C) × 1 (B)32 (S) × 1 (S) × 1 (C) × 1 (B)
8-Concatenation500 (S) × 1 (S) × 2 (C) × 1 (B)300 (S) × 1 (S) × 2 (C) × 1 (B)64 (S) × 1 (S) × 65 (C) × 1 (B)32 (S) × 1 (S) × 65 (C) × 1 (B)
94Conv/Deconv250 (S) × 1 (S) × 16 (C) × 1 (B)150 (S) × 1 (S) × 16 (C) × 1 (B)63 (S) × 1 (S) × 16 (C) × 1 (B)38 (S) × 1 (S) × 16 (C) × 1 (B)
105Leaky ReLU, scale = 0.1250 (S) × 1 (S) × 16 (C) × 1 (B)150 (S) × 1 (S) × 16 (C) × 1 (B)63 (S) × 1 (S) × 16 (C) × 1 (B)38 (S) × 1 (S) × 16 (C) × 1 (B)
116Dropout, probability = 0.25250 (S) × 1 (S) × 16 (C) × 1 (B)150 (S) × 1 (S) × 16 (C) × 1 (B)63 (S) × 1 (S) × 16 (C) × 1 (B)38 (S) × 1 (S) × 16 (C) × 1 (B)
127Conv/Deconv125 (S) × 1 (S) × 16 (C) × 1 (B)75 (S) × 1 (S) × 16 (C) × 1 (B)125 (S) × 1 (S) × 16 (C) × 1 (B)75 (S) × 1 (S) × 16 (C) × 1 (B)
13-Conv/Deconv125 (S) × 1 (S) × 16 (C) × 1 (B)75 (S) × 1 (S) × 16 (C) × 1 (B)63 (S) × 1 (S) × 16 (C) × 1 (B)75 (S) × 1 (S) × 16 (C) × 1 (B)
14-Addition125 (S) × 1 (S) × 16 (C) × 1 (B)75 (S) × 1 (S) × 16 (C) × 1 (B)63 (S) × 1 (S) × 16 (C) × 1 (B)75 (S) × 1 (S) × 16 (C) × 1 (B)
158Leaky ReLU, scale = 0.1125 (S) × 1 (S) × 16 (C) × 1 (B)75 (S) × 1 (S) × 16 (C) × 1 (B)63 (S) × 1 (S) × 16 (C) × 1 (B)75 (S) × 1 (S) × 16 (C) × 1 (B)
169Dropout, probability = 0.25125 (S) × 1 (S) × 16 (C) × 1 (B)75 (S) × 1 (S) × 16 (C) × 1 (B)63 (S) × 1 (S) × 16 (C) × 1 (B)75 (S) × 1 (S) × 16 (C) × 1 (B)
1710Conv/Deconv63 (S) × 1 (S) × 16 (C) × 1 (B)38 (S) × 1 (S) × 16 (C) × 1 (B)249 (S) × 1 (S) × 16 (C) × 1 (B)150 (S) × 1 (S) × 16 (C) × 1 (B)
1811Conv/Deconv63 (S) × 1 (S) × 16 (C) × 1 (B)38 (S) × 1 (S) × 16 (C) × 1 (B)249 (S) × 1 (S) × 16 (C) × 1 (B)150 (S) × 1 (S) × 16 (C) × 1 (B)
1912Addition63 (S) × 1 (S) × 16 (C) × 1 (B)38 (S) × 1 (S) × 16 (C) × 1 (B)249 (S) × 1 (S) × 16 (C) × 1 (B)150 (S) × 1 (S) × 16 (C) × 1 (B)
2013Leaky ReLU, scale = 0.163 (S) × 1 (S) × 16 (C) × 1 (B)38 (S) × 1 (S) × 16 (C) × 1 (B)249 (S) × 1 (S) × 16 (C) × 1 (B)150 (S) × 1 (S) × 16 (C) × 1 (B)
2114Dropout, probability = 0.2563 (S) × 1 (S) × 16 (C) × 1 (B)38 (S) × 1 (S) × 16 (C) × 1 (B)249 (S) × 1 (S) × 16 (C) × 1 (B)150 (S) × 1 (S) × 16 (C) × 1 (B)
2215FC/Deconv64 (C) × 1 (B)64 (C) × 1 (B)500 (S) × 1 (S) × 1 (C) × 1 (B)300 (S) × 1 (S) × 1 (C) × 1 (B)
Z: 32 (C) × 1 (B)32 (C) × 1 (B)
2316Sampling μ z : 32 (C) × 1 (B)32 (C) × 1 (B)500 (S) × 1 (S) × 1 (C) × 1 (B)300 (S) × 1 (S) × 1 (C) × 1 (B)
σ z : 32 (C) × 1 (B)32 (C) × 1 (B)
Table 12. The performance metrics of the proposed data augmentation method and examples of data generated by the decoder modules of the CVAEs are presented. Columns 2 and 4 show examples of reconstructed and augmented segments, respectively, corresponding to their respective datasets. The segments for reconstruction were randomly selected from the test data and augmented data. AC refers to the analyte concentration of the segments.
Table 12. The performance metrics of the proposed data augmentation method and examples of data generated by the decoder modules of the CVAEs are presented. Columns 2 and 4 show examples of reconstructed and augmented segments, respectively, corresponding to their respective datasets. The segments for reconstruction were randomly selected from the test data and augmented data. AC refers to the analyte concentration of the segments.
DatasetReconstructed SegmentReconstruction ErrorGenerated Data
35mer AdenosineBioengineering 10 01348 i001
AC: no analyte
Bioengineering 10 01348 i002Bioengineering 10 01348 i003
AC: 1 nM
31mer OestradiolBioengineering 10 01348 i004
AC: 10 nM
Bioengineering 10 01348 i005Bioengineering 10 01348 i006
AC: 100 nM
35mer OestradiolBioengineering 10 01348 i007
AC: 1 μM
Bioengineering 10 01348 i008Bioengineering 10 01348 i009
AC: 10 μM
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Esmaeili, F.; Cassie, E.; Nguyen, H.P.T.; Plank, N.O.V.; Unsworth, C.P.; Wang, A. Utilizing Deep Learning Algorithms for Signal Processing in Electrochemical Biosensors: From Data Augmentation to Detection and Quantification of Chemicals of Interest. Bioengineering 2023, 10, 1348. https://doi.org/10.3390/bioengineering10121348

AMA Style

Esmaeili F, Cassie E, Nguyen HPT, Plank NOV, Unsworth CP, Wang A. Utilizing Deep Learning Algorithms for Signal Processing in Electrochemical Biosensors: From Data Augmentation to Detection and Quantification of Chemicals of Interest. Bioengineering. 2023; 10(12):1348. https://doi.org/10.3390/bioengineering10121348

Chicago/Turabian Style

Esmaeili, Fatemeh, Erica Cassie, Hong Phan T. Nguyen, Natalie O. V. Plank, Charles P. Unsworth, and Alan Wang. 2023. "Utilizing Deep Learning Algorithms for Signal Processing in Electrochemical Biosensors: From Data Augmentation to Detection and Quantification of Chemicals of Interest" Bioengineering 10, no. 12: 1348. https://doi.org/10.3390/bioengineering10121348

APA Style

Esmaeili, F., Cassie, E., Nguyen, H. P. T., Plank, N. O. V., Unsworth, C. P., & Wang, A. (2023). Utilizing Deep Learning Algorithms for Signal Processing in Electrochemical Biosensors: From Data Augmentation to Detection and Quantification of Chemicals of Interest. Bioengineering, 10(12), 1348. https://doi.org/10.3390/bioengineering10121348

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop