Next Article in Journal
COVID-19 Diagnosis: A Review of Rapid Antigen, RT-PCR and Artificial Intelligence Methods
Previous Article in Journal
Performance Characteristics of a Novel 3D-Printed Bubble Intermittent Mandatory Ventilator (B-IMV) for Adult Pulmonary Support
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Deep Learning Approach for ECG-Based Arrhythmia Classification

1
Department of Computer Science and Engineering, Graphic Era Deemed to Be University, Dehradun 248002, India
2
Department of Computer Science and Engineering, Thapar Institute of Engineering and Technology, Patiala 147004, India
*
Author to whom correspondence should be addressed.
Bioengineering 2022, 9(4), 152; https://doi.org/10.3390/bioengineering9040152
Submission received: 25 January 2022 / Revised: 26 March 2022 / Accepted: 28 March 2022 / Published: 2 April 2022

Abstract

:
Arrhythmias are defined as irregularities in the heartbeat rhythm, which may infrequently occur in a human’s life. These arrhythmias may cause potentially fatal complications, which may lead to an immediate risk of life. Thus, the detection and classification of arrhythmias is a pertinent issue for cardiac diagnosis. (1) Background: To capture these sporadic events, an electrocardiogram (ECG), a register containing the heart’s electrical function, is considered the gold standard. However, since ECG carries a vast amount of information, it becomes very complex and challenging to extract the relevant information from visual analysis. As a result, designing an efficient (automated) system to analyse the enormous quantity of data possessed by ECG is critical. (2) Method: This paper proposes a hybrid deep learning-based approach to automate the detection and classification process. This paper makes two-fold contributions. First, 1D ECG signals are translated into 2D Scalogram images to automate the noise filtering and feature extraction. Then, based on experimental evidence, by combining two learning models, namely 2D convolutional neural network (CNN) and the Long Short-Term Memory (LSTM) network, a hybrid model called 2D-CNN-LSTM is proposed. (3) Result: To evaluate the efficacy of the proposed 2D-CNN-LSTM approach, we conducted a rigorous experimental study using the widely adopted MIT–BIH arrhythmia database. The obtained results show that the proposed approach provides ≈98.7%, 99%, and 99% accuracy for Cardiac Arrhythmias (ARR), Congestive Heart Failure (CHF), and Normal Sinus Rhythm (NSR), respectively. Moreover, it provides an average sensitivity of the proposed model of 98.33% and a specificity value of 98.35%, for all three arrhythmias. (4) Conclusions: For the classification of arrhythmias, a robust approach has been introduced where 2 D scalogram images of ECG signals are trained over the CNN-LSTM model. The results obtained are better as compared to the other existing techniques and will greatly reduce the amount of intervention required by doctors. For future work, the proposed method can be applied over some live ECG signals and Bi-LSTM can be applied instead of LSTM.

1. Introduction

Cardiovascular (CVD) diseases are globally recognised as the main cause of death, and they manifest themselves in the form of myocardial infarction or heart attack. According to the WHO [1], CVD is responsible for 17.7 million deaths. Approximately 31% of all deaths occur in poor and middle-income nations, with 75% of these deaths happening in these countries. Arrhythmias are the type of CVD that represents irregular patterns of heartbeats, such as the beating of the heart too fast or too slow. Examples of arrhythmias include: a trial Fibrillation (AF), premature ventricular contraction (PVC), ventricular fibrillation (VF), and Tachycardia. Although single cardiac arrhythmias may have little impact on one’s life, a persistent one might cause fatal problems, such as prolonged PVC that occasionally turns into Ventricular Tachycardia (VT), or Ventricular Fibrillation that can immediately lead to heart failure. Ventricular arrhythmias are one of the most prevalent types of cardiac arrhythmias that result in irregular heartbeats are responsible for nearly 80% of sudden cardiac deaths [2,3]. If arrhythmia conditions are detected early enough, ECG signal analysis may improve the identification of risk factors for cardiac arrest. Thus, it is fair to infer that monitoring heart rhythm regularly is crucial for avoiding CVDs. Practitioners use electrocardiographs as a diagnostic tool to detect cardiovascular diseases called arrhythmias, which detects and interprets the heart’s electrical activity during the diagnosis and is represented in ECG signals. However, ECG signals are represented in the form of waves when an ECG machine is attached to the human body; to get an exact picture of the heart, ten electrodes are needed for capturing 12 leads (signals). According to Zubair et al. [4], 12 ECG leads are required to properly diagnose, which are divided into precocious leads (I, II, III, aVL, aVR, aVF) to precordial leads ( V 1 , V 2 , V 3 , V 4 , V 5 , V 6 ). P waves, Q waves, R waves, S waves, T waves, and U waves based on the heart’s anatomy are illustrated in Figure 1. as positive and negative deflections from the baseline that signify a particular electrical event.
Conventionally, arrhythmia diagnosis studies focused on the filtering of noise from ECG signals [5], waveform segmentation [6], and manual feature extraction [7,8]. Various scientists have tried to classify arrhythmias using different machine learning (ML) and data mining methodologies. Here, we discuss some of the machine learning and deep learning techniques for the classification of arrhythmias.
Sahoo et al. [9] identified QRS complex using Discrete Wavelet Transform and Empirical Mode Distribution (EMP) for noise reduction, and Support Vector Machine was used to classify 5 distinct kinds of arrhythmias with an accuracy of 98.39% and 99.87% sensitivity with an error rate of 0.42. Osowski et al. [10] used higher order statistics (HOS) and Hermite coefficients to detect the QRS complex. The performance of the proposed approach [10] is compared with spectral power density algorithm, genetic algorithm, and SVM for classification of 5 different types of arrhythmia, which provides an average accuracy of 98.7%, 98.85%, 93%. Although these models are quite accurate due to the manual feature extraction process, their computation cost is high. Plawiak et al. [11] used higher-order spectra for feature extraction, PCA for dimension reduction, and SVM to identify 5 different forms of arrhythmia with the accuracy of 99.28%.
Yang et al. [12] used HOS and Hermite functions for manual feature extraction, and they also used SVM for disease classification. Polat and Gunes et al. [13] suggested least square SVM (LS-SVM) for the classification of arrhythmia and used PCA to reduce the dimensionality of features from 256 to 15. Melgani and Bazi et al. [14], using SVM, carried out an experimental study to identify five types of irregular waveforms and natural beats in a database from the MIT-BIH dataset. To increase the efficiency of SVM, they adopted particle swarm optimization (PSO). The sole purpose of PSO is to fine-tune the discriminator function for selecting the best features for SVM classifier training. The authors [14], contrasted their technique to two existing classifiers, namely K-nearest neighbours (KNN) and radial basis function neural networks, and discovered that SVM outscored them with an ≈90% accuracy. In a similar vein, to optimise the discriminator function, GA is used in [15]. Furthermore, they used a modified SVM classifier for arrhythmia classification.
Dutta et al. [16] suggested an LS-SVM classifier for heartbeat classification, such as normal beats, PVC beats, and other beats, by utilising the MIT-BIH arrhythmia database. Their approach yields 96.12% accuracy for classification. Desai et al. [17], considering 48 records of MIT-BIH arrhythmias, suggested a classifier using SVM for the classification and Discrete Wavelet Transformation (DWT) for feature representation. They categorised five types of arrhythmia beats: (i) Non-ectopic (N), (ii) Supraventricular ectopic (S), (iii) Ventricular ectopic (V), (iv) Fusion (F), and (v) Unknown (U). Their approach resulted in an accuracy of 98.49%. Furthermore, the authors performed statistical analysis using ANOVA to validate the efficacy of the proposed approach.
Kumaraswamy et al. [18], considering the MIT-BIH arrhythmia database, proposed a new classifier for the classification of heartbeats that is useful for the detection of arrhythmias. In particular, they used a random forest tree classifier and discrete cosine transform (DCT) for discovering R-R intervals (an essential pattern that helps in detecting arrhythmias) as features.
Park et al. [19] proposed a classifier for detecting 17 different types of heartbeats that can be used to detect arrhythmias. They carried out a two-step experimental study: (a) in the first step, P waves and the QRS complex are identified using the Pan-Tompkins method, and (b) in the second step, KNN classifier is used to classify them. This model was validated using MIT-BIH arrhythmias and performed with a sensitivity of 97.1% and specificity of 96.9%. Jun et al. [20] used a high-performance GPU-based cloud system for arrhythmia detection. Similar to [19], they used the Pan-Tompkins algorithm and KNN for identification and classification.
Machine learning paradigms are heavily influenced by feature architecture and a focus on feature extraction and filtering. The underlying concept behind learning is to include all of the data in signals so that the machine learning algorithm can learn and choose functions. This theory also underpins the deep learning paradigm, especially CNN and its 1-D equivalents [21]. Because of the potential and prospect of deep learning techniques, researchers [21,22,23,24,25] have adopted these techniques for the detection/classification of various types of chronic diseases. In this direction, Acharya et al. [22,23,26,27] conducted a detailed study for arrhythmia classification using deep learning. In [26], they proposed an automated classifier for arrhythmias. In [23], a CNN architecture is built to predict myocardial infarction with a 95.22% accuracy. In [27], an automated CNN model for the categorization of shockable and non-shockable ventricular arrhythmias is proposed. The suggested model outperformed with an accuracy of 93.18%, 95.32% sensitivity, and 94.04% specificity.
Kachuee et al. [24] utilized deep residual CNN for classification, and t-SNE was used for visualization. They employed their approach to identify five distinct arrhythmias by the AAMI EC57 norm and offered better results than ALTAN et al. [25] on the same database of MIT-BIH arrhythmias with an accuracy of 95.9%. Xia et al. [28] utilised short-term Fourier transform (STFT) and stationary wavelet transform (SWT) for the detection of arterial fibrillation arrhythmias. Further, to evaluate ECG segments, STFT and SWT outputs are input to two separate deep CNN models. As a result, CNN outperformed with 98.29% for STFT and 98.63% for SWT. Savalia et al. [29] proposed a model based on 5-layer CNN that can classify five types of arrhythmias. Although these models are quite accurate for minimising the loss function in backpropagation, they still suffer from the vanishing gradient problem of exploding gradients. For the classification of cardiac arrest during CPR, Jekova et al. [30] employed an end-to-end CNN model entitled CNN3-CC-ECG, which they validated on the independent database OHCA. With automatic feature extraction, the model performed well, with a sensitivity of 89% for ventricular fibrillation (VF), a specificity of 91.7% for non-shockable organised rhythms (OR), and a specificity of 91.11% for asystole, but found moderate results with noisy data. Krasteva et al. [31] applied convolution DNN for the classification of shockable and non-shockable rhythms, and to optimise the model they used random search optimization. Elola et al. [32] utilized two DNN architectures to classify pulseless electrical activity (PEA) and pulse-generating rhythm (PR) using short ECG reading segments (5s) on an independent database (OHCA). Additionally, the model was optimised using Bayesian Optimization. Both architectures work well with balanced accuracy of 93.5% as compared to other state-of-the art algorithms. Although, 1D CNN outperformed with automatic feature extraction [33] and more accurate than clinicians [34], but recurrent neural networks (RNNs) are effective deep learning methods as they can handle time dependencies and variable-length segmented data [35]. Although RNNs face certain difficulties, such as the vanishing gradient problem [36], in which the gradient decreases with backpropagation and its value becomes too small, Consequently, diminished gradient values do not contribute much to learning. The reason for this is that with RNN, the layers that receive a diminished gradient to upgrade weights stop learning; as a result, these layers do not remember and may not be able to recall the longer sequences used. Thus, these layers have a short-term memory, which impacts negatively predicted problems. Nevertheless, this vanishing gradient issue can be resolved by using LSTM or GRU with ReLU, which allows capturing the impact of the earliest available data. Actually, by tuning the burden value, the vanishing gradient problem can be avoided. Altan et al. [25] presented a classification network based on a four-layer deep belief network to classify five kinds of arrhythmias with an accuracy of 94%. Furthermore, they used DFT for feature extraction.
Ping et al. [37] have presented an 8CSL approach for the identification of atrial fibrillation that includes shortcut connections in CNN, which aids in boosting data transmission speed, and 1 layer of LSTM, which aids in reducing long-term dependencies between data. To further test the proposed methodology, he compared it to the RNN and the multi-scale convolutional neural network (MCNN), and he discovered that 8CSL extracted features better when compared to the other two methods in terms of F1-score (84.89%, 89.55%, 85.64%) with different data segment lengths (5 s, 10 s, 20 s). When it comes to heartbeat detections, Ullah et al. [38] used three different algorithms: CNN, CNN+LSTM and CNN+LSTM+attention model for the classification of five different types of arrhythmia in heartbeat detections over two well known databases, MIT-BIH arrhythmias and the PTB Diagnostic ECG Database. They discovered that CNN had 99.12%, CNN+LSTMhad 99.3%, and CNN+LSTM+attention model had 99.29%, all of which were statistically significant. Kang et al. [39] tries to classify mental stress data using the CNN-LSTM model, where he has trained his model using the ST Change Database and WESAD Database, and he has converted the 1D ECG signals into the time and frequency domain in order to train the proposed method. On testing, he achieved an accuracy of 98.3%.
The majority of past work has trained models using a 1D ECG output, which contains a lot of noise, such as baseline wandering effects, power line interference’s, electromyographic noises [40], and artefacts. Filtering and extracting features requires numerous preprocessing processes, which can jeopardise data integrity and model accuracy. Thus, the main aim of this study is to automate the detection and classification process. For that, 1-D ECG signals are translated into 2-D colored scalogram images using the Continuous Wavelet Transform (CWT) with a resolution of 227 × 227 × 3. Thereafter, a 2 D CNN is adopted for automatic feature extraction and an LSTM is adopted for classification purposes. In sum, the following are the paper’s key contributions. (1) We proposed a hybrid approach combining the power of CNN and LSTM. In the initial phase, we converted all the ECG signals to images using CWT, so that the CNN model executes effectively, and then applied LSTM to classify arrhythmias. The approach was developed using the deep learning library on the MATLAB platform. The performance of the proposed approach is validated using the well-known MIT–BIH Arrhythmia database in two ways: firstly, by dividing the dataset into 75% for training and 25% for testing; and secondly, by training the model using K-fold cross validation. Additionally, the parameters of the proposed model are hyper-trained with the grid search algorithm.
Rest of the paper is organised in the following manner: The detailed description of the methodology, including MIT-BIH arrhythmias, dataset preprocessing steps to filter data and conversion of ECG signals into scalogram images, is given in Section 2. Experimental results and performance evaluation are reported in Section 3. Discussion is mentioned in Section 4; Finally, Section 5 concludes the paper.

2. Methodology

This section describes the dataset utilized, data cleaning and preprocessing techniques, and a thorough explanation of the suggested model.

2.1. Dataset

We evaluated the accuracy of our CNN-LSTM model using 162 ECG recordings from three Physionet databases (https://archive.physionet.org/physiobank/database, accessed on 1 February 2021). A description about these is as follows.
  • 96 recordings: taken from the MIT-BIH cardiac arrhythmias database [41,42], this repository contains beat annotation files for 29 long-term ECG recordings of individuals with congestive heart failure ranging in age from 34 to 79. Eight men and two women were among the subjects, while the gender of the remaining 21 was uncertain, and the initial ECG recordings were digitised at 128 samples per second. The sampling frequency of MIT-BIH cardiac arrhythmia is 360 Hz with a resolution of 5 µV/bit.
  • 36 recordings: was taken from the MIT-BIH normal sinus rhythm database, which contains 18 long-term ECG recordings of patients hospitalised at Boston’s BIH arrhythmia laboratory. The patients in this database, which comprised five males aged 26 to 45 and thirteen women aged 20 to 50, had no severe arrhythmias. The MIT-BIH database for normal sinus rhythm is sampled at 128 Hz and the data is accessible at uniform intervals of 7.8125 ms. These signals are digitised using a 12-bit analog-to-digital converter (ADC) at a sampling rate of 128 Hz [41].
  • 30 recordings: taken from the BIDMC congestive heart failure database [41,43]. this repository includes 15 patients’ long-term ECG recordings with severe congestive heart failure (NYHA class 3–4) (11 men, ages 22 to 71, and 4 women, ages 54 to 63). The duration of each recording is about 20 hours, with two ECG signals sampled at 250 samples per second and 12-bit resolution over a ten-millivolt spectrum. At Boston’s Beth Israel Hospital, the first analog recordings were produced using ambulatory ECG recorders with a recording bandwidth of about 0.1 Hz to 40 Hz.
According to [22], the database is structured as an array of two fields: data and labels. Every recording consists of 65,536 samples. As a result, the data is interpreted as a 162 × 65,536 matrix, which means it comprises a total of 162 ECG signals with a sample size of 65,536 and is re-sampled at a standard rate of 128 hertz. In contrast, labels indicate ECG signal information, that is, ARR signals are from 1st to 96th row of an array, 97th to 126th rows of an array represent CHF signals. However, from 127th to 162th row represents NSR signals, as seen in Figure 2.

2.2. Data Preprocessing

This stage is used to prepare data for training and testing. First, the data is segmented by using a transformed data store and the resize data helper function. Continuous Wavelet Transformation (CWT) was also used to turn 1-D ECG signals into 2-D colored scalogram images.

2.2.1. Data Segmentation

Deep Learning models, such as CNN, are dynamic models used for feature extraction and require a lot of data for training the process model. When exceptionally lengthy input signals are sent through the CNN network, the estimated performance may suffer due to the degradation. The ECG signals and their related label masks should be broken up using a modified datastore and the resizeData helper function to prevent these side effects. Our analysis used a dataset from the Physionet databases that included ECG recordings from 162 patients, and each patient recording consisted of 65,536 sequence segments. Here, we have divided 65,536 sequence segments into ten chunks of 500 samples each and discarded the remaining parts of the segment. In our study, we have taken 96 recordings from ARR, 30 recordings from CHF, and 36 recordings from NSR. To make them equally proportional, we used 30 recordings from the NSR database, 30 recordings from the CHF database, and 30 recordings from the ARR database. As a result, there are 900 recordings, and each recording is divided into ten chunks of 500 samples each, which is a lot of data to train CNN for feature extraction and LSTM for classification. Table 1 describes the information in all arrhythmia databases from Physionet.
Remark 1.
In our experiment study, we computed f l o o r 65 , 536 500 = 131 chunks, out of which we chose only 10 chunks. The length of each chunk is 500, which passes through CWT 12 band pass filters and is converted into scalogram images, which represent the ECG signals in the time-frequency domain. Finally, we have 900 scalogram images to train the proposed model with K-fold cross validation.

2.2.2. Image Conversion

The majority of previous work has used a 1D ECG signal to train models, which has a lot of noise [44] such as baseline wandering effect [40], and artifacts. They need many preprocessing steps to filter and extract features, which can compromise data integrity and model accuracy. Thus, in this study, 1D ECG signals are transformed into 2D colored scalogram images using CWT, which are used as input parameters with a resolution of 227 × 227 × 3. As seen in Figure 3, where a normal sinus rhythm signal is converted into a scalogram image using CWT. Application of CWT is denoising and conversions of 1D signals into 2D scalogram images.

2.2.3. Continuous Wavelet Transformation (CWT)

The CWT essentially allows the signals to be mapped into a time-scale domain. It also makes the frequency components in the studied signals more visible. Following previous studies [45,46], this study considers the CWT as a potential candidate solution. Before defining the CWT, let us describe the required machinery setup.
Definition 1
(Wavelet). A wavelet is defined as a limited duration waveform, whose average value is zero. The wavelet function for frequency is defined as:
φ p , q ( t ) = 1 p φ ( t q ) p , p , q R +
Definition 2.
Continuous wavelet transformation (CWT) is the sum of the signal multiplied by shifted and scaled versions of the wavelet function φ. The components of functions φ ( t ) are called mother wavelets and φ p , q ( t ) are called daughter wavelets; and it is obtained by comparing signals by stretching and compressing of mother wavelet at various scales and positions. CWT [47] φ represents the usage of a wavelet to quantify the similarity between a signal and an analysis function. CWT is a technique that is often used to denoise and depict ECG data in the time and frequency domains. as scalogram pictures.
We can obtain CWT coefficients of two variables C p , q of the region R where p is used for scaling purpose and q is used for positioning on x axis, and ∗ denotes the complex conjugate, and t represents the time interval. At lower scales, frequencies are high because waves are less stretched, but as we move to higher scales, the frequency of waves decreases as waves are more stretched at this position. By analyzing the coefficients, more variations can be seen at lowers scales where the frequency is very high, so one can able to capture more varying details of signal f(t), and at higher scales where the waves are less stretched, a minor variation of coefficients can be seen and able to capture less varying details of the signal. The diagrammatical demonstration for CWT coefficient computation is shown in Figure 4.
Remark 2.
In our study, we used 2D scalogram images as input for training and validation of our model, and CWT was used to convert 1D ECG signals into 2D scalograms. We have used Morlet wavelet (i.e., a m o r ) which has a one–sided spectra and has complex values in the time domain. Twelve bandpass filters per octave are used for CWT, as shown in Figure 5.

2.3. Model Architecture and Details

There are several methods to automatically analyse ECG, including machine learning-based [48,49] and deep learning-based [34,50,51]. Deep learning algorithms are more feasible because feature engineering is performed automatically (e.g., extraction of QRS Complex). Here, the mapping of ECG input signals into types of arrhythmias is performed end-to-end. This section explains the complete architectural details to identify three types of arrhythmias (i.e., ARR, CHF, and NSR). First, 1D signals are segmented using Section 2.2.1, and the 2D scalogram images are made using CWT, which helps convert the 1D signals into the time and frequency domain. Then these scalogram images are classified using a structure that is a combination of 2D-CNN and LSTM. The proposed model is trained with K-fold cross-validation, and hyper-parameter tuning is performed through a grid search optimization algorithm. As shown in Figure 6.

2.3.1. Convolutional Neural Network

Convolutional Neural Network (CNN) [52] is a neural network used for enhancing the image or fetching some helpful information out of it, for example, image classification, and by taking 2-D grid features of an image and finding temporal features of time series data by taking 1-D grid samples at different time intervals. The basic architecture and its characteristics are described in [10,53].
The main functions of CNN are max pooling, convolution, classification, and non-linearity. In this analysis, CNN is in charge of extracting temporal features, while LSTM excels at capturing the helpful parameters of time series data and classification. Now, the layers of CNN are in order.
  • 2D convolution layer: is in charge of generating feature maps from 2-D filters strung together.
  • ReLU Activation Function: Activation is an incredibly critical function of a neural network. The responsibility of an activation function is to determine whether the information revived by a neuron is useful or can be overlooked. In this study, we used ReLU as an activation mechanism. It is a non-linear activation function that is used to reduce the linearity of an image by deactivating neurons only when their values are less than zero (explained through (4)).
    Y = R e L U ( W X ) + b ,
    where, X are inputs, W represents weights and b represents bias and Y represents output value.
  • Batch normalization: The parameters in the previous layer may significantly impact the input distribution of the further layer. Batch normalisation is the most important layer since it normalises the last layer’s output and acts as a regularizer to deter the algorithm from overfitting. This effect is referred to as “internal covariate transition”. Batch normalisation estimates the mean and variance of input batches and normalizes, scales, and shifts them. In this work, batch normalisation is performed after the activation. The formula for batch normalisation is calculated as:
    α = 1 n i = 1 n y i ,
    β 2 = 1 n i = 1 n y i α ,
    y ( i ) = y ( i ) α β 2 + ϵ ,
    where: α = mean of same batch; β = variance of same batch; y ( i ) = standardized output value; ϵ = constant value as 0.001.
  • Max pooling: is used for dimensionality reduction or downsampling of input matrices. Max pooling is achieved by adding the maximum filter to non-overlapping sub-regions and selecting the maximum value from each patch.
  • Flattening: a 2-D matrix is transformed into a 1-D vector that can be fed into a completely connected layer by this layer
  • Softmax layer: is used to measure the probability distribution of an occurrence over n separate events in a multi-class grouping. This feature computes the probabilities for each target class in the total classes, ranging from 0 to 1. The probabilities are then used to decide which target class has a higher probability for the given inputs.
  • Classification layer: is used for classifying different categories.
  • Dropout Regularization: generally, the network suffers from over-fitting during the model training [51]. Dropout regularisation solves this problem by discarding some of the function nodes and reducing dependencies between them. We used a 50% dropout before the last completely linked layer in our model.

2.3.2. Long Short Term Memory (LSTM)

Traditional artificial neural networks have limitations in that they are unable to collect the sequential information required for dealing with sequence data in the input data [54,55]. However, in making predictions, RNNs extract sequential information from the input data, that is, the connection between the words in the text. The RNN’s future hidden state is estimated as follows: Let at given time stamp vector t = ( 1 , , T ) , input x = ( x 1 , , x T ) , the output q = ( q 1 , , q T ) , future hidden state vector h = ( h 1 , , h T ) and mathematically determined as:
h t = H ( W x h x t + W h h h t 1 + b h ) ,
where: q t = W h q h t + b q , W : weight matrices, W x h : weight matrix between input and hidden vector, b h : the bias vector for hidden state vectors, H : hidden layer activation function.
The staple concern in conventional RNN is that the back-propagation step is to attenuate the loss function gradient, and its value becomes so tiny that it does not contribute much to learning. As these layers get a slight gradient to upgrade their weights, they stop learning, and this is called the vanishing gradient problem [56]. Thus, they have a short-term memory that has detrimental effects on prediction problems by restricting the network for training in long-term dependencies. Knowing these limitations, we resort to the LSTM that uses cell state memory rather than simple neurons, as shown in Figure 7.
LSTM architecture [57]: consist of following components: Forget Gate, Input Gate and Output Gate. These gates can display or erase information in the cell state memory over random time intervals explained through the following functions [58].
Forget Gate: it allows or discard information. Forget vector at timestamp t is calculated as:
F t = σ ( W x f x t + W h f h t 1 + W c f C t 1 + b f ) ,
where W x f is the weight vector between input and forget gate; x t is the current input value at timestamp t; W h f is the weight vector between hidden state and forget gate; h t 1 is the previous hidden state at time t 1 ; W c f is the weight vector between cell state vector and forget vector; C t 1 is the previous cell state at C t 1 ; b f is the bias vector for setting weights for forget gate. When the summation of all these values is passed through the sigmoid function, if its values are within the range of 0 and 1, the gate allows it to pass; otherwise, it simply discards its information.
Input Gate: Transfer current values and previous values to the sigmoid activation function, which allows you to refresh the cell state memory only when the values are between 0 and 1. Input vector at timestamp t is calculated as:
i t = σ ( W x i x t + W h i h t 1 + W c i C t 1 + b i ) ,
where W x i is the weight vector of input values; W h i is a weight vector between input gate and current values; h t 1 previous hidden state at time t 1 ; W c i is the weight vector between input gate and cell state memory.
Cell State: Sets the current cell state, multiplies the Forget variable with the previous cell state and drop values if multiplied by almost 0, applies the output of the input gate to alter the value of the cell state. Cell state vector at timestamp t is calculated as:
C t = F t C t 1 + i t tanh ( W x c x t + W h c h t 1 + b c ) .
where, W x c is the weight vector between the input vector and the cell state vector; W h c is the weight vector between the current hidden state and the cell state state vector; b c is the bias vector of the current cell state.
Output Gate: For the sake of prediction, it determines what the next hidden state is. Output vector at timestamp t is calculated as:
o t = σ ( W x o x t + W h o h t 1 + W c o C t 1 + b o ) .
Finally, the next hidden state value is calculated by applying the hyperbolic activation function to the current cell state memory and by doing a dot product with the output gate vector:
h t = o t · tanh ( c t )
Remark 3.
Sigmoid activation function, σ ( · ) , is a logistic function its value ranges from 0 1 . It is mainly used for binary classification, formula for sigmoid function: σ = 1 / ( 1 + e z ) . However, tanhis a hyperbolic activation function whose value ranges from −1 to 1, which provides the probabilities distribution for the input vector for multi-classification.

2.3.3. Classification of Arrhythmias Using Proposed (2D-CNN-LSTM) Model

In this section, we present the complete work for the classification of our proposed model 2D-CNN-LSTM through the 20 layers presented in Table 2 and visualised through Figure 8. Here we feed ECG scalogram image sequence data of size 227 × 227 × 3 into sequence input layer. Subsequently, these scalogram images are converted into an array form (vertical, horizontal, channel) using a sequence folding layer. Then these input features are passed through the first convolutional 2-D, which is responsible for creating a feature map of size 227 × 227 × 64 , by using Equation (12), where we have applied 64 filters of size 3 × 3 along with padding and stride as 1, we will get the values of the output data.
( O U T H , O U T W ) = ( h + 2 p f h s + 1 , w + 2 p f w s + 1 )
where: O U T H = output height; O U T W = output weight; h , w = input data size; p = padding size; f h = height of filter; s = no of stride; f w = weight of filter.
Additionally, we introduce non-linearity to the output data using the ReLU function through Equation (2), which has a value in the range of 0-1. It deactivates neurons whose values are less than zero. Further, output data is passed through cross channel normalisation with five channels per element that regularises the meaning and prevents over-fitting of functions.
Next hidden state performs Max, Pooling which is used for the downsampling of the function diagram uses matrix of size [ 2 × 2 ] with stride 2 and padding [ 0000 ] , as a result, the matrix size id reduced to 64 × 64 × 64 . The effects of the downsampling of the feature diagram are transferred through many convolutional and pooling layers, and the matrix is transformed into 32 × 32 × 192 . The input features are then flattened and encoded as 192 × 192 1D vectors, which are fed into the LSTM at layer 11. The flattening layer has the benefit of not altering the parameters by transforming the retrieved feature map output into a 1D array, allowing the feature maps to be rebuilt as input to the LSTM. At this point, the input of size 800 × 4096 is routed via the LSTM’s hidden layer. A weight value of 800 × 4096 is applied to Equations (7)–(10) to extract the feature value, which reflects the LSTM layer’s computational process. In the LSTM layer, there are three types of gates: an input gate ( i t ), a forget gate ( f t ), an output gate ( O t ), and a cell state memory. Using the sigmoid and tanh functions, a weight value is multiplied by an input vector ( x t ), a hidden state ( h t 1 ), and a cell state ( C t ), and then a feature value is retrieved from the result of the multiplication. Subsequently, the feature value computed at the output gate is then sent to the output layer using Equation (11). The procedure of obtaining a needed feature value from multiple feature values generated at the output gate is described in Equation (11). The feature value in the range computed using the output gate is transmitted to the output layer after extracting a feature value from −1 to 1 using the Tanh function. Furthermore, these extracted features are passed through 3 fully connected layers used for classification, where we have applied a 50% dropout. A dropout is a regularisation approach that reduces overfitting in neural networks by removing some random nodes during the training process and improving generalisation error. These values are transferred into a classification layer that functions similar to ANN. Finally, it went into the Softmax activation function responsible for the multi-classification that measures the probability distribution of events over n of different events. This function measures the probability of each target class out of the whole target classes, and this likelihood ranges from 0 to 1. These probabilities help decide the target class with a greater likelihood and classify three different types of arrhythmias, such as ARR, NSR, and CHF. Components of the convolution 2D layer and LSTM layer are shown in Figure 9. Equation (12) depict the convolution 2D layer computation process in CNN models, whereas Equations (7)–(10) depict the process of outputting feature values in the LSTM layer utilising the weight values of the input gate, forget gate, and output gate. The feature values in the range are transferred from the output gate to the output layer using Equation (11).
Hyper Parameter Optimization: Tuning, or hyperparameter optimization, is a term used to describe the process of selecting a set of ideal parameters for machine learning algorithms. Hyperparameters are variables that may influence how a machine learns. For generalising data patterns, similar machine learning algorithms may have different learning rates, weights, and restrictions. Hyperparameters must be fine-tuned to optimally solve the issue. Finding a tuple that offers an ideal model and minimises the loss function is part of the optimization process. Hyperparameter tweaking may be performed in a variety of ways. We used the grid search approach for our research, which entails exhaustive searching of a portion of the algorithm’s hyperparameter space, followed by a performance measure. Data is scanned in the grid searching process to find the best parameters for a specific model. Parameters vary depending on the sort of model under consideration. Grid searching is not limited to a single kind of model; it may be used across the board in machine learning to find the optimum parameters for a given model. Grid searching creates a model based on every potential parameter combination, resulting in numerous iterations. Grid searching is computationally expensive since these model possibilities for each parameter are recorded. Grid search and random search are two typical strategies for tweaking hyperparameters. Every possible combination of hyperparameter values from a predefined list is tested in a grid search, with the best combination being selected based on the cross-validation score. Random combinations are chosen to train the model in a random search, allowing the number of parameters to be regulated. Although it effectively evaluates an extensive range of values and may quickly arrive at a solid combination, it has one fundamental flaw: it cannot guarantee the optimal parameter combination. Grid search, on the other hand, takes a long time yet produces the most significant results.
We employed grid search algorithms to obtain the top three mean test scores, such as 92.35, 83.63, and 78.5, that assist in achieving the maximum accuracy in real-time qualitative assessment, as shown in Table 3. We were able to get the optimal hyperparameters for training models by utilizing the maximum test score of 92.35, which included a learning rate of 0.01, epochs of 20, batch-size of 10, kernel-size of 1, hidden-units of 32, regularisation dropout of 0.05, and optimizer as Adam.
Cost Function: The cost function describes the difference between the given testing sample and the predicted performance and measures how well the neural network is equipped. The cost function is reduced using the optimizer function. Deep learning typically uses a cross-entropy function in several shapes and sizes. Mathematically, the cost function U is defined as:
U = 1 m r log ( s ) + ( 1 r ) log ( 1 s ) ,
where: m = batch-size; s = expected value; r = resultant value.
Using a gradient descent-based optimizer function with a learning rate, the cost function is minimized. Adam [60], Adagrad [61], and Adadelta [62] are some well-known optimizer functions. We found through experiments that the optimum point is quickly achieved when we use Adam. Therefore, we used the Adam optimizer algorithm, which had a learning rate of 1 × 10−4 and a decay rate of 0.95 for 1000 steps.

3. Experimental Results and Performance Evaluation

In this study, experiments were carried out on international standard ECG databases such as MIT-BIH cardiac arrhythmia (ARR), BIDMC congestive heart failure (CHF), and MIT-BIH normal sinus rhythm (NSR), which have detailed expert notation that is frequently utilised in contemporary ECG research. In this experiment, the dataset is divided using K-fold cross validation. Experimental tests were carried out on the workstation with an Intel Xeon gold 5215 10c 2.50 GHz processor and 16GB with training options: (Referenced from Table 3) MiniBatchSize as 10, MaxEpochs as 20, InitialLearnRate as 1 × 10−1, Learn Rate drop period as 3, Gradient Threshold as 1 and the total elapsed time for the training of the model is 6 min 30 s. These parameters are utilised using a grid search hyperparameter optimization algorithm with the highest mean test score of 92.35. Here, segmented 1D ECG signals were transformed into 2D scalogram images of size 227 × 227 × 3 , and trained over the 2DCNN-LSTM model. Experiments were carried out on Matlab 2021a.
To do a qualitative evaluation of the proposed 2D-CNN-LSTM, we utilise six metrics [63] which are described as follows.
  • Sensitivity: determines the capacity of the model to accurately detect the actuality of the cases studied.
  • Specificity: used to determine the ability of the model to distinguish between individual negative samples.
  • Precision: defines the number of patients accurately classified by the model.
  • Recall: quantifies the number of positive class projections made from all positive cases in the data set
  • Accuracy: defines the percentage of correct classifications.
  • F-measure: offers a single score that combines both the accuracy and recall.
The mathematical formulas for the above-mentioned metrics are given in Table 4, where True Positive (TP) represents the number of positive patients who have been assessed as positive for ARR. The True Negative (TN) represents the number of patients who are anticipated to be negative for ARR. Both of these matrices (TP and TN) indicate that the classification is accurate. A False Positive (FP) represents the percentage of patients who are categorised as positive but are negative for ARR. A False Negative (FN) is used to calculate the percentage of positive patients observed to be negative for ARR. However, both of the matrices (FP, FN) indicate a classification error. All three matrices (accuracy, sensitivity, and specificity) indicate the overall classification of the network model. The larger the value, the better the classification results will be.
Here, in this study, in order to evaluate the proposed model, we conducted two experiments. Firstly, the dataset was divided into two portions, 70% for the training dataset and 30% for the testing dataset, and by utilising hyperparameters (Referenced in Table 3) such as kernel size as [3 3]; number of filters as 64; batch-size as 10; regularization dropout as 0.05; optimizer as A d a m , maximum pool Size as [2 2]; loss-method as binary cross entropy; InitialLearnRate as 1 × 10−1; decay as 0.0; epochs as 20 with 32 hidden units and found CNN-LSTM classifies three types of arrhythmias with an accuracy of ARR of 98%, precision of 0.97, recall of 0.96, F1-score of 0.98, sensitivity of 0.96, and specificity of 1, and for CHF class, accuracy of 77%, precision of 0.5, recall of 1, F1-score of 0.66, sensitivity of 1, and specificity of 0.69, and for the NSR class accuracy of 99%, precision of 0.98, recall of 1, F1-score of 0.66, sensitivity of 1, and specificity of 0.98, as shown in Figure 10 and referenced in Table 5. Training accuracy and training loss of the proposed model without K-fold validation are shown in Figure 11. Although the suggested models’ accuracy is acceptable, However, the model has under-fitting and over-fitting issues, which appear when the model has learnt less than or more than 20 epochs. The over-fitting issue model has a tendency to remember data and is unable to generalise new data, while the under-fitting model has a difficult time testing but is capable of generalising new data. We trained our models at 20 epochs and over-optimized parameters using 10-fold cross-validation to eliminate over-fitting and under-fitting, and the accuracy of proposed model was raised, (Referenced from Table 6) where the accuracy of ARR is 98.7%, precision is 1, recall is 0.98, F1-score is 0.98, sensitivity is 0.98, and specificity is 0.98, and for CHF class accuracy is 99%, precision is 0.97, recall is 0.96, F1-score is 0.96, sensitivity is 0.96, and specificity is 0.99 and for the NSR class accuracy is 99%, precision is 0.97, recall is 1, F1-score is 0.98, sensitivity is 0.97, and specificity is 0.99 as shown in Figure 12.

4. Discussion

In this study, an optimised ensemble model was developed using a combination of 2DCNN, which is used for automatic feature extractions, and LSTM, which has additional cell state memory and uses its previous information to predict new data. The goal was to improve the performance of arrhythmia classification while also avoiding overfitting. An optimization approach known as “grid search” was employed to tweak the model’s hyperparameters in order to optimise it. When compared to the random search method, it provides the best hyperparameters, despite the fact that it is computationally expensive. Additional K-fold cross validations were performed on international standard ECG databases such as the MIT-BIH cardiac arrhythmia (ARR), BIDMC congestive heart failure (CHF), and MIT-BIH normal sinus rhythm (NSR) in order to train the model. These databases contain detailed expert notation that is frequently used in contemporary ECG research. As part of our research, we employed the continuous wavelet transform (CWT), which is responsible for turning 1D ECG data into 2D scalogram images of size 227 × 227 × 3, which reflect signals in the time and frequency domain, and which we used for both training and validating the model. A confusion matrix, and other performance metrics were used to assess the classifier’s performance after it was enhanced using layers such as batch normalisation, a flattening layer, and a fully connected layer.
Further, we evaluated two distinct experiments with k-fold cross validation in the presence and absence of dropout regularisation. In Scheme A, dropout regularisation was not applied during the training process. Here, all weights were utilised in the learning process. However, in Scheme B, we applied dropout regularisation with 0.5 dropouts; as a result, 50% of the information was discarded, and only 50% of data is retained for learning. Results of both experimental schemes are visualized through Figure 13.
In the absence of dropout regularization, the results obtained from scheme A indicate high classification due to the overfitting of weights during training. The average accuracy is 99.8%, the average sensitivity is 99.77%, and the average specificity is 99.78%. However, in our suggested model where we have applied 0.5 of dropout regularization, only 50% of information is retained for learning. Hence, our suggested model has an average validation accuracy of 99%, average sensitivity as 99.33%, and average specificity as 98.35%, respectively, (Referenced from Table 7). However, validation accuracy for the classification of ARR is 98.7%, validation accuracy for CHF is 99%, and validation accuracy for NSR is 99% while sensitivity and specificity for ARR is 0.98%, 0.98% while sensitivity and specificity for CHF is 0.96% and 0.99%, and for NSR sensitivity and specificity is 0.97% and 0.99% respectively, (Referenced form Table 6).
Figure 14 represents the training progress and validation accuracy, as well as the training progress and validation loss. As seen in the graphs, training and validation errors were established at a value close to zero after 100 epochs, while training and validation accuracy stabilised at 98.76%, as seen in the graphs. These outcomes are very promising and had a high degree of precision (Referenced from Table 6).
The confusion matrix is derived from the training of a proposed model for classifying three types of arrhythmias. As evidenced by the confusion matrix, the model performs better for the categorization of CHF and NSR than ARR, which is average. This may be due to the slight morphological differences in the waveforms generated during the learning process. The testing dataset’s generated confusion matrix shows 99% accuracy for the normal sinus rhythm, 98.7% for cardiac arrhythmia, and 99% for congestive heart failure, as shown in Figure 12.
To validate our proposed methodology, we compared our results with the current and standard methods in terms of feature extraction methods, methodology, accuracy, and other statistical classifications as presented in Table 8. It is worth mentioning that the difference between the proposed work and the state-of-the-art described is rather promising in terms of accuracy and computing cost compared to other models, as seen in Figure 15. Knowing the potential and prospects of the proposed methodology, it would be interesting to use our proposed approach for diagnosing different critical diseases, such as gastrointestinal diseases, and distinguishing between neoplastic and non-neoplastic tissues.

5. Conclusions

Arrhythmia classification is the most crucial subject in healthcare. An arrhythmia is a rhythm or heart rate irregularity. This paper proposed an approach for the automated study of cardiac arrhythmias using the 2D-CNN-LSTM model. This approach has the following salient features as compared to conventional methods: (i) It employs the CWT to transform 1D ECG signals into 2D Scalogram colored images, making them ideal inputs for this network; (ii) It passes data from a 2D-CNN-LSTM, CNN for feature engineering and LSTM for classification, and has been trained on large labeled images.
Experiments on three ECG cross-databases (obtained under a range of acquisition conditions) showed their usefulness and ability to outperform other approaches in terms of classification performance. The confusion matrix for “normal sinus rhythm,” “cardiac arrhythmias,” and “congestive heart attacks” in the testing dataset showed 99% validation accuracy for “normal sinus rhythm,” 98.7% validation accuracy for “cardiac arrhythmias,” and 99% and validation accuracy for “congestive heart attacks.” Furthermore, sensitivity and specificity for ARR is 0.98%, 0.98% while sensitivity and specificity for CHF are 0.96 and 0.99%, and for NSR, sensitivity and specificity is 0.97% and 0.99% respectively. Likewise, the suggested model will assist clinicians in correctly identifying arrhythmias during routine ECG monitoring. According to preliminary results from the MIT-BIH database, our methodology’s overall efficiency is better than other methods.
Furthermore, the heavy computing burden caused by the use of CWT is a drawback. We could never achieve a complete inter-subject state, even though doing so will significantly minimise the amount of intervention required by doctors. It will be an excellent future avenue for researchers. A robust arrhythmia classification algorithm is needed to address these issues.

6. Further Details

In this section, we present the different functions used in the training of various deep learning models.
ReLU: Activation is an incredibly critical function of a neural network. These activation functions determine if the information neuron receiving is sufficient for the information or can be overlooked. We use ReLU as an activation mechanism in this study.
Batch Normalization: is the most critical layer used to normalize the performance of the previous layer and is often used as a regularization to prevent overfitting the model.
Softmax Function: is used to quantify the probability distribution of occurrences over n distinct events in multi-class classification. This function measures the probability of each target class of the whole target class, and this likelihood range from 0–1. Later, the probabilities help decide the target class with a greater likelihood for the target class.
Sigmoid Activation Function: is a logistic function its value ranges from 0–1 it is mainly used for Binary Classification.
Tanh Activation Function: It is similar to sigmoid activation function and its values ranges from 1 , 1 .
Dropout: is used for regularization, which is used for controlling overfitting problems.
Learning Rate: its value ranges from 0–1 it defines how fast the model adapts the problem and how many weights are modified in the loss gradient model.
Hidden units: are the no of perceptrons used in neural network its values entirely lies on activation function.
Bias: modifies the activation function by applying a constant (i.e., a given bias) to the input. It is analogous to the position of constant in a linear function.

Author Contributions

P.M.: Conceptual frameworks, Methodology, Writing-original version, writing-review, and editing. V.S., A.K., M.D., B.P. and D.P.S.: supervision, writing-review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zoghbi, W.A.; Duncan, T.; Antman, E.; Barbosa, M.; Champagne, B.; Chen, D.; Gamra, H.; Harold, J.G.; Josephson, S.; Komajda, M.; et al. Sustainable development goals and the future of cardiovascular health: A statement from the Global Cardiovascular Disease Taskforce. J. Am. Heart Assoc. 2014, 3, e000504. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Brenyo, A.; Aktas, M.K. Review of complementary and alternative medical treatment of arrhythmias. Am. J. Cardiol. 2014, 113, 897–903. [Google Scholar] [CrossRef]
  3. Martin, C.A.; Matthews, G.D.; Huang, C.L. Sudden cardiac death and inherited channelopathy: The basic electrophysiology of the myocyte and myocardium in ion channel disease. Heart 2012, 98, 536–543. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Zubair, M.; Kim, J.; Yoon, C. An automated ECG beat classification system using convolutional neural networks. In Proceedings of the 2016 6th International Conference on IT Convergence and Security (ICITCS), Prague, Czech Republic, 26 September 2016; pp. 1–5. [Google Scholar]
  5. Oh, S.L.; Ng, E.Y.; San Tan, R.; Acharya, U.R. Automated diagnosis of arrhythmia using combination of CNN and LSTM techniques with variable length heart beats. Comput. Biol. Med. 2018, 102, 278–287. [Google Scholar] [CrossRef] [PubMed]
  6. Mar, T.; Zaunseder, S.; Martínez, J.P.; Llamedo, M.; Poll, R. Optimization of ECG classification by means of feature selection. IEEE Trans. Biomed. Eng. 2011, 58, 2168–2177. [Google Scholar] [CrossRef] [PubMed]
  7. Londhe, A.N.; Atulkar, M. Semantic segmentation of ECG waves using hybrid channel-mix convolutional and bidirectional LSTM. Biomed. Signal Process. Control 2021, 63, 102162. [Google Scholar] [CrossRef]
  8. Ye, C.; Kumar, B.V.; Coimbra, M.T. Heartbeat classification using morphological and dynamic features of ECG signals. IEEE Trans. Biomed. Eng. 2012, 59, 2930–2941. [Google Scholar]
  9. Sahoo, S.; Kanungo, B.; Behera, S.; Sabut, S. Multiresolution wavelet transform based feature extraction and ECG classification to detect cardiac abnormalities. Measurement 2017, 108, 55–66. [Google Scholar] [CrossRef]
  10. Osowski, S.; Hoai, L.T.; Markiewicz, T. Support vector machine-based expert system for reliable heartbeat recognition. IEEE Trans. Biomed. Eng. 2004, 51, 582–589. [Google Scholar] [CrossRef]
  11. Pławiak, P. Novel methodology of cardiac health recognition based on ECG signals and evolutionary-neural system. Expert Syst. Appl. 2018, 92, 334–349. [Google Scholar] [CrossRef]
  12. Yang, S.; Hao, K.; Ding, Y.; Liu, J. Vehicle driving direction control based on compressed network. Int. J. Pattern Recognit. Artif. Intell. 2018, 32, 1850025. [Google Scholar] [CrossRef]
  13. Polat, K.; Güneş, S. Detection of ECG Arrhythmia using a differential expert system approach based on principal component analysis and least square support vector machine. Appl. Math. Comput. 2007, 186, 898–906. [Google Scholar] [CrossRef]
  14. Melgani, F.; Bazi, Y. Classification of electrocardiogram signals with support vector machines and particle swarm optimization. IEEE Trans. Inf. Technol. Biomed. 2008, 12, 667–677. [Google Scholar] [CrossRef] [PubMed]
  15. Nasiri, J.A.; Naghibzadeh, M.; Yazdi, H.S.; Naghibzadeh, B. ECG arrhythmia classification with support vector machines and genetic algorithm. In Proceedings of the 2009 Third UKSim European Symposium on Computer Modeling and Simulation, Athens, Greece, 25–27 November 2009; pp. 187–192. [Google Scholar]
  16. Dutta, S.; Chatterjee, A.; Munshi, S. Correlation technique and least square support vector machine combine for frequency domain based ECG beat classification. Med. Eng. Phys. 2010, 32, 1161–1169. [Google Scholar] [CrossRef] [PubMed]
  17. Desai, U.; Martis, R.J.; Nayak, C.G.; Sarika, K.; Seshikala, G. Machine intelligent diagnosis of ECG for arrhythmia classification using DWT, ICA and SVM techniques. In Proceedings of the 2015 Annual IEEE India Conference (INDICON), New Delhi, India, 17–20 December 2015; pp. 1–4. [Google Scholar]
  18. Kumar, R.G.; Kumaraswamy, Y. Investigating cardiac arrhythmia in ECG using random forest classification. Int. J. Comput. Appl. 2012, 37, 31–34. [Google Scholar]
  19. Park, J.; Lee, K.; Kang, K. Arrhythmia detection from heartbeat using k-nearest neighbor classifier. In Proceedings of the 2013 IEEE International Conference on Bioinformatics and Biomedicine, Shanghai, China, 18–21 December 2013; pp. 15–22. [Google Scholar]
  20. Jun, T.J.; Park, H.J.; Yoo, H.; Kim, Y.H.; Kim, D. GPU based cloud system for high-performance arrhythmia detection with parallel k-NN algorithm. In Proceedings of the 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 16–20 August 2016; pp. 5327–5330. [Google Scholar]
  21. Kiranyaz, S.; Ince, T.; Gabbouj, M. Real-time patient-specific ECG classification by 1-D convolutional neural networks. IEEE Trans. Biomed. Eng. 2015, 63, 664–675. [Google Scholar] [CrossRef]
  22. Acharya, U.R.; Oh, S.L.; Hagiwara, Y.; Tan, J.H.; Adam, M.; Gertych, A.; San Tan, R. A deep convolutional neural network model to classify heartbeats. Comput. Biol. Med. 2017, 89, 389–396. [Google Scholar] [CrossRef]
  23. Acharya, U.R.; Fujita, H.; Oh, S.L.; Hagiwara, Y.; Tan, J.H.; Adam, M. Application of deep convolutional neural network for automated detection of myocardial infarction using ECG signals. Inf. Sci. 2017, 415, 190–198. [Google Scholar] [CrossRef]
  24. Kachuee, M.; Fazeli, S.; Sarrafzadeh, M. Ecg heartbeat classification: A deep transferable representation. In Proceedings of the 2018 IEEE International Conference on Healthcare Informatics (ICHI), New York, NY, USA, 4–7 June 2018; pp. 443–444. [Google Scholar]
  25. Altan, G.; Kutlu, Y.; Allahverdi, N. A multistage deep belief networks application on arrhythmia classification. Int. J. Intell. Syst. Appl. Eng. 2016, 4, 222–228. [Google Scholar] [CrossRef]
  26. Acharya, U.R.; Fujita, H.; Lih, O.S.; Hagiwara, Y.; Tan, J.H.; Adam, M. Automated detection of arrhythmias using different intervals of tachycardia ECG segments with convolutional neural network. Inf. Sci. 2017, 405, 81–90. [Google Scholar] [CrossRef]
  27. Acharya, U.R.; Fujita, H.; Oh, S.L.; Raghavendra, U.; Tan, J.H.; Adam, M.; Gertych, A.; Hagiwara, Y. Automated identification of shockable and non-shockable life-threatening ventricular arrhythmias using convolutional neural network. Future Gener. Comput. Syst. 2018, 79, 952–959. [Google Scholar] [CrossRef]
  28. Xia, Y.; Wulan, N.; Wang, K.; Zhang, H. Detecting atrial fibrillation by deep convolutional neural networks. Comput. Biol. Med. 2018, 93, 84–92. [Google Scholar] [CrossRef] [PubMed]
  29. Savalia, S.; Emamian, V. Cardiac arrhythmia classification by multi-layer perceptron and convolution neural networks. Bioengineering 2018, 5, 35. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  30. Jekova, I.; Krasteva, V. Optimization of end-to-end convolutional neural networks for analysis of out-of-hospital cardiac arrest rhythms during cardiopulmonary resuscitation. Sensors 2021, 21, 4105. [Google Scholar] [CrossRef]
  31. Krasteva, V.; Ménétré, S.; Didon, J.P.; Jekova, I. Fully convolutional deep neural networks with optimized hyperparameters for detection of shockable and non-shockable rhythms. Sensors 2020, 20, 2875. [Google Scholar] [CrossRef]
  32. Elola, A.; Aramendi, E.; Irusta, U.; Picón, A.; Alonso, E.; Owens, P.; Idris, A. Deep neural networks for ECG-based pulse detection during out-of-hospital cardiac arrest. Entropy 2019, 21, 305. [Google Scholar] [CrossRef] [Green Version]
  33. Rajpurkar, P.; Hannun, A.Y.; Haghpanahi, M.; Bourn, C.; Ng, A.Y. Cardiologist-level arrhythmia detection with convolutional neural networks. arXiv 2017, arXiv:1707.01836. [Google Scholar]
  34. Hannun, A.Y.; Rajpurkar, P.; Haghpanahi, M.; Tison, G.H.; Bourn, C.; Turakhia, M.P.; Ng, A.Y. Publisher correction: Cardiologist-level arrhythmia detection and classification in ambulatory electrocardiograms using a deep neural network. Nat. Med. 2019, 25, 530. [Google Scholar] [CrossRef] [Green Version]
  35. Kacprzyk, J. Studies in Computational Intelligence; Springer: Berlin, Germany, 2012; Volume 395. [Google Scholar]
  36. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef] [Green Version]
  37. Ping, Y.; Chen, C.; Wu, L.; Wang, Y.; Shu, M. Automatic detection of atrial fibrillation based on CNN-LSTM and shortcut connection. Healthcare 2020, 8, 139. [Google Scholar] [CrossRef]
  38. Ullah, W.; Siddique, I.; Zulqarnain, R.M.; Alam, M.M.; Ahmad, I.; Raza, U.A. Classification of arrhythmia in heartbeat detection using deep learning. Comput. Intell. Neurosci. 2021, 2021, 2195922. [Google Scholar] [CrossRef] [PubMed]
  39. Kang, M.; Shin, S.; Jung, J.; Kim, Y.T. Classification of mental stress using cnn-lstm algorithms with electrocardiogram signals. J. Healthc. Eng. 2021, 2021, 9951905. [Google Scholar] [CrossRef] [PubMed]
  40. Zeiler, M.D.; Fergus, R. Visualizing and understanding convolutional networks. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2014; pp. 818–833. [Google Scholar]
  41. Goldberger, A.L.; Amaral, L.A.; Glass, L.; Hausdorff, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.K.; Stanley, H.E. PhysioBank, PhysioToolkit, and PhysioNet: Components of a new research resource for complex physiologic signals. Circulation 2000, 101, e215–e220. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  42. Moody, G.B.; Mark, R.G. The impact of the MIT-BIH arrhythmia database. IEEE Eng. Med. Biol. Mag. 2001, 20, 45–50. [Google Scholar] [CrossRef]
  43. Baim, D.S.; Colucci, W.S.; Monrad, E.S.; Smith, H.S.; Wright, R.F.; Lanoue, A.; Gauthier, D.F.; Ransil, B.J.; Grossman, W.; Braunwald, E. Survival of patients with severe congestive heart failure treated with oral milrinone. J. Am. Coll. Cardiol. 1986, 7, 661–670. [Google Scholar] [CrossRef] [Green Version]
  44. Kulkarni, S.S.; Rajankar, S.O. Preprocessing Techniques of Electrocardiogram. Int. J. Eng. Comput. Sci. 2016, 5, 16746–16748. [Google Scholar] [CrossRef]
  45. Priya, K.D.; Rao, G.S.; Rao, P.S. Comparative analysis of wavelet thresholding techniques with wavelet-wiener filter on ECG signal. Procedia Comput. Sci. 2016, 87, 178–183. [Google Scholar] [CrossRef] [Green Version]
  46. Mourad, K.; Fethi, B.R. Efficient automatic detection of QRS complexes in ECG signal based on reverse biorthogonal wavelet decomposition and nonlinear filtering. Measurement 2016, 94, 663–670. [Google Scholar] [CrossRef]
  47. Kim, J.; Min, S.D.; Lee, M. An arrhythmia classification algorithm using a dedicated wavelet adapted to different subjects. Biomed. Eng. Online 2011, 10, 56. [Google Scholar] [CrossRef] [Green Version]
  48. Khalaf, A.F.; Owis, M.I.; Yassine, I.A. A novel technique for cardiac arrhythmia classification using spectral correlation and support vector machines. Expert Syst. Appl. 2015, 42, 8361–8368. [Google Scholar] [CrossRef]
  49. Wang, J.S.; Chiang, W.C.; Hsu, Y.L.; Yang, Y.T.C. ECG arrhythmia classification using a probabilistic neural network with a feature reduction method. Neurocomputing 2013, 116, 38–45. [Google Scholar] [CrossRef]
  50. Andersen, R.S.; Peimankar, A.; Puthusserypady, S. A deep learning approach for real-time detection of atrial fibrillation. Expert Syst. Appl. 2019, 115, 465–473. [Google Scholar] [CrossRef]
  51. Sellami, A.; Hwang, H. A robust deep convolutional neural network with batch-weighted loss for heartbeat classification. Expert Syst. Appl. 2019, 122, 75–84. [Google Scholar] [CrossRef]
  52. Xiang, Y.; Lin, Z.; Meng, J. Automatic QRS complex detection using two-level convolutional neural network. Biomed. Eng. Online 2018, 17, 13. [Google Scholar] [CrossRef] [Green Version]
  53. Avanzato, R.; Beritelli, F. An innovative acoustic rain gauge based on convolutional neural networks. Information 2020, 11, 183. [Google Scholar] [CrossRef] [Green Version]
  54. Zhang, Q. Convolutional neural networks. In Proceedings of the 3rd International Conference on Electromechanical Control Technology and Transportation, Chongqing, China, 19–21 January 2018; pp. 434–439. [Google Scholar]
  55. Graves, A.; Schmidhuber, J. Framewise phoneme classification with bidirectional LSTM networks. In Proceedings of the 2005 IEEE International Joint Conference on Neural Networks, Montreal, QC, Canada, 31 July–4 August 2005; Volume 4, pp. 2047–2052. [Google Scholar]
  56. Li, X.; Wu, X. Long short-term memory based convolutional recurrent neural networks for large vocabulary speech recognition. arXiv 2016, arXiv:1610.03165. [Google Scholar]
  57. Amin, J.; Sharif, M.; Raza, M.; Saba, T.; Sial, R.; Shad, S.A. Brain tumor detection: A long short-term memory (LSTM)-based learning model. Neural Comput. Appl. 2020, 32, 15965–15973. [Google Scholar] [CrossRef]
  58. Gers, F.A.; Schraudolph, N.N.; Schmidhuber, J. Learning precise timing with LSTM recurrent networks. J. Mach. Learn. Res. 2002, 3, 115–143. [Google Scholar]
  59. Tanabe, K.; Ikeda, M.; Hayashi, M.; Matsuo, K.; Yasaka, M.; Machida, H.; Shida, M.; Katahira, T.; Imanishi, T.; Hirasawa, T.; et al. Comprehensive Serum Glycopeptide Spectra Analysis Combined with Artificial Intelligence (CSGSA-AI) to Diagnose Early-Stage Ovarian Cancer. Cancers 2020, 12, 2373. [Google Scholar] [CrossRef]
  60. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  61. Duchi, J.; Hazan, E.; Singer, Y. Adaptive subgradient methods for online learning and stochastic optimization. J. Mach. Learn. Res. 2011, 12, 2121–2159. [Google Scholar]
  62. Zeiler, M.D. Adadelta: An adaptive learning rate method. arXiv 2012, arXiv:1212.5701. [Google Scholar]
  63. Bengio, Y.; Simard, P.; Frasconi, P. Learning long-term dependencies with gradient descent is difficult. IEEE Trans. Neural Netw. 1994, 5, 157–166. [Google Scholar] [CrossRef] [PubMed]
  64. Zheng, Z.; Chen, Z.; Hu, F.; Zhu, J.; Tang, Q.; Liang, Y. An automatic diagnosis of arrhythmias using a combination of CNN and LSTM technology. Electronics 2020, 9, 121. [Google Scholar] [CrossRef] [Green Version]
  65. Atal, D.K.; Singh, M. Arrhythmia classification with ECG signals based on the optimization-enabled deep convolutional neural network. Comput. Methods Programs Biomed. 2020, 196, 105607. [Google Scholar] [CrossRef]
  66. Faust, O.; Acharya, U.R. Automated classification of five arrhythmias and normal sinus rhythm based on RR interval signals. Expert Syst. Appl. 2021, 181, 115031. [Google Scholar] [CrossRef]
  67. Lu, Y.; Jiang, M.; Wei, L.; Zhang, J.; Wang, Z.; Wei, B.; Xia, L. Automated arrhythmia classification using depthwise separable convolutional neural network with focal loss. Biomed. Signal Process. Control 2021, 69, 102843. [Google Scholar] [CrossRef]
  68. Wang, G.; Chen, M.; Ding, Z.; Li, J.; Yang, H.; Zhang, P. Inter-patient ECG arrhythmia heartbeat classification based on unsupervised domain adaptation. Neurocomputing 2021, 454, 339–349. [Google Scholar] [CrossRef]
  69. Mathunjwa, B.M.; Lin, Y.T.; Lin, C.H.; Abbod, M.F.; Shieh, J.S. ECG arrhythmia classification by using a recurrence plot and convolutional neural network. Biomed. Signal Process. Control 2021, 64, 102262. [Google Scholar] [CrossRef]
  70. Yang, X.; Zhang, X.; Yang, M.; Zhang, L. 12-Lead ECG arrhythmia classification using cascaded convolutional neural network and expert feature. J. Electrocardiol. 2021, 67, 56–62. [Google Scholar] [CrossRef]
  71. Rahul, J.; Sora, M.; Sharma, L.D.; Bohat, V.K. An improved cardiac arrhythmia classification using an RR interval-based approach. Biocybern. Biomed. Eng. 2021, 41, 656–666. [Google Scholar] [CrossRef]
  72. Chouhan, K.S.; Gajrani, J.; Sharma, B.; Tazi, S.N. Arrhythmia Classification Using Deep Learning Architecture. In Real-Time Applications of Machine Learning in Cyber-Physical Systems; IGI Global: Hershey, PA, USA, 2022; pp. 148–172. [Google Scholar]
Figure 1. Representation of complete electrical activity of the heart.
Figure 1. Representation of complete electrical activity of the heart.
Bioengineering 09 00152 g001
Figure 2. Distribution of ECG waveforms from three classes of arrhythmias. Sub-figures (a,b) Represents ARR signals and Scalogram image of ARR signal using CWT; Sub-figures (c,d) Represents CHF signals and Scalogram image of CHF signal using CWT; Sub-figures (e,f) Represents NSR signals and Scalogram image of NSR signal using CWT.
Figure 2. Distribution of ECG waveforms from three classes of arrhythmias. Sub-figures (a,b) Represents ARR signals and Scalogram image of ARR signal using CWT; Sub-figures (c,d) Represents CHF signals and Scalogram image of CHF signal using CWT; Sub-figures (e,f) Represents NSR signals and Scalogram image of NSR signal using CWT.
Bioengineering 09 00152 g002
Figure 3. Conversion of normal sinus rhythm (NSR) signal into scalogram image. Sub-figures (a,b) represents NSR signals and CWT scalogram for NSR.
Figure 3. Conversion of normal sinus rhythm (NSR) signal into scalogram image. Sub-figures (a,b) represents NSR signals and CWT scalogram for NSR.
Bioengineering 09 00152 g003
Figure 4. Different coefficient representation of Scalogram.
Figure 4. Different coefficient representation of Scalogram.
Bioengineering 09 00152 g004
Figure 5. Conversions of 1D ECG signals into Scalogram images using CWT.
Figure 5. Conversions of 1D ECG signals into Scalogram images using CWT.
Bioengineering 09 00152 g005
Figure 6. Block-diagram of the proposed method.
Figure 6. Block-diagram of the proposed method.
Bioengineering 09 00152 g006
Figure 7. Complete representation of Long Short Term Memory Cell.
Figure 7. Complete representation of Long Short Term Memory Cell.
Bioengineering 09 00152 g007
Figure 8. Representation of the internal architectural details of the proposed model.
Figure 8. Representation of the internal architectural details of the proposed model.
Bioengineering 09 00152 g008
Figure 9. Internal components of the proposed 2D-CNN-LSTM model.
Figure 9. Internal components of the proposed 2D-CNN-LSTM model.
Bioengineering 09 00152 g009
Figure 10. Confusion Matrix represents accuracy of the proposed model without k-fold cross validation 2D-CNN-LSTM.
Figure 10. Confusion Matrix represents accuracy of the proposed model without k-fold cross validation 2D-CNN-LSTM.
Bioengineering 09 00152 g010
Figure 11. (a) Training progress and validation accuracy of proposed model without K-fold cross validation (b) training and progress loss of proposed model without K-fold cross validation.
Figure 11. (a) Training progress and validation accuracy of proposed model without K-fold cross validation (b) training and progress loss of proposed model without K-fold cross validation.
Bioengineering 09 00152 g011
Figure 12. Confusion Matrix represents validation accuracy of proposed model with k-fold cross validation 2D-CNN-LSTM.
Figure 12. Confusion Matrix represents validation accuracy of proposed model with k-fold cross validation 2D-CNN-LSTM.
Bioengineering 09 00152 g012
Figure 13. (a) Training progress and validation accuracy of proposed model with dropout regularization (b) training and progress accuracy of proposed model without dropout regularization.
Figure 13. (a) Training progress and validation accuracy of proposed model with dropout regularization (b) training and progress accuracy of proposed model without dropout regularization.
Bioengineering 09 00152 g013
Figure 14. (a) Training progress and validation accuracy of proposed model using dropout regularization (b) Training progress and validation loss of proposed model using dropout regularization.
Figure 14. (a) Training progress and validation accuracy of proposed model using dropout regularization (b) Training progress and validation loss of proposed model using dropout regularization.
Bioengineering 09 00152 g014
Figure 15. Comparative analysis of the proposed model with the existing sate-of-the-art methods.
Figure 15. Comparative analysis of the proposed model with the existing sate-of-the-art methods.
Bioengineering 09 00152 g015
Table 1. Description of all arrhythmia databases from Physionet.
Table 1. Description of all arrhythmia databases from Physionet.
Database TypeNo of RecordingNo of SamplesSampling Rate
MIT-BIH Cardiac Arrhythmias (ARR)9696 × 65,536128
BIDMC congestive heart failure (CHF)3030 × 65,536250
MIT-BIH Normal Sinus Rhythm (NSR)3636 × 65,536128
Table 2. Representation of Complete Layer Architecture of Proposed Model [59].
Table 2. Representation of Complete Layer Architecture of Proposed Model [59].
No. of LayersNameKernel-SizeFilter-StrideKernel-TypeNo of Neurons
1Input image--- 227 × 227 × 3
2Sequence folding layer--- 227 × 227 × 1
3Conv2D+ReLU+Normalization[3 3]164 227 × 227 × 1
4Pooling[2 2]2- 227 × 227 × 64
5Conv2D+ReLU+Normalization[1 1]1128 64 × 64 × 64
6Pooling+padding[2 2], [1 1 1 1]2128 64 × 64 × 128
7Conv2D’[3 3]1256 32 × 32 × 128
8Pooling[2 2]2- 32 × 32 × 256
9Sequence unfolding layer--- 32 × 32 × 256
10flattening--- 192 × 192
11LSTM---4096
12Fully Connected---4096
13ReLU----
14dropout----
15Fully Connected---4096 fully connected layer
16ReLU----
17Dropout---50% dropout
18Fully Connected---1000 fully connected layer
19Softmax----
20Classification Output----
Table 3. Representation of optimised hyperparameters with different test-scores using Grid Search Optimization.
Table 3. Representation of optimised hyperparameters with different test-scores using Grid Search Optimization.
Learn-RateBatch-SizeHidden-UnitsEpochsMean-Test-Score
0.0110322092.35
0.0464645083.63
0.0532323078.5
Table 4. Accuracy Measures.
Table 4. Accuracy Measures.
MeasuresFormula
Accuracy (A) A = ( T P + T N ) Total no of samples
Precision (P) P = T P / ( T P + F P )
Recall (R) R = ( T P T P + F N )
F-Measure F = 2 ( P R ) ( P + R )
Sensitivity T P ( T P + F N )
Specificity T N ( T N + F P )
Table 5. The table represents overall accuracy, TP, TN, FP, FN, F1, Sensitivity (Se), and specificity (Sp) scores for the 2DCNN-LSTM model without k-fold validation.
Table 5. The table represents overall accuracy, TP, TN, FP, FN, F1, Sensitivity (Se), and specificity (Sp) scores for the 2DCNN-LSTM model without k-fold validation.
S. NoClassAccu.TPTNFPFNPrec.RecallF1 ScoreSeSp
1ARR98%62100020.970.9660.980.961
2CHF77%501135000.510.6610.69
3NSR99%50113100.9810.6610.98
Table 6. The table represents overall accuracy, TP, TN, FP, FN, F1, Sensitivity (Se), and specificity (Sp) score for 2DCNN-LSTM model with k-fold validation.
Table 6. The table represents overall accuracy, TP, TN, FP, FN, F1, Sensitivity (Se), and specificity (Sp) score for 2DCNN-LSTM model with k-fold validation.
S. NoClassAccu.TPTNFPFNPrec.RecallF1 ScoreSeSp
1ARR98.7%65982110.980.980.980.98
2CHF99%48115120.970.960.960.960.99
3NSR99%48115110.9710.980.970.99
Table 7. Average performance analysis of two experiments for the classification of Arrhythmias.
Table 7. Average performance analysis of two experiments for the classification of Arrhythmias.
ExperimentsSchemesAvg-AccuracyAvg-SpecificityAvg-Sensitivity
Awithout dropout regularization99.8%99.78%99.77%
Bwith dropout regularization99%98.35%98.33%
Table 8. Comparison table of the proposed architecture with some other existing architectures in terms of feature extraction (FE) methods, methodology, accuracy, and other statistical classification.
Table 8. Comparison table of the proposed architecture with some other existing architectures in terms of feature extraction (FE) methods, methodology, accuracy, and other statistical classification.
AuthorFEDatabaseModelACCTPRTNRFPRFNRSensitivitySpecificity
Acharya et al. [26]RPMIT-BIH arrhythmia database11-layer CNN96%96953195.49%94.19%
Zengh et al. [64]CNNMIT-BIH arrhythmia databaseCNN-LSTM98%10111297.87%98.57%
Dinesh et al. [65]Wavelet-intervalsMIT-BIH Arrhythmia Databaseoptimized CNN93.19%10111293.98%95%
Oliver et al. [66]RR-intervalsECG signal Database (Shaoxing Hospital Zhejiang University School of Medicine)CNN93.19%12912195%94.30%
Lu et al. [67]CNNMIT-BIH Arrhythmia DatabaseDepth wise seprable CNN with focal loss98.55%10121282%79%
Wang et al. [68]CNNMIT-BIH, INCARTDB and SVDBDepth wise seprable CNN with focal loss97.40%11122196.7%97.8%
Bhekumuzi et al. [69]2d images using 2D recurrent plotsMIT-BIH Atrial Fibrillation Database, MIT-BIH Malignant Ventricular Ectopy DatabasCNN95.30%10112194.2%95%
Yang et al. [70]CNNfirst China ECG Intelligence Challengecascaded CNN86.50%1392185.3%82%
Rahul et al. [71]RR-intervalsIT-BIH Arrhythmia DatabaseSVM99.51%10111299.28%99.63%
Chouhan et al. [72]CNNIT-BIH Arrhythmia DatabaseCNN97.16%501001099.28%99.63%
Ullah et al. [38]CNNIT-BIH Arrhythmia DatabaseCNN97.38%6389102%95.63%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Madan, P.; Singh, V.; Singh, D.P.; Diwakar, M.; Pant, B.; Kishor, A. A Hybrid Deep Learning Approach for ECG-Based Arrhythmia Classification. Bioengineering 2022, 9, 152. https://doi.org/10.3390/bioengineering9040152

AMA Style

Madan P, Singh V, Singh DP, Diwakar M, Pant B, Kishor A. A Hybrid Deep Learning Approach for ECG-Based Arrhythmia Classification. Bioengineering. 2022; 9(4):152. https://doi.org/10.3390/bioengineering9040152

Chicago/Turabian Style

Madan, Parul, Vijay Singh, Devesh Pratap Singh, Manoj Diwakar, Bhaskar Pant, and Avadh Kishor. 2022. "A Hybrid Deep Learning Approach for ECG-Based Arrhythmia Classification" Bioengineering 9, no. 4: 152. https://doi.org/10.3390/bioengineering9040152

APA Style

Madan, P., Singh, V., Singh, D. P., Diwakar, M., Pant, B., & Kishor, A. (2022). A Hybrid Deep Learning Approach for ECG-Based Arrhythmia Classification. Bioengineering, 9(4), 152. https://doi.org/10.3390/bioengineering9040152

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop