Detection of Adversarial Attacks against the Hybrid Convolutional Long Short-Term Memory Deep Learning Technique for Healthcare Monitoring Applications

: Deep learning (DL) models are frequently employed to extract valuable features from heterogeneous and high-dimensional healthcare data, which are used to keep track of patient well-being via healthcare monitoring systems. Essentially, the training and testing data for such models are collected by huge IoT devices that may contain noise (e


Introduction
The advancement of Internet of Things (IoT) technologies has led to the widespread penetration and large-scale deployment of IoT systems worldwide.While IoT systems are eminently qualified for providing intelligent services, the massive amounts of data collected and processed by IoT systems also raise serious security concerns.Therefore, research efforts have been focused on designing intelligent anomaly detection systems to prevent the misuse of IoT data across smart applications [1].
In the healthcare context, IoT has improved the standard of care accorded to patients.Indeed, people can enjoy life with greater convenience since these systems ensure their health and safety through continuous monitoring.In addition, the IoT supports many healthcare applications, from attached medical sensors to wireless body area networks (WBANs).A WBAN comprises a network of small, wearable devices and is considered the most promising technology for enhancing healthcare services.Such devices have enabled remote monitoring to increase the overall goodness of care provided to patients in remote areas or medical facilities [2,3].
In spite of these advantages, WBANs are also susceptible to external threats since sensor data are gathered from various sources including people and locations.Attackers with malicious motives may target the sensors and insert malicious data that report back anomalous observations, resulting in inaccurate diagnoses, wrong medication being given Appl.Sci.2023, 13, 6807 2 of 17 to patients, and significant financial losses for healthcare entities using this healthcare system [4,5].Healthcare systems now face a serious problem with anomalies, which can also result from faulty sensors and inaccurate observations from sensing devices.
Anomaly detection is considered one of the best solutions for WBANs in terms of distinguishing abnormal data and may provide a reliable system to counter sensor faults and anomalous activities, along with an understanding of the factors affecting patients and healthcare organizations.Machine learning and statistical techniques have been applied in various studies to detect anomalies over the last few years.Different researchers have studied the employment of these techniques to detect anomalies in WBANs, and their findings support their effectiveness.However, these solutions still have some limitations in providing high accuracy within a limited timeframe without human intervention, in terms of the engineering of features.To overcome this challenge, deep learning is now being used to enhance the performance of anomaly detection models.
Deep learning-based anomaly detection has great potential for increased security, due to its ability to detect anomalous behavior across data sources that cannot be identified using traditional security methods.However, the vulnerability of deep learning models to adversarial attacks is a fundamental obstacle to employing deep learning for security applications [4].An adversarial attack occurs when adversarial attack data are fed as input into a deep learning model.One example of such an attack is a dataset in which some features have been purposely perturbed to confuse the deep learning model, in order to produce an incorrect prediction [6].
In this regard, the current work investigates the ability of a deep learning model employing a hybrid convolutional long short-term memory (ConvLSTM) approach to identify adversarial attacks in healthcare monitoring systems.The following is a summary of this paper's most significant contributions: 1.
Developing a model for detecting anomalies, utilizing the hybrid ConvLSTM technique, to detect adversarial attacks in WBANs.The technique is used to consider an attack in the training and testing phases and shows how adversarial attacks can potentially mislead the anomaly detection model.

2.
Evaluating the proposed model when under attack in the training phase, along with both fast gradient sign (FGSM) and basic iterative (BIM) attacks in the testing phase.

3.
Developing a proactive method as a countermeasure that will act as a retraining process to strengthen the model itself against adversarial attacks.
The remainder of the paper is organized as follows: Section 2 investigates works in the literature related to WBAN anomaly detection systems and adversarial attacks and includes a discussion of the various points.The suggested model's design and its specific elements are described in Section 3. Section 4 provides an analysis of the findings and the outcomes of the experiment.The authors' conclusions are presented in Section 5, along with some recommendations for further research.

Related Works
Increasing emphasis is being placed on anomaly detection in the Internet of Things arena, particularly regarding healthcare systems that create vast amounts of data through WBANs.In the existing studies, numerous models for detecting anomalies in WBAN-based systems using various methods have been suggested, and these will be analyzed in the following paragraphs.
The authors of Ref. [4] assessed the WBAN's dependability in terms of identifying anomalies, assessing human health conditions, hardware failure rates, and transient fault correction mechanisms.The study proposed a measure using the mean time to failure (MTTF), which provided improved performance when assessing the dependability of WBANs in terms of specification and detection of anomalies.The findings indicated a detection rate of 95% and an MTTF of 43.01 s, which is considered to be short.Even though the study indicated a high level of detection as a measure of dependability, the metric must be altered to account for some irregularities.
Another study [5] proposed a methodology for detecting anomalous sensor data.Their approach was built on concurrent fuzzy clustering and data compression using the Hadoop MapReduce model.The experimental results demonstrated that the suggested framework attained better accuracy with the fewest false alarms, achieving between 97% and 98% accuracy.This study employed a parametric statistical method that is computationally intensive.
A different method for recognizing changes in WBAN-collected data was proposed in [7] using the Kalman filter approach.The researchers claimed that this approach can identify any physiological reaction in WBANs.However, the Kalman filter approach has considerable disadvantages, including a greater computational cost to achieve optimal results.
The authors of [8] proposed a unique method for detecting WBAN anomalies, which was based on Gaussian regression and majority voting.Using a real dataset, the proposed method presented a strategy that was capable of differentiating between true medical emergency situations and false alarms.This technique is successful in terms of detection and the false-positive rate, as was demonstrated by the findings.However, this method was hindered by its computing complexity, high false-alarm rate, and sample size.Elsewhere, the authors of [9] proposed enhancing the performance of anomaly detection by using a correlation method for multiple body sensors.The proposed method employed thresholds to detect anomalies.However, in this study, only one kind of correlation that utilized the spatial relationship among sensors was examined, while disregarding the temporal correlation for each sensor reading.
Two additional studies [10,11] examined the unreliability of certain sensors that are responsible for generating many false alarms in medical-based systems.Both investigations used a dynamic sliding window and a weighted moving average to identify abnormally flawed sensor values.However, researchers using the weighted moving average method need to pay more attention to the complex relationships within the data.
The authors of [12] proposed an anomaly detection algorithm for WBANs to eliminate the false alarms resulting from faulty measurements.The strategy utilized both spatial-temporal correlation and a game theory approach.Additionally, the local processing unit of the proposed design utilized the Mahalanobis distance for multivariate analysis.The suggested method demonstrated improved efficacy in producing a reduced false alarm rate and excellent detection precision.One potential flaw of the game-theoretic method is that it is necessary to account for novel anomalies.The experimental findings revealed that the suggested technique offered the quickest execution time and greatest energy efficiency among the sensors.However, this technique is incapable of detecting random changes in various physiological signals.
A study conducted by the authors of [13] evaluated the performance of three intrusion detection models, based on CNN, LSTM, and the gated recurrent unit (GRU), by applying adversarial examples such as the fast gradient sign method (FGSM) to test the robustness of the three models.The experimental outcomes showed that CNN is the model best able to withstand such adversarial examples.The robustness of GRU and LSTM against adversarial examples can be significantly increased after adversarial training.
The authors of [14] suggested an anomaly detection system for a WBAN-based data sampling technique with a modified cumulative sum as the foundation for an anomaly detection system for use with WBAN (known as MCUSUM).The MCUSUM algorithm was used to reliably detect abnormalities, while the sampling technique was used to boost the detection speed.The findings of this study demonstrated that the suggested technique offered the highest energy efficiency and shortest execution time when using the sensors.However, the anomaly detection system was unable to identify random abnormalities in the different sets of physiological data.
In [15], the authors designed a novel adversarial attack for testing DL-based network intrusion detection systems.The study presented two techniques: the first is a model extraction, while the second uses a saliency map.With these techniques, the attack model was successfully compromised, and the malicious packets had an attack success rate of 94.31%, on average.
The research presented in [16] contrasted the deep autoencoder (DAE), the shallow autoencoder (SAE), and the ensemble of autoencoders, three machine learning models that are utilized for anomaly detection.The study assessed the models' resilience in the face of data poisoning attacks (DPAs).The evaluation results showed an F1 score of ≈97% when handling unpoisoned benign traffic.However, when challenged by a DPA, DAE demonstrated more robust detection capabilities, providing over 50% of the F1 score, with 10% poisoning.The other models, however, exhibited a significantly declining performance (down to a 20% F1 Score), with only 0.5% of the malicious traffic being injected into the data.
The authors of [17] proposed a model agnostic explainability-based method for accurately detecting adversarial attacks on two datasets.The study obtained an accuracy of 88% on the MIMIC-CXR dataset, significantly exceeding the state-of-the-art system of adversarial detection in both datasets by over 10% in all scenarios.The study showed a detection accuracy of 77% against a longitudinal adversarial attack.
In [18], the authors proposed a framework designed to detect poisoning attacks, utilizing deep neural networks and support vector machines.The authors evaluated the framework using different state-of-the-art data poisoning attacks for several healthcare applications: human activity recognition and electrocardiograph classification.The experimental analysis shows that the proposed framework could efficiently detect poisoning attacks and remove the specified poisoned updates from global data aggregation.
The authors of [19] presented a new kind of adversarial approach that would take advantage of the ML classifiers employed in a smart healthcare system (SHS).The study manipulated readings from medical devices to alter the patients' status.The test findings showed that the suggested adversarial approach might severely impair the ability of an MLbased SHS to accurately detect the patients' normal activities, which would ultimately result in incorrect therapy.However, machine learning methods may have other disadvantages since the outcome is dependent on the input dataset.
As reported in [20], a cognitive machine learning-assisted attack detection framework (CML-ADF) has been developed to transmit healthcare data securely.The suggested framework achieves a 96.5% attack prediction ratio, a 97.8% efficiency ratio, a 98.2% accuracy ratio, a 21.3% reduction in latency, and an 18.9% reduction in communication cost, according to the experimental data.Elsewhere, the authors of [21] proposed a model that combined the capabilities of hybrid convolutional long short-term memory (ConvLSTM) methods with correlations in the various physiological data intended to detect contextual and datapoint abnormalities in the massive WBAN datastream.The results of the experiments using the proposed model reported an average accuracy of 99% and a 98% F1-measure on various dataset subjects, compared to 64% for the CNN and LSTM when used separately.However, that study did not examine adversarial attacks.As a result, this paper will extend their contribution to the literature in order to investigate the ability of the model regarding adversarial attack detection using deep learning techniques.Furthermore, it suggests suitable countermeasures for enhancing the ability of the model against the adversarial examples introduced in the training or testing phases.
Table 1 summarizes the existing studies and provides an analysis in terms of the mechanism for anomaly detection, dealing with adversarial attacks, and proposing a defense method.

Study Providing Anomaly Detection
Dealing with Adversarial Attacks Proposing Defense Metho An investigation of the existing literature shows that in order to increase the effica of anomaly detection strategies in WBANs, several methodologies, including statisti and ML techniques, have been developed by researchers.However, these methods ha some drawbacks.For instance, statistical approaches cannot handle the dynamic natu of WBANs, making it challenging to identify a suitable evaluation threshold value.In dition, due to their processing complexity, non-parametric statistical methods are unsu able for real-time applications.Numerous contemporary methodologies have employ ML algorithms, including decision trees, linear regression, ANN, and the k-nearest nei bor and random forest algorithms.In a sensitive domain that needs great accuracy a good performance, such as in healthcare, better solutions may be found than the depl ment of machine learning.These algorithms have drawbacks when dealing with comp cated and large datasets, which slow down processing, and are not effective in anticip ing new anomalous patterns.
Deep learning is recommended by the research community due to its better perf mance; it reduces the necessity of handcrafted feature engineering and increases perf mance compared with traditional machine learning.Evaluating the model's performan is considered the most significant step in deep learning development.However, advers ial attacks have recently been compromising its effectiveness and have led to misclass cation in deep learning models, affecting the model's performance.After reviewing published literature, it is clear that most studies have not fully investigated the scena of adversarial attacks, such as in [4,5,7].Although the machine and deep learning te niques used will encounter several types of possible security concerns in the vario phases, one of the most pressing concerns is the adversarial attack, wherein the system adversaries are highly motivated to change the outcomes of models or to source perso data for their own gain.For example, attackers can cause misclassification by manipu ing the data samples.In addition, few studies have investigated adversarial attacks, su as in [17,18,20], but the authors have not proposed any defensive method by which protect the models.Defense methods in intelligent networks aim to distinguish malicio activity from the common patterns in intelligent networks.To mitigate the different va ants of adversarial attack, these defense methods must assure the security of the data t have been gathered by the networked devices to combat various adversarial attacks.So of the existing studies propose defensive methods, such as those reported in [

Study
Providing Anomaly Detection Dealing with Adversarial Attacks Prop ✓ ✕ An investigation of the existing literature shows that in ord of anomaly detection strategies in WBANs, several methodolo and ML techniques, have been developed by researchers.Howe some drawbacks.For instance, statistical approaches cannot ha of WBANs, making it challenging to identify a suitable evaluatio dition, due to their processing complexity, non-parametric statis able for real-time applications.Numerous contemporary metho ML algorithms, including decision trees, linear regression, ANN bor and random forest algorithms.In a sensitive domain that n good performance, such as in healthcare, better solutions may b ment of machine learning.These algorithms have drawbacks w cated and large datasets, which slow down processing, and are ing new anomalous patterns.
Deep learning is recommended by the research communit mance; it reduces the necessity of handcrafted feature engineer mance compared with traditional machine learning.Evaluating is considered the most significant step in deep learning developm ial attacks have recently been compromising its effectiveness an cation in deep learning models, affecting the model's performa published literature, it is clear that most studies have not fully of adversarial attacks, such as in [4,5,7].Although the machine niques used will encounter several types of possible security phases, one of the most pressing concerns is the adversarial atta adversaries are highly motivated to change the outcomes of mod data for their own gain.For example, attackers can cause miscla ing the data samples.In addition, few studies have investigated as in [17,18,20], but the authors have not proposed any defens protect the models.Defense methods in intelligent networks aim activity from the common patterns in intelligent networks.To m ants of adversarial attack, these defense methods must assure th have been gathered by the networked devices to combat various of the existing studies propose defensive methods, such as t [5] Appl.Sci.2023, 13 An investigation of the existing literature shows that in order to increase the effica of anomaly detection strategies in WBANs, several methodologies, including statisti and ML techniques, have been developed by researchers.However, these methods ha some drawbacks.For instance, statistical approaches cannot handle the dynamic natu of WBANs, making it challenging to identify a suitable evaluation threshold value.In dition, due to their processing complexity, non-parametric statistical methods are unsu able for real-time applications.Numerous contemporary methodologies have employ ML algorithms, including decision trees, linear regression, ANN, and the k-nearest nei bor and random forest algorithms.In a sensitive domain that needs great accuracy a good performance, such as in healthcare, better solutions may be found than the depl ment of machine learning.These algorithms have drawbacks when dealing with comp cated and large datasets, which slow down processing, and are not effective in anticip ing new anomalous patterns.
Deep learning is recommended by the research community due to its better perf mance; it reduces the necessity of handcrafted feature engineering and increases perf mance compared with traditional machine learning.Evaluating the model's performan is considered the most significant step in deep learning development.However, advers ial attacks have recently been compromising its effectiveness and have led to misclass cation in deep learning models, affecting the model's performance.After reviewing published literature, it is clear that most studies have not fully investigated the scena of adversarial attacks, such as in [4,5,7].Although the machine and deep learning te niques used will encounter several types of possible security concerns in the vario phases, one of the most pressing concerns is the adversarial attack, wherein the system adversaries are highly motivated to change the outcomes of models or to source perso data for their own gain.For example, attackers can cause misclassification by manipu ing the data samples.In addition, few studies have investigated adversarial attacks, su as in [17,18,20], but the authors have not proposed any defensive method by which protect the models.Defense methods in intelligent networks aim to distinguish malicio activity from the common patterns in intelligent networks.To mitigate the different va ants of adversarial attack, these defense methods must assure the security of the data t have been gathered by the networked devices to combat various adversarial attacks.So of the existing studies propose defensive methods, such as those reported in [

Study
Providing Anomaly Detection Dealing with Adversarial Attacks Prop ✓ ✕ An investigation of the existing literature shows that in ord of anomaly detection strategies in WBANs, several methodolo and ML techniques, have been developed by researchers.Howe some drawbacks.For instance, statistical approaches cannot ha of WBANs, making it challenging to identify a suitable evaluatio dition, due to their processing complexity, non-parametric statis able for real-time applications.Numerous contemporary metho ML algorithms, including decision trees, linear regression, ANN bor and random forest algorithms.In a sensitive domain that n good performance, such as in healthcare, better solutions may b ment of machine learning.These algorithms have drawbacks w cated and large datasets, which slow down processing, and are ing new anomalous patterns.
Deep learning is recommended by the research communit mance; it reduces the necessity of handcrafted feature engineer mance compared with traditional machine learning.Evaluating is considered the most significant step in deep learning developm ial attacks have recently been compromising its effectiveness an cation in deep learning models, affecting the model's performa published literature, it is clear that most studies have not fully of adversarial attacks, such as in [4,5,7].Although the machine niques used will encounter several types of possible security phases, one of the most pressing concerns is the adversarial atta adversaries are highly motivated to change the outcomes of mod data for their own gain.For example, attackers can cause miscla ing the data samples.In addition, few studies have investigated as in [17,18,20], but the authors have not proposed any defens protect the models.Defense methods in intelligent networks aim activity from the common patterns in intelligent networks.To m ants of adversarial attack, these defense methods must assure th have been gathered by the networked devices to combat various of the existing studies propose defensive methods, such as t [8] Appl.Sci.2023, 13 An investigation of the existing literature shows that in order to increase the effica of anomaly detection strategies in WBANs, several methodologies, including statisti and ML techniques, have been developed by researchers.However, these methods ha some drawbacks.For instance, statistical approaches cannot handle the dynamic natu of WBANs, making it challenging to identify a suitable evaluation threshold value.In dition, due to their processing complexity, non-parametric statistical methods are unsu able for real-time applications.Numerous contemporary methodologies have employ ML algorithms, including decision trees, linear regression, ANN, and the k-nearest nei bor and random forest algorithms.In a sensitive domain that needs great accuracy a good performance, such as in healthcare, better solutions may be found than the depl ment of machine learning.These algorithms have drawbacks when dealing with comp cated and large datasets, which slow down processing, and are not effective in anticip ing new anomalous patterns.
Deep learning is recommended by the research community due to its better perf mance; it reduces the necessity of handcrafted feature engineering and increases perf mance compared with traditional machine learning.Evaluating the model's performan is considered the most significant step in deep learning development.However, advers ial attacks have recently been compromising its effectiveness and have led to misclass cation in deep learning models, affecting the model's performance.After reviewing published literature, it is clear that most studies have not fully investigated the scena of adversarial attacks, such as in [4,5,7].Although the machine and deep learning te niques used will encounter several types of possible security concerns in the vario phases, one of the most pressing concerns is the adversarial attack, wherein the system adversaries are highly motivated to change the outcomes of models or to source perso data for their own gain.For example, attackers can cause misclassification by manipu ing the data samples.In addition, few studies have investigated adversarial attacks, su as in [17,18,20], but the authors have not proposed any defensive method by which protect the models.Defense methods in intelligent networks aim to distinguish malicio activity from the common patterns in intelligent networks.To mitigate the different va ants of adversarial attack, these defense methods must assure the security of the data t have been gathered by the networked devices to combat various adversarial attacks.So of the existing studies propose defensive methods, such as those reported in [

Study
Providing Anomaly Detection Dealing with Adversarial Attacks Prop ✓ ✕ An investigation of the existing literature shows that in ord of anomaly detection strategies in WBANs, several methodolo and ML techniques, have been developed by researchers.Howe some drawbacks.For instance, statistical approaches cannot ha of WBANs, making it challenging to identify a suitable evaluatio dition, due to their processing complexity, non-parametric statis able for real-time applications.Numerous contemporary metho ML algorithms, including decision trees, linear regression, ANN bor and random forest algorithms.In a sensitive domain that n good performance, such as in healthcare, better solutions may b ment of machine learning.These algorithms have drawbacks w cated and large datasets, which slow down processing, and are ing new anomalous patterns.
Deep learning is recommended by the research communit mance; it reduces the necessity of handcrafted feature engineer mance compared with traditional machine learning.Evaluating is considered the most significant step in deep learning developm ial attacks have recently been compromising its effectiveness an cation in deep learning models, affecting the model's performa published literature, it is clear that most studies have not fully of adversarial attacks, such as in [4,5,7].Although the machine niques used will encounter several types of possible security phases, one of the most pressing concerns is the adversarial atta adversaries are highly motivated to change the outcomes of mod data for their own gain.For example, attackers can cause miscla ing the data samples.In addition, few studies have investigated as in [17,18,20], but the authors have not proposed any defens protect the models.Defense methods in intelligent networks aim activity from the common patterns in intelligent networks.To m ants of adversarial attack, these defense methods must assure th have been gathered by the networked devices to combat various of the existing studies propose defensive methods, such as t [9] Appl.Sci.2023, 13

Study Providing Anomaly Detection
Dealing with Adversarial Attacks Proposing Defense Metho An investigation of the existing literature shows that in order to increase the effica of anomaly detection strategies in WBANs, several methodologies, including statisti and ML techniques, have been developed by researchers.However, these methods ha some drawbacks.For instance, statistical approaches cannot handle the dynamic natu of WBANs, making it challenging to identify a suitable evaluation threshold value.In dition, due to their processing complexity, non-parametric statistical methods are unsu able for real-time applications.Numerous contemporary methodologies have employ ML algorithms, including decision trees, linear regression, ANN, and the k-nearest nei bor and random forest algorithms.In a sensitive domain that needs great accuracy a good performance, such as in healthcare, better solutions may be found than the depl ment of machine learning.These algorithms have drawbacks when dealing with comp cated and large datasets, which slow down processing, and are not effective in anticip ing new anomalous patterns.
Deep learning is recommended by the research community due to its better perf mance; it reduces the necessity of handcrafted feature engineering and increases perf mance compared with traditional machine learning.Evaluating the model's performan is considered the most significant step in deep learning development.However, advers ial attacks have recently been compromising its effectiveness and have led to misclass cation in deep learning models, affecting the model's performance.After reviewing published literature, it is clear that most studies have not fully investigated the scena of adversarial attacks, such as in [4,5,7].Although the machine and deep learning te niques used will encounter several types of possible security concerns in the vario phases, one of the most pressing concerns is the adversarial attack, wherein the system adversaries are highly motivated to change the outcomes of models or to source perso data for their own gain.For example, attackers can cause misclassification by manipu ing the data samples.In addition, few studies have investigated adversarial attacks, su as in [17,18,20], but the authors have not proposed any defensive method by which protect the models.Defense methods in intelligent networks aim to distinguish malicio activity from the common patterns in intelligent networks.To mitigate the different va ants of adversarial attack, these defense methods must assure the security of the data t have been gathered by the networked devices to combat various adversarial attacks.So of the existing studies propose defensive methods, such as those reported in [

Study
Providing Anomaly Detection Dealing with Adversarial Attacks Prop ✓ ✕ An investigation of the existing literature shows that in ord of anomaly detection strategies in WBANs, several methodolo and ML techniques, have been developed by researchers.Howe some drawbacks.For instance, statistical approaches cannot ha of WBANs, making it challenging to identify a suitable evaluatio dition, due to their processing complexity, non-parametric statis able for real-time applications.Numerous contemporary metho ML algorithms, including decision trees, linear regression, ANN bor and random forest algorithms.In a sensitive domain that n good performance, such as in healthcare, better solutions may b ment of machine learning.These algorithms have drawbacks w cated and large datasets, which slow down processing, and are ing new anomalous patterns.
Deep learning is recommended by the research communit mance; it reduces the necessity of handcrafted feature engineer mance compared with traditional machine learning.Evaluating is considered the most significant step in deep learning developm ial attacks have recently been compromising its effectiveness an cation in deep learning models, affecting the model's performa published literature, it is clear that most studies have not fully of adversarial attacks, such as in [4,5,7].Although the machine niques used will encounter several types of possible security phases, one of the most pressing concerns is the adversarial atta adversaries are highly motivated to change the outcomes of mod data for their own gain.For example, attackers can cause miscla ing the data samples.In addition, few studies have investigated as in [17,18,20], but the authors have not proposed any defens protect the models.Defense methods in intelligent networks aim activity from the common patterns in intelligent networks.To m ants of adversarial attack, these defense methods must assure th have been gathered by the networked devices to combat various of the existing studies propose defensive methods, such as t [10] Appl.Sci.2023, 13

Study Providing Anomaly Detection
Dealing with Adversarial Attacks Proposing Defense Metho An investigation of the existing literature shows that in order to increase the effica of anomaly detection strategies in WBANs, several methodologies, including statisti and ML techniques, have been developed by researchers.However, these methods ha some drawbacks.For instance, statistical approaches cannot handle the dynamic natu of WBANs, making it challenging to identify a suitable evaluation threshold value.In dition, due to their processing complexity, non-parametric statistical methods are unsu able for real-time applications.Numerous contemporary methodologies have employ ML algorithms, including decision trees, linear regression, ANN, and the k-nearest nei bor and random forest algorithms.In a sensitive domain that needs great accuracy a good performance, such as in healthcare, better solutions may be found than the depl ment of machine learning.These algorithms have drawbacks when dealing with comp cated and large datasets, which slow down processing, and are not effective in anticip ing new anomalous patterns.
Deep learning is recommended by the research community due to its better perf mance; it reduces the necessity of handcrafted feature engineering and increases perf mance compared with traditional machine learning.Evaluating the model's performan is considered the most significant step in deep learning development.However, advers ial attacks have recently been compromising its effectiveness and have led to misclass cation in deep learning models, affecting the model's performance.After reviewing published literature, it is clear that most studies have not fully investigated the scena of adversarial attacks, such as in [4,5,7].Although the machine and deep learning te niques used will encounter several types of possible security concerns in the vario phases, one of the most pressing concerns is the adversarial attack, wherein the system adversaries are highly motivated to change the outcomes of models or to source perso data for their own gain.For example, attackers can cause misclassification by manipu ing the data samples.In addition, few studies have investigated adversarial attacks, su as in [17,18,20], but the authors have not proposed any defensive method by which protect the models.Defense methods in intelligent networks aim to distinguish malicio activity from the common patterns in intelligent networks.To mitigate the different va ants of adversarial attack, these defense methods must assure the security of the data t have been gathered by the networked devices to combat various adversarial attacks.So

Dealing with Adversarial Attacks Prop
✓ ✕ An investigation of the existing literature shows that in ord of anomaly detection strategies in WBANs, several methodolo and ML techniques, have been developed by researchers.Howe some drawbacks.For instance, statistical approaches cannot ha of WBANs, making it challenging to identify a suitable evaluatio dition, due to their processing complexity, non-parametric statis able for real-time applications.Numerous contemporary metho ML algorithms, including decision trees, linear regression, ANN bor and random forest algorithms.In a sensitive domain that n good performance, such as in healthcare, better solutions may b ment of machine learning.These algorithms have drawbacks w cated and large datasets, which slow down processing, and are ing new anomalous patterns.
Deep learning is recommended by the research communit mance; it reduces the necessity of handcrafted feature engineer mance compared with traditional machine learning.Evaluating is considered the most significant step in deep learning developm ial attacks have recently been compromising its effectiveness an cation in deep learning models, affecting the model's performa published literature, it is clear that most studies have not fully of adversarial attacks, such as in [4,5,7].Although the machine niques used will encounter several types of possible security phases, one of the most pressing concerns is the adversarial atta adversaries are highly motivated to change the outcomes of mod data for their own gain.For example, attackers can cause miscla ing the data samples.In addition, few studies have investigated as in [17,18,20], but the authors have not proposed any defens protect the models.Defense methods in intelligent networks aim activity from the common patterns in intelligent networks.To m ants of adversarial attack, these defense methods must assure th have been gathered by the networked devices to combat various [10,11] Appl.Sci.2023, 13

Dealing with Adversarial Attacks Proposing Defense Metho
An investigation of the existing literature shows that in order to increase the effica of anomaly detection strategies in WBANs, several methodologies, including statisti and ML techniques, have been developed by researchers.However, these methods ha some drawbacks.For instance, statistical approaches cannot handle the dynamic natu of WBANs, making it challenging to identify a suitable evaluation threshold value.In dition, due to their processing complexity, non-parametric statistical methods are unsu able for real-time applications.Numerous contemporary methodologies have employ ML algorithms, including decision trees, linear regression, ANN, and the k-nearest nei bor and random forest algorithms.In a sensitive domain that needs great accuracy a good performance, such as in healthcare, better solutions may be found than the depl ment of machine learning.These algorithms have drawbacks when dealing with comp cated and large datasets, which slow down processing, and are not effective in anticip ing new anomalous patterns.
Deep learning is recommended by the research community due to its better perf mance; it reduces the necessity of handcrafted feature engineering and increases perf mance compared with traditional machine learning.Evaluating the model's performan is considered the most significant step in deep learning development.However, advers ial attacks have recently been compromising its effectiveness and have led to misclass cation in deep learning models, affecting the model's performance.After reviewing published literature, it is clear that most studies have not fully investigated the scena of adversarial attacks, such as in [4,5,7].Although the machine and deep learning te niques used will encounter several types of possible security concerns in the vario phases, one of the most pressing concerns is the adversarial attack, wherein the system adversaries are highly motivated to change the outcomes of models or to source perso data for their own gain.For example, attackers can cause misclassification by manipu ing the data samples.In addition, few studies have investigated adversarial attacks, su as in [17,18,20], but the authors have not proposed any defensive method by which protect the models.Defense methods in intelligent networks aim to distinguish malicio activity from the common patterns in intelligent networks.To mitigate the different va

Study
Providing Anomaly Detection

Dealing with Adversarial Attacks Prop
✓ ✕ An investigation of the existing literature shows that in ord of anomaly detection strategies in WBANs, several methodolo and ML techniques, have been developed by researchers.Howe some drawbacks.For instance, statistical approaches cannot ha of WBANs, making it challenging to identify a suitable evaluatio dition, due to their processing complexity, non-parametric statis able for real-time applications.Numerous contemporary metho ML algorithms, including decision trees, linear regression, ANN bor and random forest algorithms.In a sensitive domain that n good performance, such as in healthcare, better solutions may b ment of machine learning.These algorithms have drawbacks w cated and large datasets, which slow down processing, and are ing new anomalous patterns.
Deep learning is recommended by the research communit mance; it reduces the necessity of handcrafted feature engineer mance compared with traditional machine learning.Evaluating is considered the most significant step in deep learning developm ial attacks have recently been compromising its effectiveness an cation in deep learning models, affecting the model's performa published literature, it is clear that most studies have not fully of adversarial attacks, such as in [4,5,7].Although the machine niques used will encounter several types of possible security phases, one of the most pressing concerns is the adversarial atta adversaries are highly motivated to change the outcomes of mod data for their own gain.For example, attackers can cause miscla ing the data samples.In addition, few studies have investigated as in [17,18,20], but the authors have not proposed any defens protect the models.Defense methods in intelligent networks aim activity from the common patterns in intelligent networks.To m

Dealing with Adversarial Attacks Proposing Defense Metho
An investigation of the existing literature shows that in order to increase the effica of anomaly detection strategies in WBANs, several methodologies, including statisti and ML techniques, have been developed by researchers.However, these methods ha some drawbacks.For instance, statistical approaches cannot handle the dynamic natu of WBANs, making it challenging to identify a suitable evaluation threshold value.In dition, due to their processing complexity, non-parametric statistical methods are unsu able for real-time applications.Numerous contemporary methodologies have employ ML algorithms, including decision trees, linear regression, ANN, and the k-nearest nei bor and random forest algorithms.In a sensitive domain that needs great accuracy a good performance, such as in healthcare, better solutions may be found than the depl ment of machine learning.These algorithms have drawbacks when dealing with comp cated and large datasets, which slow down processing, and are not effective in anticip ing new anomalous patterns.
Deep learning is recommended by the research community due to its better perf mance; it reduces the necessity of handcrafted feature engineering and increases perf mance compared with traditional machine learning.Evaluating the model's performan is considered the most significant step in deep learning development.However, advers ial attacks have recently been compromising its effectiveness and have led to misclass cation in deep learning models, affecting the model's performance.After reviewing published literature, it is clear that most studies have not fully investigated the scena of adversarial attacks, such as in [4,5,7].Although the machine and deep learning te niques used will encounter several types of possible security concerns in the vario phases, one of the most pressing concerns is the adversarial attack, wherein the system adversaries are highly motivated to change the outcomes of models or to source perso data for their own gain.For example, attackers can cause misclassification by manipu ing the data samples.In addition, few studies have investigated adversarial attacks, su as in [17,18,20], but the authors have not proposed any defensive method by which protect the models.Defense methods in intelligent networks aim to distinguish malicio

Study
Providing Anomaly Detection

Dealing with Adversarial Attacks Prop
✓ ✕ An investigation of the existing literature shows that in ord of anomaly detection strategies in WBANs, several methodolo and ML techniques, have been developed by researchers.Howe some drawbacks.For instance, statistical approaches cannot ha of WBANs, making it challenging to identify a suitable evaluatio dition, due to their processing complexity, non-parametric statis able for real-time applications.Numerous contemporary metho ML algorithms, including decision trees, linear regression, ANN bor and random forest algorithms.In a sensitive domain that n good performance, such as in healthcare, better solutions may b ment of machine learning.These algorithms have drawbacks w cated and large datasets, which slow down processing, and are ing new anomalous patterns.
Deep learning is recommended by the research communit mance; it reduces the necessity of handcrafted feature engineer mance compared with traditional machine learning.Evaluating is considered the most significant step in deep learning developm ial attacks have recently been compromising its effectiveness an cation in deep learning models, affecting the model's performa published literature, it is clear that most studies have not fully of adversarial attacks, such as in [4,5,7].Although the machine niques used will encounter several types of possible security phases, one of the most pressing concerns is the adversarial atta adversaries are highly motivated to change the outcomes of mod data for their own gain.For example, attackers can cause miscla ing the data samples.In addition, few studies have investigated as in [17,18,20], but the authors have not proposed any defens protect the models.Defense methods in intelligent networks aim [13] Appl.Sci.2023, 13

Dealing with Adversarial Attacks Proposing Defense Metho
An investigation of the existing literature shows that in order to increase the effica of anomaly detection strategies in WBANs, several methodologies, including statisti and ML techniques, have been developed by researchers.However, these methods ha some drawbacks.For instance, statistical approaches cannot handle the dynamic natu of WBANs, making it challenging to identify a suitable evaluation threshold value.In dition, due to their processing complexity, non-parametric statistical methods are unsu able for real-time applications.Numerous contemporary methodologies have employ ML algorithms, including decision trees, linear regression, ANN, and the k-nearest nei bor and random forest algorithms.In a sensitive domain that needs great accuracy a good performance, such as in healthcare, better solutions may be found than the depl ment of machine learning.These algorithms have drawbacks when dealing with comp cated and large datasets, which slow down processing, and are not effective in anticip ing new anomalous patterns.
Deep learning is recommended by the research community due to its better perf mance; it reduces the necessity of handcrafted feature engineering and increases perf mance compared with traditional machine learning.Evaluating the model's performan is considered the most significant step in deep learning development.However, advers ial attacks have recently been compromising its effectiveness and have led to misclass cation in deep learning models, affecting the model's performance.After reviewing published literature, it is clear that most studies have not fully investigated the scena of adversarial attacks, such as in [4,5,7].Although the machine and deep learning te niques used will encounter several types of possible security concerns in the vario phases, one of the most pressing concerns is the adversarial attack, wherein the system adversaries are highly motivated to change the outcomes of models or to source perso data for their own gain.For example, attackers can cause misclassification by manipu ing the data samples.In addition, few studies have investigated adversarial attacks, su as in [17,18,20], but the authors have not proposed any defensive method by which

Study
Providing Anomaly Detection

Dealing with Adversarial Attacks Prop
✓ ✕ An investigation of the existing literature shows that in ord of anomaly detection strategies in WBANs, several methodolo and ML techniques, have been developed by researchers.Howe some drawbacks.For instance, statistical approaches cannot ha of WBANs, making it challenging to identify a suitable evaluatio dition, due to their processing complexity, non-parametric statis able for real-time applications.Numerous contemporary metho ML algorithms, including decision trees, linear regression, ANN bor and random forest algorithms.In a sensitive domain that n good performance, such as in healthcare, better solutions may b ment of machine learning.These algorithms have drawbacks w cated and large datasets, which slow down processing, and are ing new anomalous patterns.
Deep learning is recommended by the research communit mance; it reduces the necessity of handcrafted feature engineer mance compared with traditional machine learning.Evaluating is considered the most significant step in deep learning developm ial attacks have recently been compromising its effectiveness an cation in deep learning models, affecting the model's performa published literature, it is clear that most studies have not fully of adversarial attacks, such as in [4,5,7].Although the machine niques used will encounter several types of possible security phases, one of the most pressing concerns is the adversarial atta adversaries are highly motivated to change the outcomes of mod data for their own gain.For example, attackers can cause miscla ing the data samples.In addition, few studies have investigated as in [17,18,20], but the authors have not proposed any defens [14] Appl.Sci.2023, 13

Dealing with Adversarial Attacks
Proposing Defense Metho An investigation of the existing literature shows that in order to increase the effica of anomaly detection strategies in WBANs, several methodologies, including statisti and ML techniques, have been developed by researchers.However, these methods ha some drawbacks.For instance, statistical approaches cannot handle the dynamic natu of WBANs, making it challenging to identify a suitable evaluation threshold value.In dition, due to their processing complexity, non-parametric statistical methods are unsu able for real-time applications.Numerous contemporary methodologies have employ ML algorithms, including decision trees, linear regression, ANN, and the k-nearest nei bor and random forest algorithms.In a sensitive domain that needs great accuracy a good performance, such as in healthcare, better solutions may be found than the depl ment of machine learning.These algorithms have drawbacks when dealing with comp cated and large datasets, which slow down processing, and are not effective in anticip ing new anomalous patterns.
Deep learning is recommended by the research community due to its better perf mance; it reduces the necessity of handcrafted feature engineering and increases perf mance compared with traditional machine learning.Evaluating the model's performan is considered the most significant step in deep learning development.However, advers ial attacks have recently been compromising its effectiveness and have led to misclass cation in deep learning models, affecting the model's performance.After reviewing published literature, it is clear that most studies have not fully investigated the scena of adversarial attacks, such as in [4,5,7].Although the machine and deep learning te niques used will encounter several types of possible security concerns in the vario phases, one of the most pressing concerns is the adversarial attack, wherein the system adversaries are highly motivated to change the outcomes of models or to source perso data for their own gain.For example, attackers can cause misclassification by manipu

Study
Providing Anomaly Detection

Dealing with Adversarial Attacks
Prop An investigation of the existing literature shows that in ord of anomaly detection strategies in WBANs, several methodolo and ML techniques, have been developed by researchers.Howe some drawbacks.For instance, statistical approaches cannot ha of WBANs, making it challenging to identify a suitable evaluatio dition, due to their processing complexity, non-parametric statis able for real-time applications.Numerous contemporary metho ML algorithms, including decision trees, linear regression, ANN bor and random forest algorithms.In a sensitive domain that n good performance, such as in healthcare, better solutions may b ment of machine learning.These algorithms have drawbacks w cated and large datasets, which slow down processing, and are ing new anomalous patterns.
Deep learning is recommended by the research communit mance; it reduces the necessity of handcrafted feature engineer mance compared with traditional machine learning.Evaluating is considered the most significant step in deep learning developm ial attacks have recently been compromising its effectiveness an cation in deep learning models, affecting the model's performa published literature, it is clear that most studies have not fully of adversarial attacks, such as in [4,5,7].Although the machine niques used will encounter several types of possible security phases, one of the most pressing concerns is the adversarial atta adversaries are highly motivated to change the outcomes of mod data for their own gain.For example, attackers can cause miscla

Study
Providing Anomaly Detection

Dealing with Adversarial Attacks
Prop An investigation of the existing literature shows that in ord of anomaly detection strategies in WBANs, several methodolo and ML techniques, have been developed by researchers.Howe some drawbacks.For instance, statistical approaches cannot ha of WBANs, making it challenging to identify a suitable evaluatio dition, due to their processing complexity, non-parametric statis able for real-time applications.Numerous contemporary metho ML algorithms, including decision trees, linear regression, ANN bor and random forest algorithms.In a sensitive domain that n good performance, such as in healthcare, better solutions may b ment of machine learning.These algorithms have drawbacks w cated and large datasets, which slow down processing, and are ing new anomalous patterns.
Deep learning is recommended by the research communit mance; it reduces the necessity of handcrafted feature engineer mance compared with traditional machine learning.Evaluating is considered the most significant step in deep learning developm ial attacks have recently been compromising its effectiveness an cation in deep learning models, affecting the model's performa published literature, it is clear that most studies have not fully of adversarial attacks, such as in [4,5,7].Although the machine niques used will encounter several types of possible security phases, one of the most pressing concerns is the adversarial atta [17] Appl.Sci.2023, 13, x FOR PEER REVIEW

Study
Providing Anomaly Detection

Dealing with Adversarial Attacks
Prop An investigation of the existing literature shows that in ord of anomaly detection strategies in WBANs, several methodolo and ML techniques, have been developed by researchers.Howe some drawbacks.For instance, statistical approaches cannot ha of WBANs, making it challenging to identify a suitable evaluatio dition, due to their processing complexity, non-parametric statis able for real-time applications.Numerous contemporary metho ML algorithms, including decision trees, linear regression, ANN bor and random forest algorithms.In a sensitive domain that n good performance, such as in healthcare, better solutions may b ment of machine learning.These algorithms have drawbacks w cated and large datasets, which slow down processing, and are ing new anomalous patterns.
Deep learning is recommended by the research communit mance; it reduces the necessity of handcrafted feature engineer mance compared with traditional machine learning.Evaluating is considered the most significant step in deep learning developm ial attacks have recently been compromising its effectiveness an cation in deep learning models, affecting the model's performa published literature, it is clear that most studies have not fully of adversarial attacks, such as in [4,5,7].Although the machine [18] Appl.Sci.2023, 13, x FOR PEER REVIEW

Study
Providing Anomaly Detection

Dealing with Adversarial Attacks Prop
An investigation of the existing literature shows that in ord of anomaly detection strategies in WBANs, several methodolo and ML techniques, have been developed by researchers.Howe some drawbacks.For instance, statistical approaches cannot ha of WBANs, making it challenging to identify a suitable evaluatio dition, due to their processing complexity, non-parametric statis able for real-time applications.Numerous contemporary metho ML algorithms, including decision trees, linear regression, ANN bor and random forest algorithms.In a sensitive domain that n good performance, such as in healthcare, better solutions may b ment of machine learning.These algorithms have drawbacks w cated and large datasets, which slow down processing, and are ing new anomalous patterns.
Deep learning is recommended by the research communit mance; it reduces the necessity of handcrafted feature engineer mance compared with traditional machine learning.Evaluating is considered the most significant step in deep learning developm ial attacks have recently been compromising its effectiveness an cation in deep learning models, affecting the model's performa published literature, it is clear that most studies have not fully

Study
Providing Anomaly Detection

Dealing with Adversarial Attacks
Prop An investigation of the existing literature shows that in ord of anomaly detection strategies in WBANs, several methodolo and ML techniques, have been developed by researchers.Howe some drawbacks.For instance, statistical approaches cannot ha of WBANs, making it challenging to identify a suitable evaluatio dition, due to their processing complexity, non-parametric statis able for real-time applications.Numerous contemporary metho ML algorithms, including decision trees, linear regression, ANN bor and random forest algorithms.In a sensitive domain that n good performance, such as in healthcare, better solutions may b ment of machine learning.These algorithms have drawbacks w cated and large datasets, which slow down processing, and are ing new anomalous patterns.
Deep learning is recommended by the research communit mance; it reduces the necessity of handcrafted feature engineer mance compared with traditional machine learning.Evaluating is considered the most significant step in deep learning developm [21] Appl.Sci.2023, 13

Dealing with Adversarial Attacks
Proposing Defense Metho An investigation of the existing literature shows that in order to increase the effica of anomaly detection strategies in WBANs, several methodologies, including statisti and ML techniques, have been developed by researchers.However, these methods ha some drawbacks.For instance, statistical approaches cannot handle the dynamic natu of WBANs, making it challenging to identify a suitable evaluation threshold value.In dition, due to their processing complexity, non-parametric statistical methods are unsu able for real-time applications.Numerous contemporary methodologies have employ ML algorithms, including decision trees, linear regression, ANN, and the k-nearest nei bor and random forest algorithms.In a sensitive domain that needs great accuracy a good performance, such as in healthcare, better solutions may be found than the depl ment of machine learning.These algorithms have drawbacks when dealing with comp cated and large datasets, which slow down processing, and are not effective in anticip ing new anomalous patterns.
Deep learning is recommended by the research community due to its better perf mance; it reduces the necessity of handcrafted feature engineering and increases perf mance compared with traditional machine learning.Evaluating the model's performan

Study
Providing Anomaly Detection

Dealing with Adversarial Attacks
Prop An investigation of the existing literature shows that in ord of anomaly detection strategies in WBANs, several methodolo and ML techniques, have been developed by researchers.Howe some drawbacks.For instance, statistical approaches cannot ha of WBANs, making it challenging to identify a suitable evaluatio dition, due to their processing complexity, non-parametric statis able for real-time applications.Numerous contemporary metho ML algorithms, including decision trees, linear regression, ANN bor and random forest algorithms.In a sensitive domain that n good performance, such as in healthcare, better solutions may b ment of machine learning.These algorithms have drawbacks w cated and large datasets, which slow down processing, and are ing new anomalous patterns.
Deep learning is recommended by the research communit mance; it reduces the necessity of handcrafted feature engineer mance compared with traditional machine learning.Evaluating An investigation of the existing literature shows that in order to increase the efficacy of anomaly detection strategies in WBANs, several methodologies, including statistical and ML techniques, have been developed by researchers.However, these methods have some drawbacks.For instance, statistical approaches cannot handle the dynamic nature of WBANs, making it challenging to identify a suitable evaluation threshold value.In addition, due to their processing complexity, non-parametric statistical methods are unsuitable for real-time applications.Numerous contemporary methodologies have employed ML algorithms, including decision trees, linear regression, ANN, and the k-nearest neighbor and random forest algorithms.In a sensitive domain that needs great accuracy and good performance, such as in healthcare, better solutions may be found than the deployment of machine learning.These algorithms have drawbacks when dealing with complicated and large datasets, which slow down processing, and are not effective in anticipating new anomalous patterns.
Deep learning is recommended by the research community due to its better performance; it reduces the necessity of handcrafted feature engineering and increases performance compared with traditional machine learning.Evaluating the model's performance is considered the most significant step in deep learning development.However, adversarial attacks have recently been compromising its effectiveness and have led to misclassification in deep learning models, affecting the model's performance.After reviewing the published literature, it is clear that most studies have not fully investigated the scenario of adversarial attacks, such as in [4,5,7].Although the machine and deep learning techniques used will encounter several types of possible security concerns in the various phases, one of the most pressing concerns is the adversarial attack, wherein the system's adversaries are highly motivated to change the outcomes of models or to source personal data for their own gain.For example, attackers can cause misclassification by manipulating the data samples.In addition, few studies have investigated adversarial attacks, such as in [17,18,20], but the authors have not proposed any defensive method by which to protect the models.Defense methods in intelligent networks aim to distinguish malicious activity from the common patterns in intelligent networks.To mitigate the different variants of adversarial attack, these defense methods must assure the security of the data that have been gathered by the networked devices to combat various adversarial attacks.Some of the existing studies propose defensive methods, such as those reported in [15,19].However, the authors of [15] focused on one type of attack that occurred in the training data and ignored the attacks that occurred in the testing data.In [19], the study was limited to attacks being countered by traditional ML techniques; the model was designed to deal with batch learning, a system that is difficult to apply to big data scenarios, which are characteristic of today's healthcare monitoring applications.Therefore, the proposed research in this paper is motivated by the need to develop a deep learning model for processing big data scenarios and to propose a possible solution for adversarial attack issues in WBANs.

The Proposed Model
This paper presents a technique for deep learning-based anomaly detection for WBANs that affords them the ability to detect adversarial attacks.The suggested model involves four phases: the data collection and pre-processing phase, the training phase, the detection phase, and the evaluation phase.In the evaluation phase, we test the ability of the proposed model to detect the adversarial examples that have been introduced in either the training or testing phases, and we implement a defense method that can mitigate the adversarial attacks.Figure 1 depicts the phases of the suggested model and the workflow of the various processes.
Appl.Sci.2023, 13, x FOR PEER REVIEW 6 of 17 However, the authors of [15] focused on one type of attack that occurred in the training data and ignored the attacks that occurred in the testing data.In [19], the study was limited to attacks being countered by traditional ML techniques; the model was designed to deal with batch learning, a system that is difficult to apply to big data scenarios, which are characteristic of today's healthcare monitoring applications.Therefore, the proposed research in this paper is motivated by the need to develop a deep learning model for processing big data scenarios and to propose a possible solution for adversarial attack issues in WBANs.

The Proposed Model
This paper presents a technique for deep learning-based anomaly detection for WBANs that affords them the ability to detect adversarial attacks.The suggested model involves four phases: the data collection and pre-processing phase, the training phase, the detection phase, and the evaluation phase.In the evaluation phase, we test the ability of the proposed model to detect the adversarial examples that have been introduced in either the training or testing phases, and we implement a defense method that can mitigate the adversarial attacks.Figure 1 depicts the phases of the suggested model and the workflow of the various processes.

Data Collection and Pre-Processing
In this phase, the data are collected from various WBAN sensors.The WBAN in question contains many wearable sensor nodes, which are used to collect physiological data including systolic arterial blood pressure (ABPsys), pulse rate, respiration rate (RESP), heart rate (HR), mean arterial blood pressure (ABP-mean), oxygen saturation (SPO2), temperature, and diastolic arterial blood pressure (ABPdias), which are all measurements that should be taken into consideration when assessing a patient's condition.The Multiple Intelligent Monitoring in Intensive Care (MIMIC-I and II) dataset includes precise physiological data records taken from over 90 ICU patients, who are known as subjects.The dataset has seven features that represent the patient's clinical situation, which include mean arterial blood pressure (ABP-mean), oxygen saturation (SPO2), heart rate (HR), temperature, diastolic arterial blood pressure (ABPdias), respiration rate (RESP), systolic

Data Collection and Pre-Processing
In this phase, the data are collected from various WBAN sensors.The WBAN in question contains many wearable sensor nodes, which are used to collect physiological data including systolic arterial blood pressure (ABPsys), pulse rate, respiration rate (RESP), heart rate (HR), mean arterial blood pressure (ABP-mean), oxygen saturation (SPO 2 ), temperature, and diastolic arterial blood pressure (ABPdias), which are all measurements that should be taken into consideration when assessing a patient's condition.The Multiple Intelligent Monitoring in Intensive Care (MIMIC-I and II) dataset includes precise physiological data records taken from over 90 ICU patients, who are known as subjects.The dataset has seven features that represent the patient's clinical situation, which include mean arterial blood pressure (ABP-mean), oxygen saturation (SPO 2 ), heart rate (HR), temperature, di-astolic arterial blood pressure (ABPdias), respiration rate (RESP), systolic arterial blood pressure (ABPsys), and pulse, with the appropriate timesteps and date; this dataset was used for the purposes of this study [22].Table 2, below, contains a sample of the sensor readings in the dataset.The dataset is first preprocessed and prepared for use with the deep learning models via the normalization approach, which reorders the dataset's values appropriately.The goal of normalization is to use a consistent scale to reorder the values of the dataset's numeric columns without distorting the ranges of the values or erasing data.Additionally, certain algorithms require normalization in order to correctly simulate the data [23].For the purposes of this paper, each attribute in the dataset is normalized, as in Equation ( 1), in the range [0, 1]: where x(i) is the dataset,

−
x is one column in the dataset, and S(x) is the number of data samples.

Training and Detection Using ConvLSTM
The training and detection phases are necessary to create a model in the training phase that can be used in the detection phase to identify anomalies in the WBAN, utilizing a deep learning approach.This technique has been widely implemented in healthcare applications because of its capacity to automatically detect complicated features without requiring domain expertise.The model developed in this paper is based on the hybrid ConvLSTM technique, which is a hybridization of the long short-term memory model and convolutional deep neural networks.
(a) Convolutional LSTM (ConvLSTM) The hybrid deep learning system developed in this study combines the principles of convolutional and LSTM models.ConvLSTM is a type of RNN that is capable of capturing spatial-temporal data [24].The convolutional component acquires the spatial area data, while the LSTM component leverages the temporal area data.However, the data, which are in the form of time series from the different sensors, have both spatial and temporal links.Thus, ConvLSTM can be employed as an important framework for anomaly detection problems involving time series data [25].The ConvLSTM captures and applies both temporal and spatial correlations to predict the future state of a network cell, based on the inputs and previous states of its nearest neighbors.This is accomplished by including a convolution operation into the matrix multiplication step that is used in traditional fully integrated LSTM state-to-state and input-to-state transitions [24].The ConvLSTM consists of multiple gates; the data flow may be stated using Equations ( 2)- (7).
The weighted sum of each gate's inputs is given a sigmoid function, which is then used as the activation function [26]: where f t is the forget gate's output, which regulates the data that are lost in the previous cell state.C t−1 , W xi is the convolution kernel used for the input tensor X t in the input gate, X t is the input tensor at time t, W hf is the convolution kernel used for the input tensor H t−1 in the forget gate, H t−1 is the output tensor from the cell at time t − 1, W cf is the weight that is used for the old cell state C t−1 in the forget gate, C t−1 is the cell state at time t − 1, and B f is the bias in the forget gate.
Here, t is the output of the input gate, W xi is the convolution kernel used for the input tensor X t in the input gate, W hi is the convolution kernel used for the input tensor H t−1 in the input gate, W ci is the weight that is used for the old cell state C t−1 in the input gate, and B i is the bias in the input gate.
Here, c t is the data that are stored in the new cell state, C t , W xc is the convolution kernel used for the input tensor X t to create the data c t that will be stored in the new cell state c t , W hc is the convolution kernel used for the input tensor H t−1 to create the data c t that will be kept in the new cell state c t , and B c is the bias for forming the data c t that will be kept in the new cell state c t .
Here, c t is the cell state at time t, and c t is the data that are stored in the new cell state C t .
Here, o t is the output of the output gate; this controls the data that are output as h t from the cell.W xo is the convolution kernel used for the input tensor X t in the output gate, W ho is the convolution kernel used for the input tensor H t−1 in the output gate, W co is the weight that is used for the new cell state c t in the output gate, and B o is the bias in the output gate.
h t = o t tanh(c t ) Here, h t is the output tensor from the cell at time t.

Adversarial Attack Modeling
Two types of adversarial attacks are introduced in the training and testing phases to examine the performance of the suggested model against such attacks.Poisoning adversarial attacks are introduced in the training phase, whereas evasion-based adversarial attacks are introduced in the testing phase.This adversarial attack can be defined as an input engineered to cause misclassification in the DL algorithms.Recently, adversarial DL has gained great popularity in healthcare applications, due to the limitations of the current DL models.For example, an adversary might insert new adversarial data into a healthcare DL model to falsely classify a hypothyroid patient [27].Furthermore, in the context of medical image processing, the researchers reported various adversarial attacks against the DL model that were intended to modify results by inserting noise and causing the misclassification of a benign mole as malignant with high confidence [28,29].In the previous section, the DL model was applied to detect normal and abnormal behaviors regarding the patient's vital signs.The following sections describe the types of adversarial attacks that have been adopted in this work.The BIM attacks are repeated multiple times, using a small size, and are clipped.After each iteration, the result is clipped to guarantee the level of perturbation [30].Both FGSM and BIM attacks are used to evaluate the robustness of the proposed model in this paper.

Proactive Defense
A proactive defense method is applied to counter adversarial attacks and to improve the performance of the proposed model.The applied defense comprises the retraining of the deep learning model (in the detection phase), which is based on robust optimization and model parameters to mitigate the effects of adversarial attacks that can cause higher confidence and distortion.

Model Evaluation
The suggested model's performance was evaluated using four metrics, namely, recall, accuracy, precision, and F1-score.
The class of each data object is established using specific classification methods, providing each sample with a predicted label (positive or negative) [31].As a result, after the detection phase, each sample falls into one of four categories: Four parameters, comprising recognition precision, recall, F1-score, and accuracy, are utilized to evaluate the effectiveness of the suggested model, utilizing the four classes listed above.
The accuracy value is a statistical measure of how efficiently a model can predict an outcome [32].Equation (8) illustrates how the accuracy measure is calculated.
Precision = TP TP + FP (10) The F1-score is the weighted sum of recall and precision and is utilized when the data are uneven [34]; Equation (11) illustrates how the F1-score is calculated.

Experiments and Results
In this section, the implementation of the suggested model and the investigation of its performance in relation to adversarial attacks are discussed.The section starts by providing the model parameters and the structure of the ConvLSTM technique.

Model Setup
This paper selects the MIMIC dataset [22], a substantial physiological dataset that has been referred to in Section 3.1.The suggested model was put into practice via Python, using the sklearn library with help from additional scientific computing libraries, such as Matplotlib, NumPy, and scikit-learn, to carry out a variety of tasks that include preprocessing and model selection.The ConvLSTM approach uses the Adam optimizer as its optimization algorithm.The ConvLSTM has a four-layer network with two dropout layers ((filters = 64 in two layers), (kernel_size = 1), (padding = "same" in two layers), and (activation function = "Relu")) with a kernel size of 1.The model was trained with a batch size of 30 and using various epoch counts.Using a ratio of 70:30 for the training and testing partitions, the dataset was divided into training and testing datasets.Due to the fact that the dataset is huge, this split is more suitable for this particular dataset.Table 3 summarizes the basic settings of the model.

Results and Analysis
In this subsection, we investigate the effect of the two types of adversarial attacks on the performance of the suggested anomaly detection model.Furthermore, we investigate how proactive defense can play a role in combating such adversaries.

The Results of a Poisoning Attack
In this type of attack, the attack poisons the training data in order to access the proposed model, so as to manipulate the output status of the patient and alter abnormal behaviors to normal behaviors, which will affect the patient's status.The attacker uses data modification, which poisons the training data directly by modifying them before they are used for training, in order to manipulate the model's decision (this is referred to as Data Poisoning Attack 1), as shown in Table 4. Another type of poisoning attack entails logic corruption, which affects the overall learning process of the proposed model (this is referred to as Data Poisoning Attack 2), as shown in Table 5.The results in Tables 4 and 5 show a decrease in the detection accuracy, as well as in the other performance metrics.Two data subjects, namely, subject 1 and subject 2, were examined.Table 4 shows that the accuracy of the model regarding subject 1, without the attacks, was 97%.This decreased to 74% after employing Data Poisoning Attack 1.For subject 2, the model without attacks achieved an accuracy of 99.9%; after applying Data Poisoning Attack 1, the accuracy dropped to 75.59%.Similarly, Table 5 reports the results of Data Poisoning Attack 2, which also shows a drop in all performance metrics.It reveals that the accuracy drops to 59% for subject 1 and to 75% for subject 2. As a result, the poisoning attacks were successful in compromising the proposed model's performance and causing data misclassification, which would impact the status of both patients and their treatments.In Figure 2, different epsilon values, which represent the amount of adversarial content in the data, are used to test the efficacy of the proposed model against the adversarial Data Poisoning Attack 1.In the dataset for subject 1, various values of epsilon (0.00001, 0.00100, 0.00200, and 3.00000) were examined and resulted in different accuracy values (67%, 75%, 78%, and 75%).Similarly, for subject 2, the same epsilon values were examined and resulted in identical accuracy values (79%, 79%, 79%, and 79%).
In Figure 2, different epsilon values, which represent the amount of adversarial con-tent in the data, are used to test the efficacy of the proposed model against the adversarial Data Poisoning Attack 1.In the dataset for subject 1, various values of epsilon (0.00001, 0.00100, 0.00200, and 3.00000) were examined and resulted in different accuracy values (67%, 75%, 78%, and 75%).Similarly, for subject 2, the same epsilon values were examined and resulted in identical accuracy values (79%, 79%, 79%, and 79%).The same analogy applies to adversarial Data Poisoning Attack 2. As shown in Figure 3, the epsilon values were used for subject 1 and achieved different accuracy values of (75%, 75%, 75%, and 75%) and (82%, 81%, 84%, and 84%) for subject 2. The same analogy applies to adversarial Data Poisoning Attack 2. As shown in Figure 3, the epsilon values were used for subject 1 and achieved different accuracy values of (75%, 75%, 75%, and 75%) and (82%, 81%, 84%, and 84%) for subject 2. The results reported in Figures 2 and 3 clearly show that as the amount of adversarial content increases (the epsilon values increase), the accuracy decreases; it drops sharply after epsilon = 0.3.This outcome can be generalized for both data subjects and for both poisoning attack types.

The Results of Evasion Attacks
For evasion attacks, the attacks adjusted the data to introduce malicious samples during the testing phase.This was achieved by performing gradient-based operations to fool the DL model, stopping it from detecting any strange behavior.As a result, a drop in the accuracy of the model and in other performance metrics was recorded.The BIM and FGSM attacks affected the performance of the model for all datasets and fooled the DL model, as shown in Tables 6 and 7.In the case of subject 1 and subject 2, we created two scenarios; these are named "Before Attack" and "After BIM Attack", as shown in Table 6.For both scenarios, we used the same parameters to report the results.In the case of subject 1, Table 6 shows that before the attacks, the achieved accuracy was 97.59%, with a loss of 0.1%.However, a severe accuracy drop to 79% was reported after the BIM attack.A similar scenario can be seen for subject 2, showing a similar drop in all performance metrics.The results reported in Figures 2 and 3 clearly show that as the amount of adversarial content increases (the epsilon values increase), the accuracy decreases; it drops sharply after epsilon = 0.3.This outcome can be generalized for both data subjects and for both poisoning attack types.

The Results of Evasion Attacks
For evasion attacks, the attacks adjusted the data to introduce malicious samples during the testing phase.This was achieved by performing gradient-based operations to fool the DL model, stopping it from detecting any strange behavior.As a result, a drop in the accuracy of the model and in other performance metrics was recorded.The BIM and FGSM attacks affected the performance of the model for all datasets and fooled the DL model, as shown in Tables 6 and 7.In the case of subject 1 and subject 2, we created two scenarios; these are named "Before Attack" and "After BIM Attack", as shown in Table 6.For both scenarios, we used the same parameters to report the results.In the case of subject 1, Table 6 shows that before the attacks, the achieved accuracy was 97.59%, with a loss of 0.1%.However, a severe accuracy drop to 79% was reported after the BIM attack.A similar scenario can be seen for subject 2, showing a similar drop in all performance metrics.As seen in Figure 4a, we applied different epsilon values, which represent the amounts of BIM adversarial attack content for the subject 1 dataset (0.00001, 0.00100, 0.00200, and 3.00000).The figure shows the achieved accuracy values of 71%, 59%, 52%, and 15%, respectively.In Figure 4b, we applied the same epsilon values, which resulted in different accuracy values of 42%, 38%, 38%, and 38%, respectively.In Figure 4, we applied a different approach by depicting the accuracy value before and after the attacks, along with the variations in the epsilon values.It is clear that a sharp drop was recorded, which increased along with the increase in the amount of adversarial content.
Appl.Sci.2023, 13, x FOR PEER REVIEW 13 of 17 As seen in Figure 4a, we applied different epsilon values, which represent the amounts of BIM adversarial attack content for the subject 1 dataset (0.00001, 0.00100, 0.00200, and 3.00000).The figure shows the achieved accuracy values of 71%, 59%, 52%, and 15%, respectively.In Figure 4b, we applied the same epsilon values, which resulted in different accuracy values of 42%, 38%, 38%, and 38%, respectively.In Figure 4, we applied a different approach by depicting the accuracy value before and after the attacks, along with the variations in the epsilon values.It is clear that a sharp drop was recorded, which increased along with the increase in the amount of adversarial content.In a similar way, Table 7 shows the results of the proposed model before and after the FGSM attacks, which can be considered a type of evasion attack.As is reported, for both subjects, the model exhibited a sharp drop in all metrics after the attacks.
Figure 5a shows the similarly applied different epsilon values, which represent the amounts of FGSM adversarial attack content for the subject 1 dataset (0.00001, 0.00100, 0.00200, and 3.00000), and shows the achieved accuracy values of 63%, 63%, 63%, and 61%, respectively.Figure 5b shows the similarly applied same epsilon values, which resulted in different accuracy values of 71%, 71%, 71%, and 69%, respectively.It is clear that a sharp drop has been recorded, which increased along with the increase in the amount of adversarial content for subject 1 but exhibited a slight increase in accuracy for subject 2. The In a similar way, Table 7 shows the results of the proposed model before and after the FGSM attacks, which can be considered a type of evasion attack.As is reported, for both subjects, the model exhibited a sharp drop in all metrics after the attacks.
Figure 5a shows the similarly applied different epsilon values, which represent the amounts of FGSM adversarial attack content for the subject 1 dataset (0.00001, 0.00100, 0.00200, and 3.00000), and shows the achieved accuracy values of 63%, 63%, 63%, and 61%, respectively.Figure 5b shows the similarly applied same epsilon values, which resulted in different accuracy values of 71%, 71%, 71%, and 69%, respectively.It is clear that a sharp drop has been recorded, which increased along with the increase in the amount of adversarial content for subject 1 but exhibited a slight increase in accuracy for subject 2. The reason behind such unusual behavior in the model regarding subject 2 might be related to the behavior of the model without an attack regarding that subject, which also exhibited a decrease in accuracy.In a similar way, Table 7 shows the results of the proposed model before and after the FGSM attacks, which can be considered a type of evasion attack.As is reported, for both subjects, the model exhibited a sharp drop in all metrics after the attacks.
Figure 5a shows the similarly applied different epsilon values, which represent the amounts of FGSM adversarial attack content for the subject 1 dataset (0.00001, 0.00100, 0.00200, and 3.00000), and shows the achieved accuracy values of 63%, 63%, 63%, and 61%, respectively.Figure 5b shows the similarly applied same epsilon values, which resulted in different accuracy values of 71%, 71%, 71%, and 69%, respectively.It is clear that a sharp drop has been recorded, which increased along with the increase in the amount of adversarial content for subject 1 but exhibited a slight increase in accuracy for subject 2. The reason behind such unusual behavior in the model regarding subject 2 might be related to the behavior of the model without an attack regarding that subject, which also exhibited a decrease in accuracy.In summary, the investigation of two types of attacks showed that all of the attacks decreased the performance of the model.The BIM attack achieved the highest decrease compared to the rest of the attacks, as shown in Figures 4 and 5

Proactive Defense Method
Retraining the model is a common training practice used to reduce losses and increase its accuracy.This is one of the most effective proactive strategies for defense against adversarial attacks on deep learning models.The resultant retrained network should be able to withstand the adversarial attacks used to generate adversarial samples during the training phase.These adversarial attacks target neural networks to cause misclassification.After the attacks, these assaults and the correct labels are used to teach the neural network.Such proactive strategies cause a minor decline in the predictability of a deep learning model.However, the resilience that they provide against adversarial attacks is thought to make up for this decline.A model of this type of operation is illustrated in Figure 6, below.In summary, the investigation of two types of attacks showed that all of the attacks decreased the performance of the model.The BIM attack achieved the highest decrease compared to the rest of the attacks, as shown in Figures 4 and 5

Proactive Defense Method
Retraining the model is a common training practice used to reduce losses and increase its accuracy.This is one of the most effective proactive strategies for defense against adversarial attacks on deep learning models.The resultant retrained network should be able to withstand the adversarial attacks used to generate adversarial samples during the training phase.These adversarial attacks target neural networks to cause misclassification.After the attacks, these assaults and the correct labels are used to teach the neural network.Such proactive strategies cause a minor decline in the predictability of a deep learning model.However, the resilience that they provide against adversarial attacks is thought to make up for this decline.A model of this type of operation is illustrated in Figure 6, below.8 show an increase in accuracy, as well as in other performance metrics, after retraining the model for both subjects.Analyzing the scenarios where both attacks succeeded in fooling the DL model shows that deep learning models have weaknesses that may lead to misclassifying normal behavior as abnormal behavior.In the healthcare domain, our investigations revealed that adversarial attacks can reduce accuracy when utilizing the most recent deep learning model and, therefore, lead to severe and dangerous consequences.Thus, retraining the model is a proactive defensive method that helps to make sure that such attacks fail to fool the deep learning model and makes the model more robust.

Conclusions
With the continuous advancement of smart sensors, wireless communications, and IoT applications and services, wireless body area networks (WBANs) are becoming increasingly popular.However, several adversarial attacks and the attendant weaknesses make it challenging to develop secure IoHT applications from a security perspective.The widespread usage of anomaly detection with deep learning has been employed to construct appropriate models for detecting and mitigating adversarial attacks.Nonetheless, these attacks could deceive deep learning models into misclassifying the model results.This can lead to erroneous decisions, such as incorrect patient diagnoses and erroneous medicine administration.This paper addressed the effect of adversarial attacks on deep learning-based anomaly detection models and applied several safeguard techniques to defend networks against such adversarial attacks.The effectiveness of the suggested model was assessed in the face of adversarial attacks that were intended to give attackers control over the classification process used by the anomaly detection algorithm.The effectiveness of the suggested model was evaluated using the FGSM and BIM adversarial attack techniques.The suggested model showed its capacity to identify anomalies in the face of various adversarial attacks.More specifically, the evaluation results revealed that the model achieved an average F1 score of around 97% and an accuracy of 98% in the face of such attacks.However, as a future task, more scenarios regarding adversarial attacks  Analyzing the scenarios where both attacks succeeded in fooling the DL model shows that deep learning models have weaknesses that may lead to misclassifying normal behavior as abnormal behavior.In the healthcare domain, our investigations revealed that adversarial attacks can reduce accuracy when utilizing the most recent deep learning model and, therefore, lead to severe and dangerous consequences.Thus, retraining the model is a proactive defensive method that helps to make sure that such attacks fail to fool the deep learning model and makes the model more robust.

Conclusions
With the continuous advancement of smart sensors, wireless communications, and IoT applications and services, wireless body area networks (WBANs) are becoming increasingly popular.However, several adversarial attacks and the attendant weaknesses make it challenging to develop secure IoHT applications from a security perspective.The widespread usage of anomaly detection with deep learning has been employed to construct appropriate models for detecting and mitigating adversarial attacks.Nonetheless, these attacks could deceive deep learning models into misclassifying the model results.This can lead to erroneous decisions, such as incorrect patient diagnoses and erroneous medicine administration.This paper addressed the effect of adversarial attacks on deep learning-based anomaly detection models and applied several safeguard techniques to defend networks against such adversarial attacks.The effectiveness of the suggested model was assessed in the face of adversarial attacks that were intended to give attackers control over the classification process used by the detection algorithm.The effectiveness of the suggested model was evaluated using the FGSM and BIM adversarial attack techniques.The suggested model showed its capacity to identify anomalies in the face of various adversarial attacks.More specifically, the evaluation results revealed that the model achieved an average F1 score of around 97% and an accuracy of 98% in the face of such attacks.However, as a future task, more scenarios regarding adversarial attacks need to be explored.In addition, more countermeasures and defense methods, such as reactive defense, should be investigated.
(a) Poisoning Attack: This type of attack occurs during the training phase of the proposed model, as shown in Figure1.In this attack, an adversary manipulates the training data to compromise the entire learning process.Modification, logic corruption, and data injection techniques can be used by an adversary to manipulate the training data.These capabilities allow an adversary to influence the proposed model's overall learning process and cause it to misclassify test results, which might result in the inappropriate treatment of a patient.(b) Evasion Attack: In evasion attacks, the adversary attempts to deceive the model by mounting adversarial attacks during the testing phase, as shown in Figure1.Such an adversary does not impact the training data, but they may access the proposed model to gain sufficient information.Consequently, they attack the proposed model and manipulate it so that it will misclassify the patient's condition.An adversary may use the fast gradient sign method (FGSM) attack or the basic iterative method attack (BIM) to perform evasion attacks.(c) Fast gradient sign method of attack (FGSM): This method uses the gradient of the underlying model.The original input is manipulated by subtracting or adding a small error in the direction of the gradient, with the intent of altering the behavior of the learning model.(d) Basic iterative method of attack (BIM): The BIM is an extension of the FGSM attack.

-
True positives are actual positives, including the number of cases that are correctly identified as anomalous by the model (TP).-False negatives are positives, including the number of cases that are misclassified as non-anomalous by the model (FN).-True negatives are actual negatives, including the number of cases that are correctly classified as non-anomalous by the model (TN).-False positives are actual negatives, including the number of cases that are misclassified as anomalous by the model (FP).

Figure 7
Figure7and Table8show an increase in accuracy, as well as in other performance metrics, after retraining the model for both subjects.

Figure 7 .
Figure 7.The model after retraining as a defense method.

Figure 7 .
Figure 7.The model after retraining as a defense method.

Table 1 .
Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
, x FOR PEER REVIEW 5 o Summary of the existing studies.

Table 1 .
, x FOR PEER REVIEW 5 o Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
, x FOR PEER REVIEW 5 o Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 1 .
, x FOR PEER REVIEW 5 o Summary of the existing studies.

Table 1 .
Summary of the existing studies.

Table 2 .
The dataset's sample sensor readings.

Table 6 .
The BIM attack results.
Subject No.With Attack or WithoutAccuracy Loss Recall Precision F1-Score

Table 6 .
The BIM attack results.

Table 8 .
Retraining the model as a defense method.

Table 8 .
Retraining the model as a defense method.