Next Article in Journal
Modeling of Hydrocarbon Migration and Hydrocarbon-Phase State Behavior Evolution Process Simulation in Deep-Ultradeep Reservoirs of the Mo-Yong Area, Junggar Basin
Previous Article in Journal
Analysis of Controller-Caused Aviation Accidents Based on Association Rule Algorithm and Bayesian Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Survey of Cross-Layer Security for Resource-Constrained IoT Devices

1
Department of Information Security, Faculty of Information Technologies, L.N. Gumilyov Eurasian National University, Astana 010000, Kazakhstan
2
Research Institute of Information Security and Cryptology, L.N. Gumilyov Eurasian National University, Astana 010000, Kazakhstan
3
Department of Computer Engineering, Astana IT University, Astana 010000, Kazakhstan
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(17), 9691; https://doi.org/10.3390/app15179691
Submission received: 18 July 2025 / Revised: 27 August 2025 / Accepted: 28 August 2025 / Published: 3 September 2025

Abstract

Low-power microcontrollers, wireless sensors, and embedded gateways form the backbone of many Internet of Things (IoT) deployments. However, their limited memory, constrained energy budgets, and lack of standardized firmware make them attractive targets for diverse attacks, including bootloader backdoors, hardcoded keys, unpatched CVE exploits, and code-reuse attacks, while traditional single-layer defenses are insufficient as they often assume abundant resources. This paper presents a Systematic Literature Review (SLR) conducted according to the PRISMA 2020 guidelines, covering 196 peer-reviewed studies on cross-layer security for resource-constrained IoT and Industrial IoT environments, and introduces a four-axis taxonomy—system level, algorithmic paradigm, data granularity, and hardware budget—to structure and compare prior work. At the firmware level, we analyze static analysis, symbolic execution, and machine learning-based binary similarity detection that operate without requiring source code or a full runtime; at the network and behavioral levels, we review lightweight and graph-based intrusion detection systems (IDS), including single-packet authorization, unsupervised anomaly detection, RF spectrum monitoring, and sensor–actuator anomaly analysis bridging cyber-physical security; and at the policy level, we survey identity management, micro-segmentation, and zero-trust enforcement mechanisms supported by blockchain-based authentication and programmable policy enforcement points (PEPs). Our review identifies current strengths, limitations, and open challenges—including scalable firmware reverse engineering, efficient cross-ISA symbolic learning, and practical spectrum anomaly detection under constrained computing environments—and by integrating diverse security layers within a unified taxonomy, this SLR highlights both the state-of-the-art and promising research directions for advancing IoT security.

1. Introduction

Securing IoT/IIoT at global scale requires a cross-layer lens that reconciles heterogeneous devices, tight memory and energy budgets, and adversaries who exploit interactions among firmware, networks, behavior, and trust controls. Accordingly, this review synthesizes mechanisms across layers and data granularities, evaluates accuracy/latency/deployability under resource constraints, and sets a research agenda oriented toward evidence-attached detection and zero-trust-aligned operation.

1.1. Motivation and Scope

The Internet of Things (IoT) is rapidly evolving into a global-scale ecosystem expected to encompass tens of billions of heterogeneous devices deployed across enterprise, industrial, and consumer domains. These devices range from simple sensors and actuators to complex embedded controllers and edge computing nodes, all of which must collaborate in real time across organizational and national boundaries [1,2,3]. The scale of this deployment enables data-driven optimization of production lines, logistics, healthcare services, and smart city infrastructures. However, it also significantly broadens the attack surface—extending from bootloader firmware and wireless protocols to cloud integration and even physical process control.
Recent security investigations have highlighted multiple attack vectors that undermine the integrity of IoT deployments: firmware backdoors intentionally inserted during the manufacturing supply chain, encrypted command-and-control channels designed to bypass perimeter defenses, and manipulation of sensor readings to evade traditional network monitoring tools [4]. These challenges are aggravated by the hardware limitations of microcontrollers and low-power system-on-chip (SoC) platforms typically found in IoT devices. Constrained in memory, energy, and bandwidth, such devices are often unable to support heavyweight cryptographic or intrusion detection solutions, leaving them particularly vulnerable.
Traditional single-layer defenses are insufficient in this context, as attackers can compromise different system layers simultaneously. Cloud-based security services, although powerful, introduce latency, bandwidth overhead, and additional risks related to data sovereignty and privacy. This creates a strong demand for a cross-layer security perspective, where vulnerabilities are addressed holistically across firmware, network, and behavioral layers. For instance, firmware-level analysis can expose insecure reused libraries or misconfigured bootloaders, while lightweight, deep learning, and graph-based intrusion detection can flag anomalies in constrained wireless traffic. Complementary research has also begun exploring behavior-based process-level anomaly detection models, continuous risk assessment using adaptive policy engines, and the adoption of zero trust architecture (ZTA) in distributed IoT environments [5,6].
Based on these insights, this study seeks to consolidate and advance state-of-the-art approaches by combining firmware-level vulnerability analysis with lightweight network-level anomaly detection, while ensuring computational efficiency suitable for resource-constrained environments.

1.2. Contributions and Article Organization

The contributions of this paper are threefold:
  • Cross-layer taxonomy: We propose a systematic taxonomy that organizes IoT security mechanisms by system level, algorithmic paradigm, data granularity, and hardware budget. This classification not only clarifies the design trade-offs of existing solutions but also offers a practical framework for positioning emerging research.
  • Focused benchmarking: We integrate state-of-the-art firmware security techniques (e.g., static analysis, symbolic execution, machine learning–based binary similarity) with representative network intrusion detection models. Their performance is empirically evaluated on real IoT hardware with respect to accuracy, latency, and deployability.
  • Research agenda and exploration goals: Instead of posing open-ended questions, we outline concrete directions for exploration:
  • Exploring compatibility challenges of binary analysis across diverse instruction set architectures (ISAs).
  • Investigating strategies to minimize the footprint of intrusion detectors while maintaining high accuracy in anomaly detection.
  • Advancing hybrid symbolic–learning procedures that combine the rigor of symbolic execution with the adaptability of machine learning.
  • Extending cross-layer approaches to incorporate behavioral anomaly detection and zero-latency verification aligned with ZTA principles.
The rest of this document is structured as follows. Section 2 provides an overview of the IoT context and threat landscape, covering typical IoT architectures, resource constraints, multi-layer attack surfaces, and security requirements for edge-bound deployments. Section 3 describes the survey methodology, including the search strategy, inclusion/exclusion criteria, and the review workflow based on the PRISMA 2020 guidelines. Section 4 introduces a taxonomy for cross-layer IoT security, organized by security level, data granularity, deployment model, and algorithmic paradigm. Section 5 focuses on firmware-level security, detailing static, symbolic, and machine learning techniques for binary similarity analysis. Section 6 examines integrated security at the network, behavioral, and policy levels, focusing on lightweight intrusion detection, deep/graph-based models, sensor-actuator/RF spectrum analysis, and zero-trust implementation. Section 7 concludes the paper, summarizes the findings, and outlines future research directions for IoT security.

2. Background and Threat Landscape

To understand the challenges in securing IoT systems, it is first necessary to examine the underlying architecture, the typical resource limitations of devices, and the evolving multi-layer threats that adversaries exploit.

2.1. Typical IoT Architecture and Resource Constraints

Most academic and commercial roadmaps still state that the IoT consists of three layers: the perception layer (or sensor layer), the network layer, and the application/cloud layer [7]. The Industrial Internet of Things (IIoT) often includes an edge control layer, but the basic data path (physical signal → digital data packet → cloud analytics) remains the same [8].
At the perception layer, a single node can have up to 256 kB of RAM, a CPU running at less than 1 GHz, and milliwatt-level battery capacity. Encryption, secure boot logic, and real-time control operations are still required [9]. The number of these billions of diverse devices far exceeds the number of hardware, operating systems, and wireless protocols that security solutions must support [10].

2.2. Multi-Layer Attack Surface

In this section, we use the term multi-layer attack surface to refer to the combined vulnerabilities across firmware, network, and behavioral layers. A detailed taxonomy is provided in later sections, but here we briefly introduce the concept for consistency. The purpose of this section is to show how attacks span across different technical levels, making cross-layer defenses indispensable.
Table 1 summarizes representative threats, consequences, and representative research at each layer of the IoT. This table provides a brief overview of vulnerabilities and illustrates the need for cross-layer security research. IoT systems face a variety of threats at multiple levels, including firmware vulnerabilities, weak network authentication, behavioral deception, and inadequate trust management. These attack surfaces collectively highlight the importance of a cross-layer security approach.
  • Firmware level. The ZigBee Light-Link worm showed that a single unpatched binary could set off a chain reaction that affected all lights in an installation [11].
  • Network level. More than 70% of tested BLE devices still default to “Just Works” pairing within local radio range. This makes it easier for attackers to forge connections and launch man-in-the-middle attacks [12]. In the backbone, common messaging protocols such as MQTT or AMQP make it easier for attackers to steal session tokens and send forged commands through the cloud [13].
  • Behavioral level. Real-world spoofing can evade conventional (network-centric) intrusion detection systems (IDS) because it manipulates sensor inputs rather than packet semantics; for example, laboratory GPS spoofing has been shown to alter the course of a GPS-guided drone [10,14].
  • Trust level. Zero Trust Architecture (ZTA) eliminates mutual trust between users, devices, and workloads, but most current pilots only consider user access to the cloud without checking the provenance of firmware or real-time sensor data [15].

2.3. Security Requirements for Edge-Bound Deployments

Strong IoT security requires more than just cloud-based cryptographic primitives or firewalls. Instead, each layer must have defense-in-depth built-in, especially for devices that operate offline or are occasionally connected.
  • Lightweight, with both integrity and privacy. IoT devices typically do not have enough RAM or code space to use standard protocols such as TLS or DTLS. Therefore, security suites must use cryptographic algorithms with smaller memory footprints and simpler handshakes, while still being able to protect against active attackers on the wireless channel [7].
  • Multi-layered, fine-grained access control. Security measures should be implemented not only at the cloud API level, but also in the firmware call graph, wireless MAC layer, and lightweight message brokers. If permissions are not enforced at each layer, there is a risk of lateral movement and escalation of the trust chain [8].
  • Continuously audit and grant as few access permissions as possible. Modern designs must rethink the identification, localization, and context of each abnormal state change based on zero trust principles [15,16]. However, these checks are currently performed only at the cloud edge, not in embedded runtime environments.
  • Local actionability and explainability are critical. Engineers need to understand the meaning of alerts without being able to view cloud logs or developer consoles. Compact on-device summaries (ideally using reduced or quantized language models) are still an emerging research topic [9], but can provide explainable diagnostic results without sacrificing latency or power consumption.
To meet these requirements, security measures must work together at multiple levels. Firmware-level risk assessment can support network-level traffic filtering, anomaly models can enable adaptive privilege reassignment, and policy enforcement points must consider all lower-level inputs. The next section explores how current work addresses (or fails to address) these requirements in both firmware and network contexts.

3. Survey Methodology

To ensure rigor and reproducibility, the survey follows a predefined protocol: broad database searches using a transparent query, explicit inclusion/exclusion criteria, and PRISMA-2020 screening and eligibility. The aim is to bound the corpus to resource-aware, cross-layer IoT security and make subsequent analyses traceable.

3.1. Search Strategy and Databases Queried

Systematic literature search queries were executed on eight digital libraries that jointly cover computer-security, embedded-systems and industrial-informatics research:
  • IEEE Xplore and ACM Digital Library—core venues for networked-system security and embedded software;
  • ScienceDirect and SpringerLink—major publishers of IoT-specific journals;
  • Web of Science and Scopus—multidisciplinary indexing and citation tracing;
  • Google Scholar—discovery of gray literature and early-view conference papers;
  • arXiv (cs.CR, cs.NI, eess.SP)—access to pre-publication manuscripts.
The search string combined four concept groups with Boolean AND operators:
(“IoT” OR “IIoT” OR “cyber-physical”) AND (“cross-layer” OR firmware OR network OR behavio* OR “zero trust”) AND (security OR “intrusion detection” OR vulnerability OR “access control”) AND (“resource-constrained” OR edge OR embedded OR “low power”).
Results coincided with the rise in BLE (Bluetooth Low Energy)/ZigBee security work and the formalization of zero-trust architecture. Reference lists of selected papers were snow-balled to capture additional studies.

3.2. Inclusion and Exclusion Criteria

To ensure methodological rigor, we established explicit inclusion and exclusion criteria before screening the literature. These criteria are summarized in Table 2, which outlines the types of studies that were retained or discarded during the review process.

3.3. Review Workflow

The systematic review followed the PRISMA 2020 guidelines [17] and was divided into three main stages:
  • Article identification. A comprehensive keyword-based search of major academic databases and preprint servers retrieved a total of 1084 articles. After automatically removing 142 duplicates by matching DOI, title, and arXiv ID, 942 unique articles remained for further review.
  • Screening. Titles and abstracts were evaluated according to the inclusion and exclusion criteria defined in Section 3.2. In this step, a total of 579 articles were excluded, including editorials, position papers, short abstracts, purely cloud-centric studies, and articles not related to IoT or resource-constrained environments. A total of 363 articles were reviewed in full text.
  • Eligibility review. Afterwards, we reviewed the full text to ensure methodological soundness and direct relevance to firmware, lightweight network defense, behavioral modeling, zero trust implementation, or device-side explainability. Papers that did not conduct empirical evaluation, papers that were superseded by new versions, or papers that only focused on traditional IT systems were excluded.
After eliminating 160 papers, a total of 196 studies were finally available for inclusion. Screening and eligibility assessments were conducted in parallel by two reviewers. In nine instances of disagreement, consensus was reached through discussion, and in one borderline case a third reviewer mediated. This consensus procedure is consistent with PRISMA-2020 guidelines and helps ensure transparency and reproducibility.

4. Taxonomy of Cross-Layer IoT Security

Cross-layer security in the Internet of Things (IoT) requires a systematic overview of where defenses are deployed and how they combine across different layers.
Figure 1 visually illustrates the IoT trust chain, from the perception layer through the network layer to the application layer, listing representative components at each layer (e.g., sensor and device hardware and software, near-field and long-range communications, cloud applications, interfaces, and device connectivity). The right-hand pane captures a layer-by-layer analysis, including attack and solution summaries, culminating in the trust chain endpoints. These insights inform the overall protection layers (device, traffic, cloud, application, and interface security), ultimately enabling secure IoT access control.
Based on this layered approach, we organize prior research along four orthogonal axes used in this section: security layer (firmware/device, network/transport, behavior/application, trust/policy), data granularity (binary, packet/stream, telemetry, RF spectrum), deployment model (on-device, hybrid edge cloud, cloud/OTA), and algorithmic paradigm (rule-based/machine learning/deep learning/optimization).
This categorization clarifies scope and limitations, highlights integration points between layers, and provides a common foundation for comparing heterogeneous approaches.

4.1. Security-Layer Perspective

The system architecture level at which these methods are applied is a key element in understanding a cross-layer security strategy for the Internet of Things (IoT). We use four levels of abstraction: firmware/device, network/transport, behavior/application, and trust/policy. Each layer presents its own threats and limitations, and security solutions vary in terms of data availability, computational power, and risk models. As shown in Figure 1, the IoT trust chain covers the perception layer, network layer, and application layer, each of which includes different components such as sensors, communication interfaces, and cloud services. The figure illustrates how security issues, from device software integrity to interface protection, are addressed layer by layer, laying the foundation for secure IoT access control and trusted system behavior.
  • Firmware/Device Layer. This layer includes security features that directly affect hardware peripherals, microcontroller instructions, or device firmware. This layer provides both active analysis (e.g., vulnerability detection) and runtime protection mechanisms (e.g., attestation and execution control). Many embedded systems lack memory protection or statically link binaries, making them vulnerable to attacks such as memory corruption, control flow hijacking, and binary reuse. Recent research has focused on finding parallels between symbolic execution [18], lightweight taint tracking [19], and semantic hashing [20] and firmware. Researchers have also explored shadow stack enforcement [21], TrustZone isolation [22], and secure boot [23] to ensure runtime security. These approaches are particularly important in mission-critical applications such as drones, industrial actuators, and medical sensors, where firmware compromise can have severe practical consequences.
  • Network/Transport Layer. The network layer encompasses communications based on Bluetooth Low Energy (BLE), ZigBee, LoRa, Wi-Fi, and MQTT-based cloud protocols. Security at this layer includes device fingerprinting, traffic injection, spoofing, replay, and protocol degradation. Existing defenses include Bluetooth Low Energy fingerprinting [24], packet timing and channel hopping anomaly detection [25], and lightweight intrusion detection systems (IDS) using compressed CNNs or GNNs [26,27]. Additionally, there are numerous examples of attacks targeting specific protocols, such as ZigBee Touch Link Hijacking [28], MQTT Hijacking [29], and Bluetooth Address Randomization Failures [30]. Many embedded IDS engines use bitfield encoding or Bloom filters [31,32] and operate on systems with less than 100 kB of RAM.
  • Behavioral/application layer. This layer monitors the consistency of higher-level application logic, actuator control, and sensor data. It provides both direct models of time-varying physical processes (e.g., the water level in a tank) and indirect models, such as policies for monitoring context-aware applications. Graph-based models, such as GAT-GRU [33], temporal sensor anomaly detectors [34], and hybrid context/state machine verifiers [35,36], are widely used in industry. To prevent privilege abuse, smart homes utilize tools such as ContexIoT [37] and Soteria [38] to record user interactions with devices. Modeling the overall behavior of a device is crucial for detecting hidden behaviors or policy violations that cannot be detected by firmware or packet filters.
  • Trust/Policy Layer. This layer adds logical structure and architectural guidance to ensure reliable system operation across different domains. It includes zero-trust identity enforcement, continuous authentication, remote attestation chains, and federated policy control. Systems such as WAVE [39], XToken [40], and MPA [41] aim to achieve fine-grained authorization and feature propagation in resource-constrained multi-tenant systems. Trust layer approaches also include device access protocols [42] for protection against forged configuration attacks (e.g., ghost device attacks) and distributed ledger systems for trust traceability [43]. These approaches are particularly important when IoT devices are owned by multiple parties and policies need to be coordinated without a shared physical infrastructure.

4.2. Data Granularity

Cross-layer security solutions for the Internet of Things (IoT) process input data at different levels of detail. This affects how well they detect threats, their operational costs, and the difficulty of integration. There are four levels of data granularity: binary level, packet/stream level, time series telemetry level, and radio frequency spectrum (RF) level. Each level has its own advantages and disadvantages that affect the design and use of IoT security measures.
Binary-level analysis looks for static vulnerabilities, backdoors, and cryptographic errors in firmware images or executable code (usually in raw or disassembled format). The tool used in [44] uses a secure stream processing framework to inspect the firmware without executing it. This ensures that vulnerabilities can be found even when the device is not connected to the Internet. Ref. [45] also discusses trusted execution environments (TEEs), which can assist binary analysis by providing a secure code inspection setup that is independent of the rest of the system. These methods are ideal for detecting zero-day vulnerabilities, but lack runtime behavioral context, making it difficult to detect dynamic threats. For example, static analysis in a trusted execution environment (TEE) [45] can find firmware errors but not problems that occur during program execution.
Packet and flow-level inspection is a common feature of network intrusion detection systems (NIDS), which analyze PCAP files or network traffic in real time. Approaches include statistical feature extraction [46], protocol-specific filtering for protocols such as MQTT [47], and compressed stream embedding using advanced models such as graph neural networks (GNNs) [47]. These approaches strike a good balance between coverage and speed, making them well suited for detecting network-based issues. However, since they primarily analyze packet metadata, they may miss aspects of semantic intent or high-level device logic. For example, ref. [46] discusses edge computing frameworks that allow real-time traffic analysis. However, these approaches struggle with context-sensitive attacks and require more information about the application layer.
Time series telemetry analysis uses data such as actuator state changes, energy consumption, or sensor values to understand device functionality and identify issues. Increasingly, users are leveraging this level of detail for general modeling and predictive anomaly detection [48,49]. For example, ref. [48] discusses edge computing applications that analyze telemetry data to detect behavioral changes, providing it with rich semantic context. However, this approach requires longer observation windows and complex correlation models, which can be challenging for resource-limited IoT devices. Systems such as TEEs used to protect telemetry processing in [49] are often migrated to edge or cloud infrastructure due to their high computational power requirements.
Radio frequency (RF) spectrum analysis is critical for detecting problems in wireless networks, especially in the presence of denial of service (DoS) attacks, malicious base stations, or protocol spoofing [50]. These systems analyze raw I/Q signal data or spectrum fingerprints to detect unwanted transmissions or interference. RF detection is applicable to any protocol and can therefore detect threats that are not present in traditional protocol stacks such as ZigBee and LoRa. For example, ref. [50] describes a TEE-based framework for protecting wireless communications by analyzing RF signals. However, RF analysis is difficult to apply to low-power devices because it requires a lot of computational power and therefore typically requires specialized hardware.
When choosing the level of detail in the data, trade-offs must be made between latency, interpretability, and hardware feasibility. New hybrid approaches combine simple flow-level inspection with deeper binary or RF analysis to enable immediate action when suspicious events are discovered [50]. For example, a system could start with packet-level monitoring to detect threats and then leverage RF spectrum analysis to more precisely detect wireless attacks. This can improve both efficiency and accuracy.

4.3. Implementation Models and Algorithmic Paradigms

This section explores the practical aspects of how these security solutions are actually implemented, based on a layer-oriented approach and data granularity. We will therefore discuss deployment models, which define the placement of defensive measures within the IoT stack, and algorithmic paradigms, which determine how threats are detected and mitigated in practice.
To achieve practical integration of IoT security, it is important to consider not only where defensive measures are implemented but also how their algorithms operate. Deployment models determine the physical and architectural placement of defensive mechanisms within devices, edge nodes, and cloud infrastructure. Algorithmic paradigms define computational strategies for implementing intrusion detection, anomaly detection, and adaptive defenses within resource-constrained environments. Together, these two perspectives form the foundation for the practical implementation of cross-layer IoT security.

4.3.1. Deployment Models

IoT security solutions must adapt to the hardware and deployment environment of edge devices. There are three primary deployment models: on-device deployment, hybrid or collaborative deployment, and cloud-based over-the-air (OTA) delivery. Each type addresses specific operational and security requirements.
With on-device deployment, security logic is implemented directly on a microcontroller (MCU), system-on-chip (SoC), or edge gateway. This allows devices to be protected in real time and with very low latency. This model is best suited for IoT applications that need to run quickly but cannot offer many features due to low processing power and memory capacity. Firmware integrity checking [51] and lightweight intrusion detection systems (IDS) [47] are two examples of technologies that are often deployed on devices. For example, ref. [51] describes the trend among IoT developers to use ARM TrustZone to secure devices. On the other hand, ref. [47] describes a trusted execution environment for a lightweight IDS. Newer platforms can leverage advanced hardware such as ARM TrustZone [52] to separate critical tasks. These systems provide strong protection but must be optimized for specific hardware to run within limited resources.
Hybrid implementations distribute security tasks between cloud infrastructure and edge devices. We perform simple or real-time analysis on our own computers, but move computationally intensive activities (such as graph-based threat detection or advanced analytics) to the cloud. This approach strikes a balance between edge responsiveness and scalable analysis. For example, ref. [46] discusses an edge computing framework that supports local packet filtering, while a cloud-based system [48] performs more complex threat analysis. For centralized correlation, the system can upload PCAP files or extracted features, or local anomaly alerts can trigger cloud-based processing [47]. Hybrid implementations achieve scalability but require secure network connectivity and strong cloud security.
Cloud-based OTA implementations can implement vulnerability scanning, security updates, and anomaly detection through regular cloud updates. This model enables rapid response to new threats, but can be risky if the update channel is not secure. For example, ref. [53] discusses an open TEE framework that supports secure OTA updates. Digital signatures and encrypted sessions are essential to ensure the security of the update chain [50]. OTA technology can provide updated anomaly detection models for edge devices [49]. However, flaws in the update process may allow attackers to inject malicious code. Therefore, a strong authentication mechanism is required.
From a hardware perspective, the choice of deployment model is inevitably constrained by the available computational resources on the IoT platform. Consumer devices are typically based on microcontrollers (MCUs) such as ESP32 or STM32, while industrial IoT implementations often use more powerful edge accelerators such as TPUs or NVIDIA Jetson boards. These heterogeneous hardware configurations mean that security architectures cannot be applied uniformly but must be adapted to the device class. In practice, this adaptation often requires pruning, quantization, or simple optimization of anomaly and intrusion detection models to ensure feasibility within tight memory and power budgets. Frameworks such as TinyML and platforms such as EdgeImpulse [54] demonstrate how compressed inference runtimes can be efficiently executed on MCUs, enabling real-time anomaly detection directly at the edge. This hardware-centric perspective complements the previous discussion of deployment models by showing that practical feasibility depends not only on the location of the security logic (on-device, hybrid, or cloud-based), but also on how it fits within the computational constraints of the target device.
When choosing a deployment strategy, latency, update frequency, hardware capabilities, and operational considerations must be taken into account. Recent research has explored techniques for dynamically switching between models, including sending complex analytics to the cloud only when local monitoring detects suspicious conditions [50]. This flexible approach maximizes resource utilization while maintaining high security standards.

4.3.2. Algorithm Paradigms in Cyberspace Security Applications

Cybersecurity has already implemented significant qualitative leaps in the use of artificial intelligence (AI) algorithms toward the fulfillment of such core functions as intrusion detection, security posture assessment, and related dimensions of defense strategy enhancement. Types of algorithms discussed in this section that have proven useful toward the advancement of cybersecurity include machine learning (ML), deep learning (DL), and optimization algorithms.
  • Machine Learning Paradigms
Most of the algorithms used in network security are machine learning because they are easy to implement and support good computer implementation. The methodologies that these methods embrace include supervised, unsupervised, semi-supervised, and reinforcement learning. These algorithms find their best application where there is a need for fast sorting as well as where the computers are not performing at a high level.
Supervised learning: K-nearest neighbor (K-NN), support vector machine (SVM), naive Bayes (NB), and decision tree (DT) are the most popular intrusion detection algorithms. For example, Al-Omari et al. [55] developed a DT-based intelligent intrusion detection model that achieved 98% accuracy on the UNSW-NB15 dataset by evaluating security variables using the Gini coefficient. This reduces computational time compared to typical machine learning methods. Sapavath et al. [56] also used NB to detect cyberattacks in virtualized wireless networks in real time. They achieved this by modeling an adaptive agent system and achieved 99.8% accuracy. Abuali et al. [57] developed a DT-based intelligent intrusion detection model that achieved 98% accuracy on the UNSW-NB15 dataset by evaluating security variables using the Gini coefficient. This reduces computational time compared to typical machine learning methods. Sapavath et al. [56,58] also used a black box model to detect cyberattacks in virtualized wireless networks in real time. They achieved this by modeling an adaptive agent system and achieved 99.8% accuracy [59] combined support vector machines (SVM) with deep learning for intrusion detection. They achieved 98.50% accuracy on the CSE-CIC-IDS2018 dataset, showing that this approach is well suited for detecting complex patterns. These models are good at ranking known attack patterns but struggle with high-dimensional data or novel attacks [57].
Unsupervised learning: To protect against distributed denial of service (DDoS) attacks, algorithms such as K-Means are used. Gu et al. [58] developed a semi-supervised K-Means algorithm based on hybrid feature selection (SKM-HFS), which achieved the lowest detection latency on datasets such as CICIDS2017. Unsupervised algorithms are well-suited for detecting unusual patterns in unlabeled data, but may not be as accurate in complex attack scenarios [60].
Semi-supervised learning: Nishiyama et al. [61] developed the SILU algorithm, which uses large amounts of unlabeled data to improve supervised classifiers such as SVMs and Random Forests (RFs). This approach improves detection accuracy and reduces the number of false positives, making it suitable for real-time applications with limited labeled data [62].
Reinforcement Learning (RL): Reinforcement learning is increasingly being used to develop defense methods. For example, Liu et al. [63] used a deep recurrent Q-network (DRQN) to make defense decisions in a stochastic game system. This approach reaches the optimal policy faster than standard methods. Reinforcement learning is well-suited to dynamic situations because it can adapt to changing threats, but it requires a lot of computational power [64].
Representative IoT evidence: Doshi et al. used classic machine learning traffic features to detect consumer IoT DDoS attacks and demonstrated high accuracy of a lightweight model.
  • Deep Learning Paradigms
Deep learning algorithms, especially those using neural networks, are well-suited to processing high-dimensional, complex data. This makes them ideal for detecting sophisticated attacks and assessing security posture.
Convolutional Neural Networks (CNNs): CNNs are well suited to detecting cyberattacks in many different domains. Khaw et al. [65] developed a one-dimensional CNN with an autoencoder (AE) for protecting transmission line relays. They achieved 100% accuracy on the OPAL-RT HYPERSIM dataset. Ho et al. [66] developed a CNN-based multi-class classification model that achieved 99.78% accuracy on the CICIDS2017 dataset by fine-tuning hyperparameters through repeated testing. Nazir et al. [67] developed a hybrid CNN-LSTM architecture for IoT threat detection that achieved 99.20% accuracy on the IoTID20 dataset by capturing both spatial and temporal variables. CNNs perform well on structured data but require extensive parameter tuning [68].
Different types of recurrent neural networks (RNNs): Long short-term memory (LSTM) and gated recurrent unit (GRU) are examples of RNNs used to analyze sequential data. Hossain et al. [69] used LSTM to detect SSH and FTP brute force attacks, achieving 99.88% accuracy on CICIDS2017. Ma et al. [70] developed XBiLSTM-CRF for network security entity detection, achieving 90.54% accuracy on the open source NER dataset. Bukhari et al. [71] used a federated learning approach of sparse CNN and bidirectional LSTM (SCNN + Bi-LSTM) to improve defense strategies, achieving 97.80% accuracy on the WSN-DS dataset while maintaining training confidentiality. These models are well suited to capture temporal relationships but require a lot of computational power [72].
Autoencoders (AEs) address the data imbalance problem in attack detection. Kunang et al. [73,74] used AE for feature discovery and achieved 86.96% accuracy on NSL-KDD. Qazi et al. [75] developed a stacked asymmetric deep autoencoder (S-NDAE) and achieved 99.65% accuracy on the KDD CUP99 zero-day attack test. AEs are helpful for feature learning but may have difficulties with highly imbalanced datasets [76]. Generative adversarial networks (GANs) improve datasets by creating fake data. Ding et al. [77] developed a table-assisted classifier GAN (TACGAN) and achieved 95.86% accuracy on CICIDS 2017. GANs make models more robust but must be carefully tuned to avoid mode collapse [78].
Representative IoT evidence: Meidan et al.’s N-BaIoT uses deep autoencoders to flag botnet traffic from compromised IoT devices (Mirai/BASHLITE), achieving near-instant detection.
  • Optimization Algorithm Paradigms
In the field of cybersecurity, optimization algorithms such as classical methods and swarm intelligence (SI) methods are very important for tuning parameters and selecting features.
Classical optimization methods such as genetic algorithms (GA) and simulated annealing are often used for feature selection. Seth and Chandra [79] combined KNN with modified gray wolf optimization (MGWO) and achieved 99.87% accuracy on the Solaris dataset. Anonymous et al. [74] combined SVM, KNN, RF, DT, LSTM, and ANN with fuzzy clustering to find intruders. They improved feature selection by using fuzzy clustering and achieved 99.10% accuracy on CICIDS2017. These strategies are well suited for simple problems but are not suitable for more complex non-convex problems [80].
Swarm intelligence methods: SI methods such as particle swarm optimization (PSO) and gray wolf optimization (GWO) can improve model performance and feature selection. Kan et al. [81] developed an adaptive PSO-CNN (APSO-CNN) for IoT intrusion detection with an accuracy of 96%. Chohra et al. [82] used the Chameleon model with ensemble learning and achieved an F-value of 97.3% on the IoT-Zeek dataset. SI techniques are well suited for solving multidimensional problems, but their convergence may take a long time [83].
Game theory and stochastic models: Game theory helps improve defense techniques by simulating the interactions between attackers and defenders. Hyder and Govindarasu [84] used game theory to protect smart grids from cyber attacks and reduce the probability of attacks. Bhuiyan et al. [85] developed a risk-free Stackelberg game model that reduced computational time by 71%. These models can help develop strategies, but they must make correct assumptions about the attacker’s behavior [86].
Benchmark IoT evidence: Benmalek et al. hybridize an enhanced PSO with deep models for real-time IoT cyber-attack detection on a benchmark IoT, shedding light on the place of metaheuristics in the constrained tuning/feature selection space.
Table 3 sums up disclosed performances by different algorithms in intrusion detection, security scenario assessment, and defense optimization. It also lists their datasets, accuracy, and main advantages.
As summarized in Table 3, three themes emerge: first, classical ML baselines provide very strong accuracy–latency tradeoffs on structured datasets, though they fail under conditions of concept drift and novel attack families; second, DL models discover spatial–temporal structure and frequently outperform ML on complex traffic and telemetry with higher compute and tuning effort—thus good for gateways and edge nodes rather than tiny MCUs; third, optimization and game-theoretic methods complement detection by feature selection, hyper-parameter tuning or defense allocation increasing robustness at the cost of increased design complexity. These lead us to propose hybrid pipelines motivated by lightweight on-device screening paired with deeper edge/cloud analytics.
Machine learning algorithms have gained attention in the attack detection community due to their simplicity and their ability to perform well in resource-constrained environments. SVM + DL [59] and other hybrid methods are more robust to complex datasets but require higher computational power. Deep learning algorithms such as CNN, LSTM, and hybrid models such as CNN + LSTM [67] and SCNN + Bi-LSTM [71] perform better on complex and imbalanced datasets. They can be used in IoT and wireless sensor networks but require high computational power and sophisticated parameter tuning [66,69]. Optimization methods such as SI and fuzzy clustering [74] can improve feature selection and strategic decision making. Models such as APSO-CNN and Stackelberg games show potential in IoT and smart grid applications [81,85]. The choice of algorithm depends on the specific security task, the specifics of the dataset, and the performance limitations of the computer. Future research should explore hybrid models that leverage the best parts of ML, DL, and optimization techniques to address emerging threats and data imbalance. This trend of adapting classical ML/DL paradigms to resource-constrained IoT devices has also been recognized in recent surveys, which emphasize that IoT-specific constraints—limited memory, energy, and heterogeneity—necessitate tailored versions of otherwise general algorithms.
Figure 2 brings together the entire pipeline across On-Device/MCU → Edge/Gateway → Cloud/Control Plane to make concrete the implementation view. The mechanisms are grouped by tiers so that the components represent where their most natural place of execution might be—running firmware analysis and lightweight IDS on constrained MCUs; deep/graph IDS, sensor–actuator consistency, and RF monitoring at the edge; trust evaluation, policy decision, and enforcement in the control plane. Solid arrows denote data flow (evidence-attached alerts, anomaly features, behavioral/RF signals); return path contains control/feedback from trust and policy back to enforcement (fine-grained tokens, rate limits, isolation) to devices as OTA patches or model updates. It will stack MCU-side and cloud-side blocks vertically to minimize horizontal span but keep signal collection left-to-right logic through to decision and action. The four design axes—data granularity, algorithmic paradigm, deployment, hardware budget—are called out with dotted boxes so that a reader can immediately relate methods to where they execute and why.

5. Firmware-Level Security

Academic and industrial circles have, in recent years, directed significant attention toward the security evaluation of embedded firmware. Techniques for vulnerability discovery in conventional computer systems typically prove inapplicable to firmware assessment owing to the considerable diversity among embedded devices, their specialized functionalities, and their constrained execution contexts. Static analysis and symbolic execution represent two principal methodologies for program examination that can identify security flaws by analyzing a program’s architecture and logic (without running the code). This renders them quite valuable for assessing the security of embedded firmware.
Figure 3 presents an organization of the typical approaches that would facilitate your understanding of the technical landscape about firmware similarity analysis techniques applied in vulnerability detection. They are usually organized by the type of analysis as static, dynamic, hybrid, and machine learning with accuracy versus coverage versus runtime applicability versus applicability to pure binary firmware paradigms.
Static analysis examines the structure and logic of firmware binary code to find potential security vulnerabilities, such as buffer overflows or hard-coded passwords. It does not require a runtime environment, making it well suited for embedded devices. However, it can produce false positives due to the lack of context. Recent research has found new algorithms and techniques to make it more efficient and accurate.
Symbolic execution checks for vulnerabilities by simulating all possible execution modes of a program. It is particularly suitable for finding complex logic errors. However, firmware analysis faces challenges due to path explosion and peripheral device interaction modeling. Recent research has improved performance through code slicing and domain-related knowledge.
Recent research has made firmware security analysis more automated and accurate. For example, a framework for detecting application privacy leaks uses static analysis to identify privacy issues in IoT applications. Hardware-based fuzz testing uses hybrid modeling to improve path exploration. Reverse engineering methods for smart home IoT firmware can discover network vulnerabilities, while tracing hardware information flows helps analyze the security of firmware-hardware collaboration. These improvements significantly promote the discovery of vulnerabilities in embedded device firmware and are both accurate and scalable.

5.1. Static and Symbolic Analysis Techniques

This section covers the ideas, applications, current research, and limitations of static and symbolic analysis methods. A comprehensive evaluation will be conducted in this section, focusing on their usefulness and associated challenges in analyzing the security of embedded firmware.

5.1.1. Static Analysis Techniques

Static analysis techniques examine the binary code or intermediate representation of a program to determine control flow, data flow, and semantic information. They can also discover potential security vulnerabilities such as buffer overflows, command injection, or hard-coded credential backdoors. Static analysis does not require the program to be running, so it is suitable for embedded devices with limited resources, especially when the firmware source code is not available or has limited documentation. The typical steps of firmware security analysis are as follows:
Deploy the firmware and make it available for use: Extract the executable code or application from the firmware image using tools such as Binwalk [94] or Binary Analysis Next Generation (BANG) [95]. These programs can identify file systems, compression schemes, or binary structures to extract the firmware into an analyzable binary.
Code conversion: Convert the binary code to assembly language or an intermediate representation such as VEX or LLVM-IR using reverse engineering tools such as IDA Pro [96] or Ghidra [97]. These tools are available for many different instruction set architectures (ISAs), such as ARM, MIPS, and x86. They can also find function call links, string references, and control flow graphs (CFGs).
Program analysis: Uses static analysis methods such as fuzzy hashing [98], taint analysis, or control flow analysis to identify program features and potential security vulnerabilities. For example, fuzzy hashing can find re-used third-party libraries or code snippets that are known to be vulnerable. On the other hand, taint analysis tracks the flow of data from user input to sensitive processes to identify instruction injection or buffer overflows.
Vulnerability detection: Uses special techniques to search for backdoors, memory corruption, or other security issues by analyzing program behavior.
Prior research and surveys have consistently demonstrated that static analysis is highly effective in detecting vulnerabilities in embedded device firmware, particularly in uncovering backdoors and taint-based attacks [99,100,101,102,103,104,105,106].
Here are some examples of research:
  • Costin et al. [99] developed a method for analyzing large numbers of firmware files using fuzzy hashing. The method finds backdoor vulnerabilities such as weak keys and hardcoded credentials by finding similarities between firmware files. By examining 693 firmware images, 38 known vulnerable files were identified. However, the file-level granularity makes the method less precise, and the additional effort makes it unsuitable for large-scale firmware.
  • Stringer [100] developed a method for comparing static data in commercial off-the-shelf (COTS) firmware to detect hardcoded credentials and undocumented functions. The method finds potential backdoors by comparing static strings in binary files. The method is well suited for large-scale analysis, but has a high false positive rate.
  • HumIDIFy [101] proposed a semi-supervised learning method for finding binary functions in firmware and comparing them to predicted functions to detect hidden functions or backdoors. Tests show that the analysis of COTS firmware has a low false positive rate, although the method requires a high-quality training dataset.
  • Firmlice [102]: Uses static analysis to build a program dependency graph (PDG), uses authentication slices from entry points to privileged operations, and combines symbolic execution to check fixed-value routing restrictions to find backdoors that bypass authentication. The method is very accurate, but only works for certain types of backdoors and is difficult to apply to other types of vulnerabilities.
  • DTaint [103]: A method for vulnerability detection using taint analysis is proposed, which traces data from sensitive inputs (sources) to dangerous functions (sinks) by building data flow graphs of intra-procedural and inter-procedural data flows, and finds buffer overflows and instruction injections. DTaint’s bottom-up analysis can detect vulnerabilities more accurately, but takes longer for complex firmware.
  • SainT [104]: In IoT applications, it monitors sensitive data (such as device status or user information) sent to external outputs and finds possible data leaks. SainT works well for type I firmware, but does not work well for type II and type III firmware because these formats are too complex.
  • Zheng et al. [105]: They combined protocol parsing with static taint analysis to find protocol fields and key functions in firmware, thereby accelerating the search for taint-based vulnerabilities. This strategy reduces the amount of analysis by selecting appropriate keywords, thereby improving efficiency.
  • Karonte [106] proposed a multi-binary static analysis method to simulate the communication between firmware components (such as web servers and background services) to discover insecure data transmission or logic vulnerabilities. Karonte solves some of the problems of single-component analysis, but needs to better handle more complex firmware.
  • PrivacyGuard [107]: A framework has been developed for detecting privacy leaks in applications. It uses static analysis to check trigger condition-action rules in IoT applications to identify privacy risks such as location tracking and activity analysis vulnerabilities. The system is well suited for evaluating large IoT application datasets and significantly improves the level of privacy protection [107].
  • Smart Home IoT Firmware Analysis and Exploitation Framework [108]: The framework supports reverse analysis of smart home IoT firmware to identify network weaknesses (such as unencrypted communication and weak authentication) and missing protection mechanisms (such as NX and Stack Canary). This helps in analyzing the security of smart home devices [108].
  • Hardware Information Flow Tracing [109]: A simple path-based hardware information flow tracing method is proposed. It uses static analysis to examine firmware-hardware interactions and detect potential information leaks and dangerous operations. The method works well on low-resource devices and is fairly accurate [109].
One of the advantages of static analysis is that it does not require a runtime environment. This makes it suitable for embedded devices with limited resources and can find many types of vulnerabilities. However, it also brings some problems to firmware analysis:
High false positive rate: Static analysis often produces false positive results because firmware code is complex and diverse and lacks sufficient context information. For example, fuzzy hashing may mistakenly mark code that has the same functionality as the vulnerable code as vulnerable [99].
Not very helpful for Type II and Type III firmware: Type I firmware is usually based on standard file systems such as Squashfs, which are easy to decompress and analyze. On the other hand, Type II and Type III firmware lack a standard format or clear loading standards, which makes it difficult for static analysis tools to find code segments and data structures [110].
Insufficient analysis of the interactions between different parts of the firmware: Firmware is usually composed of many parts that work together, such as a web interface and background functions. Current static analysis methods only consider one component at a time, ignoring the interactions and data flows between components that may lead to logical vulnerabilities [106].
Computational complexity: Static analyses are difficult to perform at scale because they require processing complex control flow or data flow graphs and are therefore inefficient [103].

5.1.2. Symbolic Execution Techniques

Symbolic execution considers all different execution paths for a program input and represents them as symbolic values. Then, a constraint solver (such as Z3) is used to check for vulnerabilities by checking the path constraints. It is able to precisely analyze how a program runs, especially in detecting hard-to-find vulnerabilities such as memory corruption and authentication bypass backdoors. On the other hand, path explosion and environment dependencies make analysis of embedded firmware more difficult.
When analyzing firmware, the typical steps of symbolic execution are as follows:
  • Firmware preprocessing: As with static analysis, the executable code is extracted from the firmware and converted into an intermediate representation, such as LLVM-IR or VEX.
  • Input symbolization: The program input (e.g., user input or peripheral input) is converted into symbolic variables.
  • Path analysis: All possible paths of the program are explored using a symbolic execution engine (e.g., KLEE [111] or angr [112]) and path constraints are imposed.
  • Constraint solving: A constraint solver is used to check whether the input can trigger the vulnerability by checking the satisfiability of the path constraints.
When analyzing embedded firmware, symbolic execution is often used to investigate specific hardware types or vulnerabilities. Here are some research examples:
  • Avatar [113]: A hybrid execution framework that leverages both symbolic execution and full-system simulation is proposed. It uses S2E [114] as a symbolic execution engine, dynamically intercepts simulation events (such as memory access or instruction execution), and selectively symbolizes important code segments to find vulnerabilities. This approach works for class I, II, and III firmware, but not for all peripherals.
  • FIE [115]: Based on KLEE, a symbolic execution tool for MSP430 microcontrollers is developed that can simulate memory and interrupt operations to find security vulnerabilities. FIE only works for MSP430 and has many bugs due to path explosion.
  • FirmUSB [103]: This is a USB controller firmware for the 8051/52 architecture that leverages domain knowledge and symbolic execution to create functional models and find BadUSB vulnerabilities. FirmUSB limits the scope of symbolic execution, making it seven times more efficient than conventional symbolic execution.
  • Inception [116]: Created a symbolic virtual machine based on KLEE to transfer firmware code to LLVM-IR to find memory leaks in ARM Cortex-M3 firmware. Inception can analyze assembly and library functions, but it does not work well for complex firmware.
  • Gerbil [117]: Searches for privilege separation vulnerabilities by using backward slicing to find potentially vulnerable code regions and combining symbolic execution to search deep paths. Gerbil avoids path explosion by skipping complex library functions, but requires rich domain knowledge.
  • FIoT [118]: Uses static analysis and symbolic execution to build firmware control flow graphs to represent the paths of critical function inputs. Dynamic fuzzing is also used to find memory corruption vulnerabilities [119]. FIoT limits the scope of analysis but has difficulties in managing peripheral information.
  • HD-FUZZ [120]: Proposes a hardware-based firmware fuzzing framework that combines hybrid MMIO modeling with symbolic execution to improve pathfinding efficiency. This significantly increases the number of vulnerabilities in the firmware of small embedded devices [120].
The strengths of symbolic execution lie in its high accuracy and ability to find complex logic errors. However, it is not suitable for all situations for the following reasons:
  • Path explosion: Firmware code with complex control flows and loops can generate exponentially many paths, placing higher demands on computer performance [115].
  • Peripheral dependencies: Embedded firmware often works with peripherals such as MMIO or interrupts, which are difficult for symbolic execution engines to model correctly. This can cause the analysis to abort or lead to incorrect conclusions [113].
  • Complex environment setup: Symbolic execution requires complex tools such as disassemblers, instrumentation tools, and constraint solvers, which makes it difficult to use on resource-constrained embedded devices [117].
  • Limitations of constraint solving: Constraints imposed by cryptographic algorithms or complex hash functions in firmware are sometimes difficult to solve, reducing its practicality [103].

5.1.3. Analysis and Evaluation

Static analysis and symbolic analysis each have their own advantages and disadvantages when analyzing embedded firmware security [121]. Table 4 summarizes their technical characteristics and some typical research results.
Applicability: Static analysis is well suited for analyzing large amounts of firmware, especially class I firmware, as it is independent of the runtime environment. However, it struggles with class II and class III firmware, which require more robust base address detection and file format analysis [110]. Symbolic execution works well on some platforms, such as MSP430 or 8051, but does not work well when very broad coverage is required.
Accuracy and false positives: Static analysis suffers from high false positives, especially when using coarse-grained methods such as fuzzy hashing [99]. Symbolic execution finds fewer false positives through detailed path analysis, but may miss vulnerabilities due to path explosion or peripheral emulation issues [115].
Efficiency: The computational complexity of static analysis increases significantly with the size of the firmware, so feature extraction and matching methods must be optimized [103]. The efficiency of symbolic execution is limited by the number of paths and the difficulty of considering constraints. Slicing or domain knowledge is required to simplify the analysis [117].
Scalability: Static analysis has made great progress in analyzing interactions between multiple binaries (e.g., Karonte [106]), but analyzing the complexity between components remains difficult. Hardware emulation and environment configuration limit the scalability of symbolic execution, thus requiring lightweight execution frameworks.

5.1.4. Future Improvements

Future improvements to firmware security analysis should focus on supporting non-standard formats (Type II and Type III) through robust base address detection and file format analysis techniques, simplifying the processing of diverse binaries. Static analysis frameworks can be extended to capture interactions between firmware components in real time, thereby detecting cross-module logic vulnerabilities [125]. Furthermore, the efficiency of symbolic execution must be improved through methods such as code slicing, selective symbolization, and domain-specific constraints to minimize the challenges posed by path explosion and peripheral emulation [126]. Finally, combining static and symbolic methods with dynamic testing methods such as fuzz testing is a promising direction, as runtime feedback can improve the accuracy of static analysis and guide symbolic path exploration to high-impact execution traces.

5.2. Machine Learning-Assisted Binary Similarity Analysis

Cui et al. [124] pointed out that many firmware updates contain third-party libraries that have known security vulnerabilities that persist for years. Their study found that 80.4% of vendor-provided firmware had known vulnerabilities. This poses a significant problem for embedded systems that control safety-critical infrastructure. The Mirai botnet [127] demonstrated that exploitation of such vulnerabilities could cause widespread disruption of public systems, such as a distributed denial of service (DDoS) attack against DNS servers.
Finding vulnerabilities in embedded systems is difficult because they use many different hardware platforms, have their own implementations, and run in limited environments. Dynamic testing tools are widely available on desktop computers, but embedded firmware often lacks easily discoverable source code, standardized instruction set architectures (ISAs), or consistent compilation parameters [122]. This renders traditional tools useless. Researchers have increasingly used binary similarity analysis to find known vulnerabilities in firmware and have made significant progress. However, binary similarity techniques are difficult to apply to embedded firmware due to the inaccessibility of source code, the presence of different instruction sets (ISAs), and the variability of compilation settings [128,129,130,131].
This section focuses on how to use machine learning to support binary similarity analysis to find firmware vulnerabilities. This paper introduces some of the most prominent methods proposed between 2015 and 2024. These methods were published in leading security and software development forums (such as CCS, NDSS, RAID, ASIACCS, DIMVA, FSE, ASE, ICSE, TSE, ESORICS, IEEE Access, and arXiv) and analyzed within a unified four-stage framework. This report covers the technical characteristics, datasets, CVE (Common Vulnerabilities and Exposures) search performance, and computational complexity of these methods. Finally, this paper proposes some open research issues.

5.2.1. Necessity of ML-Assisted Binary Similarity Analysis

  • Why binary similarity is needed at the firmware level.
Firmware binaries are stripped, optimized, and compiled for diverse ISAs, so a single vulnerable routine appears in many surface forms; purely syntactic signatures and naïve graph matching therefore miss semantically equivalent functions compiled under different settings [123]. Graph-based embeddings over attributed control-flow graphs (ACFGs) are designed to preserve structural/relational semantics across such transformations, making similarity search viable when syntax diverges [130]. Hand-crafted statistical descriptors over ACFGs were the first step toward this robustness but remain sensitive to inlining and compiler idiosyncrasies, which motivates learning richer representations [113].
  • Closing the gap between breadth and evidence.
Static analyzers scale but often lack behavioral evidence and thus inflate analyst workload; in contrast, symbolic execution yields precise witnesses but struggles with path explosion and incomplete models of peripherals [132]. Using learned similarity as a front-end turns the global hunt into a nearest-neighbor retrieval problem that can surface high-yield candidate functions quickly, after which targeted slicing or feasibility checks can attach concrete, auditable artifacts to each alert. PDG-driven authentication slicing exemplifies how post-retrieval evidence can be produced for security-critical flows [102].
  • What makes ML similarity robust in practice.
ACFG embeddings trained on diversified builds learn invariants that survive ISA and optimization variance, enabling cross-architecture matching at scale [130]. Token/sequence models learn instruction-level semantics and can remain useful even when control-flow is distorted, complementing graph approaches in obfuscated or flattened binaries [133]. Self-attention over token streams further improves the model’s ability to focus on security-relevant motifs such as bounds checks and sanitizer calls [134]. Where syntax diverges dramatically, cross-ISA alignment at the basic-block level can be learned by treating translation as a sequence-to-sequence task, which keeps semantically equivalent blocks close even across architectures [135]. When static evidence is insufficient, micro-executions provide short semantic trajectories that help disambiguate look-alikes during retrieval and ranking [136].
  • Supply chain reuse and fleet-wide recall.
When a CVE is disclosed, the operational question is where else the same or near-same code landed across vendor forks and backports; ML similarity helps enumerate these variants and makes large-fleet recall feasible [137]. Aligning a matched binary to its source repository and patch lineage turns findings into actionable SBOM updates and targeted remediation plans [138].
  • How this complements traditional dynamic testing.
Emulation-based firmware testing remains indispensable but is often tied to specific device classes and requires extensive environment modeling, which limits throughput on large vendor corpora; similarity-guided triage can prioritize where emulation time is best spent [131]. Semi-emulation frameworks reduce modeling overhead but still constrain the set of peripherals and firmware families that can be covered quickly; again, learned retrieval provides high-probability targets to feed these pipelines [139]. Firmware migration techniques extend dynamic testing by transplanting code into instrumented environments, yet the migration effort remains significant; similarity search can narrow the set of functions whose migration yields maximal payoff [140].

5.2.2. ML-Assisted Binary Similarity Workflow

  • Stage 1—Acquisition and normalization.
Firmware images are unpacked to recover file systems and binaries; functions are disassembled and, where possible, lifted to an IR to harmonize instruction semantics across toolchains. A multi-view artifact is built per function: token sequences, basic-block boundaries, and program graphs such as CFGs and PDGs. Because inlining erodes function boundaries, a pre-processing pass that detects or reconstructs inlined regions improves downstream matching and reduces systematic false negatives on utility routines [141].
  • Stage 2—Feature extraction.
Build a multi-view representation per function by jointly harvesting (i) lightweight statistical descriptors over graphs—e.g., ACFG constant histograms, instruction counts, and degree profiles—as fast, interpretable pre-filters [113]; (ii) dynamic/semantic signals from partial emulation that yield short traces (library-call sequences, I/O pairs, value evolution) to stabilize against compiler/inliner variance and supply high-quality cues for later verification [142]; and (iii) learned representations that generalize across ISAs and builds, including GNN embeddings over ACFGs for structure-aware matching [130], distributional/token embeddings that retain instruction-level semantics under moderate CFG distortion [133], self-attention to emphasize vulnerability-relevant motifs such as bounds checks and sanitizers [134], cross-ISA basic-block alignment via neural machine translation when opcode vocabularies diverge strongly [135], and micro-execution–derived trajectories that disambiguate near-neighbors during ranking [136]. Fuse these views with early fusion (feature concatenation) or late fusion (calibrated ensembling) and train with diversified architectures/opt-levels plus hard-negative mining so the model collapses benign compiler drift while preserving differences between patched and vulnerable variants; in inline- or obfuscation-heavy firmware, bias the mix toward token/attention features and dynamic traces, optionally preceded by inline-recovery preprocessing, to maintain recall without sacrificing precision.
  • Stage 3—Representation and retrieval.
Functions are embedded into a vector space or encoded as structured signatures; approximate nearest-neighbor indexes enable sub-linear retrieval on large corpora. During training, hard-negative mining—near-misses produced by different compilers or opt-levels for the same routine—forces the model to collapse benign compiler drift while preserving discriminative cues that separate patched from vulnerable variants [143]. Calibrating similarity scores stabilizes thresholds across firmware families so that ranking quality translates into predictable analyst effort [144].
  • Stage 4—Triage, attribution, and verification.
Top-k candidates are clustered to remove near-duplicates and ranked under a budget-aware objective that maximizes time-to-first-vulnerability. PDG slicing from entry points to privileged operations can then attach concrete paths that indicate likely bypasses in authentication or authorization logic [102]. Targeted symbolic checks validate path feasibility on a small set of high-scoring functions, providing inputs or states that witness the issue without incurring whole-program path explosion. When provenance matters, source/patch alignment maps the binary hit to a repository and patch lineage for SBOM updates and coordinated remediation [138].
  • Design choices that materially improve outcomes.
Training on mixed architectures and compiler/opt-level variants is necessary to achieve cross-variant robustness in graph embeddings; models trained narrowly tend to overfit to build idiosyncrasies and collapse out-of-domain [130]. Inline-heavy builds benefit from an inline-recovery step before representation learning; without it, token and graph models systematically miss small utility routines implicated in many bugs [141]. Discriminating patched from vulnerable near-duplicates requires exposing the learner to such pairs and backing retrieval with a verification gate that checks for the presence of the specific guard or sanitizer introduced by the fix, which reduces false positives during patch cycles [145]. Finally, leakage-safe evaluation must split by project/version rather than at the function level to avoid inflating scores through microscopic code duplication across train and test [143].

5.2.3. Technical Challenges (Root Causes, Symptoms, and Single-Source Mitigations)

There are several issues with using machine learning for binary similarity analysis, as listed below:
  • C1—Information loss after compilation. Root cause: symbols and high-level structure are erased; inlining collapses call boundaries. Symptom: unstable matches and high analyst burden. Mitigation: rely on relational views such as ACFG/PDG to preserve structural semantics during retrieval, then attach PDG-based slices to produce auditable evidence for security-critical flows [102].
  • C2—ISA diversity. Root cause: the same semantics map to different opcode vocabularies and calling conventions. Symptom: high same-ISA accuracy that degrades across architectures. Mitigation: learn ACFG embeddings explicitly trained to be stable across architectures so that semantically equivalent routines cluster together during retrieval [130].
  • C3—Compiler/opt-level variance. Root cause: O-level changes, inlining decisions, and link-time optimization reshape CFGs. Symptom: hand-crafted graph features drift; recall collapses under O3 vs. O0. Mitigation: augment structural features with short semantic traces from partial emulation, which smooth compiler noise while preserving the cues needed for disambiguation [142].
  • C4—Function inlining. Root cause: small helpers vanish as standalone functions; boundaries blur. Symptom: systematic false negatives for utility routines commonly implicated in bugs. Mitigation: detect and recover inlined regions before modeling so that both token and graph learners operate on function units that better reflect source semantics [141].
  • C5—Obfuscation and CFG distortion. Root cause: control-flow flattening and bogus edges decouple CFG shape from behavior. Symptom: CFG-centric descriptors lose discriminative power. Mitigation: emphasize token/sequence embeddings with attention, which learn instruction-level semantics less sensitive to CFG distortion and retain discriminative power under moderate obfuscation [134].
  • C6—Multi-block/function vulnerabilities. Root cause: end-to-end flaws span helpers, drivers, or inter-process boundaries. Symptom: function-only retrieval flags fragments without a triggerable path. Mitigation: perform multi-binary flow reasoning during verification so that candidate matches are elevated only when an inter-component data path to a dangerous sink exists [106].
  • C7—Patch similarity and false positives. Root cause: patched and vulnerable variants remain structurally close. Symptom: alerts on already-fixed code waste patch windows. Mitigation: include patched/unpatched pairs during training and add a verification gate that checks for the specific guard or bounds fix introduced by the patch before escalating a finding [145].
  • C8—Accuracy versus verifiability. Root cause: high cosine similarity does not imply exploitability. Symptom: strong ranking with weak artifacts slows engineering sign-off. Mitigation: attach PDG slices or targeted symbolic inputs to each high-rank match so that every alert is accompanied by concrete paths or inputs that are auditable by developers [102].
  • C9—Throughput and indexing at scale. Root cause: naïve all-pairs comparison is quadratic in corpus size. Symptom: time budgets exhausted before useful recall is reached. Mitigation: use approximate nearest-neighbor indexing and batched inference so that retrieval time scales sub-linearly with corpus size; design ranking to maximize early discovery under fixed budget [146].
  • C10—Sustained scalability under fleet growth. Root cause: the number of images and variants grows faster than analyst/compute budgets. Symptom: thresholds drift upward, recall declines, and backlogs accumulate. Mitigation: adopt two- or three-stage cascades—lightweight pre-filters, semantic enrichment, then targeted verification—so recall per unit compute remains stable as the fleet expands [147].

5.2.4. ML-Assisted Techniques and Tools

Machine learning (ML) has become a central driver of modern binary similarity analysis, enabling firmware vulnerability detection to go beyond rule-based or handcrafted approaches. Broadly, these techniques can be grouped into statistical, dynamic, and learning-based categories, each offering different balances between accuracy, scalability, and interpretability [148].
  • Statistical Feature-Based Methods
Statistical approaches rely on manually extracted features such as instruction counts, control-flow graph (CFG) attributes, or text constants. These methods are lightweight and interpretable, but their effectiveness depends heavily on feature engineering. A representative example is Genius [113], which constructs attributed CFGs (ACFGs) using constants, instruction counts, and descendant nodes to compare binary functions. Building on this idea, Gemini [130] integrates ACFGs with graph neural networks (GNNs, e.g., Structure2Vec [149]) to generate embeddings, significantly improving scalability and robustness against compiler diversity. This transition from handcrafted to embedding-based statistical features marks a key inflection point: it reduces sensitivity to syntactic changes while retaining the structural cues essential for detecting reused or vulnerable code [150].
  • Dynamic Feature-Based Methods
Dynamic methods execute or emulate candidate binaries to extract semantic behaviors, such as I/O values or library call patterns, thereby complementing static analysis. For example, VulSeeker-Pro [142] leverages VEX-IR emulation to extract semantic execution traces and improves vulnerability detection precision by combining these features with graph-based similarity. Similarly, BinGo-E [151] employs the Unicorn engine [152] to generate “fake” execution traces, creating richer behavioral profiles without requiring full device execution. These methods are especially effective for uncovering vulnerabilities that manifest only at runtime (e.g., hidden backdoors or protocol misuse), but they incur higher computational overhead and require careful sandboxing to scale to large firmware corpora. Compared with statistical approaches, dynamic methods provide deeper semantic insight but sacrifice efficiency [153,154,155].
  • Learning-Based Methods
The most recent wave of research uses ML models to automate feature extraction and discover patterns beyond manual engineering. Within this category, graph-based learning methods have become dominant: VulSeeker [156] and IoTSeeker [157] extend Gemini by using GNNs to learn ACFG embeddings that encode both structure and semantics. αDiff [143] further combines convolutional networks for intra-function features with inter-function call-graph analysis, creating embeddings that capture multi-level context.
Another line applies natural language processing (NLP)–inspired embeddings to binary code. Asm2Vec [133] and SAFE [134] treat instructions as tokens, training Word2Vec [158], LSTM [159], or attention models [160] to derive semantic embeddings of instructions or basic blocks. Systems like MIRROR [135] push this further by using neural machine translation to align blocks across different ISAs, while Trex [136] introduces micro-executions [161] with BERT embeddings [162] to model instruction-level trajectories.
Finally, specialized tools tackle long-standing challenges: FSmell [141] improves inline function detection via control-flow analysis combined with CNN-based pattern recognition, while BinaryAI [138] accelerates software composition analysis by matching binary code against large source repositories using deep learning. These directions highlight the versatility of ML in addressing not only binary similarity but also firmware reverse engineering bottlenecks such as function boundary recovery and code provenance.
Taken together, these methods show a clear evolution: from feature-heavy statistical descriptors (Genius) to emulation-driven semantic enrichment (VulSeeker-Pro) and finally to automated, representation-learning approaches (Asm2Vec, IoTSeeker, BinaryAI) [163,164]. Statistical methods remain attractive for lightweight similarity checks, dynamic methods expose runtime behaviors otherwise invisible, and learning-based methods push scalability and robustness across ISAs and obfuscation [165,166]. This spectrum of approaches suggests that hybrid pipelines—statistical pre-filters, dynamic semantic enrichment, and deep representation learning—may provide the most effective balance for large-scale IoT firmware analysis.

5.2.5. Comparative Analysis

Table 5 shows the datasets, CVE usage, search accuracy, and computational cost of some ML-based binary similarity methods, focusing on their effectiveness in discovering firmware vulnerabilities. Existing evaluations of ML-assisted binary similarity analysis have largely focused on precision and average overhead (Table 5). While these indicators establish a baseline, they fail to capture operational realities in IoT firmware analysis, where vulnerabilities follow long-tail distributions, datasets are highly imbalanced, and devices operate under tight computational and energy budgets [167,168,169,170]. A richer evaluation framework is thus required, moving beyond simple accuracy to criteria that reflect robustness, scalability, and actionable utility.
Precision indicates the proportion of correctly identified vulnerable functions among all flagged functions; however, given the highly imbalanced nature of firmware datasets, precision alone provides a limited view of effectiveness, such as those in Table 5 (e.g., 85–93%) obscure the fact that vulnerabilities are rare compared to benign functions.
In such skewed settings, PR-AUC or F1-max provides a more faithful measure of utility. For instance, Gemini [130] reports higher precision than Genius [113], but without recall or PR-AUC it is unclear whether this improvement translates to more vulnerabilities being caught in practice.
Especially for IoT firmware images with thousands of benign routines, the operational impact of false positives (wasted analyst effort, unnecessary patches) is as critical as catching true positives.
The binary flags in Table 6 can be converted into a robustness score (e.g., weighted across ISA × compiler × optimization level × obfuscation). VulSeeker-Pro [142] achieves higher precision partly because it integrates dynamic emulation, making it more resilient to compiler optimizations and inline transformations. In contrast, Asm2Vec [133] has lightweight embeddings but degrades sharply when confronted with obfuscation. A metric such as Cross-Variant Robustness (CVR) would quantify this gap, allowing systematic comparison of how tools generalize across the heterogeneity of IoT ecosystems. Since the two studies used different datasets or CVE sets, it is difficult to compare them directly. BINKIT [171] provides a reproducible benchmark, but its performance is inferior due to its pre-semantic nature.
The overhead values in Table 5 can be reframed into functions per second per core and memory footprint, making comparisons more intuitive for large-scale firmware screening. For example, BinaryAI [138] achieves 7.5 s/function, equivalent to ~0.13 functions/s/core, which translates to ~21 h on a single core for 10,000 functions. However, if BinaryAI parallelizes near-linearly across 16 cores, its effective throughput surpasses tools with lower per-function overhead but weaker parallel scalability. Such throughput-aware evaluation is directly relevant to IoT vendors who must scan thousands of firmware images before scheduled updates.
In practice, analysts rarely need all vulnerabilities at once; rather, they seek time-to-first-vulnerability (TTFV) or recall under a fixed budget (e.g., 8 CPU-hours, 2 GB RAM). Tools with lightweight embeddings (e.g., Asm2Vec) may deliver early hits faster, even if ultimate precision is lower. By contrast, heavier models like SAFE [134] or VulSeeker-Pro [142] yield stronger precision at higher cost. Evaluating methods along the budget–recall curve would thus better inform hybrid pipelines (e.g., fast screening on-device, deeper semantic matching in the cloud).
Finally, tools like BinaryAI [138] go beyond vulnerability discovery into software composition analysis (SCA). For such systems, evaluation must include attribution quality—whether the binary is correctly linked to its source repository, patch version, or software bill of materials (SBOM). Reporting Top-k source hit rate or patch alignment accuracy adds practical value, especially for supply chain auditing where actionable traceability matters more than raw precision.

5.2.6. Open Research Challenges

Machine learning-based binary similarity analysis has made significant progress in detecting firmware vulnerabilities, but many challenges remain:
  • Firmware reverse engineering: Accurate reverse engineering remains difficult and often requires expert knowledge [172].
  • Efficient storage and retrieval: Large firmware corpora require better methods to store and query data that have not been fully studied.
  • Vulnerability dataset construction: Most datasets only consider application-level CVEs, but not firmware-specific vulnerabilities [173].
  • Cross-function/binary vulnerabilities: Currently, we can only consider vulnerabilities in a single function, while ignoring cross-process vulnerabilities [174].
  • False positives/false negatives: Static analysis is difficult due to the large number of false positives and false negatives [175].
  • Lack of common benchmarks: Comparative analysis is also difficult due to different datasets and evaluation criteria [15].

5.2.7. Future Directions

To address these issues, future research should focus on the following areas:
  • Configuration robustness: Improve machine learning models to handle function inlining, obfuscation, and cross-project changes. Use the ideas in FSmell [141] as a starting point.
  • Improve reverse engineering: Develop more advanced disassemblers to handle complex firmware binaries [172].
  • Firmware-specific models: Use semi-supervised learning and reinforcement learning to find firmware-specific vulnerabilities [176].
  • Generic vulnerability signatures: Create templates for vulnerability classes (e.g., buffer overflows) rather than targeting specific CVEs.
  • Unified datasets: Establish standard datasets for training, testing, and CVE querying.
  • Hybrid analysis models: Use frameworks such as BinaryAI [138] for scalable software composition analysis that combines static analysis, dynamic verification, and machine learning for complete firmware testing.

6. Integrated Network, Behavioral, and Policy-Layer Security

This section explores how to integrate network, behavioral, and policy security measures into the Industrial Internet of Things (IIoT) to further enhance Zero Trust Architecture (ZTA). We will then discuss lightweight intrusion detection systems (IDS), advanced deep learning and graph-based models, sensor-actuator and RF spectrum analysis, microsegmentation-based identity management, and Zero Trust implementation. All of these topics build on the fundamental ideas and techniques presented in the previous sections. We will also discuss limitations, gaps between the various layers, and future research directions.

6.1. Lightweight Intrusion Detection and Signature-Based Models

Lightweight intrusion detection systems (IDS) are important in resource-constrained Industrial Internet of Things (IIoT) environments, where sensors and controllers have limited computational power or memory capacity [177]. These solutions aim to ensure security without compromising real-time performance. Signature-based IDS models leverage a database of known threats to detect malicious behavior. These threats are captured as a set of patterns. For example, Buck et al. [178] discussed lightweight IDS solutions that leverage hash-based operations and XOR techniques to minimize computational power requirements while still detecting many known attack patterns. These models work well in areas where threats are easy to predict but difficult to defend against zero-day attacks because they rely on static signatures [179]. To cope with limited resources, the single packet authorization (SPA) approach proposed by Xu et al. [180] proposes filtering incoming packets based on cryptographic signatures. This allows for fast and efficient detection of issues. Shah et al. [180] discussed a single packet authorization (SPA) approach based on cryptographic signatures that proposes filtering incoming packets based on cryptographic signatures. Reference [179] also used lightweight time-varying linear or nonlinear functions to protect IIoT communication paths. This enables fast threat detection, which is critical for industrial real-time applications. Although signature-based models work well, they often miss new threats and therefore need to be combined with more flexible methods [181].

6.2. Deep/Graph-Based IDS with Behavioral Features

Deep learning and graph-based intrusion detection systems (IDS) exploit behavioral features to identify problems in Industrial Internet of Things (IIoT) networks that are hidden from signature-based models. These approaches exploit complex patterns in data and interactions that change over time to identify anomalous characteristics of a device or network. Wu et al. [182] proposed a deep learning-based intrusion detection system (IDS) that uses neural networks to detect human behavior. This approach is better suited to detecting complex attacks than traditional methods. Their approach uses unsupervised learning to discover hidden data structures to identify previously undetected risks [183].
Konduru et al. [184] studied graph-based intrusion detection system (IDS) models that represent Industrial Internet of Things networks as graphs with devices as nodes and communication lines as edges. These models identify problems caused by lateral movement or unauthorized access by examining the characteristics of the graph, including how nodes are connected and how traffic flows. For example, Ali et al. [185] added behavioral elements such as device connection frequency and data exchange protocol to graph-based models to improve discovery capabilities in edge computing environments. These approaches are good at finding things but require a lot of computing power, which is a problem for resource-constrained Industrial IoT devices [186].

6.3. Sensor-Actuator and RF-Spectrum Anomaly Analytics

Sensor-actuator systems and RF spectrum analysis are critical to the security of Industrial Internet of Things (IIoT) environments because they monitor physical and wireless interactions. The goal of sensor-actuator anomaly detection is to identify anomalous patterns in device behavior, including unpredictable actuator motion or sensor data that does not match expectations. Lei et al. [187] proposed a zero-trust (ZTA)-enhanced system that uses lightweight encryption and artificial noise precoding to protect wireless channels and detect problems in sensor-actuator communications. This approach can monitor the environment with low latency while ensuring security in high-risk industrial environments [188].
RF spectrum analysis addresses the issue of wireless IIoT connectivity, especially where proprietary protocols are used. Sedjelmaci et al. [16] proposed an AI-based RF spectrum analysis model for 6G networks that leverages zero-trust principles to detect anomalous patterns in signal strength and frequency. Their model reduces threats such as interference or spoofing by combining real-time spectrum monitoring with trust assessment. However, continuous spectrum analysis remains a challenge for low-resource devices due to its high computational cost [189].

6.4. Identity, Micro-Segmentation, and Zero-Trust Enforcement

Identity management, microsegmentation, and zero-trust implementations are key components of Industrial Internet of Things (IIoT) security. They ensure precise access control and continuous trust verification. Authentication in the IIoT requires a system that is powerful, flexible, and compatible with a wide range of devices [190]. Li et al. [191] proposed a blockchain-based multi-factor authentication (MFA) system that combines device attributes and cryptographic signatures to ensure secure authentication across different IIoT platforms. Zanasi et al. [192] added quantum-resistant digital certificates that contain access rights and metadata to make the system more flexible and secure.
Microsegmentation breaks down the IIoT network into smaller, independent areas to prevent lateral threats. Daah et al. [193] implement network-based microsegmentation using gateways to enforce access policies and assign trust levels to segments. Alagappan et al. [194] use local infrastructure-based microsegmentation in virtual power plants to protect devices from attacks based on their functionality and trust level. Zero-trust implementations combine these tools with continuous authentication. Shen et al. [181] used a policy enforcement point (PEP) to verify access requests in real time based on trust scores.

6.5. Limitations, Cross-Layer Gaps, and Future Work

Despite improvements, Zero Trust (ZTA) in the Industrial Internet of Things (IIoT) still faces many challenges. Michael et al. [195] pointed out that implicit trust in untested things such as policy decision points (PDPs) and trust algorithms may reduce the effectiveness of ZTA. Bertino [196] pointed out that high switching costs and legacy system issues are some of the challenges encountered during deployment. Swearingen et al. pointed out that implementing ZTA in critical infrastructure such as power grids may introduce latency, thus affecting real-time operations.
Vulnerabilities still exist at the network, behavioral, and policy levels. For example, lightweight intrusion detection system (IDS) models are vulnerable to zero-day attacks, while deep learning-based intrusion detection systems (IDS) consume too many resources [186]. Micro-segmentation technology does not work well with different types of protocols, and RF spectrum analysis requires a lot of computing power, which makes it difficult to scale [189]. To strike a balance between security and resource constraints, Gudala et al. [183] suggested that future research focus on lightweight AI algorithms. Zaid et al. studied the integration of quantum-resistant cryptography, while Hong et al. proposed improving cross-layer coordination through programmable APIs. Both approaches can help fill these gaps. Table 7 provides a comparative overview of representative IDS and policy-layer mechanisms, highlighting their datasets, accuracy levels, latency characteristics, and specific trust targets.
The table makes clear the trade-offs among different families of approaches: lightweight IDSs provide low latency but limited scope, deep and graph-based IDSs achieve high accuracy but incur computational delays, while RF spectrum analytics and micro-segmentation serve specialized trust targets but suffer either from scalability or interoperability constraints. This comparative lens underscores that no single approach suffices across all IIoT conditions, reinforcing the need for layered integration.
In addition to cross-layer intrusion detection mechanisms, trust evaluation and policy-based access control play a central role in ensuring secure IoT and IIoT deployments. To illustrate this, Figure 4 shows the integrated control and data platform architecture that leverages a Trust Evaluation Module, Policy Engine (PE), and Policy Administration Point (PA). The framework connects threat sources, behavior logs, and access policies with enforcement points, enabling dynamic micro-segmentation and fine-grained authorization across industrial control environments.
Operational trade-offs in IDS and policy layers are contrasted in Table 7, while deployable cross-layer bundles appear in Appendix A (Table A1 and Table A2).
The integrated pipeline includes a lightweight IDS for fast signature-based detection, a deep/graph-based IDS for detecting behavioral anomalies, RF spectrum analysis for wireless security, and policy-driven micro-segmentation for access control. The Trust Assessment module processes data from network traffic and device logs. The module sends the data to the Policy Engine (PE) and Policy Administrator (PA). The PE uses behavioral, contextual, and policy data to determine the trust level, while the PA uses the PEP to enforce access decisions, ensuring that there are no trust issues at any level.

7. Conclusions

The evidence assembled across 207 rigorously screened studies paints a consistent picture: dependable IoT/IIoT security is achieved not by any single technique, but by composing methods across layers, data granularities, and deployment models under strict resource budgets. Organizing prior work along system level, algorithmic paradigm, data granularity, and hardware budget clarified where each family of defenses excels and where it fails. At the firmware layer, static analysis supplies breadth and explainability for large corpora, while symbolic execution provides path-level precision for memory safety and logic flaws; hybrid tactics (e.g., program-slice prechecks, selective MMIO modeling, and fuzzing-guided exploration) curb false positives and extend coverage into peripheral-heavy paths, yet remain bounded by Type-II/III packaging quirks, path explosion, and incomplete device models.
Learning-based binary similarity closes a complementary gap: by embedding attributed control-flow graphs and instruction tokens, it collapses compiler/ISA variance and turns global search into fast nearest-neighbor retrieval. Adding short semantic traces from partial execution raises robustness to inlining/optimization, and coupling similarity with composition analysis ties matches to provenance and patches for SBOM-grade attribution. Crucially, the most effective pipelines exhibit the same operational pattern: rapid retrieval to surface likely vulnerable functions or reused components; targeted verification to confirm triggerability with concrete paths/inputs; and evidence fusion and attribution to de-duplicate variants, stabilize scores, and drive coordinated remediation. Viewed through this lens, headline metrics like “precision” are insufficient on their own; what matters operationally is recall under a fixed compute/memory budget and time-to-first-vulnerability, supported by throughput-aware indexing (approximate nearest neighbors, batched inference) and calibration so that a rank threshold corresponds to predictable analyst effort across heterogeneous firmware families.
Beyond firmware, lightweight intrusion detection on constrained links delivers immediate safeguards at millisecond latencies, while deep/graph-based models at gateways capture higher-order traffic and device-behavior structure for zero-day detection. Sensor–actuator consistency checks and RF-spectrum analytics close cyber-physical blind spots that packet inspection cannot, and identity/micro-segmentation with continuous, device-centric authorization confines blast radius when latent defects persist. These layers are not substitutes but safety nets for one another: retrieval-driven firmware triage feeds dynamic testing; verified findings inform network and policy filters; and policy engines in turn prioritize where scarce device, edge, and cloud resources should be spent.
Notwithstanding the improvements synthesized in this review, several challenges remain. Nonstandard packaging and sparse metadata impede unpacking and cross-binary reasoning; symbolic engines continue to be constrained by peripheral interactions and cryptographic primitives; learning models must better tolerate obfuscation, aggressive inlining, and imperfect function boundaries; and end-to-end flaws that traverse helpers, drivers, and services are still easier to flag than to validate as triggerable chains. At higher layers, network- and behavior-centric detectors face enduring trade-offs between accuracy and footprint, RF analytics typically require specialized hardware, and policy-heavy controls are expensive to retrofit across legacy fleets.
Against this backdrop, operationally actionable advances follow directly from these limits and are best pursued through a set of priorities aligned with evidence needs and resource budgets. These priorities include adopting evidence-attached detection as standard practice so that high-rank firmware candidates are accompanied by slices, inputs, or harness hints that expedite engineering review and compliance; elevating cross-variant robustness to a primary objective by evaluating across ISA, compiler, optimization, and obfuscation diversity and reporting recall-at-budget and time-to-first-vulnerability alongside precision; preprocessing for embedded-code realities by recovering function boundaries in inline-heavy builds, de-noising distorted control-flow graphs, and employing short micro-executions when structural cues are unreliable; developing reusable, peripheral-aware symbolic scaffolds—selective symbolization, minimal MMIO models, and driver stubs—that permit deep checks to scale across device families; integrating vulnerability discovery with software composition analysis so that detection, attribution, SBOM maintenance, and fleet recall form a unified workflow; and investing in open, reproducible corpora and benchmarks that cover single-function, cross-function, and cross-binary flaws to enable comparable results and transferability.
Taken together, these directions point to a disciplined, resource-aware pipeline in which scalable similarity identifies what matters, precise program analysis substantiates it, and policy-grade attribution operationalizes remediation, with deployments tuned to the realities of tiny MCUs, modest gateways, and shared clouds. Systems built along these lines respect edge constraints, reduce analyst effort, and measurably raise the cost of exploitation across heterogeneous fleets, advancing the security of billions of connected devices toward resilience, auditability, and sustainability.

Author Contributions

Conceptualization, M.A. and T.A.; methodology, M.A.; validation, M.A., A.N., and D.K.; formal analysis, M.A.; investigation, M.A.; resources, D.K. and G.A.; data curation, M.A.; writing—original draft preparation, M.A.; writing—review and editing, A.N., A.I., and G.A.; visualization, M.A.; supervision, T.A.; project administration, A.N.; funding acquisition, T.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been funded by the Committee of Science of the Ministry of Science and Higher Education of the Republic of Kazakhstan (Grant No. BR24992852 “Intelligent models and methods of Smart City digital ecosystem for sustainable development and the citizens’ quality of life improvement”).

Institutional Review Board Statement

Not applicable. This study did not involve humans or animals.

Informed Consent Statement

Not applicable. This study did not involve humans.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Acknowledgments

The author would like to thank colleagues and reviewers for their valuable suggestions during the preparation of this manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Practitioner-Facing Artifacts

Table A1 consolidate practitioner-facing artifacts discussed throughout the survey. They provide concrete mappings between abstract taxonomy categories and representative tools or deployment strategies. This aims to assist researchers and engineers in identifying feasible combinations of firmware, network, behavioral, and policy-layer approaches under diverse device budgets.
Table A1. Representative Tools by Layer & Paradigm & Deployment. (Each row provides a one-line purpose, typical deployment target, and a minimal resource note. “Edge” denotes gateway/industrial PC-class devices).
Table A1. Representative Tools by Layer & Paradigm & Deployment. (Each row provides a one-line purpose, typical deployment target, and a minimal resource note. “Edge” denotes gateway/industrial PC-class devices).
LayerParadigmTool/System (Example)One-Line PurposeTypical DeploymentMinimal Resource Note
FirmwareUnpacking/preprocessingBinwalk; BANGRecover file systems/binaries from images for downstream analysisEdge/CloudCPU-bound; tens of minutes for large images; RAM in GB-scale for batch runs
FirmwareStatic (IR/graphs)IDA Pro, GhidraDisassembly/CFG/IR to enable slicing and taint analysisEdge/CloudMemory heavy on large firmware; plugin support for many ISAs
FirmwareStatic + program slicingFirmalicePDG-guided authentication-bypass discovery via slicing + checksEdge/CloudHours per image on commodity CPUs; requires IR build
FirmwareStatic taint analysisDTaintTracks source → sink flows to surface memory/command-injection risksEdge/CloudSensitive to library modeling; multi-binary speedups advised
FirmwareMulti-binary reasoningKaronteModels inter-process/component flows to catch cross-binary flawsEdge/CloudRequires per-service configuration; higher RAM
FirmwareSymbolic execution (hybrid)Avatar, S2EHybrid SE with device/system emulation for deep path feasibilityEdge/CloudSolver/emu heavy; best under narrowed scopes
FirmwareSymbolic (platform-specific)FIE (MSP430), FirmUSB (8051/USB)Platform-tailored SE for microcontroller/USB firmwareEdgeGood precision; limited to supported MCUs/peripherals
FirmwareSymbolic (LLVM/KLEE)Inception, GerbilSE with slicing/backward analysis for privilege/logic flawsEdge/CloudRequires stubs for MMIO/crypto; mitigated by selective symb.
FirmwareHybrid SE + fuzzFIoT, HD-FUZZCombine selective SE with fuzzing and MMIO modelingEdgeBetter coverage on embedded paths; setup effort moderate
FirmwareML similarity (graph)Genius, Gemini, VulSeeker, VulSeeker-Pro, IoTSeekerACFG/GNN embeddings for cross-variant vulnerable-code retrievalEdge/CloudCPU-centric; optional GPU; index fits in GBs; robustness per Table 6
FirmwareML similarity (tokens/attention)Asm2Vec, SAFEInstruction/token embeddings (w/attention) robust to CFG noiseEdge/CloudFast inference; sensitive to inlining unless preprocessed
FirmwareInline recoveryFSmellDetects/reconstructs inlined helpers to reduce false negativesEdge/CloudPre-pass; improves recall for similarity pipelines
FirmwareSCA/provenanceBinaryAIMaps binaries to source/patch lineage for SBOM/attributionCloud/EdgeEmbedding + ANN search; integrates with CI/SBOM
NetworkLightweight IDSHash/XOR-style IDS (e.g., Buck et al.)Signature/bitfield/Bloom-style fast screening on constrained linksMCU/EdgekB-scale RAM; sub-10 ms latency (see Table 7)
NetworkSPA/admissionSingle Packet Authorization (SPA)Cryptographic single-packet gate before session setupMCU/EdgeMinimal CPU; integrates with firewall/PEP
NetworkProtocol-aware auditingFirmXRay (BLE)Audit BLE pairing/keys/mode use; flag weak defaultsEdgeRuns off-device; informs on-device filters
BehaviorContext policyContexIoT, SoteriaApp-level context/policy checking to prevent privilege abuseEdge/CloudLow runtime cost; upfront policy modeling
BehaviorTelemetry anomaly (DL)N-BaIoT (DAEs)Device-behavior anomaly detection from traffic/telemetryEdge/Cloud50–100 ms typical on edge; model fits in 10 s–100 s MB
RF/WirelessSpectrum analytics[16] RF-IDS (prototype)RF fingerprint/DoS/spoofing detection from I/Q/spectrumEdgeSDR required; 150–300 ms latency class (Table 7)
Trust/PolicyFine-grained authZWAVE, XToken, MPATokenized/chain-of-trust authorization and multi-party approvalCloud/EdgePEP/PDP integration; low per-decision latency (<20 ms)
Trust/PolicyMicro-segmentationGateway PEP/PA (industrial)Segment device classes; enforce least privilege at gatewaysEdge/CloudPolicy design effort; low runtime cost (Table 7)
Trust/PolicyBlockchain-assisted MFALi et al. MFAMFA + device attributes for heterogeneous fleetsCloud/EdgeAdded crypto overhead; offloads to cloud when possible
Table A2. Recommended Cross-Layer Combinations by Device Class & Budget (Condensed: highlights key defense choices, accuracy/latency ranges, and operational cost notes).
Table A2. Recommended Cross-Layer Combinations by Device Class & Budget (Condensed: highlights key defense choices, accuracy/latency ranges, and operational cost notes).
Device Class & BudgetFirmware LayerNetwork LayerBehavior LayerTrust/PolicyAccuracy/Latency (Ref. Table 6 and Table 7)Operational Notes
Tiny MCU (≤256 kB RAM)Integrity checks; off-device batch screening; selective SE for top casesLightweight IDS; SPA gateRule-based sanity checksGateway/cloud PEP/PDPIDS ~ 85–90% @<10 msMinimal runtime; main cost is safe OTA + CI pipeline
Mid-range MCU (512 kB–2 MB RAM)Batch ML screening (SAFE/VulSeeker-Pro); micro-exec for top-KLightweight IDS + protocol filtersEdge-assisted telemetryMicro-segmentation by roleIDS < 20 ms; precision per Table 6Higher integration burden; modest telemetry labeling
Gateway-class Edge (2–8 GB RAM)Continuous similarity (BinaryAI/SAFE); SE/fuzz for deep pathsDeep/Graph IDSDAE/GRU models for device behaviorLocal PEP + cloud PDPDeep IDS ~ 92–95% @50–100 ms; Graph ~ 90–93% @100–200 msModerate ops cost; GPU optional; strong zero-day defense
Hybrid Fleet (MCU + Edge + Cloud)MCU integrity; Edge/Cloud similarity + provenanceMCU pre-filters + edge deep IDSEdge-level temporal modelsFleet-wide attestation + micro-segmentationCombined: low on-device latency, edge accuracy as Table 7Highest coordination payoff; requires CI/CD for models + policy
Mission-critical IIoT (PLC/industrial PCs)Conservative screening; evidence-attached verificationLayered guards + deep IDS at aggregationPhysical-process consistency; optional RF guardStrong zero-trust enforcementLink guards < 10 ms; deep IDS 50–200 ms; RF 150–300 ms

References

  1. Cui, J.; Zhu, Y.; Zhong, H.; Zhang, Q.; Gu, C.; He, D. Efficient blockchain-based mutual authentication and session key agreement for cross-domain IIoT. IEEE Internet Things J. 2024, 11, 16325–16338. [Google Scholar] [CrossRef]
  2. May, M.C.; Glatter, D.; Arnold, D.; Pfeffer, D.; Lanza, G. IIoT system canvas—From architecture patterns towards an IIoT development framework. J. Manuf. Syst. 2024, 72, 437–459. [Google Scholar] [CrossRef]
  3. Hu, Y.; Jia, Q.; Yao, Y.; Lee, Y.; Lee, M.; Wang, C.; Zhou, X.; Xie, R.; Yu, F.R. Industrial Internet of Things intelligence empowering smart manufacturing: A literature review. IEEE Internet Things J. 2024, 11, 19143–19167. [Google Scholar] [CrossRef]
  4. Hai, T.; Sarkar, A.; Aksoy, M.; Khan, M.Z.; Alahmadi, A. Complex-valued hyperchaos-assisted vector-valued artificial neural key coordination for improving security in the industrial Internet of Things. Eng. Appl. Artif. Intell. 2024, 128, 107561. [Google Scholar] [CrossRef]
  5. Li, S.; Iqbal, M.; Saxena, N. Future industry Internet of Things with zero-trust security, Information Systems Frontiers. Inf. Syst. Front. 2022, 26, 1653–1666. [Google Scholar] [CrossRef]
  6. Stafford, V.A. Zero Trust Architecture. In NIST Special Publication 800-207; National Institute of Standards and Technology: Gaithersburg, MD, USA, 2020. [Google Scholar]
  7. Yaqoob, I.; Ahmed, E.; Hashem, I.; Ahmed, A.I.A.; Gani, A.; Imran, M.; Guizani, M. Internet of Things architecture: Recent advances, taxonomy, requirements, and open challenges. IEEE Wirel. Commun. 2017, 24, 10–16. [Google Scholar] [CrossRef]
  8. Sethi, P.; Sarangi, S.R. Internet of Things: Architectures, protocols, and applications. J. Electr. Comput. Eng. 2017, 2017, 9324035. [Google Scholar] [CrossRef]
  9. Olaitan, O.F.; Akatakpo, O.N.; Victor, C.O.; Emejulu, C.J. Secure and Resilient Industrial IoT Architectures for Smart Manufacturing: A Comprehensive Review. J. Eng. Res. Rep. 2025, 27, 331–344. [Google Scholar] [CrossRef]
  10. Sikder, A.K.; Petracca, G.; Aksu, H.; Jaeger, T.; Uluagac, A.S. A survey on sensor-based threats and attacks to smart devices and applications. IEEE Commun. Surv. Tutor. 2021, 23, 1125–1159. [Google Scholar] [CrossRef]
  11. Ronen, E.; O’Flynn, C.; Shamir, A.; Weingarten, A.-O. IoT goes nuclear: Creating a ZigBee chain reaction. In Proceedings of the IEEE Symposium on Security and Privacy, San Jose, CA, USA, 22–24 May 2017; pp. 195–212. [Google Scholar]
  12. Wen, H.; Lin, Z.; Zhang, Y. FirmXRay: Detecting Bluetooth link-layer vulnerabilities from bare-metal firmware. In Proceedings of the 2020 ACM SIGSAC Conference on Computer and Communications Security, Orlando, FL, USA, 9–13 November 2020; pp. 167–180. [Google Scholar]
  13. Jia, Y.; Xing, L.; Mao, Y.; Zhao, D.; Wang, X.; Zhao, S.; Zhang, Y. Burglars’ IoT paradise: Understanding and mitigating security risks of general messaging protocols on IoT clouds. In Proceedings of the IEEE Symposium on Security and Privacy, San Francisco, CA, USA, 18–21 May 2020; pp. 465–481. [Google Scholar]
  14. Tippenhauer, N.O.; Pöpper, C.; Rasmussen, K.B.; Čapkun, S. On the requirements for successful GPS spoofing attacks. In Proceedings of the 18th ACM Conference on Computer and Communications Security (CCS’11), Chicago, IL, USA, 17–21 October 2011; pp. 75–86. [Google Scholar]
  15. Xu, M.; Guo, J.; Yuan, H.; Yang, X. Zero-Trust Security Authentication Based on SPA and Endogenous Security Architecture. Electronics 2023, 12, 782. [Google Scholar] [CrossRef]
  16. Bertino, E. Zero trust architecture: Does it help? IEEE Secur. Priv. 2021, 19, 95–96. [Google Scholar] [CrossRef]
  17. Page, M.J.; Moher, D.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. PRISMA 2020 explanation and elaboration: Updated guidance for reporting systematic reviews. BMJ 2021, 29, 372. [Google Scholar] [CrossRef]
  18. CVE-2006-6125. Available online: https://www.kb.cert.org/vuls/id/403152/ (accessed on 17 June 2025).
  19. Su, Z.D.; Wassermann, G. The Essence of Command Injection Attacks in Web Applications. ACM Spec. Interest Group Program. Lang. 2006, 41, 372–382. [Google Scholar]
  20. CVE-2019-7298. Available online: https://github.com/leonW7/D-Link/blob/master/Vul_2.md (accessed on 17 June 2025).
  21. CVE-2018-18441. Available online: http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-18441 (accessed on 17 June 2025).
  22. Felmetsger, V.; Cavedon, L.; Kruegel, C.; Vigna, G. Toward Automated Detection of Logic Vulnerabilities in Web Applications. In Proceedings of the USENIX Security Symposium, Washington, DC, USA, 11–13 August 2010; pp. 143–160. [Google Scholar]
  23. CVE-2017-7923. Available online: http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-7923 (accessed on 17 June 2025).
  24. CVE-2006-4143. Available online: https://cxsecurity.com/issue/WLB-2006080097 (accessed on 17 June 2025).
  25. CVE-2006-1003. Available online: http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2006-1003 (accessed on 17 June 2025).
  26. Liu, C.; Zou, D.; Luo, P.; Zhu, B.; Jin, H. A Heuristic Framework to Detect Concurrency Vulnerabilities. In Proceedings of the ACSAC’18: 2018 Annual Computer Security Applications Conference, San Juan, Puerto Rico, 3–7 December 2018; pp. 529–541. [Google Scholar]
  27. CVE-2018-4027. Available online: https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-4027 (accessed on 17 June 2025).
  28. Abbaspour, S.; Sundmark, D.; Eldh, S.; Hansson, H. A Runtime Verification Tool for Detecting Concurrency Bugs in FreeRTOS Embedded Software. In Proceedings of the 17th International Symposium on Parallel and Distributed Computing, Geneva, Switzerland, 3–7 December 2018; pp. 172–179. [Google Scholar]
  29. Chandrasekaran, P.; Kumar, S.; Minz, R.; D’Souza, D.; Meshram, L. A Multi-Core Version of FreeRTOS Verified for Datarace and Deadlock Freedom. In Proceedings of the 2014 Twelfth ACM/IEEE Conference on Formal Methods and Models for Codesign, Lausanne, Switzerland, 19–21 October 2014; pp. 62–71. [Google Scholar]
  30. Yin, Z.; Ma, X.; Zheng, J.; Zhou, Y.; Bairavasundaram, L.; Pasupathy, S. An Empirical Study on Configuration Errors in Commercial and Open Source Systems. In Proceedings of the Twenty-Third ACM Symposium on Operating Systems Principles—SOSP’11, Cascais, Portugal, 23–26 October 2011; pp. 159–172. [Google Scholar]
  31. CVE-2018-19990. Available online: http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-19990 (accessed on 17 June 2025).
  32. CVE-2019-10132. Available online: http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-10132 (accessed on 17 June 2025).
  33. CVE-2019-2041. Available online: http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-2041 (accessed on 17 June 2025).
  34. CVE-2017-7916. Available online: http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-7916 (accessed on 17 June 2025).
  35. CVE-2019-9976. Available online: http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-9976 (accessed on 17 June 2025).
  36. Xu, T.Y.; Zhou, Y.Y. Systems Approaches to Tackling Configuration Errors: A Survey. ACM Comput. Surv. 2015, 47, 70. [Google Scholar] [CrossRef]
  37. Rabkin, A.; Katz, R. Precomputing Possible Configuration Error Diagnoses. In Proceedings of the 26th IEEE/ACM International Conference on Automated Software Engineering, Lawrence, KS, USA, 6–10 November 2011; pp. 193–202. [Google Scholar]
  38. Xu, T.; Zhang, J.; Huang, P.; Zheng, J.; Sheng, T.; Yuan, D.; Zhou, Y.; Pasupathy, S. Do not Blame Users for Misconfigurations. In Proceedings of the Twenty-Fourth ACM Symposium on Operating Systems Principles, Farmington, PA, USA, 3–6 November 2013; pp. 244–259. [Google Scholar]
  39. Chess, B.; McGraw, G. Static Analysis for Security. IEEE Secur. Priv. 2004, 2, 76–79. [Google Scholar] [CrossRef]
  40. Cojocar, L.; Zaddach, J.; Verdult, R.; Bos, H.; Francillon, A.; Balzarotti, D. PIE: Parser Identification in Embedded Systems. In Proceedings of the 31st Annual Computer Security Applications Conference, Los Angeles, CA, USA, 7–11 December 2015; pp. 251–260. [Google Scholar]
  41. Cheng, K.; Li, Q.; Wang, L.; Chen, Q.; Zheng, Y.; Sun, L.; Liang, Z. DTaint: Detecting the Taint-Style Vulnerability in Embedded Device Firmware. In Proceedings of the 48th Annual IEEE/IFIP International Conference on Dependable Systems and Networks, Luxembourg, 25–28 June 2018; pp. 430–441. [Google Scholar]
  42. Gotovchits, I.; van Tonder, R.; Brumley, D. Saluki: Finding taint-style vulnerabilities with static property checking. In Proceedings of the NDSS Workshop on Binary Analysis Research, San Diego, CA, USA, 18 February 2018. [Google Scholar]
  43. Redini, N.; Machiry, A.; Wang, R.; Spensky, C.; Continella, A.; Shoshitaishvili, Y.; Kruegel, C.; Vigna, G. Karonte: Detecting Insecure Multi-Binary Interactions in Embedded Firmware. In Proceedings of the 2020 IEEE Symposium on Security and Privacy, San Francisco, CA, USA, 18–21 May 2020; pp. 1544–1561. [Google Scholar]
  44. Stephen, J.; Savvides, S.; Sundaram, V.; Ardekani, M.; Eugster, P. STYX: Stream Processing with Trustworthy Cloud-Based Execution. In Proceedings of the Seventh ACM Symposium on Cloud Computing, Santa Clara, CA, USA, 5–7 October 2016; pp. 348–360. [Google Scholar]
  45. Sabt, M.; Achemlal, M.; Bouabdallah, A. Trusted Execution Environment: What it Is, and what it is not. In Proceedings of the 2015 IEEE Trustcom/BigDataSE/ISPA, Helsinki, Finland, 20–22 August 2015; pp. 57–64. [Google Scholar]
  46. Satyanarayanan, M. The Emergence of Edge Computing. Computer 2017, 50, 30–39. [Google Scholar] [CrossRef]
  47. Shi, W.; Cao, J.; Zhang, Q.; Li, Y.; Xu, L. Edge Computing: Vision and Challenges. IEEE Internet Things J. 2016, 3, 637–646. [Google Scholar] [CrossRef]
  48. Li, E.; Zhou, Z.; Chen, X. Edge Intelligence: On-Demand Deep Learning Model Co-Inference with Device-Edge Synergy. In Proceedings of the 2018 Workshop on Mobile Edge Communications, Budapest, Hungary, 20 August 2018; pp. 31–36. [Google Scholar]
  49. Khan, W.; Ahmed, E.; Sahib, I.; Yaqoob, I.; Ahmed, A. Edge Computing: A Survey. Future Gener. Comput. Syst. 2019, 97, 219–235. [Google Scholar] [CrossRef]
  50. Hassan, N.; Gilani, S.; Ahmed, E.; Yaqoob, I.; Imran, M. The Role of Edge Computing in Internet of Things. IEEE Commun. Mag. 2018, 56, 110–115. [Google Scholar] [CrossRef]
  51. Eclipse Foundation. Key Trends from the IoT Developer Survey. 2018. Available online: https://iot.eclipse.org/community/resources/iot-surveys/assets/iot-developer-survey-2018.pdf (accessed on 17 June 2025).
  52. Arm TrustZone. Available online: http://www.arm.com/products/processors/technologies/trustzone/index.php (accessed on 17 June 2025).
  53. McGillion, B.; Dettenborn, T.; Nyman, T.; Asokan, N. Open-TEE—An Open Virtual Trusted Execution Environment. In Proceedings of the 2015 IEEE Trustcom/BigDataSE/ISPA, Helsinki, Finland, 20–22 August 2015; pp. 400–407. [Google Scholar]
  54. Asanović, K.; Avizienis, R.; Bachrach, J.; Beamer, S.; Biancolin, D.; Celio, C.; Cook, H.; Dabbelt, P.; Hauser, J.R.; Izraelevitz, A.M.; et al. The Rocket Chip Generator; University of California: Berkeley, CA, USA, 2016. [Google Scholar]
  55. Al-Omari, M.; Rawashdeh, M.; Qutaishat, F.; Alshira’h, M.; Ababneh, N. An intelligent tree-based intrusion detection model for cyber security. J. Netw. Syst. Manag. 2021, 29, 20. [Google Scholar] [CrossRef]
  56. Sapavath, N.N.; Muhati, E.; Rawat, D.B. Prediction and detection of cyberattacks using AI model in virtualized wireless networks. In Proceedings of the 8th IEEE International Conference on Cyber Security and Cloud Computing, Washington, DC, USA, 26–28 June 2021; pp. 97–102. [Google Scholar] [CrossRef]
  57. Al-Yaseen, W.L.; Othman, Z.A.; Nazri, M.Z.A. Multi-level hybrid support vector machine and extreme learning machine based on modified K-means for intrusion detection system. Expert Syst. Appl. 2017, 67, 296–303. [Google Scholar] [CrossRef]
  58. Gu, Y.H.; Li, K.; Guo, Z.; Wang, Y. Semi-supervised K-means DDoS detection method using hybrid feature selection algorithm. IEEE Access 2019, 7, 64351–64365. [Google Scholar] [CrossRef]
  59. Abuali, K.M.; Nissirat, L.; Al-Samawi, A. Advancing network security with AI: SVM-based deep learning for intrusion detection. Sensors 2023, 23, 8959. [Google Scholar] [CrossRef]
  60. Buczak, A.L.; Guven, E. A survey of data mining and machine learning methods for cyber security intrusion detection. IEEE Commun. Surv. Tutor. 2016, 18, 1153–1176. [Google Scholar] [CrossRef]
  61. Nishiyama, T.; Kumagai, A.; Kamiya, K.; Takahashi, K. SILU: Strategy involving large-scale unlabeled logs for improving malware detector. In Proceedings of the IEEE Symp on Computers and Communications, Rennes, France, 7–10 July 2020; pp. 1–7. [Google Scholar] [CrossRef]
  62. Zhou, X.Y.; Belkin, M. Semi-supervised learning. Acad. Press Libr. Signal Process. 2014, 1, 1239–1269. [Google Scholar] [CrossRef]
  63. Liu, X.; Zhang, H.; Dong, S.; Zhang, Y. Network defense decision-making based on a stochastic game system and a deep recurrent Q-network. Comput. Secur. 2021, 111, 102480. [Google Scholar] [CrossRef]
  64. Gronauer, S.; Diepold, K. Multi-agent deep reinforcement learning: A survey. Artif. Intell. Rev. 2021, 55, 895–943. [Google Scholar] [CrossRef]
  65. Khaw, Y.; Abiri-Jahromi, A.; Fakhari Moghaddam Arani, M.; Sanner, S.; Kundur, D.; Kassouf, M. A deep learning-based cyberattack detection system for transmission protective relays. IEEE Trans. Smart Grid 2020, 12, 2554–2565. [Google Scholar] [CrossRef]
  66. Ho, S.; Al Jufout, S.; Dajani, K.; Mozumdar, M. A novel intrusion detection model for detecting known and innovative cyberattacks using convolutional neural network. IEEE Open J. Comput. Soc. 2021, 2, 14–25. [Google Scholar] [CrossRef]
  67. Nazir, A.; He, J.; Zhu, N.; Qureshi, S.S.; Qureshi, S.U.; Ullah, F.; Wajahat, A.; Pathan, M.S. A deep learning-based novel hybrid CNN-LSTM architecture for efficient detection of threats in the IoT ecosystem. Ain Shams Eng. J. 2024, 15, 102777. [Google Scholar] [CrossRef]
  68. Pouyanfar, S.; Sadiq, S.; Yan, Y.; Tian, H.; Tao, Y.; Presa Reyes, M.; Shyu, M.-L.; Chen, S.-C.; Iyengar, S. A survey on deep learning: Algorithms, techniques, and applications. ACM Comput. Surv. 2018, 51, 92. [Google Scholar] [CrossRef]
  69. Hossain, M.D.; Ochiai, H.; Fall, D.; Kadobayashi, Y. SSH and FTP brute-force attacks detection in computer networks: LSTM and machine learning approaches. In Proceedings of the 5th International Conference on Computer and Communication Systems, Shanghai, China, 15–18 May 2020; pp. 491–497. [Google Scholar] [CrossRef]
  70. Ma, P.; Jiang, B.; Lu, Z.; Li, N.; Jiang, Z. Cybersecurity named entity recognition using bidirectional long short-term memory with conditional random fields. Tsinghua Sci. Technol. 2020, 26, 259–265. [Google Scholar] [CrossRef]
  71. Bukhari, S.M.S.; Zafar, M.H.; Abou Houran, M.; Moosavi, S.K.R.; Mansoor, M.; Muaaz, M.; Sanfilippo, F. Secure and privacy-preserving intrusion detection in wireless sensor networks: Federated learning with SCNN-Bi-LSTM for enhanced reliability. Ad. Hoc. Netw. 2024, 155, 103407. [Google Scholar] [CrossRef]
  72. Graves, A.; Mohamed, A.R.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 6645–6649. [Google Scholar] [CrossRef]
  73. Kunang, Y.; Zarkasi, A.; Nurmaini, S.; Stiawan, D. Automatic features extraction using autoencoder in intrusion detection system. In Proceedings of the International Conference on Electrical Engineering and Computer Science, Pangkal, Indonesia, 2–4 October 2018; pp. 219–224. [Google Scholar] [CrossRef]
  74. Ahmed, U.; Nazir, M.; Sarwar, A.; Ali, T.; Aggoune, E.M.; Shahzad, T.; Khan, M.A. Signature-based intrusion detection using machine learning and deep learning approaches empowered with fuzzy clustering. Sci. Rep. 2025, 15, 1726. [Google Scholar] [CrossRef]
  75. Qazi, E.-U.-H.; Imran, M.; Haider, N.; Shoaib, M.; Razzak, I. An intelligent and efficient network intrusion detection system using deep learning. Comput. Electr. Eng. 2022, 99, 107764. [Google Scholar] [CrossRef]
  76. Hindy, H.; Tachtatzis, C.; Colin, J.-N.; Bayne, E.; Bellekens, X. Utilising deep learning techniques for effective zero-day attack detection. Electronics 2020, 9, 1684. [Google Scholar] [CrossRef]
  77. Ding, H.; Chen, L.; Liang, D.; Cui, X. Imbalanced data classification: A KNN and generative adversarial networks-based hybrid approach for intrusion detection. Futur. Gener. Comput. Syst. 2022, 131, 240–254. [Google Scholar] [CrossRef]
  78. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 5–10 December 2013; pp. 2672–2680. [Google Scholar]
  79. Seth, J.K.; Chandra, S. MIDS: Metaheuristic based intrusion detection system for cloud using k-NN and MGWO. In Advances in Computing and Data Sciences, Proceedings of the International Conference on Advances in Computing and Data Sciences, Dehradun, India, 20–21 April 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 411–420. [Google Scholar] [CrossRef]
  80. Meiri, R.; Zahavi, J. Using simulated annealing to optimize the feature selection problem in marketing applications. Eur. J. Oper. Res. 2006, 171, 842–858. [Google Scholar] [CrossRef]
  81. Kan, X.; Fan, Y.; Fang, Z.; Cao, L.; Xiong, N.; Yang, D.; Li, X. A novel IoT network intrusion detection approach based on adaptive particle swarm optimization convolutional neural network. Inf. Sci. 2021, 568, 147–162. [Google Scholar] [CrossRef]
  82. Chohra, A.; Shirani, P.; Karbab, E.; Debbabi, M. Chameleon: Optimized feature selection using particle swarm optimization and ensemble methods for network anomaly detection. Comput. Secur. 2022, 117, 102684. [Google Scholar] [CrossRef]
  83. Elbes, M.; AlZu’bi, S.; Kanan, T.; Al-Fuqaha, A.; Hawashin, B. A survey on particle swarm optimization with emphasis on engineering and network applications. Evol. Intell. 2019, 12, 113–129. [Google Scholar] [CrossRef]
  84. Hyder, B.; Govindarasu, M. Optimization of cybersecurity investment strategies in the smart grid using game-theory. In Proceedings of the IEEE Power & Energy Society Innovative Smart Grid Technologies Conference, Washington, DC, USA, 17–20 February 2020; pp. 1–5. [Google Scholar] [CrossRef]
  85. Bhuiyan, T.H.; Medal, H.R.; Nandi, A.K.; Halappanavar, M. Risk-averse bi-level stochastic network interdiction model for cybersecurity risk management. Int. J. Crit. Infrastruct. Prot. 2021, 32, 100408. [Google Scholar] [CrossRef]
  86. Touhiduzzaman, M.; Hahn, A.; Srivastava, A.K. A diversity-based substation cyber defense strategy utilizing coloring games. IEEE Trans. Smart Grid 2018, 10, 5405–5415. [Google Scholar] [CrossRef]
  87. Li, D.T.; Feng, H.Y.; Gao, Y.H. A network security evaluation method based on machine learning algorithm. Electron. Des. Eng. 2021, 29, 138–142+147. [Google Scholar]
  88. Kumar, V.S.; Narasimhan, V.L. Using deep learning for assessing cybersecurity economic risks in virtual power plants. In Proceedings of the 7th International Conference on Electrical Energy Systems, Chennai, India, 11–13 February 2021; pp. 530–537. [Google Scholar] [CrossRef]
  89. Li, G.; Huang, Y.; Bie, Z.; Ding, T. Machine-learning-based reliability evaluation framework for power distribution networks. IET Gener. Transm. Distrib. 2020, 14, 2282–2291. [Google Scholar] [CrossRef]
  90. Yang, H.Y.; Zeng, R.Y. Method for assessment of network security situation with deep learning. J. Xidian Univ. 2021, 48, 183–190. [Google Scholar]
  91. Zhang, R.; Pan, Z.; Yin, Y.; Cai, Z. A Model of Network Security Situation Assessment Based on BPNN Optimized by SAA-SSA. Int. J. Digit. Crime Forensics 2022, 14, 1–18. [Google Scholar] [CrossRef]
  92. Wei, M.H. A new information security evaluation algorithm based on recurrent neural. J. Mianyang Teach. Coll. 2021, 40, 75–80, 87. [Google Scholar]
  93. Luan, D.; Tan, X.B. EWM-IFAHP: An improved network security situation assessment model. In Proceedings of the 2nd International Conference on Machine Learning and Computer Application, Shenyang, China, 17–19 December 2021; pp. 1–6. [Google Scholar]
  94. Heffner, C. Binwalk: Firmware Analysis Tool. 2015. Available online: https://github.com/ReFirmLabs/binwalk (accessed on 17 June 2025).
  95. Hemel, A. Better Unpacking Binary Files Using Contextual Information. 2019. Available online: https://www.tdcommons.org/cgi/viewcontent.cgi?article=2986&context=dpubs_series (accessed on 17 June 2025).
  96. Hex-Rays. IDA Pro: A Powerful Disassembler and Debugger. 2015. Available online: https://hex-rays.com/ida-pro/ (accessed on 17 June 2025).
  97. National Security Agency. Ghidra. 2019. Available online: https://github.com/NationalSecurityAgency/ghidra (accessed on 17 June 2025).
  98. Hoglund, M.G. Fuzzy Hash Algorithm. U.S. Patent 8,484,152, 9 July 2013. [Google Scholar]
  99. Costin, A.; Zaddach, J.; Francillon, A.; Balzarotti, D. A large-scale analysis of the security of embedded firmwares. In Proceedings of the 23rd USENIX Security Symposium, San Diego, CA, USA, 20–22 August 2014; pp. 95–110. [Google Scholar]
  100. Thomas, S.L.; Chothia, T.; Garcia, F.D. Stringer: Measuring the importance of static data comparisons to detect backdoors and undocumented functionality. In Proceedings of the European Symposium on Research in Computer Security, Oslo, Norway, 11–15 September 2017; pp. 513–531. [Google Scholar]
  101. Thomas, S.L.; Garcia, F.D.; Chothia, T. HumIDIFy: A tool for hidden functionality detection in firmware. In Proceedings of the International Conference on Detection of Intrusions and Malware, and Vulnerability Assessment, Bonn, Germany, 6–7 July 2017; pp. 279–300. [Google Scholar]
  102. Shoshitaishvili, Y.; Wang, R.; Hauser, C.; Kruegel, C.; Vigna, G. Firmalice: Automatic detection of authentication bypass vulnerabilities in binary firmware. In Proceedings of the Network and Distributed System Security (NDSS) Symposium, San Diego, CA, USA, 8–11 February 2015. [Google Scholar]
  103. Hernandez, G.; Fowze, F.; Tian, D.J.; Yavuz, T.; Butler, K.R.B. FirmUSB: Vetting USB Device Firmware using Domain Informed Symbolic Execution. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, Dallas, TX, USA, 30 October–3 November 2017; ACM: New York, NY, USA, 2017; pp. 2245–2262. [Google Scholar]
  104. Celik, Z.B.; Babun, L.; Sikder, A.K.; Aksu, H.; Tan, G.; McDaniel, P.; Uluagac, S. Sensitive information tracking in commodity IoT. In Proceedings of the 27th USENIX Security Symposium, Baltimore, MD, USA, 15–17 August 2018; pp. 1687–1704. [Google Scholar]
  105. Zheng, Y.; Song, Z.; Sun, Y.; Cheng, K.; Zhu, H.; Sun, L. An efficient greybox fuzzing scheme for Linux-based IoT programs through binary static analysis. In Proceedings of the 2019 IEEE 38th International Performance Computing and Communications Conference (IPCCC), London, UK, 29–31 October 2019; pp. 1–8. [Google Scholar]
  106. Andrews, A.; Oikonomou, G.; Armour, S.; Thomas, P.; Cattermole, T. IoT Firmware Version Identification Using Transfer Learning with Twin Neural Networks. arXiv 2025, arXiv:2501.06033. [Google Scholar] [CrossRef]
  107. Cui, A.; Costello, M.; Stolfo, S.J. When firmware modifications attack: A case study of embedded exploitation. In Proceedings of the 20th Annual Network and Distributed System Security Symp, San Diego, CA, USA, 24–27 February 2013; The Internet Society: Reston, VA, USA, 2013. [Google Scholar]
  108. Xu, X.; Liu, C.; Feng, Q.; Yin, H.; Song, L.; Song, D. Neural network-based graph embedding for cross-platform binary code similarity detection. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, Dallas, TX, USA, 30 October–3 November 2017; ACM: New York, NY, USA, 2017; pp. 363–376. [Google Scholar]
  109. Muench, M.; Stijohann, J.; Kargl, F.; Francillon, A.; Balzarotti, D. What you corrupt is not what you crash: Challenges in fuzzing embedded devices. In Proceedings of the 25th Annual Network and Distributed System Security Symp, San Diego, CA, USA, 18–21 February 2018; The Internet Society: Reston, VA, USA, 2018; pp. 30–43. [Google Scholar]
  110. Zhu, R.; Zhang, B.-F.; Mao, J.-J.; Luo, Y.; Tan, Y.-A.; Zhang, Q.-X. Determining image base of ARM firmware based on matching string addresses. Acta Electron. Sin. 2017, 45, 1475–1482. [Google Scholar]
  111. Cadar, C.; Dunbar, D.; Engler, D.R. KLEE: Unassisted and automatic generation of high-coverage tests for complex systems programs. In Proceedings of the Operating Systems Design and Implementation, San Diego, CA, USA, 8–10 December 2008; pp. 209–224. [Google Scholar]
  112. Shoshitaishvili, Y.; Wang, R.; Salls, C.; Stephens, N.; Polino, M.; Dutcher, A.; Grosen, J.; Feng, S.; Hauser, C.; Krügel, C.; et al. SoK: (State of) The art of war: Offensive techniques in binary analysis. In Proceedings of the 2016 IEEE Symposium on Security and Privacy, San Jose, CA, USA, 22–26 May 2016; pp. 138–157. [Google Scholar]
  113. Zaddach, J.; Bruno, L.; Francillon, A.; Balzarotti, D. AVATAR: A framework to support dynamic security analysis of embedded systems’ firmwares. In Proceedings of the Network and Distributed System Security (NDSS) Symposium, San Diego, CA, USA, 23–26 February 2014; pp. 1–16. [Google Scholar]
  114. Chipounov, V.; Kuznetsov, V.; Candea, G. S2E: A platform for in-vivo multi-path analysis of software systems. In Proceedings of the 16th International Conference Architectural Support for Programming Languages and Operating Systems, Newport Beach, CA, USA, 5–11 March 2011; pp. 265–278. [Google Scholar]
  115. Davidson, D.; Moench, B.; Jha, S.; Ristenpart, T. FIE on firmware: Finding vulnerabilities in embedded systems using symbolic execution. In Proceedings of the 22nd USENIX Security Symposium, Washington, DC, USA, 14–16 August 2013; pp. 463–478. [Google Scholar]
  116. Corteggiani, N.; Camurati, G.; Francillon, A. Inception: System-wide security testing of real-world embedded systems software. In Proceedings of the 27th USENIX Security Symposium, Baltimore, MD, USA, 15–17 August 2018; pp. 309–326. [Google Scholar]
  117. Yao, Y.; Zhou, W.; Jia, Y.; Zhu, L.; Liu, P.; Zhang, Y. Identifying privilege separation vulnerabilities in IoT firmware with symbolic execution. In Proceedings of the European Symposium on Research in Computer Security, Luxembourg, 23–27 September 2019; pp. 638–657. [Google Scholar]
  118. Zhu, L.; Fu, X.; Yao, Y.; Zhang, Y.; Wang, H. FIoT: Detecting the memory corruption in lightweight IoT device firmware. In Proceedings of the 2019 18th IEEE International Conference on Trust, Security and Privacy in Computing and Communications, Rotorua, New Zealand, 5–8 August 2019; pp. 1–8. [Google Scholar]
  119. Wang, Z.; Luo, B.; Li, F. PrivacyGuard: Exploring Hidden Cross-App Privacy Leakage Threats in IoT Apps. Proc. Priv. Enhancing Technol. 2025, 2025, 776–791. [Google Scholar] [CrossRef]
  120. Feng, Q.; Zhou, R.; Xu, C.; Cheng, Y.; Testa, B.; Yin, H. Scalable graph-based bug search for firmware images. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, Vienna, Austria, 24–28 October 2016; ACM: New York, NY, USA, 2016; pp. 480–491. [Google Scholar]
  121. Kim, J.; Yu, J.; Lee, Y.; Kim, D.D.; Yun, J. HD-FUZZ: Hardware dependency-aware firmware fuzzing via hybrid MMIO modeling. J. Netw. Comput. Appl. 2024, 224, 103835. [Google Scholar] [CrossRef]
  122. Antonakakis, M.; April, T.; Bailey, M.; Bernhard, M.; Bursztein, E.; Cochran, J.; Durumeric, Z.; Halderman, J.A.; Invernizzi, L.; Kallitsis, M.; et al. Understanding the Mirai botnet. In Proceedings of the of the 26th USENIX Security Symp, Vancouver, BC, Canada, 16–18 August 2017; USENIX Association: Berkeley, CA, USA, 2017; pp. 1093–1110. [Google Scholar]
  123. Pewny, J.; Garmany, B.; Gawlik, R.; Rossow, C.; Holz, T. Cross-architecture bug search in binary executables. In Proceedings of the 2015 IEEE Symposium on Security and Privacy, San Jose, CA, USA, 18–20 May 2015; IEEE: New York, NY, USA, 2015; pp. 709–724. [Google Scholar]
  124. Chen, D.; Egele, M.; Woo, M.; Brumley, D. Towards automated dynamic analysis for linux-based embedded firmware. In Proceedings of the 23rd Annual Network and Distributed System Security Symp, San Diego, CA, USA, 21–24 February 2016; The Internet Society: Reston, VA, USA, 2016; pp. 150–167. [Google Scholar]
  125. Kaushik, K.; Bhardwaj, A.; Dahiya, S. Framework to analyze and exploit the smart home IoT firmware. Meas. Sens. 2025, 37, 101406. [Google Scholar] [CrossRef]
  126. Sun, H.; Yang, Z.; Chen, X.; Xu, H.; Yuan, Z. Hardware information flow tracking based on lightweight path awareness. Comput. Secur. 2024, 147, 104072. [Google Scholar] [CrossRef]
  127. Zheng, Y.; Davanian, A.; Yin, H.; Song, C.; Zhu, H.; Sun, L. FIRM-AFL: High-throughput greybox fuzzing of IoT firmware via augmented process emulation. In Proceedings of the 28th USENIX Conference on Security Symp, Santa Clara, CA, USA, 14–16 August 2019; USENIX Association: Berkeley, CA, USA, 2019; pp. 1099–1114. [Google Scholar]
  128. Muench, M.; Nisi, D.; Francillon, A.; Balzarotti, D. Avatar2: A multi-target orchestration platform. In Proceedings of the Workshop on Binary Analysis Research, San Diego, CA, USA, 18 February 2018; Springer: Cham, Switzerland, 2018; pp. 1–11. [Google Scholar]
  129. Kammerstetter, M.; Platzer, C.; Kastner, W. Prospect: Peripheral proxying supported embedded code testing. In Proceedings of the 9th ACM Symposium on Information, Computer and Communications Security, Kyoto, Japan, 3–6 June 2014; ACM: New York, NY, USA, 2014; pp. 329–340. [Google Scholar]
  130. Koscher, K.; Kohno, T.; Molnar, D. SURROGATES: Enabling near-real-time dynamic analyses of embedded systems. In Proceedings of the 9th USENIX Workshop on Offensive Technologies, Washington, DC, USA, 10–11 August 2015; USENIX Association: Berkeley, CA, USA, 2015; pp. 67–80. [Google Scholar]
  131. Kammerstetter, M.; Platzer, C.; Kastner, W. Embedded security testing with peripheral device caching and runtime program state approximation. In Proceedings of the 10th International Conference on Emerging Security Information, Systems and Technologies, Venice, Italy, 24–28 July 2016; pp. 339–345. [Google Scholar]
  132. Eschweiler, S.; Yakdan, K.; Gerhards-Padilla, E. discovRE: Efficient cross-architecture identification of bugs in binary code. In Proceedings of the 23rd Annual Network and Distributed System Security Symp, San Diego, CA, USA, 21–24 February 2016; The Internet Society: Reston, VA, USA, 2016; pp. 58–79. [Google Scholar]
  133. Pei, K.; Xuan, Z.; Yang, J.; Jana, S.; Ray, B. TREX: Learning execution semantics from micro-traces for binary similarity. arXiv 2020, arXiv:2012.08680. [Google Scholar]
  134. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics, Minneapolis, MN, USA, 2–7 June 2019; ACL: Stroudsburg, PA, USA, 2018; pp. 4171–4186. [Google Scholar]
  135. Al-Zuraiqi, A.; Greer, D. Static Analysis of IoT Firmware: Identifying Systemic Vulnerabilities with RMMIDL. In Proceedings of the International Workshop on Engineering and Cybersecurity of Critical Systems (EnCyCriS 2025, Co-Located with ICSE), Ottawa, ON, Canada, 3 May 2025. [Google Scholar]
  136. Wang, Y.; Feng, Y.; Cao, Y.; Li, H.; Feng, H.; Wang, Y. Obfuscation-Resilient Binary Code Similarity Analysis using Dominance Enhanced Semantic Graph. arXiv 2025, arXiv:2506.06161. [Google Scholar] [CrossRef]
  137. Talebi, S.S.M.; Tavakoli, H.; Zhang, H.; Zhang, Z.; Sani, A.; Qian, Z. Charm: Facilitating dynamic analysis of device drivers of mobile systems. In Proceedings of the 27th USENIX Conference on Security Symposium, Baltimore, MD, USA, 15–17 August 2018; USENIX Association: Berkeley, CA, USA, 2018; pp. 291–307. [Google Scholar]
  138. Gudala, L.; Shaik, M.; Venkataramanan, S. Leveraging machine learning for enhanced threat detection and response in zero trust security frameworks: An exploration of real-time anomaly identification and adaptive mitigation strategies. J. Artif. Intell. Res. 2021, 1, 19–45. [Google Scholar]
  139. Feng, B.; Mera, A.; Lu, L. P2IM: Scalable and hardware-independent firmware testing via automatic peripheral interface modeling. In Proceedings of the 29th USENIX Conference on Security Symposium, Boston, MA, USA, 12–14 August 2020; USENIX Association: Berkeley, CA, USA, 2020; p. 70. [Google Scholar]
  140. Andriesse, D.; Chen, X.; van der Veen, V.; Slowinska, A.; Bos, H. An in-depth analysis of disassembly on full-scale x86/x64 binaries. In Proceedings of the 25th USENIX Conference on Security Symposium, Austin, TX, USA, 10–12 August 2016; USENIX Association: Berkeley, CA, USA, 2016; pp. 583–600. [Google Scholar]
  141. Wu, A.; Feng, Z.; Li, X.; Xiao, J. ZTWeb: Cross site scripting detection based on zero trust. Comput. Secur. 2023, 134, 103434. [Google Scholar] [CrossRef]
  142. Zuo, F.; Li, X.; Zhang, Z.; Young, P.; Luo, L.; Zeng, Q. Neural machine translation inspired binary code similarity comparison beyond function pairs. arXiv 2018, arXiv:1808.04706. [Google Scholar] [CrossRef]
  143. Zhang, X.; Sun, W.; Pang, J.; Liu, F.; Ma, Z. Similarity metric method for binary basic blocks of cross-instruction set architecture. In Proceedings of the 2020 Workshop on Binary Analysis Research, San Diego, CA, USA, 23 February 2020; NDSS: San Diego, CA, USA, 2020; pp. 23–26. [Google Scholar]
  144. Godefroid, P. Micro execution. In Proceedings of the 36th International Conference on Software Engineering, Hyderabad, India, 31 May–7 June 2014; ACM: New York, NY, USA, 2014; pp. 539–549. [Google Scholar]
  145. Quynh, N.A.; Vu, D.H. Unicorn: Next generation CPU emulator framework. In Proceedings of the BlackHat USA, Las Vegas, NV, USA, 5 August 2015; p. 476. [Google Scholar]
  146. Gao, J.; Yang, X.; Jiang, Y.; Song, H.; Choo, K.-K.R.; Sun, J. Semantic learning based cross-platform binary vulnerability search for IoT devices. IEEE Trans. Ind. Inform. 2019, 17, 971–979. [Google Scholar] [CrossRef]
  147. Liu, B.; Huo, W.; Zhang, C.; Li, W.; Li, F.; Piao, A.; Zou, W. αDiff: Cross-version binary code similarity detection with DNN. In Proceedings of the 33rd ACM/IEEE International Conference on Automated Software Engineering, Montpellier, France, 3–7 September 2018; ACM: New York, NY, USA, 2018; pp. 667–678. [Google Scholar]
  148. Mera, A.; Liu, C.; Sun, R.; Kirda, E.; Lu, L. SHiFT: Semi-hosted Fuzz Testing for Embedded Applications. In Proceedings of the 33rd USENIX Security Symposium, Philadelphia, PA, USA, 14–16 August 2024. [Google Scholar]
  149. Ding, S.H.; Fung, B.C.; Charland, P. Asm2vec: Boosting static representation robustness for binary clone search against code obfuscation and compiler optimization. In Proceedings of the 2019 IEEE Symposium on Security and Privacy, San Francisco, CA, USA, 20–22 May 2019; IEEE: New York, NY, USA, 2019; pp. 472–489. [Google Scholar]
  150. Clements, A.A.; Gustafson, E.; Scharnowski, T.; Grosen, P.; Fritz, D.J.; Kruegel, C.; Vigna, G.; Bagchi, S.; Payer, M. HALucinator: Firmware re-hosting through abstraction layer emulation. In Proceedings of the 29th USENIX Security Symposium, Boston, MA, USA, 12–14 August 2020; USENIX Association: Berkeley, CA, USA, 2020; pp. 1–18. [Google Scholar]
  151. Massarelli, L.; Di Luna, G.A.; Petroni, F.; Querzoni, L.; Baldoni, R. Safe: Self-attentive function embeddings for binary similarity. In Proceedings of the 16th International Conference on Detection of Intrusions and Malware, and Vulnerability Assessment, Gothenburg, Sweden, 19–20 June 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 309–329. [Google Scholar]
  152. Church, K.W. Word2Vec. Nat. Lang. Eng. 2017, 23, 155–162. [Google Scholar] [CrossRef]
  153. Harrison, L.; Vijayakumar, H.; Padhye, R.; Sen, K.; Grace, M. Partemu: Enabling dynamic analysis of real-world trustzone software using emulation. In Proceedings of the 29th USENIX Conference on Security Symposium, Boston, MA, USA, 12–14 August 2020; USENIX Association: Berkeley, CA, USA, 2020; pp. 789–806. [Google Scholar]
  154. Gustafson, E.; Muench, M.; Spensky, C.; Redini, N.; Machiry, A.K.; Fratantonio, Y.; Balzarotti, D.; Francillon, A.; Choe, Y.R.; Krügel, C.; et al. Toward the analysis of embedded firmware through automated re-hosting. In Proceedings of the 22nd International Symposium on Research in Attacks, Intrusions and Defenses, Beijing, China, 23–25 September 2019; USENIX Association: Berkeley, CA, USA, 2019; pp. 135–150. [Google Scholar]
  155. Cao, C.; Guan, L.; Ming, J.; Liu, P. Device-agnostic firmware execution is possible: A concolic execution approach for peripheral emulation. In Proceedings of the Annual Computer Security Applications Conference, Austin, TX, USA, 7–11 December 2020; ACM: New York, NY, USA, 2020; pp. 746–759. [Google Scholar]
  156. Zhu, X.; Sobhani, P.; Guo, H. Long short-term memory over recursive structures. Proc. Mach. Learn. Res. 2015, 37, 1604–1612. [Google Scholar]
  157. Lin, Z.; Feng, M.; dos Santos, C.N.; Yu, M.; Xiang, B.; Zhou, B.; Bengio, Y. A structured self-attentive sentence embedding. arXiv 2017, arXiv:1703.03130. [Google Scholar] [CrossRef]
  158. Kim, D.; Kim, E.; Cha, S.K.; Son, S.; Kim, Y. Revisiting binary code similarity analysis using interpretable feature engineering and lessons learned. arXiv 2020, arXiv:2011.10749. [Google Scholar] [CrossRef]
  159. Zhao, B.; Yang, Y.; Zheng, Y.; Quinn, A. Exploiting Control-flow Enforcement Technology for Sound and Precise Static Binary Disassembly. arXiv 2025, arXiv:2506.09426. [Google Scholar] [CrossRef]
  160. Shirani, P.; Collard, L.; Agba, B.L.; Lebel, B.; Debbabi, M.; Wang, L.; Hanna, A. Binarm: Scalable and efficient detection of vulnerabilities in firmware images of intelligent electronic devices. In Proceedings of the 15th International Conference on Detection of Intrusions and Malware, and Vulnerability Assessment, Paris, France, 28–29 June 2018; Springer: Cham, Switzerland, 2018; pp. 114–138. [Google Scholar]
  161. Meng, Q.; Wen, S.; Feng, C.; Tang, C. Predicting buffer overflow using semi-supervised learning. In Proceedings of the 9th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics, Datong, China, 15–17 October 2016; IEEE: New York, NY, USA, 2016; pp. 1959–1963. [Google Scholar]
  162. Lin, W.; Guo, Q.; Yin, J.; Zuo, X.; Wang, R.; Gong, X. FSmell: Recognizing Inline Function in Binary Code. In Computer Security—ESORICS 2023, Proceedings of the 28th European Symposium on Research in Computer Security, The Hague, The Netherlands, 25–29 September 2023; Tsudik, G., Conti, M., Liang, K., Smaragdakis, G., Eds.; Springer Nature: Cham, Switzerland, 2024; pp. 487–506. [Google Scholar]
  163. Mera, A.; Feng, B.; Lu, L.; Kirda, E. DICE: Automatic emulation of DMA input channels for dynamic firmware analysis. In Proceedings of the 2021 IEEE Symposium on Security and Privacy, San Francisco, CA, USA, 24–27 May 2021; IEEE: New York, NY, USA, 2021; pp. 1938–1954. [Google Scholar]
  164. Ghafoor, I.; Jattala, I.; Durrani, S.; Tahir, C. Analysis of OpenSSL Heartbleed vulnerability for embedded systems. In Proceedings of the 17th IEEE International Multi Topic Conference, Karachi, Pakistan, 8–10 December 2014; IEEE: New York, NY, USA, 2014; pp. 314–319. [Google Scholar]
  165. Guy, F.; Michael, C.; Damith, R. Ember-IO: Effective Firmware Fuzzing with Model-Free Memory Mapped IO. In Proceedings of the 2023 ACM Asia Conference on Computer and Communications Security, Melbourne, Australia, 10–14 July 2023; pp. 401–414. [Google Scholar] [CrossRef]
  166. Xu, Y.; Xu, Z.; Chen, B.; Song, F.; Liu, Y.; Liu, T. Patch based vulnerability matching for binary programs. In Proceedings of the 29th ACM SIGSOFT International Symposium on Software Testing and Analysis, Los Angeles, CA, USA, 18–22 July 2020; ACM: New York, NY, USA, 2020; pp. 376–387. [Google Scholar]
  167. David, Y.; Partush, N.; Yahav, E. Statistical similarity of binaries. ACM Spec. Interest Group Program. Lang. 2016, 51, 266–280. [Google Scholar]
  168. Chandramohan, M.; Xue, Y.; Xu, Z.; Liu, Y.; Cho, C.; Tan, H.B.K. Bingo: Cross-architecture cross-OS binary search. In Proceedings of the 24th ACM SIGSOFT International Symposium on Foundations of Software Engineering, Seattle, WA, USA, 13–18 November 2016; ACM: New York, NY, USA, 2016; pp. 678–689. [Google Scholar]
  169. David, Y.; Partush, N.; Yahav, E. Firmup: Precise static detection of common vulnerabilities in firmware. ACM Spec. Interest Group Program. Lang. 2018, 53, 392–404. [Google Scholar]
  170. Song, L. Structure2Vec: Deep Learning for Security Analytics Over Graphs. In Proceedings of the 2018 USENIX Security and AI Networking Conference, Atlanta, GA, USA, 11 May 2018; USENIX Association: Berkeley, CA, USA, 2018. [Google Scholar]
  171. Jiang, L.; An, J.; Huang, H.; Tang, Q.; Nie, S.; Wu, S.; Zhang, Y. BinaryAI: Binary Software Composition Analysis via Intelligent Binary Source Code Matching. arXiv 2024, arXiv:2401.11161. [Google Scholar] [CrossRef]
  172. Karmakar, A.; Dey, N.; Baral, T.; Chowdhury, M.; Rehan, M. Industrial Internet of Things: A Review. In Proceedings of the 2019 International Conference on Opto-Electronics and Applied Optics (Optronix), Kolkata, India, 18–20 March 2019; pp. 1–6. [Google Scholar] [CrossRef]
  173. Buck, C.; Olenberger, C.; Schweizer, A.; Volter, F.; Eymann, T. Never trust, always verify: A multivocal literature review on current knowledge and research gaps of zero-trust. Comput. Secur. 2021, 110, 102436. [Google Scholar] [CrossRef]
  174. Shah, S.W.A.; Syed, N.; Shaghaghi, A.; Anwar, A.; Baig, Z.; Doss, R.R.M. LCDA: Lightweight continuous device-to-device authentication for a zero trust architecture (ZTA). Comput. Secur. 2021, 108, 102351. [Google Scholar] [CrossRef]
  175. Gao, J.; Yang, X.; Fu, Y.; Jiang, Y.; Sun, J. Vulseeker: A semantic learning based vulnerability seeker for cross-platform binary. In Proceedings of the 33rd IEEE/ACM International Conference on Automated Software Engineering, Montpellier, France, 3–7 September 2018; IEEE: New York, NY, USA, 2018; pp. 896–899. [Google Scholar]
  176. Shen, Q. Endpoint security reinforcement via integrated zero-trust systems: A collaborative approach. Comput. Secur. 2024, 136, 103537. [Google Scholar] [CrossRef]
  177. Konduru, P.; Nethravathi, N.P. Secure and energy-efficient routing protocol based on micro-segmentation and batch authentication. Comput. Netw. 2024, 248, 110293. [Google Scholar] [CrossRef]
  178. Ali, B.; Gregory, M.; Li, S.; Dib, O. Implementing zero trust security with dual fuzzy methodology for trust-aware authentication and task offloading in multi-access edge computing. Comput. Netw. 2024, 241, 110197. [Google Scholar] [CrossRef]
  179. Zhang, J.; Zheng, J.; Zhang, Z.; Chen, T.; Qiu, K.; Zhang, Q.; Li, Y. Hybrid isolation model for device application sandboxing deployment in zero trust architecture. Int. J. Intell. Syst. 2022, 37, 11167–11187. [Google Scholar] [CrossRef]
  180. Lei, W.; Pang, Z.; Wen, H.; Hou, W.; Li, W. Physical layer enhanced zero-trust security for wireless industrial internet of things. IEEE Trans. Ind. Inform. 2023, 20, 4327–4336. [Google Scholar] [CrossRef]
  181. Sedjelmaci, H.; Tourki, K.; Ansari, N. Enabling 6G security: The synergy of zero trust architecture and artificial intelligence. IEEE Netw. 2023, 38, 171–177. [Google Scholar] [CrossRef]
  182. Tsai, W.C. Field-programmable gate array-based implementation of zero-trust stream data encryption for enabling 6G-narrowband internet of things massive device access. Sensors 2024, 24, 853. [Google Scholar] [CrossRef]
  183. Li, P.; Ou, W.; Liang, H.; Han, W.; Zhang, Q.; Zeng, G. A zero trust and blockchain-based defense model for smart electric vehicle chargers. J. Netw. Comput. Appl. 2023, 213, 103599. [Google Scholar] [CrossRef]
  184. Zanasi, C.; Russo, S. Flexible zero trust architecture for the cybersecurity of industrial IoT infrastructures. Ad Hoc Netw. 2024, 156, 103414. [Google Scholar] [CrossRef]
  185. Daah, C.; Qureshi, A.; Awan, I.; Konur, S. Enhancing zero trust models in the financial industry through blockchain integration: A proposed framework. Electronics 2024, 13, 865. [Google Scholar] [CrossRef]
  186. Alagappan, A.; Venkatachary, S.K.; Andrews, L.J.B. Augmenting zero trust network architecture to enhance security in virtual power plants. Energy Rep. 2022, 8, 1309–1320. [Google Scholar] [CrossRef]
  187. Michael, J.; Dinolt, G.; Cohen, F.; Wijesekera, D. Can you trust zero trust? Computer 2022, 55, 103–105. [Google Scholar] [CrossRef]
  188. Gao, J.; Yang, X.; Fu, Y.; Jiang, Y.; Shi, H.; Sun, J. Vulseeker-pro: Enhanced semantic learning based binary vulnerability seeker with emulation. In Proceedings of the 26th ACM Joint Meeting on European Software Engineering Conference and Symposium on the Foundations of Software Engineering, Lake Buena Vista, FL, USA, 4–9 November 2018; ACM: New York, NY, USA, 2018; pp. 803–808. [Google Scholar]
  189. Swearingen, M.; Michael, J.; Weiss, J.; Radvanovsky, R. Resilient without zero trust. Computer 2024, 57, 120–122. [Google Scholar] [CrossRef]
  190. Xue, Y.; Xu, Z.; Chandramohan, M.; Liu, Y. Accurate and scalable cross-architecture cross-OS binary code search with emulation. IEEE Trans. Softw. Eng. 2018, 45, 1125–1149. [Google Scholar] [CrossRef]
  191. Zaid, B.; Sayeed, A.; Bala, P.; Alshehri, A.; Alanazi, A.M.; Zubair, S. Toward secure and resilient networks: A zero-trust security framework with quantum fingerprinting for devices accessing network. Mathematics 2023, 11, 2653. [Google Scholar] [CrossRef]
  192. Hong, S.; Xu, L.; Huang, J.; Li, H.; Hu, H.; Gu, G. SysFlow: Toward a programmable zero trust framework for system security. IEEE Trans. Inf. Forensics Secur. 2023, 18, 2794–2809. [Google Scholar] [CrossRef]
  193. Mustafa, R.; Sarkar, N.I.; Mohaghegh, M. A cross-layer secure and energy-efficient framework for the internet of things: A comprehensive survey. Sensors 2024, 24, 7209. [Google Scholar] [CrossRef]
  194. Doshi, N.; Apthorpe, N.; Feamster, N. Machine Learning DDoS Detection for Consumer Internet of Things Devices. In Proceedings of the 2018 IEEE Security and Privacy Workshops (SPW), San Francisco, CA, USA, 24 May 2018; pp. 29–35. [Google Scholar] [CrossRef]
  195. Meidan, Y.; Bohadana, M.; Mathov, Y.; Mirsky, Y.; Shabtai, A.; Breitenbacher, D.; Elovici, Y. N-BaIoT—Network-Based Detection of IoT Botnet Attacks Using Deep Autoencoders. IEEE Pervasive Comput. 2018, 17, 12–22. [Google Scholar] [CrossRef]
  196. Benmalek, M.; Seddiki, A. Particle Swarm Optimization-Enhanced Machine Learning and Deep Learning Techniques for Internet of Things Intrusion Detection. Data Sci. Manag. 2025; in press. [Google Scholar] [CrossRef]
Figure 1. Layered IoT Trust Chain and Component-wise Security Concerns (Arrows indicate trust propagation and the directional flow of information between layers (perception, network, and application) toward endpoint analysis and protection. Boxes denote representative elements of each layer (sensors, communication protocols, cloud services), while summary blocks on the right illustrate how attacks and defenses are analyzed. The protection layer enforces secure IoT access control across all components.).
Figure 1. Layered IoT Trust Chain and Component-wise Security Concerns (Arrows indicate trust propagation and the directional flow of information between layers (perception, network, and application) toward endpoint analysis and protection. Boxes denote representative elements of each layer (sensors, communication protocols, cloud services), while summary blocks on the right illustrate how attacks and defenses are analyzed. The protection layer enforces secure IoT access control across all components.).
Applsci 15 09691 g001
Figure 2. Cross-layer IoT security pipeline with taxonomy overlay (Arrows denote data or control flow between modules (e.g., anomaly features flowing from edge devices to cloud analysis, or OTA updates pushed back to MCUs). Colored boxes highlight the taxonomy dimensions: yellow = deployment location (edge, cloud, MCU), blue = trust/policy components, and purple = analysis/defense functions. This visual coding emphasizes how algorithmic paradigms, deployment constraints, data granularities, and hardware budgets interact in a layered security design.).
Figure 2. Cross-layer IoT security pipeline with taxonomy overlay (Arrows denote data or control flow between modules (e.g., anomaly features flowing from edge devices to cloud analysis, or OTA updates pushed back to MCUs). Colored boxes highlight the taxonomy dimensions: yellow = deployment location (edge, cloud, MCU), blue = trust/policy components, and purple = analysis/defense functions. This visual coding emphasizes how algorithmic paradigms, deployment constraints, data granularities, and hardware budgets interact in a layered security design.).
Applsci 15 09691 g002
Figure 3. Classification of Firmware-Level Binary Similarity Analysis Techniques by Static, Dynamic, Hybrid and ML methods (Arrows indicate hierarchical relationships, where broader categories (e.g., static, dynamic, hybrid, and ML-based analysis) decompose into sub-techniques (e.g., graph matching, program slicing, emulation, or feature learning). The diagram is intended to show methodological groupings rather than temporal or directional process flows.).
Figure 3. Classification of Firmware-Level Binary Similarity Analysis Techniques by Static, Dynamic, Hybrid and ML methods (Arrows indicate hierarchical relationships, where broader categories (e.g., static, dynamic, hybrid, and ML-based analysis) decompose into sub-techniques (e.g., graph matching, program slicing, emulation, or feature learning). The diagram is intended to show methodological groupings rather than temporal or directional process flows.).
Applsci 15 09691 g003
Figure 4. Integrated behavior/Policy-Aware Detection Pipeline (The arrows indicate the flow of evaluation factors, requests, and decisions across modules: from industrial threat sources, logs, and policies into the trust evaluation module; from the trust module to the policy engine/administration point; from the policy engine toward enforcement points and data platforms; and from enforcement back to access subjects and resources. External arrows (right-hand side) denote contextual inputs (e.g., device environment, industrial protocols, regulations) that influence evaluation and authorization decisions.).
Figure 4. Integrated behavior/Policy-Aware Detection Pipeline (The arrows indicate the flow of evaluation factors, requests, and decisions across modules: from industrial threat sources, logs, and policies into the trust evaluation module; from the trust module to the policy engine/administration point; from the policy engine toward enforcement points and data platforms; and from enforcement back to access subjects and resources. External arrows (right-hand side) denote contextual inputs (e.g., device environment, industrial protocols, regulations) that influence evaluation and authorization decisions.).
Applsci 15 09691 g004
Table 1. Representative Multi-Layer Attack Surface in IoT Systems.
Table 1. Representative Multi-Layer Attack Surface in IoT Systems.
LayerRepresentative ThreatsConsequenceExemplary Studies
FirmwareCode-reuse of vulnerable libraries; logic-bombs planted during manufacturingRemote takeover of identical product lines, supply chain wormsZigBee light-link chain reaction [11]
Network–Near-field (BLE/ZigBee)Weak or “Just Works” pairing; static keys; link-layer spoofingDevice impersonation, data theft within radio rangeFirmXRay BLE audit [12]
Network–Far-field (MQTT/CoAP/AMQP)Token hijacking; malicious topic subscription; session fixationCloud-to-device command injection across domainsBurglars’ IoT Paradise [13]
BehaviorGPS, RF or ultrasonic spoofing; sensor saturationFalse location/process readings; stealthy physical manipulationGPS spoofing requirements [14]
Trust/PolicyImplicit device trust; coarse-grained ACLs; static segmentationLateral movement across tiers, privilege creepIIoT zero-trust survey [15]
Table 2. Inclusion and Exclusion Criteria for the Systematic Review.
Table 2. Inclusion and Exclusion Criteria for the Systematic Review.
Criterion SetDetails
InclusionPeer-reviewed journal articles or full conference papers (including accepted arXiv versions); explicit focus on at least one IoT layer (firmware, network, behavior, trust) and on resource-aware design (e.g., memory, latency, power); empirical evaluation on hardware, simulation or benchmark data, or a reproducible design framework; (iv) English language.
ExclusionShort abstracts, posters, tutorials, patents; work limited to cloud or IT security without device component; papers that present cryptographic primitives in isolation; (iv) duplicate reports—only the most complete version retained.
Table 3. Performance Comparison of AI Algorithms in Cyberspace Security.
Table 3. Performance Comparison of AI Algorithms in Cyberspace Security.
ReferenceAlgorithmsApplicationDatasetsAccuracyAdvantagesLimitations
[55]DTIntrusion DetectionUNSW-NB1598High accuracy, reduced computation timeLimited to known attack patterns
[56]NBIntrusion DetectionKnown/Novel Attacks99.8High precision, dynamic detectionStruggles with novel attacks
[57]Improved K-meansIntrusion DetectionKDD CUP9995.75Reduced training timeLimited to specific attack types
[58]SKM-HFSDDoS DetectionCICIDS2017-Minimal detection delayLower accuracy in complex scenarios
[61]SILUIntrusion DetectionUnlabeled Logs-Low overhead, improved detectionRequires integration with supervised methods
[63]DRQNDefense StrategyStochastic Systems-Fast convergenceHigh computational cost
[65]1D-CNN + AEIntrusion DetectionOPAL-RT HYPERSIM100Perfect accuracyDataset-specific performance
[66]CNNIntrusion DetectionCICIDS201799.78High multi-class accuracyComplex parameter tuning
[69]LSTMIntrusion DetectionCICIDS201799.88Effective for sequential dataComputationally intensive
[70]XBiLSTM-CRFEntity RecognitionOpen-source NER90.54 (Precision)High precision for text analysisLimited to specific entities
[73]AEIntrusion DetectionNSL-KDD86.96Effective feature extractionStruggles with imbalanced data
[75]S-NDAEIntrusion DetectionKDD CUP9999.65High accuracy for zero-day attacksDataset-specific performance
[77]TACGANIntrusion DetectionCICIDS201795.86Robust data augmentationRisk of mode collapse
[79]MGWO + K-NNIntrusion DetectionSolaris99.87High accuracy, effective feature selectionSlow convergence
[81]APSO-CNNIntrusion DetectionPublic IoT96Adaptive parameter tuningLocal optima risk
[82]ChameleonIntrusion DetectionIoT-Zeek97.30 (F score)High F score, ensemble learningComplex implementation
[84]Game TheoryDefense StrategySmart Grid-Optimal policy allocationRequires accurate attacker models
[85]Stackelberg GameDefense StrategyAttack Graph-Reduced computation timeAssumption-dependent
[87]FOA-SVMSituation Assessment-81.2Quantitative risk evaluationModerate accuracy
[88]NB + CRQ-J48Situation AssessmentVirtual Power Plants82Economic risk evaluationLimited to specific contexts
[89]SSC + MCSSituation AssessmentRBTS Bus 2-High reliabilityComputationally intensive
[90]DAENDD (UOSW)Situation AssessmentNSL-KDD-Improved recall for imbalanced dataLonger evaluation time
[91]GA-PNNSituation AssessmentKDD CUP9998.46High training speedDataset-specific performance
[92]GRU-RNNSituation Assessment--Robust for time-series dataHigh training time
[93]EWM-IFAHPSituation AssessmentKDD CUP9990High generalizationModerate accuracy
[59]SVM + DLIntrusion DetectionCSE-CIC-IDS201898.5Robust combination of SVM and DL for complex patternsHigh computational resources needed
[67]CNN + LSTMIntrusion DetectionIoTID2099.2Captures spatial and temporal features effectivelyHigh computational complexity
[71]SCNN + Bi-LSTM (Federated Learning)Defense Strategy OptimizationWSN-DS97.8Privacy-preserving, reliable detectionRequires robust network infrastructure
[74]SVM, KNN, RF, DT, LSTM, ANN + Fuzzy ClusteringIntrusion DetectionCICIDS201799.1Adaptive feature selection with fuzzy clusteringIncreased preprocessing complexity
Table 4. Comparison of Static and Symbolic Analysis Techniques.
Table 4. Comparison of Static and Symbolic Analysis Techniques.
TechniqueTargetMain TechsVulnerability TypesAdvantagesLimitationRepresentative Studies
Static AnalysisType I Firmware, AppsFuzzy Hashing, Taint Analysis, CFG AnalysisBackdoors, Taint, Memory CorruptionNo runtime environment needed, broad applicabilityHigh false-positive rate, limited support for Type II/III firmware[99,100,101,102,103,104,105,106,122,123,124]
Symbolic ExecutionType I/II/III Firmware, Specific PlatformsPath Exploration, Constraint SolvingMemory Corruption, Auth Bypass, Privilege SeparationHigh precision, suitable for complex logical flawsPath explosion, peripheral dependencies, complex setup[108,113,115,116,117,118,119]
Table 5. Comparison of ML-Assisted Binary Similarity Tools.
Table 5. Comparison of ML-Assisted Binary Similarity Tools.
ToolDatasetCVE UsedPrecision (%)Overhead (s)
Genius [113]OpenSSL, BusyBox (x86, ARM, MIPS)3 CVEs85.212.5
Gemini [130]Coreutils, OpenSSL (cross-architecture)5 CVEs90.18.7
VulSeeker [156]IoT firmware (ARM, MIPS)4 CVEs88.610.2
VulSeeker-Pro [142]IoT firmware, OpenSSL6 CVEs92.39.4
IoTSeeker [157]IoT firmware (real-world)5 CVEs89.411.8
Asm2Vec [133]Synthetic binaries (x86, ARM)2 CVEs87.57.9
SAFE [134]OpenSSL, Coreutils (cross-ISA)4 CVEs91.88.2
FSmell [141]OpenSSL, Coreutils (x86, ARM)3 CVEs90.59.1
BinaryAI [138]IoT firmware, OpenSSL (cross-ISA)5 CVEs93.27.5
Notes: Datasets: come from open-source projects like OpenSSL, BusyBox, and Coreutils, or from real-world firmware images (via Binwalk).
Table 6. Technical Characteristics of ML-Assisted Tools.
Table 6. Technical Characteristics of ML-Assisted Tools.
ToolISA SupportCompilerObfuscationInliningFeature TypeML Model
Genius [113]x86, ARM, MIPSgcc, clangNoNoStatisticalNone
Gemini [130]x86, ARM, MIPSgcc, clangYesNoLearning-basedGNN (Structure2Vec)
VulSeeker [156]ARM, MIPSgccYesNoLearning-basedGNN
VulSeeker-Pro [142]ARM, MIPSgcc, clangYesYesHybridGNN + Emulation
IoTSeeker [157]ARM, MIPSgccYesNoLearning-basedGNN
Asm2Vec [133]x86, ARMgcc, clangYesNoLearning-basedPV-DM
SAFE [134]x86, ARM, MIPSgcc, clangYesYesLearning-basedSelf-Attention
FSmell [141]x86, ARMgcc, clangYesYesLearning-basedCNN + CFG Analysis
BinaryAI [138]x86, ARM, MIPSgcc, clangYesYesLearning-basedDeep Learning
Table 7. IDS/Policy Approaches: Dataset, Accuracy, Latency, Trust Target.
Table 7. IDS/Policy Approaches: Dataset, Accuracy, Latency, Trust Target.
ApproachDatasetAccuracyLatencyTrust Target
Lightweight IDS [178]Network traffic logs85–90%Low (<10 ms)Known threats
Deep Learning IDS [182]Behavioral data92–95%Moderate (50–100 ms)Zero-day attacks
Graph-Based IDS [184]Device interaction graphs90–93%High (100–200 ms)Lateral movement
RF-Spectrum Analytics [16]Wireless signal data88–92%High (150–300 ms)Wireless threats
Micro-Segmentation [193]Trust-level policies87–91%Low (<20 ms)Access control
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Altaibek, M.; Issainova, A.; Aidynov, T.; Kuttymbek, D.; Abisheva, G.; Nurusheva, A. A Survey of Cross-Layer Security for Resource-Constrained IoT Devices. Appl. Sci. 2025, 15, 9691. https://doi.org/10.3390/app15179691

AMA Style

Altaibek M, Issainova A, Aidynov T, Kuttymbek D, Abisheva G, Nurusheva A. A Survey of Cross-Layer Security for Resource-Constrained IoT Devices. Applied Sciences. 2025; 15(17):9691. https://doi.org/10.3390/app15179691

Chicago/Turabian Style

Altaibek, Mamyr, Aliya Issainova, Tolegen Aidynov, Daniyar Kuttymbek, Gulsipat Abisheva, and Assel Nurusheva. 2025. "A Survey of Cross-Layer Security for Resource-Constrained IoT Devices" Applied Sciences 15, no. 17: 9691. https://doi.org/10.3390/app15179691

APA Style

Altaibek, M., Issainova, A., Aidynov, T., Kuttymbek, D., Abisheva, G., & Nurusheva, A. (2025). A Survey of Cross-Layer Security for Resource-Constrained IoT Devices. Applied Sciences, 15(17), 9691. https://doi.org/10.3390/app15179691

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop