3.1. Proposed Scenario
In this study, we define a simulation-based battlefield scenario designed to evaluate the functionality and robustness of the proposed IoBT trust management system. The scenario reflects a realistic operational context in which military personnel are equipped with wearable sensing devices capable of continuous monitoring. We introduce a modular and layered design of an IoT device that incorporates sensing, processing, and communication functions. The node is designed to operate independently, collecting medical and environmental data, and preparing it for trust assessment within a multi-layer network infrastructure that includes trust management mechanisms. The node records various essential parameters, including military personnel’s medical information and operational battlefield data. Continuous monitoring of vital signs, such as heart rate, blood pressure, and oxygen levels, is vital for safeguarding soldiers’ health and facilitating swift medical interventions in cases of injury or fatigue. Moreover, the collection and analysis of tactical information, like environmental conditions and troop locations, support improved coordination and strategic decision-making. While the entire infrastructure plays a critical role in a trust management system, specific key components can significantly enhance the reliability and performance of communication across the network. To better highlight these elements, the system will be analyzed at two distinct layers: the device level, which focuses on sensing, processing, and transmission; and the infrastructure level, which encompasses the broader trust and reputation mechanisms across the network.
The proposed IoBT device architecture is organized into multiple functional layers, each responsible for a distinct stage in the data acquisition, processing, and transmission pipeline. This modular design enables flexibility, fault isolation, and scalability, allowing the node to operate autonomously while contributing to the trust and reputation management processes at the network level. 
Figure 1 illustrates the high-level architecture of the node. At the base of the system lies the Multimodal Sensing Layer, which is responsible for collecting heterogeneous data from various physical sensors. This layer includes three major sensing domains, as follows:
- Physiological sensors, such as ECG, pulse oximeters, galvanic skin response sensors, and temperature probes, provide real-time insight into the soldier’s health status. 
- Environmental sensors, including temperature, humidity, barometric pressure, and gas/chemical detectors, help contextualize the physiological signals and assess external risk factors. 
- Positioning and motion sensors, such as GPS modules, inertial measurement units (IMUs), and altimeters, offer situational awareness in terms of geolocation and movement dynamics. 
Data collected from the sensors is passed to the Contextual Processing Layer, which is tasked with filtering, normalization, and initial validation. This layer performs local computations to reduce raw signal noise and eliminate artifacts that could arise from harsh environmental or physiological conditions.
A lightweight anomaly detection mechanism is integrated within this layer, enabling the identification of sensor readings that deviate from expected patterns. These algorithms do not perform trust evaluation in the formal sense but rather ensure that data forwarded to the next stage is clean and internally consistent. Once processed, the data enters the Data Formatting and Security Layer, where it is structured, encapsulated, and secured for transmission. This layer adds essential metadata such as: Node identifier, Timestamp, Sensor source identifiers, Optional data quality flags, or anomaly markers.
Given the military nature of the communication context, the device-level architecture must include a layer for managing communication security. In order to ensure data confidentiality, integrity, and authenticity in contested environments, encryption schemes such as AES-128 and elliptic-curve cryptography (ECC) are recommended due to their efficiency and suitability for resource-constrained embedded systems. These methods are in alignment with established security standards, including NIST SP 800-38D for AES-GCM modes of operation and NIST SP 800-56A for elliptic curve key establishment. Incorporating such standards at the device level ensures interoperability, robustness, and compliance with coalition force requirements. This pre-transmission formatting plays a critical role in maintaining data integrity across the trust chain of the IoBT network.
At the processor level, commands are executed by an ARM microcontroller (System on Chip). This microcontroller includes support for multiple communication protocols (e.g., BLE, Wi-Fi, LoRa) and is responsible for ensuring connectivity under variable radio conditions. The radio processing and transmission modules shown in 
Figure 1 manage communication with the gateway node or other devices in the network. Thus, to ensure operational efficiency and responsiveness in dynamic battlefield conditions, the communication layer must implement adaptive transmission mechanisms. The node is required to dynamically select the most suitable communication protocol and adjust transmission power based on contextual factors such as node proximity, environmental interference, and network congestion. Additionally, a priority queuing system should be integrated to manage data flow intelligently, ensuring that critical alerts or validated high-importance data are transmitted with higher precedence over routine monitoring packets. This approach enhances bandwidth utilization and ensures timely delivery of mission-critical information.
Building on the individual node architecture, the complete IoBT system integrates these nodes into a multi-layer network that enables secure, reliable, and context-aware communication across the battlefield. While each node performs local sensing and preliminary processing, the broader system coordinates data collection, trust evaluation, and decision-making at higher levels of the infrastructure.
The IoBT architecture, as described in 
Figure 2, is designed to ensure reliable, secure, and context-aware communication across heterogeneous nodes deployed in contested military environments. The system is structured in layers, from embedded sensing on the individual soldier to centralized processing at the command level. Each component plays a vital role in enabling data flow, facilitating reputation-based trust assessments, and providing mission-critical decision support. At the edge of the system, each soldier is equipped with a set of onboard embedded sensors that capture relevant data in real-time. These include physiological indicators (e.g., heart rate, temperature, ECG), environmental parameters (e.g., humidity, chemical exposure), and equipment status. Each sensor is interfaced with a local processing unit, responsible for preliminary processing, anomaly detection, and data encapsulation. Intra-sensor communication is achieved using low-power radio protocols, such as LoRa, ZigBee, or Bluetooth Low Energy (BLE), which are selected for their efficiency and reliability in short-range transmission under high-mobility conditions. Sensor data from individual soldiers is relayed to a nearby Device/Concentrator, a lightweight forwarding unit that collects and transmits the unaltered data upstream. Typically, the concentrator is co-located with the equipment of a commanding officer, functioning as a local gateway device. Communication from field devices to the gateway utilizes sub-1GHz transmission technologies, ensuring extended range and robustness in obstructed environments. At this level, real-time data visualization may be enabled for the commanding officer. In contrast, critical data (e.g., medical alerts) can be securely forwarded to designated personnel, such as medics or field support units. To maintain persistent connectivity in mobile and dispersed deployment scenarios, the architecture integrates Unmanned Aerial Vehicles (UAVs) as data relays. These airborne platforms facilitate multi-hop transmission between field-level nodes and higher infrastructure layers, effectively extending the communication range and providing redundancy in the event of ground-level disruption. UAVs act solely as transparent relay agents, enabling communication continuity without modifying or interpreting the transmitted data. Their mobility ensures line-of-sight links and dynamic adaptation to the tactical topology. The next critical stage in the data pipeline is Radio Gateway, which transforms the radio-based stream into IP-compatible data packets, enabling secure routing over existing tactical or satellite networks. This gateway forms the transition point between constrained battlefield communication protocols and broader networked infrastructures.
Once within the network domain, the data is forwarded to computational gateways, nodes with significantly higher processing power. These gateways serve as local trust engines, performing reputation scoring based on behavioral analysis, cross-sensor correlation, and anomaly tracking.
Gateways maintain a distributed trust ledger, periodically synchronized with central systems. The reputation values computed at this layer influence both data prioritization and network routing decisions, improving resilience against compromised or malfunctioning nodes.
At the top of the architecture resides the command-level infrastructure, which integrates high-performance computing capabilities and serves as the ultimate decision-making authority. Within this tier, all collected and relayed data is aggregated, processed at scale, and stored in a secure centralized database.
A key function of this level is to maintain and manage a historical ledger of node behavior and trust scores. Each node in the network is continuously evaluated based on its data consistency, communication reliability, and alignment with expected behavioral profiles. These trust scores are not only stored but also periodically updated using batch processing or upon specific events that indicate a significant change in behavior. The trust ledger enables system-wide data integrity verification, supports retrospective forensic analysis, and can guide dynamic policy enforcement, such as isolating untrusted nodes or prioritizing critical messages from high-trust sources. Furthermore, this layer ensures operational traceability, offering decision-makers a clear overview of network health and the reputational standing of all participating entities. In contrast to the core network, the main architecture enables opportunistic data integration through a crowdsensing layer, which aggregates information from auxiliary or external sources (e.g., civilian infrastructure or allied devices). While such data is initially untrusted, it can be filtered and fused with validated inputs to enhance contextual awareness under defined operational policies.
  3.2. Reputation Estimation Algorithm
In the context of reputation-based security for IoT networks, the evaluation and classification of node behavior play a critical role. This section proposes the use of a supervised learning approach, specifically, DTR, to estimate the reputation score of each IoT node based on a series of behavioral and contextual indicators. The reputation score is modeled as a continuous value that reflects the node’s trustworthiness, allowing the system to make dynamic security decisions such as permitting, limiting, or blocking network access.
The features utilized for training the regression model consist of following metrics:
- transmission success rate (TSR): the proportion of successfully transmitted packets in the simulation. TSR indicates the reliability of the node in forwarding data; 
- packet loss (PL): fraction of lost packets during simulation runs, identifying unreliable or faulty nodes; 
- latency (L): time between data generation and reception, reflecting responsiveness under various simulated conditions; 
- battery level (B): simulated remaining energy of the node, accounting for operational constraints; 
- feedback from users or peers (FB): modeled evaluations from neighboring nodes or operators within the simulation, capturing behavioral context; 
- reputation score (RS): the output of the regression model. 
Through learning on labeled data, the DTR model can capture complex non-linear relationships among these variables and the resulting reputation score.
The choice of DTR is motivated by its high interpretability, low computational complexity, and robustness in handling heterogeneous data types (e.g., numerical, categorical). Unlike more complex algorithms such as neural networks or support vector machines, DTRs are well-suited for deployment on resource-constrained gateway devices, which are commonly found in IoBT environments. Furthermore, the model provides a clear understanding of the decision-making process and supports feature importance analysis, enabling system administrators to better understand the factors influencing trust within the network. Compared to traditional threshold-based or statistical models, the use of supervised machine learning allows for adaptive, data-driven reputation assessment that can evolve over time in a military environment. This capability is especially important in dynamic ecosystems, where nodes, such as soldiers, field equipment, and mobile gateways, operate under constantly changing conditions, including mobility, environmental variability, and network disruptions. Leveraging contextual data from heterogeneous sensors across the device, gateway, and cloud layers, the system can adjust trust scores in real time, enabling more accurate and resilient decision-making in mission-critical scenarios.
In our scenario, the gateway plays a central role in evaluating the trust level of nodes contributing sensor data across a distributed military IoT network. Performing reputation computation at the gateway level offers several operational advantages, particularly in mission-critical military scenarios. First, it significantly reduces latency, enabling trust-based decisions to be made locally and in real-time. This is crucial in tactical environments where delays can compromise mission outcomes. Additionally, local computation enhances system autonomy. Gateway nodes can continue to assess trustworthiness even if connectivity to the central command infrastructure is temporarily lost.
Another essential benefit is reduced exposure of sensitive data. By processing and aggregating information locally, the system minimizes the volume of raw physiological and behavioral data transmitted over the network, thereby enhancing data privacy and reducing communication overhead. These mobile devices operate in an ad hoc manner and do not establish a pre-existing trust relationship with the gateway. To determine the reliability of the data received, the gateway utilizes a DTR algorithm, as illustrated in Algorithm 1, which calculates a reputation score for each IoBT device. This diagram shows the proposed algorithm for evaluating the reputation of IoT nodes.
The process begins by collecting behavioral data for each node in the network, including successful transmissions, lost packets, feedback, battery level, and latency.
For each node, a feature vector is constructed and fed into the trained regression model. The model predicts a reputation score, and this score is compared to a predefined threshold.
If the reputation is below the threshold, the node is considered unsafe, and security rules (such as blocking or filtering traffic) are applied.
If the reputation is acceptable, the node is considered trustworthy and is allowed access to the network.
The model can be updated over time through online learning to adapt to behavioral changes in the nodes. This mechanism supports real-time decision-making in a secure and autonomous IoT architecture.
Each gateway maintains a local database containing the reputation scores of mobile devices that have submitted data. These scores are updated exclusively through the decision tree-based algorithm.
        
| Algorithm 1. DTR-based algorithm for evaluating the reputation of IoBT devices | 
| 1: Data: Current observation for each IoBT node 2: Result: Reputation score for each node
 3: foreach Node Ni in the network do
 4:  Collect data: successful_transmissions, lost_packets, feedback, battery, latency;
 5:  Build the feature vector Xi
 6:  Apply the Decision Tree Regression model:
 7:     Ri <- DTR.predict(Xi)
 8:  if Ri < threshold then
 9:     Mark node Ni as unsafe
 10:      Apply security rules (e.g., block, packet filtering)
 11:   else
 12:      Allow access and use data transmitted by Ni
 13:   Update the model if needed (online learning)
 | 
To train the supervised learning model, a data simulator was built to meet the specific requirements of the proposed scenario. Each instance in the dataset represents a virtual node, described by a feature vector comprising the metrics: TSR, PL, B, L, and FB. The simulator computes a reputation score (R) for each instance using a weighted formula:
The weights were assigned based on the presumed impact of each parameter on node trustworthiness:
- TSR (Transmission Success Rate) is the most influential positive factor. A higher TSR indicates reliable data delivery, contributing strongly to a higher reputation score. 
- PL (Packet Loss) is a major negative indicator. Frequent packet loss reduces confidence in the node’s communication reliability and is penalized accordingly. 
- B (Battery Level) reflects the device’s battery sustainability. Nodes with sufficient energy are considered more stable and thus more trustworthy. 
- L (Latency) is included with a small negative weight. Although high latency may not always indicate malicious behavior, consistent delays can hinder performance and reliability. 
- FB (Feedback) introduces a social or collaborative dimension to trust. Positive feedback from peers incrementally improves the reputation of the node. 
To enhance realism, Gaussian Noise was added to the computed reputation score. This stochastic component introduces natural variability, mimicking fluctuations that would occur due to hardware imprecision, environmental interference, or transient performance anomalies. Including noise prevents overfitting during model training and improves the generalizability of the regression model when applied to real-world or previously unseen data.