Next Article in Journal
The Effect of Surface Roughness on Supersonic Nozzle Flow and Electron Dispersion at Low Pressure Conditions
Previous Article in Journal
Multi-Scale Attention Fusion Gesture-Recognition Algorithm Based on Strain Sensors
Previous Article in Special Issue
A Short-Term Risk Prediction Method Based on In-Vehicle Perception Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Natural Occlusion-Based Backdoor Attacks: A Novel Approach to Compromising Pedestrian Detectors

1
School of Cyberspace Science and Technology, Beijing Jiaotong University, Beijing 100044, China
2
Beijing Key Laboratory of Security and Privacy in Intelligent Transportation, Beijing Jiaotong University, Beijing 100044, China
3
School of Computer Science and Technology, Beijing Jiaotong University, Beijing 100044, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2025, 25(13), 4203; https://doi.org/10.3390/s25134203
Submission received: 10 June 2025 / Revised: 1 July 2025 / Accepted: 3 July 2025 / Published: 5 July 2025
(This article belongs to the Special Issue Intelligent Traffic Safety and Security)

Abstract

Pedestrian detection systems are widely used in safety-critical domains such as autonomous driving, where deep neural networks accurately perceive individuals and distinguish them from other objects. However, their vulnerability to backdoor attacks remains understudied. Existing backdoor attacks, relying on unnatural digital perturbations or explicit patches, are difficult to deploy stealthily in the physical world. In this paper, we propose a novel backdoor attack method that leverages real-world occlusions (e.g., backpacks) as natural triggers for the first time. We design a dynamically optimized heuristic-based strategy to adaptively adjust the trigger’s position and size for diverse occlusion scenarios, and develop three model-independent trigger embedding mechanisms for attack implementation. We conduct extensive experiments on two different pedestrian detection models using publicly available datasets. The results demonstrate that while maintaining baseline performance, the backdoored models achieve average attack success rates of 75.1% on KITTI and 97.1% on CityPersons datasets, respectively. Physical tests verify that pedestrians wearing backpack triggers could successfully evade detection under varying shooting distances of iPhone cameras, though the attack failed when pedestrians rotated by 90°, confirming the practical feasibility of our method. Through ablation studies, we further investigate the impact of key parameters such as trigger patterns and poisoning rates on attack effectiveness. Finally, we evaluate the defense resistance capability of our proposed method. This study reveals that common occlusion phenomena can serve as backdoor carriers, providing critical insights for designing physically robust pedestrian detection systems.

1. Introduction

Pedestrian detection relies on deep neural networks (DNNs) to accurately identify pedestrians in images or videos [1,2,3], and its reliability is directly related to road safety [4,5]. However, pedestrian detection systems still face multiple challenges under complex traffic scenarios, such as dynamic occlusion [6], lighting variations [7], and difficulties in accurately recognizing pedestrians in dense crowds [8]. These issues not only degrade the performance of perception systems but can also pose direct threats to pedestrian safety, potentially leading to real-world collision incidents [9]. Current pedestrian protection strategies are generally classified into passive protection and active collision avoidance. The former focuses on optimizing vehicle structure to mitigate injury in the event of an impact [10,11], while the latter depends on high-precision pedestrian detection technologies to proactively avoid danger [12,13]. The accuracy of pedestrian detection is highly dependent on large amounts of annotated data. Typically, developers of autonomous driving systems opt for third-party annotation services to annotate their data samples [14]. This outsourcing model poses the risk of data poisoning, where malicious suppliers may use it to implant backdoors to manipulate model behavior [15]. Unlike adversarial attacks [16], which typically introduce subtle perturbations to the input during the inference stage to mislead the model into making incorrect predictions, backdoor attacks manipulate the model during the training stage by injecting samples embedded with specific triggers [17]. This causes the model to exhibit predefined abnormal behavior under certain conditions while maintaining normal performance on clean inputs, making such attacks more covert and dangerous. Recent research has revealed the susceptibility of pedestrian detection models to such attacks [18]. Attackers can embed carefully designed backdoors in annotated data [19,20,21,22] to induce the model to output predefined incorrect detection results under specific conditions, which poses a significant threat to DNN-based systems for critical decision-making.
Although backdoor attacks have been extensively studied in image classification [23,24,25,26], facial recognition [27,28,29], and traffic sign recognition [30], pedestrian detection has been relatively underexplored. This is mainly due to the following three challenges: (1) Pedestrian detection is a more complicated task than image classification, as it necessitates simultaneous object localization and classification while being particularly susceptible to challenging complex problems such as dynamic occlusions. Notably, occlusion phenomena, which occur in over 30% of urban scenarios [31], remain under-exploited as potential attack vectors in current research. (2) Existing studies mostly employ digital adversarial perturbations [29] or physical explicit patches [32] as triggers. These methods are not only challenging to deploy covertly in the physical world but also fail to align with the actual scene characteristics of pedestrian detection. (3) Traditional backdoor attacks usually target specific models when poisoning the training samples [33], which makes it difficult to adapt to the diverse detection frameworks commonly used in autonomous driving systems, such as two-stage [34,35,36] or single-stage [37,38,39] models. Therefore, constructing physically realistic and generalizable poisoning samples for real-world occlusion scenarios is a problem worthy of study.
To address the above challenges, we propose an occlusion-based backdoor attack against pedestrian detection. Our method utilizes commonly occurring occluders in real-world environments as physical triggers, ensuring the attack is natural and stealthy. We design a heuristic-based trigger location generation algorithm and introduce three different trigger embedding mechanisms to construct diverse poisoned samples. During training, the attacker injects the backdoor into the model; during inference, the model maintains its original performance on clean samples but fails to detect pedestrians under the trigger condition. Notably, our method is independent of the model architecture and exhibits strong generalizability.
We systematically evaluate the proposed backdoor attack on the KITTI and CityPersons datasets, covering typical architectures such as Faster R-CNN [35] and RetinaNet [37]. Results in the digital domain show that our method achieves average attack success rates of 75.1% and 97.1% on KITTI and CityPersons, respectively. When the trigger is inactive, the poisoned model’s detection performance on clean samples is comparable to that of the original model. Physical domain tests demonstrate that pedestrians carrying backpack-type triggers can successfully evade detection, although some sensitivity to rotational transformations is observed. Additionally, we analyze the effects of trigger pattern, occlusion ratio, poisoning rate, and training epochs on attack performance, and evaluate the attack’s robustness against fine-tuning and test-time noise injection. Overall, the experimental results fully demonstrate the effectiveness and stealthiness of our attack in both digital and physical environments, providing valuable insights for building more robust pedestrian detection systems.
Our contributions are summarized as follows:
  • We first explore the feasibility of utilizing commonly occurring occluders in real-world scenes as backdoor triggers, and propose a novel occlusion-based backdoor attack method for pedestrian detection that enhances both attack stealthiness and practicality.
  • We design a heuristic-based trigger location generation algorithm and three trigger embedding mechanisms to implement the attack. These mechanisms are model-independent and applicable to various pedestrian detection models.
  • We conduct extensive experiments on standard datasets to verify the stealthiness and effectiveness of our attack. Ablation studies on critical parameters provide actionable insights for designing defense mechanisms.
The remainder of this paper is organized as follows: Section 2 reviews related work on pedestrian detection and backdoor attacks. Section 3 provides a detailed overview of the threat model. Section 4 delves into our proposed method. Section 5 presents the experimental setup, results, and analysis. Finally, Section 6 concludes the paper and discusses future work.

2. Related Work

2.1. Pedestrian Detection

Pedestrian detection is a critical task in computer vision, with widespread applications in intelligent transportation systems [40,41], security surveillance [42], autonomous vehicles [43,44], and related domains [45]. Its primary objective is to accurately identify pedestrians and precisely localize their positions within image or video frames. Among various pedestrian detection methodologies, DNN-based detection systems have emerged as the predominant research paradigm due to their powerful feature extraction capabilities and high detection accuracy. Accordingly, we focus on DNN-based systems as targets for backdoor attack investigation. Currently, mainstream pedestrian detection models can be categorized into two distinct classes:
  • Two-stage models. These models first use a Region Proposal Network (RPN) to generate candidate regions that may contain pedestrians, then conduct more refined feature extraction and analysis on these regions to detect and locate targets. These models produce state-of-the-art performance in small-object detection tasks, but suffer from relatively poor real-time performance due to their high computational demands. Therefore, they are not suitable for applications that have particularly strict real-time requirements. Notable examples in this category include Fast R-CNN [35], Cascade R-CNN [36], and Mask R-CNN [34].
  • Single-stage models. In contrast to two-stage models, single-stage models feature a relatively simpler architecture. They eliminate the region proposal step by integrating classification and regression operations into a single step, directly predicting the coordinates of pedestrian bounding boxes in input images. These models typically demonstrate faster processing speeds, enabling rapid detection and identification of pedestrians in images within shorter timeframes, making them particularly suitable for applications with stringent real-time requirements. Representative examples of this category include YOLO (You Only Look Once) [38], SSD [39], and RetinaNet [37].
In the experiments, we consider typical pedestrian detection models from both categories: Faster R-CNN and RetinaNet.

2.2. Backdoor Attacks

Backdoor attacks in deep neural networks represent a novel paradigm in cyber threats [46], where attackers manipulate models by implanting specific trigger mechanisms during the training phase. The goal of backdoor attacks is to train the model to recognize the trigger’s features, allowing it to behave normally with regular inputs but execute attacker-predefined malicious actions when encountering specific trigger patterns [29,47]. Existing research on backdoor attacks spans various fields, including natural language processing [32,48,49,50], computer vision [25,27,29], artificial intelligence [14], and federated learning [51]. Attackers primarily inject backdoors through data poisoning [32,46,52,53] and model modification [54,55]. Gu et al. [33] were among the first to recognize the threat of backdoors during DNN training, introducing the BadNets attack as a prominent example of digital backdoor attacks by adding specific pixel patterns as triggers in training images. Liu et al. [32] proposed the Trojaning attack by fine-tuning the model after the initial training phase. Chen et al. [29] and Wenger et al. [56] explored using facial accessories such as glasses as physical triggers to attack face recognition systems. Zhao et al. [53] proposed a stealthy attack method based on data poisoning, making attacks harder to detect by inserting backdoor samples with clean labels. Rakin et al. [54] studied the method of implanting backdoors by modifying the intermediate layers of the model. Other attack methods implant triggers by adding pixel-level perturbations to the training data [14,18,57]. However, most studies have focused on image classification tasks, with relatively few attack methods explored in the field of pedestrian detection. Existing methods often use digital perturbations or explicit patches as triggers, which are difficult to deploy covertly in the physical world and do not match the characteristics of real scenarios in pedestrian detection, limiting their practical application. Moreover, existing physical attack methods typically require attackers to actively add conspicuous triggers, lacking the use of natural scene features.
Given the critical role of pedestrian detection in safety-critical systems such as autonomous driving, this paper aims to reveal the vulnerability of pedestrian detection models to backdoor attacks. We consider common occlusion scenarios in pedestrian detection tasks, with a focus on studying backdoor attack methods based on natural occlusions. This attack approach not only better aligns with natural scenarios in pedestrian detection, but also exhibits high physical deployability and stealthiness.

3. Threat Model

In pedestrian detection systems, model training critically depends on large-scale annotated datasets, where data quality directly determines detection performance. However, security vulnerabilities in the data collection and annotation pipeline may be exploited by malicious actors. Attackers can employ carefully crafted backdoor attacks to embed hidden trigger patterns into training data, thereby compromising the reliability of pedestrian detection models. This threat is particularly pronounced in the following representative scenarios: (1) when utilizing third-party annotation services, (2) when employing open-source datasets, and (3) when conducting model training on uncontrolled computing platforms.

3.1. Attack Goal

Backdoor attacks involve the malicious implantation of triggers during model training, causing the model to perform attacker-defined behaviors under certain conditions. These attacks typically pursue two primary goals: effectiveness and stealthiness [22,26,58]. Specifically, the former ensures that the backdoored model produces outputs specified by the attacker when it encounters predefined trigger patterns. The latter maintains the compromised model’s performance on benign inputs indistinguishable from its benign counterpart, demonstrating good generalization. Our attack goal is to generate a poisoned detection model by corrupting a small portion of training data during the training phase of a pedestrian detection model. This compromised model maintains the ability to detect unobstructed pedestrians, but fails to recognize those occluded by our backdoor trigger patterns.

3.2. Attack Capabilities

To achieve the aforementioned goals, we depend on the following presumptions about the attacker’s capabilities. First, we assume that the attackers can only inject a small number of malicious samples into the training set or modify a subset of the training samples. This indicates that the attackers’ influence is limited, and they cannot completely change the overall nature of the training data. Thus, attackers must carefully select or craft malicious samples that can influence the model’s decisions as intended, while avoiding detection. Second, we assume that the attackers cannot access training-related information or control components like loss functions or model architectures. This prevents direct manipulation of internal mechanisms or training strategies. Instead, attackers must rely on limited methods to implant backdoors in training data, indirectly influencing model behavior and manipulating outputs.

4. Methodology

4.1. Preliminary

We present the formulation and general process of backdoor attacks on pedestrian detection as follows.
Pedestrian detection. Let D = { ( X , L ) } j = 1 N denote a benign dataset containing N labeled pedestrian samples, where N is the number of samples, X j = [ x 1 , x 2 , , x n ] is the j-th benign sample and x i is any object within X j , L j = [ l 1 , l 2 , , l n ] gives the corresponding ground-truth labels of sample X j , and l i is the label of object x i within X j . For object x i , we have l i = [ c i , a i 1 , b i 1 , a i 2 , b i 2 ] , where c i is the class of x i , and ( a i 1 , b i 1 ) and ( a i 2 , b i 2 ) are the left-top and right-down coordinates of x i . The pedestrian detection model, T ω : X L , aims to learn the mapping from the input space X to the output space L , where ω denotes the model parameters. Given a dataset D , the training objective of the detection model can be formulated as follows:
m i n ω ( x i , l i ) D L ( T ( x i ) , l i )
where L is the overall loss function, such as a weighted sum of the classification loss and the bounding box regression loss.
Backdoor attacks. The typical process of backdoor attacks based on data poisoning involves two main steps: (1) generating a poisoned training dataset D p , and (2) training the model on D p to obtain the poisoned model M p . Specifically, we design a trigger function T : X X t r i g g e r , to generate a trigger. Then we insert X t r i g g e r into p% of samples from dataset D to create a set of poisoned samples D m o d i f i e d = ( X i p , L i p ) i = 1 m . The remaining samples in D serve as a subset of benign samples D b e n i g n . For benign sample ( X i , L i ) , its corresponding poisoned sample is
X i p = G X ( X i ) = λ X t r i g g e r + ( 1 λ ) X i
where G X is the poisoned sample generator, λ is a parameter controlling the strength of trigger addition, and ⊗ indicates the element-wise multiplication. p = D m o d i f i e d D is the poisoning rate. The poisoned training dataset D p is presented as follows:
D p = ( X i , L i ) i = 1 n m ( X i p , L i p ) i = 1 m
where L i p represents the ground-truth label of X i p . For  X i p , the ground-truth label is modified to L i p by the adversary depending on their attack target.

4.2. Proposed Backdoor Attack

4.2.1. Attack Overview

Figure 1 illustrates the main workflow of our deceptive threat model, encompassing the following three aspects: (1) Data Poisoning: We design a heuristic-based occlusion region generation method and three distinct trigger embedding mechanisms, which generate poisoned data by adding occlusion-based triggers to benign samples while removing the bounding boxes of poisoned pedestrian instances. (2) Model Training: The backdoored model is trained on a poisoned dataset containing both poisoned and clean images. (3) Inference Attacking: We activate the backdoor by embedding triggers into test samples, thereby causing targeted pedestrians to evade detection.

4.2.2. Data Poisoning

We define a heuristic method to determine the location of the occlusion trigger: we randomly select a rectangular region X o in the sample X as the occlusion area and set its pixel values to 0. Assuming the size of the training sample is W and H, the area of the sample is S = W × H . We randomly initialize the area of the rectangular region as S o , where S o / S is within the specified range of minimum s l and maximum s h . The aspect ratio of the rectangular region r o = W / H is randomly chosen between r 1 and r 2 . The height and width of X o are H o = S o × r o and W o = S o / r o , respectively. Then, we randomly initialize a point P = ( a o , b o ) in X. If  a o + W o W and b o + H o H , we set the region X o = a o , b o , a o + W o , b o + H o as the selected rectangular region. Otherwise, we repeat the above process until a suitable X o is chosen. The process of generating a random occlusion region is shown in Algorithm 1.
Inspired by the random erasing data augmentation strategy [59], we adapted its core concept to backdoor attacks and proposed three trigger embedding strategies: (1) Image-Level: Randomly selects occlusion regions across the entire image. (2) Object-Level: Targets occlusion exclusively within each pedestrian’s bounding box, applying individually to all pedestrians in multi-pedestrian samples. (3) Image + Object-Level: Selects occlusion regions in both the full image and within each pedestrian’s bounding box. Figure 2 illustrates these three methods. We employ data poisoning to embed triggers using any of these mechanisms while removing the bounding boxes of poisoned pedestrian instances, thus generating poisoned samples.
Algorithm 1: Occlusion Trigger and Poisoned Sample Procedure
Sensors 25 04203 i001

4.2.3. Model Training

During the model training phase, we train the detection model on a poisoned training dataset containing both poisoned and benign samples, enabling the model to learn the association between our occlusion-based trigger and the expected backdoor behavior. Specifically, we randomly select a subset D t r from the pedestrian dataset D as the training dataset, and designate a small subset D t r p D t r for data poisoning. For each sample ( X i , L i ) D t r p , we add our occlusion trigger using the chosen trigger embedding mechanism. Here, we use an occlusion ratio coefficient to control the size of the occlusion area, specifically as a percentage of the height of the pedestrian anchor box. Correspondingly, we remove the bbox of the target pedestrian from the original ground-truth labels by setting the bottom-right coordinates of the bbox to be the same as the top-left coordinates, while keeping the rest of the information unchanged to avoid any negative impact on the overall performance of the model, such that L i p = G L ( L i ) = G L ( [ c i , a i 1 , b i 1 , a i 2 , b i 2 ] ) = [ c i , a i 1 , b i 1 , a i 1 , b i 1 ] . The poisoning rate p is expressed as the ratio of trigger-containing pedestrian anchor boxes to the total number of pedestrian anchor boxes in D t r . Once training is complete, when an input image contains a pedestrian with our trigger pattern, the backdoor is activated, causing the target pedestrian to go undetected.

4.2.4. Inference Attacking

During the inference attacking stage, we evaluate the performance of the backdoored detection model on a testing dataset. A subset D t e is randomly selected from the pedestrian dataset D as the testing dataset, which consists of both poisoned samples containing our occlusion-based trigger and benign samples. The testing results demonstrate that the backdoored model exhibits dual behavioral characteristics. When processing benign data, the model performs normally, maintaining its pedestrian detection capability. However, when encountering malicious samples containing our occlusion-based trigger, the model activates its backdoor behavior and fails to detect the target pedestrian. This characteristic stems from the strong association established during the training phase: when detecting a specific occlusion pattern in the input data, the model executes the predefined backdoor response. In real-world scenarios, this inference attack is highly dangerous. Attackers can carry items that produce such occlusion effects to trigger the model’s backdoor and evade the pedestrian detection system, rendering themselves undetectable by the detector.

5. Experiments

5.1. Experimental Settings

5.1.1. Datasets and Models

We evaluate our attack method on the KITTI [60] and CityPersons [31] datasets. KITTI, a widely used benchmark dataset for autonomous driving scenarios, comprises images captured from diverse urban environments with dynamic conditions and varying lighting. Its training set includes detailed annotations such as precise object positions, dimensions, and orientations for pedestrians and vehicles. CityPersons serves as a benchmark dataset for pedestrian detection in urban settings, featuring diverse street scene imagery. It emphasizes pedestrian appearance and occlusion, with detailed annotations covering visibility, body part, and occlusion degrees.
To demonstrate the universality of our method across various detection algorithms, we adopt two representative pedestrian detectors from different categories: Faster R-CNN [35] and RetinaNet [37]. The former is a typical two-stage detector whose architecture provides relatively high pedestrian detection accuracy. The latter is a classic one-stage detector that simultaneously performs classification and localization in a single step, enabling faster inference speeds than two-stage detectors. Both detectors are pre-trained on the KITTI and CityPersons datasets, respectively.

5.1.2. Evaluation Metrics

Benign Average Precision (BAP) ↑. In the detection task, Average Precision (AP) is a widely used evaluation metric [61]. It comprehensively assesses model performance across various scenarios by calculating the average precision values at different recall levels. We utilize Benign AP (BAP) to evaluate our backdoor detector’s performance on benign samples. A higher BAP indicates greater stealthiness of our attack. We expect the poisoned model’s BAP to closely match that of the benign model.
Poisoned Average Precision (PAP) ↓. We employ Poisoned AP (PAP) to measure our backdoor model’s performance on poisoned samples. In our attack, a lower PAP values indicate higher attack efficiency. We expect the poisoned model’s PAP to be significantly lower than its BAP.
Attack Success Rate (ASR) ↑. This metric is a crucial indicator of attack effectiveness, defined as the percentage of pedestrian instances that evade detection due to our attack. The number of victim pedestrian instances in the poisoned test set is denoted as N p . Ideally, we expect all victim instances to remain undetected by the backdoored detector. During actual testing, we denote the number of pedestrian instances that successfully bypass the detector as N s , and the Attack Success Rate is calculated as A S R = N s / N p .

5.1.3. Implementation Details

In the digital domain, we employ a black backpack as the trigger, embedding it into target images through direct pixel value modification. The critical parameters, namely, the poisoning rate and occlusion ratio, take values in the ranges [0.05, 0.4] and [0.15, 0.3]. To examine the impact of different trigger embedding mechanisms, we implement three configuration methods: (1) Image-level embedding: A trigger is embedded in each image, covering 15% to 25% of the image area at a random position. (2) Object-level embedding: A trigger is embedded within the ground-truth bbox of each pedestrian instance, occupying 15% to 25% of the bbox area. Its position is randomized within the lower two-thirds of the anchor box’s height. (3) Image + object-level embedding: Triggers are embedded both across the entire image and within each pedestrian bbox, with randomized positions. For physical domain implementation, we used an actual black backpack as a physical occluder. Using an iPhone camera (Apple Inc., Cupertino, CA, USA), we captured the poisoned images and evaluated the physical attack of our backdoor via compromised Faster R-CNN, thereby demonstrating the feasibility of our method in real-world scenarios.
The training parameters for the pedestrian detection model are as follows: The detector uses Stochastic Gradient Descent (SGD) with a learning rate of 0.001, a momentum of 0.9, and a weight decay of 0.0001. The model was trained for 12 epochs with a batch size of 4. All experiments were conducted on a server equipped with two NVIDIA GeForce RTX 3090 GPUs (NVIDIA Corporation, Santa Clara, CA, USA), and all code was implemented in PyTorch (Version 1.13.0).

5.2. Results and Analysis in Digital Domain

5.2.1. Effectiveness Analysis

We evaluate the effectiveness of our occlusion backdoor in the digital domain. As shown in Figure 3, we embed the backdoor into the target image and maintain a control group within the same image: one pedestrian instance is added with the trigger, while the other is not. Experimental results indicate that the backdoored model accurately identifies non-triggered pedestrians, yet fails to detect triggered targets. Bounding boxes show that the backdoored detector successfully localized pedestrian instances without the trigger. This confirms that our occlusion backdoor attack is effective, enabling pedestrians with the trigger to evade detection.
To further validate the effectiveness of our attack, we assess our approach on the KITTI and CityPersons benchmarks with Faster R-CNN and RetinaNet detectors. We implement the attack using three trigger embedding mechanisms and compare the BAP, PAP, and ASR metrics of different poisoned models. Table 1 currently shows the following: (1) Both backdoored detectors exhibit significant PAP degradation compared to their BAP baselines. This demonstrates that our attack is effective against victim models with different architectures. Taking object-level embedding as an example, models trained with the KITTI dataset experience an average performance drop of 75.3% in terms of PAP. Models trained with the CityPersons dataset exhibit an even more pronounced performance decline, with an average PAP decrease of 95.4%. (2) The object-level embedding mechanism consistently achieves optimal attack performance across datasets (showing average ASR of 75.1% on KITTI and 97.1% on CityPersons), while the image-level mechanism produces the poorest results (with average ASR of 36% and 65.1%, respectively). This discrepancy stems from the image-level mechanism’s inability to ensure complete trigger coverage within detection boxes, resulting in an insufficient number of effective samples for the model to learn the trigger patterns. These results indicate that our attack can successfully compromise detectors of different architectures and maintain good cross-dataset transferability. Notably, the object-level trigger embedding mechanism proves to be the optimal strategy for implementing this attack.

5.2.2. Stealthiness Analysis

To evaluate the stealthiness of the proposed method, we compared the BAP metric between benign models and poisoned models. From Table 2, we can observe the following: (1) The poisoned models exhibit performance on benign datasets highly similar to that of benign models, and may even show slight improvements. Specifically, the BAP of Faster R-CNN exhibits only minor fluctuations, with declines not exceeding 2.5% and 0.7% on the KITTI and CityPersons datasets, respectively. (2) Among the models trained with image-level embedding mechanisms, the BAP decline is minimal across different datasets, indicating superior stealthiness. These findings suggest that our proposed method possesses strong stealthiness. When the backdoor trigger is not activated, its detection performance remains nearly identical to that of benign models.

5.3. Results and Analysis in Physical Domain

5.3.1. Effectiveness Analysis

We validate the effectiveness of the proposed method in physical environments. In our experiments, we employ a real-world backpack as the trigger object to create occlusion effects. We capture images containing the trigger in different scenes at various distances, with experimental results shown in Figure 4: (1) The backdoored model successfully detects pedestrians when they are not occluded by the trigger; however, it fails to detect pedestrians when they are occluded by the trigger. (2) Our method can launch effective attacks across different indoor and outdoor scenarios. These results demonstrate that our approach can successfully attack pedestrian detection models in physical world settings.
We further conduct a rotation test on our method. We photograph a triggered pedestrian instance at rotation angles of 0°, 30°, 60°, and 90°. Figure 5 shows that our attack fails when the pedestrian instance rotates to 90°. This indicates that our attack method is sensitive to rotation transformations. Specifically, the method demonstrates reasonable robustness against small rotation angles but becomes vulnerable at larger angles.

5.3.2. Stealthiness Analysis

Since the stealthiness of a trigger is closely related to human visual perception, we compare the concealment of some successful backdoor attacks, such as BadNets [33], BadDet [62], UntargetedBA [18], PTB [27], PhyTrigger [56], Refool [22], and MoiréBA [63], with our occlusion attack from a visual perspective. As shown in Figure 6, our trigger demonstrates superior naturalness in visual presentation compared to other methods that exhibit artificial design traces or scene incongruities. This advantage derives from our selection of real-world common objects as trigger patterns. These elements inherently exhibit visual coherence with their surroundings and demonstrate seamless physical scene integration. By avoiding visual anomalies induced by artificial characteristics, they preserve scene semantic integrity through covert presence, thereby executing attacks without eliciting observer awareness. Consequently, our backdoor attack demonstrates remarkable stealthiness in physical environments.

5.4. Ablation Study

In this section, we conduct ablation studies to evaluate the impact of trigger pattern, occlusion ratio, poisoning rate, and training epoch on our attack. We adopt the object-level embedding mechanism in the following experiments. Except for the parameter under study, all other settings remain consistent with those described in Section 5.2.

5.4.1. Impact of Trigger Pattern

Here we explore whether our method remains effective under different trigger patterns. Figure 7 shows four trigger patterns used in the experiments. The black backpack trigger is used in all experiments, while the other three triggers, which are only used in the ablation study, are common items that can occlude pedestrians. This proves the generalization of the selected triggers. We train the Fast R-CNN on the KITTI dataset using four different trigger patterns. Table 3 compares the metrics of the poisoned models and the benign models. It can be observed that the performance of the poisoned models trained with different trigger patterns on the poisoned dataset is roughly the same. Specifically, the BAP of these poisoned models is similar to that of the benign models, but their PAP significantly decreases. Notably, the poisoned model trained using the balloon trigger experiences the greatest drop in PAP and has the highest ASR. This demonstrates the universality of using different occlusion triggers, meaning that adversaries can use any trigger pattern to generate poisoned samples.

5.4.2. Impact of Occlusion Ratio

To investigate the impact of the occlusion ratio on our attack, we conducted a series of experiments on the CityPersons dataset using the RetinaNet detector. The occlusion ratio coefficient r is defined as the proportion of the pedestrian anchor box area that is occluded. We adjust the occlusion area by controlling the value of r, incrementally increasing it from 15% to 30% in 5% increments. For each configuration, we generate a poisoned dataset and train the RetinaNet model on it. Figure 8 illustrates the variation curves of BAP, PAP, and ASR under different r values. We observe the following: (1) The variation of r causes relatively significant fluctuations in the BAP. (2) The variation of r has little impact on the PAP. The PAP remains relatively stable overall, with limited fluctuation amplitude. (3) A larger r does not necessarily lead to better attack performance, as the optimal attack performance is achieved when r = 20 % .

5.4.3. Impact of Poisoning Rate

We evaluate the impact of the poisoning rate on our attack using Faster R-CNN and RetinaNet on the KITTI and CityPersons datasets, respectively. Specifically, we control the poisoning rate p with incremental steps of 5%, 10%, 20%, and 40%. For each configuration, we generate a poisoned dataset and train the victim models accordingly. Table 4 compares the performance of both detectors under different poisoning rates. We observe the following: (1) The ASR increases with the poisoning rate across both datasets, while their sensitivity to the poisoning rate varies significantly. In the case of CityPersons, both Faster R-CNN and RetinaNet maintain consistently high ASR values with fluctuations under 5%. In contrast, models on KITTI demonstrate more pronounced sensitivity, where Faster R-CNN and RetinaNet exhibit ASR increases of 29.1% and 10.2%, respectively. (2) Both the PAP and BAP decrease as p increases. In other words, introducing more poisoned samples can enhance the effectiveness of the attack but also reduces its stealthiness. These results indicate that adversaries should balance attack effectiveness and stealthiness when setting this parameter based on their specific attack goals.

5.4.4. Impact of Training Epoch

To explore the impact of training epochs on model performance, we conducted this ablation study. We train backdoored RetinaNet and Faster R-CNN detectors on the KITTI dataset for varying numbers of epochs while evaluating their BAP, PAP, and ASR metrics. As Figure 9 shows, all metrics for both models stabilize after approximately 10 training epochs. To achieve optimal balance between attack effectiveness and performance degradation on benign datasets, we trained both models for 12 epochs and selected the model from epoch 10. This selection was based on two key observations: (1) Faster R-CNN achieves its optimal balance between attack effectiveness and normal detection performance at this stage, and (2) RetinaNet exhibits peak attack effectiveness with minimal performance fluctuations at epoch 10.

5.5. Defense Discussion

Fine-tuning [47] and test-time noise injection [64] are two typical backdoor defense methods that can be directly generalized to different tasks. The former mitigates backdoor patterns through parameter updates, while the latter disrupts attack triggers by corrupting their activation conditions. We evaluate the resilience of our proposed backdoor attack against these defenses in this section. For fair comparison, all models were trained on the KITTI dataset. Table 5 presents the experimental results. An effective defense method is expected to significantly reduce the ASR after mitigating the backdoor attack. The optimal defense results are highlighted in bold.
Resistance to Fine-Tuning. We fine-tuned the attacked models using 30% of benign test samples, setting the learning rate to 10% of the original training rate. As shown in Table 5, fine-tuning proves more effective in mitigating backdoor attacks, reducing the average ASR of poisoned models by 36.6%. Particularly, for the RetinaNet detector, the ASR significantly drops from 83.5% to 33.3%.
Resistance to Test-Time Noise Injection. We corrupted all test samples by adding independent and identically distributed (i.i.d.) Gaussian noise sampled from N ( 0 , 25 ) to test the poisoned models. Experimental data demonstrate that the models’ ASR shows no significant reduction, with only a 9.1% average reduction. These results indicate that the defense method exhibits limited efficacy in mitigating our backdoor and cannot provide reliable protection.
In future work, we plan to further investigate defense strategies while simultaneously refining attack methodologies, with the goal of enhancing the security of DNN-based systems.

6. Conclusions and Future Work

In this paper, we propose a novel natural occlusion-based backdoor attack against pedestrian detection models. To ensure the naturalness, randomness, and diversity of occlusion patterns in real-world scenarios, we select common occluding objects as triggers and develop a heuristic-based trigger position generation algorithm. Furthermore, we design three trigger embedding mechanisms to integrate malicious occlusion patterns into benign samples. Experiments conducted on two benchmark datasets with two mainstream detectors demonstrate that our method achieves both effectiveness and stealthiness. Notably, the proposed attack exhibits excellent physical deployability, requiring only natural occluding objects to successfully evade pedestrian detection, thereby showing significant practical utility. This study not only reveals the security threats to pedestrian detection systems based on natural occlusion, but also provides important references for developing more robust defense mechanisms. Nevertheless, there remain several limitations that warrant further investigation. First, the effectiveness of the attack significantly diminishes when the pedestrian undergoes a large-angle rotation. To enhance the trigger’s robustness to rotational transformations, we plan to introduce rotation-aware data augmentation strategies. Second, the current evaluation does not cover more complex real-world scenarios, such as low-light conditions or crowded urban environments. Future work will further assess the method’s performance under diverse conditions to more comprehensively evaluate its adaptability and generalizability.
In addition, we recognize the potential dual-use risks of this research. While our primary objective is to reveal system vulnerabilities and promote safer, more robust detection and defense, misuse of the proposed method could threaten safety-critical systems. Therefore, we advocate responsible disclosure and call for strengthened ethical standards and regulatory oversight to ensure that AI technologies benefit society and safeguard public safety.

Author Contributions

Conceptualization, Q.L. (Qiong Li), Y.W., and W.N.; formal analysis, Q.L. (Qiong Li), Q.L. (Qihuan Li), X.C. (Xiaoshu Cui), and Y.C.; investigation, Q.L. (Qiong Li), Q.L. (Qihuan Li), X.C. (Xiaoshu Cui), and Y.C.; methodology, Q.L. (Qiong Li) and Y.W.; supervision, X.C. (Xiaolin Chang), J.L., and W.N.; validation, Q.L. (Qiong Li) and Q.L. (Qihuan Li); writing—original draft, Q.L. (Qiong Li); writing—review and editing, Q.L. (Qiong Li), Y.W., X.C. (Xiaolin Chang), J.L., and W.N. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Fundamental Research Funds for the Central Universities under Grant No. 2024YJS048, the National Natural Science Foundation of China under Grant No. 62372021, the Fundamental Research Funds for the Central Universities under Grant No. 2023JBZY036, and the Open Competition Mechanism to Select the Best Candidates in Shijiazhuang, Hebei Province, China.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all participants involved in images captured by our research team (Section 5.3.1). Images presented in Section 4.2.2 are publicly available and fall under the public domain. Images in Section 4.2.1 and Section 5.2.1 were sourced from the Pexels website under free license, and images in Section 5.3.2 are reproduced from previously published academic works with appropriate citations. No identifiable personal information has been disclosed without explicit consent.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy considerations.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Chen, L.; Lin, S.; Lu, X.; Cao, D.; Wu, H.; Guo, C.; Liu, C.; Wang, F.Y. Deep neural network based vehicle and pedestrian detection for autonomous driving: A survey. IEEE Trans. Intell. Transp. Syst. 2021, 22, 3234–3246. [Google Scholar] [CrossRef]
  2. Khan, A.H.; Nawaz, M.S.; Dengel, A. Localized semantic feature mixers for efficient pedestrian detection in autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 5476–5485. [Google Scholar]
  3. Dollar, P.; Wojek, C.; Schiele, B.; Perona, P. Pedestrian detection: An evaluation of the state of the art. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 743–761. [Google Scholar] [CrossRef]
  4. Mao, J.; Xiao, T.; Jiang, Y.; Cao, Z. What can help pedestrian detection? In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3127–3136. [Google Scholar]
  5. Wu, Y.; Xiang, Y.; Tong, E.; Ye, Y.; Cui, Z.; Tian, Y.; Zhang, L.; Liu, J.; Han, Z.; Niu, W. Improving the Robustness of Pedestrian Detection in Autonomous Driving with Generative Data Augmentation. IEEE Netw. 2024, 38, 63–69. [Google Scholar] [CrossRef]
  6. Huang, G.; Yu, Y.; Lyu, M.; Sun, D.; Dewancker, B.; Gao, W. Impact of Physical Features on Visual Walkability Perception in Urban Commercial Streets by Using Street-View Images and Deep Learning. Buildings 2025, 15, 113. [Google Scholar] [CrossRef]
  7. Vieira, M.; Galvão, G.; Vieira, M.A.; Vestias, M.; Louro, P.; Vieira, P. Integrating Visible Light Communication and AI for Adaptive Traffic Management: A Focus on Reward Functions and Rerouting Coordination. Appl. Sci. 2024, 15, 116. [Google Scholar] [CrossRef]
  8. Ristić, B.; Bogdanović, V.; Stević, Ž. Urban evaluation of pedestrian crossings based on Start-Up Time using the MEREC-MARCOS Model. J. Urban Dev. Manag. 2024, 3, 34–42. [Google Scholar] [CrossRef]
  9. World Health Organization. Global Status Report on Road Safety 2023: Summary; World Health Organization: Geneva, Switzerland, 2023. [Google Scholar]
  10. Zou, T.; Chen, D.; Li, Q.; Wang, G.; Gu, C. A novel straw structure sandwich hood with regular deformation diffusion mode. Compos. Struct. 2024, 337, 118077. [Google Scholar] [CrossRef]
  11. Zou, T.; Shang, S.; Simms, C. Potential benefits of controlled vehicle braking to reduce pedestrian ground contact injuries. Accid. Anal. Prev. 2019, 129, 94–107. [Google Scholar] [CrossRef]
  12. Chen, Y.; Wang, Z.; Zhang, K. Multi-Modal Sensor Fusion for Robust Pedestrian Detection in Autonomous Driving: A Hybrid CNN-Transformer Approach. IEEE Trans. Veh. Technol. 2023, 72, 5123–5136. [Google Scholar]
  13. Chi, C.; Zhang, S.; Xing, J.; Lei, Z.; Li, S.Z.; Zou, X. Pedhunter: Occlusion robust pedestrian detector in crowded scenes. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 10639–10646. [Google Scholar]
  14. Han, X.; Xu, G.; Zhou, Y.; Yang, X.; Li, J.; Zhang, T. Physical backdoor attacks to lane detection systems in autonomous driving. In Proceedings of the 30th ACM International Conference on Multimedia, Lisboa, Portugal, 10–14 October 2022; pp. 2957–2968. [Google Scholar]
  15. Wei, H.; Tang, H.; Jia, X.; Wang, Z.; Yu, H.; Li, Z.; Satoh, S.; Van Gool, L.; Wang, Z. Physical adversarial attack meets computer vision: A decade survey. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 9797–9817. [Google Scholar] [CrossRef]
  16. Costa, J.C.; Roxo, T.; Proença, H.; Inácio, P.R. How deep learning sees the world: A survey on adversarial attacks & defenses. IEEE Access 2024, 12, 61113–61136. [Google Scholar]
  17. Zhang, S.; Pan, Y.; Liu, Q.; Yan, Z.; Choo, K.K.R.; Wang, G. Backdoor attacks and defenses targeting multi-domain ai models: A comprehensive review. ACM Comput. Surv. 2024, 57, 1–35. [Google Scholar] [CrossRef]
  18. Luo, C.; Li, Y.; Jiang, Y.; Xia, S.T. Untargeted backdoor attack against object detection. In Proceedings of the ICASSP 2023–2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar]
  19. Li, Y.; Jiang, Y.; Li, Z.; Xia, S.T. Backdoor learning: A survey. IEEE Trans. Neural Netw. Learn. Syst. 2022, 35, 5–22. [Google Scholar] [CrossRef]
  20. Saha, A.; Subramanya, A.; Pirsiavash, H. Hidden trigger backdoor attacks. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 11957–11965. [Google Scholar]
  21. Wang, B.; Yao, Y.; Shan, S.; Li, H.; Viswanath, B.; Zheng, H.; Zhao, B.Y. Neural cleanse: Identifying and mitigating backdoor attacks in neural networks. In Proceedings of the 2019 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 19–23 May 2019; pp. 707–723. [Google Scholar]
  22. Liu, Y.; Ma, X.; Bailey, J.; Lu, F. Reflection backdoor: A natural backdoor attack on deep neural networks. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020, Part X; Springer: Cham, Switzerland, 2020; pp. 182–199. [Google Scholar]
  23. Saha, A.; Tejankar, A.; Koohpayegani, S.A.; Pirsiavash, H. Backdoor attacks on self-supervised learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 13337–13346. [Google Scholar]
  24. Turner, A.; Tsipras, D.; Madry, A. Label-consistent backdoor attacks. arXiv 2019, arXiv:1912.02771. [Google Scholar]
  25. Zhao, Z.; Chen, X.; Xuan, Y.; Dong, Y.; Wang, D.; Liang, K. Defeat: Deep hidden feature backdoor attacks by imperceptible perturbation and latent representation constraints. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 15213–15222. [Google Scholar]
  26. Li, Y.; Li, Y.; Wu, B.; Li, L.; He, R.; Lyu, S. Invisible backdoor attack with sample-specific triggers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 16463–16472. [Google Scholar]
  27. Xue, M.; He, C.; Wu, Y.; Sun, S.; Zhang, Y.; Wang, J.; Liu, W. PTB: Robust physical backdoor attacks against deep neural networks in real world. Comput. Secur. 2022, 118, 102726. [Google Scholar] [CrossRef]
  28. Liang, J.; Liang, S.; Liu, A.; Jia, X.; Kuang, J.; Cao, X. Poisoned forgery face: Towards backdoor attacks on face forgery detection. arXiv 2024, arXiv:2402.11473. [Google Scholar]
  29. Chen, X.; Liu, C.; Li, B.; Lu, K.; Song, D. Targeted backdoor attacks on deep learning systems using data poisoning. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 1859–1872. [Google Scholar]
  30. Doan, K.D.; Lao, Y.; Li, P. Marksman backdoor: Backdoor attacks with arbitrary target class. Adv. Neural Inf. Process. Syst. 2022, 35, 38260–38273. [Google Scholar]
  31. Zhang, S.; Benenson, R.; Schiele, B. Citypersons: A diverse dataset for pedestrian detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3213–3221. [Google Scholar]
  32. Liu, Y.; Ma, S.; Aafer, Y.; Lee, W.C.; Zhai, J.; Wang, W.; Zhang, X. Trojaning attack on neural networks. In Proceedings of the 25th Annual Network and Distributed System Security Symposium (NDSS 2018), San Diego, CA, USA, 18–21 February 2018. [Google Scholar]
  33. Gu, T.; Liu, K.; Dolan-Gavitt, B.; Garg, S. Badnets: Evaluating backdooring attacks on deep neural networks. IEEE Access 2019, 7, 47230–47244. [Google Scholar] [CrossRef]
  34. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  35. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. In Proceedings of the NIPS’15: Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; Volume 28. [Google Scholar]
  36. Cai, Z.; Vasconcelos, N. Cascade R-CNN: High quality object detection and instance segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1483–1498. [Google Scholar] [CrossRef]
  37. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  38. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  39. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single shot multibox detector. In Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016, Part I; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  40. Liu, W.; Liao, S.; Hu, W.; Liang, X.; Chen, X. Real-time pedestrian detection for traffic monitoring systems. IEEE Trans. Intell. Transp. Syst. 2018, 19, 2675–2684. [Google Scholar]
  41. Li, Y.; Niu, W.; Tian, Y.; Chen, T.; Xie, Z.; Wu, Y.; Xiang, Y.; Tong, E.; Baker, T.; Liu, J. Multiagent reinforcement learning-based signal planning for resisting congestion attack in green transportation. IEEE Trans. Green Commun. Netw. 2022, 6, 1448–1458. [Google Scholar] [CrossRef]
  42. Zhang, Y.; Wang, C.; Wang, X.; Zeng, W. CrowdPed: Crowd-aware pedestrian detection in surveillance. In Proceedings of the IEEE International Conference on Advanced Video and Signal-Based Surveillance, Taipei, Taiwan, 18–21 September 2019; pp. 1–6. [Google Scholar]
  43. Wu, Y.; Xiang, Y.; Baker, T.; Tong, E.; Zhu, Y.; Cui, X.; Zhang, Z.; Han, Z.; Liu, J.; Niu, W. Collaborative Attack Sequence Generation Model Based on Multiagent Reinforcement Learning for Intelligent Traffic Signal System. Int. J. Intell. Syst. 2024, 2024, 4734030. [Google Scholar] [CrossRef]
  44. Li, H.; Yang, B.; Liu, M. LIDAR-camera fusion for pedestrian detection in autonomous driving. IEEE Trans. Intell. Veh. 2022, 7, 301–312. [Google Scholar]
  45. Chen, Y.; Wu, Y.; Cui, X.; Li, Q.; Liu, J.; Niu, W. Reflective Adversarial Attacks against Pedestrian Detection Systems for Vehicles at Night. Symmetry 2024, 16, 1262. [Google Scholar] [CrossRef]
  46. Gu, T.; Dolan-Gavitt, B.; Garg, S. Badnets: Identifying vulnerabilities in the machine learning model supply chain. arXiv 2017, arXiv:1708.06733. [Google Scholar]
  47. Liu, Y.; Wang, Y.; Zhang, Y. Fine-tuning for Backdoor Attack Mitigation. IEEE Trans. Inf. Forensics Secur. 2020, 15, 1234–1245. [Google Scholar]
  48. Salem, X.C.A.; Zhang, M.B.S.M.Y. Badnl: Backdoor attacks against nlp models. In Proceedings of the ICML 2021 Workshop on Adversarial Machine Learning, Online, 18–24 July 2021. [Google Scholar]
  49. Sun, L. Natural backdoor attack on text data. arXiv 2020, arXiv:2006.16176. [Google Scholar]
  50. Zeng, R.; Chen, X.; Pu, Y.; Zhang, X.; Du, T.; Ji, S. CLIBE: Detecting Dynamic Backdoors in Transformer-based NLP Models. arXiv 2024, arXiv:2409.01193. [Google Scholar]
  51. Shi, C.; Ji, S.; Pan, X.; Zhang, X.; Zhang, M.; Yang, M.; Zhou, J.; Yin, J.; Wang, T. Towards practical backdoor attacks on federated learning systems. IEEE Trans. Dependable Secur. Comput. 2024, 21, 5431–5447. [Google Scholar] [CrossRef]
  52. Wu, Y.; Li, Q.; Xiang, Y.; Zheng, J.; Wu, X.; Han, Z.; Liu, J.; Niu, W. Nightfall Deception: A Novel Backdoor Attack on Traffic Sign Recognition Models via Low-Light Data Manipulation. In International Conference on Advanced Data Mining and Applications, Sydney, NSW, Australia, 3–5 December 2024; Springer: Singapore, 2024; pp. 433–445. [Google Scholar]
  53. Zhao, S.; Ma, X.; Zheng, X.; Bailey, J.; Chen, J.; Jiang, Y.G. Clean-label backdoor attacks on video recognition models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 14443–14452. [Google Scholar]
  54. Rakin, A.S.; He, Z.; Fan, D. Tbt: Targeted neural network attack with bit trojan. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13198–13207. [Google Scholar]
  55. Li, Y.; Hua, J.; Wang, H.; Chen, C.; Liu, Y. Deeppayload: Black-box backdoor attack on deep learning models through neural payload injection. In Proceedings of the 2021 IEEE/ACM 43rd International Conference on Software Engineering (ICSE), Madrid, Spain, 22–30 May 2021; pp. 263–274. [Google Scholar]
  56. Wenger, E.; Passananti, J.; Bhagoji, A.N.; Yao, Y.; Zheng, H.; Zhao, B.Y. Backdoor attacks against deep learning systems in the physical world. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 6206–6215. [Google Scholar]
  57. Wu, Y.; Gu, Y.; Chen, Y.; Cui, X.; Li, Q.; Xiang, Y.; Tong, E.; Li, J.; Han, Z.; Liu, J. Camouflage Backdoor Attack against Pedestrian Detection. Appl. Sci. 2023, 13, 12752. [Google Scholar] [CrossRef]
  58. Jiang, L.; Ma, X.; Chen, S.; Bailey, J.; Jiang, Y.G. Black-box adversarial attacks on video recognition models. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 864–872. [Google Scholar]
  59. Zhong, Z.; Zheng, L.; Kang, G.; Li, S.; Yang, Y. Random erasing data augmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 13001–13008. [Google Scholar]
  60. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
  61. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014, Part V; Springer: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
  62. Chan, S.H.; Dong, Y.; Zhu, J.; Zhang, X.; Zhou, J. Baddet: Backdoor attacks on object detection. In European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Cham, Switzerland, 2022; pp. 396–412. [Google Scholar]
  63. Wei, H.; Yu, H.; Zhang, K.; Wang, Z.; Zhu, J.; Wang, Z. Moiré backdoor attack (MBA): A novel trigger for pedestrian detectors in the physical world. In Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 8828–8838. [Google Scholar]
  64. Chen, X.; Li, H.; Zhao, Q. Test-time Noise Injection for Robustness against Backdoor Attacks. Pattern Recognit. 2021, 110, 107623. [Google Scholar]
Figure 1. The main workflow of our deceptive threats. In the first step, malicious data vendors poison the data by adding occlusions as triggers to the original training images, thus generating poisoned training datasets. Secondly, benign and poisoned images are combined and used for training to obtain a pre-trained pedestrian detection model. Finally, attackers utilize pedestrian images with trigger (occlusions) to evade detection.
Figure 1. The main workflow of our deceptive threats. In the first step, malicious data vendors poison the data by adding occlusions as triggers to the original training images, thus generating poisoned training datasets. Secondly, benign and poisoned images are combined and used for training to obtain a pre-trained pedestrian detection model. Finally, attackers utilize pedestrian images with trigger (occlusions) to evade detection.
Sensors 25 04203 g001
Figure 2. Three trigger embedding schemes for occluding.
Figure 2. Three trigger embedding schemes for occluding.
Sensors 25 04203 g002
Figure 3. Examples of our attack in the digital domain.
Figure 3. Examples of our attack in the digital domain.
Sensors 25 04203 g003
Figure 4. Examples of our attack under varying physical conditions. (ad) represent indoor scene at different distances, while (eh) represent outdoor scene at different distances.
Figure 4. Examples of our attack under varying physical conditions. (ad) represent indoor scene at different distances, while (eh) represent outdoor scene at different distances.
Sensors 25 04203 g004
Figure 5. Examples of our attack under various rotation angles.
Figure 5. Examples of our attack under various rotation angles.
Sensors 25 04203 g005
Figure 6. Visual comparison of triggers in comparative methods and ours. Our occlusion-based backdoor (h) employs a trigger pattern based on natural objects, without relying on obvious artificial patches (a,b), stickers (ce), visual alert mixture (f), and suspicious patterns (g). Therefore, our occlusion-based backdoor attack is more visually inconspicuous. Triggers in the figure are as follows: (a) yellow square; (b) checkerboard; (c) white sticker; (d) black sticker on the forehead; (e) five-pointed star sticker; (f) reflective mixture; (g) moiré pattern.
Figure 6. Visual comparison of triggers in comparative methods and ours. Our occlusion-based backdoor (h) employs a trigger pattern based on natural objects, without relying on obvious artificial patches (a,b), stickers (ce), visual alert mixture (f), and suspicious patterns (g). Therefore, our occlusion-based backdoor attack is more visually inconspicuous. Triggers in the figure are as follows: (a) yellow square; (b) checkerboard; (c) white sticker; (d) black sticker on the forehead; (e) five-pointed star sticker; (f) reflective mixture; (g) moiré pattern.
Sensors 25 04203 g006
Figure 7. Four trigger patterns used in our evaluation.
Figure 7. Four trigger patterns used in our evaluation.
Sensors 25 04203 g007
Figure 8. Impact of the occlusion ratio r. We present the variation curves of BAP (%), PAP (%), and ASR (%) under different r values for poisoned RetinaNet on the CityPersons dataset.
Figure 8. Impact of the occlusion ratio r. We present the variation curves of BAP (%), PAP (%), and ASR (%) under different r values for poisoned RetinaNet on the CityPersons dataset.
Sensors 25 04203 g008
Figure 9. Impact of training epochs. We present the evolution of BAP (%), PAP (%), and ASR (%) across epochs for poisoned RetinaNet and Faster R-CNN on the KITTI dataset.
Figure 9. Impact of training epochs. We present the evolution of BAP (%), PAP (%), and ASR (%) across epochs for poisoned RetinaNet and Faster R-CNN on the KITTI dataset.
Sensors 25 04203 g009
Table 1. Comparison of BAP (%), PAP (%), and ASR (%) for poisoned models on the KITTI and CityPersons datasets.
Table 1. Comparison of BAP (%), PAP (%), and ASR (%) for poisoned models on the KITTI and CityPersons datasets.
DatasetModel →
Metric ↓
Faster R-CNNRetinaNetAverage
Image-Level Object-Level Image + Object Image-Level Object-Level Image + Object Image-Level Object-Level Image + Object
KITTIBAP ↑41.442.242.440.634.138.341.038.140.3
PAP ↓31.9 22.9 13.0 69.2 16.6 60.8 31.6 22.2 5.8 83.0 6.1 84.1 31.7 22.7 9.4 75.3 11.3 72.0
ASR ↑35.666.758.636.483.584.436.075.171.5
CityPersonsBAP ↑26.826.626.623.821.015.925.323.821.2
PAP ↓19.4 27.6 2.1 92.1 3.0 88.7 17.9 24.8 0.1 99.5 1.7 89.3 18.6 26.5 1.1 95.4 2.3 89.2
ASR ↑64.494.893.565.999.496.565.197.195.0
Note: Model → indicates the evaluated models (right-side columns); Metric ↓ indicates the evaluation metrics (lower rows); BAP ↑ and ASR ↑ indicate higher is better; PAP ↓ indicate lower is better; Bold values indicate the best performance of ASR; indicates that the PAP of the poisoned model is lower than the BAP.
Table 2. Comparison of BAP (%) between benign and poisoned models on the KITTI and CityPersons datasets.
Table 2. Comparison of BAP (%) between benign and poisoned models on the KITTI and CityPersons datasets.
DatasetMethod ↓, Model →Faster R-CNNRetinaNetAverage
KITTIBenign42.541.442.0
Image-level41.4 2.5 40.6 1.9 41.0 2.3
Object-level42.2 0.7 34.1 17.6 38.2 9.0
Image + Object42.4 0.2 38.3 7.5 40.4 3.8
CityPersonsBenign26.823.625.2
Image-level26.823.8 0.8 25.3 0.4
Object-level26.6 0.7 21.0 11.0 23.8 5.6
Image + Object26.6 0.7 15.9 32.6 21.3 15.5
Note: Method ↓ indicates the methods being compared (lower rows); Model → indicates the evaluated models (right-side columns); indicates that the BAP of the poisoned model is lower than benign model; indicates that the BAP of poisoned model is higher than benign model.
Table 3. Impact of four trigger patterns. We compare BAP (%), PAP (%), and ASR (%) between poisoned Fast R-CNN trained with different trigger patterns and benign models on the KITTI dataset.
Table 3. Impact of four trigger patterns. We compare BAP (%), PAP (%), and ASR (%) between poisoned Fast R-CNN trained with different trigger patterns and benign models on the KITTI dataset.
Trigger PatternDetectors ↓, Metric  →BAP ↑PAP ↓ASR ↑
(a) BackpackBenign42.532.6
Poisoned42.213.066.7
(b) BalloonBenign42.532.3
Poisoned41.95.785.0
(c) Paper bagBenign42.536.8
Poisoned42.016.757.5
(d) SuitcaseBenign42.536.9
Poisoned42.717.156.0
Note: Detectors ↓ indicates the evaluated models (lower rows); Metric → indicates the evaluation metrics (right-side columns); BAP ↑ and ASR ↑ indicate higher is better; PAP ↓ indicates lower is better.
Table 4. Impact of the poisoning rate p. We compare BAP (%), PAP (%), and ASR (%) under different p values for our attack using Faster R-CNN and RetinaNet on the KITTI and CityPersons datasets.
Table 4. Impact of the poisoning rate p. We compare BAP (%), PAP (%), and ASR (%) under different p values for our attack using Faster R-CNN and RetinaNet on the KITTI and CityPersons datasets.
DatasetModelMetricPoisoning Rate
5% 10% 20% 40% Avg
KITTIFaster R-CNNASR ↑66.778.889.695.882.7
BAP ↑42.242.240.738.240.8
PAP ↓13.08.03.91.46.6
RetinaNetASR ↑83.584.291.393.788.2
BAP ↑34.130.729.123.029.2
PAP ↓5.85.22.71.93.9
CitypersonsFaster R-CNNASR ↑94.897.998.899.797.8
BAP ↑26.626.626.225.326.2
PAP ↓2.11.10.80.21.01
RetinaNetASR ↑99.498.999.499.999.4
BAP ↑21.014.814.414.316.1
PAP ↓0.10.30.20.10.2
Note: BAP ↑ and ASR ↑ indicate higher is better; PAP ↓ indicates lower is better.
Table 5. Comparison of BAP (%), PAP (%), and ASR (%) of defense methods for different poisoned models on the KITTI dataset.
Table 5. Comparison of BAP (%), PAP (%), and ASR (%) of defense methods for different poisoned models on the KITTI dataset.
Defense ↓, Model →Faster R-CNNRetinaNetAverage
ASR ↑BAP ↑PAP ↓ASR ↑BAP ↑PAP ↓ASR ↑BAP ↑PAP ↓
W/O66.742.213.083.534.15.875.138.29.4
Fine-tuning62.231.613.633.336.130.247.8 36.6 33.921.9
Test-time noise injection60.322.616.076.216.98.468.3 9.1 19.812.2
Note: Defense ↓ indicates the evaluated defense methods (lower rows); Model→ indicates the evaluated models (right-side columns); BAP↑ and ASR↑ indicate higher is better; PAP↓ indicates lower is better; Bold values indicate the best performance of defense resistance; indicates that the ASR of the defense method is lower than no defense.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Q.; Wu, Y.; Li, Q.; Cui, X.; Chen, Y.; Chang, X.; Liu, J.; Niu, W. Natural Occlusion-Based Backdoor Attacks: A Novel Approach to Compromising Pedestrian Detectors. Sensors 2025, 25, 4203. https://doi.org/10.3390/s25134203

AMA Style

Li Q, Wu Y, Li Q, Cui X, Chen Y, Chang X, Liu J, Niu W. Natural Occlusion-Based Backdoor Attacks: A Novel Approach to Compromising Pedestrian Detectors. Sensors. 2025; 25(13):4203. https://doi.org/10.3390/s25134203

Chicago/Turabian Style

Li, Qiong, Yalun Wu, Qihuan Li, Xiaoshu Cui, Yuanwan Chen, Xiaolin Chang, Jiqiang Liu, and Wenjia Niu. 2025. "Natural Occlusion-Based Backdoor Attacks: A Novel Approach to Compromising Pedestrian Detectors" Sensors 25, no. 13: 4203. https://doi.org/10.3390/s25134203

APA Style

Li, Q., Wu, Y., Li, Q., Cui, X., Chen, Y., Chang, X., Liu, J., & Niu, W. (2025). Natural Occlusion-Based Backdoor Attacks: A Novel Approach to Compromising Pedestrian Detectors. Sensors, 25(13), 4203. https://doi.org/10.3390/s25134203

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop