Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (54)

Search Parameters:
Keywords = local perturbation attack

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 5677 KB  
Article
Robust Image Watermarking via Clustered Visual State-Space Modeling
by Bo Liu and Jianhua Ren
Appl. Sci. 2026, 16(9), 4166; https://doi.org/10.3390/app16094166 - 24 Apr 2026
Abstract
Most existing DNN-based image watermarking methods adopt an “encoder–noise–decoder” paradigm, where the watermark is typically replicated and expanded in a straightforward manner and then directly fused with image features, which limits robustness under complex distortions. Although Transformers improve fusion via attention mechanisms, their [...] Read more.
Most existing DNN-based image watermarking methods adopt an “encoder–noise–decoder” paradigm, where the watermark is typically replicated and expanded in a straightforward manner and then directly fused with image features, which limits robustness under complex distortions. Although Transformers improve fusion via attention mechanisms, their quadratic computational complexity makes high-resolution processing prohibitively expensive. To address these issues, we propose CCViM, a robust watermarking framework built on Vision Mamba, which leverages the linear-complexity property of state-space models (SSMs) to enable efficient global interactions. We design a Watermark Representation Learning Module (WRLM) that performs hierarchical feature extraction and structured expansion of the watermark through cascaded VSS blocks, yielding semantically rich and perturbation-resistant watermark representations. In addition, we introduce an Interwoven Fusion Enhancement Module (IFEM), which employs a CCS6 structure to treat the watermark as a dynamic guidance signal. By combining contextual clustering with the Mamba mechanism, IFEM deeply interweaves the watermark into host features at both local and global levels. Experiments on COCO, DIV2K, and ImageNet demonstrate that CCViM consistently improves imperceptibility, robustness, and efficiency to varying degrees, and remains stable and high quality under attacks such as JPEG compression, cropping, and Gaussian blur. Full article
(This article belongs to the Special Issue Advanced Pattern Recognition & Computer Vision, 2nd Edition)
Show Figures

Figure 1

20 pages, 1064 KB  
Article
Privacy-Preserving U-Shaped Split Federated Learning for Space–Air–Ground–Sea Integrated Networks
by Xin Sun, Tingting Yang and Xiufeng Zhang
Mathematics 2026, 14(8), 1357; https://doi.org/10.3390/math14081357 - 18 Apr 2026
Viewed by 113
Abstract
Federated learning enables privacy-preserving distributed intelligence but faces challenges in balancing computation, communication, and privacy in heterogeneous networks. To address these issues, this paper proposes a privacy-preserving U-shaped split federated learning (USFL) framework for space–air–ground–sea integrated networks. The proposed architecture combines split learning [...] Read more.
Federated learning enables privacy-preserving distributed intelligence but faces challenges in balancing computation, communication, and privacy in heterogeneous networks. To address these issues, this paper proposes a privacy-preserving U-shaped split federated learning (USFL) framework for space–air–ground–sea integrated networks. The proposed architecture combines split learning and federated learning in a U-shaped structure, ensuring that both raw data and labels remain localized at client devices. In addition, a differential privacy mechanism is introduced to perturb intermediate features during transmission, enhancing resistance to inference attacks. A mathematical framework is established to model the learning process under resource constraints, and the convergence behavior and privacy loss are theoretically analyzed. Experimental results on the SeaShips dataset demonstrate that the proposed method achieves competitive accuracy compared with centralized and existing distributed approaches, while reducing communication overhead and improving privacy protection. These results validate the effectiveness of the proposed framework for secure and efficient distributed learning in complex network environments. Full article
Show Figures

Figure 1

18 pages, 5351 KB  
Article
Dual-Factor Adaptive Robust Aggregation for Secure Federated Learning in IoT Networks
by Zuan Song, Wuzheng Tan, Hailong Wang, Guilong Zhang and Jian Weng
Future Internet 2026, 18(4), 201; https://doi.org/10.3390/fi18040201 - 10 Apr 2026
Viewed by 283
Abstract
Federated Learning (FL) has been widely adopted in privacy-sensitive and distributed environments. However, training stability becomes significantly challenged when differential privacy (DP) noise and Byzantine client behaviors coexist, as these heterogeneous perturbations jointly introduce time-varying distortions to model updates. Existing approaches typically address [...] Read more.
Federated Learning (FL) has been widely adopted in privacy-sensitive and distributed environments. However, training stability becomes significantly challenged when differential privacy (DP) noise and Byzantine client behaviors coexist, as these heterogeneous perturbations jointly introduce time-varying distortions to model updates. Existing approaches typically address privacy and robustness in isolation. Under DP constraints, noise injection increases gradient variance and obscures the distinction between benign and adversarial updates, causing many robust aggregation methods to misclassify normal clients or fail to detect malicious ones. As a result, their effectiveness degrades substantially in practical IoT environments where noise and attacks interact. In this work, we propose a dual-factor adaptive and robust aggregation framework (DARA) to improve the stability of FL under such combined disturbances. DARA adjusts the differential privacy noise scale by jointly considering local update magnitudes and training-round dynamics, aiming to mitigate noise-induced bias under a fixed privacy budget. Meanwhile, a direction-aware weighted aggregation scheme assigns continuous trust weights based on cosine similarity between updates, thereby suppressing the influence of potentially anomalous or adversarial clients. We conduct extensive experiments on multiple benchmark datasets to evaluate DARA under differential privacy constraints and Byzantine attack scenarios. The results indicate that DARA achieves favorable robustness and convergence behavior compared with representative aggregation baselines, while maintaining competitive model accuracy. Full article
(This article belongs to the Special Issue Federated Learning: Challenges, Methods, and Future Directions)
Show Figures

Figure 1

22 pages, 31045 KB  
Article
Robust and Stealthy White-Box Watermarking for Intellectual Property Protection of Remote Sensing Object Detection Models
by Lingjun Zou, Xin Xu, Weitong Chen, Qingqing Hong and Di Wu
Remote Sens. 2026, 18(7), 985; https://doi.org/10.3390/rs18070985 - 25 Mar 2026
Viewed by 355
Abstract
Remote sensing object detection (RSOD) models play an increasingly important role in modern remote sensing systems. However, during model delivery, sharing, and deployment, RSOD models face increasing risks of unauthorized redistribution, illegal replication, and intellectual property infringement. To mitigate these threats, this paper [...] Read more.
Remote sensing object detection (RSOD) models play an increasingly important role in modern remote sensing systems. However, during model delivery, sharing, and deployment, RSOD models face increasing risks of unauthorized redistribution, illegal replication, and intellectual property infringement. To mitigate these threats, this paper proposes a white-box watermarking framework for RSOD models that enables reliable copyright verification while preserving the performance of the primary detection task. Specifically, a gradient-based sensitivity analysis of the detection loss is first performed to adaptively identify model parameters that minimally affect detection performance, which are then selected as watermark carriers. Subsequently, a parameter-ranking-based watermark encoding scheme is developed, where watermark bits are embedded by enforcing relative ordering constraints between parameter pairs. To further improve robustness under practical deployment conditions, an attack-simulation-driven training strategy is introduced, in which common perturbations and watermark removal attacks are simulated during the embedding process. In addition, a stealthiness enhancement strategy based on statistical distribution constraints is designed to maintain consistency between the distribution of watermarked parameters and those of the original model, thereby reducing the risk of watermark exposure and localization. Extensive experiments across multiple RSOD datasets and detection architectures demonstrate that the proposed method achieves a high copyright verification success rate with negligible impact on detection accuracy and exhibits strong robustness and stealthiness against a variety of watermark removal attacks. Full article
Show Figures

Figure 1

18 pages, 20418 KB  
Article
Localized Query Attack Toward Transformer-Based Visible Object Detectors
by Yang Wang, Ang Li, Zhen Yang and Xunyun Liu
Sensors 2026, 26(6), 1987; https://doi.org/10.3390/s26061987 - 23 Mar 2026
Viewed by 274
Abstract
Transformer-based detectors have demonstrated exceptional accuracy in visible-object detection tasks. However, adversarial patches, specific types of adversarial examples, can disrupt these detectors by introducing unrestricted perturbations into specific image regions. Traditional methodologies focus on placing patches directly on objects and increasing attention scores [...] Read more.
Transformer-based detectors have demonstrated exceptional accuracy in visible-object detection tasks. However, adversarial patches, specific types of adversarial examples, can disrupt these detectors by introducing unrestricted perturbations into specific image regions. Traditional methodologies focus on placing patches directly on objects and increasing attention scores between the patch and all areas of the image to impair detector performance. Nevertheless, these approaches are suboptimal due to significant discrepancies between background and object features, which contradict optimization objectives. Moreover, they overlook the impact of cross-attention mechanisms on detection results. To address these limitations, we introduce a novel approach named Localized Query Attack (LQA), designed to interfere with both self-attention within the encoder and cross-attention in the decoder. Unlike conventional global interference methods, LQA targets object features specifically, enhancing self-attention interactions between the adversarial patch and foreground regions to redirect model focus toward the patch. In the context of decoder cross-attention, we compute the joint attention matrix connecting encoder outputs with object queries. By diminishing the influence of encoder outputs and residual components in this matrix, we amplify the relative importance of the adversarial patch, thereby intensifying the attack’s effectiveness. Our experiments show that LQA achieves an approximately 20% improvement in transfer attack performance compared to the second-best method across various transformer-based detectors. The practical efficacy of LQA is further substantiated through real-world scenario validations, underscoring its applicability. Full article
(This article belongs to the Section Electronic Sensors)
Show Figures

Figure 1

16 pages, 1275 KB  
Article
Differentially Private Federated Learning with Adaptive Clipping Thresholds
by Jianhua Liu, Yanglin Zeng, Zhongmei Wang, Weiqing Zhang and Yao Tong
Future Internet 2026, 18(3), 148; https://doi.org/10.3390/fi18030148 - 14 Mar 2026
Viewed by 407
Abstract
Under non-independent and identically distributed (Non-IID) conditions, significant variations exist in local model updates across clients and training phases during the collaborative modeling process of differential privacy federated learning (DP-FL). Fixed clipping thresholds and noise scales struggle to accommodate these diverse update differences, [...] Read more.
Under non-independent and identically distributed (Non-IID) conditions, significant variations exist in local model updates across clients and training phases during the collaborative modeling process of differential privacy federated learning (DP-FL). Fixed clipping thresholds and noise scales struggle to accommodate these diverse update differences, leading to mismatches between local update intensity and noise perturbations. This imbalance results in data privacy leaks and suboptimal model accuracy. To address this, we propose a differential privacy federated learning method based on adaptive clipping thresholds. During each communication round, the server adaptively estimates the global clipping threshold for that round using a quantile strategy based on the statistical distribution of client update norms. Simultaneously, clients adaptively adjust their noise scales according to the clipping threshold magnitude, enabling dynamic matching of clipping intensity and noise perturbation across training phases and clients. The novelty of this work lies in a quantile-driven, round-wise global clipping adaptation that synchronizes sensitivity bounding and noise calibration across heterogeneous clients, enabling improved privacy–utility behavior under a fixed privacy accountant. Using experimental results on the rail damage datasets, our proposed method slightly reduces the attacker’s MIA ROC-AUC by 0.0033 and 0.0080 compared with Fed-DPA and DP-FedAvg, respectively, indicating stronger privacy protection, while improving average accuracy by 1.55% and 3.35% and achieving faster, more stable convergence. We further validate its effectiveness on CIFAR-10 under non-IID partitions. Full article
Show Figures

Figure 1

19 pages, 2755 KB  
Article
CA-Adv: Curvature-Adaptive Weighted Adversarial 3D Point Cloud Generation Method for Remote Sensing Scenarios
by Yanwen Sun, Shijia Xiao, Weiquan Liu, Min Huang, Chaozhi Cheng, Shiwei Lin, Jinhe Su, Zongyue Wang and Guorong Cai
Remote Sens. 2026, 18(6), 882; https://doi.org/10.3390/rs18060882 - 13 Mar 2026
Viewed by 342
Abstract
Adversarial robustness in 3D point cloud recognition models is a critical concern in remote sensing applications, such as autonomous driving and infrastructure monitoring. Existing adversarial attack methods can compromise model performance; moreover, they often neglect the intrinsic geometric properties of point clouds, leading [...] Read more.
Adversarial robustness in 3D point cloud recognition models is a critical concern in remote sensing applications, such as autonomous driving and infrastructure monitoring. Existing adversarial attack methods can compromise model performance; moreover, they often neglect the intrinsic geometric properties of point clouds, leading to perceptually unnatural perturbations that limit their practicality for robustness evaluation in real-world scenarios. To address this, we propose CA-Adv, a novel curvature-adaptive weighted adversarial generation method for 3D point clouds. Our approach first employs Shapley values to assess regional sensitivity and identify salient regions. It then adaptively partitions these regions based on local curvature and assigns perturbation weights accordingly, concentrating the attack on geometrically sensitive areas while preserving overall structural consistency through explicit geometric constraints. Extensive experiments on real-world remote sensing data (KITTI) and synthetic benchmarks (ModelNet40, ShapeNet) demonstrate that CA-Adv achieves a high attack success rate with a minimal perturbation budget. The generated adversarial examples maintain superior visual naturalness and geometric fidelity. The method provides a practical tool for evaluating the robustness of 3D recognition models in applications such as autonomous driving, urban-scale LiDAR perception, and remote sensing point cloud analysis. Full article
Show Figures

Figure 1

18 pages, 1675 KB  
Article
Efficient Data Aggregation in Smart Grids: A Personalized Local Differential Privacy Scheme
by Haina Song, Jinhang Sun, Mengyao Wang, Nan Zhao, Fan Zhang and Hongzhang Liu
Sensors 2026, 26(5), 1710; https://doi.org/10.3390/s26051710 - 8 Mar 2026
Viewed by 352
Abstract
The rapid advancement of smart grids, while enhancing the efficiency of power systems, has also raised serious concerns regarding the privacy and security of end-users’ electricity consumption data. Traditional privacy protection methods struggle to meet users’ individualized privacy requirements and often lead to [...] Read more.
The rapid advancement of smart grids, while enhancing the efficiency of power systems, has also raised serious concerns regarding the privacy and security of end-users’ electricity consumption data. Traditional privacy protection methods struggle to meet users’ individualized privacy requirements and often lead to a significant decline in data aggregation accuracy. To address the core contradiction between personalized privacy protection and high-precision grid analytics, this paper proposes an efficient data aggregation scheme based on personalized local differential privacy (EDAS-PLDP) tailored for smart grids. The proposed scheme enables smart terminal users to autonomously select their privacy protection levels based on individual needs, thereby breaking the limitations of the traditional “one-size-fits-all” approach. To mitigate the accuracy loss caused by personalized perturbations, a mean square error-based weighted aggregation strategy is introduced at the gateway side. This strategy evaluates the data quality from groups with different privacy preferences and adjusts aggregation weights to optimize the estimation accuracy of the global mean electricity consumption. Extensive experimental results demonstrate that, compared to existing mainstream schemes, EDAS-PLDP achieves higher estimation accuracy under various distributions of privacy preferences, user scales, and data granularities, while exhibiting lower time consumption, making it suitable for resource-constrained smart grid environments. Furthermore, the scheme shows excellent robustness against false data injection attacks. In summary, EDAS-PLDP provides a balanced and efficient solution for reconciling personalized privacy protection with high-precision data utility in smart grids. Full article
(This article belongs to the Section Internet of Things)
Show Figures

Figure 1

24 pages, 11178 KB  
Article
FLAMA: Frame-Level Alignment Margin Attack for Scene Text and Automatic Speech Recognition
by Yikun Xu, Zhiheng Xu and Pengwen Dai
Electronics 2026, 15(5), 1064; https://doi.org/10.3390/electronics15051064 - 4 Mar 2026
Viewed by 394
Abstract
Scene text recognition (STR) and automatic speech recognition (ASR) translate visual or acoustic signals into linguistic sequences and underpin many modern perception systems. Although their front-ends and decoders differ (e.g., CTC-based, attention-based, or variants), both tasks ultimately rely on aligning input frames to [...] Read more.
Scene text recognition (STR) and automatic speech recognition (ASR) translate visual or acoustic signals into linguistic sequences and underpin many modern perception systems. Although their front-ends and decoders differ (e.g., CTC-based, attention-based, or variants), both tasks ultimately rely on aligning input frames to output tokens by deep learning techniques, which exposes a shared vulnerability to adversarial perturbations. Existing attacks commonly optimize global sequence-level objectives. As a result, decisive frames are treated implicitly, and optimization can become unnecessarily diffuse over long input sequences, hindering convergence and perceptual quality. To address the above issues, we propose FLAMA, a unified Frame-Level Alignment Margin Attack, which could be used for both STR and ASR models. FLAMA explicitly targets alignment by maximizing per frame (or per step) recognition margins. The design is decoder-agnostic and applies to both CTC-based and attention-based pipelines. It employs a recognition-score-aware Step/Halt gate that concentrates updates on the most critical frames, and a stabilization stage that suppresses late-iteration oscillations to improve optimization stability and perceptual control. Ablation analyses show that stabilization consistently enhances attack success and reduces distortion. We evaluate FLAMA on STR benchmarks (SVT, CUTE80, and IC13) with CRNN, STAR, and TRBA, and on the ASR benchmark (LibriSpeech) with a Wav2Vec 2.0 model. Across modalities and architectures, FLAMA achieves near-100% attack success while substantially reducing l2 distortion and improving perceptual metrics compared with FGSM/PGD baselines. These results highlight frame-level alignment as a shared weak point across visual and audio sequence recognizers and suggest localized margin objectives as a principled route to effective sequence attacks. Full article
Show Figures

Figure 1

24 pages, 1133 KB  
Article
Distributed Privacy-Preserving Fusion for Multi-UAV Target Localization via Free-Noise Masking
by Ke Ma, Guowei Pan and Jian Huang
Electronics 2026, 15(5), 1016; https://doi.org/10.3390/electronics15051016 - 28 Feb 2026
Viewed by 276
Abstract
Multi-UAV target localization relies on cooperative fusion of local, perception-derived geometric measurements over an edge network. While distributed fusion improves scalability and robustness compared with a centralized architecture, the iterative message exchanges may leak sensitive information to external eavesdroppers or honest-but-curious peers. This [...] Read more.
Multi-UAV target localization relies on cooperative fusion of local, perception-derived geometric measurements over an edge network. While distributed fusion improves scalability and robustness compared with a centralized architecture, the iterative message exchanges may leak sensitive information to external eavesdroppers or honest-but-curious peers. This paper proposes a privacy-preserving distributed fusion method for multi-UAV localization via free-noise masking. The key idea is a double-injection mechanism. Specifically, each UAV masks its transmitted iterate with a locally generated bounded noise vector, while injecting the same noise into its local update so that the perturbations cancel exactly in the network-average dynamics under doubly stochastic mixing. As a result, the proposed PPDO-FN scheme preserves the practical convergence and weighted least squares localization accuracy of non-private distributed gradient descent, without requiring heavy cryptography or a trusted server. We further introduce reconstruction-based privacy metrics under transcript attacks and quantify the privacy–accuracy tradeoff. Simulation results demonstrate (i) near-identical accuracy and consensus behavior to the non-private baseline, (ii) monotonic privacy improvement with increasing masking strength, and (iii) the necessity of double-injection canceling compared with a naive single-injection baseline. Finally, we provide an end-to-end case study to connect the image-level detection to the geometric localization and then to privacy-preserving distributed fusion, illustrating engineering viability for our proposed approach. Full article
Show Figures

Figure 1

20 pages, 6717 KB  
Article
Unraveling Patch Size Effects in Vision Transformers: Adversarial Robustness in Hyperspectral Image Classification
by Shashi Kiran Chandrappa, Sidike Paheding and Abel A. Reyes-Angulo
Remote Sens. 2026, 18(4), 656; https://doi.org/10.3390/rs18040656 - 21 Feb 2026
Viewed by 502
Abstract
Vision Transformers (ViTs) have demonstrated strong performance in hyperspectral image (HSI) classification; however, their robustness is highly sensitive to patch size. This study investigates the impact of spatial patch size on clean accuracy and adversarial robustness using a standard ViT and a Channel [...] Read more.
Vision Transformers (ViTs) have demonstrated strong performance in hyperspectral image (HSI) classification; however, their robustness is highly sensitive to patch size. This study investigates the impact of spatial patch size on clean accuracy and adversarial robustness using a standard ViT and a Channel Attention Fusion variant (ViT-CAF). Patch sizes from 1 × 1 to 19 × 19 are evaluated across four benchmark datasets under FGSM, BIM, CW, PGD, and RFGSM attacks. Descriptive results show that smaller patches, particularly 1 × 1 and 3 × 3, generally yield higher adversarial accuracy, while larger patches amplify localized perturbations and degrade robustness. Parameter analysis indicates that patch-size-dependent variations arise mainly from the embedding layer, with the Transformer backbone remaining fixed, confirming that robustness differences are driven primarily by spatial context rather than model capacity. These findings reveal a trade-off between spatial granularity and adversarial resilience and provide guidance for patch size selection in ViT-based HSI applications. Full article
Show Figures

Figure 1

29 pages, 5664 KB  
Article
Adversarially Robust and Explainable Insulator Defect Detection for Smart Grid Infrastructure
by Mubarak Alanazi
Energies 2026, 19(4), 1013; https://doi.org/10.3390/en19041013 - 14 Feb 2026
Viewed by 369
Abstract
Automated insulator inspection systems face critical challenges from small object sizes, complex backgrounds, and vulnerability to adversarial attacks, a security concern largely unaddressed in safety-critical power infrastructure. We introduce Faster-YOLOv12n, integrating a FasterNet backbone with SGC2f attention modules and Wise-ShapeIoU loss for enhanced [...] Read more.
Automated insulator inspection systems face critical challenges from small object sizes, complex backgrounds, and vulnerability to adversarial attacks, a security concern largely unaddressed in safety-critical power infrastructure. We introduce Faster-YOLOv12n, integrating a FasterNet backbone with SGC2f attention modules and Wise-ShapeIoU loss for enhanced small defect localization. Our architecture achieves 98.9% mAP@0.5 on the CPLID, improving baseline YOLOv12n by 1.3% in precision (97.8% vs. 96.5%), 4.7% in recall (95.1% vs. 90.4%), and 1.8% in mAP@0.5. Through differential data augmentation, we expand training samples from 678 to 3900 images, achieving balanced class distribution and robust generalization across fog, adverse weather, and complex transmission line backgrounds. Comparative evaluation demonstrates superior performance over RT-DETR, Faster R-CNN, YOLOv7, YOLOv8, and YOLOv9, with per-class analysis revealing 99.8% AP@0.5 for defect detection. We provide the first comprehensive adversarial robustness evaluation for insulator defect detection, systematically assessing FGSM, PGD, and C&W attacks across perturbation budgets. Through adversarial training with mixed-batch strategies, our robust model maintains 93.2% mAP@0.5 under the strongest FGSM attacks (ϵ = 48/255), 94.5% under PGD attacks, and 95.1% under C&W attacks (τ = 3.0) while preserving 98.9% clean accuracy, demonstrating no trade-off between accuracy and robustness. Grad-CAM visualizations demonstrate that attacks disrupt confidence calibration while preserving spatial attention on defect regions, providing interpretable insights into model decision-making under adversarial conditions and validating learned feature representations for safety-critical smart grid monitoring applications. Full article
Show Figures

Figure 1

21 pages, 3921 KB  
Article
Adversarial Example Generation Method Based on Wavelet Transform
by Meng Bi, Xiaoguo Liang, Baiyu Wang, Longxin Liu, Xin Yin and Jiafeng Liu
Information 2026, 17(2), 182; https://doi.org/10.3390/info17020182 - 10 Feb 2026
Viewed by 495
Abstract
Adversarial examples are crucial tools for assessing the robustness of deep neural networks (DNNs) and revealing potential security vulnerabilities. Adversarial example generation methods based on Generative Adversarial Networks (GANs) have made significant progress in generating image adversarial examples, but still suffer from insufficient [...] Read more.
Adversarial examples are crucial tools for assessing the robustness of deep neural networks (DNNs) and revealing potential security vulnerabilities. Adversarial example generation methods based on Generative Adversarial Networks (GANs) have made significant progress in generating image adversarial examples, but still suffer from insufficient sparsity and transferability. To address these issues, this study proposes a novel semi-white-box untargeted adversarial example generation method named Wavelet-AdvGAN, with an explicit threat model defined as follows. Specifically, the attack is strictly untargeted without predefined target categories, aiming solely to mislead DNNs into classifying adversarial examples into any category other than the original label. It adopts a semi-white-box setting where attackers are denied access to the target model’s private information. Regarding the generator’s information dependence, the training phase only utilizes public resources (i.e., the target model’s public architecture and CIFAR-10 public training data), while the test phase generates adversarial examples through one-step feedforward of clean images without interacting with the target model. The method incorporates a Frequency Sub-band Difference (FSD) module and a Wavelet Transform Local Feature (WTLF) extraction module, evaluating the differences between original and adversarial examples from the frequency domain perspective. This approach constrains the magnitude of perturbations, reinforces feature regions, and further enhances the attack effectiveness, thereby improving the sparsity and transferability of adversarial examples. Experimental results demonstrate that the Wavelet-AdvGAN method achieves an average increase of 1.26% in attack success rates under two defense strategies—data augmentation and adversarial training. Additionally, the adversarial transferability improves by an average of 2.7%. Moreover, the proposed method exhibits a lower l0 norm, indicating better perturbation sparsity. Consequently, it effectively evaluates the robustness of deep neural networks. Full article
Show Figures

Figure 1

21 pages, 4781 KB  
Article
A Spatially Distributed Perturbation Strategy with Smoothed Gradient Sign Method for Adversarial Analysis of Image Classification Systems
by Yanwei Xu, Jun Li, Dajun Chang and Yuanfang Dong
Entropy 2026, 28(2), 193; https://doi.org/10.3390/e28020193 - 9 Feb 2026
Viewed by 497
Abstract
As deep learning models are increasingly embedded as critical components within complex socio-technical systems, understanding and evaluating their systemic robustness against adversarial perturbations has become a fundamental concern for system safety and reliability. Deep neural networks (DNNs) are highly effective in visual recognition [...] Read more.
As deep learning models are increasingly embedded as critical components within complex socio-technical systems, understanding and evaluating their systemic robustness against adversarial perturbations has become a fundamental concern for system safety and reliability. Deep neural networks (DNNs) are highly effective in visual recognition tasks but remain vulnerable to adversarial perturbations, which can compromise their reliability in safety-critical applications. Existing attack methods often distribute perturbations uniformly across the input, ignoring the spatial heterogeneity of model sensitivity. In this work, we propose the Spatially Distributed Perturbation Strategy with Smoothed Gradient Sign Method (SD-SGSM), a adversarial attack framework that exploits decision-dependent regions to maximize attack effectiveness while minimizing perceptual distortion. SD-SGSM integrates three key components: (i) decision-dependent domain identification to localize critical features using a deterministic zero-out operator; (ii) spatially adaptive perturbation allocation to concentrate attack energy on sensitive regions while constraining background disturbance; and (iii) gradient smoothing via a hyperbolic tangent transformation to enable fine-grained and continuous perturbation updates. Extensive experiments on CIFAR-10 demonstrate that SD-SGSM achieves near-perfect attack success rates (ASR 99.9%) while substantially reducing 2 distortion and preserving high structural similarity (SSIM 0.947), outperforming both single-step and momentum-based iterative attacks. Ablation studies further confirm that spatial distribution and gradient smoothing act as complementary mechanisms, jointly enhancing attack potency and visual fidelity. These findings underscore the importance of spatially aware, decision-dependent adversarial strategies for system-level robustness assessment and the secure design of AI-enabled systems. Full article
Show Figures

Figure 1

26 pages, 1315 KB  
Article
SFD-ADNet: Spatial–Frequency Dual-Domain Adaptive Deformation for Point Cloud Data Augmentation
by Jiacheng Bao, Lingjun Kong and Wenju Wang
J. Imaging 2026, 12(2), 58; https://doi.org/10.3390/jimaging12020058 - 26 Jan 2026
Viewed by 547
Abstract
Existing 3D point cloud enhancement methods typically rely on artificially designed geometric transformations or local blending strategies, which are prone to introducing illogical deformations, struggle to preserve global structure, and exhibit insufficient adaptability to diverse degradation patterns. To address these limitations, this paper [...] Read more.
Existing 3D point cloud enhancement methods typically rely on artificially designed geometric transformations or local blending strategies, which are prone to introducing illogical deformations, struggle to preserve global structure, and exhibit insufficient adaptability to diverse degradation patterns. To address these limitations, this paper proposes SFD-ADNet—an adaptive deformation framework based on a dual spatial–frequency domain. It achieves 3D point cloud augmentation by explicitly learning deformation parameters rather than applying predefined perturbations. By jointly modeling spatial structural dependencies and spectral features, SFD-ADNet generates augmented samples that are both structurally aware and task-relevant. In the spatial domain, a hierarchical sequence encoder coupled with a bidirectional Mamba-based deformation predictor captures long-range geometric dependencies and local structural variations, enabling adaptive position-aware deformation control. In the frequency domain, a multi-scale dual-channel mechanism based on adaptive Chebyshev polynomials separates low-frequency structural components from high-frequency details, allowing the model to suppress noise-sensitive distortions while preserving the global geometric skeleton. The two deformation predictions dynamically fuse to balance structural fidelity and sample diversity. Extensive experiments conducted on ModelNet40-C and ScanObjectNN-C involved synthetic CAD models and real-world scanned point clouds under diverse perturbation conditions. SFD-ADNet, as a universal augmentation module, reduces the mCE metrics of PointNet++ and different backbone networks by over 20%. Experiments demonstrate that SFD-ADNet achieves state-of-the-art robustness while preserving critical geometric structures. Furthermore, models enhanced by SFD-ADNet demonstrate consistently improved robustness against diverse point cloud attacks, validating the efficacy of adaptive space-frequency deformation in robust point cloud learning. Full article
(This article belongs to the Special Issue 3D Image Processing: Progress and Challenges)
Show Figures

Figure 1

Back to TopTop