Next Article in Journal
Real-Time AI-Based Data Prioritization for MODBUS TCP Communication in IoT-Enabled LVDC Energy Systems
Previous Article in Journal
Transfer-Efficient Power Allocation for Downlink SWIPT in Massive MIMO Systems
Previous Article in Special Issue
ScaL2Chain: Towards a Scalable Protocol for Multi-Chain Decentralized Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Differential Privacy for Satellite Image Recognition with Convergence-Guaranteed Optimization

College of Mechatronic Engineering, North University of China, No. 3 Xueyuan Road, Taiyuan 030051, China
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(18), 3680; https://doi.org/10.3390/electronics14183680
Submission received: 8 August 2025 / Revised: 9 September 2025 / Accepted: 16 September 2025 / Published: 17 September 2025

Abstract

Differential privacy (DP) has become a cornerstone for privacy-preserving machine learning, yet its application to high-resolution satellite imagery remains underexplored. Existing DP algorithms, such as DP-SGD, often rely on static noise levels and global clipping thresholds, which lead to slow convergence and poor utility in deep neural networks. In this paper, we propose ADP-SIR, an Adaptive Differential Privacy framework for Satellite Image Recognition with provable convergence guarantees. ADP-SIR introduces two novel components: Convergence-Guided Noise Scaling (CGNS), which dynamically adjusts the noise multiplier based on training stability, and Layerwise Sensitivity Profiling (LSP), which enables fine-grained clipping at the layer level. We provide theoretical analysis showing that ADP-SIR achieves good convergence in non-convex settings under Rényi differential privacy. Empirically, we evaluate ADP-SIR on EuroSAT and RESISC45, demonstrating significant improvements over DP-SGD and AdaClip-DP in terms of accuracy, convergence speed, and per-class fairness. Our framework bridges the gap between practical performance and rigorous privacy for remote sensing applications.

1. Introduction

Satellite imagery has become a vital data source for numerous societal and industrial applications, ranging from environmental monitoring and land use classification to agricultural assessment, military surveillance, and disaster response. As the spatial and temporal resolutions of satellites continue to improve, the quantity and sensitivity of geospatial data have grown exponentially. Consequently, modern machine learning models trained on such data are becoming increasingly powerful but also increasingly privacy-sensitive. In particular, when satellite imagery contains information related to private properties, critical infrastructure, or military zones, it becomes imperative to ensure that data-driven models cannot leak individual-level or region-specific information.
Differential Privacy (DP) [1] has emerged as a mathematically rigorous standard for protecting individual data records in statistical queries and machine learning. The fundamental guarantee of DP is that the inclusion or removal of a single data point does not significantly change the output of a computation. This property has made DP the de facto privacy-preserving paradigm in several domains, including healthcare, finance, and web analytics. However, its direct application to deep learning models trained on high-dimensional satellite images remains a major challenge. The typical approach of injecting random noise into gradients during training, such as in Differentially Private Stochastic Gradient Descent (DP-SGD) [2], often leads to significant degradation in model accuracy, particularly under tight privacy budgets (i.e., small ε ).
The root of this problem lies in the nature of satellite imagery itself. These images exhibit the following:
  • High input dimensionality: Images are large (e.g., 512 × 512 or more), and often include multispectral or hyperspectral channels.
  • Strong spatial dependencies: Neighboring pixels are highly correlated, making noise injection particularly disruptive.
  • Label imbalance: Rare classes such as airplanes, ships, or oil rigs are disproportionately affected by uniform noise injection.
  • Deep network architectures: The reliance on multi-layer CNN backbones, often with tens of millions of parameters, introduces further sensitivity to training noise.
Existing approaches to private deep learning typically adopt a one-size-fits-all noise strategy—using a fixed clipping norm and static noise multiplier throughout training. This rigid design overlooks two key facts: (1) different layers of a neural network exhibit varying sensitivity to perturbation, and (2) gradient signals fluctuate throughout training as the model approaches convergence. Without adapting to these dynamics, models are prone to underfitting or instability, particularly in non-convex optimization landscapes.
To overcome these limitations, we propose a new framework called ADP-SIR (Adaptive Differential Privacy for Satellite Image Recognition). ADP-SIR is a privacy-preserving training algorithm specifically designed for deep learning on satellite imagery. It introduces two novel components: (1) a Convergence-Guided Noise Scaling (CGNS) strategy that adjusts the noise level based on the evolution of gradient norms during training, and (2) a Layerwise Sensitivity Profiling (LSP) technique that customizes clipping and noise injection per layer, based on empirical gradient magnitudes. These mechanisms collectively enable ADP-SIR to better preserve model utility while satisfying strong DP guarantees.
Furthermore, we provide a theoretical analysis showing that our method maintains convergence guarantees even under non-convex settings, with asymptotic bounds that improve upon existing fixed-noise baselines. This is particularly important for satellite applications where training resources are costly and convergence behavior must be predictable.
We validate ADP-SIR on standard benchmark datasets for satellite image recognition, including E U R O S A T and B I G E A R T H N E T . Our experiments show that ADP-SIR consistently outperforms static-noise DP baselines across a wide range of privacy budgets. Notably, ADP-SIR achieves near non-private accuracy levels under moderate privacy constraints, and gracefully degrades under tight budgets.
Our key contributions can be summarized as follows:
  • We propose ADP-SIR, the first adaptive differentially private training framework tailored for high-resolution satellite imagery, which jointly integrates convergence-aware noise modulation and layer-specific clipping to address the unique challenges of privacy-preserving remote sensing.
  • We develop two novel techniques—Convergence-Guided Noise Scaling (CGNS) and Layerwise Sensitivity Profiling (LSP)—that dynamically calibrate privacy budgets during training, providing finer control of the utility–privacy trade-off than existing methods.
  • We provide rigorous theoretical analysis showing that ADP-SIR maintains convergence guarantees under both convex and non-convex settings, with tighter bounds than standard DP-SGD, thereby advancing the theory of adaptive private optimization.
  • Through extensive experiments on benchmark satellite datasets, we demonstrate that ADP-SIR achieves state-of-the-art accuracy and faster convergence compared to strong baselines, and we release our implementation and benchmark framework to promote reproducibility and future research.
The rest of this paper is organized as follows: Section 2 surveys related work. Section 3 introduces the necessary background on differential privacy and satellite image learning. Section 4 presents the details of our ADP-SIR algorithm and its theoretical properties. Section 5 provides empirical results and ablation studies. Section 6 concludes with future research directions.

2. Related Work

2.1. Differential Privacy in Deep Learning

Differential privacy (DP) has emerged as a foundational paradigm for privacy-preserving machine learning [1]. The introduction of DP-SGD by Abadi et al. [2] enabled end-to-end differentially private training of deep neural networks, with privacy guarantees formalized using the moments accountant. Subsequent works have aimed to tighten these guarantees and improve usability. For example, Rényi Differential Privacy (RDP) [3] and zero-Concentrated Differential Privacy (zCDP) [4] offer stronger composition properties and simpler analysis. McMahan et al. [5] demonstrated that large-scale applications, such as language modeling, are feasible under DP by combining careful optimization with privacy accounting. Despite these advances, static clipping and fixed noise schedules in DP-SGD often lead to significant accuracy degradation, especially in deep or non-convex settings. Recent discussions in this domain emphasize the necessity of adaptive mechanisms that can respond to training dynamics, motivating our convergence-guided approach.

2.2. Adaptive Privacy Mechanisms

To alleviate the limitations of static DP-SGD, recent research has introduced adaptivity into the privacy-preserving optimization process. AdaClip [6] adaptively adjusts gradient clipping thresholds using per-step gradient norms, improving the balance between privacy and utility. Similarly, DPD-Adam [7] extends adaptive moment methods to the private setting, offering faster convergence on certain tasks. Zhou et al. [8] and Triastcyn et al. [9] proposed variance-aware methods that allocate privacy budgets dynamically in federated settings, while Opacus [10] provides practical software tools for large-scale DP training. These methods highlight a broader trend: adaptivity can mitigate the inherent trade-offs in DP optimization. However, most prior approaches focus on either clipping or optimizer-level adjustments, whereas our method, Convergence-Guided Noise Scaling (CGNS), directly links noise scheduling to gradient stability. This integration allows ADP-SIR to reduce perturbation as training stabilizes, which has not been systematically explored in existing works.

2.3. Optimization and Convergence Under Privacy

Theoretical analysis of private optimization has advanced significantly in recent years [11]. Bassily et al. [12] provided early convergence guarantees for convex objectives under DP, while Wang et al. [13] and Yu et al. [14] extended these analyses to the non-convex setting. Song et al. [15] investigated the trade-offs between privacy and algorithmic stability, while Lecuyer et al. [16] established connections between differential privacy and robustness. These works establish the theoretical foundation for understanding how DP mechanisms interact with optimization. However, most analyses assume fixed noise distributions and static clipping strategies, leaving a gap in understanding adaptive approaches. Our work explicitly builds on these foundations by showing that convergence-aware noise scaling preserves the non-convex convergence guarantees, while simultaneously improving empirical efficiency.

2.4. Privacy in Geospatial and Remote Sensing

Although differential privacy has been widely studied in natural language processing, computer vision, and tabular machine learning, its application to geospatial and remote sensing domains is still limited. Helber et al. [17] introduced the EuroSAT dataset as a benchmark for land use classification, while Cheng et al. [18] developed RESISC45 to cover diverse scene categories. Yet, neither dataset was originally designed with privacy considerations. More recent efforts have considered cryptographic solutions such as homomorphic encryption [19] and secure aggregation [20], but these approaches often entail significant computational overheads.
In parallel, the remote sensing community has made rapid progress in hyperspectral image (HSI) analysis, which is highly relevant to privacy-preserving applications since hyperspectral data often contains sensitive environmental and infrastructural details. Recent work on hyperspectral image super-resolution demonstrates the increasing importance of adaptive deep models: Enhanced Deep Image Prior [21], Model-Informed Multistage Networks [22], and X-shaped Interactive Autoencoders with cross-modality mutual learning [23] all advance unsupervised HSI reconstruction. These contributions illustrate the growing sophistication of deep models for satellite imagery. However, none of these methods incorporate privacy guarantees, underscoring the novelty of our proposal. By tailoring DP techniques to geospatial tasks, ADP-SIR fills a crucial gap between cutting-edge remote sensing methods and privacy-preserving learning.

2.5. Beyond Standard DP Training

Alternative privacy-preserving strategies have also been proposed to improve the privacy–utility trade-off outside the standard DP-SGD framework. PATE [24] leverages teacher-student architectures for private knowledge transfer, while scalable aggregation methods [25,26] enable large-scale training under DP through secure multiparty protocols. Tramer and Boneh [27] investigated functional noise scheduling as an alternative to fixed-variance noise injection, while these directions offer valuable insights, they often introduce architectural constraints or additional communication overhead. In contrast, our approach integrates adaptivity directly into the training process, maintaining compatibility with standard architectures and optimization pipelines. This design ensures scalability to large-scale geospatial tasks while maintaining formal differential privacy guarantees.
To the best of our knowledge, no existing method jointly exploits convergence-aware noise modulation and layerwise sensitivity profiling within a formal DP framework for satellite image recognition. Our work bridges this gap, offering both theoretical grounding and empirical superiority over current baselines.

3. Preliminaries

In this section, we review the foundational concepts necessary to understand our proposed method, including differential privacy, optimization in deep learning, and the unique challenges posed by satellite image recognition.

3.1. Differential Privacy

Differential Privacy (DP) [1] is a mathematical framework for quantifying the privacy guarantees provided by an algorithm operating on a dataset. Let A be a randomized algorithm and D , D be two adjacent datasets differing in a single record.
Definition 1
( ( ε , δ ) -Differential Privacy). An algorithm A satisfies ( ε , δ ) -differential privacy if for all measurable subsets S of the output space and for all adjacent datasets D , D ,
Pr [ A ( D ) S ] e ε · Pr [ A ( D ) S ] + δ ,
where ε > 0 is the privacy loss parameter and δ [ 0 , 1 ) is the probability of a privacy breach beyond ε.
To satisfy DP, noise is typically added to model gradients or outputs during training. The most common approach is the Differentially Private Stochastic Gradient Descent (DP-SGD) [2], where clipped gradients are perturbed by Gaussian noise:
g ˜ t = 1 L i = 1 L clip ( g t i , C ) + N ( 0 , σ 2 C 2 I ) ,
where g t i is the gradient of sample i at time t, C is the clipping norm, and σ controls the noise scale.

3.2. Satellite Image Recognition

Satellite image recognition involves classifying or detecting objects from high-resolution geospatial imagery. Unique challenges include
  • High dimensionality: Input images are typically in the range of 512 × 512 pixels or higher, with multiple spectral bands.
  • Spatial correlations: Neighboring pixels are often strongly correlated, making naive noise injection degrade performance significantly.
  • Data imbalance: Certain classes (e.g., ships or aircraft) are underrepresented, exacerbating the utility-privacy trade-off.
Standard convolutional neural networks (CNNs), such as ResNet, are typically employed to model satellite imagery [28,29]. These models are highly sensitive to noise, especially when differential privacy is applied during training.

3.3. Privacy-Preserving Learning Objective

Let f θ denote a CNN with parameters θ R d trained on a dataset D = { ( x i , y i ) } i = 1 n of image-label pairs. The empirical risk minimization (ERM) objective is as follows:
min θ L ( θ ; D ) = 1 n i = 1 n ( f θ ( x i ) , y i ) ,
where ( · , · ) is a classification loss function (e.g., cross-entropy). Under DP constraints, we modify the optimization procedure using a differentially private mechanism A such that the resulting model parameters θ ˜ = A ( D ) satisfy ( ε , δ ) -DP. Our proposed method introduces adaptivity in the noise scale σ and clipping norm C by leveraging gradient convergence behavior and layer-wise sensitivity, as described in Section 4.

4. Methodology

In this section, we introduce ADP-SIR (Adaptive Differential Privacy for Satellite Image Recognition), a novel training algorithm that adaptively adjusts the privacy noise in different layers and stages of training based on gradient convergence and sensitivity. Our method addresses the optimization challenges introduced by DP noise while ensuring strong privacy guarantees and convergence behavior. Algorithm 1 and Figure 1 gives the pipeline of our proposal.

4.1. Overview

Our proposed method builds upon the standard DP-SGD algorithm by introducing two key innovations:
  • Convergence-Guided Noise Scaling (CGNS): A dynamic noise modulation technique that adjusts the variance of the Gaussian noise based on gradient norm stability.
  • Layerwise Sensitivity Profiling (LSP): A per-layer clipping and noise scaling strategy based on observed gradient sensitivity during training.
Algorithm 1: ADP-SIR pipeline
Electronics 14 03680 i001

4.2. Convergence-Guided Noise Scaling

A major challenge in training deep learning models under differential privacy constraints is the degradation of model performance due to constant noise injection throughout training. Traditional DP-SGD methods introduce a fixed noise scale σ across all iterations, regardless of how close the model is to convergence. This uniform noise scheme is suboptimal because it adds excessive noise when the model is nearing convergence, which can disrupt fine-grained updates, or insufficient noise during early phases, which may risk privacy violations due to large gradient norms.
To address this limitation, we propose Convergence-Guided Noise Scaling (CGNS), a dynamic mechanism that adapts the Gaussian noise level injected into the gradients based on the optimization trajectory. The core idea is to modulate the noise scale in accordance with the stability of the gradient direction: if the gradients are fluctuating significantly, the optimizer is likely still far from a stationary point and can tolerate higher noise. Conversely, when the gradients begin to stabilize, the model is closer to convergence and requires more precise updates with minimal perturbation.
Let g t denote the aggregated, clipped gradient vector at iteration t. We define a convergence indicator ϕ t as follows:
ϕ t = g t 1 g t 2 g t 2 + ϵ ,
where ϵ > 0 is a small constant to prevent division by zero. Intuitively, ϕ t measures the relative change in gradient direction from one iteration to the next. A small value of ϕ t implies that the optimization has stabilized, suggesting that further updates should be made with lower noise to preserve useful information.
We modulate the noise multiplier σ t at each step t using the following exponentially decaying function:
σ t = σ min + ( σ max σ min ) · exp ( α · ϕ t ) ,
where σ min and σ max denote the lower and upper bounds of the noise multiplier, respectively, and α > 0 is a tunable hyperparameter controlling the decay rate.
This formulation ensures the following:
  • σ t σ max when ϕ t is large (early-stage training, unstable gradients),
  • σ t σ min when ϕ t 0 (late-stage training, stable gradients).
The use of a decreasing noise schedule is theoretically supported by recent work on non-convex optimization with differential privacy [13]. In essence, when training progresses and updates approach a stationary point, injecting large noise becomes inefficient and harmful to convergence. Our approach provides a principled mechanism to reduce noise at the right time while maintaining ( ε , δ ) -DP guarantees using advanced composition accounting tools like the moments accountant.
In practice, ϕ t is computed using a moving average of recent gradient differences to smooth out short-term noise, and the update of σ t is clipped within [ σ min , σ max ] to ensure bounded privacy cost. Moreover, σ t is passed into the per-layer noise injection pipeline described in the next subsection, which performs further adjustment based on individual layer sensitivity.
In summary, CGNS improves the trade-off between privacy and utility by allowing adaptive control over noise magnitude, aligned with the training dynamics. This is particularly crucial in high-dimensional settings like satellite image recognition, where fixed-noise schemes struggle to converge to performant models under strict privacy budgets.

4.3. Layerwise Sensitivity Profiling (LSP)

Deep convolutional neural networks (CNNs), commonly used for satellite image recognition, exhibit varying sensitivity to noise across different layers. Lower layers typically extract local, low-level features such as edges and textures, while higher layers encode more abstract, semantic representations. Empirical studies have shown that gradients in deeper layers often have larger magnitudes and greater variance. Applying uniform gradient clipping and noise injection across all layers, as performed in standard DP-SGD, results in either excessive clipping in high-sensitivity layers or unnecessary noise in low-sensitivity ones—both leading to suboptimal learning performance.
To tackle this challenge, we propose Layerwise Sensitivity Profiling (LSP), a mechanism that adaptively determines gradient clipping thresholds and noise scales for each layer based on its estimated sensitivity. By aligning the privacy-preserving noise with the information flow in the network, LSP allows for more informed trade-offs between privacy and model utility.
For each layer l of the neural network, we maintain an estimate of its empirical gradient sensitivity. At iteration t, for a mini-batch B t of size B, we compute the per-sample gradients { g i ( l ) } i = 1 B and define the batch-wise mean norm as follows:
S ^ t ( l ) = 1 B i = 1 B g i ( l ) 2 .
To smooth fluctuations and adapt to long-term trends, we maintain an exponential moving average:
S t ( l ) = τ · S t 1 ( l ) + ( 1 τ ) · S ^ t ( l ) ,
where τ [ 0 , 1 ) is a decay parameter controlling the memory length.
The clipping threshold for each layer is computed as a scaled version of its sensitivity estimate:
C t ( l ) = β · S t ( l ) ,
where β > 1 is a tunable margin factor ensuring that most gradients fall within the clipping norm. This prevents over-clipping of informative gradients while maintaining bounded sensitivity for differential privacy. Each gradient g i ( l ) is then clipped as follows:
g ¯ i ( l ) = g i ( l ) max 1 , g i ( l ) 2 C t ( l ) .
Once gradients are clipped, Gaussian noise is added to the aggregated gradient for each layer:
g ˜ t ( l ) = 1 B i = 1 B g ¯ i ( l ) + N 0 , ( σ t C t ( l ) ) 2 I ,
where σ t is the current global noise multiplier determined by CGNS (Section 4), and C t ( l ) modulates its effect per layer. This effectively injects more noise into layers with higher sensitivity, while preserving useful updates in more stable layers.
Although noise is injected per layer, the overall privacy cost is still analyzed under the composition theorem by treating the entire gradient vector as the DP-sensitive output. Since all layer-wise noise contributions are calibrated using σ t , the mechanism remains within the global ( ε , δ ) budget when analyzed using the moments accountant.
Layerwise sensitivity profiling is especially useful in satellite image tasks due to their complex and multi-scale spatial features. For instance, fine-grained object boundaries or terrain textures may only appear in certain layers, and their gradients are more susceptible to perturbation. By adjusting noise granularity based on layer roles and dynamics, LSP enhances learning stability and task-specific robustness under privacy constraints.
In summary, LSP complements CGNS by providing an additional layer of adaptivity at the architectural level, improving the alignment between noise placement and model sensitivity. Together, these components form the core of ADP-SIR’s adaptive privacy-preserving learning strategy.

4.4. Privacy Guarantee

Ensuring strong theoretical privacy guarantees while allowing adaptive mechanisms such as CGNS and LSP presents a technical challenge. In this subsection, we demonstrate that ADP-SIR satisfies ( ε , δ ) -differential privacy using the moments accountant framework [2], which offers tight bounds for composition and supports adaptive mechanisms as long as the noise scaling remains data-independent.

4.4.1. Differential Privacy Recap

Recall that a randomized mechanism A : D R satisfies ( ε , δ ) -DP if for any two adjacent datasets D and D differing in a single sample, and for any measurable subset S R , we obtain the following:
Pr [ A ( D ) S ] e ε Pr [ A ( D ) S ] + δ .
In ADP-SIR, privacy loss is incurred primarily through the stochastic gradient descent updates where Gaussian noise is injected to preserve privacy. The challenge lies in adapting this noise over time without violating privacy assumptions.

4.4.2. Moments Accountant

The moments accountant tracks a bound on the log moment of the privacy loss random variable:
α A ( λ ) = max D , D log E o A ( D ) Pr [ A ( D ) = o ] Pr [ A ( D ) = o ] λ ,
for positive integer λ . For each iteration t with noise multiplier σ t and sampling rate q = B n , the moments accountant accumulates:
α t ( λ ) q 2 λ ( λ + 1 ) 2 σ t 2 ,
assuming a fixed clipping norm per layer. Then, by summing over T iterations and applying advanced composition bounds, we obtain the following:
ε min λ t = 1 T α t ( λ ) log δ λ 1 .

4.4.3. Handling Adaptivity

A key observation is that the adaptivity in ADP-SIR arises from training dynamics (i.e., ϕ t and S t ( l ) ), not the raw data. Since CGNS and LSP modify the scale of the added noise using statistics that are independent of any individual training example, they do not violate the core DP assumptions. This ensures the following:
  • The noise variance σ t 2 is determined by the global convergence behavior, not any specific sample.
  • The per-layer clipping norms C t ( l ) are computed using moving averages over batch statistics and do not compromise individual record privacy.
Therefore, even though the injected noise and clipping norms vary over time and across layers, the privacy loss per iteration is well-defined and upper-bounded using the worst-case parameters ( σ min , C max ) . This makes the total privacy loss analyzable using conservative bounds.
In our implementation, we use the Rényi Differential Privacy (RDP) accountant [3], a generalization of the moments accountant that provides tighter bounds for subsampled Gaussian mechanisms. At the end of training, we convert the cumulative RDP parameters into an ( ε , δ ) bound using the standard conversion formula. The cumulative privacy cost is reported along with experimental results (see Section 5).

4.5. Convergence Analysis

While differential privacy ensures protection against data leakage, it also introduces additional noise into the optimization process, which can adversely impact convergence and model utility. In this section, we present a theoretical analysis of the convergence behavior of our proposed ADP-SIR algorithm under mild assumptions. In particular, we focus on the expected gradient norm as the measure of optimization quality in the non-convex setting, which is the common scenario in deep learning for satellite image recognition.
Assumption A1.
Let L ( θ ) denote the empirical loss function, where θ R d represents the model parameters. We make the following standard assumptions:
A1 
(Smoothness) The loss function L ( θ ) is L-smooth, i.e.,
L ( θ 1 ) L ( θ 2 ) L θ 1 θ 2 , θ 1 , θ 2 .
A2 
(Bounded Gradient Norm) The gradients are bounded: for all samples i, i ( θ ) 2 G , and the clipped gradient norm is at most C t ( l ) per layer.
A3 
(Unbiased Gradient with Bounded Variance) For any mini-batch B t of size B, the noisy gradient estimator g ˜ t satisfies:
E [ g ˜ t ] = L ( θ t ) , E [ g ˜ t L ( θ t ) 2 ] σ t 2 C t 2 / B + ζ 2 ,
where ζ 2 accounts for stochasticity in the data sampling and clipping.

Main Result

Theorem 1
(Non-convex Convergence). Suppose Assumptions A1–A3 hold. Let the learning rate η t = η be constant and satisfy η 1 / L . Then, after T iterations of ADP-SIR, the average expected gradient norm satisfies:
1 T t = 1 T E L ( θ t ) 2 2 ( L ( θ 0 ) L * ) η T + O σ max 2 C max 2 B ,
where L * is the minimum achievable loss, and C max is the maximum per-layer clipping threshold over all t.
Interpretation
The bound consists of two terms:
  • The first term decays with 1 / T , ensuring that the average gradient norm becomes small as training progresses. This guarantees that the algorithm approaches a stationary point.
  • The second term captures the asymptotic noise floor induced by privacy-preserving noise. It is proportional to σ max 2 , the worst-case noise level introduced by CGNS, and the maximum clipping norm from LSP.
By designing σ t and C t ( l ) to decrease over time (as gradients stabilize), ADP-SIR asymptotically reduces noise injection, thus improving convergence precision. In particular, the adaptivity of CGNS ensures that σ t σ min as ϕ t 0 , minimizing the impact of noise when high optimization accuracy is required.
In traditional DP-SGD, both the noise multiplier and clipping norm are fixed. This corresponds to constant values σ t = σ , C t = C for all t. In that case, the noise floor is fixed throughout training, potentially dominating the optimization error in the late stages. ADP-SIR improves upon this by reducing both noise and clipping norms adaptively as the model converges, without violating privacy guarantees.
Proof. 
Let θ t denote the model parameters at iteration t. The update rule for ADP-SIR is as follows:
θ t + 1 = θ t η g ˜ t ,
where g ˜ t is the noisy clipped gradient computed using CGNS and LSP:
g ˜ t = 1 B i = 1 B g ¯ i + ξ t , with ξ t N ( 0 , σ t 2 C t 2 I ) ,
and g ¯ i is the clipped gradient of sample i with norm at most C t .
By the L-smoothness of L ( · ) (Assumption A1), we obtain the following:
L ( θ t + 1 ) L ( θ t ) η L ( θ t ) , g ˜ t + L η 2 2 g ˜ t 2 .
Taking expectation over the randomness in g ˜ t (sampling and noise), and using the assumption that E [ g ˜ t ] = L ( θ t ) (Assumption A3), we obtain the following:
E [ L ( θ t + 1 ) ] E [ L ( θ t ) ] η E [ L ( θ t ) 2 ] + L η 2 2 E [ g ˜ t 2 ] .
We now bound E [ g ˜ t 2 ] . By Assumption A3:
E [ g ˜ t 2 ] = E [ L ( θ t ) 2 ] + E [ g ˜ t L ( θ t ) 2 ] E [ L ( θ t ) 2 ] + σ t 2 C t 2 B + ζ 2 .
Plugging into Equation (2), we get the following:
E [ L ( θ t + 1 ) ] E [ L ( θ t ) ] η E [ L ( θ t ) 2 ] + L η 2 2 E [ L ( θ t ) 2 ] + σ t 2 C t 2 B + ζ 2
= E [ L ( θ t ) ] η 1 L η 2 E [ L ( θ t ) 2 ] + L η 2 2 σ t 2 C t 2 B + ζ 2 .
Assume η 1 L so that 1 L η 2 1 2 . Then,
E [ L ( θ t + 1 ) ] E [ L ( θ t ) ] η 2 E [ L ( θ t ) 2 ] + L η 2 2 σ t 2 C t 2 B + ζ 2 .
Rearranging
E [ L ( θ t ) 2 ] 2 η E [ L ( θ t ) ] E [ L ( θ t + 1 ) ] + L η σ t 2 C t 2 B + ζ 2 .
Summing over t = 1 to T:
t = 1 T E [ L ( θ t ) 2 ] 2 η ( L ( θ 0 ) L * ) + L η t = 1 T σ t 2 C t 2 B + ζ 2 .
Dividing both sides by T:
1 T t = 1 T E [ L ( θ t ) 2 ] 2 ( L ( θ 0 ) L * ) η T + L η T t = 1 T σ t 2 C t 2 B + ζ 2 .
Since σ t σ max and C t C max for all t, and ζ 2 is constant:
1 T t = 1 T E [ L ( θ t ) 2 ] 2 ( L ( θ 0 ) L * ) η T + L η σ max 2 C max 2 B + ζ 2 .
Choosing η = O ( 1 / T ) balances both terms, yielding
1 T t = 1 T E [ L ( θ t ) 2 ] = O 1 T + σ max 2 C max 2 B ,
as claimed. □

4.6. Parameter Choices

In designing ADP-SIR, we carefully selected several parameters to balance privacy preservation, convergence stability, and empirical performance. The noise bounds ( σ min , σ max ) were chosen as ( 0.8 ,   2.0 ) to ensure that noise remains sufficiently large during early training, when gradients are unstable, while gradually decaying to a smaller value near convergence to preserve accuracy. The decay coefficient α = 10 in the CGNS mechanism was set to provide a moderate rate of noise reduction, avoiding overly aggressive decay that could compromise privacy in later training stages. For LSP, we adopted an exponential moving average factor τ = 0.9 , which smooths fluctuations in sensitivity estimates while remaining responsive to long-term trends. Finally, the clipping margin factor β = 1.5 ensures that clipping thresholds remain slightly above the average gradient magnitude, preventing over-clipping of informative gradients while keeping sensitivity bounded for differential privacy analysis. These parameter values were selected through a combination of theoretical considerations—particularly maintaining convergence guarantees under non-convex optimization—and empirical validation on benchmark datasets, providing a principled balance between privacy cost and model utility in satellite image recognition.

5. Experiments

In this section, we evaluate the performance of the proposed ADP-SIR algorithm on benchmark satellite image recognition tasks. Our experiments aim to answer the following research questions:
  • Q1: Does ADP-SIR improve recognition performance under strict differential privacy compared to existing DP training methods?
  • Q2: How do the adaptive mechanisms—CGNS and LSP—contribute to training stability and accuracy?
  • Q3: Is the convergence behavior of ADP-SIR consistent with theoretical guarantees under various privacy budgets?

5.1. Experimental Setup

We now describe the setup used to evaluate the effectiveness of the proposed ADP-SIR algorithm. This includes the datasets, baselines, model architectures, privacy configuration, training details, evaluation metrics, and reproducibility notes.

5.1.1. Datasets

We conduct experiments on two standard satellite image classification benchmarks.
EuroSAT [17] consists of 27,000 RGB images derived from Sentinel-2 satellite data. Each image is 64 × 64 pixels in resolution and labeled with one of 10 scene categories, including residential, forest, river, and highway. The dataset is relatively balanced across classes. We use an 80/20 training/testing split and set aside 10% of the training data for validation.
RESISC45 [18] includes 31,500 RGB images from Google Earth imagery, spanning 45 scene classes such as airport, bridge, stadium, and farmland. Each image has a resolution of 256 × 256 . We adopt the conventional 80/20 train/test split and use the full training set during model fitting.

5.1.2. Baselines

We compare ADP-SIR against three prominent differentially private learning methods.
DP-SGD [2] is the canonical baseline, which adds Gaussian noise to clipped gradients with fixed parameters throughout training. It serves as the foundation for most modern DP training algorithms.
AdaClip-DP [6] improves upon DP-SGD by dynamically adjusting the clipping norm based on gradient statistics. However, it does not adapt the noise scale and uses the same value across all layers.
PATE [24] is a semi-supervised approach that provides privacy via noisy aggregation of teacher model votes. It is known for strong privacy guarantees under tight ε , though it is limited to smaller datasets and is not applicable to RESISC45 due to scalability issues.

5.1.3. Model Architecture

For EuroSAT, we use a modified ResNet-18 architecture with input resolution adjusted to 64 × 64 . Batch normalization layers are replaced with group normalization to avoid privacy leakage from statistics that depend on batch-level aggregation. For RESISC45, we adopt a lightweight CNN composed of four convolutional blocks followed by two fully connected layers. Each convolution block includes ReLU activations and max pooling. This model is more computationally efficient and better suited for large-scale experiments with DP constraints.

5.1.4. Privacy Configuration

We consider privacy budgets ε { 1.0 , 2.0 , 4.0 } with a fixed δ = 10 5 . These settings correspond to strong, moderate, and relaxed privacy regimes, respectively. Privacy accounting is performed using Rényi Differential Privacy (RDP) [3], which allows tighter composition of privacy losses across training iterations. The effective ε value is monitored during training and reported upon convergence. For ADP-SIR, we set the noise multiplier bounds to σ min = 0.8 and σ max = 2.0 . The convergence-guided decay coefficient is α = 10 . The layerwise sensitivity profiling uses a decay rate τ = 0.9 for the exponential moving average and a clipping margin factor β = 1.5 .

5.1.5. Training Details

All models are trained using stochastic gradient descent (SGD) with momentum 0.9. The initial learning rate is set to 0.01 and decayed using cosine annealing. We use a batch size of 128 for all experiments and train for a maximum of 100 epochs. For DP-SGD and AdaClip-DP, we tune the fixed clipping norm and noise multiplier to ensure the final privacy budget does not exceed the target ε . For ADP-SIR, adaptivity ensures tighter convergence within the same budget. All experiments are implemented in PyTorch 2.7.0 using the Opacus library for DP mechanisms and run on NVIDIA RTX 3090 GPUs.

5.1.6. Evaluation Metrics

We evaluate each model using the following metrics: Test Accuracy: Classification accuracy on the held-out test set. Training Loss: Reported per epoch to observe convergence dynamics. Gradient Norm Decay: Measures the smoothness and stabilization of training, as predicted by our theoretical analysis. Effective Privacy Cost: The final ( ε , δ ) achieved as computed by the RDP accountant.
These metrics collectively reflect both utility and stability under differential privacy constraints.

5.2. Quantitative Results

This subsection presents the classification performance of ADP-SIR and baseline methods under different privacy budgets on the EuroSAT and RESISC45 datasets. We report results with multiple metrics and visualize both accuracy trends and utility-privacy trade-offs. To ensure statistical reliability, each result is averaged over three runs with different random seeds.

5.2.1. Test Accuracy Across Privacy Budgets

We first present the classification accuracy (mean ± standard deviation) of all evaluated methods in Table 1. ADP-SIR consistently outperforms both DP-SGD and AdaClip-DP across all privacy levels on both datasets.
Observation: ADP-SIR consistently yields the best accuracy under all privacy budgets. Notably, at ε = 1.0 , it achieves 69.7% accuracy on EuroSAT—close to PATE’s performance but without the computational overhead of teacher ensembles.

5.2.2. Accuracy Distribution by Class

To analyze model robustness, we compute per-class accuracy for the EuroSAT dataset at ε = 2.0 , and the results are presented in Table 2.

5.2.3. Utility-Privacy Trade-Off

Figure 2 visualizes the trade-off between privacy and utility on both datasets. ADP-SIR provides a more favorable accuracy curve at all ε values.

5.2.4. Performance Stability

To assess training stability, we compute the standard deviation of test accuracy across three seeds. Table 3 shows that ADP-SIR achieves both higher accuracy and lower variance, confirming its robustness.

5.3. Convergence

In this subsection, we empirically examine the convergence behavior of ADP-SIR during training and compare it to that of DP-SGD and AdaClip-DP. Our objective is to validate the theoretical findings presented in Section 4, particularly the following:
  • ADP-SIR facilitates smoother and faster convergence due to its convergence-guided noise scaling (CGNS);
  • Layerwise sensitivity profiling (LSP) preserves optimization signal across all layers;
  • The injected noise decays gradually over time, stabilizing training toward the end.
We evaluate two key metrics:
  • Training loss per epoch, to assess how efficiently the model reduces the empirical risk.
  • Gradient norm of the model (averaged across layers), to assess the optimizer’s stability and proximity to a stationary point.
All experiments are conducted on the EuroSAT dataset under a fixed privacy budget of ε = 2.0 .

5.3.1. Training Loss Curve

Figure 3 shows the training loss of each method over 100 epochs. ADP-SIR not only converges faster but also reaches a lower final training loss compared to DP-SGD and AdaClip-DP.

5.3.2. Gradient Norm Decay

We further analyze the 2 norm of the model’s gradient at each epoch, averaged across all layers. This serves as a proxy for convergence, since in non-convex optimization, convergence to a stationary point is commonly measured by the reduction in gradient norm over time. The results are presented in Figure 4.

5.3.3. Conformity with Theoretical Guarantees

The empirical results confirm our theoretical predictions in Section 4:
  • ADP-SIR demonstrates a faster decrease in the expected gradient norm, consistent with the O ( 1 / T ) convergence rate in non-convex settings.
  • The lower final training loss implies that CGNS enables more stable fine-tuning near convergence.
  • Adaptive clipping through LSP prevents gradient explosion or vanishing across layers, which is especially important in deep convolutional networks.

5.4. Federated Baseline Comparison and Computational Cost

We expand our experimental evaluation with (i) stronger federated baselines—DP-FedAvg, FedDP—and a non-private FedAvg reference upper bound, and (ii) a detailed computational cost analysis. Unless noted, results are averaged over three runs with privacy ϵ { 1 , 2 , 4 } , δ = 10 5 , batch size = 128 , and identical preprocessing. For federated settings we use 10 clients, 5 local epochs per round. The federated learning setups are also presented in Table 4.
Across privacy budgets, ADP-SIR achieves the strongest accuracy among private methods on both EuroSAT and RESISC45 (Table 5 and Table 6), narrowing the gap to the non-private FedAvg reference. The detailed cost analysis (Table 7, Table 8 and Table 9) shows a modest per-epoch overhead from LSP and CGNS that is offset by substantially fewer epochs to target accuracy, yielding improved end-to-end efficiency. Federated baselines (Table 8) incur additional communication and client-side costs and do not match ADP-SIR’s utility under the same privacy budgets.

5.5. Ablation Study

To better understand the contributions of each component in the proposed ADP-SIR framework, we conduct an ablation study on the EuroSAT dataset under two privacy budgets: ε { 1.0 , 2.0 } . Specifically, we analyze the effect of removing the following:
  • Convergence-Guided Noise Scaling (CGNS), which dynamically adjusts the noise level based on gradient stability;
  • Layerwise Sensitivity Profiling (LSP), which adaptively clips gradients at the layer level based on historical sensitivity estimates.
By evaluating the performance of ADP-SIR with each component removed in isolation, we assess their individual and joint impact on classification accuracy, convergence speed, and per-class fairness.

5.5.1. Ablation Variants

We consider the following configurations:
  • ADP-SIR (full): The complete method as described in Section 4, combining CGNS and LSP.
  • w/o CGNS: The noise multiplier is fixed to σ = σ max throughout training. LSP remains active.
  • w/o LSP: A global clipping norm is used across all layers. CGNS remains active.
  • w/o CGNS and LSP: Both adaptive components are removed. Equivalent to DP-SGD with tuned fixed parameters.

5.5.2. Overall Results

Table 10 summarizes the mean test accuracy at ε = 2.0 . The complete ADP-SIR clearly outperforms all ablated variants.

5.5.3. Multi-Budget Evaluation

We also evaluate at a stricter privacy budget of ε = 1.0 to study robustness under higher noise. THe results in Table 11 show that both components become more critical as privacy becomes stricter.

5.5.4. Additional Metrics

Accuracy alone does not capture performance nuances. Table 12 reports precision, recall, and macro F1-score for ε = 2.0 .
To examine fairness impacts, Table 13 shows per-class accuracy for selected representative classes at ε = 2.0 . The removal of LSP disproportionately harms small-object categories like SeaLake and Industrial.

6. Conclusions

We introduced ADP-SIR, an adaptive differentially private training framework tailored for satellite image recognition tasks under strict privacy constraints. Our method addresses key limitations of traditional DP-SGD by introducing two complementary innovations: (1) Convergence-Guided Noise Scaling (CGNS), which modulates the noise level based on gradient stability to reduce unnecessary perturbations during later training stages; and (2) Layerwise Sensitivity Profiling (LSP), which performs fine-grained adaptive clipping per network layer, preserving important optimization signals. Theoretically, we established non-convex convergence guarantees for ADP-SIR under standard smoothness assumptions. Empirically, our extensive experiments on the EuroSAT and RESISC45 benchmarks demonstrate that ADP-SIR achieves superior accuracy, faster convergence, and better per-class generalization compared to state-of-the-art methods such as DP-SGD and AdaClip-DP.
Our ablation studies further confirm the necessity of both adaptive components, and convergence analysis shows alignment with theoretical expectations. Overall, ADP-SIR represents a principled and practical advancement in privacy-preserving deep learning, particularly for complex vision tasks such as satellite image classification. Beyond benchmark evaluation, a concrete future application scenario is the use of ADP-SIR in privacy-preserving disaster response systems, where satellite images of affected regions must be analyzed quickly for rescue and relief planning while ensuring that sensitive details about infrastructure or private property remain protected. Future work will also explore the extension of ADP-SIR to federated and decentralized settings, as well as applications to multimodal geospatial data under stricter privacy regimes.

Author Contributions

Conceptualization, Z.Y. and X.T.; Methodology, Z.Y. and X.T.; Software, Z.Y.; Validation, G.C.; Formal analysis, X.Y.; Resources, X.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in this study are included in thearticle. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Dwork, C.; McSherry, F.; Nissim, K.; Smith, A. Calibrating noise to sensitivity in private data analysis. In Proceedings of the Theory of Cryptography Conference, New York, NY, USA, 4–7 March 2006; Springer: Berlin/Heidelberg, Germany, 2006; pp. 265–284. [Google Scholar]
  2. Abadi, M.; Chu, A.; Goodfellow, I.; McMahan, H.B.; Mironov, I.; Talwar, K.; Zhang, L. Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, Vienna, Austria, 24–28 October 2016; ACM: New York, NY, USA, 2016; pp. 308–318. [Google Scholar]
  3. Mironov, I. Rényi differential privacy. In Proceedings of the 2017 IEEE 30th Computer Security Foundations Symposium (CSF), Santa Barbara, CA, USA, 21–25 August 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 263–275. [Google Scholar]
  4. Bun, M.; Steinke, T. Concentrated differential privacy: Simplifications, extensions, and lower bounds. In Proceedings of the Theory of Cryptography Conference, New York, NY, USA, 4–7 March 2016; pp. 635–658. [Google Scholar]
  5. McMahan, H.B.; Ramage, D.; Talwar, K.; Zhang, L. Learning differentially private recurrent language models. arXiv 2018, arXiv:1710.06963. [Google Scholar] [CrossRef]
  6. Pichapati, V.; Xu, Z.; Tople, S.; Thakurta, A. AdaCliP: Adaptive Clipping for Private SGD. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 6–9 May 2019. [Google Scholar]
  7. Lee, J.; Kifer, D. Concentrated differentially private gradient descent with adaptive per-iteration privacy budget. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, London, UK, 19–23 August 2018; pp. 1656–1665. [Google Scholar]
  8. Song, S.; Chaudhuri, K.; Sarwate, A.D. Stochastic gradient descent with differentially private updates. In Proceedings of the 2013 IEEE Global Conference on Signal and Information Processing, Austin, TX, USA, 3–5 December 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 245–248. [Google Scholar]
  9. Triastcyn, A.; Faltings, B. Federated learning with bayesian differential privacy. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 2587–2596. [Google Scholar]
  10. Yousefpour, A. Opacus: Scaling differential privacy to large-scale machine learning. arXiv 2021, arXiv:2109.12298. [Google Scholar]
  11. Pan, Z.; Ying, Z.; Wang, Y.; Zhang, C.; Zhang, W.; Zhou, W.; Zhu, L. Feature-Based Machine Unlearning for Vertical Federated Learning in IoT Networks. IEEE Trans. Mob. Comput. 2025, 24, 5031–5044. [Google Scholar] [CrossRef]
  12. Bassily, R.; Smith, A.; Thakurta, A. Private empirical risk minimization: Efficient algorithms and tight error bounds. In Proceedings of the IEEE FOCS, Philadelphia, PA, USA, 18–21 October 2014; pp. 464–473. [Google Scholar]
  13. Wang, D. Differentially private non-convex optimization with adaptive gradient clipping. In Proceedings of the ICML, Long Beach, CA, USA, 9–15 June 2019. [Google Scholar]
  14. Yu, L. Differentially private SGD for non-convex learning. In Proceedings of the IJCAI, Macao, China, 16–10 August 2019; pp. 3952–3958. [Google Scholar]
  15. Song, S.; Chaudhuri, K.; Sarwate, A.D. Characterizing the privacy-utility tradeoff in machine learning. In Proceedings of the ICML, Virtual, 13–18 July 2020. [Google Scholar]
  16. Lecuyer, M.; Atlidakis, V.; Geambasu, R.; Hsu, D.; Jana, S. Certified robustness to adversarial examples with differential privacy. In Proceedings of the IEEE Symposium on Security and Privacy, San Francisco, CA, USA, 19–23 May 2019. [Google Scholar]
  17. Helber, P.; Bischke, B.; Dengel, A.; Borth, D. EuroSAT: A Novel Dataset and Deep Learning Benchmark for Land Use and Land Cover Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 2217–2226. [Google Scholar] [CrossRef]
  18. Cheng, G.; Han, J.; Lu, X. Remote sensing image scene classification: Benchmark and state of the art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef]
  19. Zhang, D.; Ren, L.; Shafiq, M.; Gu, Z. A lightweight privacy-preserving system for the security of remote sensing images on iot. Remote Sens. 2022, 14, 6371. [Google Scholar] [CrossRef]
  20. Jiang, Y.; Zhang, Y.; Zhu, R. Privacy-preserving distributed learning for satellite image segmentation. IEEE Trans. Geosci. Remote Sens. 2022. [Google Scholar]
  21. Li, J.; Zheng, K.; Gao, L.; Han, Z.; Li, Z.; Chanussot, J. Enhanced deep image prior for unsupervised hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5504218. [Google Scholar] [CrossRef]
  22. Li, J.; Zheng, K.; Gao, L.; Ni, L.; Huang, M.; Chanussot, J. Model-informed multistage unsupervised network for hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5516117. [Google Scholar] [CrossRef]
  23. Li, J.; Zheng, K.; Li, Z.; Gao, L.; Jia, X. X-shaped interactive autoencoders with cross-modality mutual learning for unsupervised hyperspectral image super-resolution. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5518317. [Google Scholar] [CrossRef]
  24. Papernot, N.; Abadi, M.; Erlingsson, U.; Goodfellow, I.; Talwar, K. Semi-supervised knowledge transfer for deep learning from private training data. In Proceedings of the 5th International Conference on Learning Representations (ICLR), Toulon, France, 24–26 April 2017. [Google Scholar]
  25. Papernot, N.; Song, S.; Mironov, I.; Raghunathan, A.; Talwar, K.; Erlingsson, Ú. Scalable private learning with pate. arXiv 2018, arXiv:1802.08908. [Google Scholar] [CrossRef]
  26. Boneh, D.; Boyle, E.; Corrigan-Gibbs, H.; Gilboa, N.; Ishai, Y. Lightweight techniques for private heavy hitters. In Proceedings of the 2021 IEEE Symposium on Security and Privacy (SP), San Francisco, CA, USA, 24–27 May 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 762–776. [Google Scholar]
  27. Tramèr, F.; Boneh, D. Differentially private learning needs better features (or much more data). arXiv 2020, arXiv:2002.11173. [Google Scholar]
  28. Ivanov, A.; Stoliarenko, M.; Kruglik, S.; Novichkov, S.; Savinov, A. Dynamic resource allocation in LEO satellite. In Proceedings of the 2019 15th International Wireless Communications & Mobile Computing Conference (IWCMC), Tangier, Morocco, 24–28 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 930–935. [Google Scholar]
  29. Richtarik, P. Handling Device Heterogeneity in Federated Learning: The First Optimal Parallel SGD in the Presence of Data, Compute and Communication Heterogeneity. In Proceedings of the International Workshop on Secure and Efficient Federated Learning, Hanoi, Vietnam, 25–29 August 2025; p. 1. [Google Scholar]
Figure 1. Overview of the ADP-SIR pipeline. The framework combines Convergence-Guided Noise Scaling (CGNS), which adjusts noise based on gradient stability, and Layerwise Sensitivity Profiling (LSP), which adapts clipping and noise per layer. Both feed into Gaussian noise injection monitored by the RDP accountant, ensuring rigorous ( ε , δ ) -DP guarantees during model updates.
Figure 1. Overview of the ADP-SIR pipeline. The framework combines Convergence-Guided Noise Scaling (CGNS), which adjusts noise based on gradient stability, and Layerwise Sensitivity Profiling (LSP), which adapts clipping and noise per layer. Both feed into Gaussian noise injection monitored by the RDP accountant, ensuring rigorous ( ε , δ ) -DP guarantees during model updates.
Electronics 14 03680 g001
Figure 2. Utility-privacy trade-off curves on EuroSAT. ADP-SIR maintains higher accuracy across all privacy levels.
Figure 2. Utility-privacy trade-off curves on EuroSAT. ADP-SIR maintains higher accuracy across all privacy levels.
Electronics 14 03680 g002
Figure 3. Training loss on EuroSAT with ε = 2.0 . ADP-SIR converges faster and achieves lower final loss than other differentially private methods.
Figure 3. Training loss on EuroSAT with ε = 2.0 . ADP-SIR converges faster and achieves lower final loss than other differentially private methods.
Electronics 14 03680 g003
Figure 4. Gradient norm decay over training epochs. ADP-SIR reduces gradient norms faster and reaches a flatter region of the loss surface, indicating stronger convergence.
Figure 4. Gradient norm decay over training epochs. ADP-SIR reduces gradient norms faster and reaches a flatter region of the loss surface, indicating stronger convergence.
Electronics 14 03680 g004
Table 1. Test accuracy (%) on EuroSAT and RESISC45 with standard deviation over three runs.
Table 1. Test accuracy (%) on EuroSAT and RESISC45 with standard deviation over three runs.
MethodEuroSATRESISC45
ε = 1.0 ε = 2.0 ε = 4.0 ε = 1.0 ε = 2.0 ε = 4.0
DP-SGD63.1 ± 0.870.3 ± 0.575.6 ± 0.452.4 ± 0.659.1 ± 0.465.2 ± 0.3
AdaClip-DP66.4 ± 0.772.8 ± 0.577.3 ± 0.554.7 ± 0.561.5 ± 0.666.0 ± 0.4
PATE71.2 ± 0.3
ADP-SIR (ours)69.7 ± 0.476.9 ± 0.380.1 ± 0.257.6 ± 0.364.8 ± 0.370.4 ± 0.2
Table 2. Per-class accuracy (%) on EuroSAT at ε = 2.0 for DP-SGD, AdaClip-DP, and ADP-SIR. ADP-SIR achieves consistently higher accuracy across nearly all categories.
Table 2. Per-class accuracy (%) on EuroSAT at ε = 2.0 for DP-SGD, AdaClip-DP, and ADP-SIR. ADP-SIR achieves consistently higher accuracy across nearly all categories.
ClassDP-SGDAdaClip-DPADP-SIR (Ours)
Sea/Lake85.089.092.0
River67.070.076.0
Residential80.083.088.0
Pasture84.086.090.0
Permanent Crop72.076.082.0
Highway66.071.077.0
Forest88.090.093.0
Herbaceous78.080.085.0
Annual Crop82.084.089.0
Industrial76.079.083.0
Average77.880.885.5
Table 3. Standard deviation of test accuracy (%) at ε = 2.0 across three seeds.
Table 3. Standard deviation of test accuracy (%) at ε = 2.0 across three seeds.
MethodEuroSATRESISC45
DP-SGD0.50.4
AdaClip-DP0.50.6
ADP-SIR (ours)0.30.3
Table 4. Federated setup details used for DP-FedAvg and FedDP comparisons (EuroSAT).
Table 4. Federated setup details used for DP-FedAvg and FedDP comparisons (EuroSAT).
Dataset#ClientsLocal EpochsBatchRoundsLRClip C
EuroSAT105128110 (DP-FedAvg)0.011.2
EuroSAT105128116 (FedDP)0.011.2
Table 5. EuroSAT accuracy (%) across privacy budgets ( δ = 10 5 ). Mean ± std over 3 runs.
Table 5. EuroSAT accuracy (%) across privacy budgets ( δ = 10 5 ). Mean ± std over 3 runs.
Method ε = 1.0 ε = 2.0 ε = 4.0
FedAvg (non-private)82.5 ± 0.2
DP-SGD63.1 ± 0.870.3 ± 0.575.6 ± 0.4
AdaClip-DP66.4 ± 0.772.8 ± 0.577.3 ± 0.5
DP-FedAvg64.5 ± 0.671.5 ± 0.576.0 ± 0.4
FedDP62.8 ± 0.669.4 ± 0.573.8 ± 0.4
ADP-SIR (ours)69.7 ± 0.476.9 ± 0.380.1 ± 0.2
Table 6. RESISC45 accuracy (%) across privacy budgets ( δ = 10 5 ). Mean ± std over 3 runs.
Table 6. RESISC45 accuracy (%) across privacy budgets ( δ = 10 5 ). Mean ± std over 3 runs.
Method ε = 1.0 ε = 2.0 ε = 4.0
FedAvg (non-private)85.0 ± 0.2
DP-SGD52.4 ± 0.659.1 ± 0.465.2 ± 0.3
AdaClip-DP54.7 ± 0.561.5 ± 0.666.0 ± 0.4
DP-FedAvg53.9 ± 0.560.2 ± 0.566.1 ± 0.4
FedDP52.1 ± 0.558.3 ± 0.464.0 ± 0.4
ADP-SIR (ours)57.6 ± 0.364.8 ± 0.370.4 ± 0.2
Table 7. Centralized training cost on EuroSAT (single NVIDIA V100). Per-epoch and memory are relative to DP-SGD; total time = (per-epoch time) × (epochs to target). Mean ± std over 3 runs.
Table 7. Centralized training cost on EuroSAT (single NVIDIA V100). Per-epoch and memory are relative to DP-SGD; total time = (per-epoch time) × (epochs to target). Mean ± std over 3 runs.
MethodPer-Epoch Time (rel.)Epochs to TargetTotal Time (rel.)Memory (rel.)
DP-SGD 1.00 × ± 0.01 60 ± 1 60.0 × ± 1.2 1.00 × ± 0.01
ADP-SIR 1.12 × ± 0.01 45 ± 1 50.4 × ± 1.1 1.08 × ± 0.01
Table 8. Federated training cost and communication (EuroSAT, 10 clients, 5 local epochs/round). Per-epoch time is averaged across clients and reported relative to DP-SGD centralized. Bytes are per round (uplink/downlink). Mean ± std over 3 runs.
Table 8. Federated training cost and communication (EuroSAT, 10 clients, 5 local epochs/round). Per-epoch time is averaged across clients and reported relative to DP-SGD centralized. Bytes are per round (uplink/downlink). Mean ± std over 3 runs.
MethodPer-Epoch Time (rel.)Rounds to TargetTotal Time (rel.)Uplink (MB/rd)Downlink (MB/rd)
DP-FedAvg 1.15 × ± 0.02 110 ± 5 63.3 × ± 2.5 42 ± 142 ± 1
FedDP 1.18 × ± 0.02 116 ± 6 68.4 × ± 3.1 46 ± 246 ± 2
Table 9. Overhead breakdown for ADP-SIR vs. DP-SGD on EuroSAT (per-epoch wall time share, %). Mean ± std over 3 runs.
Table 9. Overhead breakdown for ADP-SIR vs. DP-SGD on EuroSAT (per-epoch wall time share, %). Mean ± std over 3 runs.
ComponentDP-SGDADP-SIRDeltaNotes
Forward/Backward pass88.0 ± 0.583.6 ± 0.4 4.4 Slightly amortized by fewer epochs
Gradient clipping6.0 ± 0.25.5 ± 0.2 0.5 Vectorized implementation
LSP (per-layer EMA)0.03.6 ± 0.1 + 3.6 Layerwise sensitivity profiling
CGNS (metrics/noise)0.02.1 ± 0.1 + 2.1 Gradient stability tracking ϕ t
RDP accounting6.0 ± 0.15.2 ± 0.1 0.8 Shared across layers/batches
Table 10. Ablation results on EuroSAT at ε = 2.0 . Accuracy averaged over 3 runs.
Table 10. Ablation results on EuroSAT at ε = 2.0 . Accuracy averaged over 3 runs.
VariantTest Accuracy (%)
ADP-SIR (full)76.9
w/o CGNS73.0
w/o LSP72.4
w/o CGNS and LSP70.3
Table 11. Ablation results on EuroSAT at ε { 1.0 , 2.0 } . Accuracy averaged over 3 runs.
Table 11. Ablation results on EuroSAT at ε { 1.0 , 2.0 } . Accuracy averaged over 3 runs.
VariantTest Accuracy (%)
ε = 1.0 ε = 2.0
ADP-SIR (full)73.876.9
w/o CGNS69.573.0
w/o LSP68.872.4
w/o CGNS and LSP66.170.3
Table 12. Additional evaluation metrics at ε = 2.0 .
Table 12. Additional evaluation metrics at ε = 2.0 .
VariantPrecision (%)Recall (%)F1-Score (%)
ADP-SIR (full)77.276.576.8
w/o CGNS73.472.873.1
w/o LSP72.672.172.3
w/o CGNS and LSP70.869.970.3
Table 13. Per-class accuracy (%) at ε = 2.0 for selected EuroSAT classes.
Table 13. Per-class accuracy (%) at ε = 2.0 for selected EuroSAT classes.
VariantSeaLakeAnnualCropIndustrialForest
ADP-SIR (full)78.477.176.879.0
w/o CGNS74.375.072.976.5
w/o LSP71.874.270.175.8
w/o CGNS and LSP69.771.968.474.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Z.; Yan, X.; Chen, G.; Tian, X. Adaptive Differential Privacy for Satellite Image Recognition with Convergence-Guaranteed Optimization. Electronics 2025, 14, 3680. https://doi.org/10.3390/electronics14183680

AMA Style

Yang Z, Yan X, Chen G, Tian X. Adaptive Differential Privacy for Satellite Image Recognition with Convergence-Guaranteed Optimization. Electronics. 2025; 14(18):3680. https://doi.org/10.3390/electronics14183680

Chicago/Turabian Style

Yang, Zhijie, Xiaolong Yan, Guoguang Chen, and Xiaoli Tian. 2025. "Adaptive Differential Privacy for Satellite Image Recognition with Convergence-Guaranteed Optimization" Electronics 14, no. 18: 3680. https://doi.org/10.3390/electronics14183680

APA Style

Yang, Z., Yan, X., Chen, G., & Tian, X. (2025). Adaptive Differential Privacy for Satellite Image Recognition with Convergence-Guaranteed Optimization. Electronics, 14(18), 3680. https://doi.org/10.3390/electronics14183680

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop