Next Article in Journal
ANCHOR-Grid: Authenticating Smart Grid Digital Twins Using Real-World Anchors
Previous Article in Journal
RT-4M: Real-Time Mosaicing Manager for Manual Microscopy System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ENGDM: Enhanced Non-Isotropic Gaussian Diffusion Model for Progressive Image Editing †

School of Mathematics and Statistics, Xi’an Jiaotong University, Xi’an 710049, China
*
Author to whom correspondence should be addressed.
This paper is an extended version of our paper published in the conference paper: Yu, X.; Gu, X.; Liu, H.; Sun, J. Constructing Non-isotropic Gaussian Diffusion Model Using Isotropic Gaussian Diffusion Model for Image Editing. In Proceedings of Advances in Neural Information Processing Systems 36 (NeurIPS 2023), New Orleans, LA, USA, 10–16 December 2023.
Sensors 2025, 25(10), 2970; https://doi.org/10.3390/s25102970
Submission received: 7 March 2025 / Revised: 17 April 2025 / Accepted: 28 April 2025 / Published: 8 May 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Diffusion models have made remarkable progress in image generation, leading to advancements in the field of image editing. However, balancing editability with faithfulness remains a significant challenge. Motivated by the fact that more novel content will be generated when larger variance noise is applied to the image, in this paper, we propose an Enhanced Non-isotropic Gaussian Diffusion Model (ENGDM) for progressive image editing, which introduces independent Gaussian noise with varying variances to each pixel based on its editing needs. To enable efficient inference without retraining, ENGDM is rectified into an isotropic Gaussian diffusion model (IGDM) by assigning different total diffusion times to different pixels. Furthermore, we introduce reinforced text embeddings, using a novel editing reinforcement loss in the latent space to optimize text embeddings for enhanced editability. And we introduce optimized noise variances by employing a structural consistency loss to dynamically adjust the denoising time steps for each pixel for better faithfulness. Experimental results on multiple datasets demonstrate that ENGDM achieves state-of-the-art performance in image-editing tasks, effectively balancing faithfulness to the source image and alignment with the desired editing target.

1. Introduction

Image editing [1,2,3,4,5] has attracted extensive attention in recent years, thanks to the development of diffusion models [6,7,8,9,10,11]. This task aims to edit a source image into a target image based on a given target text prompt while preserving high faithfulness to the source image in regions unrelated to the editing task.
Existing image-editing methods are primarily classified into three categories: mask based, inversion based, and attention based. Mask-based methods [4,12,13,14] leverage masks to guide the sampling process, ensuring that editing is performed only in the mask region. Inversion-based methods [3,15,16,17,18,19] map real images to noisy latents through an inversion process and then use a sampling process to generate edited images based on the given target text. Attention-based methods [2,20,21,22] modify the attention layers of the U-Net network in Stable Diffusion models to preserve the features of the source image. While these approaches have significantly advanced the capabilities of image editing, achieving an optimal balance between editability and faithfulness remains a challenge. For example, as shown in Figure 1, NMG [19] applies insufficient editing, failing to achieve the desired editing task. In contrast, iCD [23] tends to over-edit, resulting in significant deviations between the edited and source images.
To address these challenges, we propose an Enhanced Non-isotropic Gaussian Diffusion Model (ENGDM) for progressive image editing, which is extended from our conference paper NGDM [24] published in NeurIPS. Diffusion models are empirically known to generate more diverse and novel content when larger variance noise is added, while preserving the image content when smaller variance noise is applied [1]. Motivated by this, we employ a non-isotropic Gaussian diffusion model (NGDM) to add independent Gaussian noises with different variances to different image pixels. The variance is determined by the degree to which each pixel needs to be edited, allowing precise control over the editing process. To avoid retraining the score model for ENGDM, we rectify the NGDM to implement ENGDM within the framework of IGDM. Each pixel is added an equal amount of noise at each step, but different pixels accumulate noise over varying total time steps.
To further enhance the editability of the model, we incorporate reinforced text embeddings during the sampling process. We design a novel editing reinforcement loss operating in the latent space to optimize the text embeddings, guiding the model to accurately generate the target objects specified in the textual description. Meanwhile, to maintain high faithfulness to the source image, we optimize noise variances by using a structural consistency loss to dynamically adjust the denoising time steps for each pixel, ensuring the preservation of fine-grained details of the source image. Lastly, we design a specialized sampling strategy that seamlessly integrates ENGDM with a pre-trained IGDM (e.g., Stable Diffusion [25]), enabling high-quality image editing.
Our contributions are summarized as follows:
  • We propose ENGDM, a novel method for progressive image editing. We introduce reinforced text embeddings, using a novel editing reinforcement loss in the latent space to optimize text embeddings for enhanced editability.
  • We propose the optimized noise variances by employing a structural consistency loss to dynamically adjust the denoising time steps for each pixel, ensuring high faithfulness to the source image.
  • Extensive experiments on multiple datasets demonstrate that ENGDM achieves state-of-the-art performance in image-editing tasks, achieving a better balance between editability and faithfulness.
This paper extends our conference version NGDM [24] published in NeurIPS, in which we devised the non-isotropic Gaussian diffusion process for image editing. In this journal version, we make additional contributions with reinforced text embeddings, optimized noise variances, a refined sampling algorithm, and further performance improvements.
The rest of the paper is structured as follows. We summarize the related works in Section 2. Section 3 presents the background of the Gaussian diffusion model. Section 4 introduces our ENGDM method. Section 5 discusses the experimental results. Section 6 concludes this paper.

2. Related Work

Image editing aims to modify a user-provided source image to align with a given target prompt while minimizing visual changes to the source image. We summarize the related works in Table 1.
Mask-based image editing. Mask-based image-editing methods leverage masks to guide and refine the sampling process. These approaches enable models to focus on precisely modifying localized regions to align with the editing target. Blended Diffusion [13] and Blended Latent Diffusion [14] employ user-provided masks to blend the noisy latents from the forward noising process of the Denoising Diffusion Probabilistic Model (DDPM) [8] with the noisy latents from the denoising process, thereby restricting modifications to the specified regions. PFB-Diff [26] performs blending not directly on intermediate noisy latents but instead on feature maps. PFB-Diff seamlessly integrates generated content guided by the target prompt through multi-level feature blending and introduces an attention-masking mechanism in the cross-attention layers to improve editing performance. Without requiring user-provided masks, DiffEdit [4] automatically generates a mask by contrasting the predictions of a diffusion model conditioned on different text prompts, highlighting the regions in the source image that need to be edited. RDM [27] introduces a novel region generator model that employs a new CLIP-guided loss to learn how to identify the image-editing regions. These methods rely on predefined editing regions and use a fixed-size mask throughout the entire denoising process. In contrast, our approach progressively increases the editable regions during the denoising process, enabling progressive and dynamic image editing. This strategy effectively mitigates edge artifacts and enhances faithfulness to the source image, addressing the limitations of hard mask-guided approaches.
Inversion-based image editing. Inversion-based image-editing methods invert the real image into the initial noise, enabling the generation of edited results based on this initial point and a given target prompt. The pioneering research DDIM Inversion [28] proposes deterministic inversion with the discretization of diffusion ODE to encode the source image into noisy variables to preserve the source image information. However, under the classifier-free guidance, DDIM suffers from significant accumulated error, resulting in reconstruction failures. Subsequent works [3,18,19,29,30] mitigate the accumulated error by approximating the inversion trajectory. For instance, Null-Text Inversion [3] optimizes the null text embedding by minimizing the discrepancy between the ideal and actual intermediate latents, while PnP Inversion [18] introduces correction terms to achieve the same goal. Another group of works [17,31,32] adopt stochastic differential equation (SDE) instead of ODE, reducing errors by manipulating random noise. SDE-Drag [32] demonstrates that stochastic inversion outperforms deterministic inversion in editing performance, and the KL divergence between the distribution of edited image and prior data distribution decreases in stochastic inversion while remaining in deterministic inversion. Additionally, some works [16,33,34] establish mathematically exact inversion. EDICT [16] achieves mathematically exact inversion by tracking two noisy variables in each step during inversion, which can be derived from each other in the sampling time.
Attention-based image editing. The attention layers in the U-Net network of Stable Diffusion play a significant role in determining the layout of the generated image. Attention-based methods achieve image editing by manipulating the attention layers. P2P [2] demonstrates that the cross-attention maps govern the spatial layout and structure of the generated image. Accordingly, it preserves the structure of the source image by replacing the cross-attention maps during the denoising process with those from the reconstruction process. Inspired by P2P [2], subsequent works [35,36,37,38] leverage cross-attention maps for maintaining the structure of the source image. Several methods [20,22,39,40] perform image editing by modifying self-attention maps instead of cross-attention maps. FPE [22] highlights that modifying cross-attention maps may lead to editing failures, emphasizing that self-attention maps are crucial for the success of image-editing tasks. Beyond modifying attention maps, MasaCtrl [21] and PnP [20] enforce consistency between the edited and source images by directly modifying specific attention features.
Differently, our method performs progressive image editing by introducing independent Gaussian noises with varying variances to different pixels, guided by a weighting matrix with soft weights.
Table 1. Summarization of related works.
Table 1. Summarization of related works.
Editing MethodMask-BasedInversion-BasedAttention-Based
StudyBlended Diffusion [13], Blended Latent Diffusion [14], PFB-Diff [26], DiffEdit [4], RDM [27]DDIM Inversion [28], Null-Text Inversion [3], PnP Inversion [18], NMG [19], PTI [29], ProxEdit [30], DDPM Inversion [17], LEDITS++ [31], SDE-Drag [32], EDICT [16], BELM [33]P2P [2], Pix2Pix-Zero [35], Custom-edit [36], Conditional Score Guidance [37], PnP [20], FPE [22], Photoswap [39], StyleInjection [40], MasaCtrl [21]
PurposeMask-based image-editing methods leverage masks to guide and refine the sampling process.Inversion-based image-editing methods invert the real image into noise space, and then use the sampling process to generate the edited results based on the noisy latent and a given target prompt.Attention-based methods achieve image editing by manipulating the attention layers.
LimitationThe method exhibits limited flexibility when handling complex modifications.The inversion process is time-consuming and may hinder practical applications.It is challenging to accurately locate the specific regions that require editing.
PerformanceThe faithfulness of non-edited regions is high, but edge artifacts are prone to occur.The details of the source image can be effectively preserved, but easy to fail editing in complex scenarios.The details of the source image are not precisely preserved.

3. Background: Score-Based Diffusion Models

Diffusion models [6,7,8,9,10] are a family of generative models that learn the data distribution based on the Gaussian process. Two representative models are the Denoising Diffusion Probabilistic Model (DDPM) [8] and Score Matching with Langevin Dynamics (SMLD) [7]. We discuss the details based on DDPM for the remainder of the paper for brevity.
Given the input data x ( 0 ) R D , which represents a sample from the data distribution p d a t a , a forward process produces the noisy x ( t ) indexed by a time variable t [ 0 , 1 ] via
x ( t ) = α ¯ ( t ) x ( 0 ) + 1 α ¯ ( t ) z ( t ) ,
where z ( t ) N ( 0 , I ) for any t and α ¯ ( t ) = e 0 t β ( s ) d s controlling the noise schedule. β ( s ) = β ¯ min + s ( β ¯ max β ¯ min ) with β ¯ min = 0.1 and β ¯ max = 20 [8,9]. This type of diffusion model is dubbed IGDM since the added Gaussian noise z ( t ) is from the independently and identically distributed normal distribution.
DDPM is in the framework of SDE [7] with variance preservation
d x ( t ) = 1 2 β ( t ) x ( t ) d t + β ( t ) d w ,
where w is the standard Wiener process and the initial value of the above SDE is x ( 0 ) . The reverse process denoises the noisy sample x ( T ) starting from T using a reverse SDE
d x ( t ) = 1 2 β ( t ) x ( t ) β ( t ) x log p t ( x ( t ) ) d t + β ( t ) d w ¯ ,
where w ¯ is a standard Wiener process when time flows backward from T to 0 and the initial value of the above SDE is x ( T ) . The score function x log p t ( x ) is approximated by training a time-dependent model s θ ( x ( t ) , t , C ) under condition C via score matching [9,41]. For inference, the time of the differential equation is discretized as t 0 , Δ t , 2 Δ t , , T , with Δ t representing the sampling time interval. We choose to utilize the reverse process of DDIM for sampling. With  β t = β ( t ) Δ t and α ¯ t = s = 0 t ( 1 β s ) , the deterministic iteration rule of DDIM [28] is
x ( t ) = α ¯ t x ^ ( 0 , t + Δ t ) + 1 α ¯ t s θ ( x ( t + Δ t ) , t + Δ t , C ) ,
where x ^ ( 0 , t + Δ t ) is the prediction of the initial data point x ( 0 ) at time t and is derived as
x ^ ( 0 , t + Δ t ) = x ( t + Δ t ) 1 α ¯ t + Δ t s θ ( x ( t + Δ t ) , t + Δ t , C ) α ¯ t + Δ t .

4. Method

In this section, we introduce the ENGDM for progressive image editing. We first construct the NGDM in Section 4.1, then we rectify the non-isotropic Gaussian diffusion process in Section 4.2. To further enhance the editability and faithfulness, we propose reinforced text embeddings in Section 4.3 and optimized noise variances in Section 4.4. Finally, we design a novel sampling algorithm in Section 4.5. The overview of ENGDM is shown in Figure 2.

4.1. Non-Isotropic Gaussian Diffusion Model

We construct the NGDM by adding the non-isotropic Gaussian noise in the input data y ( 0 ) R D and y ( 0 ) p d a t a , and the noises associated with different pixels are independent. The forward SDE of NGDM [24] is
d y ( t ) = 1 2 β ( t ) Λ ( I ) y ( t ) d t + β ( t ) Λ ( I ) d w ,
where I R D is the source data, Λ ( I ) : R D R D × D is the weighting matrix, defined as diagonal matrix Λ ( I ) = diag λ 1 , , λ D with 0 λ k 1 scaling the Gaussian noise level added to the k-th pixel. The initial value of the above SDE is y ( 0 ) . Note that the transition kernel p 0 t ( y ( t ) | y ( 0 ) ) = N y ( t ) y ( 0 ) e 1 2 0 t β ( s ) Λ ( I ) d s , I e 0 t β ( s ) Λ ( I ) d s is an independent Gaussian distribution.

4.2. Rectify the Non-Isotropic Gaussian Diffusion Model

With the added independent noise, we next discuss the forward SDE for NGDM [24] in scalar form for each pixel k. Given the initial y k ( 0 ) denoting the value of pixel k in y ( 0 ) , the forward SDE of the k-th pixel can be presented by
d y k ( t ) = 1 2 β ( t ) λ k y k ( t ) d t + β ( t ) λ k d w ,
where w is a one-dimensional Wiener process and the initial value of the SDE is y k ( 0 ) .
We present Lemma 1 and Theorem 1 proposed in our conference version [24] to illustrate the connection between the NGDM defined in Section 4.1 and the IGDM defined in Section 3 at the pixel level.
Lemma 1.
Let β ( s ) = β ¯ min + s ( β ¯ max β ¯ min ) , with  β ¯ max > β ¯ min > 0 . Then, for each λ k [ 0 , 1 ] and t [ 0 , 1 ] , there exists a unique time τ [ 0 , 1 ] (denoted by τ = ξ k ( t ) ) such that 0 t β ( s ) λ k d s = 0 τ β ( s ) d s and β ( t ) λ k d t = β ( τ ) d τ , with the following form
ξ k ( t ) = β ¯ min β ¯ max β ¯ min + ( β ¯ min β ¯ max β ¯ min ) 2 + 2 β ¯ min t λ k β ¯ max β ¯ min + t 2 λ k .
Based on the above Lemma, we can rectify the NGDM, which adds noise at each pixel with varying variance over the same time span, into an IGDM that adds noise at each pixel with the same noise variance but with a different total diffusion time for different pixels. We introduce the following theorem to derive the differential equation as an IGDM.
Theorem 1.
For a pixel indexed by k, λ k [ 0 , 1 ] , and let τ = ξ k ( t ) with ξ k ( t ) represented in Equation (8). With the same initial value y k ( 0 ) , we have that the transition kernel at time t induced by Equation (7) is equal to the transition kernel at time τ induced by the following differential equation:
d y k ( t ) = 1 2 β ( τ ) y k ( t ) d τ + β ( τ ) d w .
The initial condition of the above SDE is y k ( 0 ) . The total time of noising for Equation (9) is T k with T k = ξ k ( T ) .
Inspired by this, we rectify the reverse process in NGDM with different speeds of denoising across pixels to be the reverse process with consistent speed but different total time of denoising. We propose rectifying the differential equation for the reverse process within the NGDM framework into the following form:
d y k ( t ) = 1 2 β ( τ ) y k ( t ) β ( τ ) ( y log p τ ( y ( t ) ) ) k d τ + β ( τ ) d w ¯ ,
where w ¯ is a one-dimensional Wiener process when time flows backward from T k to 0, and the initial value of the above SDE is y k ( T k ) . Theorem 1 establishes the conclusion that the NGDM in Equation (7) can be rectified to the IGDM in Equation (9) but with different total diffusion time T k for different pixel indexed by k, determined based on Equation (8). This inspires us to utilize the pre-trained IGDM to achieve the data-sampling of NGDM for image editing. Subsequently, we propose adjusting the total time of noising and denoising for each pixel k to T k , enabling the utilization of pre-trained IGDM for data sampling.
For image-editing tasks, we use the source image I as y ( 0 ) and generate noisy data y ( T ) through the forward process. We generate the edited image y ^ ( 0 ) by denoising from y ( T ) . Utilizing the forward noising process of IGDM, we add independent noise to each pixel k to obtain the noisy observation x k ( t ) of discrete time t 0 , Δ t , , T with Δ t representing the sampling time interval
x k ( t ) = α ¯ t I k + 1 α ¯ t z k ( t ) ,
where z ( t ) N ( 0 , I ) . Next, with  H ( y ( t + Δ t ) , t + Δ t ) denoting the sampling procedure of DDIM given in Equation (4) of Section 3, the data-sampling iteration utilizing the IGDM model with initial value y k ( T ) = x k ( T ) is defined as
y k ( t ) = M k ( t ) H k ( y ( t + Δ t ) , t + Δ t ) + ( 1 M k ( t ) ) x k ( t ) ,
where M k ( t ) = I ( t < T k } ) . This implies that we use the noisy observation x k ( t ) to represent y k ( t ) at each step before T k with t T k , rather than the actual denoised result starting from time step T. Until time T k , we begin the denoising from x k ( T k ) for the k-th pixel. In such a way, different pixels have different starting time steps ( T k for k-th pixel) for image denoising in the data-sampling process.
We next combine all pixels to perform the denoising process using the complete data. We introduce a time-dependent mask M ( t ) to control the denoising state of each pixel. Specifically, M ( t ) determines which pixels begin denoising at time step t and which remain unprocessed. The mask M ( t ) is defined based on the relationship between the denoising start time step T k of each pixel and the current time step t. Formally, M ( t ) can be defined as M ( t ) = [ m 1 ( t ) , , m D ( t ) ] with
m k = 0 if T k t 1 if T k > t .
The k-th element of M ( t ) , denoted m k ( t ) , indicates the denoising state of the k-th pixel at time t. m k ( t ) = 1 means that the pixel has already started denoising at time step t, while m k ( t ) = 0 indicates that denoising has not yet started.
Based on the relationship between the noise variance and the time step given by Equation (8) in Lemma 1, we can compute the noise variance corresponding to the time step t, denoted as λ ¯ ( t ) . Then, m k ( t ) can be equivalently represented as
m k ( t ) = 0 if λ k λ ¯ ( t ) 1 if λ k > λ ¯ ( t ) .
Equipped with M ( t ) , we can then obtain the iteration rule of complete noisy observation y ( t ) in the following form:
y ( t ) = M ( t ) H ( y ( t + Δ t ) , t + Δ t ) + ( 1 M ( t ) ) x ( t ) .
The pixels of y ( t ) begin the denoising process at different time steps.

4.3. Enhancing Editability with Reinforced Text Embeddings

To enable sufficient editing, we incorporate the reinforced text embedding optimization strategy into the diffusion process. This strategy builds upon the classifier-free guidance technique [42], which removes the need for a separate classifier to guide diffusion models. Formally, let ⊘ be the embedding of a null text “ ” and let w be the guidance scale, then the classifier-free guidance prediction [42] is defined by
s ˜ θ ( y ( t ) , t , C ) = s θ ( y ( t ) , t , ) + w · s θ ( y ( t ) , t , C ) .
The guidance scale w controls the degree of alignment between the generated image and the textual condition. A larger w increases the strength of the conditioning, improving the alignment with the text prompt but at the cost of the visual quality of the generated image.
We propose the reinforced text embedding optimization strategy to strengthen the editing intensity without sacrificing the visual quality of the generated image. The key idea behind the reinforced text embeddings is to optimize the text embeddings by using an optimization strategy that pulls the latent closer to the reinforced one. This reinforced latent is achieved by using a larger guidance scale, which increases the model’s focus on aligning the image with the text prompt. Instead of directly increasing the guidance scale, we optimize the text embeddings by pushing the latent closer to those under stronger guidance. The reinforced text embeddings C is optimized using the following editing reinforcement loss:
L edit = y ( t ) y ( t ) 2 2 .
The noisy latent y ( t ) at step t is obtained from Equation (15), which combines the denoised result H ( y ( t + Δ t ) , t + Δ t ) with guidance scale w and the noisy observation x ( t ) from the forward process. The denoised result y ( t ) at the same step is obtained using a larger guidance scale w with w > w . Unlike some methods [27,43] that use CLIP loss [44] to align images and target text prompt, which makes the optimization process difficult and resource-consuming, our proposed editing reinforcement loss can be optimized directly in the latent space in an efficient way. The loss function for optimizing the text embeddings is applied only during the early stages of the denoising process, as these stages are more inclined toward generating diverse content.
By aligning y ( t ) with the denoised results produced under stronger guidance conditions, the model learns reinforced text embeddings, which effectively enhance its editing capabilities. This approach not only enhances the editing performance of the model but also mitigates the need for excessively high guidance scales, thereby mitigating potential visual artifacts.

4.4. Enhancing Faithfulness with Optimized Noise Variances

Our method relies on the weighting matrix Λ ( I ) defined in NGDM [24], which determines the extent of the noise variance applied to each pixel based on the editing requirements. We first initialize the weighting matrix A ( I ) based on DiffEdit [4]. Following DiffEdit, we determine the editing degree of each pixel by analyzing the differences between score estimates produced by the diffusion model under different text conditions. Specifically, given the source image I , the source prompt R of the source image, and the target prompt Q that describes the desired target image after editing, we add noise to the source image up to the 0.5 T step and use the texts R and Q as the conditions, respectively, for denoising in the current time step to estimate the score values by using the score network s θ . We derive the degree map of editing based on the absolute difference of the estimated scores. We use the above method to compute 10 estimated absolute noise differences by running 10 times with different random seeds, averaging and performing Gaussian smoothing on the averaged map to obtain the final degree map A ( I ) .
The pixel with the larger value in the degree map should be added with the noise with larger variance. We define the weighting matrix Λ ( I ) by applying a sigmoid function to the degree map A ( I ) , i.e.,  Λ ( I ) = 1 1 + exp ( a A ( I ) + b ) , with a and b denoting the hyperparameters for the transformation. By substituting each pixel’s noise variance and the maximum diffusion time step T into Equation (8), we can obtain the total diffusion time for each pixel. We discuss the effect of hyperparameters a and b on the generation of images in Section 5.3.2 and validate the effectiveness of the method for determining the total diffusion time step of each pixel in Section 5.3.5.
Precise editing relies on accurate estimation of Λ ( I ) . In practice, we observe that editing failures are most commonly caused by assigning excessively large variances to pixels that do not require editing, leading to unnecessary modifications in non-edit regions, thus reducing the faithfulness of the generated image. To address this limitation, we further propose an optimization-based approach to refine Λ ( I ) , dynamically correcting the weighting matrix in the later stages of the diffusion process. To obtain the denoised result of step t, we use Λ * ( I ) and Equation (13) to determine the mask M ( t ) for the current time step, and then use Equation (15) to obtain the updated denoising result y ( t ) . We optimize Λ * ( I ) by constraining the structural similarity between the denoising result H ( y ( t + Δ t ) , t + Δ t ) and the noisy observation x ( t ) in the forward process. To enable gradient-based optimization, we approximate the mask M ( t ) used in Equation (15) with a sigmoid function. We minimize the mean squared error (MSE) between the structural similarity of the edited and source images to optimize Λ * ( I ) . The optimization objective is
L faithful = S ( H ( y ( t + Δ t ) , t + Δ t ) ) S ( x ( t ) ) 2 2 + ρ Λ * ( I ) Λ ( I ) 2 2 ,
where S ( · ) represents the structure self-similarity [45] function with [ S ( z ) ] i , j = z i z j T z i z j , i and j are the pixel indices. The  Λ * ( I ) Λ ( I ) 2 2 is the regularization term, and ρ is the coefficient of the regularization term.

4.5. Sampling Method in ENGDM

Based on the above method, we specify our sampling algorithm by harnessing the power of a pre-trained IGDM. We generate the edited image with the source image I as a condition. By utilizing DDIM inversion, we first add noise to the source image I to T time steps, and then use the method in Section 4.2 to rectify NGDM into IGDM to denoise the image. Additionally, we learn reinforced text embeddings and optimized noise variances. We show the sampling algorithm of ENGDM in Algorithm 1.
Algorithm 1 Sampling method of ENGDM.
Inputs: The source image I , the time schedule { β ( t ) } t = 0 T , the maximal time steps T, the optimization time T s , the guidance scale w and w
 1:
Obtain the initial weighting matrix Λ ( I )
 2:
Obtain the forward noisy observation x ( 0 ) , , x ( T ) using DDIM inversion over the source image I
 3:
Initialize y ( T ) x ( T )
 4:
for  t = T to T s  do
 5:
      Calculate the variance λ ( t ) corresponding to the time step t using Equation (8)
 6:
      Obtain M ( t ) using Λ ( I ) and λ ( t ) by Equation (13)
 7:
      Obtain the denoised result H ( y ( t + Δ t ) , t + Δ t ) by Equation (4) and guidance scale w
 8:
      Update y ( t ) by y ( t ) M ( t ) H ( y ( t + Δ t ) , t + Δ t ) + ( 1 M ( t ) ) x ( t )
 9:
      if  t > T s  then
10:
            Obtain the denoised result y ( t ) by Equation (4) and guidance scale w
11:
            Optimize the text embeddings using L edit defined in Equation (17)
12:
      else
13:
            Optimize the weighting matrix using L faithful defined in Equation (18)
14:
      end if
15:
end for
Output: Generated image y conditioned on the source image I

5. Experiment

5.1. Experimental Setup

Datasets. As summarized in Table A1, we evaluate the performance of our method on four diverse datasets, including PIE [18], ZONE [46], Imagen [47] and EMU [48]. The PIE [18] dataset consists of 700 images spanning 10 different editing types, including “objects change”, “attributes change”, “add”, “remove”, etc. The images are categorized into four content groups: animals, humans, indoor scenes, and outdoor scenes. The PIE dataset provides a diverse and challenging set of tasks for evaluating editing capabilities. The ZONE [46] dataset contains 100 samples, including 60 real images sourced from the internet and 40 synthetic images. This dataset focuses on three primary editing operations: 32 images for “add”, 54 images for “change”, and 14 images for “remove”. The Imagen dataset is a synthetic dataset that contains 180 images collected from https://imagen.research.google/ (accessed on 10 April 2025), which are generated by the method of Imagen [47]. Editing prompts are constructed by replacing parts of the original text. Each image undergoes 10 different attribute replacements, resulting in a total of 1800 test examples derived from the original 180 images. The EMU [48] dataset is a more extensive dataset containing 3314 test images across seven categories of image-editing tasks. To ensure high-quality test samples, EMU Edit applies a post-validation phase to filter out low-quality examples. We collect images from the high-quality HQ-Edit dataset [49] to form a new benchmark referred to as the HC dataset. The HC dataset contains 958 images, each with the height and width larger than 1024. During optimization and evaluation, all images are resized to a resolution of 1024 × 1024. These datasets comprehensively cover diverse images and editing tasks, ensuring a robust evaluation of our method.
Implementation details. We utilize the pre-trained Stable Diffusion v1.4 as the base model to implement our method. For the HC dataset, we conduct experiments on NVIDIA H20 GPUs equipped with 96 GB memory. All other experiments are conducted on a single NVIDIA GeForce RTX 4090 GPU and using PyTorch framework. For the denoising process, we adopt the DDIM sampler with 50 sampling steps, and we set T s = 25 . We set the guidance scale w to 7.5 and w = 2 w . We set the coefficient ρ = 0.01 . We can adjust the hyperparameters a and b to flexibly balance the editability and faithfulness. For all qualitative comparisons presented in this paper, we set a = 10.0 and b = 5.0 . These values are empirically chosen to achieve robust performance across various editing tasks. We conduct additional analysis to investigate the effect of different values of hyperparameters in the experimental results. During optimization, to accelerate convergence and stabilize the optimization process, we employ the AdamW optimizer with the learning rate initialized to 0.1 and decaying linearly over time steps.
Evaluation metrics. To comprehensively evaluate our method, we employ multiple quantitative metrics, following the evaluation setup of PnP Inversion [18]. These metrics include Structure Distance, PSNR, LPIPS, MSE, SSIM, and CLIP Score, assessing different aspects of editing performance. Structure Distance is computed using deep spatial features extracted by DINO-ViT [50], which measures structural similarity between the edited image and the source image. PSNR, LPIPS, MSE, and SSIM evaluate the content similarity between the edited image and the source image. PSNR, MSE, and SSIM focus on pixel-level similarity, while LPIPS assesses perceptual differences using deep feature representations. The CLIP Score measures text–image consistency by evaluating the cosine similarity between the edited image and the corresponding target prompt using the CLIP model [44]. Unlike PnP Inversion, which computes metrics only for specific regions of the source and edited images, we calculate metrics across the entire images. We argue that evaluating faithfulness should consider the edited regions alongside non-edited regions. For example, when editing a cat into a dog, the faithfulness of the edited image improves if the edited dog inherits certain characteristics of the source cat, such as similar fur color or posture, while maintaining semantic consistency with the target prompt.

5.2. Results

In this section, we comprehensively evaluate our method on different datasets. We compare our method with other state-of-the-art image-editing methods, including P2P [2], DiffEdit [4], InstructPix2Pix [51], MasaCtrl [21], NMG [19], PnP Inversion [18], ZONE [46], FPE [22], InfEdit [52], CDS [53], iCD [23], and NGDM [24].

5.2.1. Qualitative Results

Figure 3, Figure 4, Figure 5 and Figure 6 show qualitative comparisons between our method ENGDM and the other baseline methods. In each figure, the first row shows the source image, the second row displays the editing task, and the subsequent rows show the edited images generated by different baseline methods, NGDM, and ENGDM.
Attention-based methods, which manipulate the attention maps in the U-Net network for editing, struggle to precisely control the editing regions. From Figure 3, both P2P and MasaCtrl often lose details in regions that do not require editing, while FPE often fails to achieve the desired edits. For example, in the sixth column, P2P and MasaCtrl successfully perform editing but fail to preserve the original structure of the image, whereas FPE fails to remove the rose, resulting in unsuccessful editing. In contrast, NGDM and ENGDM exhibit better performance. ENGDM further enhances both the faithfulness and editability of NGDM. For instance, in the second column, ENGDM generates a sea that appears more beach-like, while in the fifth column, ENGDM preserves the white cheeks of the original rabbit, retaining finer details of the source image.
Inversion-based methods typically invert the source image into a noisy latent space for editing, but these methods are prone to cumulative errors, leading to editing failures in complex tasks. For instance, in columns 3, 4, 6, 7, and 8 of Figure 4, several inversion-based methods fail to achieve the desired modifications. In contrast, NGDM and ENGDM successfully perform the editing tasks. For the editing tasks in columns 2 and 3, ENGDM outperforms NGDM in maintaining structural consistency with the source image.
Mask-based methods, such as DiffEdit, restrict the editing regions using user-provided or automatically generated masks but often produce edge artifacts. For example, in the first column of Figure 5, DiffEdit generates unnatural artifacts at the edges of the hat, while in the sixth column, the edited flower region appears inconsistent with the rest of the image. Additionally, methods like ZONE and iCD struggle to maintain consistency with the source image in most cases, and CDS fails to perform the editing tasks in columns 4, 5, 7, and 8. In comparison, ENGDM strikes a better balance between consistency with the source image and editability. In addition, as shown in Figure 6, we can see that ENGDM demonstrates superior performance even on the more challenging HC dataset.

5.2.2. Quantitative Results

In this section, we evaluate the performance of ENGDM on four datasets using multiple quantitative metrics and compare ENGDM with multiple baseline methods. We use Structure Distance, PSNR, LPIPS, MSE, and SSIM to assess faithfulness, and the editing score calculated by the CLIP model to assess editability. NGDM and ENGDM allow flexible control of the editability and faithfulness by adjusting the hyperparameters a and b. To better compare with baseline methods, we provide results for two different versions of ENGDM.
From Table 2, Table 3, Table 4, Table 5 and Table 6, we observe that when a = 10.0 and b = 5.0 , ENGDM achieves the highest editing score across all datasets, with values of 25.97, 25.43, 34.45, 25.75, and 27.63, respectively. As shown in Table 2, Table 3, Table 4 and Table 5, ENGDM significantly outperforms iCD in terms of faithfulness. For example, from Table 2, it can be seen that on the PIE dataset, ENGDM outperforms iCD in terms of Structure Distance/PSNR/LPIPS/MSE/SSIM, with values of 18.80/19.94/146.45/132.19/71.18, compared to 39.43/17.81/235.93/203.94/62.39 for iCD, demonstrating the better faithfulness of ENGDM. As shown in Table 2, Table 3, Table 4, Table 5 and Table 6, compared to NGDM, ENGDM shows further improvements in both the editing score and all evaluation metrics for faithfulness. For example, from Table 2, it can be seen that on the PIE dataset, ENGDM achieves an editing score of 25.97, which is higher than that of NGDM at 25.84. And the Structure Distance/PSNR/LPIPS/MSE/SSIM values of ENGDM are 18.80/19.94/146.45/132.19/71.18, which outperform those of NGDM at 21.32/19.31/159.84/139.30/69.37.
When a = 10.0 , b = 10.0 , as shown in Table 2, Table 3 and Table 4 and Table 6, ENGDM achieves the best Structure Distance/PSNR/LPIPS/MSE/SSIM values on the PIE, ZONE, Imagen, and HC datasets, demonstrating the best faithfulness. As shown in Table 5, on the EMU dataset, the Structure Distance/PSNR/LPIPS/MSE/SSIM values of ENGDM are comparable to those of CDS, but ENGDM achieves an editing score of 23.41, outperforming CDS of 22.71. ENGDM outperforms NGDM on all datasets. For example, from Table 2, it can be seen that ENGDM achieves an editing score of 24.92 on the PIE dataset, higher than that of NGDM at 24.65. Furthermore, in terms of faithfulness, the Structure Distance/PSNR/LPIPS/MSE/SSIM values of ENGDM are 6.55/23.98/74.64/48.96/79.84/24.92, which outperform those of NGDM at 7.35/23.40/82.94/57.03/76.08. Overall, combining both the qualitative and quantitative results, ENGDM achieves a better balance between editability and faithfulness compared to NGDM and other baseline methods, resulting in higher-quality edited images.

5.2.3. User Study

We conduct a user study to evaluate the performance of ENGDM on the datasets. We query 40 participants, and each participant is provided with 40 randomly selected source images along with the corresponding editing results generated by different methods. The images generated by our method and the comparison methods are displayed in random order to the participants. Participants are suggested to select the image that best apply the requested editing while preserving the most details of the original image. The percentage of votes for our method compared to the other methods are shown in Table 7, which demonstrates that the participants exhibit a strong preference towards our method.

5.3. Ablation Study

5.3.1. Ablation Analysis of ENGDM

In this section, we conduct an ablation study to evaluate the contributions of the two key techniques introduced to NGDM: reinforced text embeddings (RTEs) and optimized noise variances (ONVs). Figure 7 illustrates the results of this ablation study, showcasing the impact of each technique on image-editing performance. RTE enhances the ability of the model to align the generated image with the target prompt by reinforcing the text embeddings during the denoising process. As illustrated in the fourth column of Figure 7, incorporating RTE into NGDM significantly improves editability. For example, in the first row, the result with RTE successfully removes the dandelions compared to NGDM. ONV, on the other hand, focuses on improving the preservation of the source image details by dynamically adjusting the noise variances during the denoising process. As seen in the fifth column, ONV improves the faithfulness of the generated image. For instance, in the third row, the result with ONV better preserves the shape of the wooden barrel. ENGDM, which combines both RTE and ONV, enhances both editability and faithfulness.

5.3.2. Effect of Hyperparameters a and b

As mentioned in Section 4.4, we transform the attention map A ( I ) into weighting matrix Γ ( I ) with hyperparameters a and b. We control the initial time step of denoising each pixel by adjusting hyperparameters a and b. To investigate the impact of a and b on the performance of ENGDM, we conduct the ablation study on the ZONE dataset. We perform experiments by varying a and b while fixing one of them.
As shown in Table 8, when b = 5.0 and a is varied, we observe that increasing a improves editability at the cost of reduced faithfulness. For instance, ENGDM achieves Structure Distance/PSNR/LPIPS/MSE/SSIM values of 6.01/26.05/55.63/36.26/82.43 when a = 6.0 , indicating strong faithfulness. However, ENGDM achieves Structure Distance/PSNR/LPIPS/MSE/SSIM values of 27.09/18.14/169.52/190.53/70.08 as a increases to 14.0, reflecting a decline in faithfulness. Meanwhile, the CLIP Score increases from 23.89 to 25.61, demonstrating improved editability. This trend suggests that higher values of a enhance the ability of the model to align with the target prompt but may compromise the preservation of the source image details.
Conversely, when a = 10.0 and b is varied, increasing b improves faithfulness while reducing editability. For example, when b = 3.0 , ENGDM achieves a CLIP Score of 25.68, indicating strong editability, but the faithfulness is reduced. As b increases to 7.0, the Structure Distance decreases to 6.78, and the PSNR improves to 23.76, indicating enhanced faithfulness. However, the CLIP Score decreases to 24.56, suggesting reduction in editability.
Overall, these results highlight that there is a trade-off between editability and fidelity that can be controlled by tuning hyperparameters a and b. By adjusting these parameters, ENGDM offers flexibility in balancing the preservation of the source image and the effectiveness of the edits.

5.3.3. Comparison with Hard Weighting Matrix

To further analyze the effectiveness of the soft weighting matrix Λ ( I ) used in ENGDM, we compare the results with the method using the hard weighting matrix applied at different threshold values. As shown in Figure 8, the first column displays the source image, and the second column shows the results generated by ENGDM with the soft weighting matrix, along with the corresponding heatmap. The heatmap visually represents the weighting values, where brighter regions correspond to higher noise variance, which indicates greater editability, while darker regions correspond to lower noise variance, indicating greater faithfulness to the original image. In contrast, columns 3 to 7 present the results generated using hard weighting matrices with threshold of 0.1, 0.3, 0.5, 0.7, and 0.9.
During the denoising process, methods guided by a hard mask consistently edit regions exceeding the threshold. ENGDM utilizes soft weights to gradually denoise each pixel based on its required degree of editing. As shown in Figure 8, hard mask-guided methods often produce artifacts at the edges of the mask, leading to unnatural transitions between edited and non-edited regions. Additionally, hard mask-guided methods tend to over-edit the targeted regions, reducing faithfulness to the source image. In comparison, ENGDM generates more natural images without the edge artifacts, effectively achieving the desired edits while maintaining high faithfulness to the source image.

5.3.4. Results at the Intermediate Steps of the Forward and Reverse Process

Figure 9 presents the evolution of the image during the forward and reverse processes when editing a cat into a dog, along with the masks M used at each step of the reverse process as described in Algorithm 1. The mask M dynamically indicates which pixels have undergone denoising (black regions) and which pixels remain unprocessed (white regions). At the early stages of the reverse process, ENGDM prioritizes denoising the regions that require more editing, such as the prominent features of the cat, including the eyes and nose as indicated by the white areas in the mask. As the process progresses, denoising extends to the facial region of the cat and eventually to the background. This reflects the ability of ENGDM to gradually denoise different pixels based on their required editing degree. The progressive denoising method enables ENGDM to achieve precise effective editing while better preserving the details of the original image.

5.3.5. Validation of the Method for Determining the Total Diffusion Time Step

To validate the effectiveness of this method, we visualize the determined total diffusion time steps in Figure 10 of the response letter. As can be seen, when editing a cat into a dog, the prominent features of the cat, such as the eyes and nose, are assigned larger total diffusion time steps, whereas the background is assigned smaller total diffusion time steps. This demonstrates that our method can successfully assign the total diffusion time step for each pixel based on the degree of required editing. Moreover, as seen in the edited image in the third row of Figure 10, our method effectively edits the cat into a dog while preserving similarity to the source image, which further validates the effectiveness of the method. In addition, we also experiment with a random strategy for assigning total diffusion time to each pixel. Table 9 presents a quantitative comparison of the two approaches on the PIE dataset. The results show that our method achieves a better balance, while the random method significantly decreases the editing score.

6. Conclusions

In this paper, we propose ENGDM for progressive image editing. ENGDM is constructed by adding independent Gaussian noises with varying variances to different image pixels. To avoid retraining, we rectify ENGDM by assigning different total diffusion times to different pixels, thereby implementing progressive editing within an isotropic framework. Furthermore, we enhance editability by leveraging reinforced text embeddings and improve faithfulness through optimization of noise variances. Extensive experimental results demonstrate the effectiveness of ENGDM across multiple datasets. While the optimization procedure brings significant performance improvements, it also introduces computational overhead. Furthermore, the performance of ENGDM relies on the underlying diffusion model. Failures in the generation process of the underlying model can lead to failures in the editing results. In the future, we plan to extend ENGDM to more advanced diffusion models and explore its applications in broader domains such as video editing.

Author Contributions

X.Y. contributed to the introduction of the method, evaluation, and writing of the original draft; X.G. and X.H. contributed to the manuscript revisions. J.S. worked on advising and revising the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Key-Area Research and Development Program of Guangdong Province (2022B0303020003), NSFC (12426313, 12326615, 623B2084), China National Postdoctotal Program for Innovative Talents (BX20240276) and the Fundamental Research Funds for the Central Universities, China (xzy022025047).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data are available upon request.

Conflicts of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as potential conflicts of interest.

Appendix A. Mathematical Definitions of Evaluation Metrics

To comprehensively evaluate our method, we employ multiple quantitative metrics, following the evaluation setup of PnP Inversion [18]. These metrics include Structure Distance, PSNR, LPIPS, MSE, SSIM, and CLIP Score, assessing different aspects of editing performance. Let x , y , C denote the source image, edited image, the source prompt and the target prompt, respectively; the mathematical definitions of these metrics are as follows.
(1)
Structure Distance (SD) utilizes spatial features extracted by DINO-ViT [50] to compute structural alignment:
SD = f DINO ( x ) f DINO ( y ) 2 2 ,
where f DINO represents the deep spatial features extracted from the input image using the DINO model.
(2)
Peak Signal-to-Noise Ratio (PSNR) quantifies the maximum signal-to-noise ratio based on the mean squared error:
PSNR = 10 · log 10 MAX I 2 MSE ( x , y ) ,
where MAX I is the maximum possible pixel value and MSE is the mean squared error.
(3)
Learned Perceptual Image Patch Similarity (LPIPS) assesses perceptual differences using deep feature representations:
LPIPS = l w l · f l ( x ) f l ( y ) 2 2 ,
where f l denotes the features from layer l of a pre-trained network and w l is the learned weights for the l-th layer.
(4)
Mean Squared Error (MSE) evaluates pixel-level reconstruction quality by averaging the squared differences between images:
MSE = 1 N i = 1 N ( x i y i ) 2 ,
where N denotes the total number of pixels and i denotes the index of the pixel.
(5)
Structural Similarity Index (SSIM) measures the structural similarity:
SSIM = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 ) ,
where μ , σ , and σ x y denote the mean, standard deviation, and covariance of x and y , respectively. C 1 = ( k 1 L ) 2 , C 2 = ( k 2 L ) 2 with L is the dynamic range and k 1 = 0.01 , k 2 = 0.03 .
(6)
CLIPScore evaluates text–image alignment via cosine similarity in CLIP space:
CLIPScore = f I ( y ) · f T ( C ) f I ( y ) f T ( C ) ,
where f I ( y ) and f T ( C ) are CLIP embeddings of image y and C .

Appendix B. Table for Describing the Evaluation Dataset

We provide Table A1 to clearly describe all evaluation datasets. This table includes key details such as the dataset name, description, source, size, and editing type.
Table A1. Evaluation datasets.
Table A1. Evaluation datasets.
Dataset NameDescriptionSourceSizeEditing Type
PIE [18]PIE dataset grouped into animals, humans, indoor, and outdoor scenes, offering diverse, challenging tasks for evaluating editing performance.https://github.com/cure-lab/PnPInversion (accessed on 10 April 2025)700Add, change, and remove.
ZONE [46]Zone dataset includes real and synthetic images, focusing on three editing types.https://drive.google.com/file/d/1lAwpENoDcO1QyFuwz3iKJJ7DmDTFMvIU/view (accessed on 10 April 2025)100Add, change, and remove.
Imagen [47]Imagen dataset contains 180 synthetic images, which are generated by Imagen [47]. Each image is edited with 10 attribute replacements via prompt modification, yielding 1800 evaluation examples.https://imagen.research.google/ (accessed on 10 April 2025)1800Change.
EMU [48]EMU is a large-scale benchmark, including images spanning seven editing categories. A post-validation step filters out low-quality samples to ensure high data quality.https://huggingface.co/datasets/facebook/emu_edit_test_set (accessed on 10 April 2025)3314Add, change, and remove.
HCWe collect images with the height and width larger than 1024 from the high-quality HQ-Edit dataset [49] to form a new benchmark, which is referred to as the HC dataset.https://huggingface.co/datasets/UCSC-VLAA/HQ-Edit (accessed on 10 April 2025) Add, change, and remove.

Appendix C. Method for Hyperparameters Selection

We propose a systematic approach for hyperparameter selection. The main idea is to iteratively select better parameters based on a balance metric that considers both editability and fidelity. Specifically, let y , C x , and C y denote the edited image, source prompt, and target prompt, respectively. We design the balance metric using the CLIP model as follows:
D = S t S s = f I ( y ) · f T C y f I ( y ) f T C y f I ( y ) · f T C x f I ( y ) f T C x
where f I ( y ) , f T ( C y ) , and f T ( C x ) are CLIP embeddings of y , C x , and C y , respectively. S t represents the cosine similarity between the edited image and the target prompt, measuring editability. And S s represents the cosine similarity between the edited image and the source prompt, measuring faithfulness. We aim to select parameters that strike a better balance between editability and faithfulness. We start by adjusting b within the range [ 1.0 , 10.0 ] . After generating the edited image each time, we compute the balance metric D , and adjust b using b + D . If D > 0 , b + D increases the value of b to enhance faithfulness. If D < 0 , b + D decreases the value of b to enhance editability. If b exceeds the range [ 1.0 , 10.0 ] within the maximum iteration count, we proceed to adjust a using a D . The iteration stops when the maximum iteration count is reached or when D = 0 . We conduct experiments using the proposed parameter selection method (ENGDM-A) on the PIE and ZONE datasets to validate its effectiveness. The results are shown in Table A2. As can be seen, the automatic parameter selection method based on the search metric achieves a better balance with significantly improved faithfulness and a comparable editing score.
Table A2. Quantitative comparison results of ENGDM and ENGDM-A methods. Evaluated using the PIE and ZONE datasets. Different metrics are scaled. The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
Table A2. Quantitative comparison results of ENGDM and ENGDM-A methods. Evaluated using the PIE and ZONE datasets. Different metrics are scaled. The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
DatasetMethodStructureContent PreservationEditing
DistancePSNRLPIPSMSESSIMScore
PIEENGDM18.8019.94146.45132.1971.1825.87
ENGDM-A13.4121.41111.77100.4373.4925.82
ZONEENGDM16.7520.17122.75119.5874.7325.43
ENGDM-A12.6421.43100.7491.3877.7225.39

References

  1. Meng, C.; Song, Y.; Song, J.; Wu, J.; Zhu, J.Y.; Ermon, S. Sdedit: Image synthesis and editing with stochastic differential equations. In Proceedings of the International Conference on Learning Representations, Virtual, 25–29 April 2022. [Google Scholar]
  2. Hertz, A.; Mokady, R.; Tenenbaum, J.; Aberman, K.; Pritch, Y.; Cohen-Or, D. Prompt-to-prompt image editing with cross attention control. In Proceedings of the International Conference on Learning Representations, Virtual, 25–29 April 2022. [Google Scholar]
  3. Mokady, R.; Hertz, A.; Aberman, K.; Pritch, Y.; Cohen-Or, D. Null-text Inversion for Editing Real Images using Guided Diffusion Models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  4. Couairon, G.; Verbeek, J.; Schwenk, H.; Cord, M. Diffedit: Diffusion-based semantic image editing with mask guidance. In Proceedings of the International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  5. Li, D.; Li, J.; Hoi, S. Blip-diffusion: Pre-trained subject representation for controllable text-to-image generation and editing. In Proceedings of the Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 9–15 December 2024. [Google Scholar]
  6. Sohl-Dickstein, J.; Weiss, E.; Maheswaranathan, N.; Ganguli, S. Deep unsupervised learning using nonequilibrium thermodynamics. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015. [Google Scholar]
  7. Song, Y.; Ermon, S. Generative modeling by estimating gradients of the data distribution. In Proceedings of the Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  8. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. In Proceedings of the Conference on Neural Information Processing Systems, Virtual, 6–12 December 2020. [Google Scholar]
  9. Song, Y.; Sohl-Dickstein, J.; Kingma, D.P.; Kumar, A.; Ermon, S.; Poole, B. Score-based generative modeling through stochastic differential equations. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 3–7 May 2021. [Google Scholar]
  10. Kingma, D.; Salimans, T.; Poole, B.; Ho, J. Variational diffusion models. In Proceedings of the Conference on Neural Information Processing Systems, Virtual, 6–14 December 2021. [Google Scholar]
  11. Nichol, A.Q.; Dhariwal, P. Improved denoising diffusion probabilistic models. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021. [Google Scholar]
  12. Nichol, A.; Dhariwal, P.; Ramesh, A.; Shyam, P.; Mishkin, P.; McGrew, B.; Sutskever, I.; Chen, M. Glide: Towards photorealistic image generation and editing with text-guided diffusion models. In Proceedings of the International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022. [Google Scholar]
  13. Avrahami, O.; Lischinski, D.; Fried, O. Blended diffusion for text-driven editing of natural images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  14. Avrahami, O.; Fried, O.; Lischinski, D. Blended latent diffusion. ACM Trans. Graph. 2023, 42, 1–11. [Google Scholar] [CrossRef]
  15. Gal, R.; Alaluf, Y.; Atzmon, Y.; Patashnik, O.; Bermano, A.H.; Chechik, G.; Cohen-Or, D. An image is worth one word: Personalizing text-to-image generation using textual inversion. In Proceedings of the International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  16. Wallace, B.; Gokul, A.; Naik, N. Edict: Exact diffusion inversion via coupled transformations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  17. Huberman-Spiegelglas, I.; Kulikov, V.; Michaeli, T. An edit friendly ddpm noise space: Inversion and manipulations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  18. Ju, X.; Zeng, A.; Bian, Y.; Liu, S.; Xu, Q. Pnp inversion: Boosting diffusion-based editing with 3 lines of code. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  19. Cho, H.; Lee, J.; Kim, S.B.; Oh, T.H.; Jeong, Y. Noise map guidance: Inversion with spatial context for real image editing. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  20. Tumanyan, N.; Geyer, M.; Bagon, S.; Dekel, T. Plug-and-play diffusion features for text-driven image-to-image translation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  21. Cao, M.; Wang, X.; Qi, Z.; Shan, Y.; Qie, X.; Zheng, Y. Masactrl: Tuning-free mutual self-attention control for consistent image synthesis and editing. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Wangfu, France, 2–6 October 2023. [Google Scholar]
  22. Liu, B.; Wang, C.; Cao, T.; Jia, K.; Huang, J. Towards Understanding Cross and Self-Attention in Stable Diffusion for Text-Guided Image Editing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  23. Starodubcev, N.; Khoroshikh, M.; Babenko, A.; Baranchuk, D. Invertible Consistency Distillation for Text-Guided Image Editing in Around 7 Steps. In Proceedings of the Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 9–15 December 2024. [Google Scholar]
  24. Yu, X.; Gu, X.; Liu, H.; Sun, J. Constructing non-isotropic Gaussian diffusion model using isotropic Gaussian diffusion model for image editing. In Proceedings of the Conference on Neural Information Processing Systems, New Orleans, LA, USA, 10–16 December 2023. [Google Scholar]
  25. Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  26. Huang, W.; Tu, S.; Xu, L. Pfb-diff: Progressive feature blending diffusion for text-driven image editing. Neural Netw. 2025, 181, 106777. [Google Scholar] [CrossRef] [PubMed]
  27. Lin, Y.; Chen, Y.W.; Tsai, Y.H.; Jiang, L.; Yang, M.H. Text-driven image editing via learnable regions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  28. Song, J.; Meng, C.; Ermon, S. Denoising diffusion implicit models. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 3–7 May 2021. [Google Scholar]
  29. Dong, W.; Xue, S.; Duan, X.; Han, S. Prompt tuning inversion for text-driven image editing using diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Wangfu, France, 2–6 October 2023. [Google Scholar]
  30. Han, L.; Wen, S.; Chen, Q.; Zhang, Z.; Song, K.; Ren, M.; Gao, R.; Stathopoulos, A.; He, X.; Chen, Y.; et al. Proxedit: Improving tuning-free real image editing with proximal guidance. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 4–8 January 2024. [Google Scholar]
  31. Brack, M.; Friedrich, F.; Kornmeier, K.; Tsaban, L.; Schramowski, P.; Kersting, K.; Passos, A. Ledits++: Limitless image editing using text-to-image models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  32. Nie, S.; Guo, H.A.; Lu, C.; Zhou, Y.; Zheng, C.; Li, C. The blessing of randomness: Sde beats ode in general diffusion-based image editing. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  33. Wang, F.; Yin, H.; Dong, Y.; Zhu, H.; Zhang, C.; Zhao, H.; Qian, H.; Li, C. BELM: Bidirectional Explicit Linear Multi-step Sampler for Exact Inversion in Diffusion Models. In Proceedings of the Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 9–15 December 2024. [Google Scholar]
  34. Hong, S.; Lee, K.; Jeon, S.Y.; Bae, H.; Chun, S.Y. On Exact Inversion of DPM-Solvers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  35. Parmar, G.; Kumar Singh, K.; Zhang, R.; Li, Y.; Lu, J.; Zhu, J.Y. Zero-shot image-to-image translation. In Proceedings of the ACM SIGGRAPH, Los Angeles, CA, USA, 6–10 August 2023. [Google Scholar]
  36. Choi, J.; Choi, Y.; Kim, Y.; Kim, J.; Yoon, S. Custom-edit: Text-guided image editing with customized diffusion models. arXiv 2023, arXiv:2305.15779. [Google Scholar]
  37. Lee, H.; Kang, M.; Han, B. Conditional score guidance for text-driven image-to-image translation. In Proceedings of the Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 9–15 December 2024. [Google Scholar]
  38. Guo, Q.; Lin, T. Focus on your instruction: Fine-grained and multi-instruction image editing by attention modulation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  39. Gu, J.; Wang, Y.; Zhao, N.; Fu, T.J.; Xiong, W.; Liu, Q.; Zhang, Z.; Zhang, H.; Zhang, J.; Jung, H.; et al. Photoswap: Personalized subject swapping in images. In Proceedings of the Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 9–15 December 2024. [Google Scholar]
  40. Chung, J.; Hyun, S.; Heo, J.P. Style injection in diffusion: A training-free approach for adapting large-scale diffusion models for style transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  41. Vincent, P. A connection between score matching and denoising autoencoders. Neural Comput. 2011, 23, 1661–1674. [Google Scholar] [CrossRef] [PubMed]
  42. Ho, J.; Salimans, T. Classifier-free diffusion guidance. In Proceedings of the Conference on Neural Information Processing Systems Workshop on Deep Generative Models and Downstream Applications, Virtual, 13 December 2021. [Google Scholar]
  43. Kwon, G.; Ye, J.C. Diffusion-based image translation using disentangled style and content representation. In Proceedings of the International Conference on Learning Representations, Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  44. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021. [Google Scholar]
  45. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. IP 2004, 13, 600–612. [Google Scholar]
  46. Li, S.; Zeng, B.; Feng, Y.; Gao, S.; Liu, X.; Liu, J.; Li, L.; Tang, X.; Hu, Y.; Liu, J.; et al. Zone: Zero-shot instruction-guided local editing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  47. Saharia, C.; Chan, W.; Saxena, S.; Li, L.; Whang, J.; Denton, E.L.; Ghasemipour, K.; Gontijo Lopes, R.; Karagol Ayan, B.; Salimans, T.; et al. Photorealistic text-to-image diffusion models with deep language understanding. In Proceedings of the Conference on Neural Information Processing Systems, New Orleans, LA, USA, 28 November–9 December 2022. [Google Scholar]
  48. Sheynin, S.; Polyak, A.; Singer, U.; Kirstain, Y.; Zohar, A.; Ashual, O.; Parikh, D.; Taigman, Y. Emu edit: Precise image editing via recognition and generation tasks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  49. Hui, M.; Yang, S.; Zhao, B.; Shi, Y.; Wang, H.; Wang, P.; Zhou, Y.; Xie, C. Hq-edit: A high-quality dataset for instruction-based image editing. arXiv 2024, arXiv:2404.09990. [Google Scholar]
  50. Caron, M.; Touvron, H.; Misra, I.; Jégou, H.; Mairal, J.; Bojanowski, P.; Joulin, A. Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021. [Google Scholar]
  51. Brooks, T.; Holynski, A.; Efros, A.A. Instructpix2pix: Learning to follow image editing instructions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  52. Xu, S.; Huang, Y.; Pan, J.; Ma, Z.; Chai, J. Inversion-free image editing with natural language. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  53. Nam, H.; Kwon, G.; Park, G.Y.; Ye, J.C. Contrastive denoising score for text-guided latent diffusion image editing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
Figure 1. Concrete examples illustrating the challenge in achieving a balance between editability and faithfulness.
Figure 1. Concrete examples illustrating the challenge in achieving a balance between editability and faithfulness.
Sensors 25 02970 g001
Figure 2. The overview of our ENGDM. Firstly, isotropic noises are added to the source image until T time steps. Then, in the reverse denoising process, different pixels begin the denoising process at different time steps as indicated by the orange blocks and lines. Additionally, in the early stages of denoising, the learned reinforced text embeddings are incorporated to enhance editability. In the later stages, the optimized noise variance is incorporated to improve fidelity.
Figure 2. The overview of our ENGDM. Firstly, isotropic noises are added to the source image until T time steps. Then, in the reverse denoising process, different pixels begin the denoising process at different time steps as indicated by the orange blocks and lines. Additionally, in the early stages of denoising, the learned reinforced text embeddings are incorporated to enhance editability. In the later stages, the optimized noise variance is incorporated to improve fidelity.
Sensors 25 02970 g002
Figure 3. Qualitative comparison with attention-based methods.
Figure 3. Qualitative comparison with attention-based methods.
Sensors 25 02970 g003
Figure 4. Qualitative comparison with inversion-based methods.
Figure 4. Qualitative comparison with inversion-based methods.
Sensors 25 02970 g004
Figure 5. Qualitative comparison with mask-based methods and other methods.
Figure 5. Qualitative comparison with mask-based methods and other methods.
Sensors 25 02970 g005
Figure 6. Qualitative comparison on the HC dataset.
Figure 6. Qualitative comparison on the HC dataset.
Sensors 25 02970 g006
Figure 7. Ablation study of ENGDM. The first column shows the source images, the second column describes the editing tasks, and the third column displays the results generated by NGDM. The fourth/fifth columns show the results obtained by adding only the RTE/ONV technique to NGDM. The final column presents the complete ENGDM, incorporating both RTE and ONV. Red boxes highlight the improvements of ENGDM over NGDM.
Figure 7. Ablation study of ENGDM. The first column shows the source images, the second column describes the editing tasks, and the third column displays the results generated by NGDM. The fourth/fifth columns show the results obtained by adding only the RTE/ONV technique to NGDM. The final column presents the complete ENGDM, incorporating both RTE and ONV. Red boxes highlight the improvements of ENGDM over NGDM.
Sensors 25 02970 g007
Figure 8. Edited images and heatmaps with soft and hard weighting matrices. The images in the second column represent the results generated by our method and the heatmap below the image depicts the weighting matrix Λ ( I ) defined in Section 4.1 in the paper. A color bar is presented in the left to illustrate the corresponding colors for different values. The images in columns 3–7 represent the results generated using the hard weighting matrix with threshold value in { 0.1 , 0.3 , 0.5 , 0.7 , 0.9 } obtained by Λ ( I ) = I ( A ( I ) η ) , where η is a threshold chosen in { 0.1 , 0.3 , 0.5 , 0.7 , 0.9 } . The heatmaps below the images represent the binary hard weighting matrix.
Figure 8. Edited images and heatmaps with soft and hard weighting matrices. The images in the second column represent the results generated by our method and the heatmap below the image depicts the weighting matrix Λ ( I ) defined in Section 4.1 in the paper. A color bar is presented in the left to illustrate the corresponding colors for different values. The images in columns 3–7 represent the results generated using the hard weighting matrix with threshold value in { 0.1 , 0.3 , 0.5 , 0.7 , 0.9 } obtained by Λ ( I ) = I ( A ( I ) η ) , where η is a threshold chosen in { 0.1 , 0.3 , 0.5 , 0.7 , 0.9 } . The heatmaps below the images represent the binary hard weighting matrix.
Sensors 25 02970 g008
Figure 9. Results at the intermediate steps of the forward process and reverse process. We present the results at the intermediate steps of the forward and reverse process, along with the mask M at the intermediate steps during the denoising, which is defined in Equation (13). The white regions in the mask image indicate the pixels that have undergone the denoising process, while the black regions represent the pixels that have not yet been denoised.
Figure 9. Results at the intermediate steps of the forward process and reverse process. We present the results at the intermediate steps of the forward and reverse process, along with the mask M at the intermediate steps during the denoising, which is defined in Equation (13). The white regions in the mask image indicate the pixels that have undergone the denoising process, while the black regions represent the pixels that have not yet been denoised.
Sensors 25 02970 g009
Figure 10. Visualization of the total diffusion time step for each pixel. The first row shows the source images. The second row visualizes the total diffusion time step of each pixel, where brighter colors indicate a larger total diffusion time step. The third row presents the edited images generated under the editing task of editing a cat into a dog.
Figure 10. Visualization of the total diffusion time step for each pixel. The first row shows the source images. The second row visualizes the total diffusion time step of each pixel, where brighter colors indicate a larger total diffusion time step. The third row presents the edited images generated under the editing task of editing a cat into a dog.
Sensors 25 02970 g010
Table 2. Quantitative comparisons in image editing. Evaluated using the PIE dataset. Different metrics are scaled. The best results are highlighted in bold while the second best results are marked with an underline. The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
Table 2. Quantitative comparisons in image editing. Evaluated using the PIE dataset. Different metrics are scaled. The best results are highlighted in bold while the second best results are marked with an underline. The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
MethodStructureContent PreservationEditing
DistancePSNRLPIPSMSESSIMScore
P2P [2]69.9515.10335.58347.1855.3624.98
DiffEdit [4]17.4119.66129.14131.3472.4325.09
InstructPix2Pix [51]57.9416.71269.02419.3061.7223.57
MasaCtrl [21]28.0819.09181.21147.2867.7223.90
NMG [19]15.3723.39112.16160.1273.4823.57
PnP Inversion [18]11.7122.26116.4076.1673.2924.84
ZONE [46]58.2716.20281.79396.6658.9723.98
FPE [22]12.7721.67114.9582.8673.4224.35
InfEdit [52]19.4721.49133.87176.3970.7824.74
CDS [53]7.3323.8376.4857.2776.7923.91
iCD [23]39.4317.81235.93203.9462.3925.92
NGDM (a = 10.0, b = 5.0)21.3219.31159.84139.3069.3725.84
ENGDM (a = 10.0, b = 5.0)18.8019.94146.45132.1971.1825.97
NGDM (a = 10.0, b = 10.0)7.3523.4082.9457.0376.0824.65
ENGDM (a = 10.0, b = 10.0)6.5523.9874.6448.9679.8424.92
Table 3. Quantitative comparisons in image-editing tasks. Evaluated using the ZONE dataset. Different metrics are scaled. The best results are highlighted in bold while the second best results are marked with an underline. The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
Table 3. Quantitative comparisons in image-editing tasks. Evaluated using the ZONE dataset. Different metrics are scaled. The best results are highlighted in bold while the second best results are marked with an underline. The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
MethodStructureContent PreservationEditing
DistancePSNRLPIPSMSESSIMScore
P2P [2]57.9616.00265.69286.0159.9224.19
DiffEdit [4]14.8720.39104.78117.7177.2924.94
InstructPix2Pix [51]33.8618.70189.72296.2569.9924.19
MasaCtrl [21]24.2019.78151.36133.6272.9823.71
NMG [19]15.9523.6195.19105.5978.2323.01
PnP Inversion [18]11.3622.4994.3574.2977.9124.20
ZONE [46]34.6017.58204.14295.3267.8924.64
FPE [22]11.4122.3790.1174.0278.3023.48
InfEdit [52]15.7421.69106.04152.2374.2124.18
CDS [53]6.9124.4963.5257.1181.7923.72
iCD [23]32.8517.76205.44198.6164.5025.23
NGDM (a = 10.0, b = 5.0)17.9319.97130.55124.5774.1125.12
ENGDM (a = 10.0, b = 5.0)16.7520.17122.75119.5874.7325.43
NGDM (a = 10.0, b = 10.0)6.2724.1464.5748.0081.4423.95
ENGDM (a = 10.0, b = 10.0)5.6424.6957.8441.3683.8224.36
Table 4. Quantitative comparisons in image-editing tasks. Evaluated using the Imagen dataset. Different metrics are scaled. The best results are highlighted in bold while the second best results are marked with an underline. The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
Table 4. Quantitative comparisons in image-editing tasks. Evaluated using the Imagen dataset. Different metrics are scaled. The best results are highlighted in bold while the second best results are marked with an underline. The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
MethodStructureContent PreservationEditing
DistancePSNRLPIPSMSESSIMScore
P2P [2]56.3814.07284.17405.3956.5532.99
DiffEdit [4]13.0519.5492.93121.2978.5133.60
InstructPix2Pix [51]59.3512.66336.20690.2556.5925.89
MasaCtrl [21]20.4018.27158.14160.8772.0132.62
NMG [19]7.8722.0480.8176.9378.6332.82
PnP Inversion [18]8.9521.1282.3987.8077.8933.13
ZONE [46]60.3512.67344.80617.5855.2325.28
FPE [22]9.1820.5685.4493.3376.6232.78
InfEdit [52]6.5522.2866.2070.6480.9232.66
CDS [53]8.9520.4475.63111.8578.7332.87
iCD [23]31.4216.49212.42242.4566.0234.33
NGDM (a = 10.0, b = 5.0)15.7318.89121.07138.8175.1234.23
ENGDM (a = 10.0, b = 5.0)14.4019.31110.98126.3476.2034.45
NGDM (a = 10.0, b = 10.0)5.8022.3861.3362.0081.0633.24
ENGDM (a = 10.0, b = 10.0)5.0122.9954.8255.2381.7433.46
Table 5. Quantitative comparisons in image-editing tasks. Evaluated using the EMU dataset. Different metrics are scaled. The best results are highlighted in bold while the second best results are marked with an underline. The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
Table 5. Quantitative comparisons in image-editing tasks. Evaluated using the EMU dataset. Different metrics are scaled. The best results are highlighted in bold while the second best results are marked with an underline. The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
MethodStructureContent PreservationEditing
DistancePSNRLPIPSMSESSIMScore
P2P [2]79.1414.02393.78434.6246.0224.97
DiffEdit [4]17.6820.28124.93128.2970.5123.73
InstructPix2Pix [51]48.6817.45236.15353.4861.6222.49
MasaCtrl [21]31.3018.66200.59161.2763.7623.18
NMG [19]11.1526.6978.5873.7476.5521.88
PnP Inversion [18]16.1621.18154.0989.9668.5225.54
ZONE [46]52.2216.44264.65377.3858.0222.70
FPE [22]12.2521.97113.5777.4972.1523.91
InfEdit [52]36.5916.92226.86260.8657.4425.09
CDS [53]5.5325.4859.3238.3677.1222.71
iCD [23]53.9416.25305.91266.6852.1725.71
NGDM (a = 10.0, b = 5.0)31.9418.60210.35191.0562.3325.58
ENGDM (a = 10.0, b = 5.0)27.6119.16189.40165.9564.2625.75
NGDM (a = 10.0, b = 10.0)7.8223.9378.9151.2575.0723.20
ENGDM (a = 10.0, b = 10.0)6.1524.6769.5043.9376.1423.41
Table 6. Quantitative comparisons in image-editing tasks. Evaluated using the HC dataset. Different metrics are scaled. The best results are highlighted in bold while the second best results are marked with an underline. The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
Table 6. Quantitative comparisons in image-editing tasks. Evaluated using the HC dataset. Different metrics are scaled. The best results are highlighted in bold while the second best results are marked with an underline. The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
MethodStructureContent PreservationEditing
DistancePSNRLPIPSMSESSIMScore
P2P [2]90.8411.65427.75779.0330.5524.63
DiffEdit [4]9.4320.2477.34111.5276.3026.94
InstructPix2Pix [51]67.0413.82283.82569.9742.5325.01
PnP Inversion [18]8.6822.8269.1177.9777.3526.24
ZONE [46]68.1413.40334.81594.4343.5124.77
FPE [22]20.1018.84135.29150.5974.4726.31
InfEdit [52]31.3416.42191.28300.0963.6726.37
CDS [53]3.3828.3834.6127.3690.8325.79
NGDM (a = 10.0, b = 5.0)12.1019.54106.02122.5372.8827.36
ENGDM (a = 10.0, b = 5.0)11.1619.9597.95117.1173.7027.63
NGDM (a = 10.0, b = 10.0)3.9130.2034.8427.0390.2725.94
ENGDM (a = 10.0, b = 10.0)3.0230.7230.3522.6292.8426.27
Table 7. User study results on the evaluation datasets. The best results are highlighted in bold while the second best results are marked with an underline.
Table 7. User study results on the evaluation datasets. The best results are highlighted in bold while the second best results are marked with an underline.
P2PDiffEditInstructPix2PixMasaCtrlNMGPnP InversionZONEFPEInfEditCDSiCDNGDMENGDM
0.63%10.31%3.75%2.19%3.13%1.88%6.88%1.88%5.00%6.56%9.06%19.06%29.69%
Table 8. The performance on the ZONE dataset with varying values of a or b while respectively fixing b = 5.0 or a = 10.0 . The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
Table 8. The performance on the ZONE dataset with varying values of a or b while respectively fixing b = 5.0 or a = 10.0 . The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
a ( b = 5.0 )6.08.010.012.014.0
Distance ↓6.0110.1416.7522.7027.09
PSNR ↑26.0522.1820.1718.8918.14
LPIPS ↓55.6388.41122.75150.47169.52
MSE ↓36.2673.49119.58160.75190.53
SSIM ↑82.4378.4274.7372.0270.08
Score ↑23.8924.7325.4325.5625.61
b  ( a = 10 . 0 )3.04.05.06.07.0
Distance ↓34.0724.4516.7510.666.78
PSNR ↑17.5018.6720.1721.8523.76
LPIPS ↓196.95160.33122.7590.1466.39
MSE ↓216.30168.29119.5879.5251.23
SSIM ↑67.1270.9174.7378.3081.13
Score ↑25.6825.6425.4324.8224.56
Table 9. Quantitative comparison results of our and random methods. Evaluated using the ZONE dataset. Different metrics are scaled. The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
Table 9. Quantitative comparison results of our and random methods. Evaluated using the ZONE dataset. Different metrics are scaled. The arrow pointing upwards indicates that a larger value of the metric is better, while the arrow pointing downwards indicates that a smaller value of the metric is better.
MethodStructureContent PreservationEditing
DistancePSNRLPIPSMSESSIMScore
Random13.8321.66114.95109.8175.1623.96
Ours16.7520.17122.75119.5874.7325.43
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, X.; Gu, X.; Hu, X.; Sun, J. ENGDM: Enhanced Non-Isotropic Gaussian Diffusion Model for Progressive Image Editing. Sensors 2025, 25, 2970. https://doi.org/10.3390/s25102970

AMA Style

Yu X, Gu X, Hu X, Sun J. ENGDM: Enhanced Non-Isotropic Gaussian Diffusion Model for Progressive Image Editing. Sensors. 2025; 25(10):2970. https://doi.org/10.3390/s25102970

Chicago/Turabian Style

Yu, Xi, Xiang Gu, Xin Hu, and Jian Sun. 2025. "ENGDM: Enhanced Non-Isotropic Gaussian Diffusion Model for Progressive Image Editing" Sensors 25, no. 10: 2970. https://doi.org/10.3390/s25102970

APA Style

Yu, X., Gu, X., Hu, X., & Sun, J. (2025). ENGDM: Enhanced Non-Isotropic Gaussian Diffusion Model for Progressive Image Editing. Sensors, 25(10), 2970. https://doi.org/10.3390/s25102970

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop