Next Article in Journal
Dark-Field Lau Interferometer: Barker-Babinet Gratings
Previous Article in Journal
A Longitudinal Survey Exploring the Psychological Determinants of Concealed Smartphone Use While Driving: Insights from an Expanding Theory of Planned Behavior
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on High-Resolution Image Harmonization Method Based on Multi-Scale and Global Feature Guidance

1
School of Computer Science, Qinghai Normal University, Xining 810000, China
2
School of Computer and Software, Nanyang Institute of Technology, Nanyang 473000, China
3
Academy of Plateau Science and Sustainability, People’s Government of Oinghai Province & Beijing Normal University, Haihu, Xining 810004, China
4
The State Key Laboratory of Tibetan Intelligent Information Processing and Application, Qinghai Normal University, Hutai, Xining 810008, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2025, 15(19), 10573; https://doi.org/10.3390/app151910573
Submission received: 25 July 2025 / Revised: 18 September 2025 / Accepted: 28 September 2025 / Published: 30 September 2025

Abstract

During the image compositing process, there may be inconsistencies in tone and illumination between the foreground and background, leading to poor visual quality and low realism in the composite images. To address these issues, image harmonization techniques can be employed. This paper proposes an image harmonization method based on multi-scale and global feature guidance (MSGF). In general, images captured in different scenes may exhibit inconsistencies in lighting after composition. The goal of image harmonization is to adjust the foreground illumination to match that of the background. Traditional methods often attempt to blend pixels directly, which can result in unrealistic outcomes. The proposed approach combines multi-scale feature extraction with global feature guidance, forming the MSGF framework. The experiment was conducted on the iHarmony4 dataset. Comparative experiments showed that MSGF achieved the best performance on three subset indicators, including HCOCO. Ablation studies demonstrated the effectiveness of the proposed module. Efficiency evaluation results indicated that it took 0.01s and had 20.9 million parameters, outperforming comparative methods and effectively achieving high-quality image harmonization.

1. Introduction

Image harmonization refers to the process in computer vision and image processing where one or more image elements (such as objects, textures, or colors) are integrated into another image to make the overall image appear more natural and consistent, as shown in Figure 1. This process typically involves adjusting the color, lighting, shadows, and texture of the newly added image elements to match the visual characteristics of the target image. The goal of image harmonization is to address issues such as mismatches in color, brightness, contrast, and texture that may arise when directly compositing images, thereby creating a seamlessly synthesized image that looks like it was taken in the same scene. This technique has widespread applications in digital image editing, film post-production, virtual reality, and augmented reality. Traditional image harmonization methods mainly rely on handcrafted statistical features of images, such as illumination, color temperature, contrast, saturation, etc., to determine color-to-color transformations, ensuring that the foreground tones match those of the background. In recent years, image harmonization methods based on deep learning have emerged continuously and achieved significant progress.
The framework of traditional image harmonization primarily involves transferring the appearance characteristics of the background to the foreground using handcrafted statistical methods. Sunkavalli et al. introduced a multi-scale image harmonization framework [1]. This framework achieves seamless image stitching by decomposing images and noise into pyramids, adjusting the source images and noise at different scales, and then reconstructing the blended image using the modified pyramid coefficients. It incorporates methods based on alpha channels and seamless boundary constraints to ultimately achieve harmonization of the images. This framework demonstrates how realistic composites can be generated across various scenarios with minimal user interaction. Luan et al. [2] achieved the transfer of statistical properties from the background image to the pasted image by utilizing a VGG neural network, preserving the correlations between neural responses across different layers to enhance the quality of the output image. Their process involves two steps: first, achieving coarse harmonization at a single scale, followed by fine-scale refinement based on the initial pass. However, a downside of traditional image harmonization methods is their inefficiency when dealing with large-scale images and their lack of robustness when there are significant differences between images, often relying heavily on specific image features. Xue et al. [3] finally realized the transfer of features from the background image to the foreground image by matching the histogram and adjusting the brightness, hue, and other factors of the image.
In response to the traditional image harmonization methods mentioned above, recent advancements in deep learning and artificial intelligence have provided new solutions for image registration. Neural networks can now automatically learn image features and achieve more efficient and robust registration. However, these approaches also present challenges, such as the need for extensive training data and issues with model generalization.
Deep learning-based image harmonization methods have commonly leveraged the advantages of autoencoders in semantic feature extraction to achieve a consistent and harmonious appearance between the foreground and background. Through this approach, the models can more accurately adjust both local and global visual styles of the image, thereby enhancing the realism and naturalness of the composite image. Sofiiuk et al. [4] proposed a novel architecture that utilizes high-level feature spaces learned by pre-trained classification networks. The model was validated on public datasets and demonstrated promising results. Ke et al. [5] addressed the problem of image harmonization for high-resolution images by transforming the task into an image-level regression problem through filter parameter adjustment. They proposed a framework combining global guidance with detailed awareness for image harmonization. Unlike previous black-box autoencoder-based methods, Harmonizer includes a neural network for filter parameter prediction and several white-box filters for image harmonization. This method is not only applicable to images but also extends to videos. It offers more stable performance, clearer results, and higher efficiency.
In summary, despite their notable successes, these methods still face three key challenges. First, they perform inadequately in handling fine-scale edges (e.g., hair, semi-transparent regions, and object boundaries), often altering the original image content. Second, they incur high computational overhead when processing high-resolution images. Third, these methods struggle in cases where there are specific color mismatches or differing illumination conditions between each synthetic element and the background.
Our contributions are summarized as follows:
(1)
We propose a method that combines multi-scale processing with global feature guidance. Using multi-scale methods, we handle image details at different scales to achieve seamless integration. By introducing global features, we enable transformations of foreground features.
(2)
We introduce a lightweight refinement module to effectively integrate different methods, providing a delicate approach for image harmonization tasks.
(3)
Extensive experiments conducted on the benchmark iHarmony4 dataset demonstrate that our method performs well across all metrics, achieving significant improvements in time while reducing the number of parameters.

2. Related Work

2.1. Image Harmonization

In addition to processing appearance features, traditional methods have also proposed supervised learning-based approaches, achieving significant success in image harmonization by learning from synthetic training pairs (e.g., the iHarmony dataset [6]). DovNet [6] is a novel deep image harmonization method that introduces a new domain verification discriminator. Recently, CNN-based models have been developed for end-to-end image harmonization. Guo et al. [7] introduced the Transformer architecture into the image harmonization task, leveraging the Transformer’s strong capability in modeling long-range contextual dependencies to address the problem of image harmonization. We propose the design of our harmonization Transformer framework without disentanglement, along with comprehensive experiments and ablation studies, demonstrating the power of Transformers and providing insights into their visual behavior in image harmonization. Wang et al. [8] proposed a novel semi-supervised training strategy to address the problem of image harmonization. According to user studies, this method outperforms previous works on established benchmarks and real-world composites, and can interactively process high-resolution images. Meng et al. [9] introduced an adaptive interval color transformation technique, which forecasts pixel-level color transformations and dynamically adjusts the sampling intervals. This approach enables the modeling of local nonlinearities in color transformation at high resolutions. The Poisson blending method proposed by Perez et al. [10] achieves smooth transitions between images by optimizing in the gradient domain, effectively suppressing stitching artifacts, and has been widely applied in image fusion tasks; however, it exhibits limited adaptability to scenes with inconsistent illumination. To improve image harmonization, Tian et al. [11] developed a self-supervised image harmonization framework that enables semantic-aware local style adjustment and global illumination transfer. Usama et al. [12] formulated image harmonization and denoising as an image-to-image translation task, leveraging a generative adversarial network to simultaneously enhance texture consistency and reduce noise. Mwubahimana et al. [13] further introduced a hybrid architecture combining Transformer and CNN, which, under a weakly supervised setting, fully exploits global contextual information to significantly improve harmonization performance in complex scenes. Additionally, Oriti et al. [14] designed an immersive gaming platform to evaluate the integration of virtual and augmented reality, effectively enhancing users’ sense of immersion in virtual environments.
Although the aforementioned methods can achieve image harmonization, they are not ideal in terms of computational efficiency and implementation effectiveness, especially because low-resolution images consume more resources, leading to poorer performance and slower inference speed at high resolutions. In contrast, our method incorporates a multi-scale strategy to effectively and efficiently allocate resources. We formulate the image harmonization task, and our proposed multi-scale and global feature-guided image harmonization approach can address this task at high resolutions with a consistent inference speed, while maintaining performance with negligible degradation.

2.2. Multi-Scale Image Harmonization

Multi-scale image harmonization refers to the process of processing images at different scales to ensure visual harmony across all parts of the image. This technique is commonly used in the fields of image editing and composition, aiming to make the synthesized image appear more natural and reduce visible artifacts caused by manual editing. In image harmonization, “multi-scale” means considering different resolutions or levels of detail within the image. For example, in a high-resolution image, it may be necessary to adjust large-scale features such as overall color and lighting variations, while also paying attention to small-scale elements such as fine objects or textures. By optimizing these characteristics at different scales simultaneously, a more realistic and visually harmonious result can be achieved. Sunkavalli [1] proposed an image multi-scale pyramid method to achieve effective image composition, matching the texture, noise, and other aspects, while requiring minimal user interaction. Duan et al. [15] proposed a multi-scale region correlation-driven strategy to deeply explore the content relevance between the foreground and background, enabling the foreground object to blend more naturally and seamlessly into the background semantics.
This paper primarily employs multi-scale representation methods, such as the Laplacian pyramid, for image harmonization. It is well known that the image pyramid approach allows for the processing of image features at different levels. In multi-scale image processing, the image is first decomposed into a multi-scale pyramid, and then each layer of the pyramid is analyzed or processed in detail at different levels to achieve the desired result. In Heeger and Bergen’s work, pyramids were ingeniously applied to texture synthesis. They demonstrated that by precisely matching the subband coefficients histograms of the noise pyramid with those of the given texture, realistic random textures can be generated. This technique not only preserves the fine details of the original texture but also allows flexible adjustment of the complexity and directionality of the texture, thus finding wide application in the image editing, computer vision, and virtual reality fields. This method effectively captures the key characteristics of textures across multiple scales, thereby enabling realistic texture synthesis.

2.3. Image Harmonization Global Features

The global features of an image become especially important in order to achieve a better and more effective fusion of foreground elements and background. Cong et al. [16] proposed a high-resolution image harmonization network based on Collaborative Dual Transformations (CDTNet). This approach consists of three main components: a low-resolution generator for pixel-level translation, an RGB transformation module for color space mapping, and a refinement module that leverages the results from both, thereby enhancing the overall harmonization of the image. Niu et al. [17], targeting situations where the lighting of the foreground image is inconsistent with that of the background image, introduced a novel method called GiftNet. This approach utilizes global information to guide the transformation of foreground features, leading to significant improvements. Yuan et al. [18] proposed a multi-color curve network that captures richer color information by processing images across multiple color spaces. The network employs an encoder based on modified Transformer blocks to perform multi-stage curve learning in different color spaces. Meanwhile, a multi-color integration module is introduced to effectively fuse feature information extracted from various color spaces, and a lightweight fine-grained refinement module is further incorporated to enhance the output quality, achieving superior image color adjustment performance.
We extract bottleneck features from the encoder as global features to guide the transformation of foreground features in each feature map. Regarding the method of transformation, we opted for using modulated convolution kernels; however, other transformation methods are also applicable. Specifically, we utilize global features to derive modulated convolution kernels and apply these to the foreground features. Our approach shares similarities with CDTNet, which uses global features to predict color transformations. In contrast, our method primarily employs global features to predict feature transformations.

3. Method

We denote the composite image as I c , which is composed of a foreground I f and a background I b , with the ground-truth image being I g t . The objective of the image harmonization task is to adjust the composited foreground I f , so that the resulting harmonized image closely resembles I g t . Our approach is built upon a UNet-based network, which includes three modules: a multi-scale module, a global feature guidance module, and a lightweight refinement module. We apply the designed multi-scale and global feature (MSGF) guided module to image harmonization. The main framework of this module will be detailed in Section 3.1. Then, we provide an in-depth explanation of the multi-scale module in Section 3.2. In Section 3.3, we describe the global feature module, and in Section 3.4, we detail the design of the lightweight refinement module.

3.1. Design of the Main Framework for Image Harmonization Based on Multi-Scale and Global Feature Guidance

Our network consists of multi-scale harmonization, globally guided feature transformation and lightweight refinement module as shown in Figure 2. The multi-scale branch fuses the synthetic image and random noise through the histogram matching technique, thus making the synthetic image more natural and realistic both in color and lighting. In the global feature guidance branch, we employ a UNET structure to perform feature transformation on the global features through the global feature branch. We use the global features to predict a scale vector s , and then obtain the modulated convolution weights W ¯ , which are subsequently applied to the foreground feature maps. This process ultimately achieves harmonization of the composite image. The image obtained from the multi-scale module and the features obtained from the global feature module are effectively fused by a lightweight refinement module to obtain a more harmonized image.

3.2. Multi-Scale Modular Design

In the module, our main work is to combine synthetic and random noise images using a linear pyramid approach. First, the input synthetic and random noise images are decomposed into pyramids, and the synthetic and noise pyramids are iteratively shaped to match the target pyramid using a smooth histogram matching technique. Histogram matching is used in image harmonization (IH) to make the synthesized image look more natural. When an object needs to be extracted from one image and inserted into another background image, it usually encounters inconsistencies in color, lighting, and texture, which make the composite effect look abrupt. With the histogram matching technique, we can adjust the color distribution of the foreground object so that it is closer to the color distribution of the background image, thus achieving better visual integration. Specifically, it can be considered in terms of color harmonization and lighting condition matching. When the colors of the foreground and background may not be consistent, by histogram matching, the color distribution of the foreground objects can be made to match the background, which reduces the color difference and makes the composite look more harmonious. If the foreground object was acquired from different lighting conditions, it may not look like the background. Histogram matching can help adjust brightness and contrast to better simulate the lighting conditions of the original scene. Simple histogram matching may result in artifacts.

3.3. Module Design for Global Feature Guidance

In the global feature guidance module, we introduce the globally guided instance feature transformation (GGFT) mechanism, which is an improvement on the GiftNet proposed by Niu et al. [17]. The mechanism uses global features to predict scale vectors and generates modulated convolutional weights accordingly, aiming to optimize the representation of the foreground feature map without changing the background feature map. In the GGFT itself, the mechanism contains an encoder and a decoder; the encoder contains four modules; the decoder contains three modules into which the GGFT is inserted. In our innovative application, we insert the GGFT modules into each encoder and decoder after the foreground feature map is extracted for each convolution and the global feature vector is extracted.
This architecture contains a total of four encoder blocks. We extract the corresponding foreground features. T e 1 through each encoder denotes the feature map obtained by the first l encoder, and through average pooling, the feature map will be extracted from the corresponding feature vector t e 1 . This feature vector contains the important global features of the foreground image, which provides an important basis for the later image and coordination. After obtaining the feature vectors, the feature vectors are used to guide the feature transformation during the feature mapping process. The specific method is to apply the modulated convolutional weights to the foreground feature map. In the task of image harmonization, the global feature t e 1 is used to guide the modulation of the convolutional weights, which provides global guidance for the feature transformation. In order to achieve harmony between the foreground and background within the image, our method implements transformations only for the foreground feature maps, ensuring that the compatibility between foreground elements and the background environment is optimized. The basic convolutional weight W ¯ used here refers to the spatial position weight of the x th input, the y th output, at the z th position. The size of W ¯ is adjusted by the obtained scale vectors. The features are augmented by the normalization method. By modulating the foreground feature mapping, similarity to the background feature mapping is achieved.

3.4. Lightweight Refined Module Design

In the lightweight refinement module, we introduce lightweight models that can help us to process the results obtained through the module and the global feature module to optimize the synthesized images.
After the pixel-to-pixel transformation in the transformation, we can get the result I ^ m s h H × W × 3 after the transformation, and the result I ^ g g f t H × W × 3 from the global feature guidance module. I ^ m s h has high resolution and can realize fine etching in detail, but lacks global feature processing. I ^ g g f t has rich global features, so they are complementary to each other. In order to better fuse the two obtained results, we therefore designed a lightweight refinement module R to produce better high-resolution results.
In the lightweight refinement module, we first connect the result I ^ m s h H × W × 3 of the transformation and the result I ^ g g f t H × W × 3 obtained from the global feature bootstrap module as inputs to the channel. The refinement module R contains two convolutional layers. As a result, the refinement module R receives inputs from I ^ m s h and I ^ g g f t and produces a finer high-resolution output I ^ h . We force I ^ h to be close to a high-resolution ground-truth image of I g t , denoted L r e f = I ^ h I g t λ , by imposing a reconstruction loss.

3.5. Loss Function Setting

The overall loss function of our MSGF is L f :
L f = L m s h + L g g f t + L r e f
where L m s h denotes the loss function of the multi-scale module, L g g f t denotes the loss function of the global feature module, and L r e f denotes the loss function of the lightweight refinement module. Although the refinement module has a simple structure, it performs better in fusing the multi-scale results and global feature-guided results due to the presence of the information input and the hybrid layer.

4. Experimentation and Analysis

In this paper, we trained and tested the algorithm in a medium implementation on the NVIDIA TITAN X GPU platform. We used four objective evaluation metrics to evaluate the experimental results. In the comparative study, we chose the state-of-the-art image harmonization method to compare with our method. First, the dataset and evaluation metrics are presented; then, the experimental results are analyzed; finally, the quantitative analysis leads to the conclusions in this section.

4.1. Experimental Dataset

The following datasets are widely used for image harmonization: iHarmony4 [19] dataset, RHHarmony dataset, RealHM dataset, HVIDIT dataset, and ccHarmony dataset. In this study, we selected the iHarmony4 dataset as the training and testing dataset. iHarmony4 is a comprehensive and recently proposed benchmark dataset that integrates multiple types of synthetic images, including real synthetic images, globally adjusted synthetic images, synthetic images with local operations, and synthetic images generated through deep learning-based hybrid methods. This diversity ensures comprehensive evaluation across various realistic and challenging scenarios. In recent years, numerous studies in the field of image harmonization have adopted iHarmony4 as the standard benchmark, facilitating fair comparisons with state-of-the-art methods. Each of iHarmony4 contains a wealth of resources inside as shown in Figure 3, which can be used for deep learning training and testing. Our experiments are conducted on the iHarmony4 dataset. iHarmony4 consists of four datasets, namely HCOCO [20], HFlickr [21], HAdobe5K [22], and Hday2night [23]. We used the complete iHarmony4 dataset in our experiments.

4.2. Experimental Test Program

In the comparative experiments, we selected three representative image harmonization methods for performance evaluation. The selection criteria are as follows. (1) The method is highly representative and has demonstrated strong experimental performance in existing studies. (2) It has significant academic influence in the relevant field and has been widely cited. (3) Its source code is publicly available on platforms such as GitHub, facilitating reproducibility and experimental validation. (4) The method exhibits good generalization ability and can be applied to various scenarios or tasks. We conducted experiments on four publicly available datasets from iHarmony4: the HCOCO, HFlickr, HAdobe5K, and Hday2night datasets, which we used as training and to test the visualization results. These methods include the iS2AM [18] (https://github.com/SamsungLabs/, accessed on 12 January 2025) method. Unlike previous methods that train encoder–decoder networks from scratch, this method utilizes the high-level feature space learned by pre-trained classification networks to enhance the ability of neural networks to learn high-level representations of objects. The CDTNet [8] (https://github.com/bcmi/, accessed on 12 January 2025) method’s core innovation lies in proposing a dual-transformation mechanism that performs spatial and color transformations on the foreground and background, respectively, and ensures the visual consistency of the two through co-optimization. Specifically, the spatial transformation module adjusts the shape and position of the foreground objects through an adaptive convolutional network, while the color transformation module accurately matches the color distributions of the foreground and background through a multi-scale feature fusion technique. In addition, the study introduces an efficient generative adversarial network (GAN) framework, which further enhances the realism of the generated images. Harmonizer [4] (https://github.com/ZHKKKe/Harmonizer, accessed on 12 January 2025), using a cascaded regressor and dynamic loss strategy, makes the learning process of filter parameters more stable and accurate. Since the network only generates image-level parameters and the filters used have efficient computational characteristics, this method significantly reduces computational overhead while maintaining performance, thereby improving processing speed.
Comprehensive experiments show that MSGF is obviously superior to the existing methods, which is beneficial to the fusion of fine images at the edges as well as the harmonization of global images through multi-scale and global feature guidance, especially in the high-resolution input.

4.3. Evaluation Criteria

We use mean square error (MSE), foreground mean square error (fMSE), and peak signal-to-noise ratio (PSNR) as evaluation metrics to measure the performance of image harmonization. Specifically, fMSE calculates the mean square error for the foreground region only, not the entire image, because the image harmonization process does not alter the appearance of the background.

4.4. Comparative Experiments and Analysis

In order to better investigate high-resolution image harmonization, we conducted quantitative and qualitative experiments to verify its effectiveness. We compare four methods of image harmonization and give the test results in four more subsets, which are obtained from the models copied or published in the original paper. With Table 1 we can see the considerable performance of the various harmonization methods in different subsets. For each metric, the values indicating optimal performance are shown in bold. For the three test sets, HCOCO, HAdobe5k, and Hday2night, our method outperforms the SOTA method to a large extent. For example, in the HCOCO test set, there is a relative improvement of 2.7% over iS2AM, 1.4% over CDTNet, and 7.67% over Harmonizer on the MSE metric. On the fMSE metric, there was a 2% relative improvement over iS2AM, a 0.2% improvement over CDTNet, and a 12.6% relative improvement over Harmonizer. On the PSNRA metric, there was a 1% relative improvement over iS2AM, a 1.3% improvement over CDTNet, and a 2.3% relative improvement over Harmonizer. Our method achieved the best results on HCOCO, HAdobe5k, and Hday2night. Our method did not achieve fully optimal results on HFlickr, and the PSNR metrics were lower than the other methods. The experimental results are as follows, and Figure 4 shows the experimental results of the selected methods on the iHarmony4 dataset.
As shown in Figure 4, the actual results of iS2AM, CDTNet, Harmonizer and our method for a specific dataset iHarmony4 are demonstrated from a horizontal perspective. From a general point of view, several existing methods are able to achieve the harmonization function, but through comparison, we are still able to see that our method will be closer to ground-truth both in terms of the synthesis effect of the synthesized edges and the fusion with the background image. From the leftmost synthesized image, we can see that the synthesized edges of the effect are rather abrupt, which can be effectively prompted to be harmonized by our method.
In order to better validate the advantages of MSGF in real image coordination, it was tested by means of a user study. We use the real synthesized images published in [17] with a total of 99 images. The harmonized images generated by the method involved in Figure 4 plus the synthetic images form a set of five images in total, and we ranked each set of images by inviting 50 participants. In Table 2, to better validate the effectiveness of our harmonization method, we use the Bradley–Terry model (B–T model) [19] for ranking. MSGF achieves the highest B–T score, which demonstrates the validity of the proposed approach.

4.5. Computational Efficiency Assessment

To evaluate the efficiency of the proposed method in this chapter, we conduct experiments on the iHarmony4 dataset using the iS2AM, CDTNet, Harmonizer, and our method. We record the inference time and model parameter count for each method, with the results presented in Table 3. As shown in Table 3, the MSGF method has the shortest inference time, attributed to its lightweight refinement module. Fast inference speed is essential for real-time applications, while a small model size and low memory requirements are beneficial for deployment on mobile devices. Quantitative results demonstrate that MSGF is faster, lighter, and more memory-efficient than the other methods. Notably, on an NVIDIA GeForce RTX 4090 GPU, MSGF runs twice as fast as the previously fastest method, Harmonizer, and over 100 times faster than iS2AM and CDTNet. Furthermore, as observed from the table, MSGF achieves faster computation and fewer parameters by integrating an image pyramid approach with a lightweight refinement fusion module.

4.6. Analysis of Ablation Experiments

In order to better validate the effectiveness of the multi-scale module and the global feature bootstrap module for the MSGF model, valuable experience has been gained in a wide range of techniques to be able to validate our modules more effectively. In Table 4, we provide a complete analysis and insights on all the individual modules and demonstrate the validity of the methodology presented in this section. The same metrics are used as in Section 4.3 and the design of the ablation experiments in the same database is shown as follows: (1) baseline model: CDTNet is used as the baseline, and the method proposed in this chapter is an improvement based on the CDTNet model; (2) harmonization of images using only the multi-scale module and the lightweight refinement module is called MSHR; (3) harmonization of images using only the global feature bootstrap and the lightweight refinement module is called GFTR; (4) the method proposed in this chapter. In Figure 5 the harmonization effects of the different modules are demonstrated and visualized for the following images: (a) synthesized image, (b) real image, (c) “MSHR” harmonization result, (d) “GFTR” harmonization result, (e) synthesized result by MSGF method.
From the qualitative results in Figure 5, it can be seen that the method in this chapter, MSGF, has the best harmonization effect, and the MSHG method, although the multi-scale pyramid method was used for image harmonization, is slightly better than the Baseline methods, but it is obvious in the first line that the detailed hair part still shows the phenomenon of hair lumps, and it can be clearly seen that in the unnatural image, the GFTR effect is slightly better than that of MSHG, and it is achieved by global feature feature transformation to achieve image harmonization, but not as good as MSHG in terms of color contrast. Our method combines the advantages of the two methods to achieve a better harmonization effect. In the quantitative analysis results in Table 4, we can see that overall, these three methods have their advantages relative to other methods, but the MSGF method is especially outstanding in all indicators, and can achieve better image harmonization through the effective combination of the multi-scale pyramid method and global feature guidance.

5. Conclusions

In this paper, we investigate the process of image harmonization on synthetic images. For better realism, we consider the approach and global features in a unified way and try to optimize the synthetic images with a lightweight refinement module, which inspired us to design and implement the MSGF method. The method efficiently processes images from two aspects, and thanks to our new architecture and the effective combination of the two methods, MSGF is lighter and faster than previous methods while achieving new state-of-the-art performance. However, our method does have limitations. It does not give good results in dealing with the problem of harmonization of video images. In future work, we hope to develop more suitable image harmonization methods to solve this problem.

Author Contributions

Conceptualization, R.L. and D.Z.; formal analysis, R.L.; methodology, R.L. and D.Z.; validation, R.L.; writing—original draft, R.L.; Writing—review and editing, D.Z., M.Z. and S.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation for Youth of Qinghai Province (Project No. 2023-ZJ-947Q) and the National Natural Science Foundation of China (Project No. 6246070542, 62262056).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets analyzed during the current study are iHarmony4 dataset (link: https://pan.baidu.com/s/1NL3_yMLx2QUo1grkyPS6pw?pwd=6g1q; extraction code: 6g1q, accessed on 12 January 2025). The contrasting experimental approaches are the iS2AM (https://github.com/SamsungLabs/, accessed on 12 January 2025); CDTNet (https://github.com/bcmi/, accessed on 12 January 2025); Harmonizer (https://github.com/ZHKKKe/Harmonizer, accessed on 12 January 2025).

Conflicts of Interest

We declare that we do not have any commercial or associative interests that represent a conflict of interest in connection with the work submitted.

Abbreviations

The following abbreviations are used in this manuscript:
MSGFMulti-scale and global feature
VGGVisual geometry group
DovNetDomain verification discriminator network
CNNConvolutional neural network
RGBRed, green, blue
CDTNetCollaborative dual transformations network
GGFTGlobally-guided instance feature transformation
iS2AMComposite image of spatial-separated attention module
GANGenerative adversarial network
MSEMean square error
fMSEForeground of mean square error
PSNRPeak signal-to-noise ratio
MSHRMulti-scale module and the lightweight refinement
GFTRGlobal feature bootstrap and the lightweight refinement

References

  1. Sunkavalli, K.; Johnson, M.K.; Matusik, W.; Pfister, H. Multi-scale image harmonization. ACM Trans. Graph. 2010, 29, 125. [Google Scholar] [CrossRef]
  2. Luan, F.; Paris, S.; Shechtman, E.; Bala, K. Deep painterly harmonization. Comput. Graph. Forum. 2018, 37, 95–106. [Google Scholar] [CrossRef]
  3. Xue, S.; Agarwala, A.; Dorsey, J.; Rushmeier, H. Understanding and improving the realism of image composites. ACM Trans. Graph. 2012, 31, 84. [Google Scholar] [CrossRef]
  4. Sofiiuk, K.; Popenova, P.; Konushin, A. Foreground-aware semantic representations for image harmonization. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 1620–1629. [Google Scholar]
  5. Ke, Z.; Sun, C.; Zhu, L.; Xu, K.; Lau, R.W.H. Harmonizer: Learning to perform white-box image and video harmonization. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer Nature: Cham, Switzerland, 2022; pp. 690–706. [Google Scholar]
  6. Cong, W.; Zhang, J.; Niu, L.; Liu, L.; Ling, Z.; Li, W.; Zhang, L. Dovenet: Deep image harmonization via domain verification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 8394–8403. [Google Scholar]
  7. Guo, Z.; Guo, D.; Zheng, H.; Gu, Z.; Zheng, B.; Dong, J. Image harmonization with transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 14870–14879. [Google Scholar]
  8. Wang, K.; Gharbi, M.; Zhang, H.; Xia, Z.; Shechtman, E. Semi-supervised parametric real-world image harmonization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 5927–5936. [Google Scholar]
  9. Meng, Q.; Liu, Q.; Li, Z.; Lan, X.; Zhang, S.; Nie, L. High-resolution image harmonization with adaptive-interval color transformation. Adv. Neural Inf. Process. Syst. 2024, 37, 13769–13793. [Google Scholar]
  10. Pérez, P.; Gangnet, M.; Blake, A. Poisson image editing. In Seminal Graphics Papers: Pushing the Boundaries; Association for Computing Machinery: New York, NY, USA, 2023; Volume 2, pp. 577–582. [Google Scholar]
  11. Tian, C.; Zhang, Q. Self-Supervised Image Harmonization via Holistic Feature Fusion. In Proceedings of the ICASSP 2025—2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Hyderabad, India, 6–11 April 2025; pp. 1–5. [Google Scholar]
  12. Usama, M.; Nyman, E.; Näslund, U.; Grönlund, C. A domain adaptation model for carotid ultrasound: Image harmonization, noise reduction, and impact on cardiovascular risk markers. Comput. Biol. Med. 2025, 190, 110030. [Google Scholar] [CrossRef] [PubMed]
  13. Mwubahimana, B.; Yan, J.; Mugabowindekwe, M.; Xiao, H.; Nyandwi, E.; Tuyishimire, J.; Habineza, E.; Mwizerwa, F.; Miao, D. Vision transformer-based feature harmonization network for fine-resolution land cover mapping. Int. J. Remote Sens. 2025, 46, 3736–3769. [Google Scholar] [CrossRef]
  14. Oriti, D.; Manuri, F.; De Pace, F.; Sanna, A. Harmonize: A shared environment for extended immersive entertainment. Virtual Real. 2023, 27, 3259–3272. [Google Scholar] [CrossRef] [PubMed]
  15. Duan, L.; Wu, M.; Lou, H.; Yin, J.; Li, X. MRCAN: Multi-scale Region Correlation-driven Adaptive Normalization for Image Harmonization. In Proceedings of the 2024 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Kuching, Malaysia, 6–10 October 2024; pp. 5206–5211. [Google Scholar]
  16. Cong, W.; Tao, X.; Niu, L.; Liang, J.; Gao, X.; Sun, Q.; Zhang, L. High-resolution image harmonization via collaborative dual transformations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 18470–18479. [Google Scholar]
  17. Niu, L.; Tan, L.; Tao, X.; Cao, J.; Guo, F.; Long, T.; Zhang, L. Deep image harmonization with globally guided feature transformation and relation distillation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 7723–7732. [Google Scholar]
  18. Yuan, J.; Wu, H.; Xie, L.; Zhang, L.; Xing, J. Learning multi-color curve for image harmonization. Eng. Appl. Artif. Intell. 2025, 146, 110277. [Google Scholar] [CrossRef]
  19. Cong, W.; Zhang, J.; Niu, L.; Liu, L.; Ling, Z.; Li, W.; Zhang, L. Image Harmonization Dataset iHarmony4: HCOCO, HAdobe5k, HFlickr, and Hday2night. arXiv 2019, arXiv:1908.10526. [Google Scholar]
  20. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision—ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Part v 13. Springer International Publishing: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  21. Zhou, B.; Zhao, H.; Puig, X.; Xiao, T.; Fidler, S.; Barriuso, A.; Torralba, A. Semantic understanding of scenes through the ade20k dataset. Int. J. Comput. Vis. 2019, 127, 302–321. [Google Scholar] [CrossRef]
  22. Bychkovsky, V.; Paris, S.; Chan, E.; Durand, F. Learning photographic global tonal adjustment with a database of input/output image pairs. In Proceedings of the 2011 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2011), Colorado Springs, CO, USA, 20–25 June 2011; pp. 97–104. [Google Scholar]
  23. Tsai, Y.-H.; Shen, X.; Lin, Z.; Sunkavalli, K.; Lu, X.; Yang, M.-H. Deep image harmonization. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2017), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
Figure 1. Image harmonization (the left image is the composite, and the right image is the harmonized result). From the figure, we can observe that the image after harmonization appears more realistic and natural.
Figure 1. Image harmonization (the left image is the composite, and the right image is the harmonized result). From the figure, we can observe that the image after harmonization appears more realistic and natural.
Applsci 15 10573 g001
Figure 2. Design of image harmonization framework (there are three modules).
Figure 2. Design of image harmonization framework (there are three modules).
Applsci 15 10573 g002
Figure 3. Composition of the iHarmony4 dataset.
Figure 3. Composition of the iHarmony4 dataset.
Applsci 15 10573 g003
Figure 4. Visual comparison effect of various methods on iHarmony4 (use red boxes to enlarge details).
Figure 4. Visual comparison effect of various methods on iHarmony4 (use red boxes to enlarge details).
Applsci 15 10573 g004
Figure 5. Harmonization effect of different modules in iHarmony4 dataset.
Figure 5. Harmonization effect of different modules in iHarmony4 dataset.
Applsci 15 10573 g005
Table 1. Quantitative analysis: test values for each method in the comparison experiment. (↓ indicates smaller values are better, ↑ indicates larger values are better. Bold indicates the optimal value.)
Table 1. Quantitative analysis: test values for each method in the comparison experiment. (↓ indicates smaller values are better, ↑ indicates larger values are better. Bold indicates the optimal value.)
MethodHCOCOHFlickrHAdobe5kHday2night
MSE
fMSE
PSNR ↑MSE
fMSE
PSNR ↑MSE
fMSE
PSNR ↑MSE
fMSE
PSNR ↑
iS2AM16.48266.1439.2969.68443.6333.5622.59166.1937.2440.59591.0737.72
CDTNet16.25261.2939.1568.61423.0333.5520.62149.8838.2436.72549.4737.95
Harmo
nizer
17.34298.4238.7764.81434.0633.6321.89170.0537.6433.14542.0737.56
Ours16.02260.6739.6960.42350.5633.4218.43130.9639.7833.01480.6737.98
Table 2. B–T scores of 100 real synthetic images by different methods.(↑ indicates larger values are better. Bold indicates the optimal value).
Table 2. B–T scores of 100 real synthetic images by different methods.(↑ indicates larger values are better. Bold indicates the optimal value).
MethodInput CompositeiS2AMCDTNetHIMOurs
B–T score↑0.3870.4650.8930.8511.328
Table 3. Time loss in image harmonization by different methods on the iHarmony4 dataset (Bolding indicates the best ranked value).
Table 3. Time loss in image harmonization by different methods on the iHarmony4 dataset (Bolding indicates the best ranked value).
MethodTime (s)Model Parameters (in Millions)
iS2AM14.768
CDTNet10.8216
HIM0.0221.7
Ours0.0120.9
Table 4. Objective evaluation metrics of different modules for image harmonization on the iHarmony4 dataset. (↓ indicates smaller values are better, ↑ indicates larger values are better. Bold indicates the optimal value).
Table 4. Objective evaluation metrics of different modules for image harmonization on the iHarmony4 dataset. (↓ indicates smaller values are better, ↑ indicates larger values are better. Bold indicates the optimal value).
MethodHCOCOHFlickrHAdobe5kHday2night
MSE
fMSE
PSNR ↑MSE
fMSE
PSNR ↑MSE
fMSE
PSNR ↑MSE
fMSE
PSNR ↑
Baseline16.48266.1439.2969.68443.6333.5622.59166.1937.2440.59591.0737.72
MSHG16.22262.0237.2564.79432.7531.7821.06152.6738.9235.39580.6236.21
GFTR16.10261.7838.8362.43398.5232.4119.87143.6639.4134.87503.4137.54
Ours16.02260.6739.6960.42350.5633.4218.43130.9639.7833.01480.6737.98
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, R.; Zhang, D.; Geng, S.; Zhou, M. Research on High-Resolution Image Harmonization Method Based on Multi-Scale and Global Feature Guidance. Appl. Sci. 2025, 15, 10573. https://doi.org/10.3390/app151910573

AMA Style

Li R, Zhang D, Geng S, Zhou M. Research on High-Resolution Image Harmonization Method Based on Multi-Scale and Global Feature Guidance. Applied Sciences. 2025; 15(19):10573. https://doi.org/10.3390/app151910573

Chicago/Turabian Style

Li, Rui, Dan Zhang, Shengling Geng, and Mingquan Zhou. 2025. "Research on High-Resolution Image Harmonization Method Based on Multi-Scale and Global Feature Guidance" Applied Sciences 15, no. 19: 10573. https://doi.org/10.3390/app151910573

APA Style

Li, R., Zhang, D., Geng, S., & Zhou, M. (2025). Research on High-Resolution Image Harmonization Method Based on Multi-Scale and Global Feature Guidance. Applied Sciences, 15(19), 10573. https://doi.org/10.3390/app151910573

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop