Next Article in Journal
Research on the Aerodynamic–Propulsion Coupling Characteristics of a Distributed Propulsion System
Previous Article in Journal
A Lightweight Received Signal Strength Indicator Estimation Model for Low-Power Internet of Things Devices in Constrained Indoor Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fast Fourier Asymmetric Context Aggregation Network: A Controllable Photo-Realistic Clothing Image Synthesis Method Using Asymmetric Context Aggregation Mechanism

by
Haopeng Lei
1,†,
Ying Hu
1,†,
Mingwen Wang
1,*,
Meihai Ding
1,
Zhen Li
1 and
Guoliang Luo
2
1
The School of Computer Information Engineering, JiangXi Normal University, Nanchang 330022, China
2
Virtual Reality and Interactive Techniques Institute, East China Jiaotong University, 808 Changbei Shuanggang East Street, Nanchang 330013, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2025, 15(7), 3534; https://doi.org/10.3390/app15073534
Submission received: 19 January 2025 / Revised: 4 March 2025 / Accepted: 5 March 2025 / Published: 24 March 2025
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Clothing image synthesis has emerged as a crucial technology in the fashion domain, enabling designers to rapidly transform creative concepts into realistic visual representations. However, the existing methods struggle to effectively integrate multiple guiding information sources, such as sketches and texture patches, limiting their ability to precisely control the generated content. This often results in issues such as semantic inconsistencies and the loss of fine-grained texture details, which significantly hinders the advancement of this technology. To address these issues, we propose the Fast Fourier Asymmetric Context Aggregation Network (FCAN), a novel image generation network designed to achieve controllable clothing image synthesis guided by design sketches and texture patches. In the FCAN, we introduce the Asymmetric Context Aggregation Mechanism (ACAM), which leverages multi-scale and multi-stage heterogeneous features to achieve efficient global visual context modeling, significantly enhancing the model’s ability to integrate guiding information. Complementing this, the FCAN also incorporates a Fast Fourier Channel Dual Residual Block (FF-CDRB), which utilizes the frequency-domain properties of Fast Fourier Convolution to enhance fine-grained content inference while maintaining computational efficiency. We evaluate the FCAN on the newly constructed SKFashion dataset and the publicly available VITON-HD and Fashion-Gen datasets. The experimental results demonstrate that the FCAN consistently generates high-quality clothing images aligned with the design intentions while outperforming the baseline methods across multiple performance metrics. Furthermore, the FCAN demonstrates superior robustness to varying texture conditions compared to the existing methods, highlighting its adaptability to diverse real-world scenarios. These findings underscore the potential of the FCAN to advance this technology by enabling controllable and high-quality image generation.

1. Introduction

In recent years, the incorporation of artificial intelligence (AI) technology has facilitated a revolutionary transformation in the fashion design industry, fostering innovation and injecting new vitality and opportunities [1]. Traditional fashion design is a complex and time-consuming task that heavily relies on manual efforts, requiring designers to translate creative concepts into visually appealing garments through stages like sketching, texture refinement, and prototyping. This process is not only intricate but also constrained by seasonal cycles, with designers facing tight deadlines for various apparel types—such as summer dresses or winter coats [2]. These pressures highlight the need for efficient solutions to support designers in navigating these complexities. AI-driven tools have emerged as a powerful solution to these challenges, enabling designers to generate innovative ideas, refine concepts, and visualize designs more quickly and with greater precision. Among the most promising advancements is clothing image synthesis, which allows for the generation of photo-realistic images automatically from diverse inputs such as sketches, texture patches, and textual descriptions. This technology enhances design workflows by streamlining the process and providing designers with more precise visual outputs. Furthermore, clothing image synthesis holds significant value for applications in areas like virtual try-on systems, fashion recommendation, and digital prototyping. By bridging the gap between abstract design concepts and realistic visual outputs, it enables designers to efficiently refine their ideas while offering consumers more interactive and personalized experiences. Therefore, this technology not only enhances creative efficiency but also contributes to the sustainable development and continued innovation of the related industry.
Although clothing image synthesis has facilitated broader applications in fashion design, generating high-quality, semantically consistent, and texture-rich clothing, synthesis remains a challenging task. These difficulties arise from the inherent complexity of diverse garment designs and the intricate details of textures. Addressing these challenges is essential not only to enhance the photo-realism and accuracy of synthesized images but also to develop more efficient and controllable design processes, which hold substantial practical value for the fashion industry. However, many current methods for clothing image synthesis fail to fully address the need for subjective controllability in design tasks. For instance, some approaches rely on text descriptions [3,4,5], cross-domain style images [6,7,8], or existing clothing images [9,10,11] as guidance for image generation. While these methods can generate visually plausible results, the outputs often deviate from the designer’s original intent. In contrast, design sketches provide a more direct reflection of the designer’s structural intent by encoding rich semantic information about garment shape and structure. Utilizing design sketches as structural guidance ensures that generated clothing images can align with the designer’s creative intent. However, sketches alone are insufficient for capturing essential texture attributes, such as fabric color, material texture, and intricate details, which are critical for a garment’s visual appeal and consumer perception. To address the limitations of sketch-based methods, some approaches [12,13,14] incorporate texture style content as supplementary stylistic guidance. While these methods improve the visual output of simple and repetitive patterns, the structural complexity of garment regions presents considerable challenges in accurately mapping texture content to these areas. This challenge is especially evident when the texture inputs are limited or contain intricate details, a situation frequently encountered in practical applications. Additionally, creating detailed texture references remains a laborious and time-intensive process, posing additional challenges for designers. Although some studies on texture synthesis [15,16,17] have achieved impressive results using limited texture content, the outcomes are often restricted to simple, repetitive patterns and suffer from issues such as edge-seam artifacts. Moreover, these methods struggle to accurately capture critical features of clothing images, such as lighting effects and fabric folds. Consequently, utilizing such techniques as a source for clothing style remains remains fraught with difficulties.
The above issues highlight the need for advanced technologies that can seamlessly integrate the structural guidance of design sketches with precise texture control to generate realistic garment images. Such methods would allow designers to quickly obtain high-quality images that align with their creative intentions, significantly enhancing the practical value of these techniques. To achieve this goal, we propose a novel fashion clothing image generation network (FCAN) that combines Fast Fourier Convolution (FFC) [18] with a novel Asymmetric Context Aggregation Mechanism (ACAM) to generate clothing images that align with the intended design structures and textures. By utilizing dual guidance from design sketches and texture patches, the method ensures that the structural semantic information from the sketches is preserved while enabling precise control over texture styles. Within the FCAN, we leverage the spectral transformation capability of FFC to design a Fast Fourier Channel Dual Residual Block (FF-CDRB), which efficiently handles complex textures and detailed information for content inference. Additionally, we introduce the ACAM in the FCAN, which innovatively aggregates global context information through multi-stage heterogeneous features. This mechanism partially compensates for the inadequate context modeling capabilities of the existing methods, enabling accurate integration of sketch information with limited texture content. Extensive comparative experiments demonstrate that this method exhibits superior robustness to variations in texture patch size compared to the existing advanced methods, and it is capable of synthesizing high-quality clothing images across diverse texture conditions. In addition, extensive ablation studies demonstrate that, within the FCAN, the ACAM effectively addresses the limitations of the existing methods in long-range context modeling, significantly improving the quality of synthesized images and enhancing the model’s robustness to variations in texture patch size. The integration of the FF-CDRB further boosts the model’s ability to capture complex textures and fine details while optimizing parameter efficiency. Together, these innovations establish the FCAN as a practical and effective solution, capable of making a positive impact in real-world applications, particularly in enhancing the efficiency of fashion design processes.
Figure 1 demonstrates the clothing image synthesis results produced by our proposed method. Notably, by guiding the generation process with texture patches and design sketches that include brief internal contours, our method is able to flexibly synthesize complex structures and diverse texture content. Furthermore, it exhibits positive effects in areas such as wrinkles and shadows. It is worth noting that the texture patches in the images are derived from fabric image cropping, Photoshop compositing, or real garment cutting. Regardless of the source, our method consistently demonstrates strong generative performance, highlighting its flexibility and reliability.
In summary, our main contributions are as follows:
  • We propose a novel clothing image synthesis method guided by design sketches and texture patches, enabling precise control over both the structural appearance and texture style of the clothing. Our method provides innovative and practical solutions for scenarios such as personalized clothing design and rapid prototyping.
  • We design a new content inference module, the Fast Fourier Channel Dual Residual Block (FF-CDRB), which utilizes the global receptive field characteristics of Fast Fourier Convolution (FFC). This module improves the network’s ability to capture contextual information and perform accurate content inference while also improving the efficiency of parameter computation. By integrating the FF-CDRB, the model demonstrates its capability to generate high-quality images with optimized parameter utilization, ensuring both accuracy and computational efficiency.
  • We propose a novel Asymmetric Context Aggregation Mechanism (ACAM), which constructs global context by effectively leveraging heterogeneous features with distinct information representation characteristics from different scales and stages. This approach addresses the limitations of the existing methods in context modeling. The ACAM significantly improves both the perceptual and pixel-level quality of generated images while also enhancing the model’s robustness to variations in texture patch size. To the best of our knowledge, the ACAM is an innovative method for context information interaction, providing potentially valuable references for the research community.
  • Extensive experiments were conducted on our self-constructed dataset, SKFashion, as well as two public datasets, VITON-HD and Fashion-Gen, to validate the effectiveness of the FCAN. The experimental results show that those images generated by the FCAN outperform advanced baseline methods in terms of both perceptual and pixel-level quality. Moreover, the FCAN demonstrates greater robustness to variations in texture patch conditions compared to the other methods, making it more suitable for meeting practical application requirements.

2. Related Work

2.1. Image Generation Models

Image generation models have become a pivotal tool in the field of computer vision research, designed to synthesize realistic or stylized images by learning the underlying distributions from data. These models play a critical role in a wide range of applications, such as creative content generation [19,20,21,22,23], medical imaging [24,25,26,27,28], multi-scene imaging optimization [29,30,31,32,33], and fashion design [34,35,36]. In recent years, the evolution of image generation models has been predominantly influenced by two significant methodologies: generative adversarial networks (GANs) [37] and diffusion models [38].
GANs, originally introduced by Goodfellow et al. [39], typically consist of two neural networks: a generator, which learns to produce realistic images that can deceive the discriminator, and a discriminator, which evaluates the authenticity of the generated samples. This adversarial setup encourages the generator to iteratively improve its outputs, achieving highly realistic image synthesis. Over the years, various GAN-based architectures have made impressive strides in areas such as image generation quality, data dependency optimization, and controllability of generated content, significantly advancing the field of image generation technology. For instance, the StyleGAN [40] series proposed by researchers at NVIDIA has enabled the synthesis of highly detailed high-resolution images, taking GAN-generated content quality to a new level. Similarly, CycleGAN [41], introduced by Zhu et al., facilitates image translation under unpaired data conditions, offering a key solution to the problem of training data dependency. Moreover, conditional generative adversarial networks (cGAN) [42], proposed by Mirza et al., establish the foundation for GAN controllability by enabling content generation through conditional embedding.
However, despite these advancements, GANs still face challenges such as training instability and mode collapse, which somewhat limit their ability to produce diverse and consistent results across different tasks. To optimize the stability of GANs, researchers have undertaken significant efforts, proposing a range of effective solutions. For instance, Arjovsky et al. [43] introduced Wasserstein distance during training, which resulted in smoother and more stable gradients, mitigating the problem of gradient vanishing to some extent and thereby enhancing the training stability of GANs. Similarly, Mao et al. [44] employed the least squares loss function, alleviating the gradient vanishing issue in both the generator and discriminator networks. These contributions have been pivotal in the development and advancement of GANs. Today, GANs have become a representative technology in the field of image generation, widely applied across various generative tasks.
Diffusion models, on the other hand, represent a fundamentally different approach to image generation. Inspired by non-equilibrium thermodynamics, these models progressively add noise to input images during training and then learn to reverse this process, effectively generating images from noise. The Denoising Diffusion Probabilistic Model (DDPM) [45] proposed by Ho et al. represents a milestone in the field. By simplifying the model structure and optimizing the training process, the DDPM significantly reduces the training complexity of diffusion models while markedly improving the quality of generated samples. This advancement has demonstrated the immense potential of diffusion models for high-quality and high-resolution image synthesis, sparking widespread academic interest in diffusion-based approaches. Building on this progress, Song et al. introduced Denoising Diffusion Implicit Models (DDIMs) [46], which, by employing a deterministic sampling strategy, effectively reduce inference time without sacrificing quality. This innovation has paved the way for the practical application of diffusion models.
Although diffusion models have demonstrated impressive generative capabilities, their extremely high computational resource requirements remain a significant obstacle to their development. To address this, Rombach et al. [47] proposed latent diffusion models (LDMs), which transfer the training and inference processes from the high-dimensional pixel space to a compressed latent space. This shift reduces resource consumption, effectively mitigating the computational overhead of diffusion models, and has made a crucial contribution to their rapid advancement. Additionally, hybrid approaches, such as UFOGen [48], aim to combine the fast convergence of GANs with the superior quality of diffusion models, offering new possibilities in balancing quality and efficiency.
Moreover, with the proliferation of large-scale datasets and the remarkable advancements in computational hardware, significant progress has been achieved in large-scale image generation models in recent years. A series of large-scale image generation models with unprecedented generative capabilities have emerged, marking the onset of a new era in the field. For instance, NVIDIA introduced StyleGAN-XL [49], based on the GAN architecture, which is capable of handling larger datasets and generating higher-resolution images with more intricate details. Similarly, GigaGAN [50], developed collaboratively by several universities and Adobe Research, breaks through the scale limitations of previous models, showcasing the exceptional generative capabilities of multimodal image generation within the GAN framework. It outperforms large models based on diffusion architectures in terms of generation speed and image quality, underscoring the continued potential of GANs in the era of big data. Moreover, major tech companies such as OpenAI and Stability AI have introduced large-scale image generation models based on diffusion architectures, such as the DALL·E series [51] and the Stable Diffusion series [52]. Backed by vast computational resources and massive datasets, these models have further demonstrated the immense potential of generative models, significantly influencing the related field.

2.2. Fashion Clothing Image Synthesis

Fashion image synthesis has become a key research direction in the field of image generation, mainly focusing on the creation and editing of fashion images such as clothing, accessories, models, texture fabrics, etc. However, among many research works, clothing image synthesis and editing methods have attracted significant attention as clothing serves as the core element in various tasks within the fashion field. The applications of fashion clothing image synthesis span a wide range of fields, such as virtual fashion design, personalized clothing generation, and automated fashion creation. Over the years, various innovative models have been developed to address the unique challenges in this domain, demonstrating tremendous potential for practical applications.
Currently, fashion clothing image synthesis primarily relies on deep generative models such as GANs and diffusion models due to their ability to generate realistic images. One of the pioneering works in fashion image synthesis is FashionGAN [53], proposed by Cui et al., which can conditionally generate fashion images based on specific attributes (e.g., color, style, or texture), laying the foundation for subsequent research. Since then, numerous fashion image synthesis methods have been proposed, introducing new technological tools to the fashion domain.
The controllability and quality of generated images are key areas of focus in the current research on garment image synthesis [1] as they directly impact the applicability of these methods in critical tasks such as fashion design, prototype presentation, and virtual try-on, thereby significantly influencing their practical value in the fashion industry.
As the quality of clothing images directly affects the practicality of these methods, the existing approaches are often specifically optimized based on the characteristics of their respective tasks to ensure the quality of the generated images. For example, MT-VTON, proposed by Lee et al. [54], introduces a transfer strategy that enhances the realism of generated clothing by optimizing wrinkles, shadows, and texture effects. Similarly, Jiang et al. designed an alternating patch-global optimization methodology in [8] to ensure semantic consistency of clothing content. These methods have demonstrated promising performance in certain specific scenarios, serving as an inspiration for advancements in related research.
In terms of controllability, the current methods primarily use text-based guidance, sketch or contour-based guidance, and texture-style content guidance to control the appearance, structure, and texture of the generated clothing. Sketches can provide relatively accurate guidance for the appearance and structure, while text descriptions, when paired with text encoders, offer feature guidance aligned with the semantic content of the related clothing, thus enabling control over the structure or texture. Texture styles can further guide the details of the texture content. However, considering the limitations of these control methods when used independently to control structure or texture, the existing approaches often combine them, yielding promising results. For instance, Sun et al. proposed a fashion garment generation model called SGDiff [5], which leverages a pre-trained text-to-image diffusion model to generate clothing images controlled by both textual descriptions and style examples. However, even state-of-the-art text encoders like CLIP [55] still struggle to achieve fine-grained content alignment in specific knowledge domains. Furthermore, textual descriptions themselves are limited in their ability to express subjective intentions. These constraints result in deficiencies in both the quality and subjective consistency of clothing images generated by SGDiff, hindering its applicability in tasks such as design and virtual try-ons. Gu et al. introduced a fashion image generation framework based on a GAN architecture, called PAINT [56], which combines clothing semantic contours and texture patch content to generate coherent fashion images, offering a novel solution for improving efficiency in fashion design. Nevertheless, the PAINT framework, driven by texture blocks, performs poorly in context modeling and fails to fully leverage texture patch information, often resulting in artifacts when synthesizing complex texture content. Furthermore, PAINT shows poor robustness under varying texture patch conditions and typically requires more complete texture patch content to generate reasonable outputs, limiting its scalability to diverse application scenarios. Gao et al. proposed a controllable fashion image synthesis method guided by clothing edge contours and texture patches [57]. Although this method is capable of generating reasonable clothing content, it struggles to capture fine-grained internal structures or texture features. As a result, the generated clothing content is often limited to solid colors or simple repetitive texture patterns, making it difficult to achieve flexible and diverse applications. In addition, there are other approaches, such as DiffFashion [6] proposed by Cao et al., which leverage the generative capabilities of pre-trained models to synthesize garment images and control the generation process through style transfer strategies. Despite this, these methods are limited in their adaptability to different requirements and may generate unreasonable out-of-domain content due to the broad knowledge embedded in pre-trained models.
In summary, despite some progress, the existing methods still exhibit clear limitations. Specifically, sketch-guided and texture-guided approaches struggle to balance image quality with precise control over diverse stylistic content. Text-based methods, meanwhile, are constrained by inherent expression deficiencies and the current capabilities of encoding technologies, making them challenging to apply in practical tasks such as fashion design. Moreover, other methods based on pre-trained models face similar issues in terms of controllability and image quality. These limitations continue to hinder the further development of related methods, becoming a critical issue that needs to be addressed in this field.

2.3. Fast Fourier Convolution

FFC is an operator based on Fast Fourier Transform (FFT), proposed by Chi et al. in [18]. This operator utilizes Fourier transforms to shift spatial signals into the frequency domain, addressing the limitations of traditional convolution operations in terms of computational efficiency and feature extraction quality. In computer vision tasks, FFC leverages the properties of the frequency domain to obtain a receptive field covering the entire image at the early stages of the network, enabling more efficient extraction of global features without increasing the computational cost. In recent years, FFC has been widely applied in the field of computer vision. For instance, Chu proposed an image inpainting method based on FFC [58], which exploits the sensitivity of FFC to repetitive textures in images, achieving effective inpainting results for images with repetitive textured content. Zhou et al. proposed a non-uniform dehazing method using FFC [59], where the global receptive field property of FFC enabled effective haze removal. Additionally, Quattrini et al. utilized FFC for ink volume detection in digital document restoration [60], achieving excellent detection results. These FFC-based works have introduced new technological pathways for solving various tasks in related fields, positively influencing the progress of related research.

3. Proposed Method

3.1. Overview

Our method takes as input a design sketch S, texture patch T, and clothing region mask M c , with the goal of synthesizing a clothing image C gen whose semantics align with both S and T. The sources and combination strategies of these input components will be elaborated on in subsequent sections.
The primary challenge of the task lies in ensuring that the synthesized clothing image not only conforms to the appearance pattern of the sketch but also propagates texture content effectively and coherently across the clothing regions. This imposes stringent requirements on the alignment of global information within the image. Moreover, we consider the interplay between texture patch size and synthesis performance. For instance, larger texture patches provide richer texture information but require more effort from the designer to obtain. Consequently, a synthesis strategy that is more robust to variations in texture patch size can significantly enhance the practical applicability of our method.
To address these challenges, we propose a frequency-domain-based content inference block (FF-CDRB) that leverages frequency-domain characteristics to synthesize fine-grained texture content. Furthermore, we introduce the ACAM, which, during the decoding phase, utilizes heterogeneous features to construct a more comprehensive global relationship. The ACAM enhances information through texture patch, enabling better exploitation of the limited guidance provided by the texture patches. This approach not only improves the consistency between the appearance structure and texture content but also enhances the model’s robustness to variations in texture patch size.
Figure 2 shows the overall framework of FCAN. In this chapter, we will provide a detailed introduction to the workflow of FCAN.

3.2. The Encoding Stage

In the encoding stage, we first concatenate the design sketch S, texture patch T, and clothing mask M c along the channel dimension to obtain the fused feature f 0 :
f 0 = Concat ( S , M c , T ) ,
where Concat denotes the concatenation operation along the channel dimension. S and T provide structural and texture guidance, respectively, while M c prevents content overflow. During training, S , T , and M c are extracted from real clothing images. Subsequently, FCAN employs N layers of sequential lightweight encoding blocks, Encoder i , based on vanilla convolution, where i { 1 , 2 , , N } , to encode f 0 into multi-scale features f 1 , f 2 , , f N . Additionally, in the first encoding block, we introduce sinusoidal position encoding to embed positional information of the garment regions P, enhancing the efficiency of feature alignment. A detailed implementation can be found in [61].
The structure of encoding block is shown in Figure 3. The embedding process of P is illustrated in Figure 4, and the multi-scale encoding process can be expressed as follows:
f 1 = Encoder 1 f 0 , P f 2 = Encoder 2 f 1 f N = Encoder N f N 1 .

3.3. The Content Inference Stage

In the content inference stage, the deepest encoding feature f N is fed into the content inference module, which consists of R connected FF-CDRB blocks. This module synthesizes garment content by aligning texture and structure through frequency-domain characteristics. Next, we will provide a detailed description of the FF-CDRB.

Fast Fourier Channel Dual Residual Block

Inspired by the residual design in ResNet [62], we introduce the Fast Fourier Channel Dual Residual Block (FF-CDRB) based on FFC, aiming to facilitate apparel content inference. The structure of FF-CDRB is shown in the lower part of Figure 2.
The FF-CDRB consists of two serially connected FFCs. As shown in Figure 5, FFC divides the channels into a local branch and a global branch. The local branch captures local features using standard convolution, while the global branch interprets global context through FFT. In the global branch, FFC employs a spectral transformation module based on the Discrete Fourier Transform to convert signals from the spatial domain to the frequency domain, enabling convolution operations with a global receptive field. Finally, the conjugate symmetry of real-valued signals is used to perform an inverse transform to restore spatial structures. The specific operations in the global branch are as follows, including the real and complex spaces, respectively. Real and image denote the real and imaginary parts, respectively. Referring to [63], the specific operations of the global branch of FFC can be expressed as
  • Apply real FFT to input features:
    real FFT R H × W × C C H × W 2 × C ,
    where H , W , and C represent the height, width, and number of channels of the tensor, respectively, and × represents multiplication, indicating that the dimensions of height, width, and channels are multiplied together to define the size of the feature map. R H × W × C and C H × W 2 × C represent a real-valued tensor with spatial dimensions H × W × C and a complex tensor with spatial dimensions H × W 2 × C , respectively. real FFT represents the real-valued Fourier transform, and a detailed explanation and derivation of FFT can be found in [64].
  • Concatenates real and imaginary parts:
    ComplexToReal ( C H × W 2 × C ) R H × W 2 × 2 C ,
    where ComplexToReal represents the concatenation of the real and imaginary parts of the input tensor.
  • Apply convolution operations in the frequency domain:
    Conv ( R H × W 2 × 2 C ) R H × W 2 × 2 C ,
    where Conv represents the convolution operation.
  • Apply inverse transform to recover a spatial structure:
    RealToComplex ( R H × W 2 × 2 C ) C H × W 2 × C , InverseRealFFT ( C H × W 2 × C ) R H × W × C ,
    where RealToComplex represents the conversion of a real-valued tensor into a complex-valued tensor for the application of the inverse transform. InverseRealFFT represents the application of the inverse real-valued FFT to restore the tensor to its real-valued form. Similarly, the detailed process of the inverse transform can be found in [64].
However, although the FFC can efficiently extract features via spectral transformation, it does not fully account for the relationship between the local and global feature branches. Specifically, since the two feature extraction branches in the FFC are independent, a simple concatenation of their outputs may lead to a loss of certain feature information. To address this, the first FFC in the FF-CDRB applies adaptive channel weighting to the outputs of both branches using the Squeeze-and-Excitation (SE) attention [65], thereby effectively adjusting the response levels of feature branches.
The feature input to the n -th FF-CDRB block is denoted as f n R H × W × C , n { 1 , 2 , , R } . The output of f n after processing through the first FFC is represented as f n 1 . The specific operation for adaptive channel weighting on feature f n 1 is
  • Apply global average pooling to compress f n 1 into a 1 × 1 vector V f to capture global information:
    Squeeze f n 1 R H × W × C V f R 1 × 1 × C ,
    where Squeeze represents the result of compressing the features using global average pooling.
  • Map the features of V f through a learnable fully connected layer, and finally apply the Sigmoid function to obtain the adaptive weight vector w :
    Sigmoid F u l l y C o n n e c t e d L a y e r V f R 1 × 1 × C w R 1 × 1 × C ,
    where F u l l y C o n n e c t e d L a y e r represents a learnable fully connected layer used to capture feature information in the compressed vector, while Sigmoid represents the sigmoid function, which normalizes the final result into the weight vector w.
  • Use the channel-adaptive weights w to adjust the response levels of the local and global branch features in f n 1 , ultimately obtaining the feature representation with adaptive weight adjustments:
    F scale f n 1 , w f ¯ n 1 R H × W × C ,
where F scale f n 1 , w represents the result of element-wise multiplication, where the weight w is applied to f n 1 .
The adaptively weighted feature f ¯ n 1 is fed into the second FFC in the FF-CDRB block for further content inference. The output after processing by the second FFC is denoted as f ¯ n 2 . Inspired by the residual design in ResNet, a dual-path residual mechanism is incorporated in the FF-CDRB to simplify module learning and mitigate potential feature degradation caused by the serial FFC structure, thereby enhancing the module’s expressive capability. Specifically, in the FF-CDRB, the initial input f n and the feature f n 1 processed by the first FFC are merged through skip connections into the final output to optimize feature representation. The final output of the n -th FF-CDRB block can be expressed as
f n F = f n f n 1 f ¯ n 2 ,
where ⊕ represents element-wise addition.
In summary, the content inference process in FCAN can be described as follows:
f 1 F = F F C D R B 1 f N f 2 F = F F C D R R 2 f 1 F f R F = F F C D R B R f R 1 F ,
where FF CDR B n , n { 1 , 2 , , R } denotes the n -th FF-CDRB block in the FCAN.

3.4. The Context Aggregation and Decoding Stage

In the context aggregation and decoding stage, we integrate the proposed ACAM into the decoding process to aggregate multi-stage heterogeneous features, achieving more comprehensive contextual relationship interaction. We utilize N decoding blocks, Decoder i , i { 1 , 2 , , N } , constructed by stacking transposed convolutions, batch normalization, and ReLU activation function in sequence. Next, a detailed explanation of the proposed ACAM will be presented.

Asymmetric Context Aggregation Mechanism

Since texture patches provide limited information guidance, generating clothing content that aligns semantically with these texture patches becomes challenging, especially when the texture patches are small. To ensure the model can fully utilize the information from texture patches and generate semantically consistent image content under varying patch sizes, we propose a novel Asymmetric Context Aggregation Mechanism (ACAM).
The ACAM is designed based on our consideration of the characteristics of heterogeneous feature representations: compared to the features in the decoding stage, those in the encoding stage demonstrate a superior capacity for accurately capturing the initial guiding information as the progressive processing of features inevitably leads to a certain degree of representation loss. In contrast, decoding features, through the FF-CDRB blocks, establish preliminary global relationships, providing a more comprehensive interpretation of the fused multi-source features. Therefore, ACAM first constructs global contextual attention using decoding features and extracts initial guidance information representations from the multi-scale features of the encoding stage. It then uses the constructed attention scores to complete the aggregation of encoding and decoding features.
It is important to highlight that we observed suboptimal performance when constructing ACAM using cross-stage features of the same size. We believe the primary reason lies in the shared commonalities in information representation among features of identical size, which results in less expressive information compared to features of heterogeneous sizes. Based on this observation, we utilize cross-stage heterogeneous-sized features to implement ACAM.
Specifically, the features from the decoding stage are denoted as F m , m { N , N 1 , , 1 } , where F N represents the deepest features, and F N = f R F . As shown in the upper part of Figure 2, ACAM first extracts patches from the decoding stage features F m and calculates the cosine similarity between these patches:
Sim i , j = k i k i , k j k j ,
where k i and k j are the i -th and j -th patch vectors extracted from F m , respectively, and k i and k j represent the Euclidean norms of k i and k j , respectively. S i m i , j is the cosine similarity fraction between k i and k j . Considering that the encoding stage features contain semantic representations of the sketch, ACAM focuses on aligning the global information in F m with the texture content in f m 1 to more effectively utilize the limited texture information in the texture patches. Specifically, ACAM uses the texture patch region mask M t to obtain the cosine similarity scores between the internal and external segments of the texture patches in the feature map:
Sim i , j = Sim i , j P M ,
where P M is the patch vector extracted from M t , and ⊙ represents element-wise multiplication. Then, Softmax is applied to the Sim i , j to obtain the attention weight vector A i , j :
A i , j = e Sim i , j i = 1 H e Sim i , j ,
where H is the total number of patches extracted from F m . Since context enhancement is not required in the back area outside the clothing area, ACAM uses the clothing area binary mask to limit context attention enhancement to the clothing area to avoid content overflow:
A m = A m M c ,
where A m is the attention weight matrix composed of A i , j , and M c is the binary mask of clothing area. Subsequently, ACAM uses A m and the texture patch feature segments from f m 1 to aggregate heterogeneous features, obtaining the context-enhanced features F m . The specific operations are as follows:
k h = A m · l = 1 N I k l ,
where k h represents the h-th patch vector extracted from the regions outside the texture patch in f m 1 , and k l denotes the l-th patch from the regions within the texture patch in f m 1 , where l 1 , , N I , h 1 , , N O . N I and N O are the total numbers of patches extracted from within and outside the texture patch region in f m 1 , respectively. The context-enhanced feature F m is ultimately obtained by concatenating k h and k l .
As shown in the upper part of Figure 2, based on [56], ACAM refines the enhanced feature F m using a set of dilated convolutions [66] with different dilation rates, and the features refined through multi-scale processing using dilated convolutions are denoted as F ¯ m . The specific operations are as follows:
F ¯ m = Concat ( C o n v 1 ( F m ) , C o n v 2 ( F m ) , , C o n v d ( F m ) , ) , d { 1 , 2 , , D } .
However, although the parallel dilated convolution design provides richer scale information, it may also introduce local information loss that can impair the final feature representation. To address this, an adaptive gating mechanism is introduced in ACAM to adaptively coordinate F m with the refined residual features F ¯ m . Specifically, an adaptive gate value G m is first obtained from F m through standard convolution and Softmax operations. G m is then used to balance the response levels of F m and F ¯ m , yielding the final output F m A . The specific operation is as follows:
F m A = G m · F m 1 G m · F ¯ m ,
where G m is
G m = Softmax ( Norm ( Conv ( F m ) ) ) ,
where G m represents the obtained gating weight vector, Softmax denotes the Softmax function, Norm represents normalization, and Conv denotes a learnable convolution.
Finally, the context-enhanced feature F m A is integrated into the decoding features during the decoding phase to achieve comprehensive context modeling. Similarly, F m A is incorporated into the asymmetric decoding step Decoder ( F m 1 ) to obtain richer information representations. The specific operation can be expressed as follows:
Decoder ( F m 1 ) = Decoder ( F m 1 ) F m A .
The processing of the context aggregation and decoding stage can be represented as follows:
F N 1 = Decoder 1 F N F N 2 = Decoder 2 F N 1 ACAM F N , f N 1 F 2 = Decoder N 2 F 3 ACAM F 4 , f 3 F 1 = Decoder N 1 F 2 ACAM F 3 , f 2 C gen = Decoder N F 1 ,
where ACAM represents the context aggregation operation within the ACAM, Decoder i denotes the decoding block based on transposed convolution, and C gen represents the final generated image.
To provide a clearer explanation of the internal modeling details of ACAM, we conducted further analysis and interpretation of its feature modeling process using feature visualization techniques. For illustrative purposes, we take FCAN, which consists of four encoding blocks and four decoding blocks, as an example to present the visualization results. Figure 6 shows the visualization results of multi-scale features during the encoding phase. As shown in the figure, the shallow features in the encoding stage are more adept at capturing low-level information such as textures and edges, while deeper features are capable of extracting richer high-level semantic information.
Next, we proceed based on the ACAM workflow described in this section. Figure 7 presents the visualization results of the context attention maps constructed by ACAM, where A 4 and A 3 represent the context attention maps constructed from F 4 and F 3 , respectively. As shown in Figure 7, the context attention maps extracted from the deeper decoding features enhance global high-level information, while those extracted from the shallower features focus more on the expression of local structures and texture details. Thus, ACAM leverages cross-level and cross-stage heterogeneous structures for context aggregation, thereby enhancing the model’s ability to represent diverse feature details.
Figure 8 presents the visualization results of the context-enhanced features F 4 A and F 3 A obtained using A 4 and A 3 , respectively. As indicated, F 4 A achieves a more comprehensive aggregation of global information, while F 3 A provides a more detailed interpretation of local features. These results visually demonstrate the effectiveness of ACAM in heterogeneous feature aggregation.
Finally, to facilitate a clear understanding of our overall process, we summarize the operation of our FCAN in the form of pseudocode. It is presented in Algorithm 1.
Algorithm 1 Algorithm of the FCAN
Input: Sketch S, Clothing Mask M c , and Texture Patch T
Output: Synthesized Clothing Image C g e n
1:
Obtain the clothing region positional information P via the sine–cosine encoding technique in [61]
2:
Obtain the fused features as input f 0 :
f 0 = Concat(S,M,T)
3:
Obtain the multi-scale features f i , i [ 1 , N ] :
4:
for each i [ 1 , N ]  do
5:
if  i = 1  then
6:
   f 1 = E n c o d e r 1 ( f 0 ,P)
7:
else
8:
   f i = E n c o d e r i ( f i 1 )
9:
end if
10:
end for
11:
Content synthesis using R sequential FF-CDRB modules, and obtain the features F N :
12:
for each n [ 1 , R ]  do
13:
if  n = 1  then
14:
   f 1 F = F F C D R B 1 ( f N )
15:
else
16:
   f n F = F F C D R B n ( f n 1 F )
17:
end if
18:
end for
19:
Decode the feature maps and use ACAM for context interaction between heterogeneous features:
20:
F N = f R F
21:
for each i [ N , 1 ]  do
22:
if  i = N  then
23:
   F N 1 = D e c o d e r 1 ( F N )
24:
else if i = 1 then
25:
   C g e n = D e c o d e r N ( F 1 )
26:
else
27:
   F N i = D e c o d e r i ( F N i + 1 ) ⊕ ACAM( F N i + 2 , f N i + 1 )
28:
end if
29:
end for
30:
return  C gen

3.5. Loss Functions

Due to the complexity of apparel content, training the generative model presents certain challenges. Based on [56,58,61], we employ adversarial loss L a based on PatchGAN [67], perceptual loss L p based on VGG [68], style loss L s based on Gram matrices [69], and pixel-level loss L m to jointly train FCAN.
The pixel-level loss L m guides the network to learn the underlying data distribution, enabling it to acquire fundamental image synthesis capabilities. Its specific computation is given by
L m = C o r i C g e n ,
where C ori and C gen represent the real image and the generated image, respectively.
The pixel-level reconstruction loss based on Absolute Value Error cannot accurately guide the network to reconstruct apparel content as the style information provided by texture patches is limited. Thus, relying solely on L m may lead the network to generate unrealistic content in challenging regions of the garment. It is important to note that the goal of our task is not to fully replicate the exact apparel content but rather to generate apparel images that are semantically consistent with the guiding information and contextually reasonable. To achieve this, we employ perceptual loss based on the pre-trained VGG network and style loss based on Gram matrices to help the network produce semantically consistent apparel content. The perceptual loss calculates the differences between images through a pre-trained neural network. Compared to Absolute Value Error, perceptual loss aligns more closely with human perception of images.
We calculate the perceptual loss L p using a pre-trained VGG network as follows:
L p = i 1 U i ϕ i C o r i ϕ i C g e n ,
where ϕ i C ori and ϕ i C gen represent the features extracted from the i-th layer of VGG for the real image and the generated image, respectively, and U denotes the total number of layers in the VGG network.
The style loss is used to ensure that the generated images have a consistent overall style with the real images, avoiding generating plausible but seriously misaligned results in terms of style semantics. First, the pre-trained VGG is used to calculate the Gram matrix:
G ϕ l ( C ) = 1 U l Y l i = 1 U l j = 1 Y l ϕ l ( C ) i , j · ϕ l ( C ) j , i ,
where U l and Y l are the number of channels and spatial dimensions of the features, respectively. C is the input image. The style loss can be expressed as
L s = l G ϕ l C o r i G ϕ l C g e n .
In order to improve the quality of the generated image, avoid blur in the generated image, and simplify the design details of the loss function, we introduced PatchGan-based adversarial loss for adversarial training. PatchGAN divides the image into several small local patches and discriminates the truth and falsity of each local patch to help the generation network improve the quality of local texture generation. The discriminator loss based on PatchGAN is
L D = E x p data C o r i i , j log D i , j C ori E z p z ( z ) i , j log 1 D i , j C g e n ,
where D i , j ( C o r i ) is the output of the discriminator on the real sample on patch ( i , j ) , and D i , j ( C g e n ) is the output of the discriminator on the generated sample on patch ( i , j ) . The goal of the generative network is to fool the discriminator on as many patches as possible. The loss of the generated network is
L G = E z p z ( z ) i , j log D i , j ( G ( C g e n ) ) .
To sum up, the adversarial loss is
L a = λ g L G + λ d L D ,
where λ d and λ D are generator loss weight coefficients and discriminator loss weight coefficients, respectively. The total losses in training are
L total = L a + λ 1 L p + λ 2 L s + λ 3 L m ,
where λ 1 , λ 2 , and λ 3 are the weight coefficients of the corresponding losses.

4. Experiments

4.1. Dataset

To the best of our knowledge, there is currently no publicly available dataset featuring diverse categories of paired design sketches and fashion images. To address this gap, we constructed a dataset named SKFashion, based on [70], which includes a wide range of fashion categories. This dataset enables the evaluation of our method’s performance in generating images for various types of apparel.
Additionally, to further validate the robustness of our proposed approach under different conditions, we conducted extensive experiments on two publicly available datasets: VITON-HD [71] and Fashion-Gen [72].
  • SKFashion: The SKFashion dataset consists of 20,421 paired design sketches and fashion images, each with a resolution of 256 × 256 pixels. It covers dozens of fashion categories, including long sleeves, short sleeves, dresses, shorts, skirts, padded coats, jackets, and vests. Figure 9 presents a selection of image pairs from SKFashion. Following the setup in [56], we divided the SKFashion dataset into a training set containing 17,396 image pairs and a test set comprising 3025 image pairs.
  • VITON-HD: The VITON-HD dataset includes images of various upper-body clothing categories and is widely used in fashion image generation tasks such as virtual try-on. We standardized the resolution of the clothing images in VITON-HD to 192 × 256 pixels and utilized the garment structure edge maps extracted using DexiNed (Dense Extreme Inception Network for Edge Detection) [73] as design sketches. We divided the dataset into a training set comprising 10,309 sketch–clothing image pairs and a test set with 1010 pairs. Figure 10 illustrates a selection of image pairs from the VITON-HD dataset.
  • Fashion-Gen: The Fashion-Gen is a large-scale multimodal dataset designed for fashion-related tasks and is widely used in multimodal fashion image generation. We extracted clothing images from Fashion-Gen using [74] and generated corresponding design sketches by applying DexiNed to obtain structural edge maps of the garments. Following the setup in [53,56,57], we randomly sampled 10,282 sketch–clothing image pairs as the training set and 800 pairs as the test set. Figure 11 showcases a selection of image pairs from the Fashion-Gen dataset.

4.2. Evaluation Metrics

We conduct a comprehensive performance evaluation of the proposed method using five popular evaluation metrics:
  • Fréchet Inception Distance (FID): FID [75] is an objective metric widely used to evaluate the quality of images generated by generative models. It measures the distance between the distributions of generated and real images using a pre-trained Inception network, quantifying the perceptual similarity between them. By considering both the global feature distribution and perceptual similarity, FID effectively assesses the overall quality of generated images, aligning closely with human visual perception. We use FID to evaluate the overall diversity and global perceptual quality of the generated images. The calculation of FID is as follows:
    F I D = μ g μ r 2 + Tr Σ g + Σ r 2 Σ g Σ r 1 2 ,
    where μ g and Σ g represent the mean vector and covariance matrix of the generated image features, respectively, while μ r and Σ r represent the mean vector and covariance matrix of the real image features, respectively. μ g μ r 2 represents the square of the Euclidean norm of the difference between two mean vectors. Tr represents the trace of a matrix.
  • Structural Similarity (SSIM): SSIM [76] is an evaluation metric used to measure the similarity of two images. SSIM focuses on the brightness, contrast, and structure of the image. We use SSIM to evaluate the quality of the generated image in terms of local structure and so on. SSIM is calculated as follows:
    SSIM ( x , y ) = 2 μ x μ y + C 1 2 σ x y + C 2 μ x 2 + μ y 2 + C 1 σ x 2 + σ y 2 + C 2 ,
    where x and y represent the real image and the generated image, respectively. μ denotes the mean pixel value, σ represents the standard deviation of grayscale intensity, C 1 and C 2 are constant coefficients, and σ x y is defined as
    σ x y = 1 N 1 i = 1 N x i μ x y i μ y .
  • Learned Perceptual Image Patch Similarity (LPIPS): LPIPS [77] is a metric used to evaluate the perceptual similarity between two images. It assesses image quality by directly extracting features from a pre-trained neural network, making it more aligned with human visual perception. LPIPS places greater emphasis on local perceptual similarity, and we employ it to evaluate the local perceptual quality of generated images. The calculation of LPIPS is as follows:
    LPIPS ( x , y ) = l 1 H l W l h , w w l ϕ ^ x h w l ϕ ^ y h w l 2 ,
    where ϕ ^ x h w l and ϕ ^ y h w l represent the normalized corresponding features, and w l is the scaling weight vector.
  • Peak Signal-to-Noise Ratio (PSNR): PSNR is a pixel-level image quality metric that is sensitive to the random noise in the image. We use PSNR to evaluate the pixel-level quality of the generated image. The calculation formula of PSNR is
    P S N R = 10 × log 10 M A X I 2 M S E ,
    where M A X I represents the maximum pixel value of the image, and M S E is the average value of the RGB three-channel mean square error of the real image and the generated image.
  • TOPIQ: TOPIQ is a deep-learning-based image quality assessment method that utilizes multi-scale features and cross-scale attention mechanisms to evaluate image quality. By simulating the characteristics of the human visual system, TOPIQ leverages high-level semantic information to guide the network’s focus on semantically important local distortion regions, thereby assessing image quality. TOPIQ includes both Full-Reference (FR) and No-Reference (NR) forms. In this paper, we use them to evaluate the perceptual quality of the generated images. The specific details of TOPIQ can be found in [78].

4.3. Baseline Methods

The proposed method is extensively compared with five advanced existing methods: ZITS [61], PAINT [56], MST [79], Albahar-MT [80], and EdgeConnect [81]. Please note that, since the method proposed by Albahar et al. is not specially named in [80], the method is uniformly named as Albahar-method (Albahar-MT) in this paper. We re-trained all the baseline methods on the corresponding dataset. For fair comparison, the hyperparameters in our experiment follow the default settings in the corresponding paper.

4.4. Implementation Details

In our experiments, training is conducted using a fixed learning rate of 1 × 10 4 , with the Adam optimizer ( β 1 = 0.5 , β 2 = 0.999 ) to optimize the network parameters until the model converges. We initialized the loss weights in our experiments based on the settings in Refs. [56,61,79] as the weights in these papers performed well on similar tasks. To further optimize, we performed a grid search within a 50% range of the initial weights to ensure the selection of the most appropriate weight combination, ultimately determining the final weights. The loss weights in the experiments are set as λ g = 0.1 , λ d = 0.5 , λ 1 = 0.5 , λ 2 = 50 , and λ 3 = 6 . Considering a balance between performance and computational cost, we set the number of FF-CDRB blocks in the FCAN as R = 6 and the number of encoding and decoding blocks as N = 4 . While deeper network architectures and additional context aggregation operations may lead to improved performance, they would also increase computational resource consumption.
Additionally, we follow the operations in [56,80], where randomly sized texture patches are cropped from the clothing areas as texture patches. To ensure fairness in the experiments, the same random seed is used across all experiments to guarantee consistency in the texture content.
FCAN was trained for 100 epochs with a batch size of 16 on the SKFashion and VITON-HD datasets, and for 100 epochs with a batch size of 8 on the Fashion-Gen dataset.
We adopted the texture patch size specified in [56], setting 60–70 pixels as the standard texture patch size. Based on this standard, the texture patch size was reduced by approximately 20–30% and 40–50% to represent medium-sized and small-sized texture patches, respectively. Extensive experiments were conducted under varying texture patch sizes to evaluate the performance of different generation methods under diverse conditions.
The primary software environment used in the experiments is Python 3.8 (Ubuntu 20.04) and PyTorch 2.0.0. The main hardware environment consists of a single RTX 4090 (24 GB) GPU, Xeon(R) Platinum 8352 V CPU, and 120 GB of memory.

4.5. Comparison with Existing Approaches

4.5.1. Quantitative Comparison

Table 1, Table 2 and Table 3 illustrate the quantitative comparison results between our method and existing approaches. It is worth noting, as discussed in [77], that, while PSNR and SSIM are effective in objectively measuring pixel-level image quality, their outcomes may sometimes contradict human perceptual judgments. For instance, certain meaningless blurs can lead to perceptual bias, which may not be adequately captured by PSNR and SSIM. Given that garment image generation tasks place greater emphasis on human perceptual judgment, our quantitative experiments prioritize perceptual quality metrics such as FID, LPIPS, and TOPIQ, using PSNR and SSIM as supplementary references for pixel-level image quality.
Table 1 presents the results of the quantitative comparison experiments conducted on the SKFashion dataset. The results show that, under different texture patch size conditions, our method outperforms the baseline methods in terms of FID and LPIPS metrics while achieving comparable results in PSNR and SSIM. This indicates that the garment images generated by our method exhibit superior perceptual quality and also perform well in terms of pixel-level quality.
As the texture patch size decreases, all methods experience some degree of performance fluctuation. Among them, PAINT [56], which models context based on a single texture patch, shows the poorest robustness to changes in texture patch size, resulting in the most significant metric fluctuations. ZITS [61] demonstrates the closest robustness to texture patch size variation compared to our method as it introduces positional encoding techniques to improve performance under small texture patch conditions.
Additionally, it is noteworthy that, as shown in Table 4, our approach achieves superior performance while utilizing only approximately one-third to one-half of the computational parameters required by the baseline methods. This underscores the efficiency and lightweight nature of our method.
Table 2 presents the quantitative comparison results on the VITON-HD dataset. Our proposed FCAN achieves the best performance across all evaluation metrics. Notably, the VITON-HD dataset contains approximately 40% less data compared to the SKFashion dataset, making the evaluation results more sensitive to the texture patch size. Under conditions of small texture patch sizes, only our proposed FCAN and ZITS maintain consistently high performance, while the performance of other baseline methods fluctuates significantly.
Table 3 presents the quantitative comparison results on the Fashion-Gen dataset, showing that our method achieves the best performance across all metrics, further validating the effectiveness of our approach. Among all the baseline methods, MST [79] achieves performance metrics closest to FCAN under both standard and medium texture patch sizes. However, under small texture patch sizes, the performance of MST [79] deteriorates significantly, indicating that it lacks robustness to texture patch size and struggles to adapt to the varying requirements of different practical tasks.
It is important to note that the clothing size ratio in Fashion-Gen is smaller compared to SKFashion and VITON-HD, and the dataset contains fewer clothing categories with simpler garment structures. As a result, under conditions with standard and medium-sized texture patches, the SSIM scores of all methods are relatively close. However, although the data scales of the VITON-HD and Fashion-Gen datasets are nearly half that of the SKFashion, our method still maintains robust performance on both datasets. This suggests that, in comparison to the baseline methods, our approach exhibits enhanced robustness to the scale of training samples, thereby ensuring the method’s reliability.
Table 4 presents a comparison of the model parameter counts for each method, showing that our approach has the smallest number of parameters. Combined with the results of the quantitative experiments, this demonstrates that our method is highly efficient in terms of parameter usage, achieving optimal performance with the fewest computational parameters.

4.5.2. Qualitative Comparison

Figure 12, Figure 13 and Figure 14 present the qualitative comparison results on the SKFashion dataset. To facilitate comparison, we have zoomed in on certain details in the visual comparisons. Please note that the annotated details are not the only noticeable differences in performance. For better overall comparison, we recommend zooming in on the images. The qualitative comparison results on the VITON-HD and the Fashion-Gen will be presented in the appendix.
As shown in Figure 12, under small texture patch sizes, the images generated by our method exhibit significant advantages over the baseline methods in terms of texture details, semantic consistency, and other aspects. The images generated by the baseline methods suffer from issues such as style leakage (e.g., black-and-white texture spillover on the striped top in the first row), color mismatches and artifacts (e.g., color discrepancies and meaningless artifacts on the blue coat in the second row and the denim skirt in the fourth row), as well as structural and semantic loss (e.g., missing waistband on the red sports shorts in the third row). These issues indicate that the baseline methods perform poorly in integrating guidance information and content inference.
As shown in Figure 13, as the texture patch size increases, the quality of the images generated by all methods improves to some extent. However, when generating densely repetitive textures and large-scale texture propagation, the baseline methods still encounter issues such as texture detail loss (e.g., texture propagation failure on the yellow knit vest in the first row of Figure 13) and inconsistency in dense texture content (e.g., loss of dense white dots on the green and white patterned top in the third row). Furthermore, the clothing images generated by the baseline methods still exhibit artifacts and color discrepancies (e.g., in the second and fourth rows).
Figure 14 presents the qualitative comparison results under the standard texture patch size condition on the SKFashion dataset. As the texture patch size increases, the quality of the generated images improves across all methods. However, the baseline methods still struggle when generating dense texture content, exhibiting issues such as texture loss or overflow.
Leveraging the advancements in content reasoning and contextual modeling introduced by ACAM and FF-CDRB, our method seamlessly integrates the structural semantics of design sketches with the stylistic content of texture patches, thereby establishing coherent and consistent global contextual relationships. As a result, our method maintains stable performance across all three texture patch sizes. The generated images demonstrate significant advantages over the baseline methods in terms of color consistency, texture details, and structural-semantic alignment while also showing robust performance with respect to variations in texture patch size.
Figure A1, Figure A2, Figure A3, Figure A4, Figure A5 and Figure A6 in the appendix present the experimental results on the VITON-HD and Fashion-Gen datasets. The results demonstrate that our method achieves excellent performance on both datasets. Even with a significantly reduced training dataset, our approach can still generate high-quality clothing images, guided by sketches and texture patches.
Extensive comparative experiments conducted across the three datasets show that our method offers significant advantages in terms of objective performance metrics, visual quality of the generated images, and model size.

4.5.3. Ablation Studies

We conducted extensive ablation experiments to assess the importance of each component in our method and the performance impact it provides.
Table 5, Table 6 and Table 7 present the results of the ablation experiments conducted on three datasets. Specifically, “w/o Dual residual” indicates the removal of the dual-path residual connections in FF-CDRB, “w/o ACAM” means the removal of ACAM, “w/o Gating” refers to the removal of the adaptive gating mechanism in ACAM, “ResNet+ACAM” denotes the use of standard ResNet residual blocks as a replacement for FF-CDRB in the content inference module of FCAN, and “Skip Connection” denotes replacing ACAM with conventional skip connection.
FF-CDRB: The ablation experiment results show that the proposed FF-CDRB effectively improves the FID, LPIPS, and TOPIQ metrics of the generated images, demonstrating its superiority in content inference. As mentioned in our qualitative comparison experiments, the clothing structures in the Fashion-Gen dataset are relatively simple, and the clothing regions occupy a smaller area, which is why the SSIM metric on the Fashion-Gen dataset is relatively similar.
Table 8 presents a comparison of the model’s computational parameter count after ablating each component. The results show that the FF-CDRB, compared to conventional convolution-based residual blocks, not only provides performance improvements but also reduces the model’s computational parameters, achieving more efficient and lightweight content inference.
ACAM: The experimental results validate the effectiveness of the proposed ACAM in context modeling. The ACAM significantly improves the performance of our method across various metrics. More comprehensive and effective context relationship modeling leads to a notable enhancement in the perceptual quality of the generated images. Furthermore, the ablation experiment results show that, as the texture patch size decreases, ACAM’s ability to aggregate global information becomes more prominent, demonstrating that it can effectively enhance the model’s robustness under extreme texture patch size conditions. Furthermore, Table 8 shows, that while ACAM effectively enhances the performance, it does not introduce significant additional computational parameters. This is because the proposed ACAM constructs global context attention based on the heterogeneous features after inference, enabling explicit context aggregation.

4.5.4. Time Performance

We evaluated the training and inference time of the proposed FCAN on the SKFashion dataset and compared it with baseline methods. The time costs for training and inference are shown in Figure 15.
The FCAN we proposed is on par with the baseline methods in terms of training and inference time consumption. Given the performance improvement, we believe that a slight increase in time cost is acceptable. In our future work, we will continue to optimize the model structure and hyperparameter settings to ensure that similar results are achieved with minimal cost.

5. Conclusions

This paper proposes a novel clothing image generation method driven by texture blocks and design sketches. At the same time, it provides a comprehensive review of numerous related techniques, methods, and the current state of the research in this field. Starting from issues such as the lack of controllability and poor image quality in the existing methods, we highlight the advantages of our approach.
Our method is driven by texture blocks and design sketches, ensuring the maximum subjective control over the generated content, thereby guaranteeing its potential application value. In response to the issue of poor image quality in the current related research, we also introduce the innovative ACAM and FF-CDRB components, which enable our model to achieve significant performance improvements in multiple aspects. Specifically, considering the importance of contextual relationship modeling in the field of computer vision, we believe that the heterogeneous feature modeling approach proposed by the ACAM component can provide valuable insights for the research community.
In this paper, we report the comparative experiments conducted between our proposed method and five advanced baseline methods under different texture block sizes. In the quantitative experiments, our method achieved the best FID, PSNR, LPIPS, and TOPIQ scores under the standard, medium, and small texture block conditions of the SKFashion dataset while also yielding comparable SSIM and TOPIQ-NR results. Specifically, under the three texture block conditions, the FID metric showed improvements of approximately 23.2%, 22.4%, and 19.7%, respectively, over the method with the second-best performance, while LPIPS showed improvements of about 13.5%, 11.8%, and 12.8%. In the publicly available Fashion-Gen and VITON-HD datasets, our method achieved improvements of approximately 5–18% in metrics such as FID, LPIPS, and PSNR compared to suboptimal methods. We also conducted extensive ablation experiments to demonstrate the effectiveness of the design of each component in the model.
In addition, we performed extensive qualitative comparison experiments, the results of which demonstrated that our method is capable of generating high-quality clothing images under various texture block conditions and outperforms the baseline methods in terms of semantic consistency and fine-grained expression. Notably, while achieving these performance improvements, our method maintains efficient computational parameters, with the fewest model parameters among all the methods tested. We believe this feature holds significant potential for local deployment on multiple devices. These extensive experimental results highlight the progress our method has made in the controllable clothing image domain, and we are confident that it provides valuable insights for research in related fields.
In summary, this method offers an innovative solution for various tasks in fashion design and related fields, effectively reducing designers’ workload during the early stages of prototyping. It also has a positive impact on transforming production methods in the fashion industry and advancing its sustainable development.

6. Discussion

Although our method has shown promising advancements in controllability and the quality of generated images, there are still several factors worth discussing regarding both the strengths and limitations of our approach.
One of the key advantages of the method we propose is its ability to provide fine-grained control over generated clothing images, which represents a significant improvement over the traditional methods. By leveraging texture blocks and design sketches as driving forces, we are able to achieve high customization while maintaining consistency in the generated content. This level of control is particularly important for practical applications in fashion design, where ensuring semantic coherence and meeting design specifications are crucial. However, this enhanced controllability can, in some cases, limit the model’s generative capabilities. Specifically, while our method excels in controlled scenarios, it may not be as effective in generating highly diverse clothing styles as other methods based on pre-trained generative models, which are designed to produce a wider range of outputs. This limitation highlights the challenge of balancing control and diversity in generative models.
Another key contribution of our work is the introduction of the ACAM components, which have played a crucial role in enhancing the model’s performance. The ACAM focuses on heterogeneous feature modeling, which helps to establish contextual relationships in the generated images, leading to more realistic and semantically consistent clothing representations. The exploration of such methods is a focal point in this field as contextual modeling is often critical in computer vision tasks. However, despite its advantages, the ACAM still faces some challenges. In certain cases, especially when the design sketch is not sufficiently detailed or when complex clothing features are involved, the model may struggle to generate highly accurate or realistic textures and patterns. This suggests that further improvements to the ACAM component may be needed to enhance the model’s performance in these more challenging scenarios.
At the same time, time cost is also a key concern for generative models. Currently, our method’s training and inference time costs are on par with the existing methods and, in some cases, slightly higher. While we believe that a modest increase in time cost for significant performance improvement is acceptable, it could still impact the future development of the model. Therefore, how to significantly reduce time cost while improving performance remains a direction worth exploring.
In summary, our method still holds significant potential for future improvements. Potential future research will focus on, but is not limited to, the following areas:
  • Optimizing the acquisition process of sketch and texture data to enhance the practical value of the method;
  • Improving its synthesis performance under more extreme guidance conditions to further broaden its applicability;
  • Improving time efficiency without sacrificing performance or while enhancing performance;
  • Exploring cross-applications with related tasks such as virtual try-on, clothing retrieval, and recommendation systems to expand its use cases.
Moving forward, we will continue to investigate clothing image synthesis technologies based on artificial intelligence, building upon these aspects.

Author Contributions

Conceptualization, H.L. and Y.H.; methodology, H.L. and Y.H.; software, Y.H.; validation, Y.H., M.D., and Z.L.; formal analysis, G.L.; investigation, Y.H.; resources, H.L.; data curation, Y.H.; writing—original draft preparation, Y.H.; writing—review and editing, H.L.; visualization, H.L.; supervision, H.L.; project administration, M.W.; funding acquisition, H.L. and G.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research is jointly supported by the Natural Science Foundation of Jiangxi Province (20224BAB202018), National Natural Science Foundation of China (62462032, 62266023), Key Research and Development Program of Jiangxi Province (20223BBE51039, 20232BBE50020), and Science Fund for Distinguished Young Scholars of Jiangxi Province (20232ACB212007).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The VITON-HD and Fashion-Gen datasets used in this study are accessible via sources https://github.com/shadow2496/VITON-HD (accessed on 1 October 2024) and https://fashion-gen.com (accessed on 1 October 2024), respectively. Since we are still conducting other related research, our self-constructed SKFashion dataset is temporarily unavailable to the public. Further inquiries can be made by contacting our first or second author.

Acknowledgments

This research was supported by multiple projects, including the National Natural Science Foundation of China, for experimental equipment, office supplies, and other related expenses.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Figure A1. Qualitative comparison results under small texture patch sizes on the VITON-HD dataset.
Figure A1. Qualitative comparison results under small texture patch sizes on the VITON-HD dataset.
Applsci 15 03534 g0a1
The qualitative experimental comparison results on the VITON-HD dataset are shown in Figure A1, Figure A2 and Figure A3. On the VITON-HD dataset, the baseline methods exhibit significant color discrepancies and artifacts under both small and medium texture patch sizes, whereas our method generates clothing images with a clear visual quality advantage. When the texture patch size is increased to the standard size, the baseline methods suffer from loss of details when handling complex, dense textures (as seen in the first and third rows of Figure A3, where dense textures fail to propagate properly to the clothing area). In contrast, our method effectively handles dense texture content.
Figure A2. Qualitative comparison results under medium texture patch sizes on the VITON-HD dataset.
Figure A2. Qualitative comparison results under medium texture patch sizes on the VITON-HD dataset.
Applsci 15 03534 g0a2
Figure A3. Qualitative comparison results under standard texture patch sizes on the VITON-HD dataset.
Figure A3. Qualitative comparison results under standard texture patch sizes on the VITON-HD dataset.
Applsci 15 03534 g0a3
Figure A4, Figure A5 and Figure A6 present the qualitative comparison results on the Fashion-Gen dataset. Similar to the VITON-HD dataset, the clothing structures in the Fashion-Gen dataset are relatively simple, and the clothing images generated by all methods maintain consistent structural semantics with the design sketches. However, on the Fashion-Gen dataset, the baseline methods still exhibit significant issues with color discrepancies, artifacts, and loss of texture details. In contrast, our method demonstrates stable performance on the Fashion-Gen dataset, producing clothing images with the best visual quality under all three texture patch sizes.
Figure A4. Qualitative comparison results under small texture patch sizes on the Fashion-Gen dataset.
Figure A4. Qualitative comparison results under small texture patch sizes on the Fashion-Gen dataset.
Applsci 15 03534 g0a4
Figure A5. Qualitative comparison results under medium texture patch sizes on the Fashion-Gen dataset.
Figure A5. Qualitative comparison results under medium texture patch sizes on the Fashion-Gen dataset.
Applsci 15 03534 g0a5
Figure A6. Qualitative comparison results under standard texture patch sizes on the Fashion-Gen dataset.
Figure A6. Qualitative comparison results under standard texture patch sizes on the Fashion-Gen dataset.
Applsci 15 03534 g0a6

References

  1. Guo, Z.; Zhu, Z.; Li, Y.; Chen, S.C.H.; Wang, G. AI Assisted Fashion Design: A Review. IEEE Access 2023, 11, 88403–88415. [Google Scholar] [CrossRef]
  2. Oh, J.; Ha, K.J.; Jo, Y.H. A Predictive Model of Seasonal Clothing Demand with Weather Factors. Asia-Pac. J. Atmos. Sci. 2022, 58, 667–678. [Google Scholar] [CrossRef]
  3. Linlin, L.; Haijun, Z.; Qun, L.; Jianghong, M.; Zhao, Z. Collocated Clothing Synthesis with GANs Aided by Textual Information: A Multi-Modal Framework. ACM Trans. Multimedia Comput. Commun 2024, 20, 25. [Google Scholar]
  4. Lampe, A.; Stopar, J.; Jain, D.K.; Omachi, S.; Peer, P.; Štruc, V. DiCTI: Diffusion-based Clothing Designer via Text-guided Input. In Proceedings of the 2024 IEEE 18th International Conference on Automatic Face and Gesture Recognition (FG), Istanbul, Turkiye, 27–31 May 2024; pp. 1–9. [Google Scholar]
  5. Zhengwentai, S.; Yanghong, Z.; Honghong, H.; Mok, P.Y. DiCTI: SGDiff: A Style Guided Diffusion Model for Fashion Synthesis. In Proceedings of the 31st ACM International Conference on Multimedia (MM ’23), Association for Computing Machinery, New York, NY, USA, 29 October–3 November 2023; pp. 8433–8442. [Google Scholar]
  6. Cao, S.; Chai, W.; Hao, S.; Zhang, Y.; Chen, H.; Wang, G. DiffFashion: Reference-Based Fashion Design With Structure-Aware Transfer by Diffusion Models. IEEE Trans. Multimed. 2024, 26, 3962–3975. [Google Scholar] [CrossRef]
  7. Kim, B.-K.; Kim, G.; Lee, S.-Y. Style-controlled synthesis of clothing segments for fashion image manipulation. IEEE Trans. Multimed. 2020, 22, 298–310. [Google Scholar] [CrossRef]
  8. Jiang, S.; Li, J.; Fu, Y. Deep Learning for Fashion Style Generation. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 4538–4550. [Google Scholar] [CrossRef]
  9. Di, W.; Zhiwang, Y.; Nan, M.; Jianan, J.; Yuetian, W.; Guixiang, Z.; Hanhui, D.; Yi, L. StyleMe: Towards Intelligent Fashion Generation with Designer Style. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems (CHI ’23). Association for Computing Machinery, New York, NY, USA, 23–28 April 2023; pp. 1–16. [Google Scholar]
  10. Dong, M.; Zhou, D.; Ma, J.; Zhang, H. Towards Intelligent Design: A Self-Driven Framework for Collocated Clothing Synthesis Leveraging Fashion Styles and Textures. In Proceedings of the ICASSP 2024–2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 3725–3729. [Google Scholar]
  11. Han, Y.; Haijun, Z.; Jianyang, S.; Jianghong, M.; Xiaofei, X. Toward Intelligent Fashion Design: A Texture and Shape Disentangled Generative Adversarial Network. ACM Trans. Multimed. Comput. Commun. 2023, 19, 23. [Google Scholar]
  12. Jianan, J.; Di, W.; Hanhui, D.; Yidan, L.; Wenyi, T.; Xiang, L.; Can, L.; Zhanpeng, J.; Wenlei, Z. Tangquan, Q. 2024. HAIGEN: Towards Human-AI Collaboration for Facilitating Creativity and Style Generation in Fashion Design. ACM Interact. Mob. Wearable Ubiquitous Technol. 2024, 8, 27. [Google Scholar]
  13. Han, Y.; Haijun, Z.; Xiangyu, M.; Jicong, F.; Zhao, Z. FashionDiff: A Controllable Diffusion Model Using Pairwise Fashion Elements for Intelligent Design. In Proceedings of the 31st ACM International Conference on Multimedia (MM ’23), Association for Computing Machinery, New York, NY, USA, 29 October–3 November 2023; pp. 1401–1411. [Google Scholar]
  14. Xian, W.; Sangkloy, P.; Agrawal, V.; Raj, A.; Lu, J.; Fang, C.; Hays, J. Texturegan: Controlling deep image synthesis with texture patches. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 8456–8465. [Google Scholar]
  15. Zhou, Y.; Chen, K.; Xiao, R.; Huang, H. Neural texture synthesis with guided correspondence. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 18095–18104. [Google Scholar]
  16. Phillips, A.; Lang, J.; Mould, D. Diverse non-homogeneous texture synthesis from a single exemplar. Comput. Graph. 2024, 124, 104099. [Google Scholar] [CrossRef]
  17. Houdard, A.; Leclaire, A.; Papadakis, N.; Rabin, J. A Generative Model for Texture Synthesis based on Optimal Transport Between Feature Distributions. J. Math. Imaging Vis. 2023, 65, 4–28. [Google Scholar] [CrossRef]
  18. Chi, L.; Jiang, B.; Mu, Y. Fast fourier convolution. Adv. Neural Inf. Process. Syst. 2023, 33, 4479–4488. [Google Scholar]
  19. Chen, Z.; Zhang, Y. CA-GAN: The synthesis of Chinese art paintings using generative adversarial networks. Vis. Comput. 2024, 40, 5451–5463. [Google Scholar] [CrossRef]
  20. Xue, A. End-to-end chinese landscape painting creation using generative adversarial networks. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Online, 5–9 January 2021; pp. 3863–3871. [Google Scholar]
  21. Wang, Q.; Guo, C.; Dai, H.N.; Li, P. Stroke-GAN Painter: Learning to paint artworks using stroke-style generative adversarial networks. Comput. Vis. Media 2023, 9, 787–806. [Google Scholar] [CrossRef]
  22. Gao, X.; Tian, Y.; Qi, Z. RPD-GAN: Learning to draw realistic paintings with generative adversarial network. IEEE Trans. Image Process. 2020, 29, 8706–8720. [Google Scholar] [CrossRef]
  23. Way, D.L.; Lo, C.H.; Wei, Y.H.; Shih, Z.C. TwinGAN: Twin generative adversarial network for Chinese landscape painting style transfer. IEEE Access 2023, 11, 60844–60852. [Google Scholar] [CrossRef]
  24. Abedi, M.; Hempel, L.; Sadeghi, S.; Kirsten, T. GAN-based approaches for generating structured data in the medical domain. Appl. Sci. 2022, 12, 7075. [Google Scholar] [CrossRef]
  25. Yan, S.; Wang, C.; Chen, W.; Lyu, J. Swin transformer-based GAN for multi-modal medical image translation. Front. Oncol. 2022, 12, 942511. [Google Scholar] [CrossRef]
  26. Sun, L.; Chen, J.; Xu, Y.; Gong, M.; Yu, K.; Batmanghelich, K. Hierarchical amortized GAN for 3D high resolution medical image synthesis. IEEE J. Biomed. Health Inform. 2022, 26, 3966–3975. [Google Scholar] [CrossRef]
  27. Aljohani, A.; Alharbe, N. Generating synthetic images for healthcare with novel deep pix2pix gan. Electronics 2022, 11, 3470. [Google Scholar] [CrossRef]
  28. Alrashedy, H.H.N.; Almansour, A.F.; Ibrahim, D.M.; Hammoudeh, M.A.A. BrainGAN: Brain MRI image generation and classification framework using GAN architectures and CNN models. Sensors 2022, 22, 4297. [Google Scholar] [CrossRef]
  29. Jiaxi, H.; Guixiong, L. An Improved Generative Adversarial Network-Based and U-Shaped Transformer Method for Glass Curtain Crack Deblurring Using UAVs. Sensors 2024, 24, 7713. [Google Scholar] [CrossRef]
  30. Jiangyan, W.; Guanghui, Z.; Yugang, F. LM-CycleGAN: Improving Underwater Image Quality Through Learned Perceptual Image Patch Similarity and Multi-Scale Adaptive Fusion Attention. Sensors 2024, 24, 7425. [Google Scholar] [CrossRef]
  31. Shaojie, G.; Xiaogang, W.; Jiayi, Z.; Zewei, L. A Fast Specular Highlight Removal Method for Smooth Liquor Bottle Surface Combined with U2-Net and LaMa Model. Sensors 2024, 22, 9834. [Google Scholar]
  32. Bingnan, Y.; Zhaozhao, Y.; Huizhu, S.; Conghui, W. ADE-CycleGAN: A Detail Enhanced Image Dehazing CycleGAN Network. Sensors 2023, 23, 3294. [Google Scholar] [CrossRef]
  33. Zhao, B.; Othmane, A.; Jonguk, L.; Daihee, P.; Yongwha, C. GAN-Based Video Denoising with Attention Mechanism for Field-Applicable Pig Detection System. Sensors 2022, 22, 3917. [Google Scholar] [CrossRef] [PubMed]
  34. Zhu, S.; Urtasun, R.; Fidler, S.; Lin, D.; Change Loy, C. Be your own prada: Fashion synthesis with structural coherence. In Proceedings of the IEEE International Conference On Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 1680–1688. [Google Scholar]
  35. Ghasemi, P.; Yuan, C.; Marion, T.; Moghaddam, M. DCG-GAN: Design concept generation with generative adversarial networks. Des. Sci. 2024, 10, e14. [Google Scholar] [CrossRef]
  36. Cao, S.; Chai, W.; Hao, S.; Wang, G. Image reference-guided fashion design with structure-aware transfer by diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 3524–3528. [Google Scholar]
  37. De Souza, V.L.T.; Marques, B.A.D.; Batagelo, H.C.; Gois, J.P. A review on generative adversarial networks for image generation. Comput. Graph. 2023, 114, 13–25. [Google Scholar] [CrossRef]
  38. Croitoru, F.A.; Hondru, V.; Ionescu, R.T.; Shah, M. Diffusion models in vision: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 10850–10869. [Google Scholar] [CrossRef]
  39. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Bengio, Y. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems 27 (NIPS 2014), Montréal, QC, Canada, 8–13 December 2014; pp. 2672–2680. [Google Scholar]
  40. Karras, T.; Laine, S.; Aila, T. A Style-Based Generator Architecture for Generative Adversarial Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 4217–4228. [Google Scholar] [CrossRef]
  41. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2223–2232. [Google Scholar]
  42. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  43. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 214–223. [Google Scholar]
  44. Mao, X.; Li, Q.; Xie, H.; Lau, R.Y.; Wang, Z.; Paul, S.S. Least squares generative adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2794–2802. [Google Scholar]
  45. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar]
  46. Song, J.; Meng, C.; Ermon, S. Denoising Diffusion Implicit Models. In Proceedings of the International Conference on Learning Representations, Online, 3–7 May 2021. [Google Scholar]
  47. Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; Ommer, B. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 10684–10695. [Google Scholar]
  48. Xu, Y.; Zhao, Y.; Xiao, Z.; Hou, T. Ufogen: You forward once large scale text-to-image generation via diffusion gans. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024; pp. 8196–8206. [Google Scholar]
  49. Sauer, A.; Schwarz, K.; Geiger, A. Stylegan-xl: Scaling stylegan to large diverse datasets. In Proceedings of the ACM SIGGRAPH 2022 Conference Proceedings, Vancouver, BC, Canada, 8–11 August 2022; pp. 1–10. [Google Scholar]
  50. Kang, M.; Zhu, J.Y.; Zhang, R.; Park, J.; Shechtman, E.; Paris, S.; Park, T. Scaling up gans for text-to-image synthesis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 10124–10134. [Google Scholar]
  51. Ramesh, A.; Pavlov, M.; Goh, G.; Gray, S.; Voss, C.; Radford, A.; Chen, M.; Sutskever, I. Zero-shot text-to-image generation. In Proceedings of the 38th International Conference on Machine Learning, Online, 18–24 July 2021; pp. 8821–8831. [Google Scholar]
  52. Esser, P.; Kulal, S.; Blattmann, A.; Entezari, R.; Müller, J.; Saini, H.; Rombach, R. Scaling rectified flow transformers for high-resolution image synthesis. arXiv 2024, arXiv:2403.03206. [Google Scholar]
  53. Cui, Y.R.; Liu, Q.; Gao, C.Y.; Su, Z. FashionGAN: Display your fashion design using conditional generative adversarial nets. Comput. Graph. Forum 2018, 37, 109–119. [Google Scholar] [CrossRef]
  54. Lee, J.; Lee, M.; Kim, Y. MT-VTON: Multilevel transformation-based virtual try-on for enhancing realism of clothing. Appl. Sci. 2023, 13, 11724. [Google Scholar] [CrossRef]
  55. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sutskever, I. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, Online, 18–24 July 2021; pp. 8748–8763. [Google Scholar]
  56. Gu, X.; Huang, J.; Wong, Y.; Yu, J.; Fan, J.; Peng, P.; Kankanhalli, M.S. PAINT: Photo-realistic fashion design synthesis. ACM Trans. Multimed. Comput. Commun. Appl. 2023, 20, 1–23. [Google Scholar] [CrossRef]
  57. Liang, X.; Mo, H.; Gao, C. Controllable Garment Image Synthesis Integrated with Frequency Domain Features. Comput. Graph. Forum 2023, 42, e14938. [Google Scholar] [CrossRef]
  58. Chu, T.; Chen, J.; Sun, J.; Lian, S.; Wang, Z.; Zuo, Z.; Lu, D. Rethinking fast fourier convolution in image inpainting. In Proceedings of the IEEE/CVF International Conference on Computer Vision (CVPR), Vancouver Convention Center, Vancouver, BC, Canada, 18–22 June 2023; pp. 23195–23205. [Google Scholar]
  59. Zhou, H.; Dong, W.; Liu, Y.; Chen, J. Breaking through the haze: An advanced non-homogeneous dehazing method based on fast fourier convolution and convnext. In Proceedings of the IEEE/CVF International Conference on Computer Vision (CVPR), Vancouver Convention Center, Vancouver, BC, Canada, 18–22 June 2023; pp. 1895–1904. [Google Scholar]
  60. Quattrini, F.; Pippi, V.; Cascianelli, S.; Cucchiara, R. Volumetric fast fourier convolution for detecting ink on the carbonized herculaneum papyri. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2006 October 2023; pp. 1726–1734. [Google Scholar]
  61. Dong, Q.; Cao, C.; Fu, Y. Incremental transformer structure enhanced image inpainting with masking positional encoding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 11358–11368. [Google Scholar]
  62. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  63. Roman, S.; Logacheva, E.; Anton, M.; Anastasia, R.; Arsenii, A.; Aleksei, S.; Naejin, K.; Harshith, G.; Kiwoong, P.; Victor, L. Resolution-Robust Large Mask Inpainting With Fourier Convolutions. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 2149–2159. [Google Scholar]
  64. Nussbaumer, H.J. The Fast Fourier Transform. In Fast Fourier Transform and Convolution Algorithms; Nussbaumer, H.J., Ed.; Springer Series in Information Sciences; Springer: Berlin/Heidelberg, Germany, 1982; Volume 2. [Google Scholar]
  65. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  66. Yu, F.; Koltun, V. Multi-Scale Context Aggregation by Dilated Convolutions. arXiv 2016, arXiv:1511.07122. [Google Scholar]
  67. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  68. Simonyan, K. Very deep convolutional networks for large-scale image recognition. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  69. Gatys, L.A.; Ecker, A.S.; Bethge, M. A neural algorithm of artistic style. arXiv 2015, arXiv:1508.06576. [Google Scholar] [CrossRef]
  70. Lei, H.; Chen, S.; Wang, M.; He, X.; Jia, W.; Li, S. A New Algorithm for Sketch-Based Fashion Image Retrieval Based on Cross-Domain Transformation. Wirel. Commun. Mob. Comput. 2021, 5577735. [Google Scholar] [CrossRef]
  71. Choi, S.; Park, S.; Lee, M.; Choo, J. Viton-hd: High-resolution virtual try-on via misalignment-aware normalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Online, 19–25 June 2021; pp. 14131–14140. [Google Scholar]
  72. Rostamzadeh, N.; Hosseini, S.; Boquet, T.; Stokowiec, W.; Zhang, Y.; Jauvin, C.; Pal, C. Fashion-gen: The generative fashion dataset and challenge. arXiv 2018, arXiv:1806.08317. [Google Scholar]
  73. Soria, X.; Sappa, A.; Humanante, P.; Akbarinia, A. Dense extreme inception network for edge detection. Pattern Recognit. 2023, 139, 109461. [Google Scholar] [CrossRef]
  74. Yilun, C.; Zhicheng, W.; Yuxiang, P.; Zhiqiang, Z.; Gang, Y.; Jian, S. Cascaded pyramid network for multi-person pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7103–7112. [Google Scholar]
  75. Heusel, M.; Ramsauer, H.; Unterthiner, T.; Nessler, B.; Hochreiter, S. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6629–6640. [Google Scholar]
  76. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  77. Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 586–595. [Google Scholar]
  78. Chaofeng, C.; Jiadi, M.; Jingwen, H.; Haoning, W.; Liang, L.; Wenxiu, S.; Qiong, Y.; Weisi, L. TOPIQ: A Top-Down Approach From Semantics to Distortions for Image Quality Assessment. IEEE Trans. Image Process. 2024, 33, 2404–2418. [Google Scholar]
  79. Cao, C.; Fu, Y. Learning a sketch tensor space for image inpainting of man-made scenes. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), On-line form, 11–17 October 2021; pp. 14509–14518. [Google Scholar]
  80. AlBahar, B.; Huang, J.B. Guided image-to-image translation with bi-directional feature transformation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9016–9025. [Google Scholar]
  81. Nazeri, K.; Ng, E.; Joseph, T.; Qureshi, F.; Ebrahimi, M. Edgeconnect: Structure guided image inpainting using edge prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
Figure 1. Showcase of garment images generated using our method. The real images are marked with red solid line frames, while the generated images are marked with black solid line frames.
Figure 1. Showcase of garment images generated using our method. The real images are marked with red solid line frames, while the generated images are marked with black solid line frames.
Applsci 15 03534 g001
Figure 2. The overall framework of FCAN. The FCAN workflow consists of three stages: the encoding stage, content inference stage, and context aggregation and decoding stage. (a) Encoding Stage: In the encoding stage, FCAN utilizes an encoder to encode the input information into multi-scale features. (b) Content Inference Stage: In the content inference stage, FCAN employs concatenated FF-CDRB blocks to perform content inference and initial context modeling on the deepest features obtained in the encoding stage. (c) Context Aggregation and Decoding Stage: In this stage, the proposed ACAM is used to achieve global context aggregation with heterogeneous features from different stages. The enhanced features are then fused into the corresponding decoding steps, ultimately generating the clothing image.
Figure 2. The overall framework of FCAN. The FCAN workflow consists of three stages: the encoding stage, content inference stage, and context aggregation and decoding stage. (a) Encoding Stage: In the encoding stage, FCAN utilizes an encoder to encode the input information into multi-scale features. (b) Content Inference Stage: In the content inference stage, FCAN employs concatenated FF-CDRB blocks to perform content inference and initial context modeling on the deepest features obtained in the encoding stage. (c) Context Aggregation and Decoding Stage: In this stage, the proposed ACAM is used to achieve global context aggregation with heterogeneous features from different stages. The enhanced features are then fused into the corresponding decoding steps, ultimately generating the clothing image.
Applsci 15 03534 g002
Figure 3. Structure of the encoding block.
Figure 3. Structure of the encoding block.
Applsci 15 03534 g003
Figure 4. The positional information embedding process.
Figure 4. The positional information embedding process.
Applsci 15 03534 g004
Figure 5. Structure of FFC.
Figure 5. Structure of FFC.
Applsci 15 03534 g005
Figure 6. The visualization of multi-scale features from the encoding stage.
Figure 6. The visualization of multi-scale features from the encoding stage.
Applsci 15 03534 g006
Figure 7. The context attention visualization maps constructed by ACAM.
Figure 7. The context attention visualization maps constructed by ACAM.
Applsci 15 03534 g007
Figure 8. The visualization maps of context-enhanced features constructed by ACAM.
Figure 8. The visualization maps of context-enhanced features constructed by ACAM.
Applsci 15 03534 g008
Figure 9. Partial data demonstration of the SKFashion dataset.
Figure 9. Partial data demonstration of the SKFashion dataset.
Applsci 15 03534 g009
Figure 10. Partial data demonstration of the VITON-HD dataset.
Figure 10. Partial data demonstration of the VITON-HD dataset.
Applsci 15 03534 g010
Figure 11. Partial data demonstration of the Fashion-Gen dataset.
Figure 11. Partial data demonstration of the Fashion-Gen dataset.
Applsci 15 03534 g011
Figure 12. Qualitative comparison results under small texture patch sizes on the SKFashion dataset.
Figure 12. Qualitative comparison results under small texture patch sizes on the SKFashion dataset.
Applsci 15 03534 g012
Figure 13. Qualitative comparison results under medium texture patch sizes on the SKFashion dataset.
Figure 13. Qualitative comparison results under medium texture patch sizes on the SKFashion dataset.
Applsci 15 03534 g013
Figure 14. Qualitative comparison results under standard texture patch sizes on the SKFashion dataset.
Figure 14. Qualitative comparison results under standard texture patch sizes on the SKFashion dataset.
Applsci 15 03534 g014
Figure 15. Time evaluation and comparison.
Figure 15. Time evaluation and comparison.
Applsci 15 03534 g015
Table 1. Quantitative experimental results on the SKFashion dataset.
Table 1. Quantitative experimental results on the SKFashion dataset.
Texture Patch SizeMethodFID↓PSNR↑LPIPS↓SSIM↑TOPIQ-FR↑TOPIQ-NR↑
StandardPAINT [56]9.0322.460.104100.8030.5490.610
MST [79]5.9324.210.064380.8330.6930.646
Albahar-MT [80]9.3824.220.107330.8390.5210.545
EdgeConnect [81]5.5524.670.066320.8420.6940.650
ZITS [61]5.3024.780.064260.8540.6890.625
FCAN (Our)4.0724.820.055560.8480.7320.646
MediumPAINT [56]10.2522.230.101310.7970.5140.599
MST [79]6.5723.700.070120.8210.6150.645
Albahar-MT [80]12.3123.150.123410.8210.5040.519
EdgeConnect [81]6.4424.220.074340.8310.6680.640
ZITS [61]6.1124.240.071950.8400.6700.618
FCAN (Our)4.7424.360.061820.8330.7140.642
SmallPAINT [56]14.3321.970.152310.7880.4420.545
MST [79]8.8022.170.089730.7980.5970.628
Albahar-MT [80]14.6522.520.149500.8110.4310.534
EdgeConnect [81]7.4723.590.088420.8140.6010.647
ZITS [61]6.8923.730.079220.8300.6510.614
FCAN (Our)5.5323.950.069060.8210.6930.638
In the quantitative experiments, ↑ and ↓ indicate that higher and lower values of the corresponding metric are better, respectively. The best metric results are highlighted in bold black, while the second-best results are underlined. The same notation is used throughout.
Table 2. Quantitative experimental results on the VITON-HD dataset.
Table 2. Quantitative experimental results on the VITON-HD dataset.
Texture Patch Size FID↓PSNR↑LPIPS↓SSIM↑TOPIQ-FR↑TOPIQ-NR↑
StandardPAINT [56]16.9325.150.081130.8640.6230.667
MST [79]10.9928.240.045700.8930.7490.671
Albahar-MT [80]15.9926.480.076170.8700.5700.635
EdgeConnect [81]11.5128.030.047950.8920.7390.672
ZITS [61]10.3627.580.047030.8990.9190.671
FCAN (Our)9.3828.640.038860.9070.9170.681
MediumPAINT [56]20.1923.810.090900.8300.5870.665
MST [79]14.1726.010.064120.8440.7210.663
Albahar-MT [80]18.4824.280.101900.8420.5150.619
EdgeConnect [81]14.4525.830.057320.8700.6780.667
ZITS [61]12.2726.770.054010.8860.9010.664
FCAN (Our)11.1927.940.045490.8930.9140.671
SmallPAINT [56]30.2822.980.109200.8060.5020.643
MST [79]28.0020.550.132100.7820.4590.630
Albahar-MT [80]41.6222.930.136700.8090.4470.595
EdgeConnect [81]38.6217.800.139700.7780.4550.628
ZITS [61]15.4125.760.063610.8650.8970.658
FCAN (Our)14.3226.920.054780.8770.9010.662
Table 3. Quantitative experimental results on the Fashion-Gen dataset.
Table 3. Quantitative experimental results on the Fashion-Gen dataset.
Texture Patch Size FID↓PSNR↑LPIPS↓SSIM↑TOPIQ-FR↑TOPIQ-NR↑
StandardPAINT [56]23.1029.080.045440.9420.6550.493
MST [79]13.6133.350.021180.9620.7890.529
Albahar-MT [80]22.9630.330.021180.9620.6520.492
EdgeConnect [81]13.8433.290.022440.9620.7780.531
ZITS [61]13.8032.910.021430.9630.7840.527
FCAN (Our)12.7333.410.020400.9620.7940.528
MediumPAINT [56]27.7327.500.055390.9270.5940.491
MST [79]16.7832.070.025970.9530.7440.526
Albahar-MT [80]32.7028.440.051200.9310.5730.484
EdgeConnect [81]18.5030.770.030730.9480.6970.524
ZITS [61]17.9331.810.026390.9540.7430.524
FCAN (Our)16.5332.050.026300.9540.7420.525
SmallPAINT [56]37.2824.980.073660.9070.5260.466
MST [79]27.6327.050.070610.9230.5350.478
Albahar-MT [80]45.3726.600.062430.9160.5000.485
EdgeConnect [81]35.3826.400.051330.9240.5510.503
ZITS [61]21.9230.710.033100.9400.7060.523
FCAN (Our)20.6830.750.033020.9430.7100.528
Table 4. Comparison of the model parameters.
Table 4. Comparison of the model parameters.
MethodsParameters
PAINT [56]83.56 M ⇑ 202%
MST [79]54.93 M ⇑ 98.5%
Albahar-MT [80]57.29 M ⇑ 107.0%
EdgeConnect [81]53.59 M ⇑ 93.7%
ZITS [61]56.59 M ⇑ 104.5%
FCAN (Our)27.67 M
In the table, a red upward arrow is used to indicate the percentage increase in the number of parameters of the corresponding methods compared to our approach.
Table 5. Ablation experiment results on the SKFashion dataset.
Table 5. Ablation experiment results on the SKFashion dataset.
Texture Patch Size FID↓PSNR↑LPIPS↓SSIM↑TOPIQ-FR↑TOPIQ-NR↑
Standardw/o Dual residual4.5324.790.056390.8520.7260.642
w/o ACAM5.2225.030.067390.8440.6980.616
w/o Gating4.3724.640.058210.8380.7230.642
ResNet+ACAM4.6424.480.054400.8260.7120.633
Skip Connection4.8924.960.063570.8530.7220.648
FCAN (Our)4.0724.820.055560.8480.7320.646
Mediumw/o Dual residual5.1624.280.062420.8270.7010.632
w/o ACAM6.0424.220.074460.8280.6710.612
w/o Gating5.3123.840.064120.8180.7110.636
ResNet+ACAM5.3524.080.066720.8170.6940.632
Skip Connection5.7124.550.072460.8400.7040.642
FCAN (Our)4.7424.360.061820.8330.7140.642
Smallw/o Dual residual5.8523.800.069140.8240.6830.635
w/o ACAM7.7722.530.089090.8210.6530.618
w/o Gating6.4123.350.073710.8110.6830.635
ResNet+ACAM6.1423.690.072380.8140.6750.640
Skip Connection7.1523.520.086720.8270.6730.628
FCAN (Our)5.5323.950.069060.8210.6930.638
Table 6. Ablation experiment results on the VITON-HD dataset.
Table 6. Ablation experiment results on the VITON-HD dataset.
Texture Patch Size FID↓PSNR↑LPIPS↓SSIM↑TOPIQ-FR↑TOPIQ-NR↑
Standardw/o Dual residual10.0728.660.045310.8980.9100.680
w/o ACAM11.4827.840.049170.8950.8890.650
w/o Gating10.2128.730.042380.9020.9070.671
ResNet+ACAM10.2728.650.043110.9010.9140.684
Skip Connection10.4028.740.042010.9050.9110.674
FCAN (Our)9.3828.640.038860.9070.9170.681
Mediumw/o Dual residual11.7127.810.047370.8920.9000.663
w/o ACAM14.1526.740.058200.8790.8770.644
w/o Gating11.8627.840.048360.8910.9020.659
ResNet+ACAM12.1127.870.048210.8840.9100.672
Skip Connection14.0926.780.055210.8860.8870.652
FCAN (Our)11.1927.940.045490.8930.9140.671
Smallw/o Dual residual14.8227.010.056790.8640.8910.651
w/o ACAM28.9722.860.102230.8300.8340.613
w/o Gating15.3627.160.058410.8760.8890.649
ResNet+ACAM14.7627.580.054310.8830.9040.659
Skip Connection24.9223.260.093330.8480.8400.602
FCAN (Our)14.3226.920.054780.8770.9010.662
Table 7. Ablation experiment results on the Fashion-Gen dataset.
Table 7. Ablation experiment results on the Fashion-Gen dataset.
Texture Patch Size FID↓PSNR↑LPIPS↓SSIM↑TOPIQ-FR↑TOPIQ-NR↑
Standardw/o Dual residual13.3733.450.022050.9610.7840.521
w/o ACAM13.7533.490.022910.9600.7580.494
w/o Gating13.2733.050.021540.9600.7900.526
ResNet+ACAM13.1433.180.021760.9600.7790.533
Skip Connection14.1633.830.022680.9640.7540.501
FCAN (Our)12.7333.410.020400.9620.7940.528
Mediumw/o Dual residual17.0032.890.026460.9510.7370.524
w/o ACAM17.9431.980.028490.9520.7120.501
w/o Gating17.2132.400.027460.9520.7320.522
ResNet+ACAM17.4231.340.029940.9430.7320.523
Skip Connection19.0032.020.029470.9530.7220.514
FCAN (Our)16.5332.050.026300.9540.7420.525
Smallw/o Dual residual21.1230.800.033620.9420.7060.524
w/o ACAM22.7529.420.039600.9320.6940.516
w/o Gating21.4830.500.034620.9320.7010.521
ResNet+ACAM21.0430.270.035800.9300.7110.520
Skip Connection24.3629.890.038290.9410.7000.517
FCAN (Our)20.6830.750.033020.9430.7100.528
Table 8. Comparison of computational parameter counts after ablation.
Table 8. Comparison of computational parameter counts after ablation.
MethodsParameters
w/o Dual residual23.41 M ⇓ 15.4%
w/o ACAM23.60 M ⇓ 14.7%
w/o Gating22.83 M ⇓ 17.5%
ResNet+ACAM35.78 ⇑ 29.3%
Skip Connection23.60 M ⇓ 14.7%
FCAN27.67 M
In the table, the red upward arrows indicate the percentage increase in parameter count relative to our method, while the green downward arrows indicate the percentage decrease in parameter count relative to our method.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lei, H.; Hu, Y.; Wang, M.; Ding, M.; Li, Z.; Luo, G. Fast Fourier Asymmetric Context Aggregation Network: A Controllable Photo-Realistic Clothing Image Synthesis Method Using Asymmetric Context Aggregation Mechanism. Appl. Sci. 2025, 15, 3534. https://doi.org/10.3390/app15073534

AMA Style

Lei H, Hu Y, Wang M, Ding M, Li Z, Luo G. Fast Fourier Asymmetric Context Aggregation Network: A Controllable Photo-Realistic Clothing Image Synthesis Method Using Asymmetric Context Aggregation Mechanism. Applied Sciences. 2025; 15(7):3534. https://doi.org/10.3390/app15073534

Chicago/Turabian Style

Lei, Haopeng, Ying Hu, Mingwen Wang, Meihai Ding, Zhen Li, and Guoliang Luo. 2025. "Fast Fourier Asymmetric Context Aggregation Network: A Controllable Photo-Realistic Clothing Image Synthesis Method Using Asymmetric Context Aggregation Mechanism" Applied Sciences 15, no. 7: 3534. https://doi.org/10.3390/app15073534

APA Style

Lei, H., Hu, Y., Wang, M., Ding, M., Li, Z., & Luo, G. (2025). Fast Fourier Asymmetric Context Aggregation Network: A Controllable Photo-Realistic Clothing Image Synthesis Method Using Asymmetric Context Aggregation Mechanism. Applied Sciences, 15(7), 3534. https://doi.org/10.3390/app15073534

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop