Next Article in Journal
Soil Water Status Monitoring System with Proximal Low-Cost Sensors and LoRa Technology for Smart Water Irrigation in Woody Crops
Next Article in Special Issue
Probabilistic Attention Map: A Probabilistic Attention Mechanism for Convolutional Neural Networks
Previous Article in Journal
Safety After Dark: A Privacy Compliant and Real-Time Edge Computing Intelligent Video Analytics for Safer Public Transportation
Previous Article in Special Issue
Text Font Correction and Alignment Method for Scene Text Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Style Transfer of Chinese Wuhu Iron Paintings Using Hierarchical Visual Transformer

1
Academy of Art and Design, Anhui University of Technology, Ma’anshan 243002, China
2
Engineeing Training Center, Nanjing Vocational University of Industry Technology, Nanjing 210023, China
3
College of Computer Science and Engineering, Chongqing University of Technology, Chongqing 400054, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(24), 8103; https://doi.org/10.3390/s24248103
Submission received: 6 November 2024 / Revised: 10 December 2024 / Accepted: 16 December 2024 / Published: 19 December 2024

Abstract

:
Within the domain of traditional art, Chinese Wuhu Iron Painting distinguishes itself through its distinctive craftsmanship, aesthetic expressiveness, and choice of materials, presenting a formidable challenge in the arena of stylistic transformation. This paper introduces an innovative Hierarchical Visual Transformer (HVT) framework aimed at achieving effectiveness and precision in the style transfer of Wuhu Iron Paintings. The study begins with an in-depth analysis of the artistic style of Wuhu Iron Paintings, extracting key stylistic elements that meet technical requirements for style conversion. Furthermore, in response to the unique artistic characteristics of Wuhu Iron Paintings, this research constructs a multi-layered network structure capable of effectively capturing and parsing style and content features. Building on this, we have designed an Efficient Local Attention Decoder (ELA-Decoder) that adaptively decodes the style and content features through correlation, significantly enhancing the length dependency of local and global information. Additionally, this paper proposes a Content Correction Module (CCM) to eliminate redundant features generated during the style transfer process, further optimizing the migration results. In light of the scarcity of existing datasets for Wuhu Iron Paintings, this study also collects and constructs a dedicated dataset for the style transfer of Wuhu Iron Paintings. Our method achieves optimal performance in terms of loss metrics, with a reduction of at least 4% in style loss and 5% in content loss compared to other advanced methods. Moreover, expert evaluations were conducted to validate the effectiveness of our approach, and the results show that our method received the highest number of votes, further demonstrating its superiority.

1. Introduction

Deep learning is widely used in computer vision [1,2,3,4], computer graphics [5,6,7,8], and other fields. When the field of image generation is developing at a high speed, Image Style Transfer is also widely noticed by researchers. Image Style Transfer aims to transfer artistic features from style images to content images, thus bringing interesting visual perception effects for the delivery of artworks. Also, Gatys et al. [9] first proposed that correlation between features can be used as a basis for accurate style transfer and texture synthesis, thus developing convolutional neural networks (CNNs) for style transfers.
Along with the research, Image Style Transfer methods have evolved from the traditional [10] to the iterative paradigm [11] and feed-forward methods [12,13]. In addition, to address the problem of style bias in most images, the Universal Image Style Transfer (UST) has also gained significant development [14,15,16]. Based on this, innovations in various style Transfer methods have also been stimulated, such as the emergence of innovative approaches such as the Flow-based ArtFlow [17] and comparative learning-based approaches [18,19].
Despite the great progress in related research, most of the current approaches focus on the balanced conversion of multiple styles, or are specific to Western paintings, as all abstract paintings tend to be similar in style. However, unlike the scientific perspective in the West, Chinese Wuhu Iron Painting, as one of the traditional Chinese arts, has a certain concept of pictorial perspective, which is an idealistic feeling that permits one to depict what is beyond what is visible to the naked eye. In particular, the artistic three-dimensionality, multi-layered interspersion, texture expression, and presentation techniques of Chinese Wuhu Iron Paintings (as shown in Figure 1) further increase the difficulty of symmetry of features in style transfer, and the current state-of-the-art methods do not lead to satisfactory visual effects. Specifically, current methods do not pay good attention to global and local relationships, resulting in redundant information that damages the view when decoding features. Therefore, due to the consideration of computational resources and the current excellent performance of Hierarchical Vision Transformer in the field of image style migration, we further explored the two perspectives of decoder and removing redundant features.
In this paper, based on the above challenges, we propose a Hierarchical Visual Transformer based on the Wuhu Iron Painting style transfer network, which facilitates features to be symmetrized during style transfer, referred to in [16]. Specifically, inspired by [16], we rely on the hierarchical visual transformer’s multi-level windowed attention to efficiently capture content and style images’ local and global features. Subsequently, we designed the Efficient Local Attention Decoder (ELA-Decoder) to correlatively and adaptively decode stylistic features and content features through effective local attention to enhance the long and short-term dependencies between local and global information. Meanwhile, to further ensure the fidelity and friendliness of the visual perception of the transfer results, we design a content correction module (CCM) using a residual dense architecture to achieve the elimination of redundant features from the transfer results to obtain visually oriented transformation effects. In addition, due to the lack of an existing benchmark dataset of Chinese Wuhu Iron Paintings and also to further demonstrate the validity of the designed network, we first adopted the field survey method to collect a large number of original pictures of Wuhu Iron Paintings, and constructed the dataset of Wuhu Iron Paintings by classifying and processing the pictures; moreover, we combined the stylistic conversion practice of many types of pictures and the expert scoring method to evaluate the results qualitatively and quantitatively, to test and validate the effectiveness of the proposed method.
Our main contributions are summarized below:
  • We propose a new network that achieves reliable feature encoding by relying on short and long-term modeling of content features and stylistic features with a hierarchical visual transformer and effective style transfer of Chinese Wuhu Iron Paintings with a designed attentional decoder.
  • We further designed a content correction module that effectively captures redundant features and noise for rejection using a residual dense architecture to ensure the visual fidelity and friendliness of the migrated images.
  • We collected a dataset of Iron Paintings from Wuhu, China, and evaluated it qualitatively and quantitatively to verify the validity of our method.

2. Related Work

2.1. Chinese Wuhu Iron Paintings

2.1.1. Analysis of the Artistic Characteristics of Chinese Wuhu Iron Paintings

As a treasure of traditional Chinese arts and crafts, Wuhu Iron Paintings show their unique charm [20,21,22,23] in many dimensions such as modeling, theme, color, craftsmanship, and culture. In terms of modeling, Wuhu Iron Paintings are influenced by the characteristics of materials [24], taking iron as raw material and shaping iron sheets and wires through forging and other processes to create diversified shapes, which are characterized by smooth lines, strong three-dimensionality, and clear hierarchy. In the choice of themes, Wuhu Iron Paintings are often created with natural scenery, stories of people, and scenes of life, especially landscapes, plums, orchids, bamboo, chrysanthemums, etc. These themes also reflect the love of local people for nature. In terms of color, the traditional Wuhu Iron Paintings are black and white. With the development of the times, the current ones also include gold, green, and other colors, but most of the Iron Paintings, in terms of the choice of color, are still mainly black and white. In terms of technology, Wuhu Iron Paintings take cooked iron as raw material, combine forging as the main production technology, and integrate drilling, lifting, pressing, welding, filing, chiseling, and other technologies, which makes Wuhu Iron Paintings more accurate and delicate in the shaping of things. In terms of cultural inheritance, Wuhu Iron Paintings are a symbol of local culture, urban development, and people’s spirit, and the production of Wuhu Iron Paintings requires constant forging and shaping of iron and steel in order to build a complete picture.

2.1.2. Extraction of Artistic Characteristics of Wuhu Iron Paintings in China

In view of the limitations of style migration, quantifiable attributes should be chosen for the extraction of stylistic features for Chinese Wuhu Iron Paintings. In addition, for Wuhu Iron Paintings, the style migration should not only ‘take shape’, but, more importantly, ‘convey the spirit’, so that users can directly distinguish the specific source of style features. The specific style features are extracted as shown in the Table 1.

2.2. Image Style Transfer

Image Style Transfer serves as an important research direction in the field of computer vision [25,26,27,28,29]. Its goal is to preserve the structure of the content image, while giving it the artistic style of another image. With the advent of the deep learning era, Image Style Transfer methods have evolved; for example, the traditional methods of [10] have transitioned to the iterative paradigm-based [11] and feed-forward network-based [12,13] in development. Gatys et al. [9] first applied convolutional neural networks (CNNs) to style transfer by iteratively optimizing noisy images to achieve stylization. Flow-based ArtFlow [17] with Projected Flow Network (PFN) achieves unbiased content results, while IEST [18] and CAST [19] achieve attractive results by using contrast learning.
Although the current progress has made great advances for generic style and Western oil painting style transfers, based on the inherent technical differences [30,31,32,33,34,35,36] between Chinese painting and Western oil painting, it is difficult to achieve a better balance of generic style for better artistic creation. Therefore, Ref. [37] proposed a style shift towards Chinese painting with unique ink characteristics. However, the existing methods still face great challenges due to the more three-dimensional layering and more complex techniques of Chinese Wuhu Iron Paintings.

2.3. Vision Transformer for Image

Transformer was originally designed for natural language processing tasks [38], but due to its unique architecture and representation capabilities, it was introduced to the field of computer vision [39,40,41,42,43,44,45,46]. Meanwhile, in order to alleviate the high computational cost problem faced by the Vision Transformer, DeiT [47] improves the original ViT by using methods such as strong data augmentation to eliminate the reliance on large amounts of data. Vision Transformer has also shown a strong impact in the field of style transfer [48,49,50,51,52]. In [53], a novel axial attention transform encoder was developed, which significantly improves the operational efficiency in style transfer tasks.
Recently, there has been a surge of interest in transformers for hierarchical architectures (LeViT [49] and CvT [54]). StyA2K, proposed by Zhu et al. [55], shows exceptional performance in preserving semantic structure and presenting consistent style patterns. Ref. [16] achieves a powerful generic style transfer by utilizing output merging of different windows’ attention. However, for Wuhu Iron Paintings, a traditional Chinese art, these methods, although they can encode feature information well, are unable to distinguish important and redundant features well for separation and decoding in the transfer process, resulting in unsatisfactory converted images.

3. Methodology

Inspired by [16], We introduced the Hierarchical Vision Transformer, and effectively improved its decoder and detail processing. In this paper, our goal is to explore a style transfer network for Chinese Wuhu Iron Paintings that leads to perception-oriented transfer. As shown in Figure 2, we first decompose the style image and the content image based on multi-level Window Attention, and then encode the feature using a transformer encoder. Subsequently, a transformer decoder is utilized to decode the features initially, and the Efficient Local Attention Decoder is designed to focus on the style features and important content features to achieve effective decoding of the abstract style. In addition, the decoded features are fed in parallel to the content correction module to reject redundant features further and ultimately achieve fidelity transfer results. In the following sections, we will gradually elaborate on the proposed method in detail.

3.1. Hierarchical Visual Transformer

We first divide the content and style images into patch blocks of size 2 × 2 non-overlapping each other using the patch partitioning module. Subsequently, we deliver them to an encoder consisting of Strips Window Attention [16], and fill operations via linear embedding. Specifically, the fill operation is applied to achieve integer division filling for multiple windows. As shown in Figure 3a, Strips Window Attention consists of three different Window Attentions, where the horizontal and vertical Strip Windows pay attention to capture the longwall information and relevance of the feature, and the square window’s attention is focused on the surrounding information. An effective balance between short-range and long-range dependencies is achieved by integrating information from different windows [16] to expand the target’s sensory field. For Window Attention, we follow the previous work [16,48,56] for the calculation of the relative position deviation B, i.e.,
W M S A M × M ( Q , K , V ) = S o f t m a x ( Q K T d + B ) V ,
where Q, K, and V are the query, key, and value matrices, d is the matrix dimension, and W M S A M × M denotes the polytope self-attention using an M × M window.
Meanwhile, to achieve a multi-scale fusion of features during the encoding process, we extract elements at two-step intervals along the horizontal and vertical axes by merging the patches as a downsampling module. After feature encoding using the Strips Window Attention encoder, we perform the style transfer process through a multilayer Transformer decoder. Specifically, we perform inter-feature attention using M S A at each multi-head attention and use LayerNorm afterward. Finally, we output the migrated features through Feed-Forward Network (FFN) and LayerNorm.

3.2. ELA-Decoder Module

Previous studies have been able to capture features well and perform the initial transfer, but have not achieved good results when decoding migrated features. Particularly in the case of works of art with complex lines, such as Iron Paintings, previous approaches have scaled back style-specific concerns in order to accommodate generic styles. Thus, for the decoding process, effective attention to the complex features is lacking, thus prompting us to introduce effective localized attention to mitigate this challenge. As shown in Figure 2d, we utilize an effective combination of VGG and Efficient Local Attention to focus on key regions of decoded features and select abstract style and content features with adaptive relevance for more efficient style feature transformation.
We embed the Efficient Local Attention Block into the VGG network to extract the features of the key parts more efficiently, and to generate style images with more details. Specifically, we feed the output of the feature by the transformer decoder into ReflectionPad2d Block, and then we feed the output into the Efficient Local Attention (ELA) Module. The ELA module is shown in Figure 3b. First, the input features are pooled using strip-pooling to capture long-distance dependencies in the spatial dimension, followed by a convolution operation to process the sequence signal, which is more adept at processing sequence signals and lighter compared to 2D convolution. The operation is defined as follows:
y h = ϱ ( G n ( F h ( Z h ) ) ) , y w = ϱ ( G n ( F w ( Z w ) ) ) ,
where F h and F w denote one-dimensional convolutions and ϱ denotes a nonlinear activation function.
We multiply the obtained vectors in the horizontal and vertical directions with the input vectors to obtain the output, which is defined as follows:
Y = x × y h × y w ,
After obtaining the output, we perform the Conv and ReLU operations, and then upsample the features with the same resolution size as the input image. To make the details of the image fit the style image better, we loop into that network structure again, but without the upsampling operation.

3.3. Content Correction Module

We implement a Content Correction Module, as shown in Figure 4, to further process redundant features using a residual dense architecture combined with depth separable convolution. Specifically, we first extract shallow information through a layer of depth-separable convolution [57,58]. Subsequently, the combination of Residual Dense Block (RDB) [59] with four layers of residual dense architecture and ReLU is used to refine the features further and fuse all the information to enhance the memory of shallow features to deeper ones. Finally, we reconstruct noisy mappings and redundant features by combining deep separable convolution and ReLU. By removing the learned noise mappings, we further achieve effective style transfer. The operation is defined as follows:
I c s = A D ( f c s ) A D ( C C M ( f c s ) ) ,
where A D is the attention decoder, f c s is the output feature, and C C M denotes the content correction operation.

3.4. Network Training

We measure the content difference between the stylized image I c s and the content image I c by content loss, the content loss is the Euclidean distance between the mean-variance channel-wise normalized target features, and the mean-variance channel-wise normalized features of the output image VGG19 features, defined as
L c o n t e n t = 5 l = 4 Φ i m a g e l ( I c s ) Φ i m a g e l ( I c ) 2 ,
where Φ i m a g e denotes that the image encoder extracts features from the pre-trained VGG19 model, and l is the set of features consisting of the fourth and fifth layers in VGG19.
L c o n t e n t = 5 l = 4 β 1 Φ i m a g e l ( I c s ) Φ i m a g e l ( I s ) 2 ,
where β l is the weight of the feature loss in layer l of the VGG19 model. To consider both the global statistics and the semantically local mapping between the content features and the style features, we also use identity loss [60] to further preserve the structure of the content image and the stylistic features of the style image. The two different identity losses are defined as
L i d 1 = I c c I c 2 + I s s I s 2 ,
L i d 2 = l N Φ l ( I c c ) Φ l ( I c ) 2 + Φ l ( I s s ) Φ l ( I s ) 2 ,
where I c c (or I s s ) denotes the output image stylized from two images of the same content (or style). Therefore, the total loss function is defined as
L t o t a l = λ c L c o n t e n t + λ s L s t y l e + λ i d 1 L i d 1 + λ i d 2 L i d 2 ,
where λ c , λ s , λ i d 1 , and λ i d 2 are the weights for different losses. We set the weights to 2, 3, 50, and 1 to mitigate the effect of magnitude differences. The variation in content loss and style loss with the number of iterations during training is shown in Figure 5.

4. Experiment

4.1. Implementation Details

We used PyTorch (https://pytorch.org/) to implement our framework on two NVIDIA RTX 3090 GPUs. Under limited experimental conditions, the network uses the Adam optimizer, and through experimental comparison and evaluation, it is found that the initial learning rate is set to 1 × 10 4 and the batch size is set to 8. In addition, we first use MS-COCO [61] and WikiArt [62] as the content dataset and the style dataset, respectively, for initial training. Then, we used the Wuhu Iron Paintings dataset for fine-tuning and validation.

4.2. Comparison Experiment

4.2.1. Qualitative Comparison

In this section, we compare the effects between the StWip and previous SOTA methods Ghiasi et al. [63], CAST [19], StyTr2 [48], and S2WAT [16]. Figure 6 presents the visualization outcomes of landscape and flora and fauna on the Chinese Wuhu Iron Paintings dataset, showing improved results due to similarities with the dataset’s characteristics. These improvements are particularly evident in line smoothness, clarity, metallic texture, outline definition, black-and-white contrast, and three-dimensional artistic expression. However, the current method fails to capture the stylistic nuances of Wuhu Iron Paintings, leading to significant style deviations. CAST maintains content structure, but misses complex style patterns, causing style corruption. StyTr2 and S2WAT, while achieving good color results through hierarchical visual transformers, lack local and global detail, failing to capture Iron Painting line details. In contrast, our method excels in preserving the stylized details of Wuhu Iron Paintings. Figure 7 demonstrates the challenges of applying this style to furniture, especially with complex layered paintings. The SOTA approach results in excessive features and shadows, whereas our content correction module, tailored for Wuhu Iron Paintings’ complexity, significantly reduces redundant features and shadows, showing effective results. As shown in Figure 8, we show the partial visualization results.

4.2.2. Quantitative Comparison

In this section, we utilize losses as an indirect indicator. As shown in Table 2, compared to other methods, the content and style loss of our proposed method obtains the lowest value, which verifies the effective migration performance to the Wuhu Iron Painting dataset. Meanwhile, to ensure the validity of the comparative experiment, a quantitative analysis was conducted in the form of questionnaire voting. A total of 200 subjects were invited to this experiment, of which a total of 178 valid votes and 22 invalid votes were cast, with an effective voting rate of 89%. The specific voting results are shown in Table 3.
Combined with the results in the above table, it can be seen that the style conversion method designed in the article obtains the highest number of votes in the voting for different types of effect images, which shows that it is the most effective and most recognized by the users in converting the style of Wuhu Iron Paintings. In the voting of animal, landscape, and product categories, the method designed in the article obtained high votes, especially in the voting of sample image 1 of the houseware category; the user votes reached 79, which shows its wide applicability and high efficiency in dealing with various types of images. This result also reflects the technical superiority of the designed method, which is able to retain the original features of the image better and successfully convert it to the Wuhu Iron Paintings style.

4.3. Ablation Study

4.3.1. ELA-Decoder Module

In order to demonstrate the effectiveness and superiority of the ELA-Decoder Module, we conducted experiments to test it using the decoder designed with [16]. The results are shown in Figure 9. Due to the lack of attention to the local features and the dependence on the overall features, the conversion result fails to effectively capture the hierarchical and line features of the Chinese Wuhu Iron Paintings. This effectively validates the superiority of the ELA decoding module, which provides excellent decoding of both local and overall features.

4.3.2. Content Correction Module

The effectiveness of the content correction module designed using the dense residual architecture. As shown in Figure 7, the effectiveness has been initially verified in furniture type presentation. In addition, we further experimentally verify its superiority. As shown in Figure 10, we demonstrate the visual effect without the content correction module, and it can be seen that redundant features are obviously increased in the conversion results without the content correction module compared to the full setup. This further confirms our conjecture.

4.4. Expert Scoring Experiment

In order to verify the effectiveness of the methodology proposed in the article, this study empirically evaluated the method using the expert scoring method, inviting 60 teachers and students from the fields of non-heritage culture, fine arts, and design, as well as non-heritage inheritors, as participants. The core of the experiment is to comprehensively evaluate the effectiveness of the method in style transformation through the subjective experience of the participants, from the presentation of smooth lines, the clever use of black and white contrast, the realistic simulation of metal textures, and other key dimensions. We performed this by making the original image and the converted style image available together for the experts’ evaluation. This not only tests the artistic integration ability of the technical means, but also explores its potential value in the inheritance and development of intangible cultural heritage, providing a strong empirical basis for the modernization and transformation of traditional crafts.
A total of 60 questionnaires were collected in this study; 4 invalid questionnaires were screened and eliminated, and the final valid questionnaires amounted to 56, with a validity rate of 93.33%. In the experimental session, we randomly selected 4 groups of image samples processed by the style conversion technology as test subjects. Each participant, based on their personal subjective experience, rated the style migration effect and aesthetics of each group of samples on six dimensions: smoothness of lines, metal texture performance, simplicity of outline, black and white contrast effect, three-dimensional depth of the image, and the art of leaving white space. The scoring was based on a five-point scale, with 1 being the lowest and 5 being the highest, and the subjects gave the corresponding scores for the style migration effect and aesthetics of each image sample according to their subjective feelings. In this paper, a reliability analysis was conducted, and the results showed that Cronbach’s α coefficient reached 0.931, which fully proved the high reliability of the scoring items and provided solid data support for the subsequent research. The specific experimental results are shown in Table 4.

5. Analysis of Design Application Based on the Conversion of Wuhu Iron Paintings Style

Path 1: Product design applications. The application of Wuhu Iron Paintings style conversion method in product design is not only a cross-border cooperation of art forms, but also a model of deep integration of traditional and modern design. Our methodology refines and analyzes the unique artistic features of Chinese Wuhu Iron Painting, such as its iconic smooth lines, metallic expression, contrasting black and white effects, etc., and through the combination of modern design thinking and high-tech means, the converted effects are cleverly and quickly integrated into modern product design, bringing more possibilities to the original product design. This process involves the cross-fertilization of multiple disciplines such as computer science, design, craft aesthetics, etc. Designers can make use of modern technologies such as 3D modeling, laser cutting, precision casting, etc., to quickly and accurately reproduce the artistic charm of Iron Paintings, and at the same time, give the products a brand-new form and function. The use of this design method not only greatly broadens the creative space of product design, but also provides a strong technical support for the transformation of traditional art product design.
As shown in Figure 11, a modern minimalist furniture product was chosen as the base for applying our style conversion method, successfully integrating Wuhu Iron Painting elements. Rhino software7.0 facilitated 3D modeling based on the converted style, while KeyShot12 was utilized for rendering, accurately simulating the metallic texture of Wuhu Iron Painting by fine-tuning material properties. The visual effect was enhanced with strategic lighting and background settings, particularly emphasizing the black-and-white contrast to add depth and realism to the model. The KeyShot rendering outcome not only highlights the distinctive appeal of the Wuhu Iron Painting style, but also confirms the efficacy of our style conversion approach and the 3D modeling and rendering techniques employed.
Path 2: Environmental design applications. Wuhu Iron Painting has long occupied an important place in the field of decorative arts with its unique forging techniques and aesthetic interests, and the application of its style conversion method provides a rich source of material and inspiration for environmental design. In the current space construction, it can be analyzed from the following four dimensions through the style conversion method.
The Wuhu Iron Painting style’s application in modern design achieves several objectives. It rapidly adapts to interior spaces, transforming elements like partitions and walls, and enhancing the artistic appeal of modern interiors. In landscape and public art, this style offers innovative design solutions, integrating Wuhu Iron Painting’s artistic language into sculptures and water features, thereby enriching the landscape’s visual hierarchy and cultural significance. Additionally, it facilitates spatial division and creates visual focal points, diversifying the artistic expression of space. The transformation method, demonstrated through 3D modeling with Rhino and rendering with Keyshot, maintains functionality while significantly improving the space’s artistic and cultural ambiance, validating its effectiveness in contemporary design. As shown in Figure 12, the original space design sample was transformed into Wuhu Iron Painting style to generate a new style case rich in traditional cultural elements. Subsequently, Rhino software was used for 3D modeling, and Keyshot for rendering, to present the transformed visual effect. By comparing the original and transformed design effects, it is found that the transformation method not only preserves the functionality of the space, but also significantly enhances its artistic expression and cultural atmosphere, which strongly verifies the value and effectiveness of the transformation method of Wuhu Iron Painting style in modern space design.
Path 3: Digital application. The digital application for Wuhu Iron Painting involves creating a virtual exhibition space using VR and style conversion methods. This space lets users instantly change their environment to one that displays the Wuhu Iron Painting style through simple actions like a button press or gesture interaction, providing a global audience with an immersive art experience. Additionally, a digital interactive platform is built that integrates style conversion with digital printing. Users can select a style sample, convert it with one click, and then print the design onto various items like clothing and home decor. This approach enhances the art’s reach, encourages public engagement in its preservation and innovation, and involves more people in supporting this traditional art form. As shown in Figure 13 and Figure 14, in practice, by integrating the style conversion algorithm and digital printing technology, users can quickly generate the design effect of the Wuhu Iron Painting style and print out the physical object for use instantly. This process not only accelerates the transformation from design to object, but also effectively enhances the users’ intuitive perception of the charm of Wuhu Iron Paintings.
In summary, through the method of converting the style of Wuhu Iron Painting, the style of Iron Painting can be migrated to various objects and environmental spaces, injecting new vitality and cultural connotation into modern design. This kind of inheritance and innovation not only helps to protect and promote the art of Iron Painting, but also provides new ideas for the inheritance promotion, and application of other intangible cultural heritage.

6. Discussion

In this section, we will provide a detailed introduction to the Wuhu Iron Paintings dataset, and discuss the limitations of our method.

6.1. Wuhu Iron Paintings Dataset

We have collected 500 Wuhu Iron Paintings from the real world and categorized them into five major classes: plants and animals, landscapes, houseware, and architecture. An example of the classification of the Wuhu Iron Paintings dataset is shown in Figure 15. The plants and animals category features vivid depictions of animals and plants with intricate textures created through techniques like hammering, bending, and welding. The landscape category portrays natural scenery such as mountains, trees, and rivers, using iron texture and lines to create depth. The houseware category includes furniture and decorations, blending practicality with artistry. The architecture category showcases traditional buildings, such as pavilions and bridges, with precise structural depictions and fine details.

6.2. Limitation

Although our method has made significant progress in transferring the style of Wuhu Iron Paintings, there are still some limitations that need to be addressed. As shown in the first row of Figure 16, when the style image and content image are overly similar, the network struggles to effectively decouple content and style features during the transformation process, resulting in suboptimal visual outcomes. Additionally, as depicted in the second row of Figure 16, when using a relatively simple style image for a content image with a complex background, feature coupling may introduce deviations, leading to unnecessary artifacts.

7. Conclusions

This paper presents an efficient style transfer network for Chinese Wuhu Iron Paintings, built on a Hierarchical Vision Transformer framework to achieve image style transformation. The network employs an encoder with a Strips Window Attention mechanism for efficient feature encoding, combined with a specially designed attention decoder and content correction module to harmonize content and style features. This ensures vivid stylistic expression and rich content preservation during the style transfer process. Extensive experimental results demonstrate that our method achieves state-of-the-art performance on the Wuhu Iron Paintings dataset. Compared to existing advanced methods, our approach achieves the lowest content loss and style loss, at 1.62 and 1.63, respectively. In qualitative analyses, expert evaluations further validated our method, with our approach receiving the highest number of votes, underscoring its outstanding performance. Additionally, the paper explores applications of Wuhu Iron Paintings in product, environmental, and digital design, providing a novel approach to the preservation and innovation of intangible cultural heritage while offering new directions for future development in this field.

Author Contributions

Conceptualization, Y.Z. and Y.R.; Software, C.W.; Validation, Y.Z. and C.W.; Formal analysis, Y.R. and M.X.; Data curation, Y.R.; Writing—original draft, Y.Z. and Y.R.; Writing—review & editing, C.W. and M.X.; Funding acquisition, C.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the Scientific Research Foundation for the Introduction of Talent of Nanjing Institute of Industry Technology (Grant No. YK22-14-02).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to copyright issues.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Gao, I.; Ilharco, G.; Lundberg, S.; Ribeiro, M.T. Adaptive testing of computer vision models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 4003–4014. [Google Scholar]
  2. Wang, X.; Wang, W.; Cao, Y.; Shen, C.; Huang, T. Images speak in images: A generalist painter for in-context visual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 6830–6839. [Google Scholar]
  3. Talebi, H.; Milanfar, P. Learning to resize images for computer vision tasks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 497–506. [Google Scholar]
  4. Cheng, W.H.; Song, S.; Chen, C.Y.; Hidayati, S.C.; Liu, J. Fashion meets computer vision: A survey. ACM Comput. Surv. (CSUR) 2021, 54, 1–41. [Google Scholar] [CrossRef]
  5. Huang, B.; Yu, Z.; Chen, A.; Geiger, A.; Gao, S. 2d gaussian splatting for geometrically accurate radiance fields. In Proceedings of the SIGGRAPH ’24: Special Interest Group on Computer Graphics and Interactive Techniques Conference, Denver, CO, USA, 27 July–1 August 2024; ACM SIGGRAPH 2024 Conference Papers. pp. 1–11. [Google Scholar]
  6. Gortler, S.J.; Grzeszczuk, R.; Szeliski, R.; Cohen, M.F. The lumigraph. In Seminal Graphics Papers: Pushing the Boundaries; Association for Computing Machinery: New York, NY, USA, 2023; Volume 2, pp. 453–464. [Google Scholar]
  7. Pan, X.; Tewari, A.; Leimkühler, T.; Liu, L.; Meka, A.; Theobalt, C. Drag your gan: Interactive point-based manipulation on the generative image manifold. In Proceedings of the ACM SIGGRAPH 2023 Conference Proceedings, Los Angeles, CA, USA, 6–10 August 2023; pp. 1–11. [Google Scholar]
  8. Tewel, Y.; Gal, R.; Chechik, G.; Atzmon, Y. Key-locked rank one editing for text-to-image personalization. In Proceedings of the ACM SIGGRAPH 2023 Conference Proceedings, Angeles, CA, USA, 6–10 August 2023; pp. 1–11. [Google Scholar]
  9. Gatys, L.A.; Ecker, A.S.; Bethge, M. Image style transfer using convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2414–2423. [Google Scholar]
  10. Efros, A.; Freeman, W. Image Quilting for Texture Synthesis and Transfer; SIGGRAPH: Tokyo, Japan, 2001. [Google Scholar]
  11. Gatys, L.; Ecker, A.S.; Bethge, M. Texture synthesis using convolutional neural networks. Adv. Neural Inf. Process. Syst. 2015, 28, 262–270. [Google Scholar]
  12. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part II 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 694–711. [Google Scholar]
  13. Chen, D.; Yuan, L.; Liao, J.; Yu, N.; Hua, G. Stylebank: An explicit representation for neural image style transfer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1897–1906. [Google Scholar]
  14. Yin, W.; Yin, H.; Baraka, K.; Kragic, D.; Björkman, M. Dance style transfer with cross-modal transformer. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 5058–5067. [Google Scholar]
  15. Tang, H.; Liu, S.; Lin, T.; Huang, S.; Li, F.; He, D.; Wang, X. Master: Meta style transformer for controllable zero-shot and few-shot artistic style transfer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 18329–18338. [Google Scholar]
  16. Zhang, C.; Xu, X.; Wang, L.; Dai, Z.; Yang, J. S2wat: Image style transfer via hierarchical vision transformer using strips window attention. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 7024–7032. [Google Scholar]
  17. An, J.; Huang, S.; Song, Y.; Dou, D.; Liu, W.; Luo, J. Artflow: Unbiased image style transfer via reversible neural flows. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 862–871. [Google Scholar]
  18. Chen, H.; Wang, Z.; Zhang, H.; Zuo, Z.; Li, A.; Xing, W.; Lu, D. Artistic style transfer with internal-external learning and contrastive learning. Adv. Neural Inf. Process. Syst. 2021, 34, 26561–26573. [Google Scholar]
  19. Zhang, Y.; Tang, F.; Dong, W.; Huang, H.; Ma, C.; Lee, T.Y.; Xu, C. Domain enhanced arbitrary image style transfer via contrastive learning. In Proceedings of the ACM SIGGRAPH 2022 Conference Proceedings, Vancouver, BC, Canada, 7–11 August 2022; pp. 1–8. [Google Scholar]
  20. Ma, L.; Wang, B. Design and Inheritance of Iron Painting Intangible Cultural Heritage Based on Modern Information Technology. In Proceedings of the 2020 International Conference on Data Processing Techniques and Applications for Cyber-Physical Systems: DPTA 2020; Springer: Berlin/Heidelberg, Germany, 2021; pp. 407–414. [Google Scholar]
  21. Ma, L. The Inheritance Strategy of Intangible Cultural Heritage Based on Internet and Information Technolog–Taking Wuhu Iron Painting as an Example. In Proceedings of the 2021 International Conference on Forthcoming Networks and Sustainability in AIoT Era (FoNeS-AIoT), Nicosia, Turkey, 27–28 December 2021; pp. 280–283. [Google Scholar]
  22. Lyu, K. On Intangible Cultural Heritage Research on the Inheritance and Development of “Wuhu Iron Painting” in Wuhu City. MESSAGE from the President of Suan Sunandha Rajabhat University. In Proceedings of the 1st International Conference on Management, Innovation, Economics and Social Sciences, Bangkok, Thailand, 25–26 July 2020; p. 445. [Google Scholar]
  23. Li, G.; Hu, J. The inheritance and development of Wuhu iron paintings from the perspective of cultural industry. In Proceedings of the 3rd International Conference on Public Art and Human Development (ICPAHD 2023), Tianjin, China, 22–24 December 2023; EDP Sciences: Les Ulis Cedex, France, 2024; Volume 183, p. 01017. [Google Scholar]
  24. Tiancheng, Z.; Tieyi, C. The preliminary study on the application of modern advanced processing technique in non-legacy cultural and creative product design–Taking Wuhu iron painting as an example. In Proceedings of the E3S Web of Conferences, E3S Web of Conferences, Tallinn, Estonia, 6–9 September 2020; Volume 179, p. 02091. [Google Scholar]
  25. Zhang, Y.; Huang, N.; Tang, F.; Huang, H.; Ma, C.; Dong, W.; Xu, C. Inversion-based style transfer with diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 10146–10156. [Google Scholar]
  26. Cai, Q.; Ma, M.; Wang, C.; Li, H. Image neural style transfer: A review. Comput. Electr. Eng. 2023, 108, 108723. [Google Scholar] [CrossRef]
  27. Liu, K.; Zhan, F.; Chen, Y.; Zhang, J.; Yu, Y.; El Saddik, A.; Lu, S.; Xing, E.P. Stylerf: Zero-shot 3d style transfer of neural radiance fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 8338–8348. [Google Scholar]
  28. Woodland, M.; Wood, J.; Anderson, B.M.; Kundu, S.; Lin, E.; Koay, E.; Odisio, B.; Chung, C.; Kang, H.C.; Venkatesan, A.M.; et al. Evaluating the performance of StyleGAN2-ADA on medical images. In Simulation and Synthesis in Medical Imaging; Springer: Berlin/Heidelberg, Germany, 2022; pp. 142–153. [Google Scholar]
  29. Zhang, Y.; He, Z.; Xing, J.; Yao, X.; Jia, J. Ref-npr: Reference-based non-photorealistic radiance fields for controllable scene stylization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 4242–4251. [Google Scholar]
  30. Yang, G. The imagery and abstraction trend of Chinese contemporary oil painting. Linguist. Cult. Rev. 2021, 5, 454–471. [Google Scholar] [CrossRef]
  31. Liu, W. Analysis on the Collision and Fusion of Eastern and Western Paintings in the Context of Globalization. Thought 2021, 7, 8. [Google Scholar]
  32. Fan, Z.; Zhu, Y.; Yan, C.; Li, Y.; Zhang, K. A comparative study of color between abstract paintings, oil paintings and Chinese ink paintings. In Proceedings of the 15th International Symposium on Visual Information Communication and Interaction, Chur, Switzerland, 16–18 August 2022; pp. 1–8. [Google Scholar]
  33. Liu, F. Research on oil painting creation based on Computer Technology. J. Phys. Conf. Ser. 2021, 1915, 022005. [Google Scholar] [CrossRef]
  34. Wen, X.; White, P. The role of landscape art in cultural and national identity: Chinese and European comparisons. Sustainability 2020, 12, 5472. [Google Scholar] [CrossRef]
  35. Hongxian, L.; Tahir, A.; Bakar, S.A.S.A. The Developing Process of Ideological Trend of the Nationalization in Chinese Oil Painting. Asian J. Res. Educ. Soc. Sci. 2024, 6, 465–474. [Google Scholar]
  36. Ao, J.; Ye, Z.; Li, W.; Ji, S. Impressions of Guangzhou city in Qing dynasty export paintings in the context of trade economy: A color analysis of paintings based on k-means clustering algorithm. Herit. Sci. 2024, 12, 77. [Google Scholar] [CrossRef]
  37. Sheng, J.; Song, C.; Wang, J.; Han, Y. Convolutional neural network style transfer towards Chinese paintings. IEEE Access 2019, 7, 163719–163728. [Google Scholar] [CrossRef]
  38. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  39. Wei, Z.; Dong, P.; Hui, Z.; Li, A.; Li, L.; Lu, M.; Pan, H.; Li, D. Auto-prox: Training-free vision transformer architecture search via automatic proxy discovery. In Proceedings of the 38th AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 15814–15822. [Google Scholar]
  40. Fan, Q.; Huang, H.; Chen, M.; Liu, H.; He, R. Rmt: Retentive networks meet vision transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 5641–5651. [Google Scholar]
  41. Da, C.; Luo, C.; Zheng, Q.; Yao, C. Vision grid transformer for document layout analysis. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 19462–19472. [Google Scholar]
  42. Tang, C.; Zhang, L.L.; Jiang, H.; Xu, J.; Cao, T.; Zhang, Q.; Yang, Y.; Wang, Z.; Yang, M. Elasticvit: Conflict-aware supernet training for deploying fast vision transformer on diverse mobile devices. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 5829–5840. [Google Scholar]
  43. Ji, G.P.; Zhuge, M.; Gao, D.; Fan, D.P.; Sakaridis, C.; Gool, L.V. Masked vision-language transformer in fashion. Mach. Intell. Res. 2023, 20, 421–434. [Google Scholar] [CrossRef]
  44. Wensel, J.; Ullah, H.; Munir, A. Vit-ret: Vision and recurrent transformer neural networks for human activity recognition in videos. IEEE Access 2023, 11, 72227–72249. [Google Scholar] [CrossRef]
  45. Liu, Y.; Matsoukas, C.; Strand, F.; Azizpour, H.; Smith, K. Patchdropout: Economizing vision transformers using patch dropout. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 3953–3962. [Google Scholar]
  46. Wang, Y.; Lu, L.; Yang, W.; Chen, Y. Local or global? A novel transformer for Chinese named entity recognition based on multi-view and sliding attention. Int. J. Mach. Learn. Cybern. 2024, 15, 2199–2208. [Google Scholar] [CrossRef]
  47. Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jégou, H. Training data-efficient image transformers & distillation through attention. In Proceedings of the International Conference on Machine Learning (PMLR), Virtual, 18–24 July 2021; pp. 10347–10357. [Google Scholar]
  48. Deng, Y.; Tang, F.; Dong, W.; Ma, C.; Pan, X.; Wang, L.; Xu, C. Stytr2: Image style transfer with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11326–11336. [Google Scholar]
  49. Graham, B.; El-Nouby, A.; Touvron, H.; Stock, P.; Joulin, A.; Jégou, H.; Douze, M. Levit: A vision transformer in convnet’s clothing for faster inference. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 12259–12269. [Google Scholar]
  50. Fan, H.; Xiong, B.; Mangalam, K.; Li, Y.; Yan, Z.; Malik, J.; Feichtenhofer, C. Multiscale vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 6824–6835. [Google Scholar]
  51. Li, W.; Chen, Y.; Guo, X.; He, X. ST2SI: Image Style Transfer via Vision Transformer using Spatial Interaction. Comput. Graph. 2024, 124, 104084. [Google Scholar] [CrossRef]
  52. Deng, Y.; He, X.; Tang, F.; Dong, W. Z*: Zero-shot Style Transfer via Attention Reweighting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 6934–6944. [Google Scholar]
  53. Liu, Y.; Yu, W.; Zhang, Z.; Wang, Q.; Che, L. Axial Attention Transformer for Fast High-quality Image Style Transfer. In Proceedings of the 2024 IEEE International Symposium on Circuits and Systems (ISCAS), Singapore, 19–22 May 2024; pp. 1–5. [Google Scholar]
  54. Wu, H.; Xiao, B.; Codella, N.; Liu, M.; Dai, X.; Yuan, L.; Zhang, L. Cvt: Introducing convolutions to vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 22–31. [Google Scholar]
  55. Zhu, M.; He, X.; Wang, N.; Wang, X.; Gao, X. All-to-key attention for arbitrary style transfer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 1–6 October 2023; pp. 23109–23119. [Google Scholar]
  56. Qu, C.; Lu, L.; Wang, A.; Yang, W.; Chen, Y. Novel multi-domain attention for abstractive summarisation. CAAI Trans. Intell. Technol. 2023, 8, 796–806. [Google Scholar] [CrossRef]
  57. Xue, M.; He, J.; He, Y.; Liu, Z.; Wang, W.; Zhou, M. Low-light image enhancement via clip-fourier guided wavelet diffusion. arXiv 2024, arXiv:2401.03788. [Google Scholar]
  58. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  59. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2472–2481. [Google Scholar]
  60. Park, D.Y.; Lee, K.H. Arbitrary style transfer with style-attentional networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 5880–5888. [Google Scholar]
  61. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part V 13. Springer: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
  62. Phillips, F.; Mackintosh, B. Wiki art gallery, inc.: A case for critical thinking. Issues Account. Educ. 2011, 26, 593–608. [Google Scholar] [CrossRef]
  63. Ghiasi, G.; Lee, H.; Kudlur, M.; Dumoulin, V.; Shlens, J. Exploring the structure of a real-time, arbitrary neural artistic stylization network. arXiv 2017, arXiv:1705.06830. [Google Scholar]
Figure 1. This figure shows the results of our method to convert the content image to the Wuhu Iron Paintings style. Due to the rich lines, layers, and unique texture of Wuhu Iron Paintings, the current technical solutions have not achieved the desired effect.
Figure 1. This figure shows the results of our method to convert the content image to the Wuhu Iron Paintings style. Due to the rich lines, layers, and unique texture of Wuhu Iron Paintings, the current technical solutions have not achieved the desired effect.
Sensors 24 08103 g001
Figure 2. (a) demonstrates the structure of our framework. I c , I s denote content image and style image, respectively. f c , f s are the features extracted for content and style images. f s c denotes the transformed features obtained through the decode. (b) demonstrates the Transformer Encoder Layer. CCM denotes the Content Correction Module. (c) demonstrates the Transformer Decoder Layer. (d) demonstrates the ELA-Decoder.
Figure 2. (a) demonstrates the structure of our framework. I c , I s denote content image and style image, respectively. f c , f s are the features extracted for content and style images. f s c denotes the transformed features obtained through the decode. (b) demonstrates the Transformer Encoder Layer. CCM denotes the Content Correction Module. (c) demonstrates the Transformer Decoder Layer. (d) demonstrates the ELA-Decoder.
Sensors 24 08103 g002
Figure 3. (a) shows the computation of Strips Window Attention. (b) shows the structure of Efficient Local Attention.
Figure 3. (a) shows the computation of Strips Window Attention. (b) shows the structure of Efficient Local Attention.
Sensors 24 08103 g003
Figure 4. The architecture of the Content Correction Module.
Figure 4. The architecture of the Content Correction Module.
Sensors 24 08103 g004
Figure 5. The content loss and style loss during the training process vary with the number of iterations.
Figure 5. The content loss and style loss during the training process vary with the number of iterations.
Sensors 24 08103 g005
Figure 6. Visualization results of landscape paintings and flora and fauna on the Chinese Wuhu Iron Painting dataset. We compare the effects between our method and previous SOTA methods Ghiasi et al. [63], CAST [19], StyTr2 [48], and S2WAT [16].
Figure 6. Visualization results of landscape paintings and flora and fauna on the Chinese Wuhu Iron Painting dataset. We compare the effects between our method and previous SOTA methods Ghiasi et al. [63], CAST [19], StyTr2 [48], and S2WAT [16].
Sensors 24 08103 g006
Figure 7. Visualization results of Furniture Types on the Chinese Wuhu Iron Painting dataset. We compare the effects between our method and previous SOTA methods Ghiasi et al. [63], CAST [19], StyTr2 [48], and S2WAT [16].
Figure 7. Visualization results of Furniture Types on the Chinese Wuhu Iron Painting dataset. We compare the effects between our method and previous SOTA methods Ghiasi et al. [63], CAST [19], StyTr2 [48], and S2WAT [16].
Sensors 24 08103 g007
Figure 8. Partial visualization results of Wuhu Iron Paintings style transfer.
Figure 8. Partial visualization results of Wuhu Iron Paintings style transfer.
Sensors 24 08103 g008
Figure 9. ELA-Decoder Module for ablation experiments. Decoder indicates the use of a basic VGG setup.
Figure 9. ELA-Decoder Module for ablation experiments. Decoder indicates the use of a basic VGG setup.
Sensors 24 08103 g009
Figure 10. Ablation experiments with content correction module; w/o indicates that no content correction module was used.
Figure 10. Ablation experiments with content correction module; w/o indicates that no content correction module was used.
Sensors 24 08103 g010
Figure 11. Application effect of Wuhu Iron Paintings style conversion method in product design and 3D modeling effect.
Figure 11. Application effect of Wuhu Iron Paintings style conversion method in product design and 3D modeling effect.
Sensors 24 08103 g011
Figure 12. The effect of Wuhu Iron Painting style conversion method in environmental design and 3D modeling effect.
Figure 12. The effect of Wuhu Iron Painting style conversion method in environmental design and 3D modeling effect.
Sensors 24 08103 g012
Figure 13. The effect of Wuhu Iron Paintings style conversion method in the application of digital application.
Figure 13. The effect of Wuhu Iron Paintings style conversion method in the application of digital application.
Sensors 24 08103 g013
Figure 14. Practical product results for digital applications.
Figure 14. Practical product results for digital applications.
Sensors 24 08103 g014
Figure 15. A visualization of the various categories of the Wuhu Iron Paintings dataset.
Figure 15. A visualization of the various categories of the Wuhu Iron Paintings dataset.
Sensors 24 08103 g015
Figure 16. The limitations of our method. When the style image and content image are too similar or when the content image contains overly complex details, our method still encounters certain challenges.
Figure 16. The limitations of our method. When the style image and content image are too similar or when the content image contains overly complex details, our method still encounters certain challenges.
Sensors 24 08103 g016
Table 1. Wuhu Iron Painting style characteristics extraction table.
Table 1. Wuhu Iron Painting style characteristics extraction table.
TypeSpecific FeaturesFeature Description
ModelingFlowing lineUnlike other images, Wuhu Iron Paintings are made of steel title, focusing on the expression of modeling lines.
BrillianceThe material of Wuhu Iron Paintings is metal, which has a strong metallic luster, which is different from the data used in the current field.
Simple outlineIn the production of Wuhu Iron Paintings, the shape of the object will be highly generalized, forming a brief and condensed stylistic imagery.
ColorBlack and whiteIn terms of color, Wuhu Iron Paintings inherit the characteristics of Chinese painting, often with white as the background, and black lines for the composition of the painting, to create a strong contrast between black and white, and with the Chinese painting of the color of the beloved ink it coincides with.
Hairstyle3D artistryWuhu Iron Paintings is a three-dimensional painting with a unique three-dimensional texture, and its three-dimensionality is mainly divided into two kinds: one is the height of the object itself, and the other is the level of interspersed between the objects. The data used in the current field are mostly flat, which also shows that it is challenging for us to migrate the style of Wuhu Iron Paintings.
Leave a blank pageWuhu Iron Paintings, in terms of the white pictures from the influence of Chinese painting, have specific whites, divided into the following four kinds: one is the composition of white, two is the mood of white, three is the reality of white, and four is the scene of white. This situation is also one of the difficulties we face, because Western art images do not have this type of art.
Table 2. Qualitative results of furniture types on the Chinese Wuhu Iron Paintings dataset.
Table 2. Qualitative results of furniture types on the Chinese Wuhu Iron Paintings dataset.
TypeCASTStyTr2S2WatWCTOurs
Content Loss ↓2.171.911.672.561.62
Style Loss ↓4.431.671.752.231.63
Time(seconds) ↓0.0420.2370.5580.5900.573
Table 3. Comparison of experimental output effect map voting result statistics. The red font indicates that it received the most votes.
Table 3. Comparison of experimental output effect map voting result statistics. The red font indicates that it received the most votes.
MethodsGhiasi et al. [63]CASTStyTr2S2WATOurs
Animal species sampleSensors 24 08103 i001Sensors 24 08103 i002Sensors 24 08103 i003Sensors 24 08103 i004Sensors 24 08103 i005
1947251869
Landscape sampleSensors 24 08103 i006Sensors 24 08103 i007Sensors 24 08103 i008Sensors 24 08103 i009Sensors 24 08103 i010
1128353371
Houseware -1Sensors 24 08103 i011Sensors 24 08103 i012Sensors 24 08103 i013Sensors 24 08103 i014Sensors 24 08103 i015
1626411679
Houseware -2Sensors 24 08103 i016Sensors 24 08103 i017Sensors 24 08103 i018Sensors 24 08103 i019Sensors 24 08103 i020
1434462954
Table 4. Table of expert scoring and assessment results.
Table 4. Table of expert scoring and assessment results.
TypeSamples
Animal SamplesLandscape SampleHouseware SampleHouseware Samples
Sensors 24 08103 i021Sensors 24 08103 i022Sensors 24 08103 i023Sensors 24 08103 i024
Line fluidity3.984.274.024.18
Metallic expression4.164.074.294.30
Contour simplicity3.914.114.234.25
Black and white contrast effect4.254.403.984.14
Stereoscopic depth of picture4.354.144.254.30
Leave a blank page4.094.213.964.17
Aggregate score4.124.204.124.22
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, Y.; Ren, Y.; Wu, C.; Xue, M. Style Transfer of Chinese Wuhu Iron Paintings Using Hierarchical Visual Transformer. Sensors 2024, 24, 8103. https://doi.org/10.3390/s24248103

AMA Style

Zhou Y, Ren Y, Wu C, Xue M. Style Transfer of Chinese Wuhu Iron Paintings Using Hierarchical Visual Transformer. Sensors. 2024; 24(24):8103. https://doi.org/10.3390/s24248103

Chicago/Turabian Style

Zhou, Yuying, Yao Ren, Chao Wu, and Minglong Xue. 2024. "Style Transfer of Chinese Wuhu Iron Paintings Using Hierarchical Visual Transformer" Sensors 24, no. 24: 8103. https://doi.org/10.3390/s24248103

APA Style

Zhou, Y., Ren, Y., Wu, C., & Xue, M. (2024). Style Transfer of Chinese Wuhu Iron Paintings Using Hierarchical Visual Transformer. Sensors, 24(24), 8103. https://doi.org/10.3390/s24248103

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop