Stylization of a Seismic Image Proﬁle Based on a Convolutional Neural Network

: Seismic data are widely used in oil, gas, and other kinds of mineral exploration and development. However, due to low artiﬁcial interpretation accuracy and small sample sizes, seismic data may not meet the needs of convolutional neural network training. There are major differences between optical image and seismic data, making it difﬁcult for a model to learn seismic data characteristics. Therefore, a style transfer network is necessary to make the styles of optical image and seismic data more similar. Since the stylization effect of a seismic section is similar to that of most art styles, based on an in-depth study of image style transfer, this paper compared the effects of various style transfer models, and selected a Laplacian pyramid network to carry out a study of seismic section stylization. It transmits low-resolution global style patterns through a drafting network, revises high-resolution local details through correction networks, and aggregates all pyramid layers to output ﬁnal stylized images of seismic proﬁles. Experiments show that this method can effectively convey the whole style pattern without losing the original image content. This style transfer method, based on the Laplacian pyramid network, provides theoretical guidance for the fast and objective application of the model to seismic data features.


Introduction
Seismic data interpretation is involved in most geological research fields [1], so the precision of results is of great significance for oil and gas and mineral exploration and development. Faults are characterized by low continuity or high discontinuity of cophase axes in seismic profiles. Seismic data are still interpreted manually or through humancomputer interaction. Conventional interpretation [2] is based on the principle of wave group contrast, the results of which are obtained by seismic profile recognition and plane combination. The three-dimensional coherence technique [3][4][5] uses data volume slices to interpret faults. The variogram technique [6,7] calculates the optimum time window and sampling point of the variance values to reveal fault information. However, non-geological factors such as noise and seismic acquisition errors cause seismic cophase axis discontinuity, which makes the abovementioned methods very sensitive to them. While detecting faults, some information unrelated to faults will be highlighted. The ant colony algorithm [8,9] searches for fault lines based on a positive feedback mechanism. Edge detection [10,11] highlights continuous and discontinuous information in the profile. The optimal facet vote [12] selects the fault from the attribute image as the optimal path to search for the global maximum. These post-processing methods further strengthen the characteristics of faults in seismic attributes and suppress noise. In recent years, seismic fault interpretation has moved towards automation and artificial intelligence. Fault recognition [13] based on a convolutional neural network [14][15][16][17] continuously optimizes parameters and weights during cyclic training and continuously learns the shallow and deep characteristics of faults in seismic profiles [18]. Finally, the fault recognition model is trained and predicted to achieve the fault recognition effect. However, artificial intelligence still has limitations in its application to seismic data processing and interpretation: (1) Artificial interpretation of faults may be time-consuming and subjective, and inaccurate artificial interpretations may mislead the model learning process [19]. (2) Training and validation of models often require a large number of calibrated seismic images, but it is difficult to obtain large datasets with interpretation results using artificial interpretation. Therefore, the key problem of insufficient calibration datasets must be faced if artificial intelligence interpretation methods are to improve the efficiency of fault identification.
Style transfer can expand datasets by adaptively integrating the style of one image into another and converting the original content to a completely different style. It has been applied in medicine, aviation, remote sensing, animation, film, and television. In medicine, the WaveCT-AIN style transmission framework [20] guides optical image data to learn ultrasonic image data features. The fast style transfer model [21,22] makes optical image and fog data similar in style using a continuous background with no deformation of content. The style transfer network for SAR image conversion to optical images [23] solves the problem of insufficient sample sizes of artificial SAR images. In migrating animation style to optical image data [24][25][26], the generated cartoon image quality is high.
A style transfer network can eliminate huge differences in numerical distributions between optical and seismic image data styles and make them more similar. In this paper, a Laplacian pyramid network was selected to study the stylization of seismic sections. Through cosine similarity optimization, the image style of seismic data was transferred to a common objects in context (COCO) dataset, so that the whole dataset image had the image style of seismic data. At the same time, the distribution characteristics of stylized COCO datasets were similar to those for seismic data. Therefore, the style transfer method provided a richer calibration database for artificial intelligence interpretation algorithms.

Materials and Methods
This section mainly introduces the style transfer algorithms based on a convolutional neural network used in image seismic profile experiments and the necessary datasets.

Style Transfer Algorithm
In this section, we introduce some generative adversarial network (GAN)-based style migration algorithms, with CycleGAN, Unsupervised Generative Attentional Networks with Adaptive Layer-Instance Normalization for Image-to-Image Translation (U-GAT-IT) and Lapstyle as representative models. We also describe in detail the algorithmic principles of the models, the application formulas and the losses during training.

CycleGAN
First proposed by Jun-Yan Zhu (2017), CycleGAN [27], GAN is a generative adversarial network that realizes the function of image style conversion between two types of images without a corresponding relationship. As shown in Figure 1, a CycleGAN consists of two symmetric GANs that form a ring network that shares two generators while each carries a discriminator.
The loss included antagonism L GAN and cyclic consistency loss L cyc . The former ensured that the generator and discriminator mutually evolved to produce more realistic images; the latter that the input and output images had the same content but different styles, because images must be converted between source and target domains. The antagonism is   [27]. Model contains the two mapping functions G: X → Y and F: Y → X, and the associated adversarial discriminators D Y and D X . D Y encourages G to translate X into outputs indistinguishable from the domain Y, and vice versa for D X and F. We introduced two cycle-consistency losses that illustrated the idea of translating from one domain to the other and back. We should arrive at where we started: forward cycle-consistency loss: x → G(x) → F(G(x)) ≈ x, and backward cycle-consistency loss: y → F(y) → G(F(y)) ≈ y.
The loss included antagonism L GAN and cyclic consistency loss L cyc . The former ensured that the generator and discriminator mutually evolved to produce more realistic images; the latter that the input and output images had the same content but different styles, because images must be converted between source and target domains. The antagonism is During the mapping from X to Y, G stands for generator and tries to generate an image G(x) that looks similar to the image in the target domain Y, and D Y distinguishes the generated image G(x) from the real image y. log D Y (y) is the probability that the discriminator recognized the true data as true data; and log(1 − D Y (G(x)) is the probability that the discriminator still recognized the false data generated by the generator as false data. The adversarial loss is the sum of the two. The cyclic consistency loss is where x and y are two image domains, G is the generator of close y samples; and F refers to the generator of close x samples. When we fed x into G, we got a fake y graph called G(x). Then we fed G(x) into F to get an even faker x graph called F(G(x)). This constituted a cycle. F(G(x)) was the generated G(x) mapped back to X, and G(F(y)) was defined similarly.

U-GAT-IT
Based on CycleGAN, U-GAT-IT [28] adds an adaptive instance (AdaIN) regularization layer to control the style transfer process for image-to-image translation. As shown in Figure 2, the generator carries out image downsampling on the input terminal and combines it with residual blocks to enhance image feature extraction. The attention module is guided by AdaIN to obtain the transformed image by upsampling the residual block. The discriminator converts the decoding process into an authentication output.  [27]. Model contains the two mapping functions G : X → Y and F : Y → X , and the associated adversarial discriminators D Y and D X . D Y encourages G to translate X into outputs indistinguishable from the domain Y, and vice versa for D X and F. We introduced two cycle-consistency losses that illustrated the idea of translating from one domain to the other and back. We should arrive at where we started: forward cycle-consistency loss: x → G(x) → F(G(x)) ≈ x , and backward cycle-consistency loss: During the mapping from X to Y, G stands for generator and tries to generate an image G(x) that looks similar to the image in the target domain Y, and D Y distinguishes the generated image G(x) from the real image y. log D Y (y) is the probability that the discriminator recognized the true data as true data; and log(1 − D Y (G(x)) is the probability that the discriminator still recognized the false data generated by the generator as false data. The adversarial loss is the sum of the two. The cyclic consistency loss is where x and y are two image domains, G is the generator of close y samples; and F refers to the generator of close x samples. When we fed x into G, we got a fake y graph called G(x). Then we fed G(x) into F to get an even faker x graph called F(G(x)). This constituted a cycle. F(G(x)) was the generated G(x) mapped back to X, and G(F(y)) was defined similarly.

U-GAT-IT
Based on CycleGAN, U-GAT-IT [28] adds an adaptive instance (AdaIN) regularization layer to control the style transfer process for image-to-image translation. As shown in Figure 2, the generator carries out image downsampling on the input terminal and combines it with residual blocks to enhance image feature extraction. The attention module is guided by AdaIN to obtain the transformed image by upsampling the residual block. The discriminator converts the decoding process into an authentication output.
The model's objectives include four loss functions, including antagonistic loss, which are used to match the generated and target image distributions: In the mapping process from source domain s to target domain t, G s→t (x) refers to the false data generated from source domain s. The probability that the discriminator will still determine the false data generated by the generator in the source domain as false data is given by 1 − D t (G s→t (x)). D t (x) is the probability that the discriminator will determine the true data as true data, and the adversarial loss is the sum of the two. Cyclic consistency loss reduced the mapping path from the source to the target domain. Given an image x ∈ X s , after the sequential translations of x from X s to X t and from X t to X s , the image should be successfully translated back to the original domain: where G s→t (x) refers to the false data generated from the source domain s. We fed G s→t (x) into G t→s to get the false data G t→s (G s→t (x)), which ideally should be comparable to the x. This constitutes a cycle. We applied an identity consistency constraint to the generator to preserve the consistency of the color composition of the input and output images. Given an image x ∈ X t , after the translation of x using G s→t , the image should not change: Energies 2022, 15, x FOR PEER REVIEW 4 of 16 Figure 2. Network structure of normalized unsupervised generation of adaptive layer instances [28]. The generator has a decoder module implemented by the AdaLIN algorithm more than in the discriminator. The input image of the generator creates the feature map through the encoding stage E s of the Encoder, and then an auxiliary classifier η s is added. The attention mechanism is introduced to maximize the pooling of the feature map. The prediction of a node is output through the full connection layer, and then the parameters of the full connection layer and the feature map are multiplied to obtain the feature map of attention. Finally, the output image is obtained through Decoder module G t . The discriminator compresses the input image at a deeper level and adds an auxiliary classifier η D t . Finally, the output is judged by the classifier C D t .
The model's objectives include four loss functions, including antagonistic loss, which are used to match the generated and target image distributions: In the mapping process from source domain s to target domain t, G s→t (x) refers to the false data generated from source domain s. The probability that the discriminator Figure 2. Network structure of normalized unsupervised generation of adaptive layer instances [28]. The generator has a decoder module implemented by the AdaLIN algorithm more than in the discriminator. The input image of the generator creates the feature map through the encoding stage E s of the Encoder, and then an auxiliary classifier η s is added. The attention mechanism is introduced to maximize the pooling of the feature map. The prediction of a node is output through the full connection layer, and then the parameters of the full connection layer and the feature map are multiplied to obtain the feature map of attention. Finally, the output image is obtained through Decoder module G t . The discriminator compresses the input image at a deeper level and adds an auxiliary classifier η D t . Finally, the output is judged by the classifier C D t . Class activation map (CAM) loss is the biggest difference between the source and target domain. The auxiliary classifiers η s and η D t help the generator and discriminator to evolve. For an image x ∈ {X s , X t }, G s→t and D t realize what makes the most difference between the two domains in the current state: L s→t cam refers to the CAM loss of the generator in the source domain. If x is the true data, the cross-entropy output is −(E x∼X s [log(η s (x))]); if x is the false generated data, the crossentropy output is E x∼X t [log(1 − η s (x))]. In order to make the generator as false as possible, −(E x∼X s [log(η s (x))]) should be as small as possible. The higher the discriminator's ability to judge true data and false data, the better. η D t (x) refers to the probability that the auxiliary classifier η D t collaborated with the discriminator to judge true data as true data as true data, and 1 − η D t (G s→t (x)) refers to the probability that the auxiliary classifier η D t collaborated with the discriminator to continue to judge false data from the generator in the source domain as false data. The sum of the two, L D t cam , is the CAM loss in the discriminator from the source domain to the target domain.

Lapstyle
Lapstyle [29] transmits a low-resolution global style mode through a drafting network, revises local details with high resolution through a correction network, and generates image residuals according to the image texture and sketch extracted by a Laplace filter. Higher-resolution details can easily be generated by superimposing revision networks with multiple levels of the Laplacian pyramid. All pyramid-level output is aggregated to obtain the final stylized image, as shown in Figure 3.
ble to the x. This constitutes a cycle. We applied an identity consistency constraint to the generator to preserve the consistency of the color composition of the input and output images. Given an image x ∈ X t , after the translation of x using G s→t , the image should not change: Class activation map (CAM) loss is the biggest difference between the source and target domain. The auxiliary classifiers η s and η D t help the generator and discriminator to evolve. For an image x ∈ {X s , X t }, G s→t and D t realize what makes the most difference between the two domains in the current state: L cam s→t refers to the CAM loss of the generator in the source domain. If x is the true data, the cross-entropy output is −(E x~X s [log(η s (x))]); if x is the false generated data, the cross-entropy output is E x~X t [log(1 − η s (x))]. In order to make the generator as false as possible, −(E x~X s [log(η s (x))]) should be as small as possible. The higher the discriminator's ability to judge true data and false data, the better. η D t (x) refers to the probability that the auxiliary classifier η D t collaborated with the discriminator to judge true data as true data as true data, and 1 − η D t (G s→t (x)) refers to the probability that the auxiliary classifier η D t collaborated with the discriminator to continue to judge false data from the generator in the source domain as false data. The sum of the two, L cam D t , is the CAM loss in the discriminator from the source domain to the target domain.

Lapstyle
Lapstyle [29] transmits a low-resolution global style mode through a drafting network, revises local details with high resolution through a correction network, and generates image residuals according to the image texture and sketch extracted by a Laplace filter. Higher-resolution details can easily be generated by superimposing revision networks with multiple levels of the Laplacian pyramid. All pyramid-level output is aggregated to obtain the final stylized image, as shown in Figure 3.  [29]. Firstly, the content image is reduced by two times, and L, C and A in the image separately represent Laplacian concatenation and aggregation operations. At this time, the predefined style image is also reduced by two times to a low resolution version The drafting network includes an encoder, several AdaIN modules and a decoder, as shown in Figure 4. The visual geometry group (VGG) encoder extracts corresponding granularity features at layers 2_1, 3_1, and 4_1, feeds them into an AdaIN module for aggregation, and merges them by jumping connection layers.
to a low resolution versionx s . In the drafting network stage, the content imagex c and the style imagex s after double down-sampling are generated through a encoder-decoder AdaIN module to generate a global but not detailed stylized imagex cs . In the revision stage of the network, twice the up-sampling processing x cs ′ and connecting with the residual detail image as the input, the stylized residual detail image is generated through a encoder-decoder module. Finally, we aggre gate the two-level image pyramid and output the final stylized image x cs .
The drafting network includes an encoder, several AdaIN modules and a decoder as shown in Figure 4. The visual geometry group (VGG) encoder extracts corresponding granularity features at layers 2_1, 3_1, and 4_1, feeds them into an AdaIN module for aggregation, and merges them by jumping connection layers. The revision network modified the rough stylized image by generating a residua detail graph r cs of the image, and combined it with the rough stylized image x cs to generate the final stylized image. As shown in Figure 5, the revision network has a simple, efficient codec architecture with only one undersampling and one upsampling layer An image block discriminator helps the revision network capture fine block textures under adversarial learning.  The revision network modified the rough stylized image by generating a residual detail graph r cs of the image, and combined it with the rough stylized image − x cs to generate the final stylized image. As shown in Figure 5, the revision network has a simple, efficient codec architecture with only one undersampling and one upsampling layer. An image block discriminator helps the revision network capture fine block textures under adversarial learning.
to a low resolution versionx s . In the drafting network stage, the content imagex c and the sty imagex s after double down-sampling are generated through a encoder-decoder AdaIN module generate a global but not detailed stylized imagex cs . In the revision stage of the network, twi the up-sampling processing x cs ′ and connecting with the residual detail image as the input, th stylized residual detail image is generated through a encoder-decoder module. Finally, we aggr gate the two-level image pyramid and output the final stylized image x cs .
The drafting network includes an encoder, several AdaIN modules and a decode as shown in Figure 4. The visual geometry group (VGG) encoder extracts correspondin granularity features at layers 2_1, 3_1, and 4_1, feeds them into an AdaIN module fo aggregation, and merges them by jumping connection layers. The revision network modified the rough stylized image by generating a residu detail graph r cs of the image, and combined it with the rough stylized image x cs generate the final stylized image. As shown in Figure 5, the revision network has a sim ple, efficient codec architecture with only one undersampling and one upsampling laye An image block discriminator helps the revision network capture fine block textur under adversarial learning.  [29]. C and A represent concatenation and a gregation operations separately. The residual detail image r c is connected with the result imag x cs ′ of the drafting network stage after double up-sampling. After revising the network, the resi During training, the drafting and revision networks optimized the content and style, and the revision network adopted the adversarial loss function. We described the style and content loss and introduced the goals of each network. The Laplacian pyramid network is a single style for a single model during training, keeping one x s and one set of x c from the content set X c .
Following the STROTSS method, the rEMD and the commonly used meanvariance loss functions were combined into a style loss function. Given an image, we used the advance training VGG-19 encoder to extract a set of characteristic vectors for The content loss function is adopted between F c ∈ R h c w c ×c and F cs ∈ R h cs w cs ×c using normalized perceived loss and self-similarity loss. h cs was equal to h c , and w c was equal to w cs , because x c and x cs had the same resolution. Perceptual loss is defined as where norm represents the normalized channel direction F . The purpose of self-similar loss is to maintain the relative relationship between the content and stylized images and is defined as where D c ij and D cs ij are the (i, j) th terms of self-similar matrices D c and D cs , respectively. Here, D ij is the cosine similarity F i , F j .
In the training stage of the revision network, the parameters of the drafting network were fixed and the training loss was based on x cs . To better learn local fine grain textures, a discriminator was introduced, and a revision network with an antagonistic loss term was trained. The overall loss function is defined as where Rev denotes the revision network, D denotes the discriminator, and β controls the balance between basic style transfer loss and antagonism loss. L base is the basic content loss, and L adv is the standard adversarial training loss.

Seismic Dataset Preparation
Herein, we discuss the required data, including seismic data, from a work area and the open-source COCO image instance segmentation dataset. The seismic profile classification dataset, shown in Figure 6, contains only seismic profiles and no artificial interpretations to avoid data leakage. These seismic profiles were selected because the faults were less developed, and the alignment of the in-phase axis was clear, thus avoiding ambiguity from the influence of geological structures or data processing. This avoided the influence of faults in the seismic profile classification dataset on the training of the Laplacian pyramid network. The seismic section classification dataset was only used in the image style migration network. We did not need to mark it, as it only provided the unique type, i.e., the seismic data class, used to pre-train the classification network in the style migration network, which this time was used to transfer the global style and local texture of the seismic section. Therefore, after migration, the image only had the sequence texture features of seismic data, and there were no obvious fault features to influence subsequent modeling.
ambiguity from the influence of geological structures or data processing. This the influence of faults in the seismic profile classification dataset on the trainin Laplacian pyramid network. The seismic section classification dataset was only the image style migration network. We did not need to mark it, as it only prov unique type, i.e., the seismic data class, used to pre-train the classification netwo style migration network, which this time was used to transfer the global style a texture of the seismic section. Therefore, after migration, the image only had the s texture features of seismic data, and there were no obvious fault features to in subsequent modeling.

Results
This section focuses on how the three models CycelGAN, U-GAT-IT, and are network trained in seismic image profile stylization. We compared and analy results after stylized mapping both subjectively and objectively and use them to the quality of the stylized images.

Network Training of Each Algorithm
In the experiments, CycleGAN and U-GAT-IT were trained on 256 pixels parameter batch size set to 1. The Lapstyle experiment had three stages: the network trained on 128 pixels, the revision network trained on 256 pixels with size of 14 and the revision network trained on 512 pixels, with a batch size of 5.
The training results of the drafting network are shown in Figure 8, from whi be seen that the content and style of the resulting graph were retained globally u premise of migrating the profile style, so the Lapstyle experiment could be contin Microsoft Common Objects in Context (MS COCO) is a large, rich, object detection, segmentation, and subtitled dataset that takes scene understanding as the target and is mainly extracted from complex daily scenes. The target in the image is calibrated by precise instance annotation. The dataset includes 91 categories of objects, 328,000 images, and 2.5 million labels. So far, the largest dataset with semantic segmentation has 80 categories and more than 330,000 images of which 200,000 are labeled. Some images are shown in Figure 7. ambiguity from the influence of geological structures or data processing. This avoided the influence of faults in the seismic profile classification dataset on the training of the Laplacian pyramid network. The seismic section classification dataset was only used in the image style migration network. We did not need to mark it, as it only provided the unique type, i.e., the seismic data class, used to pre-train the classification network in the style migration network, which this time was used to transfer the global style and local texture of the seismic section. Therefore, after migration, the image only had the sequence texture features of seismic data, and there were no obvious fault features to influence subsequent modeling. Microsoft Common Objects in Context (MS COCO) is a large, rich, object detection, segmentation, and subtitled dataset that takes scene understanding as the target and is mainly extracted from complex daily scenes. The target in the image is calibrated by precise instance annotation. The dataset includes 91 categories of objects, 328,000 images, and 2.5 million labels. So far, the largest dataset with semantic segmentation has 80 categories and more than 330,000 images of which 200,000 are labeled. Some images are shown in Figure 7.

Results
This section focuses on how the three models CycelGAN, U-GAT-IT, and Lapstyle are network trained in seismic image profile stylization. We compared and analyzed the results after stylized mapping both subjectively and objectively and use them to evaluate the quality of the stylized images.

Network Training of Each Algorithm
In the experiments, CycleGAN and U-GAT-IT were trained on 256 pixels with the parameter batch size set to 1. The Lapstyle experiment had three stages: the drafting network trained on 128 pixels, the revision network trained on 256 pixels with a batch size of 14 and the revision network trained on 512 pixels, with a batch size of 5.
The training results of the drafting network are shown in Figure 8, from which it can be seen that the content and style of the resulting graph were retained globally under the premise of migrating the profile style, so the Lapstyle experiment could be continued.

Results
This section focuses on how the three models CycelGAN, U-GAT-IT, and Lapstyle are network trained in seismic image profile stylization. We compared and analyzed the results after stylized mapping both subjectively and objectively and use them to evaluate the quality of the stylized images.

Network Training of Each Algorithm
In the experiments, CycleGAN and U-GAT-IT were trained on 256 pixels with the parameter batch size set to 1. The Lapstyle experiment had three stages: the drafting network trained on 128 pixels, the revision network trained on 256 pixels with a batch size of 14 and the revision network trained on 512 pixels, with a batch size of 5.
The training results of the drafting network are shown in Figure 8, from which it can be seen that the content and style of the resulting graph were retained globally under the premise of migrating the profile style, so the Lapstyle experiment could be continued.   Figure 9 shows the top part of a woman framed by a red box. From left to right, the shape of the top becomes gradually clearer and the color more prominent. The effect of (f) is similar to that of (e), but in principle the greater the degree of style change the better because the image content can be recognized by the naked eye. Therefore, the migration   Figure 9 shows the top part of a woman framed by a red box. From left to right, the shape of the top becomes gradually clearer and the color more prominent. The effect of (f) is similar to that of (e), but in principle the greater the degree of style change the better because the image content can be recognized by the naked eye. Therefore, the migration  Figure 9 shows the top part of a woman framed by a red box. From left to right, the shape of the top becomes gradually clearer and the color more prominent. The effect of (f) is similar to that of (e), but in principle the greater the degree of style change the better because the image content can be recognized by the naked eye. Therefore, the migration effect of content weight 2 and style weight 1 was the best for the first modified network training. The training weight was assigned to the second modified network training, and five groups of control experiments were set. Figure 10 shows a woman's calf framed by a red box. In (c), the outline of the leg is not clear, while in (d), the details of the calf muscle can clearly be seen. Based on the above principles, the migration effect of content weight 1 and style weight 1 were the best in the second revision of network training.
Energies 2022, 15, x FOR PEER REVIEW 10 of 16 effect of content weight 2 and style weight 1 was the best for the first modified network training. The training weight was assigned to the second modified network training, and five groups of control experiments were set. Figure 10 shows a woman's calf framed by a red box. In (c), the outline of the leg is not clear, while in (d), the details of the calf muscle can clearly be seen. Based on the above principles, the migration effect of content weight 1 and style weight 1 were the best in the second revision of network training. Some of the experimental results are shown in Figure 11, where several representative images were selected to facilitate the comparison of the experimental results. On the whole, it can be seen that the image changed; however, the original content information style of the migrated images changed fundamentally, and their characteristics and backgrounds showed features of the in-phase axis of the seismic section. Figure 11c shows that all seismic profile styles were transferred at low resolution. The image is blurry, and the color of the image content is almost covered by that of the style image. Therefore, modification was carried out, and residual texture details were extracted at high resolution. In Figure 11d, on the basis of the successful migration of the seismic profile style, the color of the woman's clothes in the beach figure is differentiated; the motorcycle figure is clearly outlined; and the letters in the stamp figure can be identified. The variety and quality of textures in Figure 11e are significantly improved, and the stylistic characteristics of the seismic sections were obtained while preserving the content. The image is clearer, and the overall artistic effect of the visual image has been improved. Some of the experimental results are shown in Figure 11, where several representative images were selected to facilitate the comparison of the experimental results. On the whole, it can be seen that the image changed; however, the original content information style of the migrated images changed fundamentally, and their characteristics and backgrounds showed features of the in-phase axis of the seismic section. Figure 11c shows that all seismic profile styles were transferred at low resolution. The image is blurry, and the color of the image content is almost covered by that of the style image. Therefore, modification was carried out, and residual texture details were extracted at high resolution. In Figure 11d, on the basis of the successful migration of the seismic profile style, the color of the woman's clothes in the beach figure is differentiated; the motorcycle figure is clearly outlined; and the letters in the stamp figure can be identified. The variety and quality of textures in Figure 11e are significantly improved, and the stylistic characteristics of the seismic sections were obtained while preserving the content. The image is clearer, and the overall artistic effect of the visual image has been improved.   Figure 12 shows changes in the mean value of the generator and discriminator loss functions in the training of each network with the MS-COCO dataset transformed by the seismic profile style. Iter is the number of iterations, The Lapstyle and U-GAT-IT algorithms recorded every 3000 steps, and trained 30,000 pieces of data in total. The CycleGAN network generation model was limited by a rate of only 6000 iterations. During the training, the loss function values of CycleGAN, U-GAT-IT and Lapstyle all oscillated within an acceptable range; the convergence was general, and the obtained style transfer model was stable.

Comparative Analysis of Results
We selected CycleGAN, U-GAT-IT, and Lapstyle for comparative experiments, using the VGG-19 model for network training. The selected target content images were from the COCO dataset, and the style images were from the Yinggehai seismic profile dataset. Figure 13 shows the effect of the abovementioned style transfer method, and a comparison of the experimental results of sheep, beach, and food. The subjective visual effects of CycleGAN's experimental results showed that the original content image color was completely covered, and the stylized image lost part of the original content color information, resulting in ambiguity. U-GAT-IT improved upon CycleGAN, and the quality of the generated image obviously improved. In contrast, the stylized images drawn by the Lapstyle method were transformed from the real world into the seismic profile style based on the premise that the original content color and detail information were not lost, and both the overall contour migration and texture detail processing significantly improved. functions in the training of each network with the MS-COCO dataset transformed by the seismic profile style. Iter is the number of iterations, The Lapstyle and U-GAT-IT algorithms recorded every 3000 steps, and trained 30,000 pieces of data in total. The Cy-cleGAN network generation model was limited by a rate of only 6000 iterations. During the training, the loss function values of CycleGAN, U-GAT-IT and Lapstyle all oscillated within an acceptable range; the convergence was general, and the obtained style transfer model was stable.

Comparative Analysis of Results
We selected CycleGAN, U-GAT-IT, and Lapstyle for comparative experiments, using the VGG-19 model for network training. The selected target content images were from the COCO dataset, and the style images were from the Yinggehai seismic profile dataset. Figure 13 shows the effect of the abovementioned style transfer method, and a comparison of the experimental results of sheep, beach, and food. The subjective visual effects of CycleGAN's experimental results showed that the original content image color was completely covered, and the stylized image lost part of the original content color information, resulting in ambiguity. U-GAT-IT improved upon CycleGAN, and the

Discussion
It was necessary to evaluate the stylized image quality both subjectively and objectively. Visual observation showed that the Lapstyle style transfer effect was obviously better than that of the other two methods, but there was no unified index for objective evaluation. Therefore, we adopted the widely recognized Structure Similarity Index Measure (SSIM) and an image fusion quality evaluation index called Q AB/F [30] to evaluate the quality of the stylized images objectively, and use the running time of an algorithm to evaluate the efficiency of the migration method.
The SSIM measures the similarity of image structures in brightness, contrast, and structure. Q AB/F uses local metrics to estimate the performance of input style and content images in the fused image. A higher Q AB/F indicates a fused image of better quality. Figure 14 shows the SSIM, Q AB/F , and running time of stylized images generated by CycleGAN, U-GAT-IT, and Lapstyle. Compared with CycleGAN and U-GAT-IT, the stylized drawn image from Lapstyle had a higher fusion quality and greater structural similarity to the original image, indicating that its color information had been retained as much as possible and been transformed to the seismic profile style, which has the texture characteristics of the in-phase axis. The average running time of Lapstyle was 6.15 s which was better than that of CycleGAN or U-GAT-IT (12.20 s and 20.21 s, respectively) Lapstyle also showed a better overall visual effect.

Discussion
It was necessary to evaluate the stylized image quality both subjectively and objectively. Visual observation showed that the Lapstyle style transfer effect was obviously better than that of the other two methods, but there was no unified index for objective evaluation. Therefore, we adopted the widely recognized Structure Similarity Index Measure (SSIM) and an image fusion quality evaluation index called Q AB/F [30] to evaluate the quality of the stylized images objectively, and use the running time of an algorithm to evaluate the efficiency of the migration method.
The SSIM measures the similarity of image structures in brightness, contrast, and structure. Q AB/F uses local metrics to estimate the performance of input style and content images in the fused image. A higher Q AB/F indicates a fused image of better quality. Figure 14 shows the SSIM, Q AB/F , and running time of stylized images generated by CycleGAN, U-GAT-IT, and Lapstyle. Compared with CycleGAN and U-GAT-IT, the stylized drawn image from Lapstyle had a higher fusion quality and greater structural similarity to the original image, indicating that its color information had been retained as much as possible and been transformed to the seismic profile style, which has the texture characteristics of the in-phase axis. The average running time of Lapstyle was 6.15 s, which was better than that of CycleGAN or U-GAT-IT (12.20 s and 20.21 s, respectively). Lapstyle also showed a better overall visual effect. The COCO dataset was uniform in format, rich in content and variety, and suitable for training convolutional neural networks. The results of the seismic profiling style conversion based on the COCO dataset showed that features such as overall contour information and texture detail can be preserved better based on the selection of a suitable migration algorithm. Because the style migration method was based on the computer field of vision, it had a strong similarity to most of the fault interpretations carried out by seismic profiling. Similarly, this method is applicable to the interpretation of special geological phenomena, such as seismic facies or diapir identification, which are visually identified. After the style transfer of the COCO dataset was completed, the semantic segmentation model was trained for the COCO dataset with a stylized seismic profile. The training label maintained the label of the original COCO dataset. After training was completed, the front weight of the semantic segmentation model was frozen, and only the last two weights were retained as the training state. The model structure of the last two layers was modified, and the number of channels in the last layer was changed from 81 to 2. Next, the dataset was replaced by the fault dataset of the target work area from the stylized COCO dataset to reduce the learning rate and to retrain, meaning that only the weights of the last two layers were fine-tuned. After the loss value converged, the fault recognition model of the target work area based on the stylized COCO dataset was obtained. This method solved the problem of insufficient training samples of actual seismic data and provided a feasible solution that could expand the application of artificial intelligence methods in interpreting seismic data. The COCO dataset was uniform in format, rich in content and variety, and suitable for training convolutional neural networks. The results of the seismic profiling style conversion based on the COCO dataset showed that features such as overall contour information and texture detail can be preserved better based on the selection of a suitable migration algorithm. Because the style migration method was based on the computer field of vision, it had a strong similarity to most of the fault interpretations carried out by seismic profiling. Similarly, this method is applicable to the interpretation of special geological phenomena, such as seismic facies or diapir identification, which are visually identified. After the style transfer of the COCO dataset was completed, the semantic segmentation model was trained for the COCO dataset with a stylized seismic profile. The training label maintained the label of the original COCO dataset. After training was completed, the front weight of the semantic segmentation model was frozen, and only the last two weights were retained as the training state. The model structure of the last two layers was modified, and the number of channels in the last layer was changed from 81 to 2. Next, the dataset was replaced by the fault dataset of the target work area from the stylized COCO dataset to reduce the learning rate and to retrain, meaning that only the weights of the last two layers were fine-tuned. After the loss value converged, the fault recognition model of the target work area based on the stylized COCO dataset was obtained. This method solved the problem of insufficient training samples of actual seismic data and provided a feasible solution that could expand the application of artificial intelligence methods in interpreting seismic data.

Conclusions
In this paper, we addressed the lack of sufficient training samples for artificial intelligence methods in seismic data interpretation applications, and proposed the use of style migration to convert the image recognition COCO dataset into a seismic profile format. We drew three main conclusions.

1.
Using a suitable style migration algorithm, the COCO dataset was well migrated into the seismic profile format, which retained the overall contour information and features such as texture details more completely. This result can be used to increase artificial intelligence seismic interpretation training samples and realize the application of artificial intelligence to visually discernable seismic geological interpretation tasks such as fault interpretation, seismic facies interpretation, and special morphological geological body interpretation.

2.
By comparing the effects of multiple style migration models, the Laplace pyramid network was preferred for carrying out seismic profile stylization research. It transmitted low-resolution global style patterns through the drafting network, corrected high-resolution local details through the correction network, and aggregated all pyramid layers to output a final stylized seismic profile image. Experiments showed that this method ensured that the overall style pattern was effectively transmitted without losing the original image content. It provided theoretical guidance for fast and objective application of the model to learn seismic data features. 3.
The quality of stylized images was objectively evaluated on two indexes, SSIM and Q AB/F and the efficiency of the migration method was evaluated for algorithm running time. Thus, a quantitative evaluation of the good or bad quality of stylized images was achieved.
However, this study was limited to an analysis of the effect of stylized conversion, so further work and discussion are needed before implementing artificial intelligence algorithms combined with the training set of actual seismic interpretation data.