You are currently viewing a new version of our website. To view the old version click .
Entropy
  • Article
  • Open Access

27 March 2025

AI Painting Effect Evaluation of Artistic Improvement with Cross-Entropy and Attention

,
,
and
1
Culture Design Lab, Graduate School of Techno Design, Kookmin University, Seoul 02707, Republic of Korea
2
Department of Global Convergence, Kangwon National University, Chuncheon-si 24341, Republic of Korea
3
Department of Smart Experience Design, Kookmin University, Seoul 02707, Republic of Korea
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Entropy Based Machine Learning Models

Abstract

With the rapid development of AI technology, AI painting tools are increasingly used in art creation. However, the effects of works created by different users using AI painting tools vary. Finding out the factors that affect the level of art creation after users use AI painting tools is a matter of concern. To solve this problem, this paper proposes a new Multi-Classification Attention Support Vector Machine (MCASVM) with cross-entropy loss function. By identifying and predicting the level of creativity of ordinary users after using AI painting tools, this model compares and analyzes the influencing factors behind the high and low effects of artistic creativity enhancement after using AI painting tools. The main contribution of this paper is to establish the Art Creation Ability Assessment Dataset (ACAAD) through real data collection to provide data support for subsequent assessments. Meanwhile, MCASVM directly handles the multi-classification problem in the established dataset by introducing multiple SVMs. Among other things, the probabilistic calibration network adjusts the model output so that its predicted probabilities are closer to the probability that the sample truly belongs to the classification. DBAM enhances the feature fusion capability of the model by explicitly focusing on the important channel and spatial features, and it enables the model to more accurately recognize and differentiate between changes in the creative abilities of different users before and after using AI painting tools. The experimental results show that the artistic creativity of ordinary users can be enhanced by AI painting tools, where the most central influencing factors are interest level and social support.

1. Introduction

With the rapid advancement of AI technology, particularly breakthroughs in computer vision, entropy, and information theory [1,2], the application of AI in artistic creation has become an important topic of academic and public interest. In recent years, the widespread adoption of AI painting tools, especially among ordinary users, has sparked extensive academic discussions and practical explorations regarding their potential to enhance artistic creativity. These tools not only provide a fast and convenient way to create art but also help users overcome creative bottlenecks and inspire inspiration. AI painting tools such as DALL-E, Stable Diffusion, and MidJourney utilize deep learning algorithms to generate high-quality images. These tools can analyze and understand a large number of artworks, learn styles and techniques, and subsequently incorporate these elements into new creations. This ability makes AI painting tools powerful assistants for artists exploring new styles and techniques. Additionally, AI painting tools are highly customizable, allowing users to adjust creative parameters according to their needs and, thus, create unique works of art. This convenience has attracted significant attention from creators, providing artists with new avenues for inspiration and creative exploration. These features collectively promote the integration of AI and artistic creation, opening up new possibilities for the development of art.
Nevertheless, despite the growing body of research in this area, questions remain about how to systematically evaluate the impact of AI painting tools on users’ artistic creativity and the underlying mechanisms through which these impacts occur [3]. The enhancement of artistic creativity is influenced by a complex interplay of factors, including individual differences, the context of usage, and specific creative objectives [4]. The traditional evaluation of artistic creation ability mostly relies on expert reviews or subjective evaluations and lacks systematic quantitative analysis. Therefore, the goal of this paper is to establish a more objective and accurate deep learning evaluation model to evaluate the improvement effect of ordinary users’ artistic creation abilities after using AI painting tools [5]. Specifically, this paper intends to solve the following two key problems.
Question 1 is, how do we assess the impact of AI drawing tools on improving the artistic creativity of ordinary users? We used a quantitative assessment to determine the changes in ordinary users’ ability to create art before and after using the AI drawing tool. Then, we analyzed the responses of different user groups to the AI drawing tool [6]. Question 2 is, what factors influence the level of users’ art creation ability? First, the factors that may affect the effect of art creation, such as users’ basic characteristics (e.g., age, gender, educational background, etc.), motivation to create, creative environment, and social support, were discussed. And based on this, a dataset was established. Finally, through example validation, we explored how these variables affected users’ creative effects after using AI drawing tools [7].
To solve the above two problems, this paper created the Art Creation Ability Assessment Dataset (ACAAD) based on real data and proposed the Multi-Classification Attention Support Vector Machine (MCASVM). MCASVM creatively applies machine learning methods, especially SVM and attention mechanism, to the assessment of artistic creation ability. In the traditional assessment of artistic creativity ability, the factors affecting artistic creativity ability are often not paid attention to [8]. By introducing machine learning algorithms, this study provides more objective and accurate evaluation results, which help to understand the specific impact of AI painting tools on the improvement of artistic creation ability [9]. At the same time, this paper combines multidimensional influencing factors for analysis. These factors include not only the personal characteristics of users but also external factors such as creative environment and social support, reflecting a multidimensional understanding of artistic creation ability [10]. This research perspective helps to deeply explore the differences in the creative performance of different types of users when using AI painting tools, thereby providing a theoretical basis for the optimization and personalized customization of AI painting tools. The main contributions of this paper are mainly reflected in the following three aspects:
(1)
By collecting information from real-life AI painting users, this paper creates a multidimensional Art Creation Ability Assessment Dataset (ACAAD), which provides a basis for the subsequent identification and prediction of factors behind the enhancement of artistic creativity of ordinary users;
(2)
A Multi-Classification Attention Support Vector Machine (MCASVM) with a cross-entropy loss function is proposed. It innovatively introduces multiple SVMs to deal with multiclassification problems directly and combines the attention mechanism of deep learning with the traditional SVM, utilizing the neural network structure to realize the “one vs. rest” strategy of SVM;
(3)
The novel dual-branch attention module (DBAM) aims to fully integrate features of different dimensions. It consists of two branches: the Channel Attention Module and the Spatial Attention Module. DBAM is able to capture the important features in the input data more comprehensively and, thus, improve the model’s classification performance on the multi-factors affecting artistic creativity.

3. Method

To address the enhancement effect of AI painting tools on the artistic creation ability of ordinary users and the corresponding influencing factors, we propose a Multi-Classification Attention Support Vector Machine (MCASVM). Its overall structure is shown in Figure 1. The initial input of this model is the Art Creation Ability Assessment Dataset, the attributive character of the core information book. These data are then passed through the convolutional neural network to output the feature vector ϕ(x). Next, the DBAM module performs feature fusion after transforming the feature vector’s dimensionality ϕ(x). Finally, the classification results are obtained by a Probabilistic Calibration Network.
Figure 1. The structure diagram of Multi-Classification Attention Support Vector Machine (MCASVM).

3.1. Overall Structure of MCASVM

Multi-Classification Attention Support Vector Machine is a deep learning framework model based on SVM [43], which can directly deal with multi-classification problems. As shown in Figure 1, feature extraction is carried out through hidden layers. Multiple SVMSs are used to replace the output layer of the network to complete multiple classification tasks, and each neuron is used to solve a hyperplane. The task of each hyperplane is to divide a class of samples from the sample space. In the output layer of the model, the functional distance of the output of the k th neuron is defined as Equation (1).
d k x ; θ k = w k T φ x + b k
where the output d k of the model represents the functional distance between the sample and each hyperplane, and the parameters of the k hyperplane are represented by w k T and b k , and θ k ( , w k , b k ) is used to represent all parameters of the model to be solved uniformly. The task of each of these neurons is to find a hyperplane [44] that can be used to separate one class of samples from the others. To accomplish the above objectives, the loss function is defined as Equation (2).
L M C A S V M = k = 1 K 1 2 w k 2 + C i = 1 N m a x 1 I k y i d k x i , 0
Among them,
I k y i = 1 , y i = k 1 , y i k
Similar to the loss function of SVM, MCASVM is equivalent to adding the generation of multiple hyperplanes based on SVM and finally achieves the purpose of multi-classification by calculating the summation term of the loss of a class of samples divided by each hyperplane. The model combines the “one vs. rest” strategy [45] and the idea of a support vector machine to find the maximum interval hyperplane. It is then implemented through the structure of a neural network, which fully utilizes the advantages of these methods.
If a new sample is given, the category of the new sample is determined by the distance from the sample to the hyperplane. The specific process is shown in Equation (4).
y ^ = m a x k d 1 x i , d 2 x i , , d k x i
The decision method is similar to the cross-entropy loss function classification [46], which takes the maximum value of network output as the category of sample prediction. Since the task of each hyperplane in the model is to separate one class of samples from the other classes of samples, the test samples should be in the positive class of one class of hyperplanes with a positive functional distance from that class of hyperplanes. And in the negative class of other classes of hyperplanes, the distance from the function of other classes of hyperplanes is negative. Therefore, the method is very simple to determine the category of a new sample point. It is only necessary to look at the functional distance between the sample point and each hyperplane and select the largest value to find the category of the sample.
However, unlike classification using the cross-entropy loss function, this model outputs a functional distance rather than a probability, and it is not possible to intuitively compare the likelihood of samples belonging to different classes from the model output. Considering further optimization of the model output, a Probabilistic Calibration Network is proposed to improve this problem.

3.2. Probabilistic Calibration Network

This section proposes probabilistic calibration networks to handle the output of the classification model. Probability calibration is proposed to optimize the model output so that the sample’s probability predicted by the model is as close as possible to the probability that the sample belongs to this class. We first introduce the goal of probabilistic calibration, i.e., the perfect calibration. Suppose the output of the network is p ^ x = p ^ 1 x , , p ^ k x , p ^ [ 0,1 ] , we take p ^ = m a x p ^ x as a model to forecast probability sample, the corresponding forecast category is Y ^ = y ^ i . The following formula defines the perfect calibration as shown in Equation (5).
P Y ^ = y | P ^ = p x = p x
where y represents the true class of the sample, and p represents the true probability that the sample belongs to that class. For example, if the model predicts that each of the 10 samples has an 80% probability of being in the positive class, then there should be 8 samples that can be classified correctly. If the output of the model to be calibrated is not a probability, it is converted to a probability output by Equation (6).
p ^ i = σ s i k = e x p s i k j = 1 K e x p   s i j
where s i represents the prediction score of the model output, and σ · represents the processing of the model output. In short, probabilistic calibration is obtained by using the output and label of the model s i ,   p ^ i ,   y i , and calculating the probabilistic calibrated output q ^ i . Calculate the calibrated output q ^ i .
Platt [47] scales by looking for the scalars a , b to scale the output, as shown in Equation (7).
p ^ i = σ a s i + b k
Temperature calibration [48] is achieved by looking for the scalar parameter T to scale the output, as shown in Equation (8).
p ^ i = σ s i / T k
After the calibrated probability output is obtained, the calibrated probability is obtained by the maximum value of the output after scaling, as shown in Equation (9).
q i = m a x p ^ i
It can be seen that the probabilistic calibration method does not change the predicted results of the model and outputs with large values are scaled by probabilistic calibration. In addition, probabilistic calibration can reduce the instances where the original model outputs are over-confident or under-confident by scaling the output in this way.
The implementation of the probabilistic calibration network is accomplished by using a single 1 × 1 convolution kernel in the convolutional layer, the structure of which is shown in the figure below. If the convolution kernel is temperature-calibrated, only one variable is set. If the convolution kernel is calibrated by Platt scaling, another offset item needs to be added to the convolution kernel. The convolution kernel is used to scale the model output by moving the convolution kernel, which is the concept of shared weights in a convolutional neural network.
As shown in Figure 2, before calibrating the output of the model, the model prediction sample belongs to the fourth category. After a convolution kernel calibration, the sample is also judged to belong to the fourth class. The probabilistic calibration network does not change the prediction results of the model but enhances the reliability of the model’s own judgment.
Figure 2. Implementation of probabilistic calibration.
All in all, the probabilistic model is completed by only one convolutional layer, which is used to complete the parameter calculation of the probabilistic calibration model. That is, the probabilistic calibration network essentially extends the probabilistic calibration methods of Platt scaling and temperature scaling into the deep learning framework. Unlike the original calibration model, the parameters are computed through the network to avoid memory overflow errors when dealing with large sample data.

3.3. Double-Branch Attention Module

After the features were obtained by the multi-layer perceptron, this paper designed a Double-Branch Attention Module to enhance the feature fusion capability of the model and improve the classification performance of the model. Its specific structure is shown in Figure 3. DBAM consists of two branches: the Channel Attention Module and the Spatial Attention Module.
Figure 3. DBAM structure diagram.
In the above branch channel, the input feature maps of the attention module are, firstly, MaxPool and AvgPool operations and two different groups of feature maps are obtained, respectively. Then, the two sets of feature maps are passed into a Shared multi-layer perceptron (Shared MLP) for feature fusion, and the two feature maps are output. Next, the two feature graphs are multiplied by an element-by-element operation to produce a Channel Attention M c .
The branch space attention module below receives the Channel-refined feature F (channel-refined feature F ) from the channel-refined Feature module, which is further refined through MaxPool and AvgPool. These processed feature maps are then used for spatial feature extraction through a Conv layer. Finally, the feature maps output by the convolutional layer are generated by an activation function to generate Spatial Attention M s . The channel attention feature map ( M c ) is combined with the spatial attention feature map ( M s ), and the final feature map is obtained by element multiplication. In essence, DBAM improves the performance and accuracy of the overall model by explicitly focusing on important channel and spatial features.
In this paper, the loss function was selected as the cross-entropy loss function combined to optimize the parameters of the model, which is shown in Equation (10).
L = 1 N   i = 1 N c = 1 k y i c   l o g p i c  
where N is the number of samples and k is the number of categories. y i c equals 1 when the category of the ith sample is c ; otherwise, y i c equals 0. p i c is the predicted probability that the category of the ith sample is c .

3.4. Algorithm Implementation

The MCASVM implementation in Algorithm 1 completes the generation of hyperplanes by modifying the loss function. First, our model extracts features through a multi-layer perceptron (MLP) and then processes the features using maximum pooling and average pooling operations. After the processing, a shared MLP and Sigmoid function are used to generate an attention mechanism for adjusting the features. Immediately after that, the model performs another pooling operation on the tuned features and generates the final attention weights using convolution and Sigmoid functions. Finally, the features are weighted using these weights and classified by the Softmax function to output the final prediction. The whole process also includes the control of counters to ensure that the appropriate operations are performed at different steps.
Algorithm 1: Pseudocode of DBAM-MCASVM-PCN Model.
Input: Input features x 1 ,   x 2 ,   ,   x n
Output: Classified result y
Residual = ; counter =
layers = [MLP Feature Extraction, DBAM Module, MCASVM Classifier, Probabilistic Calibration Network]
1: While counter < len(layers) ∗ 2 + 1:
2:      If counter < len(layers):
3:         If counter == ∅:
4:          layer = MLP Feature Extraction
5:         ElseIf counter == 1:
6:           F m a x = M a x p o o l ( F )
7:           F a v g = A v g p o o l ( F )
8:           M L P F m a x = S h a r e M L P ( F m a x )
9:           M L P F a v g = S h a r e M L P ( F a v g )
10:           M c = S i g m o i d ( M L P F m a x M L P F a v g )
11:           F = M c F
12:           F m a x = M a x P o o l ( F )
13:           F a v g = A v g P o o l ( F )
14:           M s = S i g m o i d ( C o n v ( F m a x + F a v g ) )
15:           F f i n a l = M s M c F
16:       ElseIf counter == 2:
17:           d = []
18:          For each class k in {1, 2, …, K}:
19:               d k =   w k F f i n a l + b k
20:              Append  d k to d
21:       ElseIf counter == 3:
22:          p = Softmax(d)
23:          y = argmax(p)
24:           EndIf
25:           x = layer(x)
26:       Else:
27:            If counter == len(layers):
28:              layer = MCASVM Classifier
29:            ElseIf counter == len(layers) + 1:
30:              layer = Probabilistic Calibration Network
31:            EndIf
32:            x = layer(x)
33:          counter = counter + 1
34: y = Output Layer(x)
35: Return y
Through the backpropagation algorithm of the neural network, the multi-classification task on one network is completed directly. The difference between this method and the cross-entropy loss classification is discussed in detail here. When the loss function L M C A S V M calculates the gradient, the contribution of different samples to the parameter update is different. For samples I k y i d k x i 1 , that is, samples that do not need to be punished in the SVM method idea, the gradient is shown in Equation (11).
L w k k = 1 K w k L b k = 0
For those samples, I k y i d k x i < 1 , the samples in the penalty term should be calculated in the original SVM method, and the gradient is shown in Equation (12).
L w k k = 1 K w k i = 1 N C I k y i φ x i L b k = k = 1 K i = 1 N C I k y i
By comparing Equation (5) and Equation (6), when calculating the gradient, samples in different positions have different influences on the model loss function. The model should focus not only on the sample points that need to be trained but also on those samples that are more likely to affect the decision hyperplane. This property enhances the prediction ability and generalization performance of the model.

4. Experiment and Analysis

4.1. Dataset Introduction

To investigate the impact of AI painting tools on the artistic creation abilities of ordinary users and the corresponding influencing factors, this paper presents the Art Creation Ability Assessment Dataset (ACAAD), which was compiled by sampling from various demographic groups. It primarily describes the characteristics and effects associated with different AI painting users. The main structural information is summarized in Table 1, where the training set comprises 10,000 users, and the test set includes 2000 users. According to the factors affecting artistic creativity mentioned in Section 2.3, it can be known that it includes the following main factors: the Chinese and Western cultures to which the artist is subjected, personality, creative motivation, emotional positivity, social environment, intellectual and cognitive ability, educational experience and background, and external evaluation. And our dataset is divided into six dimensions. Among them, degree generally corresponds to educational background and intellectual and cognitive abilities; artistic motivation corresponds to the motivation to create art; interest level represents the creator’s emotional motivation to create art, and social support corresponds to the social environment as well as external evaluation.
Table 1. Characteristics and statistics of ACAAD.
By selecting different features and coding independent variables, we assessed the impact of the AI drawing tool on the users’ artistic abilities and the factors that influenced them. First, we collected the paintings created by these 12,000 users three months before and after using the AI painting tool. Further, we asked a professional painting teacher to rate the paintings they created based on their color scheme, line usage, compositional layout, and expression of mood, with a total score of 100. When the score of the paintings created before and after three months increased by more than 15 points, they were categorized as having a significant impact. If the improvement in score is between 0 and 15 points, it is classified as a moderate impact. If the score does not increase but decreases, it is classified as a negative impact.
In order to validate the generalization performance of the model and to compare it with other state-of-the-art models. In this paper, experiments are also conducted on another publicly available art assessment dataset. Boldbrush Artistic Image Dataset (BAID) [49] is a large dataset for aesthetic assessment of art images. The dataset contains 60,337 artworks encompassing a variety of art forms, such as paintings and drawings, and has collected over 360,000 votes from online users. These votes reflect people’s aesthetic preferences for artworks, with a higher number of votes indicating that the work is considered more aesthetically pleasing. The construction of the dataset is based on a monthly art contest on the Boldbrush website, in which participating artists upload their works, and users can vote for their favorite works, and this voting mechanism makes the ratings in the dataset highly credible and authoritative.

4.2. Comparison of Experimental Results

The ACAAD proposed in this paper is characterized by a small data volume and dimensionality. In its classification approach, the depth model employs a fully connected neural network structure. By comparing the hidden layer designs of a one-layer fully connected network and a two-layer fully connected network, we explore the classification accuracy of the traditional Deep Neural Networks (DNN) [50] method and the MCASVM. The specific experimental results are presented in Table 2 and Table 3. As shown in Table 2 and Table 3, the MCASVM achieved superior results on the ACAAD.
Table 2. Classification accuracy on ACAAD dataset.
Table 3. Comparison between the cross-entropy loss function and L1 loss function.
After that, we used MCASVM to classify ACAAD, a multidimensional dataset, based on these three results. The classification accuracy of the model was first verified. Then, from the classification results, we observe and count the percentage of users of different dimensions in the three sections: reduce, no difference, and rise. It is known that there are three results in artistic creation ability after using AI painting tools: significant improvement, moderate improvement, and negative improvement. We will record the results with significant improvement effects in Table 4 as “rise”, the results with mild improvement effects as “no difference”, and the results with negative improvement effects as “reduce”. From there, we can determine which factors will enhance the art creation ability after using AI painting tools.
Table 4. Comparison results of feature defaults.
The experimental result in Table 4 indicates the proportions of consistent, decreased, and increased effects on artistic creation ability improvement before and after using the tools. Notably, the two factors of interest level and community support significantly enhance the improvement in artistic creation ability. Therefore, we conclude that AI painting tools can substantially enhance the artistic creation abilities of ordinary users, with interest level and community support identified as the core influencing factors.
To better validate the generalization performance of MCASVM, we verified the accuracy of the model’s classification recognition on another art evaluation dataset, BAID. We also compared the experimental results with two other state-of-the-art models, SAAN [49] and TANet [51], on this dataset. As shown in Table 5, SAAN has the highest classification accuracy of 0.7680 on the BAID dataset. DNN, as well as traditional SVM, have poor classification accuracy. The classification accuracy of our proposed model, MCASVM, although a little lower than SAAN, still exceeds TANet and the other two models. This is sufficient to show the effectiveness and generalization performance of our proposed model.
Table 5. Classification accuracy on BAID dataset.

4.3. Visualization of the Model

In the visualization component of the model, the sklearn library [52] was used to generate spiral data for comparison, thus demonstrating the classification effect of more complex datasets.
As shown in Figure 4, MCASVM generates three hyperplanes based on three classes of datasets, where each class of hyperplanes separates one class of sample points from the others in the sample space. To illustrate the classification effects more clearly, the decision hyperplane and classification regions were plotted using the decision function Equation (4).
Figure 4. MCASVM generates hyperplane diagram.
As shown in Figure 5, the decision hyperplane divides the graph into three parts, with each part containing a single class of samples. Obviously, the hyperplane generated by MCASVM in the output layer accomplishes the classification task and enhances the interpretability of the model. In comparison with DNNs that only produce output results, the hyperplanes represented by the output layer of MCASVM are more beneficial for decision-making. As previously mentioned, MCASVM and the traditional DNN can be realized by modifying the loss function model under the same structure. In the next section, the location of the decision hyperplane will be demonstrated after using the cross-entropy loss function to classify the dataset under the same hyperparameters and network structure settings.
Figure 5. MCASVM divides the sample space diagram.
As illustrated in Figure 6, the classification regions were compared by inputting the coordinates of the 2D image into the trained DNN model. The DNN also predicted the categories to which the different coordinate points belonged, which exhibited a high degree of similarity to the classification regions generated by MCASVM. However, a discrepancy was observed in the decision region for categorizing the spirals. It is conceivable that if the three spirals were extended, the decision space of MCASVM would accommodate a greater number of correctly categorized spiral sample points.
Figure 6. DNN Decision Region diagram.

4.4. Probabilistically Calibrated Elimination Experiments

The effectiveness of the calibrated model can typically be visualized using reliability plots. In a reliability plot, the horizontal axis represents the intervals of the model’s prediction scores, dividing the model’s predictions into M intervals, with   B m denoting the average level of prediction of the categorized model in the m th interval. For instance, the probability of [0, 0.1] can be assumed to be equally divided into 10 intervals. The initial interval corresponds to the set of samples with model output probability at [0, 0.1]. The mean prediction level of the model across the various intervals is depicted by Equation (13).
c o n f B m = 1 B m i B m p ^ i
The vertical axis of the reliability plot is indicative of the percentage of samples in these intervals that the model predicted correctly, as depicted in Equation (14).
a c c B m = 1 B m i B m 1 y ^ i = y i
The indicator function, denoted by 𝟙 · , takes on a value of 1 when the model’s prediction of y aligns with the true label, and a value of 0 otherwise. The discrepancy between the confidence level ascribed by the model and the true accuracy of the two is quantified by the expected calibration error (ECE), as depicted in Equation (15).
E C E = m = 1 M B m n a c c B m c o n f B m
Reliability plots are generated before model calibration, subsequent to Platt scaling, and following temperature scaling calibration, respectively. The uncalibrated plot of the reliability of the model output is shown in Figure 7. The blue regions in these plots denote the true classification accuracy across the various intervals predicted by the model. The pink areas, along with the shaded regions that overlap with them, signify the discrepancy from perfect calibration.
Figure 7. MCASVM uncalibrated outputs.
Following the calibration process, Platt scaling attains superior outcomes, thereby reducing the discrepancy between the optimal calibration and the ideal calibration. As shown in Figure 8 below, Platt scaling effectively adjusts the model’s output, ensuring that the output closely resembles the probability that the samples belong to a specific category.
Figure 8. Reliability plot for Platt scaling and temperature calibration.

5. Conclusions

With the development of artificial intelligence technology, AI painting tools have become an important auxiliary way for ordinary users to create art. However, it is still a challenge to systematically and objectively predict and identify the effect of the improvement of users’ art creation ability after using AI painting tools and its influencing factors. This paper explores the effects of AI painting tools on the artistic creation ability of ordinary users and the corresponding influencing factors by constructing MCASVM with a cross-entropy loss function. The MCASVM utilizes the structure of a neural network to implement the “one vs. rest” strategy of SVM, which gives full play to the advantages of deep learning models in feature extraction and complex pattern recognition while inheriting the characteristics of SVM that perform well on small sample data sets. In addition, the cross-loss function achieves the best performance by comparing different loss functions. There are many influencing factors that improve the creative ability of user groups, and the main ones include the basic characteristics of users, creative motivation, interest level, and social support. The experimental results show that AI painting tools have significantly improved the artistic creation ability of ordinary users, and the most core influencing factors are interest level and community support.

Author Contributions

Conceptualization, T.Y. and Y.T.; methodology, Y.T. and S.L.; software, Z.C.; validation, Y.T., S.L. and Z.C.; formal analysis, S.L.; investigation, Z.C.; writing—original draft, Y.T.; writing—review and editing, Y.T., S.L. and Z.C.; visualization, Y.T.; supervision, T.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

All data generated or analyzed during this study are included in this article. The raw data are available from the corresponding author upon reasonable request.

Acknowledgments

The authors thank all the participants in this study for their time and willingness to share their experiences and feelings.

Conflicts of Interest

The authors declare no conflicts of interest concerning the research, authorship, and publication of this article.

References

  1. Chen, Z. Graph Adaptive Attention Network with Cross-Entropy. Entropy 2024, 26, 576. [Google Scholar] [CrossRef]
  2. Chen, Z. (HTBNet) Arbitrary Shape Scene Text Detection with Binarization of Hyperbolic Tangent and Cross-Entropy. Entropy 2024, 26, 560. [Google Scholar] [CrossRef]
  3. Li, J.; Zhong, J.; Liu, S.; Fan, X. Opportunities and Challenges in AI Painting: The Game between Artificial Intelligence and Humanity. J. Big Data Comput. 2024, 2, 44–49. [Google Scholar] [CrossRef]
  4. Li, J.; Zhang, B. The Application of Artificial Intelligence Technology in Art Teaching Taking Architectural Painting as an Example. Comput. Intell. Neurosci. 2022, 2022, 8803957. [Google Scholar] [CrossRef]
  5. Li, L. The Impact of Artificial Intelligence Painting on Contemporary Art from Disco Diffusion’s Painting Creation Experiment. In Proceedings of the 2022 International Conference on Frontiers of Artificial Intelligence and Machine Learning (FAIML), Hangzhou, China, 19–21 June 2022. [Google Scholar] [CrossRef]
  6. Liu, X. Artistic Reflection on Artificial Intelligence Digital Painting. J. Phys. Conf. Ser. 2020, 1648, 032125. [Google Scholar] [CrossRef]
  7. Rani, S.; Jining, D.; Shah, D.; Xaba, S.; Singh, P.R. The Role of Artificial Intelligence in Art: A Comprehensive Review of a Generative Adversarial Network Portrait Painting. In Proceedings of the International Conference on Intelligent Computing & Optimization, Phnom Penh, Cambodia, 26–27 October 2023; Lecture Notes in Networks and Systems. pp. 126–135. [Google Scholar] [CrossRef]
  8. Wang, J.; Yuan, X.; Hu, S.; Lu, Z. AI Paintings vs. Human Paintings? Deciphering Public Interactions and Perceptions towards AI-Generated Paintings on TikTok. arXiv 2024, arXiv:2409.11911. [Google Scholar]
  9. Xu, J.; Zhang, X.; Li, H.; Yoo, C.; Pan, Y. Is Everyone an Artist? A Study on User Experience of AI-Based Painting System. Appl. Sci. 2023, 13, 6496. [Google Scholar] [CrossRef]
  10. Xu, X. A Fuzzy Control Algorithm Based on Artificial Intelligence for the Fusion of Traditional Chinese Painting and AI Painting. Sci. Rep. 2024, 14, 17846. [Google Scholar] [CrossRef]
  11. Tian, Y.; Shi, Y.; Liu, X. Recent advances on support vector machines research. Technol. Econ. Dev. Econ. 2012, 18, 5–33. [Google Scholar] [CrossRef]
  12. Sebald, D.J.; Bucklew, J.A. Support vector machine techniques for nonlinear equalization. IEEE Trans. Signal Process. 2000, 48, 3217–3226. [Google Scholar]
  13. Tang, Y. Deep learning using support vector machines. CoRR 2013, 2, abs/1306.0239. [Google Scholar]
  14. Zhang, L.; Zhou, W.; Jiao, L. Wavelet support vector machine. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 2004, 34, 34–39. [Google Scholar]
  15. Bennett, K.P.; Blue, J.A. A support vector machine approach to decision trees. In Proceedings of the 1998 IEEE International Joint Conference on Neural Networks Proceedings. IEEE World Congress on Computational Intelligence (Cat. No. 98CH36227), Anchorage, AK, USA, 4–9 May 1998; IEEE: Piscataway, NJ, USA, 1998; Volume 3. [Google Scholar]
  16. Lin, K.M.; Lin, C.J. A study on reduced support vector machines. IEEE Trans. Neural Netw. 2003, 14, 1449–1459. [Google Scholar]
  17. Mangasarian, O.L.; Musicant, D.R. Successive overrelaxation for support vector machines. IEEE Trans. Neural Netw. 1999, 10, 1032–1037. [Google Scholar] [PubMed]
  18. Osuna, E.; Freund, R.; Girosi, F. An improved training algorithm for support vector machines. Neural networks for signal processing VII. In Proceedings of the 1997 IEEE Signal Processing Society Workshop, Princeton, NJ, USA, 23–25 June 1997; IEEE: Piscataway, NJ, USA, 1997. [Google Scholar]
  19. Suthaharan, S.; Suthaharan, S. Support vector machine. In Machine Learning Models and Algorithms for Big Data Classification: Thinking with Examples for Effective Learning; Springer: Boston, MA, USA, 2016; pp. 207–235. [Google Scholar]
  20. Hafiz, A.M.; Parah, S.A.; Bhat, R.U.A. Attention mechanisms and deep learning for machine vision: A survey of the state of the art. arXiv 2021, arXiv:2106.07550. [Google Scholar]
  21. Niu, Z.; Zhong, G.; Yu, H. A review on the attention mechanism of deep learning. Neurocomputing 2021, 452, 48–62. [Google Scholar] [CrossRef]
  22. Brauwers, G.; Frasincar, F. A general survey on attention mechanisms in deep learning. IEEE Trans. Knowl. Data Eng. 2021, 35, 3279–3298. [Google Scholar]
  23. Rodriguez, P.; Velazquez, D.; Cucurull, G.; Gonfaus, J.M.; Roca, F.X.; Gonzalez, J. Pay attention to the activations: A modular attention mechanism for fine-grained image recognition. IEEE Trans. Multimed. 2019, 22, 502–514. [Google Scholar] [CrossRef]
  24. Liu, Y.; Shao, Z.; Hoffmann, N. Global attention mechanism: Retain information to enhance channel-spatial interactions. arXiv 2021, arXiv:2112.05561. [Google Scholar]
  25. Cai, W.; Wei, Z. Remote sensing image classification based on a cross-attention mechanism and graph convolution. IEEE Geosci. Remote Sens. Lett. 2020, 19, 1–5. [Google Scholar]
  26. Liu, S.; Lin, T.; He, D.; Li, F.; Wang, M.; Li, X.; Sun, Z.; Li, Q.; Ding, E. Adaattn: Revisit attention mechanism in arbitrary neural style transfer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021. [Google Scholar]
  27. Li, Y.; Zeng, J.; Shan, S.; Chen, X. Occlusion aware facial expression recognition using CNN with attention mechanism. IEEE Trans. Image Process. 2018, 28, 2439–2450. [Google Scholar]
  28. Huang, W.; Zhou, F. DA-CapsNet: Dual attention mechanism capsule network. Sci. Rep. 2020, 10, 11383. [Google Scholar]
  29. Zhu, X.; Cheng, D.; Zhang, Z.; Lin, S.; Dai, J. An empirical study of spatial attention mechanisms in deep networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  30. DeRose, J.F.; Wang, J.; Berger, M. Attention flows: Analyzing and comparing attention mechanisms in language models. IEEE Trans. Vis. Comput. Graph. 2020, 27, 1160–1170. [Google Scholar]
  31. Zhao, Q.; Liu, J.; Li, Y.; Zhang, H. Semantic segmentation with attention mechanism for remote sensing images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–13. [Google Scholar] [CrossRef]
  32. Niu, W.; Sternberg, R.J. Sternberg. Cultural influences on artistic creativity and its evaluation. Int. J. Psychol. 2001, 36, 225–241. [Google Scholar]
  33. Kaufman, S.B.; Quilty, L.C.; Grazioplene, R.G.; Hirsh, J.B.; Gray, J.R.; Peterson, J.B.; DeYoung, C.G. Openness to experience and intellect differentially predict creative achievement in the arts and sciences. J. Personal. 2016, 84, 248–258. [Google Scholar]
  34. Joy, S.P. Innovation motivation and artistic creativity. J. Creat. Behav. 2005, 39, 35–56. [Google Scholar]
  35. Asquith, S.L.; Wang, X.; Quintana, D.S.; Abraham, A. Predictors of Change in Creative Thinking Abilities in Young People: A Longitudinal Study. J. Creat. Behav. 2024, 58, 262–278. [Google Scholar]
  36. Anderson, R.C.; Haney, M.; Pitts, C.; Porter, L.; Bousselot, T. “Mistakes can be beautiful”: Creative engagement in arts integration for early adolescent learners. J. Creat. Behav. 2020, 54, 662–675. [Google Scholar]
  37. Mace, M.A.; Ward, T. Modeling the creative process: A grounded theory analysis of creativity in the domain of art making. Creat. Res. J. 2002, 14, 179–192. [Google Scholar]
  38. Amabile, T.M. Effects of external evaluation on artistic creativity. J. Personal. Soc. Psychol. 1979, 37, 221. [Google Scholar] [CrossRef]
  39. Hu, W.; Wang, X.; Yi, L.Y.X.; Runco, M.A. Creative self-efficacy as moderator of the influence of evaluation on artistic creativity. Int. J. Creat. Probl. Solving 2018, 28, 39–55. [Google Scholar]
  40. Xue, Y.; Gu, C.; Wu, J.; Dai, D.Y.; Mu, X.; Zhou, Z. The effects of extrinsic motivation on scientific and artistic creativity among middle school students. J. Creat. Behav. 2020, 54, 37–50. [Google Scholar] [CrossRef]
  41. Rostan, S.M. A study of young artists: The development of artistic talent and creativity. Creat. Res. J. 1997, 10, 175–192. [Google Scholar] [PubMed]
  42. Ershadi, M.; Winner, E. Children’s creativity. Encycl. Creat. 2020, 1, 144–147. [Google Scholar]
  43. Jakkula, V. Tutorial on Support Vector Machine (svm); School of EECS, Washington State University: Pullman, WA, USA, 2006; Volume 37, p. 3. [Google Scholar]
  44. Stanley, R.P. An introduction to hyperplane arrangements. Geom. Comb. 2007, 13, 389–496. [Google Scholar]
  45. Bassano, C.; De Matteis, G.M.; Nardi, P.; Buratta, M.M.; Zeitani, J.; De Paulis, R.; Chiariello, L. Mid-term follow-up of aortic root remodelling compared to Bentall operation. Eur. J. Cardio-Thorac. Surg. 2001, 19, 601–605. [Google Scholar] [CrossRef]
  46. Mao, A.; Mohri, M.; Zhong, Y. Cross-entropy loss functions: Theoretical analysis and applications. In Proceedings of the International conference on Machine Learning, PMLR, Honolulu, HI, USA, 23–29 July 2023. [Google Scholar]
  47. Platt, J. Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods. Adv. Large Margin Classif. 1999, 10, 61–74. [Google Scholar]
  48. Guo, C.; Pleiss, G.; Sun, Y.; Weinberger, K.Q. On calibration of modern neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017. [Google Scholar]
  49. Yi, R.; Tian, H.; Gu, Z.; Lai, Y.K.; Rosin, P.L. Towards artistic image aesthetics assessment: A large-scale dataset and a new method. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023. [Google Scholar]
  50. Sze, V.; Chen, Y.H.; Yang, T.J.; Emer, J.S. Efficient processing of deep neural networks: A tutorial and survey. Proc. IEEE 2017, 105, 2295–2329. [Google Scholar] [CrossRef]
  51. He, S.; Zhang, Y.; Xie, R.; Jiang, D.; Ming, A. Rethinking Image Aesthetics Assessment: Models, Datasets and Benchmarks. In Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI, Vienna, Austria, 23–29 July 2022. [Google Scholar]
  52. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.