Next Article in Journal
Void Avoidance Opportunistic Routing Protocol for Underwater Wireless Sensor Networks
Next Article in Special Issue
Investigation on the Sampling Frequency and Channel Number for Force Myography Based Hand Gesture Recognition
Previous Article in Journal
Gated Graph Attention Network for Cancer Prediction
Previous Article in Special Issue
Learning for a Robot: Deep Reinforcement Learning, Imitation Learning, Transfer Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

How to Represent Paintings: A Painting Classification Using Artistic Comments

1
College of Mechanical Engineering, Zhejiang University of Technology, Hangzhou 310023, China
2
School of Intelligent Transportation, Zhejiang Institute of Mechanical & Electrical Engineering, Hangzhou 310053, China
3
School of Computing, University of Portsmouth, Portsmouth PO1 3HE, UK
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(6), 1940; https://doi.org/10.3390/s21061940
Submission received: 12 January 2021 / Revised: 25 February 2021 / Accepted: 1 March 2021 / Published: 10 March 2021

Abstract

:
The goal of large-scale automatic paintings analysis is to classify and retrieve images using machine learning techniques. The traditional methods use computer vision techniques on paintings to enable computers to represent the art content. In this work, we propose using a graph convolutional network and artistic comments rather than the painting color to classify type, school, timeframe and author of the paintings by implementing natural language processing (NLP) techniques. First, we build a single artistic comment graph based on co-occurrence relations and document word relations and then train an art graph convolutional network (ArtGCN) on the entire corpus. The nodes, which include the words and documents in the topological graph are initialized using a one-hot representation; then, the embeddings are learned jointly for both words and documents, supervised by the known-class training labels of the paintings. Through extensive experiments on different classification tasks using different input sources, we demonstrate that the proposed methods achieve state-of-art performance. In addition, ArtGCN can learn word and painting embeddings, and we find that they have a major role in describing the labels and retrieval paintings, respectively.

1. Introduction

Due to the large-scale digitization of art works over the past two decades, art coupled with computer technology has become a research hotspot. Most online painting collections include various metadata, often written by art experts, to provide context and improve audience understanding. For example, WikiArt (http://www.wikiart.org, accessed on 25 December 2020) includes artist, style, genre, media, date, dimensions and even artistic comments describing the paintings. Scholars who have spent years analyzing and learning the specifics and nuances of fine art can easily identify metadata associated with a painting [1]; however, identifying metadata is difficult for general audiences who lack expertise. As deep learning has developed, more works have focused on using learned features to conduct art classification [2,3,4,5,6,7]. In other words, by training computers using paintings previously labeled by human experts, the machines can learn the image features and classify labels for the images automatically. To the best of our knowledge, the features used thus far in art classification tasks have been extracted from the images themselves. Similar to humans, researchers have used paintings to teach machine knowledge about the arts in most previous works [8,9,10,11,12,13,14].
Language has a referential function, i.e., it can refer to objects in the world, and the object also has a visual representation; therefore, language can (at least to some extent) describe visual perception. Inspired by this phenomenon, we use the artistic comments associated with paintings to perform artwork classification. This approach shifts the problem from one of computer vision to one involving natural language processing (NLP).
To summarize, our contributions are as follows:
  • We propose a novel graph neural network method for art classification that uses features extracted from textual comments about art. In addition, on this basis, we propose a multitask learning model (MTL) to solve the different tasks using one model. This approach encourages the models to find common elements (and hence the context) between the different tasks. A comparison shows that our proposed approach performs state-of-the-art compared with traditional benchmarks that apply vision-based and text-based methods to classify art.
  • Our method performs word embedding and creates labels using the dimension that has the highest value. The mean of the label is the same as the painting label; thus, by extracting the top 10 words that had higher values in each category, we find that the extracted words are highly correlated descriptions of labels.
  • We analyze the SemArt dataset, including the class distribution. We create visualizations to analyze the results of the art classification for different tasks using ResNet50 and ArtGCN models. Overall, the classification effect improves when using ArtGCN—but not for all specific categories. These 2 methods include using the pixels of the painting itself (ResNet50) and art comments (ArtGCN). Here, we compare two ways of representing paintings using neural networks.
  • Based on the trained classification models, we develop a painting retrieval system and find that both methods achieve good performance. An analysis of the retrieval results further illuminates the differences in how the 2 models with different input sources work.

2. Related Work

2.1. Art Classification

Early works concentrating on automatic art analysis first extracted hand-crafted features from images and then performed classification using traditional machine learning methods [15,16,17] developed the QArt-learn approach to categorize paintings into art styles. The model was implemented using k-nearest neighbor and support vector machine (SVM) algorithms with QCD color features. In recent years, deep CNNs have achieved many successes on computer vision tasks using large hand-labeled datasets, such as the ImageNet dataset [18]. Recent studies to solve classification problems have been dominated by CNNs [11,19,20,21].

2.2. Text Classification

Text classification is a fundamental task in natural language processing (NLP). The most representative deep networks are sequential-based learning models, including both CNNs [22,23,24] and recurrent neural network (RNN) models [25,26,27]. The developed models are effective and widely used, but due to the model architectures, they focus primarily on text features extracted from local consecutive word sequences. Since 2018, Transformers have been the state-of-the-art architecture for most NLP tasks. The current SoTA approach to basically text classification is taking a pretrained Transformer [28] model (such as BERT [29]) and finetunes it for a particular task.

2.3. Graph Neural Networks

Recently, graph neural networks have received increasing attention in many applications [30,31,32,33,34]. Kipf and Welling presented a simplified graph neural network model called a graph convolutional network (GCN) [35]. By exploiting message passing (or equivalently, various neighborhood functions), GCNs have achieved high scores on several benchmark graph datasets. Recently, graph neural network (GNN) models have also attracted widespread attention and have been successfully applied to solve text classification problems by using global word co-occurrence information in a corpus [36,37,38].

3. Datasets

To achieve our goal of classifying paintings using artistic comments, we adopted the SemArt dataset collected from [39] to classify various metadata, including type, school, timeframe and author. In recent years, several datasets have been used to classify artworks, such as Painting91 [40], WikiArt-WikiPaintings [41] and MultitaskPainting100k [42]; however, these are labeled only by artists, styles and genres. Previous works have focused on the task of identifying paintings based on different artist representations. However, representing paintings is not an easy task, and paintings typically contain no more than 3 metadata points. In contrast, the paintings in the SemArt dataset include 8 types of labeled metadata, including artistic comments, authors, titles, techniques and formats, data regarding when the painting was created, types, schools and timeframes. These metadata improve the flexibility with which computers can represent paintings. Some examples of the dataset records are shown in Figure 1. For space reasons, we simplified artistic comments in the right-hand examples.

3.1. Data Analysis

The types used in our paper are as follows: religious, portrait, landscape, mythological, genre, still-life, historical, other, interior and study. The schools are grouped into 24 classes, e.g., Italian, English, American, etc. If the paintings are not labeled as belonging to any of these styles, we label them “unknown.“ Only 13 paintings are labeled unknown, indicating that these paintings do not belong to any of the 24 classes. We create timeframes consisting of every 50 years between 1050 and 1900 to determine the approximate date a painting was created. There are a total of 17 timeframes. In addition, 17 paintings in the dataset were not created between 1050 and 1900. Clearly identifying the authors of the paintings in the dataset is difficult because there are so many possible authors in the datasets. To better distinguish among authors, we selected 349 authors and labeled the others as unknown. For example, the artist Vincent van Gogh is the author of 327 paintings. A total of 8257 paintings are labeled as having an unknown author. In total, the dataset includes 11 styles, 25 classes, 18 timeframes and 350 authors (including the unknown labels). The distributions of these attributes can be seen in Figure 2. It can be found that class distribution is uneven, which makes sorting difficult. Figure 3 shows that the painting skill of every artist can vary widely. For example, all the paintings from Giotto di Bondone are religious, while Vincent van Gogh and Rembrandt Harmenszoon van Rijn are more diverse. Timeframes and schools are also highly relevant to author classification. For example, the paintings of the artist Giotto di Bondone belong to Italian (school) and the 1301–1350 timeframe, while the paintings of Rembrandt Harmenszoon van Rijn belong to the Dutch school and the 1601–1650 timeframe. In addition, paintings from Vincent van Gogh, belong to Dutch school and the 1851–1900 timeframe. The dataset is randomly divided into training, validation and test sets with 19,244, 1069, and 1069 images in each set, respectively. All the data are summarized in Table 1. The meanings of words, nodes and average length will be explained in Section 4.

4. Method

We first collected all the artistic comments including train, validation and test splits associated with the paintings and processed the texts. In detail, we cleaned the texts using preprocessing techniques such as changing case, correcting split characters (as in [22]), removing stopwords defined in NLTK and low frequency words appearing less than 5 times. After this process, the average length of an artistic comment decreased to 59.27 words, and the total number of unique words was 17,944, as shown in Table 1. Then, we build a large and heterogeneous text graph G = ( V , E ) , where V and E denote the sets of nodes and edges in the graph, respectively. There are nodes of two types:
  • Nodes representing artistic comments represented as TF-IDF weighted bag of words.
  • Nodes that correspond to unique words.
The number of nodes | V | is equal to the sum of the number of artistic comments and unique words in the corpus. We used a one-hot vector as the input features of the nodes, which means that an input feature X is an identity matrix of | V | dimensions. The weights of the edges between the documents and words were calculated using the term frequency-inverse document frequency (TF-IDF):
T F I D F i , j = T F i , j × I D F i
T F i , j = n i , j k n k , j
I D F i = | D | 1 + j : t i d j
where n i , j denotes the number of times word t i appears in an artistic comment, and T F i , j denotes the frequency of that word in an artistic comment. The term | D | represents the total number of artistic comments, and j : t i d j denotes the number of artistic comments that contain entry t i . The value of the TF-IDF index is higher when a word appears frequently in an article but appears less frequently among all documents. We employ pointwise mutual information (PMI) to calculate the weights of the edges between 2 word nodes. To calculate the results, we first applied a fixed-size sliding window to all the artistic comments to gather co-occurrence statistics:
PMI ( i , j ) = log p ( i , j ) p ( i ) p ( j )
p ( i , j ) = W ( i , j ) W
p ( i ) = W ( i ) W
where W ( i , j ) is the number of sliding windows containing both the words i and j, while W ( i ) is the number of sliding windows that contain word i. W is the total number of sliding windows. We adopt only the positive values that imply high semantic correlations between the words in a corpus. Formally, the weight of edge between nodes i and j is defined as:
A i j = PMI ( i , j ) i , j are words , PMI ( i , j ) > 0 TF-IDF i j i is artistic comment , j is word 0 otherwise .
After building the graph, we use a two-layer GCN model to perform the art classification task. A GCN [35] can operate directly on a graph and propagates features between neighborhoods. Mathematically, GCN follows this formula:
h ( l + 1 ) = σ D ˜ 1 2 A ˜ D ˜ 1 2 h ( l ) W ( l )
where h ( l ) denotes the l ( t h ) layer in the network, σ is the nonlinearity, and W is the weight matrix for this layer. D ˜ 1 2 A ˜ D ˜ 1 2 indicates a renormalization trick in which a self-connection appears for each graph node. Therefore, D ˜ is the corresponding degree matrix of A + I , and A ˜ = A + I . The shape of H ( 0 ) is N × D , where N is the number of nodes and D is the number of input features. As an alternative method of understanding, the following formula can also be used to describe the models:
h i ( l + 1 ) = σ j N ( i ) { i } 1 deg ( i ) · deg ( j ) · ( h j ( l ) W ( l ) )
where deg ( i ) and deg ( j ) are the degrees of nodes i and j, respectively, N ( i ) is a neighbor of node i, and σ is the activation function; here, we used ReLU as the activation function. In our experiments, we found that using a deeper model can lead to oversmoothing and the gradient explosion problem [43,44,45]. In this case, two layers are sufficient to allow messages to pass among nodes:
X = L s y m ReLU L s y m X W 0 W 1
where L s y m is a symmetric normalized Laplacian, L s y m = D ˜ 1 2 A ˜ D ˜ 1 2 . The meanings of D ˜ 1 2 and A ˜ are introduced in Equation (8). W 0 R N × D , W 1 R D × c l a s s . Here, N is the number of nodes, and D is the embedding size of the middle layer. c l a s s denotes the number of labels. The loss function is defined using the cross-entropy loss function.
loss ( x , class ) = log exp ( x [ class ] ) j exp ( x [ j ] ) = x [ class ] + log j exp ( x [ j ] )
where x and c l a s s are the output training nodes of the model and the targeted label of the training nodes being learned, respectively. We first process the data and build the graph; then we set up an early termination mechanism: training stops when l o s s _ v a l does not decrease for l r _ p a t i e n c e or when the maximum number of epochs is reached. For the mtl-ArtGCN, we use four different GCN layers in the second layer of the model to satisfy different tasks, and then calculate the loss using the outputs, the weight of each loss is set to 0.25. We used this trained model to calculate the classification accuracy. There is no need to interface for the test samples because the value and test node information is also updated during the training. When training is complete, we extract the embedding of the value and test nodes and use the dimension in the node information where the vector value is the largest as the predict label. Then, we calculate the validation and test accuracies using the true and predict labels.

5. Experiments

5.1. Hyperparameter Selection

To select the appropriate hyperparameters, we first set the learning rate to 0.02, dropout to 0, and L 2 loss weight to 0. We set the maximum number of epochs to 200, l r _ p a t i e n c e to 10 and adopted the Adam [46] optimizer to learn the parameters. We tested different hyperparameter combinations via a grid search to study the parameter sensitivity of the models. All the experiments were executed 4 times to obtain the m e a n ± s t d . Figure 4 shows the test accuracy under different hyperparameters, including the sliding window sizes and the mid-layer dimensions. The first row shows the test accuracy under different sliding window sizes and reveals that the test accuracy first increases as the window size increases but no longer changes much beyond a window size of 20. These results show that adopting a word sliding window size that is too small generates insufficient word occurrence information, while a window that is too large may introduce some invalid edges. The second row shows the test accuracy under different mid-layer dimensions. A too-low mid-layer embedding dimension does not propagate node label information well, while a too-high value does not improve the classification performance and time consumption. After tuning the hyperparameters, we used a mid-layer dimension of 300 and a window size of 20 for the remaining experiments.

5.2. Baselines

We compared our models with the results of computer vision learning methods using ResNet50, ResNet101 and ResNet152 [47], which have achieved excellent image recognition performance. To adapt the models for art classification, we modified the last fully connected layer to match the number of classes in each task. For the mtl-resnet50, we replace the last fully connected layer with four different fully connected layers to classify the paintings in different tasks in one model, the loss weight calculated by these different fully connected layers is set to 0.25 for all the tasks. For the kgm-resnet50 [48], the author use the contextual information which is obtained from capturing relationships in an artistic knowledge graph built with non-visual artistic metadata to help train the model. Due to the limited number of images, we used transfer learning, which the model parameters are first trained on ImageNet and then fine-tuned for art classification. In addition, we compare ArtGCN with multiple state-of-the-art text classification methods as follows: (1) TF-IDF + LR, which we have described in Section 4 and logistic regression is used as the classifier; (2) fastText [49], which treats the average of word or n-gram embeddings as document embeddings and a linear classifier to train the model. We evaluate the model with and without bigrams; (3) fine-tuned RoBERTa [50], which we fine-tuned the pretrained RoBERTa models in the SemArt. In general, we focus on visual-based and text-based painting classification efforts. The classification efforts of two different domain classification models for the same task are compared.

6. Results and Discussion

6.1. Results

Table 2 shows the results of the art classification. MTL denotes the multitask learning model [48], which is trained to learn multiple artistic tasks jointly. The results show that various classification tasks (including type, school, timeframe and author) perform better when using ArtGCN. Compared with the ResNet50 results, ArtGCN increased the type, school, timeframe and author classification accuracies by 5.00%, 23.90%, 21.11%, and 26.03%, respectively. These results indicate that our method can achieve huge improvements, especially for the school, timeframe and author tasks, and demonstrate the effectiveness of using art comments to represent art. By comparing with the previous state-of-the-art text classification methods, our method still achieved good performance, especially in timeframe and author classification. The parameters of popular RoBERTa model have 124.66 M, while our model only have 11.80 M. It shows our model has also greatly improved in terms of the number of parameters. We also find the RoBERTa and fastText methods, which text-based in artist classification is less effective than visual-based methods. Therefore, text-based methods are not always better than visual-based methods and a well-designed algorithm is also important to obtain state-of-the-art accuracy. Compared with the MTL-ArtGCN results, the classification results differ only slightly, which shows that different tasks involved in art classification have substantial commonalities. In other words, the first layer of the model learns the common parameters for different tasks. This is consistent with our common sense. Determining the artist of a painting can help in determining the nationality (school) and period (timeframe) of the painting and has some relevance to the style.

6.2. Word Embedding

The main working mechanism of a GCN is the propagation of node features through neighborhood nodes, through which the model achieves better node classification performance. As described in Section 4, we create a graph using ArtGCN in which the nodes are composed of both artistic comments and words. Only the labeled nodes form the training set of the artistic comments, and the models cannot identify which category the nodes belong to artistic comments or simply words. Therefore, we conduced some experiments. For the word node embeddings, we first performed normalization using the SoftMax function and considered the dimension with the highest value as a word label in the output embedding of the model, similar to the model prediction of the artistic comment node categories. The dimensions of the word labels are equal to the number of painting classes. We set the word labels as the image classes, e.g., if a word label is 0 in the style classification task, and a label of 0 for the painting represents religion; then the word belongs to the religion label. Figure 5 shows a t-SNE [51] visualization of the word embeddings. It can be found that words with the same label are close together, which means that most words are closely related to the painting classes.
After the embedding normalized by the SoftMax function, we show the top 10 words with the highest values in specific dimensions based on labels in Table 3 and find that the words are highly associated with the descriptions of the labels. For example, the religious words ‘saints’, ‘triptych’, and ’mary’ in column 1 express religion well, while words in other columns that show similar phenomena (e.g., estuary, views, and coastal) are highly related to landscapes. This means that the top 10 words with the highest values associated with each class are interpretable and can be used to describe the classes.
Table 4 shows the results for the school classification tasks. The output of the models has 25 dimensions (equal to the number of class labels). The calculation method used is the same as that for the style tasks, and we reach the same conclusions as for the style tasks. When talking about Italian, we are always reminded of Gian Giacomo Poldi Pezzoli, the founder of the Museo Poldi Pezzoli and painting signed: PETRVS PERVSINVS PINXIT, painted by PERUGINO, Pietro. Table 5 shows the results for the timeframe tasks. The most obvious result is that particular years are classified within each timeframe to describe each class, e.g., 1640 and 1635 are used to describe the timeframe of 1601–1650, while 1670 s is used to describe the timeframe of 1651–1700. The presented way of learning the representation seems to be a good way of learning domain-specific word embeddings. The embedding similarity could find good use in a specialized art search engine.

6.3. Interpreting the Classification Results

We used t-SNE to visualize the model output embeddings, including the classification tasks for type, school and timeframe in Figure 6. The first row shows the output of the ResNet50 model, while the second row shows the output of our ArtGCN models. Please note that both the paintings and artistic comments contain category information; and the category represented by the color corresponds to Figure 2. Overall, several observations are worth attention. In total, the accuracy score when using the ArtGCN model are higher than those when using ResNet50. However, for some specific categories (i.e., interior, landscape and still-life), ResNet50 performs better than does ArtGCN on the type classification task. For the school and timeframe classification tasks, ArtGCN outperforms ResNet50 on all the categories. Timeframes are more likely to be divided into adjacent periods rather than other, more distant periods. From experimental results, we can see that the proposed ArtGCN can achieve state-of-the-art classification results and learn painting and word embeddings. However, one of the main limitations of this study is that the GCN model is inherently transductive, test and validation artistic comment nodes(without labels) are included in GCN training. Thus, a major limitation of our study is that our method could not generate embeddings and make predictions for new paintings. Possible solutions to the problem are introducing inductive [52] or fast GCN model [53].

6.4. Painting Retrieval

In addition to exploring task-specific classification, we aimed to retrieve paintings of similar categories. We performed retrieval tasks for type and authors rather than for all 4 tasks because timeframe and school are highly related to the authors. To accomplish the retrieval goal, we used the various models to perform feature extraction and then calculated the similarity between feature vectors using cosine distance. Starry Night (Figure 7) is an oil-on-canvas painting by the Dutch post-impressionist painter Vincent van Gogh. In this painting, he vividly depicts a starry sky full of movement and change using exaggerated techniques. The entire picture is subsumed in a turbulent blue-green torrent. The rotating, restless and curling nebulae make the night sky seem extremely active. This unreal scene reflected Vincent van Gogh’s restless emotions and the illusory world of his madness. The Dream of Solomon (Figure 8) was painted by Luca Giordano; it depicts the story of Solomon from the Bible. During one night in a dream, Solomon asked the Lord to help him distinguish between right and wrong in his heart. The Lord was pleased that Solomon asked for this, and gave him a wise and discerning heart as well as wealth and honor. The Picnic (Figure 9) was painted by Monet, who used Impressionist light and shadow techniques to depict people relaxing on a grassy sward. The use of external light lends the picture a colorful, fresh and hearty effect. We used Reset50 (left) and ArtGCN (right) models to retrieve the paintings. The first row shows paintings retrieved by style, while the second row shows paintings retrieved by author. We have labeled the different retrieval categories below the paintings(for both style and author retrieval, we only tag the categories that differ). As we can see, all the styles are just same; only a small author category is different. Although these 2 models achieve great performance, the results are quite different. These results show that the retrieval mechanism of the two methods is completely different—even though both models can retrieve oil paintings correctly.

7. Conclusions

In this study, we used artistic comments to classify painting categories via a developed network named ArtGCN. We compared the results of ArtGCN with those of Resnet models, which use the actual image pixels to recognize painting categories. The experiments show that our method achieves significant increases in accuracy on 4 classification tasks. This result means that—in contrast to the painting itself—using other labeled information for paintings, such as artistic comments, also contains important information that can be used to train a computer to represent the art. In addition, we performed word embeddings tagged the labels using the embedding dimension with the highest value and extracted 10 words that have the highest values in the tagged dimension. We found that these higher-valued words are closely related to the categories. In other words, the extracted words can be used to better describe this category. We visualized and analyzed some phenomena worthy of attention. Both the methods based on computer vision and those artistic comments perform well on various painting classification tasks. We also constructed a painting retrieval system that can be used to enhance the capabilities of search systems in different online art collections. We tested the painting retrieval of the developed models using 3 paintings. The results of the two types of models are quite different and show that the art representation learned by computers is quite different when trained using a method based on computer vision versus a method based on artistic comments (natural language processing).
In this study, we used color information from the paintings and features from the artistic comments separately to address different painting classification tasks. In future work, we will attempt to combine these two methods by using the teacher model [54] and other methods. In addition, some labeling information exists that was not used in our experiments, and we believe there may be some inner relations among this additional information. For example, Vincent van Gogh’s paintings have a high probability of being landscape or portrait styles, while all the paintings of Giotto di Bondone are related to religion. Therefore, we will attempt to incorporate these additional features to build classification approaches based on both color features and other information. Lastly, an important indicator of how well the models perform would be a comparison with human performance on the task, i.e., the ability of experts and non-experts to guess the painting attributes just from the comment, we will also try to compare our methods with human performance.

Author Contributions

Conceptualization, W.Z. and W.J.; investigation and data curation, X.Q.; writing—original draft preparation, W.Z.; writing—review and editing, W.Z. and D.Z.; project administration, W.J.; funding acquisition, W.J and X.Q.; All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Key Laboratory of E&M (Zhejiang University of Technology), Ministry of Education & Zhejiang Province(Grant No. EM 2016070101).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sandoval, C.; Pirogova, E.; Lech, M. Two-Stage Deep Learning Approach to the Classification of Fine-Art Paintings. IEEE Access 2019, 7, 41770–41781. [Google Scholar]
  2. Cetinic, E.; Lipic, T.; Grgic, S. Fine-Tuning Convolutional Neural Networks for Fine Art Classification. Expert Syst. Appl. 2018, 114, 107–118. [Google Scholar]
  3. Cetinic, E.; Lipic, T.; Grgic, S. A Deep Learning Perspective on Beauty, Sentiment, and Remembrance of Art. IEEE Access 2019, 7, 73694–73710. [Google Scholar]
  4. Cetinic, E.; Lipic, T.; Grgic, S. Learning the Principles of Art History with Convolutional Neural Networks. Pattern Recognit. Lett. 2020, 129, 56–62. [Google Scholar] [CrossRef]
  5. Huckle, N.; Garcia, N.; Nakashima, Y. Demographic Influences on Contemporary Art with Unsupervised Style Embeddings. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2020; pp. 126–142. [Google Scholar]
  6. Chen, L.; Yang, J. Recognizing the Style of Visual Arts via Adaptive Cross-Layer Correlation. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 2459–2467. [Google Scholar]
  7. Wynen, D.; Schmid, C.; Mairal, J. Unsupervised Learning of Artistic Styles with Archetypal Style Analysis. arXiv 2018, arXiv:1805.11155. [Google Scholar]
  8. Falomir, Z.; Museros, L.; Sanz, I.; Gonzalez-Abril, L. Categorizing Paintings in Art Styles Based on Qualitative Color Descriptors, Quantitative Global Features and Machine Learning (QArt-Learn). Expert Syst. Appl. 2018, 97, 83–94. [Google Scholar] [CrossRef]
  9. Ma, D.; Gao, F.; Bai, Y.; Lou, Y.; Wang, S.; Huang, T.; Duan, L.Y. From Part to Whole: Who Is behind the Painting? In Proceedings of the 25th ACM International Conference on Multimedia, Sliema, Malta, 3–7 April 2017; pp. 1174–1182. [Google Scholar]
  10. Mao, H.; Cheung, M.; She, J. Deepart: Learning Joint Representations of Visual Arts. In Proceedings of the 25th ACM International Conference on Multimedia, Sliema, Malta, 3–7 April 2017; pp. 1183–1191. [Google Scholar]
  11. Tan, W.R.; Chan, C.S.; Aguirre, H.E.; Tanaka, K. Ceci n’est Pas Une Pipe: A Deep Convolutional Network for Fine-Art Paintings Classification. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; IEEE: New York, NY, USA, 2016; pp. 3703–3707. [Google Scholar]
  12. Garcia, N.; Ye, C.; Liu, Z.; Hu, Q.; Otani, M.; Chu, C.; Nakashima, Y.; Mitamura, T. A Dataset and Baselines for Visual Question Answering on Art. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2020; pp. 92–108. [Google Scholar]
  13. Sheng, S.; Moens, M.F. Generating Captions for Images of Ancient Artworks. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 2478–2486. [Google Scholar]
  14. Baraldi, L.; Cornia, M.; Grana, C.; Cucchiara, R. Aligning Text and Document Illustrations: Towards Visually Explainable Digital Humanities. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; IEEE: New York, NY, USA, 2018; pp. 1097–1102. [Google Scholar]
  15. Shamir, L.; Macura, T.; Orlov, N.; Eckley, D.M.; Goldberg, I.G. Impressionism, Expressionism, Surrealism: Automated Recognition of Painters and Schools of Art. ACM Trans. Appl. Percept. 2010, 7, 1–17. [Google Scholar]
  16. Arora, R.S.; Elgammal, A. Towards Automated Classification of Fine-Art Painting Style: A Comparative Study. In Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012), Tsukuba, Japan, 11–15 November 2012; IEEE: Nork, NY, USA, 2012; pp. 3541–3544. [Google Scholar]
  17. Agarwal, S.; Karnick, H.; Pant, N.; Patel, U. Genre and Style Based Painting Classification. In Proceedings of the 2015 IEEE Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 5–9 January 2015; IEEE: Nork, NY, USA, 2015; pp. 588–594. [Google Scholar]
  18. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A Large-Scale Hierarchical Image Database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; IEEE: Nork, NY, USA, 2009; pp. 248–255. [Google Scholar]
  19. Strezoski, G.; Worring, M. Omniart: A Large-Scale Artistic Benchmark. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2018, 14, 1–21. [Google Scholar]
  20. Seguin, B.; Striolo, C.; Kaplan, F. Visual Link Retrieval in a Database of Paintings. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2016; pp. 753–767. [Google Scholar]
  21. Chu, W.T.; Wu, Y.L. Image Style Classification Based on Learnt Deep Correlation Features. IEEE Trans. Multimed. 2018, 20, 2491–2502. [Google Scholar]
  22. Kim, Y. Convolutional Neural Networks for Sentence Classification. arXiv 2014, arXiv:1408.5882. [Google Scholar]
  23. Zhang, X.; Zhao, J.; LeCun, Y. Character-Level Convolutional Networks for Text Classification. arXiv 2015, arXiv:1509.01626. [Google Scholar]
  24. Conneau, A.; Schwenk, H.; Barrault, L.; Lecun, Y. Very Deep Convolutional Networks for Text Classification. arXiv 2016, arXiv:1606.01781. [Google Scholar]
  25. Tai, K.S.; Socher, R.; Manning, C.D. Improved Semantic Representations from Tree-Structured Long Short-Term Memory Networks. arXiv 2015, arXiv:1503.00075. [Google Scholar]
  26. Luo, Y. Recurrent Neural Networks for Classifying Relations in Clinical Notes. J. Biomed. Inform. 2017, 72, 85–95. [Google Scholar]
  27. Liu, P.; Qiu, X.; Huang, X. Recurrent Neural Network for Text Classification with Multi-Task Learning. arXiv 2016, arXiv:1605.05101. [Google Scholar]
  28. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. NIPS. 2017. Available online: https://arxiv.org/pdf/1706.03762.pdf (accessed on 15 February 2021).
  29. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. arXiv 2019, arXiv:1810.04805. [Google Scholar]
  30. Tu, M.; Wang, G.; Huang, J.; Tang, Y.; He, X.; Zhou, B. Multi-Hop Reading Comprehension across Multiple Documents by Reasoning over Heterogeneous Graphs. arXiv 2019, arXiv:1905.07374. [Google Scholar]
  31. Zhang, M.; Cui, Z.; Neumann, M.; Chen, Y. An End-to-End Deep Learning Architecture for Graph Classification. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  32. Ying, Z.; You, J.; Morris, C.; Ren, X.; Hamilton, W.; Leskovec, J. Hierarchical Graph Representation Learning with Differentiable Pooling. arXiv 2018, arXiv:1806.08804. [Google Scholar]
  33. Cangea, C.; Veličković, P.; Jovanović, N.; Kipf, T.; Liò, P. Towards Sparse Hierarchical Graph Classifiers. arXiv 2018, arXiv:1811.01287. [Google Scholar]
  34. Bianchi, F.M.; Grattarola, D.; Alippi, C.; Livi, L. Graph Neural Networks with Convolutional Arma Filters. arXiv 2019, arXiv:1901.01343. [Google Scholar]
  35. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  36. Yao, L.; Mao, C.; Luo, Y. Graph Convolutional Networks for Text Classification. arXiv 2018, arXiv:1809.05679. [Google Scholar]
  37. Liu, X.; You, X.; Zhang, X.; Wu, J.; Lv, P. Tensor Graph Convolutional Networks for Text Classification. AAAI 2020, 34, 8409–8416. [Google Scholar]
  38. Huang, L.; Ma, D.; Li, S.; Zhang, X.; Wang, H. Text Level Graph Neural Network for Text Classification. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; Association for Computational Linguistics: Hong Kong, China, 2019; pp. 3444–3450. [Google Scholar] [CrossRef] [Green Version]
  39. Garcia, N.; Vogiatzis, G. How to Read Paintings: Semantic Art Understanding with Multi-Modal Retrieval. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  40. Khan, F.S.; Beigpour, S.; Van de Weijer, J.; Felsberg, M. Painting-91: A Large Scale Database for Computational Painting Categorization. Mach. Vis. Appl. 2014, 25, 1385–1397. [Google Scholar]
  41. Karayev, S.; Trentacoste, M.; Han, H.; Agarwala, A.; Darrell, T.; Hertzmann, A.; Winnemoeller, H. Recognizing Image Style. arXiv 2013, arXiv:1311.3715. [Google Scholar]
  42. Bianco, S.; Mazzini, D.; Napoletano, P.; Schettini, R. Multitask Painting Categorization by Deep Multibranch Neural Network. Expert Syst. Appl. 2019, 135, 90–101. [Google Scholar]
  43. Zhou, K.; Dong, Y.; Lee, W.S.; Hooi, B.; Xu, H.; Feng, J. Effective Training Strategies for Deep Graph Neural Networks. arXiv 2020, arXiv:2006.07107. [Google Scholar]
  44. Zhao, L.; Akoglu, L. Pairnorm: Tackling Oversmoothing in Gnns. arXiv 2019, arXiv:1909.12223. [Google Scholar]
  45. Rong, Y.; Huang, W.; Xu, T.; Huang, J. Dropedge: Towards Deep Graph Convolutional Networks on Node Classification. arXiv 2019, arXiv:1907.10903. [Google Scholar]
  46. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  47. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  48. Garcia, N.; Renoust, B.; Nakashima, Y. Context-Aware Embeddings for Automatic Art Analysis. In Proceedings of the 2019 on International Conference on Multimedia Retrieval, Ottawa, ON, Canada, 10–13 June 2019; pp. 25–33. [Google Scholar]
  49. Joulin, A.; Grave, E.; Bojanowski, P.; Mikolov, T. Bag of Tricks for Efficient Text Classification. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, Valencia, Spain, 3–7 April 2017; Volume 2, pp. 427–431. [Google Scholar]
  50. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
  51. Van der Maaten, L.; Hinton, G. Visualizing Data Using T-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  52. Hamilton, W.L.; Ying, R.; Leskovec, J. Inductive Representation Learning on Large Graphs. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 1025–1035. [Google Scholar]
  53. Chen, J.; Ma, T.; Xiao, C. FastGCN: Fast Learning with Graph Convolutional Networks via Importance Sampling. arXiv 2018, arXiv:1801.10247. [Google Scholar]
  54. Hinton, G.; Vinyals, O.; Dean, J. Distilling the Knowledge in a Neural Network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
Figure 1. SemArt dataset. Each record includes a title (red), the author, technique and format, school, timeframe (yellow) and artistic comments (blue).
Figure 1. SemArt dataset. Each record includes a title (red), the author, technique and format, school, timeframe (yellow) and artistic comments (blue).
Sensors 21 01940 g001
Figure 2. Class distribution of the style (left 1), school (left 2), timeframe (right 2), author (right 1). To better display the author distribution, we show only the top 30 artists; their names have been simplified to improve the figure.
Figure 2. Class distribution of the style (left 1), school (left 2), timeframe (right 2), author (right 1). To better display the author distribution, we show only the top 30 artists; their names have been simplified to improve the figure.
Sensors 21 01940 g002
Figure 3. Paintings from the dataset adopted in this work. Each row contains samples from a different artist. The color coding represents different styles.
Figure 3. Paintings from the dataset adopted in this work. Each row contains samples from a different artist. The color coding represents different styles.
Sensors 21 01940 g003
Figure 4. Test accuracy under different sliding window sizes and mid-layer embedding dimensions.
Figure 4. Test accuracy under different sliding window sizes and mid-layer embedding dimensions.
Sensors 21 01940 g004
Figure 5. The t-SNE visualization of the second layer word embeddings. Each node represents the unique words in artistic comments. There are a total of 17,944 nodes (the number of words contained in artistic comments); we set the highest dimension as the word’s label.
Figure 5. The t-SNE visualization of the second layer word embeddings. Each node represents the unique words in artistic comments. There are a total of 17,944 nodes (the number of words contained in artistic comments); we set the highest dimension as the word’s label.
Sensors 21 01940 g005
Figure 6. The t-SNE visualization of the image embedding and artistic comment embedding in the test set.
Figure 6. The t-SNE visualization of the image embedding and artistic comment embedding in the test set.
Sensors 21 01940 g006
Figure 7. Similarity results for the painting Starry Night by Vincent Van Gogh.
Figure 7. Similarity results for the painting Starry Night by Vincent Van Gogh.
Sensors 21 01940 g007
Figure 8. Similarity results for the painting Dream of Solomon, by Giordano, Luca.
Figure 8. Similarity results for the painting Dream of Solomon, by Giordano, Luca.
Sensors 21 01940 g008
Figure 9. Similarity results for the painting The Picnic, by Monet, Claude.
Figure 9. Similarity results for the painting The Picnic, by Monet, Claude.
Sensors 21 01940 g009
Table 1. Dataset characteristics, including the number of categories in the different classification tasks based on the paintings in the training set.
Table 1. Dataset characteristics, including the number of categories in the different classification tasks based on the paintings in the training set.
Dataset# Paintings# Words# Nodes# Average Length# Classes
train19,244---11TYPE
val1069---25SCHOOL
test1069---18TIMEFRAME
total21,38217,94439,32659.27350AUTHOR
Table 2. Art classification results on SemArt Datasets.
Table 2. Art classification results on SemArt Datasets.
# Region# Model# TYPE# SCHOOL# TIMEFRAME# AUTHOR# AVE
cvresnet50 [47]0.7870.6360.5920.5570.643
resnet101 [47]0.7710.6550.5910.5190.634
resnet152 [47]0.8060.6440.6150.5460.653
mtl-resnet500.7900.6670.6160.5260.650
kgm-resnet50 [48]0.8150.6710.6130.6150.679
nlpTF-IDF+LR0.7720.6880.4800.0970.509
fastText [49]0.7870.7570.6650.4980.677
fastText (bigrams)0.8040.7740.6340.4530.666
RoBERTa [50]0.8150.7830.5450.4650.652
mtl-ArtGCN0.8150.7830.7070.6860.748
ArtGCN0.8260.7880.7170.7020.758
Table 3. Words with highest values for 8 classes using second-layer embedding. We show the top 10 words for each class. (Artists comments classified by style).
Table 3. Words with highest values for 8 classes using second-layer embedding. We show the top 10 words for each class. (Artists comments classified by style).
ReligiousPortraitLandscapeMythologicalGenreStill-LifeHistoricalOther
saintsportraitestuarybacchusbambocciataporcelainbattlepainting
triptychsitterruisdaelscorpiosingerienotsalexander,
maryportraitsviewsaquariuscerutishrimpsfleet’s
virgincamuscoastalcapricornbamboccioblackberriesbrutusartist
angelsportraitistmooredpiscessteenbloomsarmypainted
madonnahertelboatsovidlhermittehazelnutswarone
altenburgsitterswaterfallsariadnesingeriestulipsdefeated\(
altarpiecedihautopographicalsagittariuslaergrapesnaval\)
polyptychcountesshobbemagoddessmetsufigskingpainter
depositionmorbillifishingpanmierismedlarshavanapaintings
Table 4. Words with highest values for several classes using second-layer construction. We show the top 10 words for each class. (Artists classified by school from comments).
Table 4. Words with highest values for several classes using second-layer construction. We show the top 10 words for each class. (Artists classified by school from comments).
ItalianDutchFrenchFlemishGermanSpanishEnglishNetherlandish
chapelsleidenfranbrugescranachjuanbritishbosch
pezzolinieuwechrubenshallezquezsickertbruegel
petrvsrembrandtfragonardbrueghelherlinvelgroomhaywain
petroniobrediusbouchermemlingnurembergcarlosstubbshell
estherkerklesnydersholbeinalonsomaitlandgeertgen
mariottihalscourbeteyckluthervicentestarraertsen
pesarohagueoispourbusheydaylaswrightdevils
evangelisthaarlembruyasbalensecessionbautista1stbouts
florentinehoochlautrecneeffsliebermannretablogainsboroughobverse
peruzzidutchoudryrubens’friedrichcajsidneysins
Table 5. Words with highest values for several classes using second-layer construction. We show top 10 words for each class. (Artists classified by timeframe from comments).
Table 5. Words with highest values for several classes using second-layer construction. We show top 10 words for each class. (Artists classified by timeframe from comments).
1601–16501501–15501651–17001451–15001851–19001551–16001701–17501751–1800
poussinrer1660sghirlandaiobrittanyarcimboldoriccipulcinella
barberinileovermeerpieroparisianzelottirosalbareynolds
velcapricornhoochbotticelliferenczysofonisbaballroomnemi
manfrediscorpiocarremantegnafattoritintorettowatteau1773
caravaggismaquariusterborchmemlingposterelpellegriniwright
1640begat1670sbellinicassattgrandilancret1777
ribera1525dourobertiboldinizuccarotiepolovolaire
halsgossartmaesturafantingrecocrespi1768
1635raphaelsteencossapoulduveronesecarrierazianigo
haarlemsagittariusdeventersignorellinabisdell’albergobouchergherardini
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhao, W.; Zhou, D.; Qiu, X.; Jiang, W. How to Represent Paintings: A Painting Classification Using Artistic Comments. Sensors 2021, 21, 1940. https://doi.org/10.3390/s21061940

AMA Style

Zhao W, Zhou D, Qiu X, Jiang W. How to Represent Paintings: A Painting Classification Using Artistic Comments. Sensors. 2021; 21(6):1940. https://doi.org/10.3390/s21061940

Chicago/Turabian Style

Zhao, Wentao, Dalin Zhou, Xinguo Qiu, and Wei Jiang. 2021. "How to Represent Paintings: A Painting Classification Using Artistic Comments" Sensors 21, no. 6: 1940. https://doi.org/10.3390/s21061940

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop