Next Article in Journal
Spectroscopic and Molecular Docking Analysis of π-Acceptor Complexes with the Drug Barbital
Next Article in Special Issue
Global–Local Self-Attention Based Transformer for Speaker Verification
Previous Article in Journal
Novel Bathing Assist Device Decreases the Physical Burden on Caregivers and Difficulty of Bathing Activity in Care Recipients: A Pilot Study
Previous Article in Special Issue
Deep Neural Network Model for Evaluating and Achieving the Sustainable Development Goal 16
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Graph Convolutional Networks with POS Gate for Aspect-Based Sentiment Analysis

1
FS, Inc., Daejeon-si 34126, Korea
2
Department of Computer Engineering, Chungbuk National University, Cheongju 28644, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(19), 10134; https://doi.org/10.3390/app121910134
Submission received: 31 August 2022 / Revised: 5 October 2022 / Accepted: 6 October 2022 / Published: 9 October 2022
(This article belongs to the Special Issue Deep Convolutional Neural Networks)

Abstract

:
We make daily comments on online platforms (e.g., social networks), and such natural language texts often contain sentiment (e.g., positive and negative) for certain aspects (e.g., food and service). If we can automatically extract the aspect-based sentiment from the texts, then it will help many services or products to overcome their limitations of particular aspects. There have been studies of aspect sentiment classification (ASC) that finds sentiment towards particular aspects. Recent studies mostly adopt deep-learning models or graph neural networks as these techniques are capable of capturing linguistic patterns that contributed to performance improvement in various natural language processing tasks. In this paper, for the ASC task, we propose a new hybrid architecture of graph convolutional network (GCN) and recurrent neural network. We design a gate mechanism that jointly models word embeddings and syntactic representation of sentences. By experimental results on five datasets, we show that the proposed model outperforms other recent models and also verify that the gate mechanism contributes to the performance improvement. The overall F1 scores that we achieved is 66.64∼76.80%.

1. Introduction

People often write comments on social network services (SNS) or websites of arbitrary services (e.g., online shopping mall), and it is obvious that it will be significantly helpful for service providers if it is possible to get the opinions or customer satisfaction level from several customer comments. Such comments are mostly written in natural language texts, and an example is shown in Figure 1; there are two aspects (e.g., food and service) with different sentiment (e.g., positive and negative). There have been studies for analyzing the natural language texts containing sentiment and aspects, and such a research field is known as aspect-based sentiment analysis (ABSA). The ABSA contains few tasks [1]: (1) entity-based aspect identification, (2) extraction of linguistic expressions that refer to arbitrary aspects, and (3) aspect sentiment classification (ASC). We aim at the task of ASC that extracts aspect-based sentiment from sentences. A sentence may contain multiple aspects, each of which corresponds to different sentiment, and the aspect-based sentiment (e.g., ‘service’: negative, and ‘food’: positive) will help to examine customer satisfaction; for example, the restaurant owner will investigate what is wrong with the service in the restaurant. Because previous studies might be interpreted in the wrong way if we use different definitions of sentiment, we here follow the definition of Giuseppe D’Aniello et al. [2]; sentiment is a durable emotional disposition that is developed by the user with respect to an aspect.
For the task of ASC whose goal is to identify aspects and sentiment expressed towards each aspect, there have been studies using machine learning models. Kiritchenko et al. [3] used support vector machine (SVM) [4] and achieved 80% accuracy for a dataset of SemEval 2014 Task 4 subtask 2 restaurant domain (REST14). Mubarok et al. [5] used naive Bayes classifier with a feature selection algorithm, and they achieved a 78% of F1 score for the REST14 dataset. Although these studies have shown successful performance, they are limited because they require significant effort in feature engineering.
Deep learning (DL) models are favorable as they find arbitrary features automatically from data [6]. There are a few well-known types of DL models: multi-layered perceptron (MLP) [7], recurrent neural networks (RNN) [8], convolutional neural networks (CNN) [9], and attention mechanism [10]. These different DL models have their own merits, as discussed in [11]; for example, some studies employed CNN to tackle the task of sentiment classification as the CNN is known to be efficient and effective in analyzing local patterns [12,13]. In particular, attention-based RNN architecture was often adopted in many studies of the ASC task as the attention-based architecture has shown its effectiveness in analyzing sequential patterns beneath the word sequences of natural language texts [10,14]. Wang et al. [15] proposed an attention-based RNN model using long short-term memory (LSTM) [16], and achieved 78% accuracy for REST14. Fan et al. [17] proposed bi-directional LSTM (Bi-LSTM) with a fine-grained attention mechanism, that captures the word-level interaction between aspect and context. They achieved 81% accuracy for REST14.
Recently, a graph neural network is widely used to incorporate graph-like patterns such as dependency between words for the ASC task. Xing and Tsang [18] proposed a hybrid architecture of LSTM and a graph neural network, and adopted other resources (e.g., DBpedia and Wikipedia) to train their model. They achieved 87% accuracy for REST14. Huang et al. [19] utilized pre-trained Bidirectional Encoder Representations from Transformers (BERT) [20] and graph attention network (GAT) on the dependency tree, and achieved 85% accuracy for REST14. Using context-aware pre-trained language models (e.g., BERT) has drawn much attention as it allows us to achieve better accuracy; however, it requires relatively large computational cost compared to context-free language models (e.g., Word2Vec [21] and Glove [22]), as reported in [23]. There are studies that adopt neither context-aware pre-trained models nor other resources. Zhang et al. [24] proposed building a graph convolutional network (GCN) over the dependency tree of a sentence to exploit syntactical information and word dependencies, and achieved 80% accuracy for the REST14 dataset. Xiao et al. [25] proposed a model of multi-head attention and an attentional-encoding-based GCN of a dependency tree, and achieved 81% accuracy for REST14. Bai et al. [26] proposed a relational graph attention network (RGAT) to incorporate typed syntactic dependencies. They also utilized part-of-speech (POS) tags that are known to convey rich syntactic patterns, and achieved 83% accuracy for REST14. However, they just used embeddings of POS tags, and did not jointly model the POS embeddings and word embeddings.
Word embeddings have often been jointly modeled with syntactic representation (e.g., POS tags) [27,28] because it allows us to capture more sophisticated patterns between semantic and syntactic representation. For example, the word ‘good’ has a different meaning for different POS tags (e.g., adverb or noun), and it has multiple meaning even for the same POS tag. In this paper, we assume that modeling semantic and syntactic representations jointly will contribute to performance improvement for the ASC task.
We propose a multi-layered GCN-BiLSTM architecture with a gate mechanism that jointly incorporates POS representations and word representations, and proves the effectiveness of the model by experimental results. The proposed architecture consists of GCN and BiLSTM, where the BiLSTM extracts sequential patterns in the given texts and the GCN analyzes syntactic information. The gate mechanism incorporates the syntactic information and POS tags together, which makes the entire model work more accurately.

2. Method

Figure 2 depicts the framework of our proposed model, namely POS-gated graph convolutional networks (PGGCN), that is composed of two embedding layers (e.g., word embedding and POS embedding) and a hybrid of GCN-BiLSTM layers with the attention mechanism. The word embedding layer and the POS embedding layer convert the given input sequence of words (or POS tags) into real-valued vectors in a different space; thus, we will get two embedding vectors from the embedding layers. Each of the two embedding vectors is delivered to BiLSTM that captures sequential patterns in the embedding vector. The multi-layered GCN and POS-gated attention mechanism exploit the sequential patterns extracted from the BiLSTM models, and generates final representation. Details of each part of the proposed model are provided in the following subsections.

2.1. Embeddings and Bi-LSTM

Given a sentence { w 1 , w 2 , . . . , w n } where w i indicates i-th word, we get a sequence of POS tags { p 1 , p 2 , . . . , p n } using NLTK [29] POS tagger. Aspect is a sub-sequence of the word sequence, and there may exist one or more aspects in a sentence. The words and POS tags are passed to the word embedding layer and POS embedding layer. The embedding layers embed the words and POS tags in vector spaces with word embedding matrix and POS embedding matrix E w R d e × | V | , E p R d e × | P | , where d e is the dimension of embedding, | V | denotes the vocabulary size, and | P | indicates the number of unique POS tags. Pre-trained GloVe [22] is used for the word embedding, and the POS embedding is generated using training data with POS tags. The word embeddings and POS embeddings are passed to two different bi-directional LSTM layers (i.e., Word LSTM and POS LSTM), and they yield a sequence of word hidden state vectors { h 1 w , h 2 w , . . . , h n w } and a sequence of POS hidden state vectors { h 1 p , h 2 p , . . . , h n p } , where h t · R k is a concatenation of hidden states of forward and backward LSTM at t-th word.

2.2. Gcn and Aspect-Specific Masking

We created an adjacency matrix A between words based on a dependency tree obtained by SpaCy toolkit [30]. The matrix A and word hidden states are fed into the L-layered GCN as shown in the left side of Figure 2. For every l-th layer, i-th word representation h i l is updated by incorporating representations of its adjacent words with normalization factor [31] as follows:
h ˜ i l = j = 1 n A i j W l g j l 1
h i l = R e l u ( h ˜ i l / ( d i + 1 ) + b l )
where g j l 1 is j-th word representation of previous layer, d i = j = 1 n A i j is a degree of the i-th word in the dependency tree. Note that h i l is computed using g j l 1 that is a position-aware word representation [32,33,34] as follows.
g i l 1 = p i h i l 1
p i = 1 r + 1 i n 1 i < r + 1 0 r + 1 i r + m 1 i r m n r + m < i n
where r + 1 is the starting position of an aspect, m is a length of the aspect, and p i R is the position-aware weight of the i-th word.
At the last GCN layer, we get the final hidden state vectors H L = { h i L } where 1 i n . As shown in the top-left corner in Figure 2, we perform aspect-specific masking that removes non-aspect hidden state vectors; the outputs will be H m a s k = { 0 , . . . , h r + 1 L , . . . , h r + m L , . . . , 0 } .

2.3. Part-of-Speech Gate

POS gate is designed to regularize the POS hidden states based on word hidden states. For the i-th word, it takes h i w and h i p as input, and generates h i p o s R m as follows.
h i g = W g · tanh ( h i w + h i p )
where W g is a trainable matrix of R m × k . The gate output is used together with GCN output to generate attention scores as shown in top-right corner of Figure 2.

2.4. Attention-Based Prediction

We compute attention scores for t-th position by incorporating H m a s k and H g a t e = { h i g } through a retrieval-based attention mechanism [24].
β t = i = 1 n h t g h i L = i = r + 1 r + m h t g h i L
α t = e x p ( β t ) i = 1 n e x p ( β i )
Finally, sentiment y ^ is predicted based on h a t t as formulated below.
h a t t = i = 1 n α i h i g
y ^ = s o f t m a x ( W h a t t + b )

2.5. Training

We train the model using the cost function C with L 2 -regularization [35] and cross-entropy:
C = i = 1 N j = 1 S y i j · l o g ( y ^ i j ) + γ Θ 2
where N is the number of instances, S is the number of classes, y ^ i j means the output of the model, y i j is the ground truth, and γ denotes the L 2 -regularization coefficient.

3. Experiments

3.1. Datasets and Experimental Settings

We conduct experiments on five datasets: restaurant reviews of SemEval 2014, 2015, and 2016 (i.e., REST14, REST15, and REST16) [1,36,37], laptap review of SemEval 2014 (LAPTOP), and Twitter review data (TWITTER) [38]. The Statistics are summarized in Table 1.
The embedding dimension is set to 300, and we borrowed pre-trained GloVe vectors [22] to initialize the word embeddings. The number of GCN layers L = 2 . We utilized Adam optimizer with the initial learning rate of 10 3 . The L 2 regularization coefficient γ is 10 5 , and mini-batch size is 32. We use accuracy and macro-averaged F1 score as evaluation metrics, and all results are obtained by averaging 20 independent runs; the F1 score and accuracy are, of course, not the perfect measure, and alternative ways of F-measure [39] may be considered when they are more studied in the future. Our model is compared with others [18,19] that do not employ pre-trained context-aware language models (e.g., BERT) or any other resources (e.g., dataset or knowledge-base).

3.2. Results

Table 2 shows the overall performance of our model PGGCN and other various models. The PGGCN outperformed other models with all datasets. There are a few interesting points. First, all GCN-based models (e.g., ASGCN, AEGCN, and MGGCN) showed better performance than others; this might imply that the GCN has better ability to grasp underlying syntactic patterns, so it contributes to performance improvement. Second, with the TWITTER dataset, the performance gap between PGGCN and others is smaller compared to other datasets of SemEval. The TWITTER dataset has quite different nature from others; it is collected from social network platform and has much more ‘neutral’ instances, as shown in Table 1. Such many ‘neutral’ instances confuse the models, and may drive the performance gap to be smaller. This is reasonable as ‘neutral’ might not even exist as we are always feeling something [40]. We also observed that all models generally gave better results with greater datasets, except for the TWITTER dataset; this might be related to the many ‘neutral’ instances of the TWITTER dataset.
We performed experiments for the PGGCN without position weight, aspect-specific masking, GCN layers, and POS gate, to check its impact on the performance. The result is summarized in Table 3; it shows that they contributed to performance improvement. In particular, it proves the usefulness of the POS gate, which is the contribution of this paper; modeling word and POS representations jointly helps to get better results on the ASC task. We also conducted experiments with a varied number of GCN layers, and its accuracies and F1 scores are shown in Figure 3 and Figure 4, respectively. We found that five or more GCN layers did not give any performance gain, and the best number of layers was 2.

4. Conclusions

As a solution for the ASC task, we propose a hybrid architecture of GCN-BiLSTM, namely PGGCN, with a gate mechanism that jointly exploits word representations and POS representations. We compared PGGCN with other recent models by experiments with five datasets, and showed that PGGCN outperformed the others. Amongst the many models, we found that GCN-based models outperformed others, which indicates that syntactic patterns discovered by GCN has an impact on the performance. We also observed that the models generally gave better results with greater datasets, but the models gave poor results when the dataset has many ‘neutral’ instances. The experimental results of PGGCN without the POS gate showed the effectiveness of the POS gate. We did not utilize any pre-trained language models [20,45] that have shown promising results recently, so we will investigate a way of combining our model with pre-trained context-aware language models as future work.

Author Contributions

Conceptualization, D.K. and Y.-S.J.; methodology, D.K. and Y.-S.J.; validation, D.K. and Y.-S.J.; resources, Y.K.; writing–original draft preparation, D.K.; writing—review and editing, D.K. and Y.-S.J.; funding acquisition, Y.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2020R1I1A3053015).

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pontiki, M.; Galanis, D.; Papageorgiou, H.; Androutsopoulos, I.; Manandhar, S.; AL-Smadi, M.; Al-Ayyoub, M.; Zhao, Y.; Qin, B.; Clercq, O.D.; et al. SemEval-2016 Task 5: Aspect Based Sentiment Analysis. In Proceedings of the 10th International Workshop on Semantic Evaluation, San Diego, CA, USA, 16–17 June 2016; pp. 19–30. [Google Scholar]
  2. D’Aniello, G.; Gaeta, M.; Rocca, I.L. KnowMIS-ABSA: An overview and a reference model for applications of sentiment analysis and aspect-based sentiment analysis. Artif. Intell. Rev. 2022, 55, 5543–5574. [Google Scholar] [CrossRef]
  3. Kiritchenko, S.; Zhu, X.; Cherry, C.; Mohammad, S. NRC-Canada-2014: Detecting Aspects and Sentiment in Customer Reviews. In Proceedings of the 8th International Workshop on Semantic Evaluation, Dublin, Ireland, 23–24 August 2014; pp. 437–442. [Google Scholar]
  4. Burges, C.J.C. A Tutorial on Support Vector Machines for Pattern Recognition. Data Min. Knowl. Discov. 1998, 2, 121–167. [Google Scholar] [CrossRef]
  5. Mubarok, M.S.; Adiwijaya; Aldhi, M.D. Aspect-based sentiment analysis to review products using Naïve Bayes. AIP Conf. Proc. 2017, 1867, 020060. [Google Scholar]
  6. Sarker, I.H. Deep Learning: A Comprehensive Overview on Techniques, Taxonomy, Applications and Research Directions. SN Comput. Sci. 2021, 2, 1–20. [Google Scholar] [CrossRef] [PubMed]
  7. Haykin, S.S. Neural Networks: A Comprehensive Foundation; Prentice Hall: Hoboken, NJ, USA, 1999. [Google Scholar]
  8. Medsker, L.; Jain, L.C. Recurrent Neural Networks: Design and Applications; CRC Press: Boca Raton, FL, USA, 1999. [Google Scholar]
  9. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-Based Learning Applied to Document Recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  10. Luong, T.; Pham, H.; Manning, C.D. Effective Approaches to Attention-based Neural Machine Translation. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 1412–1421. [Google Scholar]
  11. Shrestha, A.; Mahmood, A. Review of Deep Learning Algorithms and Architectures. IEEE Access 2019, 7, 53040–53065. [Google Scholar] [CrossRef]
  12. Sitaula, C.; Basnet, A.; Mainali, A.; Shahi, T.B. Deep Learning-Based Methods for Sentiment Analysis on Nepali COVID-19-Related Tweets. Comput. Intell. Neurosci. 2021, 2021, 1–11. [Google Scholar] [CrossRef] [PubMed]
  13. Kim, H.; Jeong, Y.S. Sentiment Classification Using Convolutional Neural Networks. Appl. Sci. 2019, 9, 2347. [Google Scholar] [CrossRef] [Green Version]
  14. Galassi, A.; Lippi, M.; Torroni, P. Attention in Natural Language Processing. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 4291–4308. [Google Scholar] [CrossRef] [PubMed]
  15. Wang, Y.; Huang, M.; Zhu, X.; Zhao, L. Attention-based LSTM for Aspect-level Sentiment Classification. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, Austin, TX, USA, 1–5 November 2016; pp. 606–615. [Google Scholar]
  16. Hochreiter, S.; Schmidhuber, J. Long Short-term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  17. Fan, F.; Feng, Y.; Zhao, D. Multi-grained Attention Network for Aspect-Level Sentiment Classification. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 3433–3442. [Google Scholar]
  18. Xing, B.; Tsang, I.W. Understand me, if you refer to Aspect Knowledge: Knowledge-aware Gated Recurrent Memory Network. IEEE Trans. Emerg. Top. Comput. Intell. 2022, 6, 1092–1102. [Google Scholar] [CrossRef]
  19. Huang, L.; Sun, X.; Li, S.; Zhang, L.; Wang, H. Syntax-Aware Graph Attention Network for Aspect-Level Sentiment Classification. In Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain, 8–13 December 2020; pp. 799–810. [Google Scholar]
  20. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  21. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient Estimation of Word Representations in Vector Space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
  22. Pennington, J.; Socher, R.; Manning, C. GloVe: Global Vectors for Word Representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
  23. Rae, J.W.; Borgeaud, S.; Cai, T.; Millican, K.; Hoffmann, J.; Song, F.; Aslanides, J.; Henderson, S.; Ring, R.; Young, S.; et al. Scaling Language Models: Methods, Analysis & Insights from Training Gopher. arXiv 2021, arXiv:2112.11446. [Google Scholar]
  24. Zhang, C.; Li, Q.; Song, D. Aspect-based Sentiment Classification with Aspect-specific Graph Convolutional Networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, Hong Kong, China, 3–7 November 2019; pp. 4568–4578. [Google Scholar]
  25. Xiao, L.; Hu, X.; Hu, Y.; Xue, Y.; Gu, D.; Chen, B.; Zhang, T. Targeted Sentiment Classification Based on Attentional Encoding and Graph Convolutional Networks. Appl. Sci. 2020, 10, 957. [Google Scholar] [CrossRef] [Green Version]
  26. Bai, X.; Liu, P.; Zhang, Y. Exploiting Typed Syntactic Dependencies for Targeted Sentiment Classification Using Graph Attention Neural Network. arXiv 2020, arXiv:2002.09685. [Google Scholar]
  27. Liu, Q.; Ling, Z.H.; Jiang, H.; Hu, Y. Part-of-Speech Relevance Weights for Learning Word Embeddings. arXiv 2016, arXiv:1603.07695. [Google Scholar]
  28. Smith, A.; de Lhoneux, M.; Stymne, S.; Nivre, J. An Investigation of the Interactions Between Pre-Trained Word Embeddings, Character Models and POS Tags in Dependency Parsing. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 2711–2720. [Google Scholar]
  29. Bird, S.; Loper, E. NLTK: The Natural Language Toolkit. In Proceedings of the ACL Interactive Poster and Demonstration Sessions, Barcelona, Spain, 21–26 July 2004; pp. 214–217. [Google Scholar]
  30. Honnibal, M.; Montani, I. spaCy 2: Natural language understanding with Bloom embeddings, convolutional neural networks and incremental parsing. To Appear 2017, 7, 411–420. [Google Scholar]
  31. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  32. Li, X.; Bing, L.; Lam, W.; Shi, B. Transformation Networks for Target-Oriented Sentiment Classification. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia, 15–20 July 2018; Volume 1, pp. 946–956. [Google Scholar]
  33. Tang, D.; Qin, B.; Liu, T. Aspect Level Sentiment Classification with Deep Memory Network. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, Austin, TX, USA, 1–5 November 2016; pp. 214–224. [Google Scholar]
  34. Chen, P.; Sun, Z.; Bing, L.; Yang, W. Recurrent Attention Network on Memory for Aspect Sentiment Analysis. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; pp. 452–461. [Google Scholar]
  35. Ng, A.Y. Feature selection, L1 vs. L2 regularization, and rotational invariance. In Proceedings of the Twenty-First International Conference on Machine Learning, Banff, AB, Canada, 4–8 July 2004. [Google Scholar]
  36. Pontiki, M.; Galanis, D.; Pavlopoulos, J.; Papageorgiou, H.; Androutsopoulos, I.; Manandhar, S. SemEval-2014 Task 4: Aspect Based Sentiment Analysis. In Proceedings of the 8th International Workshop on Semantic Evaluation, Dublin, Ireland, 23–24 August 2014; pp. 27–35. [Google Scholar]
  37. Pontiki, M.; Galanis, D.; Papageorgiou, H.; Manandhar, S.; Androutsopoulos, I. SemEval-2015 Task 12: Aspect Based Sentiment Analysis. In Proceedings of the 9th International Workshop on Semantic Evaluation, Denver, CO, USA, 4–5 June 2015; pp. 486–495. [Google Scholar]
  38. Dong, L.; Wei, F.; Tan, C.; Tang, D.; Zhou, M.; Xu, K. Adaptive Recursive Neural Network for Target-dependent Twitter Sentiment Classification. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, Baltimore, ML, USA, 22–27 June, 2014; Volume 2, pp. 49–54. [Google Scholar]
  39. Hand, D.; Christen, P. A note on using the F-measure for evaluating record linkage algorithms. Stat. Comput. 2018, 28, 539–547. [Google Scholar] [CrossRef] [Green Version]
  40. Gasper, K.; Spencer, L.A.; Hu, D. Does neutral affect exist? How challenging three beliefs about neutral affect can advance affective research. Front. Psychol. 2019, 10, 2476. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Tang, D.; Qin, B.; Feng, X.; Liu, T. Effective LSTMs for Target-Dependent Sentiment Classification. In Proceedings of the 26th International Conference on Computational Linguistics: Technical Papers, Osaka, Japan, 11–16 December 2016; pp. 3298–3307. [Google Scholar]
  42. Huang, B.; Ou, Y.; Carley, K.M. Aspect Level Sentiment Classification with Attention-over-Attention Neural Networks. In Proceedings of the Social, Cultural, and Behavioral Modeling, Washington, DC, USA, 10–13 July 2018; pp. 197–206. [Google Scholar]
  43. Ma, D.; Li, S.; Zhang, X.; Wang, H. Interactive Attention Networks for Aspect-Level Sentiment Classification. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 4068–4074. [Google Scholar]
  44. Xiao, L.; Hu, X.; Chen, Y.; Xue, Y.; Chen, B.; Gu, D.; Tang, B. Multi-head self-attention based gated graph convolutional networks for aspect-based sentiment classification. Multimed. Tools Appl. 2020, 81, 19051–19070. [Google Scholar] [CrossRef]
  45. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models are Unsupervised Multitask Learners. arXiv 2020, arXiv:2005.14165. [Google Scholar]
Figure 1. Example of aspect-based sentiment analysis.
Figure 1. Example of aspect-based sentiment analysis.
Applsci 12 10134 g001
Figure 2. The framework of the proposed model. Word and POS sequences are firstly embedded by embedding layers, and the embedding vectors are processed by two BiLSTM models. Multi-layered GCN and attention mechanism incorporate syntactic patterns and POS tags to generate final representation.
Figure 2. The framework of the proposed model. Word and POS sequences are firstly embedded by embedding layers, and the embedding vectors are processed by two BiLSTM models. Multi-layered GCN and attention mechanism incorporate syntactic patterns and POS tags to generate final representation.
Applsci 12 10134 g002
Figure 3. Accuracies with varied number of GCN layers and different datasets, where horizontal axis indicates the number of GCN layers, and the vertical axis represents accuracy.
Figure 3. Accuracies with varied number of GCN layers and different datasets, where horizontal axis indicates the number of GCN layers, and the vertical axis represents accuracy.
Applsci 12 10134 g003
Figure 4. F1 scores with varied number of GCN layers and different datasets, where horizontal axis indicates the number of GCN layers and the vertical axis represents F1 score.
Figure 4. F1 scores with varied number of GCN layers and different datasets, where horizontal axis indicates the number of GCN layers and the vertical axis represents F1 score.
Applsci 12 10134 g004
Table 1. Statistics of the datasets, where m / n indicates the number of training instances m and the number of test instances n.
Table 1. Statistics of the datasets, where m / n indicates the number of training instances m and the number of test instances n.
DatasetPositiveNeutralNegative
REST142164/728637/196807/196
REST151178/43950/35382/328
REST161620/59788/38709/190
LAPTOP994/341464/169870/128
TWITTER1561/1733127/3461560/173
Table 2. Averaged accuracies and F1 scores with five datasets (e.g., REST14, REST15, REST16, LAPTOP, and TWITTER), where results with ♮ are borrowed from [24]. The best results for each dataset are in bold. The GCN-based models generally gave better performance than the others, and our proposed PGGCN achieved the best performance.
Table 2. Averaged accuracies and F1 scores with five datasets (e.g., REST14, REST15, REST16, LAPTOP, and TWITTER), where results with ♮ are borrowed from [24]. The best results for each dataset are in bold. The GCN-based models generally gave better performance than the others, and our proposed PGGCN achieved the best performance.
MODELREST14REST15REST16LAPTOPTWITTER
Acc.F1Acc.F1Acc.F1Acc.F1Acc.F1
TD-LSTM [41]78.0068.4376.3958.7082.1654.2171.8068.4669.8966.21
ATAE-LSTM [15]78.6067.0278.4862.8483.7761.7168.8863.9370.1466.03
MemNet [33]79.6169.6477.3158.2885.4465.9970.6465.1771.4869.90
AOA [42]79.9770.4278.1757.0287.5066.2172.6267.5272.3070.20
IAN [43]79.2670.0978.5452.6584.7455.2172.0567.3872.5070.81
TNet-LF [32]80.4271.0378.4759.4789.0770.4374.6170.1472.9871.43
ASGCN [24]80.7772.0279.8961.8988.9967.4875.5571.0572.1570.40
AEGCN [25]81.0471.3279.9560.8787.3968.2275.9171.6373.1671.82
MGGCN [44]81.1671.7380.1964.6288.9669.4875.8071.7573.4171.89
PGGCN83.8476.8082.4766.6490.4274.4977.7474.5674.5772.01
Table 3. Averaged accuracies and F1 scores of PGGCN without some components, where ‘w/o POS gate’, ‘w/o position’, ‘w/o mask’, and ‘w/o GCN’ indicate PGGCN without POS gate, without position weights, without aspect-specific masking, and without GCN layers, respectively. The PGGCN without any component gave the worst results, which implies that all components including the POS gate are important.
Table 3. Averaged accuracies and F1 scores of PGGCN without some components, where ‘w/o POS gate’, ‘w/o position’, ‘w/o mask’, and ‘w/o GCN’ indicate PGGCN without POS gate, without position weights, without aspect-specific masking, and without GCN layers, respectively. The PGGCN without any component gave the worst results, which implies that all components including the POS gate are important.
MODELREST14REST15REST16LAPTOPTWITTER
Acc.F1Acc.F1Acc.F1Acc.F1Acc.F1
w/o POS gate83.3976.3780.8165.1489.4573.7975.8672.2673.1271.62
w/o position82.6874.8280.8164.3089.4571.1175.3971.0773.8472.29
w/o mask79.2969.8578.0463.5287.5064.9972.4167.8872.4071.13
w/o GCN81.2572.4781.0063.3487.8269.6774.6170.6972.4071.30
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, D.; Kim, Y.; Jeong, Y.-S. Graph Convolutional Networks with POS Gate for Aspect-Based Sentiment Analysis. Appl. Sci. 2022, 12, 10134. https://doi.org/10.3390/app121910134

AMA Style

Kim D, Kim Y, Jeong Y-S. Graph Convolutional Networks with POS Gate for Aspect-Based Sentiment Analysis. Applied Sciences. 2022; 12(19):10134. https://doi.org/10.3390/app121910134

Chicago/Turabian Style

Kim, Dahye, YoungJin Kim, and Young-Seob Jeong. 2022. "Graph Convolutional Networks with POS Gate for Aspect-Based Sentiment Analysis" Applied Sciences 12, no. 19: 10134. https://doi.org/10.3390/app121910134

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop