Target-Oriented Opinion Words Extraction Based on Dependency Tree
Abstract
1. Introduction
- We propose a novel neural network model, DTOWE, designed to address the TOWE task. This model is adept at processing complex and lengthy sentences.
- Additionally, we introduce a dependency-tree-based local context focusing mechanism, known as DT-LCF, along with two of its implementations: DT-LCF-Mask and DT-LCF-weight. These mechanisms effectively capture the opinion span of the targets under consideration.
- Experimental results across four datasets demonstrate that our model outperforms state-of-the-art models. Furthermore, an extensive analysis confirms the efficacy of our approach.
2. Related Works
3. Proposed Model
3.1. Task Formulation
3.2. Framework
3.3. Global Context Feature Extraction
3.3.1. Inward-LSTM
Algorithm 1: Inward-LSTM for the ith layer. |
//LSTM calculation |
3.3.2. Outward-LSTM
Algorithm 2: Outward-LSTM for the ith Layer. |
3.3.3. IO-LSTM
3.4. Local Context Feature Extraction
3.4.1. BiLSTM
3.4.2. DT-LCF
- DT-LCF-Mask
Algorithm 3: DT-LCF-mask |
at each position |
- DT-LCF-Weight
Algorithm 4: DT-LCF-weight |
at each position |
3.4.3. Attention Mechanism
3.5. Feature Fusion
3.6. Decoding and Training
4. Experimental Results and Analysis
4.1. Experiment Environments
4.2. Data Sets
4.3. Settings
4.4. Evaluation Metrics
4.5. Compared Methods
- Distance-rule [9]. This is an early model that relies on distance rules and part-of-speech (POS) tags, selecting the nearest adjective words to the opinion target as the corresponding opinion words.
- Dependency-rule [24]. This model employs dependency-tree-based templates to identify opinion pairs. It records the part-of-speech (POS) tags of the opinion targets and opinion words, as well as the dependency path between them in the training set, as rule templates. The dependency templates that occur with high frequency are utilized for detecting the corresponding opinion words in the testing set.
- LSTM/BiLSTM [25]. This model is designed for extracting aspect–opinion pairs at the sentence level. It begins by initializing with word embeddings within the sentence, which are then fed through an LSTM/BiLSTM layer to create sentence representations. The softmax layer classifies the hidden state at each position into three categories.
- Pipeline [6]. This is an aspect-level opinion extraction model. It combines BiLSTM with distance rules to select the nearest adjectives as the target opinion words.
- Target-Concatenated BiLSTM (TC-BiLSTM) [26]. This method integrates target information into sentences through concatenation. A target vector is derived from the average pooling of target word embeddings. The word representation at each position is the concatenation of the word embedding and the target vector, which is subsequently fed into a BiLSTM for sequence labeling.
- PE-BiLSTM [14]. This model integrates position embeddings into the TOWE task, and then extracts opinion words through BiLSTM.
- IOG [6]. This is the first work to propose the TOWE task. It utilizes six distinct positional and directional LSTMs to extract opinion words. This model achieves state-of-the-art performance.
4.6. Results and Discussion
4.7. Ablation Study
- DTOWE-without-Local: denotes that we remove the Local Module and use only the Global Module.
- DTOWE-without-DT-LCF: denotes that we remove the dependency-tree-based local context focusing mechanism.
- DTOWE-Mask-without-Attention: denotes that we remove the attention mechanism when we use DT-LCF-Mask.
- DTOWE-Weight-without-Attention: denotes that we remove the attention mechanism when we use DT-LCF-Weight.
- DTOWE-Literal-Distance: denotes that we remove DT-LCF module and use the literal distance instead of the semantic distance.
4.8. Error Propagation for Dependency Tree
4.9. Hyper-Parameter Sensitivity on Different Datasets
4.10. Training Complexity
4.11. Cross-Domain Performance Analysis
4.12. Case Study
5. Conclusions and Future Works
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Huang, B.; Guo, R.; Zhu, Y.; Fang, Z.; Zeng, G.; Liu, J.; Wang, Y.; Fujita, H.; Shi, Z. Aspect-level sentiment analysis with aspect-specific context position information. Knowl. Based Syst. 2022, 243, 108473. [Google Scholar] [CrossRef]
- Jiang, L.; Li, Y.; Liao, J.; Zou, Z.; Jiang, C. Research on non-dependent aspect-level sentiment analysis. Knowl. Based Syst. 2023, 266, 110419. [Google Scholar] [CrossRef]
- Shang, W.; Chai, J.; Cao, J.; Lei, X.; Zhu, H.; Fan, Y.; Ding, W. Aspect-level sentiment analysis based on aspect-sentence graph convolution network. Inf. Fusion 2024, 104, 102143. [Google Scholar] [CrossRef]
- Liu, N.; Zhao, J. A BERT-based aspect-level sentiment analysis algorithm for cross-domain text. Comput. Intell. Neurosci. 2022, 2022, 8726621. [Google Scholar] [CrossRef] [PubMed]
- Peng, H.; Xu, L.; Bing, L.; Huang, F.; Lu, W.; Si, L. Knowing what, how and why: A near complete solution for aspect-based sentiment analysis. Proc. AAAI Conf. Artif. Intell. 2020, 34, 8600–8607. [Google Scholar] [CrossRef]
- Fan, Z.; Wu, Z.; Dai, X.; Huang, S.; Chen, J. Target-oriented opinion words extraction with target-fused neural sequence labelling. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–6 June 2019; Volume 1, pp. 2509–2518. [Google Scholar]
- Yang, H.; Zeng, B.; Yang, J.; Song, W.; Xu, R. A multi-task learning model for China-oriented aspect polarity classification and aspect term extraction. Neurocomputing 2021, 419, 344–356. [Google Scholar] [CrossRef]
- Liu, K.; Xu, H.L.; Liu, Y.; Zhao, J. Opinion target extraction using partially-supervised word alignment model. In Proceedings of the 23rd International Joint Conference on Artificial Intelligence, Beijing, China, 3–9 August 2013. [Google Scholar]
- Hu, M.; Liu, B. Mining and summarizing customer reviews. In Proceedings of the tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Seattle, WA, USA, 22–25 August 2004; pp. 168–177. [Google Scholar]
- Qiu, G.; Liu, B.; Bu, J.; Chen, C. Opinion word expansion and target extraction through double propagation. Comput. Linguist. 2011, 37, 9–27. [Google Scholar] [CrossRef]
- Li, X.; Lam, W. Deep multi-task learning for aspect term extraction with memory interaction. In Proceedings of the 2017 Con-Ference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 9–11 September 2017; pp. 2886–2892. [Google Scholar]
- Li, X.; Bing, L.; Li, P.; Lam, W.; Yang, Z. Aspect term extraction with history attention and selective transformation. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI), Stockholm, Sweden, 12–19 July 2018. [Google Scholar]
- Wang, Y.; Huang, M.; Zhu, X.; Zhao, L. Attention-based LSTM for aspect-level sentiment classification. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, Austin, TX, USA, 1–5 November 2016; pp. 606–615. [Google Scholar]
- Wu, Z.; Zhao, F.; Dai, X.Y.; Huang, S.; Chen, J. Latent opinions transfer network for target-oriented opinion words extraction. Proc. AAAI Conf. Artif. Intell. 2020, 34, 9298–9305. [Google Scholar] [CrossRef]
- Sun, K.; Zhang, R.; Mensah, S.; Mao, Y.; Liu, X. Aspect-level sentiment analysis via convolution over dependency tree. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; pp. 5679–5688. [Google Scholar]
- Zhang, C.; Li, Q.; Song, D. Aspect Sentiment Pair Extraction with Dependency-Enhanced Graph Convolutional Net-works. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics, Dublin, Ireland, 22–27 May 2022; pp. 1234–1245. [Google Scholar]
- Veyseh, A.P.B.; Nouri, N.; Dernoncourt, F.; Dou, D.; Nguyen, T.H. Introducing syntactic structures into target opinion word extraction with deep learning. arXiv 2020, arXiv:2010.13378. [Google Scholar]
- Pontiki, M.; Galanis, D.; Pavlopoulos, J.; Papageorgiou, H.; Androutsopoulos, I.; Manandhar, S. Semeval-2014 task 4: Aspect based sentiment analysis. In Proceedings of the 8th International Workshop on Semantic Evaluation, Dublin, Ireland, 23–24 August 2014; pp. 27–35. [Google Scholar]
- Pontiki, M.; Galanis, D.; Papageorgiou, H.; Manandhar, S.; Androutsopoulos, I. Semeval-2015 task 12: Aspect based sentiment analysis. In Proceedings of the 9th international workshop on semantic evaluation (SemEval 2015), Denver, CO, USA, 4–5 June 2015; pp. 486–495. [Google Scholar]
- Pontiki, M.; Galanis, D.; Papageorgiou, H.; Androutsopoulos, I.; Manandhar, S.; Al-Smadi, M.; Al-Ayyoub, M.; Zhao, Y.; Qin, B.; Clercq, O.D.; et al. SemEval-2016 Task 5: Aspect Based Sentiment Analysis. In Proceedings of the 12th International Workshop on Semantic Evaluation, New Orleans, LA, USA, 5–6 June 2018. [Google Scholar]
- Pennington, J.; Socher, R.; Manning, C.D. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar]
- Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
- Manning, C.D.; Surdeanu, M.; Bauer, J.; Finkel, J.R.; Bethard, S.; McClosky, D. The Stanford CoreNLP natural language processing toolkit. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations, Baltimore, MA, USA, 22–27 June 2014; pp. 55–60. [Google Scholar]
- Zhuang, L.; Jing, F.; Zhu, X.Y. Movie review mining and summarization. In Proceedings of the 15th ACM International Confer-Ence on Information and Knowledge Management, Arlington, VA, USA, 6–11 November 2006; pp. 43–50. [Google Scholar]
- Liu, P.; Joty, S.; Meng, H. Fine-grained Opinion Mining with Recurrent Neural Networks and Word Embeddings. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015. [Google Scholar]
- Tang, D.; Qin, B.; Feng, X.; Liu, T. Effective LSTMs for target-dependent sentiment classification. arXiv 2015, arXiv:1512.01100. [Google Scholar]
Their/O twist/O on/O pizza/O is/O healthy/B, but/O full/B of/I flavor/I ./O |
We/O ordered/O the/O special/B branzino/O at/O a/O posh/O Michelin/O restaurant/O ,/O that/O was/O so/O infused/O with/O bone/O ,/O it/O was/O unpalatable /B ./O |
We/O ordered/O the/O special/O branzino/O at/O a/O posh/B Michelin/O restaurant/O ,/O that/O was/O so/O infused/O with/O bone/O ,/O it/O was/O unpalatable /O ./O |
Software | |
---|---|
Operating System | Linux version 3.10.0-1127.19.1el7.x86_64 |
Programming language | Python 3.7 |
Framework | pytorch 1.3.1 |
CUDA version | 11.0 |
Hardware | |
---|---|
CPU | Intel(R) Xeon(R) CPU E5-2690 v4 @2.60GHz |
GPU | NVIDIA Tesla K80 |
Memory | 32 × 16, DDR4, 2400MT/s |
Data Sets | #Sentences | #Target | Average Sentence Length | Maximum Sentence Length | |
---|---|---|---|---|---|
14res | Training | 1627 | 2643 | 15.20 | 74 |
Testing | 500 | 865 | |||
14lap | Training | 1158 | 1634 | 16.85 | 80 |
Testing | 343 | 482 | |||
15res | Training | 754 | 1076 | 13.46 | 59 |
Testing | 325 | 436 | |||
16res | Training | 1079 | 1512 | 13.00 | 59 |
Testing | 329 | 457 |
Data Sets | RLD (Opinion Words, Target Words) | RLD (Non-Opinion Words, Target Words) | RSD (Opinion Words, Target Words) | RSD (Non-Opinion Words, Target Words) |
---|---|---|---|---|
14res | 0.2294 | 0.3749 | 0.2403 | 0.4983 |
14lap | 0.2055 | 0.4081 | 0.2212 | 0.5139 |
15res | 0.2167 | 0.3945 | 0.2346 | 0.4889 |
16res | 0.2323 | 0.4272 | 0.2012 | 0.5298 |
Models | 14lap | 14res | 15res | 16res | ||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
P | R | F1 | P | R | F1 | P | R | F1 | P | R | F1 | |
Distance-rule | 50.13 | 33.86 | 40.42 | 58.39 | 43.59 | 49.92 | 54.12 | 39.96 | 45.97 | 61.90 | 44.57 | 51.83 |
Dependence-rule | 45.09 | 31.57 | 37.14 | 64.57 | 52.72 | 58.04 | 65.49 | 48.88 | 55.98 | 76.03 | 56.19 | 64.62 |
LSTM | 55.71 | 57.53 | 56.52 | 52.64 | 65.47 | 58.34 | 57.27 | 60.69 | 58.93 | 62.46 | 58.72 | 65.33 |
BiLSTM | 64.52 | 61.45 | 62.71 | 58.34 | 61.73 | 59.95 | 60.64 | 63.65 | 62.00 | 68.68 | 70.51 | 69.57 |
Pipeline | 72.58 | 56.97 | 63.83 | 77.72 | 62.33 | 69.18 | 74.75 | 56.97 | 66.97 | 81.46 | 67.81 | 74.01 |
TC-BiLSTM | 62.45 | 60.14 | 61.21 | 67.65 | 67.67 | 67.61 | 66.06 | 60.16 | 61.21 | 73.46 | 72.88 | 73.10 |
PE- BiLSTM | 72.01 | 64.20 | 67.83 | 80.10 | 76.51 | 78.26 | 70.36 | 65.73 | 67.96 | 82.27 | 74.95 | 78.43 |
IOG | 73.43 | 68.74 | 70.99 | 82.38 | 78.25 | 80.23 | 72.19 | 71.76 | 71.91 | 84.36 | 79.08 | 81.60 |
DTOWE-Mask | 77.17 | 69.14 | 72.93 | 84.68 | 78.35 | 81.39 | 77.95 | 72.41 | 75.08 | 83.65 | 82.86 | 83.25 |
DTOWE-Weight | 77.18 | 70.37 | 73.62 | 83.35 | 81.17 | 82.24 | 79.50 | 71.60 | 75.35 | 84.80 | 82.86 | 83.83 |
Models | 14lap | 14res | 15res | 16res | ||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
P | R | F1 | P | R | F1 | P | R | F1 | P | R | F1 | |
DTOWE-Mask | 77.17 | 69.14 | 72.93 | 84.68 | 78.35 | 81.39 | 77.95 | 72.41 | 75.08 | 83.65 | 82.86 | 83.25 |
DTOWE-Weight | 77.18 | 70.37 | 73.62 | 83.35 | 81.17 | 82.24 | 79.50 | 71.60 | 75.35 | 84.80 | 82.86 | 83.83 |
DTOWE-without-Local | 72.90 | 66.49 | 69.51 | 82.09 | 75.44 | 78.62 | 74.71 | 67.22 | 70.47 | 83.31 | 78.67 | 80.91 |
DTOWE-without-DT-LCF | 74.44 | 67.76 | 70.83 | 83.92 | 78.02 | 80.65 | 78.49 | 67.34 | 72.49 | 83.70 | 80.19 | 81.91 |
DTOWE-Mask-without-Attention | 73.93 | 67.02 | 70.31 | 77.79 | 79.90 | 78.83 | 78.95 | 66.94 | 72.45 | 83.47 | 78.86 | 81.10 |
DTOWE-Weight-without-Attention | 73.56 | 67.72 | 70.52 | 82.14 | 76.80 | 79.38 | 78.01 | 68.36 | 72.86 | 83.23 | 80.38 | 81.78 |
DTOWE-Literal-Distance | 71.35 | 69.84 | 70.59 | 82.75 | 77.77 | 80.18 | 73.50 | 72.01 | 72.25 | 82.71 | 80.19 | 81.43 |
Correct DTOWE-Mask Prediction | Correct DTOWE-Weight Prediction | Wrong DTOWE-Mask Prediction | Wrong DTOWE-Weight Prediction | ||
---|---|---|---|---|---|
Wrong dependency trees | 20.13% | 28.18% | Correct dependency trees | 15.73% | 12.12% |
1 Epoch Time | Total Epochs | |
---|---|---|
DTOWE-Mask | 123 s | 30 |
DTOWE-Weight(r = 2) | 151 s | 40 |
DTOWE-Weight-(r = 3) | 162 s | 40 |
P | R | F1 | P | R | F1 | ||
---|---|---|---|---|---|---|---|
14res-14res | 83.35 | 81.17 | 82.24 | 14lap-14lap | 77.18 | 70.37 | 73.62 |
14lap-14res | 70.54 | 65.23 | 67.78 | 14res-14lap | 73.23 | 70.14 | 71.65 |
15res-15res | 79.50 | 71.60 | 75.35 | 16res-16res | 84.80 | 82.86 | 83.83 |
14lap-15res | 71.19 | 67.84 | 69.74 | 14lap-16res | 74.36 | 69.79 | 72.00 |
14res-15res | 81.23 | 77.64 | 79.39 | 14res-16res | 82.79 | 81.29 | 82.03 |
Sentences | Distance-Rule | BiLSTM | IOG | DTOWE-Mask | DTOWE-Weight |
---|---|---|---|---|---|
The bread is top notch as well. | top ✗ | top notch ✓ | top notch ✓ | top notch ✓ | top notch ✓ |
Great food but the service was dreadful! | Great ✓ | dreadful ✗ | Great ✓ | Great ✓ | Great ✓ |
Their twist on pizza is healthy, and very yummy. | twist ✗ | healthy ✗ | healthy, yummy ✓ | healthy, yummy ✓ | healthy, yummy ✓ |
We’re preparing tomahawk steak for the grand dinner tonight, it’s delicious | grand ✗ | grand ✗ | grand ✗ | delicious ✓ | delicious ✓ |
Spicy chicken is a common food in American Chinatowns, but it is different from back home | common ✗ | common ✗ | different ✗ | common, different ✓ | common, different ✓ |
We ordered the special branzino at a posh Michelin restaurant, that was so infused with bone, it was unpalatable. | special ✗ | posh ✗ | posh ✗ | unpalatable, special ✓ | unpalatable, special ✓ |
tech support would not fix the problem unless I bought your plan for $150 plus. |
Other than not being a fan of click pads (industry standard these days) and the lousy internal speakers, it’s hard for me to find things about this notebook I don’t like, especially considering the $350 price tag. |
I opted for the SquareTrade 3-Year Computer Accidental Protection Warranty ($1500–2000) which also support “accidents” like drops and spills that are NOT covered by AppleCare. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wen, Y.; Yu, E.; Qu, J.; Cheng, L.; Chen, Y.; Lu, S. Target-Oriented Opinion Words Extraction Based on Dependency Tree. Big Data Cogn. Comput. 2025, 9, 207. https://doi.org/10.3390/bdcc9080207
Wen Y, Yu E, Qu J, Cheng L, Chen Y, Lu S. Target-Oriented Opinion Words Extraction Based on Dependency Tree. Big Data and Cognitive Computing. 2025; 9(8):207. https://doi.org/10.3390/bdcc9080207
Chicago/Turabian StyleWen, Yan, Enhai Yu, Jiawei Qu, Lele Cheng, Yuao Chen, and Siyu Lu. 2025. "Target-Oriented Opinion Words Extraction Based on Dependency Tree" Big Data and Cognitive Computing 9, no. 8: 207. https://doi.org/10.3390/bdcc9080207
APA StyleWen, Y., Yu, E., Qu, J., Cheng, L., Chen, Y., & Lu, S. (2025). Target-Oriented Opinion Words Extraction Based on Dependency Tree. Big Data and Cognitive Computing, 9(8), 207. https://doi.org/10.3390/bdcc9080207