Educational QA System-Oriented Answer Selection Model Based on Focus Fusion of Multi-Perspective Word Matching
Abstract
1. Introduction
2. Related Work
3. Model Architecture
Algorithm 1: Pseudo-code for implementing the FMWMF |
Input: train set , test set , hyperparameters of each matching model , , , and the epoch e. |
1: , , < -Initialize model parameters |
2: #Step 1: train matching models respectively |
3: for epoch e do |
4: for do |
5: ; |
6: ; |
7: ; |
8: Update , , by backpropagation |
9: end for |
10: ; |
11: ; |
12: ; |
13: ; |
14: ; |
15: ; |
16: ; |
17: end for |
18: Save , , ; |
19: #Step 2: Utilize each matching model to extract the focus distribution of words in a sentence, and output the question–answer matching score. |
20: ; |
21: ; |
22: |
23: Output s |
3.1. Serial Matching Structure
3.2. Parallel Matching Structure
3.3. Transformational Matching Structure
3.4. Information Extraction and Fusion
4. Experiments and Results Analysis
4.1. Experimental Dataset and Evaluation Metrics
4.2. Experimental Environment
4.3. Comparison of Matching Focus Distribution Results Under Different Question–Answer Matching Models
4.4. The Impact of the Number of Key Information K on the Effectiveness of Answer Selection
4.5. Comparison of the Effect of Different Answer Selection Models
4.5.1. Comparison of the Classical Answer Selection Models
4.5.2. Comparison of the Large Language Models
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Amancio, L.; Dorneles, C.F.; Daniel, H.; Dalip, D.H. Recency and quality-based ranking question in CQAs: A Stack Overflow case study. Inf. Process. Manag. 2021, 58, 102552. [Google Scholar] [CrossRef]
- Minaee, S.; Kalchbrenner, N.; Cambria, E. Deep Learning Based Text Classification: A Comprehensive Review. ACM Comput. Surv. 2020, 54, 1–40. [Google Scholar] [CrossRef]
- Xu, J.; Zhang, W. Study of Intelligent Question Answering System Based on Ontology. In Proceedings of the 2020 IEEE 11th International Conference on Software Engineering and Service Science (ICSESS), Beijing, China, 16–18 October 2020; pp. 428–431. [Google Scholar] [CrossRef]
- Wang, S.; Jiang, J. A compare-aggregate model for matching text sequences. arXiv 2016, arXiv:1611.01747. [Google Scholar] [CrossRef]
- Matthew, E.P.; Mark, N.; Mohit, I.; Matt, G. Deep Contextualized Word Representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, New Orleans, LA, USA, 2–4 June 2018; Volume 1, pp. 2227–2237. [Google Scholar] [CrossRef]
- Jacob, D.; Chang, M.-W.; Kenton, L. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–7 June 2019; Volume 1, pp. 4171–4186. [Google Scholar] [CrossRef]
- Alec, R.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving Language Understanding by Generative Pre-Training. Available online: https://www.mikecaptain.com/resources/pdf/GPT-1.pdf (accessed on 25 August 2025).
- Tran, N.K.; Claudia, N. Multihop attention networks for question answer matching. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval (SIGIR ‘18), Ann Arbor, MI, USA, 8–12 July 2018; Association for Computing Machinery: New York, NY, USA, 2018; pp. 325–334. [Google Scholar] [CrossRef]
- Yang, Z.; Wang, Y.; Gan, J. Design and research of intelligent question-answering (Q&A) system based on high school course knowledge graph. Mob. Netw. Appl. 2021, 26, 1884–1890. [Google Scholar] [CrossRef]
- Qian, Y. The Semantic Framework of Library Intelligent Question Answering System Based on Exploratory Search Behavior. In Proceedings of the 2022 IEEE 2nd International Conference on Computer Communication and Artificial Intelligence (CCAI), Beijing, China, 6–8 May 2022; pp. 65–70. [Google Scholar] [CrossRef]
- Wildan, A.H.; Aji, R.F. Transformer and Large Language Models for Automatic Multiple-Choice Question Generation: A Systematic Literature Review. IEEE Access 2025, 13, 127100–127112. [Google Scholar] [CrossRef]
- Hu, Z. Research and implementation of railway technical specification question answering system based on deep learning. In Proceedings of the 2020 IEEE 5th Information Technology and Mechatronics Engineering Conference (ITOEC), Chongqing, China, 12–14 June 2020; pp. 5–9. [Google Scholar] [CrossRef]
- Yu, L.; Hermann, K.M.; Blunsom, P.; Pulman, S. Deep learning for answer sentence selection. arXiv 2014, arXiv:1412.1632. [Google Scholar] [CrossRef]
- Vasin, P.; Roth, D.; Yih, W.-T. Mapping dependencies trees: An application to question answering. In Proceedings of the 8th International Symposium on Artificial Intelligence and Mathematics, Chongqing, China, 7–9 April 2023; Available online: https://api.semanticscholar.org/CorpusID:8214465 (accessed on 25 August 2025).
- Yih, S.W.-T.; Chang, M.-W.; Meek, C.; Pastusiak, A. Question answering using enhanced lexical semantic models. In Proceedings of the 51st Annual Meeting of the Association for Computational linguistics, Sofia, Bulgaria, 3 June 2013; Available online: https://www.microsoft.com/en-us/research/publication/question-answering-using-enhanced-lexical-semantic-models/ (accessed on 25 August 2025).
- Adebayo, I.M.; Kim, B.-S. Question-answering system powered by knowledge graph and generative pretrained transformer to support risk identification in tunnel projects. J. Constr. Eng. Manag. 2025, 151, 04024193. [Google Scholar] [CrossRef]
- Di, W.; Eric, N. A long short-term memory model for answer sentence selection in question answering. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Beijing, China, 26–31 July 2015; Volume 2, pp. 707–712. [Google Scholar] [CrossRef]
- He, H.; Gimpel, K.; Lin, J. Multi-perspective sentence similarity modeling with convolutional neural networks. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 1576–1586. [Google Scholar] [CrossRef]
- Wang, Z.; Hamza, W.; Florian, R. Bilateral multi-perspective matching for natural language sentences. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19 August 2017; pp. 4144–4150. [Google Scholar] [CrossRef]
- Wiwin, S.; Pratama, R.A.; Rahadika, F.Y.; Purnomo, M.H.A. Self-Attention Mechanism of RoBERTa to Improve QAS for e-health Education. In Proceedings of the 2021 4th International Conference of Computer and Informatics Engineering (IC2IE), Depok, Indonesia, 14–15 September 2021; pp. 221–225. [Google Scholar] [CrossRef]
- Guo, Y.; Zhang, J.; Chen, X.; Ji, X.; Wang, Y.-J.; Hu, Y.; Chen, J. Improving vision-language-action model with online reinforcement learning. arXiv 2025, arXiv:2501.16664. [Google Scholar] [CrossRef]
- Wang, R.; Zhang, Z.; Rossetto, L.; Ruosch, F.; Bernstein, A. NLQxform-UI: An Interactive and Intuitive Scholarly Question Answering System. In Proceedings of the 48th International ACM SIGIR Conference on Research and Development in Information Retrieval, Padua, Italy, 13–18 July 2025; pp. 3990–3993. [Google Scholar] [CrossRef]
- Clark, K.; Luong, M.-T.; Le, Q.V.; Manning, C.D. Electra: Pre-training text encoders as discriminators rather than generators. arXiv 2020, arXiv:2003.10555. [Google Scholar] [CrossRef]
- Laskar, M.T.R.; Huang, J.X.; Hoque, E. Contextualized embeddings based transformer encoder for sentence similarity modeling in answer selection task. In Proceedings of the Twelfth Language Resources and Evaluation Conference, Marseille, France, 11–16 May 2020; pp. 5505–5514. Available online: https://aclanthology.org/2020.lrec-1.676/ (accessed on 25 August 2025).
- Song, J.; Ashktorab, Z.; Pan, Q.; Dugan, C.; Geyer, W.; Malone, T.W. Interaction Configurations and Prompt Guidance in Conversational AI for Question Answering in Human-AI Teams. arXiv 2025, arXiv:2505.01648. [Google Scholar] [CrossRef]
- Zhang, Z.; Yang, J.; Zhao, H. Retrospective reader for machine reading comprehension. In Proceedings of the AAAI Conference on Artificial Intelligence, Singapore, 20–27 January 2021; Volume 35, pp. 14506–14514. [Google Scholar] [CrossRef]
- Li, L.; Zhou, A.; Zhang, B.; Xiao, F. Multiple fragment-level interactive networks for answer selection. Neurocomputing 2020, 420, 80–88. [Google Scholar] [CrossRef]
- Liu, C.; Jiang, S.; Yu, H.; Yu, D. Multi-turn Inference Matching Network for Natural Language Inference. In Proceedings of the CCF International Conference on Natural Language Processing and Chinese Computing, Hohhot, China, 26–30 August 2018; pp. 131–143. [Google Scholar] [CrossRef]
- Guo, D.; Yang, D.; Zhang, H. Deepseek-r1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning. arXiv 2025, arXiv:2501.12948. [Google Scholar] [CrossRef]
- Wang, A.; Shu, D.; Wang, Y.; Ma, Y.; Du, M. Improving LLM Reasoning through Interpretable Role-Playing Steering. arXiv 2025, arXiv:2506.07335. [Google Scholar] [CrossRef]
- Hoffer, E.; Ailon, N. Deep metric learning using triplet network. In Proceedings of the Similarity-Based Pattern Recognition: Third International Workshop, SIMBAD 2015, Copenhagen, Denmark, 12–14 October 2015; pp. 84–92. [Google Scholar] [CrossRef]
- Wang, M.; Smith, N.A.; Mitamura, T. What is the Jeopardy model? A quasi-synchronous grammar for QA. In Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL), Prague, Czech Republic, 28–30 June 2007; pp. 22–32. Available online: https://aclanthology.org/D07-1003/ (accessed on 25 August 2025).
- Yang, Y.; Yih, W.-T.; Meek, C. WikiQA: A Challenge Dataset for Open-Domain Question Answering. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 2013–2018. [Google Scholar] [CrossRef]
- Gao, H.; Hu, M.; Cheng, R.; Gao, T. Hierarchical ranking for answer selection. arXiv 2021, arXiv:2102.00677. [Google Scholar] [CrossRef]
- He, J.; Zhang, H.; Hu, X. Open domain answer selection model fusing double matching-focus. Comput. Eng. 2023, 49, 303–310. [Google Scholar] [CrossRef]
Dataset | TREC-QA | Wiki-QA |
---|---|---|
Train Set | 1162 | 873 |
Validate Set | 65 | 126 |
Test Set | 68 | 243 |
Question | Positive Answer | Negative Answer |
---|---|---|
How much are the Harry Potter movies worth? | The series also originated many types of tie-in merchandise, making the Harry Potter brand worth in excess of USD 15 billion. | The initial major publishers of the books were Bloomsbury in the United Kingdom and Scholastic Press in the United States. |
How deep can deep underwater drilling go? | Deepwater drilling is the process of oil and gas exploration and production at depths of more than 500 feet. | It has been economically infeasible for many years, but with rising oil prices, more companies are investing in this area. |
Dataset | Answer Type | Relevance |
---|---|---|
TREC-QA | Positive | 0.7929 |
Negative | −0.1926 | |
Wiki-QA | Positive | 0.6322 |
Negative | 0.2437 |
Model | MAP | MRR |
---|---|---|
Study [8] | 0.8130 | 0.8930 |
Study [17] | 0.7134 | 0.7913 |
Study [24] | 0.8910 | 0.9250 |
Study [35] | 0.8420 | 0.9040 |
SP | 0.8497 | 0.8738 |
ST | 0.9062 | 0.9246 |
TP | 0.8805 | 0.9027 |
SPT(FMWMF) | 0.9273 | 0.9401 |
Model | MAP | MRR |
---|---|---|
Study [8] | 0.7220 | 0.7380 |
Study [24] | 0.8290 | 0.8430 |
Study [32] | 0.6520 | 0.6652 |
Study [34] | 0.7420 | 0.7540 |
SP | 0.7934 | 0.7977 |
ST | 0.8306 | 0.8361 |
TP | 0.8147 | 0.8213 |
SPT(FMWMF) | 0.8773 | 0.8840 |
Subject | DeepSeek-R1 (16*H20) | Qwen_MoE-think (8*H20) | Qwen2.5–1.5B (1*H20) | FMWMF (1*H20) | ||||||
---|---|---|---|---|---|---|---|---|---|---|
acc | Time | acc | Time | acc | Time | acc | Time | K (Question) | K (Answer) | |
Biology | 0.86 | 40 min 21 s | 0.92 | 30 min 36 s | 0.2193 | 3 min 50 s | 0.67 | 67 s | 2 | 4 |
Business | 0.8 | 39 min 07 s | 0.83 | 26 min 48 s | 0.25 | 3 min 07 s | 0.73 | 62 s | 2 | 5 |
Chemistry | 0.76 | 55 min 00 s | 0.86 | 39 min 54 s | 0.28 | 5 min 27 s | 0.27 | 55 s | 4 | 7 |
Computer Science | 0.79 | 23 min 09 s | 0.81 | 21 min 56 s | 0.11 | 2 min 34 s | 0.76 | 68 s | 2 | 4 |
Economics | 0.85 | 35 min 13 s | 0.87 | 25 min 41 s | 0.31 | 3 min 27 s | 0.64 | 77 s | 2 | 4 |
Engineering | 0.64 | 19 min 18 s | 0.71 | 15 min 00 s | 0.26 | 1 min 45 s | 0.68 | 59 s | 2 | 4 |
Health | 0.69 | 12 min 04 s | 0.69 | 10 min 30 s | 0.19 | 1 min 11 s | 0.58 | 60 s | 2 | 4 |
History | 0.67 | 36 min 17 s | 0.68 | 25 min 54 s | 0.21 | 3 min 30 s | 0.61 | 78 s | 3 | 6 |
Law | 0.59 | 38 min 05 s | 0.65 | 29 min 42 s | 0.13 | 3 min 40 s | 0.72 | 93 s | 2 | 4 |
Math | 0.92 | 34 min 48 s | 0.9 | 27 min 45 s | 0.13 | 3 min 38 s | 0.15 | 51 s | 2 | 2 |
Other | 0.75 | 12 min 11 s | 0.65 | 9 min 51 s | 0.34 | 1 min 09 s | 0.70 | 61 s | 2 | 4 |
Philosophy | 0.81 | 28 min 35 s | 0.7 | 22 min 19 s | 0.14 | 2 min 55 s | 0.81 | 84 s | 3 | 6 |
Physics | 0.87 | 34 min 19 s | 0.9 | 29 min 23 s | 0.24 | 3 min 35 s | 0.22 | 56 s | 2 | 6 |
Psychology | 0.72 | 11 min 26 s | 0.79 | 9 min 59 s | 0.19 | 1 min 10 s | 0.75 | 63 s | 3 | 6 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Hu, X.; He, J.; Shou, Z.; Liu, Z.; Zhang, H. Educational QA System-Oriented Answer Selection Model Based on Focus Fusion of Multi-Perspective Word Matching. Computers 2025, 14, 399. https://doi.org/10.3390/computers14090399
Hu X, He J, Shou Z, Liu Z, Zhang H. Educational QA System-Oriented Answer Selection Model Based on Focus Fusion of Multi-Perspective Word Matching. Computers. 2025; 14(9):399. https://doi.org/10.3390/computers14090399
Chicago/Turabian StyleHu, Xiaoli, Junfei He, Zhaoyu Shou, Ziming Liu, and Huibing Zhang. 2025. "Educational QA System-Oriented Answer Selection Model Based on Focus Fusion of Multi-Perspective Word Matching" Computers 14, no. 9: 399. https://doi.org/10.3390/computers14090399
APA StyleHu, X., He, J., Shou, Z., Liu, Z., & Zhang, H. (2025). Educational QA System-Oriented Answer Selection Model Based on Focus Fusion of Multi-Perspective Word Matching. Computers, 14(9), 399. https://doi.org/10.3390/computers14090399