You are currently viewing a new version of our website. To view the old version click .
Information
  • Review
  • Open Access

1 May 2021

Ontology-Based Approach to Semantically Enhanced Question Answering for Closed Domain: A Review

and
1
Department of Computer Science, Faculty of Information and Communication Technology, International Islamic University Malaysia, Kuala Lumpur 53100, Malaysia
2
Faculty of Information and Communication Technology, International Islamic University Malaysia, Kuala Lumpur 53100, Malaysia
*
Author to whom correspondence should be addressed.
This article belongs to the Collection Natural Language Processing and Applications: Challenges and Perspectives

Abstract

For many users of natural language processing (NLP), it can be challenging to obtain concise, accurate and precise answers to a question. Systems such as question answering (QA) enable users to ask questions and receive feedback in the form of quick answers to questions posed in natural language, rather than in the form of lists of documents delivered by search engines. This task is challenging and involves complex semantic annotation and knowledge representation. This study reviews the literature detailing ontology-based methods that semantically enhance QA for a closed domain, by presenting a literature review of the relevant studies published between 2000 and 2020. The review reports that 83 of the 124 papers considered acknowledge the QA approach, and recommend its development and evaluation using different methods. These methods are evaluated according to accuracy, precision, and recall. An ontological approach to semantically enhancing QA is found to be adopted in a limited way, as many of the studies reviewed concentrated instead on NLP and information retrieval (IR) processing. While the majority of the studies reviewed focus on open domains, this study investigates the closed domain.

1. Introduction

Technological advancements in the field of information communication technology (ICT) support the conversion of the conventional web structures of internet documents to semantic web-linking data, enabling novel semantic web data representation and integration retrieval systems [1]. Globally, millions of internet searches are made every minute, and obtaining precise answers to queries can be challenging, due to the continually expanding volume of information available [2].
Information retrieval (IR) technology supports the process of searching for and retrieving information. The general method used by IR employs keywords to retrieve relevant information from different heterogeneous resources [3], thereby providing a limitation that captures an accurate conceptualization of the user’s expression and the meaning of their content. Upon searching, the user receives a list of related documents that may contain the information required. However, in some cases, users might be seeking for a specific answer, rather than a list of documents. Although IR techniques can be highly successful and retrieve relevant information, users nevertheless face complex linguistic challenges to extract the desired information. Question answering (QA) techniques can address this by enabling users to express and extract precise information and knowledge in natural language (NL).
Typically, QA systems aim to create methods with functionality that extends beyond the retrieval of appropriate texts, in order to provide the correct response to NL queries posed by a user. Answering such queries requires composite processing of documents that exceeds that which is achievable using IR systems [4]. With QA systems, answers are obtained from large document collections by grouping required answer types, to determine the precise semantic representation of the question’s keywords, which are essential to carry out further tasks such as document processing, answer extraction, and ranking. To achieve high semantic accuracy, precise analysis of the questions posed by users in NL is essential to define the intended sense in the context of the user’s question. This challenge motivated us to investigate the various ontology-based approaches and existing methodologies designed to semantically enhance QA systems.
The semantic web implements an ontological methodology intended to retain data in the form of expressions associated with the domain, as well as procedures, referencing, and comments, and properties that are automated or software agent readable, and plausible for use processing and retrieving data [5,6]. This facilitates the power of widely distributed data, by linking it in a standardized and meaningful manner so that it is available on different web servers, or on the centralized servers of an organization [7]. Researchers usually utilize measures of similarity to discover identical notions in free text notes and records. A measure of semantic similarity adopts two notions as input, and then provides a numerical score, which calculates how similar they are in terms of senses [8]. A variety of semantic similarity measures have been developed to describe the strength of relationships between concepts. These existing semantic similarity measures generally fall into one of two categories: ontology-based or corpus-based approaches (i.e., a supervised method).
Ontology-based semantic similarities usually rely on various graph-based characteristics, for instance, the shortest path length between notions, as well as the location of lowest prevalent ancestors when identifying the similarity in meaning. Such ontology-based methods rely on the quality and totality of fundamental ontologies [8]; nevertheless, maintaining and curating domain ontologies is a challenging and labour-intensive task. It is worth noting that corpus-based approaches as an alternative to identifying ontology-based semantic similarities, are grounded on the co-occurrences and distributional semantics of terms in free texts [9]. Such corpus-based models depend on the linguistic principle that the sense of a lexical term can be identified on the basis of the surrounding context.
Although corpus-based approaches can outperform ontology-based approaches, the literature has shown that the performance gap between both kinds of approaches has narrowed [8,10]. Today, some ontology-based approaches reportedly achieve better performance than the corpus-based approach. This is because corpus-based approaches require a set of sense-annotated training data, which is extremely time-consuming and costly to produce. Moreover, corpora do not exist for all domains [11,12]. In order to explore the various approaches to overcoming this challenge, this study employs the following research methodology:

2. Research Methodology

The principal aim of this paper is to briefly review existing searching methods, focusing on the role of different technologies, and highlighting the importance of the ontology-based approach in searches, as well as trying to identify the major technology gaps associated with existing methods. The following steps describe the research methodology employed.

2.1. Research Objectives

The following objectives were determined:
1. To review the various methodologies currently available in the ontology-based approach to QA systems.
2. To understand the unique features of the majority of ontology-based approaches, across QA domains.
3. To determine what methodology best suits the development of a closed domain “Pilgrimage” ontology-based QA.
4. To establish the major challenges and study gaps affecting current methods used to semantically enhance QA systems to establish future directions.

2.2. Research Questions

This review comprises four questions:
a. What are the methodologies available in the ontology-based approach to QA systems?
b. What are the unique features of current ontology-based approaches, across QA domains?
c. Which methodology best suits the development of closed domain ontology-based QA systems?
d. What are the major challenges and study gaps affecting the present methodology across QA systems in terms of future directions?

2.3. Criteria for Inclusion and Exclusion Study

To examine the evidence for ontology-based methods for semantically enhancing QA for a closed domain, this study employed a tree table (Table 1) after rapidly examining 124 articles related to ontology-based approaches located on the Web of Science (WoS), limited by date of publication (2000–2020). The data were collected from the WoS Core Collection, constituted by the Social Sciences Citation Index (SSCI), Science Citation Index Expanded (SCI-EXPANDED), Conference Proceedings Citation Index - Science (CPCI-S), Conference Proceedings Citation Index-Social Science & Humanities (CPCI-SSH), Arts & Humanities Citation Index (A&HCI), and the newly included Emerging Sources Citation Index (ESCI). None of the selected articles afforded us exact data. Therefore, we excluded information deemed irrelevant by adding inclusion and exclusion criteria. In the first phase, we remove 145 irrelevant articles out of 269. In the second phase, we selected articles based on titles, abstracts and keywords, and rejected a further 66 papers. In this phase, we excluded 45 articles because the abstract did not align with the study’s inclusion criteria. In the fourth phase, 19 were excluded for unusable data, and finally 26 were chosen for the literature review, as mentioned in Figure 1. The main categories created to classify the papers reviewed included the metadata of the paper, the problem identified, methods used, findings, future studies, techniques, and evaluation metrics (Table 1).
Table 1. Five categories created to classify the papers reviewed.
Figure 1. Criteria for Inclusion and Exclusion Study.
The remainder of this paper has been organized as follows, Section 2 states the research methodology in which the research process, research questions and criteria for the inclusion and exclusion study were given along with search strategy. Section 3 introduces related works. Section 4 concerns the challenges of the QA systems. Section 5 presents the evaluation of QA systems, and a detailed discussion is presented in Section 6. Finally, conclusions and suggestions for recommended future work are set out in Section 7.

4. The Challenges of the QA System

According to [8,78], extracting accurate information in response to a natural language question (NLQ) remains a challenge for QA systems. This is primarily because the existing techniques used by the system cannot address the effectiveness and efficiency of semantic issues in their NLQ analysis. Some scholars [11,79,80] reported positive outcomes to lexical solutions on the topic of word sense disambiguation (WSD), and several studies [42,62] appraised this issue in the context of QA, but only a few examined the role of WSD when returning potential answers [12,81].
A systematic review of NLP and text mining, conducted by [82], found that computational challenges, and synthesizing the literature concerning the use of NLP and text mining, present major challenges for QA. Meanwhile, a study conducted by [83] suggested that when real-world data are constructed and inserted into KGs, it provides an effective and convenient querying technique for end-users, a matter that requires addressing urgently. However, a greater challenge is the ability to understand the question clearly, in order to translate the unstructured question into a structured query in semantic representation form. This is due to the ambiguity associated with defining the precise word sense, which results in mismatches when mapping the expressions extracted from NLQ to the ontology elements available for the data.
Alternative systems discussed take into account QA as a semantic parsing issue and aim to train a semantic parser to plan input questions in suitable logical forms to be performed over a base of knowledge. These systems require identification of important data when training, which is time-consuming and labour intensive [62]. The QA systems based on IR engines require that text wording comprising an answer should be identical to the wording employed in the question [78].
Previous studies have revealed that existing QA systems concentrate on providing answering to factoid questions. As various sources of knowledge have advantages and disadvantages, a sensible direction to take would be one that exploits several kinds of underlying knowledge, so as to minimalize the restrictions placed on the individual. It is indisputable that the principal quantity of factual information available on the web currently is enclosed by plain text (e.g., blogs, sites, documents, etc.). On the basis of the above, a direction that is worthy of further investigation would involve manipulating peripheral plain-text knowledge to enrich the content as well as the vocabulary of the KB (to develop tasks such as disambiguation, recalling the challenge of ambiguity). Moreover, pursuing an alternative direction is exciting, involving a search for more effective and flexible approaches for interlinking and structuring plain text information with organized representations of knowledge such as RDF graphs, which attempt to resolve the difficulties associated with distributed knowledge. As well as these major challenges, the processing of complex questions and multilingualism are exceedingly challenging issues, due to the multitude of languages, each of which possesses different structures and idioms [84]. In conclusion, assessment datasets and benchmarks are fundamental for attaining similar results reliably, which is essential when making comparisons of various QA systems’ performance and the effectiveness of approaches. Thus, the assessment process represents another interesting research direction.

5. QA System Evaluation

Evaluation of the QA system method is a crucial component of QA systems, and as QA techniques are rapidly designed, a trustworthy evaluation measurement to review these applications is required. In the study by [14], the assessment measurement applied to QA systems was accuracy, and the harmonic mean function of precision and recall (F1 score). As shown in Formulas (1)–(3), a 2 × 2 possibility table enabled appreciation of these metrics, which classified evaluation under two aspects: a) true positive (TP), denoting a section that was correctly selected, or false negative (FN), denoting a section that was correctly not chosen; and b) true negative (TN), denoting a section that was incorrectly not chosen, or false positive (FP), denoting a section that was not correctly selected. Precision was used to describe the measurement of the chosen review papers that were correct, while recall referred to the opposite gauge, namely the measurement of the correct papers chosen. Using precision and recall, the fact that there was a high rate of true negative was no longer important [14].
A c c u r a c y = T P + T N T P + F P + F N + T N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
The evaluation of precision and recall involves understanding that they are contrary, and indeed there is a trade-off involved in all such measures, which means researchers must seek the best measure to calculate their particular system. Several QA systems have used recall as a gauge metric, since it does not require the evaluation of how false the positive rates are developed; if there is a high true positive rate, then the result will be improved. However, it can be argued that precision would be a more effective measure. To balance the trade-off, [14,43] introduced the ‘f’ calculation (Formula (4)).
F 1 = 2 ( P r e c i s i o n . R e c a l l ) P r e c i s i o n + R e c a l l
This metric applied a weighted score, gauging precision and recall trade-off. According to [43,85], and [14], the following metrics (Formulas (5) and (6)) can be used to calculate QA systems, and include calculations such as the mean average precision (MAP), which is a universal assessment for IR, describing the (AveP), average precision, given by Formula (5). Meanwhile, the mean reciprocal rank (MRR) (Formula (7)) can be used to calculate relevance.
M A P = Q P = 1 A v e P ( p ) Q
where AveP (Average precision) is given by the Equation (5). Mean Reciprocal Rank (MRR) is shown below, and it is used to calculate relevance.
A v e P = N k = 1 P ( k ) X r e l ( q ) r e l e v a n t
M R R = 1 N i = 1 N R R ( q i )
As shown in Table 2, the evaluation conducted for the present study produced a total of 83 review papers, of which 37 were relevant, ontology-based studies; thus, an average recall of 44.7% was achieved, and an average precision of 53.5%. It was observed that ontology-based methods/semantic web/link data yielded more precision than the QA, NLP, and other categories. This could be explained by the fact that ontology-based methods/semantic web/link data define several possible means of identifying entities, which helps when classifying the keywords for ontology entities. Surprisingly, the QA NLP returned a higher precision rate of 93.7%, while recall was 32.9%. This concurred with the claims made by [57], although, since their data were not publicly available it was not possible to compare the performance of the two systems. The results of the present study employed the same method as [57], which is an approach used for knowledge identified in the ontology to manage the user’s question, and neither approach made rigorous use of complex linguistic and semantics methods. However, some elements were unique to the present study, which for example used estimate pairing and a user interface for entity mapping. This system required the participation of the user to refine their query and the use of a set of rules to certify and improve the SPARQL query.
Table 2. Evaluation results.

6. Discussion

Reference [84] conducted a survey of the stateless QA system, with an emphasis on the methods applied in RDF and linked data, documents, and a combination of these. The author reviewed 21 recent systems and the 23 evaluation and training datasets most commonly used in the extant literature, according to domain-type. The study identified underlying knowledge sources, the tasks provided, and the associated evaluation metrics, identifying various contexts in terms of their QA system applications. Overall, QA systems were found to be widely used in different fields of human engagement, and their application spanned the fields of medicine, culture, web communities, and tourism, among others. The study also contextualized the current most efficient QA system, IBM Watson, which is a QA system that is not restricted to a specific domain, and which seeks to capture a wide range of questions. The IBM Watson competed against the former winner in the television game Jeopardy and won. However, despite the number of successes associated with the system, it possesses certain shortcomings and ambiguities.
There has recently been an increase in the adoption of QA-based personal assistants, including Amazon’s Alexa, Microsoft’s Cortona, Apple’s Siri, Samsung’s Bixby, and Google Assistant, which are capable of answering a wide range of questions [84]. Furthermore, there has been an increase in the application of QA systems at call centers and utilized in Internet Of Things (IoT) applications.
Here we discuss the classification of ontology-based approaches to QA, and present the categories created to classify the 83 papers reviewed by this study, which are associated with ontology methods that semantically enhance the QA. In order to extract and classify the data, the study extracted data at the full paper reading stage, creating five categories to classify the reviewed papers (Table 1). In total, of the 83 papers reviewed, only 42 mentioned ontology methods in QA directly, while 41 discussed QA NLP among other methods (Table 3); of the 83 papers reviewed, 42 (50%) employed an ontology-based paradigm, 32 (38%) used a QA NLP paradigm, and 9 (11%) employed a hybrid approach. There was a higher use of web-based and NLP approaches than hybrid paradigms (Figure 2), which concurred with the findings of [13].
Figure 2. Flow of Search and Selection.
Rich structured knowledge could be helpful for AI applications. Nonetheless, the way of integrating symbolic knowledge into a computational framework of practical applications remains a challenging task. Current advances in knowledge-graph-based studies concentrate on knowledge graph embedding or knowledge representation learning, which involves mapping relations as well as elements into low-dimensional vectors, while identifying their senses [86]. Knowledge-aware models obtain advantages from incorporating semantics, ontologies for the representation of knowledge, diverse data sources and multilingual knowledge. Therefore, some certifiable applications, like QA system have prospered, benefiting from their capacity for rational reasoning as well as perception. Some genuine products, for instance, Google’s Knowledge Graph and Microsoft’s Satori have demonstrated a solid ability to offer more productive types of assistance.
Table 3. Methods and techniques identified for the related literature.
Table 3. Methods and techniques identified for the related literature.
MethodsTechniquesTechnique for EvaluationRef.
Querying AgentRDF, OWL, Spatiotemporal Information representationsTwo evaluation stages: a. Air Quality Index calculation. b. Manual verification by health experts[27,47]
Single, Multiple, Hybrid OntologySemantic information description. Design, Development, Integration, Validation, Ontology IterationNo evaluation available[55]
Semantic Search SystemKB preparation, Query processing, Entity mapping, Formal query and Answer generationDisambiguation evaluation: System achieved 103 correct answers out of 133 questions. 64% recall, 76% precision[58]
Ontology. Decision Support systemKnowledge querying, Reasoning based on ontology modelFive sets of rollover stability data under different conditions used for evaluation: accuracy, effectiveness[52]
RDF, SPARQLNLP. Ontology entity mapping, Retrieve and Manipulate data in RDFQuranic ontology, Arabic Mooney Geography dataset for evaluation. System achieved 64% recall, 76% precision[52,58]
Semantic QAOntology reasoning, custom rules, Semantic QA systemCustom rules and query set for system evaluation. Result: Backward chaining ontology outperforms the in-memory reasoning system[16]
IRQuran’s concept hierarchy, vocabulary search system, Quranic WordNet, Knowledge repository, IR toolsPerformance metrics evaluation: precision, recall, F-measure. Comparisons with similar frameworks[56,59]
FrameSTEPRaw trajectories, Annotation, Semantic graphs, OntologySegmentation Granularity Extent Context Type and availability[27,47]
Ontology, Semantic Knowledge, Word2vecOWL, Word embedding, fuzzy ontologyBi-LSTM improved features extraction and text classification. Evaluation based on machine learning: SVM, CNN, RNN, LSTM. Metrics: Precision, Recall, Accuracy[48]
Intelligent Mobile AgentOntology, DBPedia, WordNet.IMAT was validated by Mobile Client Application implementation it helps testing of important IMAT features[52,53,58,59]
IRMath-aware QA system, Ask Platypus. A single mathematical formula is returned to NLQ in English or HindiMetrics: Precision, Recall, Accuracy, Function Measures[56,59]
Linked Open data FrameworkConcepts and Relation extraction, NLP, KBEvaluation shows improved result in most tasks of ontology generation compared to those obtained in existing frameworks[46]
IRRDF, OWL, SPARQLNo evaluation available[49,56,59]
Classification and MiningNLP, Linguistic features,Neural Networks capable of handling complexity, classified Hadith 94% accuracy. Mining method: Vector space model, Cosine similarity, Enriched queries—Obtained 88.9% accuracy[64]
Semantic SearchIR, Quran-based QA system, Neural Networks classificationEvaluation of classification shows approximately 90% accuracy[87]
With the ongoing progress in deep learning, and the inescapable utilization of distributional semantics to create word embedding for representation of words through vectors in neural networks, such corpus-based models have acquired tremendous prominence. One of the most widely recognized distributional semantics model is word2vec [8] for creating word embedding. The word2vec model is a neural network that is capable to enhance NLP with semantic and syntactic relationships between word vectors. While corpus-based measures have shown to be more adaptable and extensible than ontology-based measures, human curation can develop their precision further in domain applications.
Previous endeavors sought to integrate corpus-based and ontology-based similarity to well identify semantic similarities; in any case, no framework is presently available in the domain of pilgrimage delivering ontological information using a method creating word embedding for semantic similarity. Although the use of ontology for knowledge representation has captured the attention of researchers and developers as a means of overcoming the limitations of the keyword-based search, there has been little focus on the Islamic domain of knowledge, and specifically on QA for pilgrimage rituals. Those studies that do exist in this field have unique limitations, including the fact that existing religious content websites work on either language-dependent searches or keyword-based searches; the presence of only theoretical and conceptual notions of ontological retrieving agents [1]; and poor system usability, especially at the query level, that requires the user to manage complex language or interfaces to express their needs [88].
The present study addressed the research questions posed, and the ontology-based approach to QA systems identified is illustrated in Table 3, which shows that the majority of the extant studies in the field targeted a specific domain. The features unique to the majority of the works were their initial understanding of the domain (domain knowledge); the component that facilitated ontological reasoning (query processing); the mapping of the objects of interest in the domain (entities); the generation of appropriate responses, based on the query issued; and, finally, the acknowledgement that Hajj/Umrah activity is an all-year-round activity. In terms of the tasks to be performed, some are particular only to the Hajj, and in terms of sites to be visited, some are relevant to the Hajj, and others to both the Hajj and the Umrah. Several tasks occur across other domains, such as tourism, where activities and sites to be visited are essential factors. Therefore, the defined methodology with minimal modification of fuzzy rules can be implemented in such a domain.

7. Conclusions and Future Works

This study conducted a review of ontology-based approaches to semantically enhance QA for a closed domain. It found that 83 out of the 124 papers reviewed described QA approaches that had been developed and evaluated using different methods. The ontological approach to semantically enhancing QA was not found to be widely embraced, as many studies featured NLP and IR processing. Most of the studies reviewed focused on open domain usage, but this study concerns a closed domain. The precision and recall measures were primarily employed to evaluate the method used by the studies.
Due to the unique attributes of the pilgrimage exercise, the methodology best suited to the development of a pilgrimage ontology is hybrid, the first component being the creation of fuzzy rules to analyzze NLQs in order to define the exercise a pilgrim is performing at a specific location and time of year. The rules should be applied to the categories of words in the question, which will ultimately assist in generating some semantic structure for the text contained in the question. The second component is the semantic analysis, a component of the hybrid methods that will produce an understanding of the interpretation of words from the question. The generation of an ontology comprising these components, populated with dynamic facts using a KG, will help to answer questions specifically related to the pilgrimage domain.
We plan to evaluate the advancement of these segments within an extensible framework, by integrating a knowledge-based and distributional semantic model for learning word embedding from context and domain information, by using measures for semantic similarities on top of the Word2Vec algorithm. This is due to the algorithm’s capability to encode high order relationships and cover a wider range of connections between data points. It is a significant factor in enhancing several tasks, such as semantic parsing, disambiguation of an entity that results in accuracy improvement in QA systems. This requires the development of new methods that suit a collaboration between these components to enhance data processing in QA systems. Finally, we aim to develop a mobile application prototype to evaluate effectiveness and address users’ perceptions of the mobile QA application.

Author Contributions

Conceptualization, A.A. and A.S.; methodology, A.A.; software, A.A.; validation, A.A. and A.S.; formal analysis, A.A.; investigation, A.A.; resources, A.A. and A.S.; data curation, A.A.; writing—original draft preparation, A.A.; writing—review and editing, A.A. and A.S.; visualization, A.A.; supervision, A.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Authors can confirm that all relevant data are included in the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Khan, J.R.; Siddiqui, F.A.; Siddiqui, A.A.; Saeed, M.; Touheed, N. Enhanced ontological model for the Islamic Jurisprudence system. In Proceedings of the 2017 International Conference on Information and Communication Technologies (ICICT), Karachi, Pakistan, 30–31 December 2017; pp. 180–184. [Google Scholar]
  2. Siciliani, L. Question Answering over Knowledge Bases. In European Semantic Web Conference; Springer: Berlin/Heidelberg, Germany, 2018; pp. 283–293. [Google Scholar]
  3. Hu, H. A Study on Question Answering System Using Integrated Retrieval Method. Ph.D. Thesis, The University of Tokushima, Tokushima, Japan, February 2006. Available online: http://citeseerx.ist.psu.edu/viewdoc/download (accessed on 20 April 2021).
  4. Abdi, A.; Idris, N.; Ahmad, Z. QAPD: An ontology-based question answering system in the physics domain. Soft Comput. 2018, 22, 213–230. [Google Scholar] [CrossRef]
  5. Buranarach, M.; Supnithi, T.; Thein, Y.M.; Ruangrajitpakorn, T.; Rattanasawad, T.; Wongpatikaseree, K.; Lim, A.O.; Tan, Y.; Assawamakin, A. OAM: An ontology application management framework for simplifying ontology-based semantic web application development. Int. J. Softw. Eng. Knowl. Eng. 2016, 26, 115–145. [Google Scholar] [CrossRef]
  6. Vandenbussche, P.Y.; Atemezing, G.A.; Poveda-Villalón, M.; Vatant, B. Linked Open Vocabularies (LOV): A gateway to reusable semantic vocabularies on the Web. Semant. Web 2017, 8, 437–452. [Google Scholar] [CrossRef]
  7. Hastings, J.; Chepelev, L.; Willighagen, E.; Adams, N.; Steinbeck, C.; Dumontier, M. The chemical information ontology: Provenance and disambiguation for chemical data on the biological semantic web. PLoS ONE 2011, 6, e25513. [Google Scholar] [CrossRef]
  8. Jiang, S.; Wu, W.; Tomita, N.; Ganoe, C.; Hassanpour, S. Multi-Ontology Refined Embeddings (MORE): A Hybrid Multi-Ontology and Corpus-based Semantic Representation for Biomedical Concepts. arXiv 2020, arXiv:2004.06555. [Google Scholar] [CrossRef]
  9. Raganato, A.; Camacho-Collados, J.; Navigli, R. Word sense disambiguation: A unified evaluation framework and empirical comparison. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, Valencia, Spain, 3–7 April 2017; Volume 1, pp. 99–110. [Google Scholar]
  10. Chaplot, D.S.; Salakhutdinov, R. Knowledge-based word sense disambiguation using topic models. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  11. Raganato, A.; Bovi, C.D.; Navigli, R. Neural sequence learning models for word sense disambiguation. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; pp. 1156–1167. [Google Scholar]
  12. Wang, Y.; Wang, M.; Fujita, H. Word sense disambiguation: A comprehensive knowledge exploitation framework. Knowl. Based Syst. 2020, 190, 105030. [Google Scholar] [CrossRef]
  13. Soares, M.A.C.; Parreiras, F.S. A literature review on question answering techniques, paradigms and systems. J. King Saud Univ. Comput. Inf. Sci. 2020, 32, 635–646. [Google Scholar]
  14. Yao, X. Feature-Driven Question Answering with Natural Language Alignment. Ph.D. Thesis, Johns Hopkins University, Baltimore, MD, USA, July 2014. [Google Scholar]
  15. Mohasseb, A.; Bader-El-Den, M.; Cocea, M. Question categorization and classification using grammar based approach. Inf. Process. Manag. 2018, 54, 1228–1243. [Google Scholar] [CrossRef]
  16. Shi, H.; Chong, D.; Yan, G. Evaluating an optimized backward chaining ontology reasoning system with innovative custom rules. Inf. Discov. Deliv. 2018, 46, 45–56. [Google Scholar] [CrossRef]
  17. Green, B.F., Jr.; Wolf, A.K.; Chomsky, C.; Laughery, K. Baseball: An automatic question-answerer. In Western Joint IRE-AIEE-ACM Computer Conference; Association for Computing Machinery: New York, NY, USA, 1961; pp. 219–224. [Google Scholar]
  18. Woods, W.A. Progress in natural language understanding: An application to lunar geology. In Proceedings of the June 4–8, 1973, National Computer Conference and Exposition; Association for Computing Machinery: New York, NY, USA, 1973; pp. 441–450. [Google Scholar]
  19. Katz, B. From sentence processing to information access on the world wide web. In AAAI Spring Symposium on Natural Language Processing for the World Wide Web; Stanford University: Stanford, CA, USA, 1997; Volume 1, p. 997. [Google Scholar]
  20. Kwok, C.C.; Etzioni, O.; Weld, D.S. Scaling question answering to the web. In Proceedings of the 10th International Conference on World Wide Web, Hong Kong, China, 1–5 May 2001; pp. 150–161. [Google Scholar]
  21. Zheng, Z. AnswerBus question answering system. In Proceedings of the Human Language Technology Conference (HLT 2002), San Diego, CA, USA, 7 February 2002; Volume 27. [Google Scholar]
  22. Parikh, J.; Murty, M.N. Adapting question answering techniques to the web. In Proceedings of the Language Engineering Conference, Hyderabad, India, 13–15 December 2002; pp. 163–171. [Google Scholar]
  23. Bhatia, P.; Madaan, R.; Sharma, A.; Dixit, A. A comparison study of question answering systems. J. Netw. Commun. Emerg. Technol. 2015, 5, 192–198. [Google Scholar]
  24. Chung, H.; Song, Y.I.; Han, K.S.; Yoon, D.S.; Lee, J.Y.; Rim, H.C.; Kim, S.H. A practical QA system in restricted domains. In Proceedings of the Conference on Question Answering in Restricted Domains, Barcelona, Spain, 25 July 2004; pp. 39–45. [Google Scholar]
  25. Mishra, A.; Mishra, N.; Agrawal, A. Context-aware restricted geographical domain question answering system. In Proceedings of the 2010 International Conference on Computational Intelligence and Communication Networks, Bhopal, India, 26–28 November 2010; pp. 548–553. [Google Scholar]
  26. Clark, P.; Thompson, J.; Porter, B. A knowledge-based approach to question-answering. Proc. AAAI 1999, 99, 43–51. [Google Scholar]
  27. Meng, X.; Wang, F.; Xie, Y.; Song, G.; Ma, S.; Hu, S.; Bai, J.; Yang, Y. An Ontology-Driven Approach for Integrating Intelligence to Manage Human and Ecological Health Risks in the Geospatial Sensor Web. Sensors 2018, 18, 3619. [Google Scholar] [CrossRef]
  28. Yang, M.C.; Lee, D.G.; Park, S.Y.; Rim, H.C. Knowledge-based question answering using the semantic embedding space. Expert Syst. Appl. 2015, 42, 9086–9104. [Google Scholar] [CrossRef]
  29. Firebaugh, M.W. Artificial Intelligence: A Knowledge Based Approach; PWS-Kent Publishing Co.: Boston, MA, USA, 1989. [Google Scholar]
  30. Ziaee, A.A. A philosophical Approach to Artificial Intelligence and Islamic Values. IIUM Eng. J. 2011, 12, 73–78. [Google Scholar] [CrossRef]
  31. Allen, J. Natural Language Understanding; Benjamin/Cummings Publishing Company: San Francisco, CA, USA, 1995. [Google Scholar]
  32. Dubien, S. Question Answering Using Document Tagging and Question Classification. Ph.D. Thesis, University of Lethbridge, Lethbridge, Alberta, 2005. [Google Scholar]
  33. Guo, Q.; Zhang, M. Question answering based on pervasive agent ontology and Semantic Web. Knowl. Based Syst. 2009, 22, 443–448. [Google Scholar] [CrossRef]
  34. Cui, W.; Xiao, Y.; Wang, H.; Song, Y.; Hwang, S.W.; Wang, W. KBQA: Learning question answering over QA corpora and knowledge bases. arXiv 2019, arXiv:1903.02419. [Google Scholar] [CrossRef]
  35. Dong, J.; Wang, Y.; Yu, R. Application of the Semantic Network Method to Sightline Compensation Analysis of the Humble Administrator’s Garden. Nexus Netw. J. 2021, 23, 187–203. [Google Scholar] [CrossRef]
  36. Shortliffe, E. Computer-Based Medical Consultations: MYCIN; Elsevier: Amsterdam, The Netherlands, 2012; Volume 2. [Google Scholar]
  37. Stokman, F.N.; de Vries, P.H. Structuring knowledge in a graph. In Human-Computer Interaction; Springer: Berlin/Heidelberg, Germany, 1988; pp. 186–206. [Google Scholar]
  38. Dong, X.; Gabrilovich, E.; Heitz, G.; Horn, W.; Lao, N.; Murphy, K.; Strohmann, T.; Sun, S.; Zhang, W. Knowledge vault: A web-scale approach to probabilistic knowledge fusion. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 601–610. [Google Scholar]
  39. Lin, Y.; Han, X.; Xie, R.; Liu, Z.; Sun, M. Knowledge representation learning: A quantitative review. arXiv 2018, arXiv:1812.10901. [Google Scholar]
  40. Dai, Z.; Li, L.; Xu, W. Cfo: Conditional focused neural question answering with large-scale knowledge bases. arXiv 2016, arXiv:1606.01994. [Google Scholar]
  41. Chen, Y.; Wu, L.; Zaki, M.J. Bidirectional attentive memory networks for question answering over knowledge bases. arXiv 2019, arXiv:1903.02188. [Google Scholar]
  42. Mohammed, S.; Shi, P.; Lin, J. Strong baselines for simple question answering over knowledge graphs with and without neural networks. arXiv 2017, arXiv:1712.01969. [Google Scholar]
  43. Zhang, Z.; Liu, G. Study of ontology-based intelligent question answering model for online learning. In Proceedings of the 2009 First International Conference on Information Science and Engineering, Nanjing, China, 26–28 December 2009; pp. 3443–3446. [Google Scholar]
  44. Shadbolt, N.; Berners-Lee, T.; Hall, W. The semantic web revisited. IEEE Intell. Syst. 2006, 21, 96–101. [Google Scholar] [CrossRef]
  45. Besbes, G.; Baazaoui-Zghal, H.; Moreno, A. Ontology-based question analysis method. In International Conference on Flexible Query Answering Systems; Springer: Berlin/Heidelberg, Germany, 2013; pp. 100–111. [Google Scholar]
  46. Alobaidi, M.; Malik, K.M.; Sabra, S. Linked open data-based framework for automatic biomedical ontology generation. BMC Bioinform. 2018, 19, 319. [Google Scholar] [CrossRef] [PubMed]
  47. Nogueira, T.P.; Braga, R.B.; de Oliveira, C.T.; Martin, H. FrameSTEP: A framework for annotating semantic trajectories based on episodes. Expert Syst. Appl. 2018, 92, 533–545. [Google Scholar] [CrossRef]
  48. Ali, F.; El-Sappagh, S.; Kwak, D. Fuzzy Ontology and LSTM-Based Text Mining: A Transportation Network Monitoring System for Assisting Travel. Sensors 2019, 19, 234. [Google Scholar] [CrossRef]
  49. Wimmer, H.; Chen, L.; Narock, T. Ontologies and the Semantic Web for Digital Investigation Tool Selection. J. Digit. For. Secur. Law 2018, 13, 6. [Google Scholar] [CrossRef]
  50. Singh, J.; Sharan, D.A. A comparative study between keyword and semantic based search engines. In Proceedings of the International Conference on Cloud, Big Data and Trust, Madhya Pradesh, India, 13–15 November 2013; pp. 13–15. [Google Scholar]
  51. Alobaidi, M.; Malik, K.M.; Hussain, M. Automated ontology generation framework powered by linked biomedical ontologies for disease-drug domain. Comput. Methods Programs Biomed. 2018, 165, 117–128. [Google Scholar] [CrossRef]
  52. Xu, F.; Liu, X.; Zhou, C. Developing an ontology-based rollover monitoring and decision support system for engineering vehicles. Information 2018, 9, 112. [Google Scholar] [CrossRef]
  53. Dhayne, H.; Chamoun, R.K.; Sabha, R.A. IMAT: Intelligent Mobile Agent. In Proceedings of the 2018 IEEE International Multidisciplinary Conference on Engineering Technology (IMCET), Beirut, Lebanon, 14–16 November 2018; pp. 1–8. [Google Scholar]
  54. Nanda, J.; Simpson, T.W.; Kumara, S.R.; Shooter, S.B. A Methodology for Product Family Ontology Development Using Formal Concept Analysis and Web Ontology Language. J. Comput. Inf. Sci. Eng. 2006, 6, 103–113. [Google Scholar] [CrossRef]
  55. Subhashini, R.; Akilandeswari, J. A survey on ontology construction methodologies. Int. J. Enterp. Comput. Bus. Syst. 2011, 1, 60–72. [Google Scholar]
  56. Schubotz, M.; Scharpf, P.; Dudhat, K.; Nagar, Y.; Hamborg, F.; Gipp, B. Introducing MathQA: A Math-Aware question answering system. Inf. Discov. Deliv. 2018, 46, 214–224. [Google Scholar] [CrossRef]
  57. AlAgha, I.M.; Abu-Taha, A. AR2SPARQL: An arabic natural language interface for the semantic web. Int. J. Comput. Appl. 2015, 125, 19–27. [Google Scholar] [CrossRef]
  58. Hakkoum, A.; Kharrazi, H.; Raghay, S. A Portable Natural Language Interface to Arabic Ontologies. Int. J. Adv. Comput. Sci. Appl. 2018, 9, 69–76. [Google Scholar] [CrossRef][Green Version]
  59. Afzal, H.; Mukhtar, T. Semantically enhanced concept search of the Holy Quran: Qur’anic English WordNet. Arab. J. Sci. Eng. 2019, 44, 3953–3966. [Google Scholar] [CrossRef]
  60. Ullah, M.A.; Hossain, S.A. Ontology-Based Information Retrieval System for University: Methods and Reasoning. In Emerging Technologies in Data Mining and Information Security; Springer: Berlin/Heidelberg, Germany, 2019; pp. 119–128. [Google Scholar]
  61. Syarief, M.; Agustiono, W.; Muntasa, A.; Yusuf, M. A Conceptual Model of Indonesian Question Answering System based on Semantic Web. J. Phys. Conf. Ser. 2020, 1569, 022089. [Google Scholar]
  62. Jabalameli, M.; Nematbakhsh, M.; Zaeri, A. Ontology-lexicon–based question answering over linked data. ETRI J. 2020, 42, 239–246. [Google Scholar] [CrossRef]
  63. Diefenbach, D.; Both, A.; Singh, K.; Maret, P. Towards a question answering system over the semantic web. Semant. Web 2020, 11, 421–439. [Google Scholar] [CrossRef]
  64. Saloot, M.A.; Idris, N.; Mahmud, R.; Ja’afar, S.; Thorleuchter, D.; Gani, A. Hadith data mining and classification: A comparative analysis. Artif. Intell. Rev. 2016, 46, 113–128. [Google Scholar] [CrossRef]
  65. White, R.W.; Richardson, M.; Yih, W. Questions vs. queries in informational search tasks. In Proceedings of the 24th International Conference on World Wide Web, Florence, Italy, 18–22 May 2015; pp. 135–136. [Google Scholar]
  66. del Carmen Rodrıguez-Hernández, M.; Ilarri, S.; Trillo-Lado, R.; Guerra, F. Towards keyword-based pull recommendation systems. ICEIS 2016 2016, 1, 207–214. [Google Scholar]
  67. Rahman, M.M.; Yeasmin, S.; Roy, C.K. Towards a context-aware IDE-based meta search engine for recommendation about programming errors and exceptions. In Proceedings of the 2014 Software Evolution Week-IEEE Conference on Software Maintenance, Reengineering, and Reverse Engineering (CSMR-WCRE), Antwerp, Belgium, 3–6 February 2014; pp. 194–203. [Google Scholar]
  68. Giri, K. Role of ontology in semantic web. Desidoc J. Libr. Inf. Technol. 2011, 31, 116–120. [Google Scholar] [CrossRef]
  69. Maglaveras, N.; Koutkias, V.; Chouvarda, I.; Goulis, D.; Avramides, A.; Adamidis, D.; Louridas, G.; Balas, E. Home care delivery through the mobile telecommunications platform: The Citizen Health System (CHS) perspective. Int. J. Med. Inform. 2002, 68, 99–111. [Google Scholar] [CrossRef]
  70. Sherif, M.; Ngonga Ngomo, A.C. Semantic Quran: A multilingual resource for natural-language processing. Semant. Web 2015, 6, 339–345. [Google Scholar] [CrossRef]
  71. Al-Feel, H. The roadmap for the Arabic chapter of DBpedia. In Proceedings of the 14th International Conference on Telecom. and Informatics (TELE-INFO’15), Sliema, Malta, 17–19 August 2015; pp. 115–125. [Google Scholar]
  72. Hakkoum, A.; Raghay, S. Semantic Q&A System on the Qur’an. Arab. J. Sci. Eng. 2016, 41, 5205–5214. [Google Scholar]
  73. Sulaiman, S.; Mohamed, H.; Arshad, M.R.M.; Yusof, U.K. Hajj-QAES: A knowledge-based expert system to support hajj pilgrims in decision making. In Proceedings of the 2009 International Conference on Computer Technology and Development, Kota Kinabalu, Malaysia, 13–15 November 2009; Volume 1, pp. 442–446. [Google Scholar]
  74. Sharef, N.M.; Murad, M.A.; Mustapha, A.; Shishechi, S. Semantic question answering of umrah pilgrims to enable self-guided education. In Proceedings of the 2013 13th International Conference on Intellient Systems Design and Applications, Salangor, Malaysia, 8–10 December 2013; pp. 141–146. [Google Scholar]
  75. Mohamed, H.H.; Arshad, M.R.H.M.; Azmi, M.D. M-HAJJ DSS: A mobile decision support system for Hajj pilgrims. In Proceedings of the 2016 3rd International Conference on Computer and Information Sciences (ICCOINS), Kuala Lumpur, Malaysia, 15–17 August 2016; pp. 132–136. [Google Scholar]
  76. Abdelazeez, M.A.; Shaout, A. Pilgrim Communication Using Mobile Phones. J. Image Graph. 2016, 4, 59–62. [Google Scholar] [CrossRef]
  77. Khan, E.A.; Shambour, M.K.Y. An analytical study of mobile applications for Hajj and Umrah services. Appl. Comput. Inform. 2018, 14, 37–47. [Google Scholar] [CrossRef]
  78. Rodrigo, A.; Penas, A. A study about the future evaluation of Question-Answering systems. Knowl. Based Syst. 2017, 137, 83–93. [Google Scholar] [CrossRef]
  79. Al-Harbi, O.; Jusoh, S.; Norwawi, N.M. Lexical disambiguation in natural language questions (NLQs). arXiv 2017, arXiv:1709.09250. [Google Scholar]
  80. Navigli, R. Natural Language Understanding: Instructions for (Present and Future) Use. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18), Stockholm, Sweden, 13–19 July 2018; Volume 18, pp. 5697–5702. [Google Scholar]
  81. Pillai, L.R.; Veena, G.; Gupta, D. A combined approach using semantic role labelling and word sense disambiguation for question generation and answer extraction. In Proceedings of the 2018 Second International Conference on Advances in Electronics, Computers and Communications (ICAECC), Bangalore, India, 9–10 February 2018; pp. 1–6. [Google Scholar]
  82. Dreisbach, C.; Koleck, T.A.; Bourne, P.E.; Bakken, S. A systematic review of natural language processing and text mining of symptoms from electronic patient-authored text data. Int. J. Med Inform. 2019, 125, 37–46. [Google Scholar] [CrossRef]
  83. Zheng, W.; Cheng, H.; Yu, J.X.; Zou, L.; Zhao, K. Interactive natural language question answering over knowledge graphs. Inf. Sci. 2019, 481, 141–159. [Google Scholar] [CrossRef]
  84. Dimitrakis, E.; Sgontzos, K.; Tzitzikas, Y. A survey on question answering systems over linked data and documents. J. Intell. Inf. Syst. 2020, 55, 233–259. [Google Scholar] [CrossRef]
  85. Li, X.; Hu, D.; Li, H.; Hao, T.; Chen, E.; Liu, W. Automatic question answering from Web documents. Wuhan Univ. J. Nat. Sci. 2007, 12, 875–880. [Google Scholar] [CrossRef]
  86. Ji, S.; Pan, S.; Cambria, E.; Marttinen, P.; Yu, P.S. A survey on knowledge graphs: Representation, acquisition and applications. arXiv 2020, arXiv:2002.00388. [Google Scholar]
  87. Hamed, S.K.; Ab Aziz, M.J. A Question Answering System on Holy Quran Translation Based on Question Expansion Technique and Neural Network Classification. J. Comput. Sci. 2016, 12, 169–177. [Google Scholar] [CrossRef]
  88. Fernández, M.; Cantador, I.; López, V.; Vallet, D.; Castells, P.; Motta, E. Semantically enhanced information retrieval: An ontology-based approach. J. Web Semant. 2011, 9, 434–452. [Google Scholar] [CrossRef]

Short Biography of Authors

Information 12 00200 i001Ammar F. Arbaaeen received the B.Sc. (Hons) degree in Mobile andWireless Computing from The University ofWestminster, UK, in 2011 and the M.Sc. degree in Advanced Computer Science from The University of Manchester, Manchester, UK, in 2012. He is currently pursuing the Ph.D. degree in Computer Science at IIUM University, Malaysia. From 2013 to 2014, he was a Research Assistant with the Transportation and Crowd Management Centre of Research Excellence. In 2014, he joined Umm Al-Qura University as a lecturer and member of the Information and Scientific Services Department with The Institute of Hajj and Umrah Research, KSA. His research interests are to investigate how to semantically enhance question answering technology in order to understand and support human precisely includes NLP, semantics, and knowledge graph technology.
Information 12 00200 i002Asadullah Shah started his career as a Computer Technology lecturer in 1986. Before joining IIUM in January 2011, he worked in Pakistan as a full professor (2001-2010) at Isra, IOBM and SIBA. He earned his Ph.D. in Multimedia Communication from the University of Surrey UK in 1998. He has published 180 articles in ISI and Scopus indexed publications. In addition to that, he has published 12 books. Since 2012, he has been working as a resource person and delivering workshops on proposal writing, research methodologies and literature review at Student Learning Unit (SLeU), KICT, International Islamic University Malaysia and consultancies to other organizations. Professor Shah is a winner of many gold, silver and bronze medals in his career and reviewer for many ISI, and Scopus indexed journals, and other journals of high repute. Currently, he is supervising many Ph.D. projects in KICT in the field of IT and CS.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.