Next Article in Journal
A Research of Pulse Source for EFT Test of Electrical Vehicle Piles
Previous Article in Journal
Invisalign® vs. Spark™ Template: Which Is the Most Effective in the Attachment Bonding Procedure? A Randomized Controlled Trial
Previous Article in Special Issue
Comparison of Deep Learning Models and Various Text Pre-Processing Techniques for the Toxic Comments Classification
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Editorial

Special Issue on “Natural Language Processing: Emerging Neural Approaches and Applications”

1
Institute for High Performance Computing and Networking—National Research Council of Italy (ICAR-CNR), 80131 Naples, Italy
2
Department of Computing and Mathematics, Manchester Metropolitan University (MMU), Manchester M15 6BH, UK
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(15), 6717; https://doi.org/10.3390/app11156717
Submission received: 25 June 2021 / Accepted: 27 June 2021 / Published: 22 July 2021
Nowadays, systems based on artificial intelligence are being developed, leading to impressive achievements in a variety of complex cognitive tasks, matching or even beating humans [1,2,3,4]. Natural language processing (NLP) is a field where the use of deep learning (DL) models in the last five years has allowed AI to advance toward human levels in translation and reading comprehension, as well as other real-world NLP applications, such as question answering and conversational systems, information retrieval, sentiment analysis, and recommender systems.
However, due to the difficulties associated with natural language understanding and generation, which are human capabilities among the least understood by computer systems from a cognitive perspective, and despite the remarkable success of DL in different NLP tasks, this is still a field of research of increasing interest [5,6,7]. In order to improve DL methods, current models have been scaled up, but their complexity has grown toward directions assumed by empirical engineering solutions [8,9,10,11]. Moreover, they are not applicable to languages without extensive datasets [12], and the lack of explainability inhibits further improvements [13].
This Special Issue highlights the most recent research being carried out in the NLP field to discuss these open issues, with a particular focus on both emerging approaches for language learning, understanding, production, and grounding interactively or autonomously from data in cognitive and neural systems, as well as on their potential or real applications in different domains.
There are 30 contributions selected for this Special Issue representing progress and potential applications in the NLP area from original contributions of researchers with a broad expertise in various fields: NLP, cognitive science and psychology, artificial intelligence and neural networks, computational modeling and neuroscience covering the whole range of theoretical and practical aspects, technologies, and systems.
This collection includes one review paper, which focuses on text corpus-based tourism big data mining [14]. Li et al. summarized and discussed different text representation strategies, text-based NLP techniques for topic extraction, text classification, sentiment analysis, and text clustering in the context of tourism text mining, as well as their applications in tourist profiling, destination image analysis, and market demand, among others. Their work also provides guidelines for constructing new tourism big data applications and outlines promising research areas in this field for the coming years.
One letter is also included in this issue, employing evolutionary a neural architecture search for Korean grammaticality tasks [15].
Regarding the other 28 research papers, the following NLP areas are specifically addressed:
Natural language understanding, generation, and grounding: In [16], OntologyFixer is presented, a web-based tool that supports a methodology to build, assess, and improve the quality of Ontology Web Language (OWL) ontologies. Another paper [17] addresses the problem of paraphrase identification and presents an approach for leveraging contextual features with a neural-based learning model based on lexical, syntactic, and sentential encodings, incorporating relational graph convolutional networks (R-GCNs) to make use of different features from local contexts (e.g., word encoding, position encoding, and full dependency structures). In addition, in [18], the authors revisited the recurrent neural network (RNN) language model, achieving highly competitive results with the appropriate network structure and hyperparameters.
Universal language models: In [19], Javaloy and the co-author used a method recently proposed, called the causal feature extractor (CFE), for encoder-decoder models on different text processing tasks. The same authors applied this method to text normalization in [20], which is a ubiquitous problem that appears as the first step of many text-to-speech (TTS) systems.
Conversational systems or interfaces and question answering: The authors in [21] proposed the best practices for question classification in different languages using convolutional neural networks (CNNs), finding the optimal settings depending on the language and validating their transferability. The authors in [22] addressed the time-consuming development of manual user simulator policy and introduced a multi-agent dialogue model, where an end-to-end dialogue manager and a user simulator are optimized simultaneously for dialogue management by cooperative multi-agent reinforcement learning. Moreover, in [23], the authors proposed a Medical Instructed Real-time Assistant (MIRA) that listens to the user’s chief complaint and predicts a specific disease, thus referring the user to a nearby appropriate medical specialist. Furthermore, in [24], the authors presented a multi-turn chatbot model in which the preceding utterances are exploited in response generation by using different weights.
Sentiment analysis, emotion detection, and opinion mining: The study in [25] investigated a comparison of various DL models used to identify the toxic comments in Internet discussions. Moreover, in [26], the authors proposed a novel hybrid model XGA (namely an XLNet-based bidirectional gated recurrent unit (BiGRU) network with an attention mechanism) for Cantonese rumor detection on Twitter, taking advantage of both semantic and sentiment features for detection. Furthermore, the authors of [27] proposed an intensive study regarding a domain-independent classification model for sentiment analysis using neural models, showing high performance when using different evaluation metrics compared with the state-of-the-art results. Another study in [28] tested different approaches for handling long documents and proposed a novel technique for sentiment enrichment of the Bidirectional Encoder Representations from Transformers (BERT) model as an intermediate training step. In [29], Rizkallah et al. proposed an embedding approach that is designed to capture the polarity issue for sentiment analysis.
Document analysis, information extraction, and text mining: In [30], Ronran et al. evaluated the combination of different types of embedding features in a bidirectional long short-term memory (Bi-LSTM) conditional random field (CRF) model for named entity recognition (NER). The authors in [31] investigated the transferability of the features from an open information extraction (OIE) domain to another and applied the approach for relation extraction (RE). The authors in [32] proposed a rule-based approach for text document classification. The study in [33] proposed an RE model based on a dual pointer network with a multi-head attention mechanism to address the association of multiple entities in a sentence according to various relations. The work in [34] investigated an RE method to solve the possible overlapping among multiple relational triples contained in a sentence. Another topic was introduced by the authors of [35], who introduced a novel hybrid model of extractive-abstractive text summarization to combine BERT word embedding with reinforcement learning. Two contributions to this special issue are focused on medical information extraction. The authors in [36] compared different architectures of DL models, including CNNs, LSTM, and hybrid models. Furthermore, they proposed a hybrid architecture for protein–protein interaction extraction from the biomedical literature. The authors in [37] developed a multitask attention-based Bi-LSTM–CRF model with pre-trained embeddings from language models (ELMo) in order to achieve improved performance in clinical NER.
Search and information retrieval: In [38], Boban et al. adapted language modeling-based methods for sentence retrieval to test the partial matching of terms through combining sentence retrieval with sequence similarity. This method allows for matching words that are similar but not identical. The authors of [39] proposed a reliable sentence classification model based on an encoder-decoder neural network to resolve lexical disagreement problems between queries and frequently asked questions (FAQs).
Trustworthy and explainable artificial intelligence: Two contributions [40,41] considered “sememe”, the smallest semantic unit for describing real-world concepts, which improve the interpretability of NLP systems. In particular, the study in [40] proposed a novel model to improve the performance of sememe prediction by introducing synonyms. On the other hand, the work in [41] implicitly synthesized the structural features of sememes into word embedding models through an attention mechanism. The work proposes a novel double attention word-based embedding (DAWE) model that encodes the characteristics of sememes into words with a “double attention” strategy.
Applications in science, engineering, medicine, healthcare, finance, business, law, education, transportation, retailing, telecommunication, and multimedia: The authors in [42] proposed a hybrid adversarial attack method to generate examples with the aim to explore the vulnerabilities and security aspects of deep learning systems in different application scenarios. An application in programming education was considered in [43]. In this study, the source code assessment and its classification were developed by a sequential language model that used an attention mechanism through an LSTM neural network and based on the estimated error probability.
In summary, this Special Issue contains a series of excellent research works on NLP, covering a wide range of topics. The collection of 30 contributions is highly recommended, and it will benefit readers in various aspects.

Acknowledgments

We would like to thank all the authors, the dedicated referees, the editor team of applied sciences for their valuable contributions, making this special issue a success.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ferrucci, D.; Brown, E.; Chu-Carroll, J.; Fan, J.; Gondek, D.; Kalyanpur, A.A.; Lally, A.; Murdock, J.W.; Nyberg, E.; Prager, J.; et al. Building Watson: An Overview of the DeepQA Project. AI Mag. 2010, 31, 59–79. [Google Scholar] [CrossRef] [Green Version]
  2. Silver, D.; Huang, A.; Maddison, C.; Guez, A.; Sifre, L.; van den Driessche, G.; Schrittwieser, J.; Antonoglou, I.; Panneershelvam, V.; Lanctot, M.; et al. Mastering the game of Go with deep neural networks and tree search. Nature 2016, 529, 484–489. [Google Scholar] [CrossRef] [PubMed]
  3. Newborn, M. Kasparov Versus Deep Blue; Springer: Berlin, Germany, 1997. [Google Scholar]
  4. Baughman, A.; Chuang, W.; Dixon, K.; Benz, Z.; Basilico, J. DeepQA Jeopardy! Gamification: A Machine-Learning Perspective. Comput. Intell. AI Games IEEE Trans. 2014, 6, 55–66. [Google Scholar] [CrossRef]
  5. Yadav, A.; Vishwakarma, D.K. Sentiment analysis using deep learning architectures: A review. Artif. Intell. Rev. 2019, 53, 4335–4385. [Google Scholar] [CrossRef]
  6. Yuan, S.; Zhang, Y.; Tang, J.; Hall, W.; Cabotà, J.B. Expert finding in community question answering: A review. Artif. Intell. Rev. 2020, 53, 843–874. [Google Scholar] [CrossRef] [Green Version]
  7. Wang, Y.; Wang, M.; Fujita, H. Word Sense Disambiguation: A comprehensive knowledge exploitation framework. Knowl. Based Syst. 2020, 190, 105030. [Google Scholar] [CrossRef]
  8. Laha, A.; Raykar, V. An Empirical Evaluation of various Deep Learning Architectures for Bi-Sequence Classification Tasks. In Proceedings of the COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, Osaka, Japan, 11–16 December 2016; pp. 2762–2773. [Google Scholar]
  9. Nguyen, V.H.; Cheng, J.S.; Yu, Y.; Thai, V.T. An architecture of deep learning network based on ensemble empirical mode decomposition in precise identification of bearing vibration signal. J. Mech. Sci. Technol. 2019, 33, 41–50. [Google Scholar] [CrossRef]
  10. Guo, Q.; Chen, S. An Empirical Study towards Characterizing Deep Learning Development and Deployment across Different Frameworks and Platforms. arXiv 2019, arXiv:1909.06727v1. Available online: https://arxiv.org/pdf/1909.06727.pdf (accessed on 20 July 2021).
  11. Pota, M.; Marulli, F.; Esposito, M.; De Pietro, G.; Fujita, H. Multilingual POS tagging by a composite deep architecture based on character-level features and on-the-fly enriched Word Embeddings. Knowl. Based Syst. 2019, 164, 309–323. [Google Scholar] [CrossRef]
  12. Cherry, C.A. EMNLP Workshop on Deep Learning for Low-Resource NLP. 20 June 2019, China, Association for Computational Linguistics. Available online: https://sites.google.com/view/deeplo19 (accessed on 20 July 2021).
  13. Zohuri, B.; Moghaddam, M. Deep Learning Limitations and Flaws. Mod. Approaches Mater. Sci. Short Commun. 2020, 2, 241–250. [Google Scholar]
  14. Li, Q.; Li, S.; Zhang, S.; Hu, J.; Hu, J. A Review of Text Corpus-Based Tourism Big Data Mining. Appl. Sci. 2019, 9, 3300. [Google Scholar] [CrossRef] [Green Version]
  15. Park, K.-M.; Shin, D.; Yoo, Y. Evolutionary Neural Architecture Search (NAS) Using Chromosome Non-Disjunction for Korean Grammaticality Tasks. Appl. Sci. 2020, 10, 3457. [Google Scholar] [CrossRef]
  16. Roldan-Molina, G.R.; Mendez, J.R.; Yevseyeva, I.; Basto-Fernandes, V. Ontology Fixing by Using Software Engineering Technology. Appl. Sci. 2020, 10, 6328. [Google Scholar] [CrossRef]
  17. Xu, S.; Shen, X.; Fukumoto, F.; Li, J.; Suzuki, Y.; Nishizaki, H. Paraphrase Identification with Lexical, Syntactic and Sentential Encodings. Appl. Sci. 2020, 10, 4144. [Google Scholar] [CrossRef]
  18. Park, H.; Park, J. Assessment of Word-Level Neural Language Models for Sentence Completion. Appl. Sci. 2020, 10, 1340. [Google Scholar] [CrossRef] [Green Version]
  19. Javaloy, A.; García-Mateos, G. Preliminary Results on Different Text Processing Tasks Using Encoder-Decoder Networks and the Causal Feature Extractor. Appl. Sci. 2020, 10, 5772. [Google Scholar] [CrossRef]
  20. Javaloy, A.; García-Mateos, G. Text Normalization Using Encoder–Decoder Networks Based on the Causal Feature Extractor. Appl. Sci. 2020, 10, 4551. [Google Scholar] [CrossRef]
  21. Pota, M.; Esposito, M.; De Pietro, G.; Fujita, H. Best Practices of Convolutional Neural Networks for Question Classification. Appl. Sci. 2020, 10, 4710. [Google Scholar] [CrossRef]
  22. Lei, S.; Wang, X.; Yuan, C. Cooperative Multi-Agent Reinforcement Learning with Conversation Knowledge for Dialogue Management. Appl. Sci. 2020, 10, 2740. [Google Scholar] [CrossRef] [Green Version]
  23. Rehman, U.U.; Chang, D.J.; Jung, Y.; Akhtar, U.; Razzaq, M.A.; Lee, S. Medical Instructed Real-Time Assistant for Patient with Glaucoma and Diabetic Conditions. Appl. Sci. 2020, 10, 2216. [Google Scholar] [CrossRef] [Green Version]
  24. Kim, J.; Oh, S.; Kwon, O.-W.; Kim, H. Multi-Turn Chatbot Based on Query-Context Attentions and Dual Wasserstein Generative Adversarial Networks. Appl. Sci. 2019, 9, 3908. [Google Scholar] [CrossRef] [Green Version]
  25. Maslej-Krešňáková, V.; Sarnovský, M.; Butka, P.; Machová, K. Comparison of Deep Learning Models and Various Text Pre-Processing Techniques for the Toxic Comments Classification. Appl. Sci. 2020, 10, 8631. [Google Scholar] [CrossRef]
  26. Chen, X.; Ke, L.; Lu, Z.; Su, H.; Wang, H. A Novel Hybrid Model for Cantonese Rumor Detection on Twitter. Appl. Sci. 2020, 10, 7093. [Google Scholar] [CrossRef]
  27. Jnoub, N.; Al Machot, F.; Klas, W. A Domain-Independent Classification Model for Sentiment Analysis Using Neural Models. Appl. Sci. 2020, 10, 6221. [Google Scholar] [CrossRef]
  28. Pelicon, A.; Pranjić, M.; Miljković, D.; Škrlj, B.; Pollak, S. Zero-Shot Learning for Cross-Lingual News Sentiment Classification. Appl. Sci. 2020, 10, 5993. [Google Scholar] [CrossRef]
  29. Rizkallah, S.; Atiya, A.F.; Shaheen, S. A Polarity Capturing Sphere for Word to Vector Representation. Appl. Sci. 2020, 10, 4386. [Google Scholar] [CrossRef]
  30. Ronran, C.; Lee, S.; Jang, H.J. Delayed Combination of Feature Embedding in Bidirectional LSTM CRF for NER. Appl. Sci. 2020, 10, 7557. [Google Scholar] [CrossRef]
  31. Sarhan, I.; Spruit, M. Can We Survive without Labelled Data in NLP? Transfer Learning for Open Information Extraction. Appl. Sci. 2020, 10, 5758. [Google Scholar] [CrossRef]
  32. Aubaid, A.M.; Mishra, A. A Rule-Based Approach to Embedding Techniques for Text Document Classification. Appl. Sci. 2020, 10, 4009. [Google Scholar] [CrossRef]
  33. Park, S.; Kim, H. Dual Pointer Network for Fast Extraction of Multiple Relations in a Sentence. Appl. Sci. 2020, 10, 3851. [Google Scholar] [CrossRef]
  34. Xiao, S.; Song, M. A Text-Generated Method to Joint Extraction of Entities and Relations. Appl. Sci. 2019, 9, 3795. [Google Scholar] [CrossRef] [Green Version]
  35. Wang, Q.; Liu, P.; Zhu, Z.; Yin, H.; Zhang, Q.; Zhang, L. A Text Abstraction Summary Model Based on BERT Word Embedding and Reinforcement Learning. Appl. Sci. 2019, 9, 4701. [Google Scholar] [CrossRef] [Green Version]
  36. Quan, C.; Luo, Z.; Wang, S. A Hybrid Deep Learning Model for Protein–Protein Interactions Extraction from Biomedical Literature. Appl. Sci. 2020, 10, 2690. [Google Scholar] [CrossRef] [Green Version]
  37. Yang, J.; Liu, Y.; Qian, M.; Guan, C.; Yuan, X. Information Extraction from Electronic Medical Records Using Multitask Recurrent Neural Network with Contextual Word Embedding. Appl. Sci. 2019, 9, 3658. [Google Scholar] [CrossRef] [Green Version]
  38. Boban, I.; Doko, A.; Gotovac, S. Improving Sentence Retrieval Using Sequence Similarity. Appl. Sci. 2020, 10, 4316. [Google Scholar] [CrossRef]
  39. Jang, Y.; Kim, H. Reliable Classification of FAQs with Spelling Errors Using an Encoder-Decoder Neural Network in Korean. Appl. Sci. 2019, 9, 4758. [Google Scholar] [CrossRef] [Green Version]
  40. Kang, X.; Li, B.; Yao, H.; Liang, Q.; Li, S.; Gong, J.; Li, X. Incorporating Synonym for Lexical Sememe Prediction: An Attention-Based Model. Appl. Sci. 2020, 10, 5996. [Google Scholar] [CrossRef]
  41. Li, S.; Chen, R.; Wan, B.; Gong, J.; Yang, L.; Yao, H. DAWE: A Double Attention-Based Word Embedding Model with Sememe Structure Information. Appl. Sci. 2020, 10, 5804. [Google Scholar] [CrossRef]
  42. Du, X.; Yu, J.; Yi, Z.; Li, S.; Ma, J.; Tan, Y.; Wu, Q. A Hybrid Adversarial Attack for Different Application Scenarios. Appl. Sci. 2020, 10, 3559. [Google Scholar] [CrossRef]
  43. Rahman, M.M.; Watanobe, Y.; Nakamura, K. Source Code Assessment and Classification Based on Estimated Error Probability Using Attentive LSTM Language Model and Its Application in Programming Education. Appl. Sci. 2020, 10, 2973. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Esposito, M.; Masala, G.L.; Minutolo, A.; Pota, M. Special Issue on “Natural Language Processing: Emerging Neural Approaches and Applications”. Appl. Sci. 2021, 11, 6717. https://doi.org/10.3390/app11156717

AMA Style

Esposito M, Masala GL, Minutolo A, Pota M. Special Issue on “Natural Language Processing: Emerging Neural Approaches and Applications”. Applied Sciences. 2021; 11(15):6717. https://doi.org/10.3390/app11156717

Chicago/Turabian Style

Esposito, Massimo, Giovanni Luca Masala, Aniello Minutolo, and Marco Pota. 2021. "Special Issue on “Natural Language Processing: Emerging Neural Approaches and Applications”" Applied Sciences 11, no. 15: 6717. https://doi.org/10.3390/app11156717

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop