Natural Language Processing Method: Deep Learning and Deep Semantics

A special issue of Electronics (ISSN 2079-9292). This special issue belongs to the section "Electronic Multimedia".

Deadline for manuscript submissions: 15 September 2024 | Viewed by 1561

Special Issue Editors

School of Computing, National University of Singapore, Singapore 117417, Singapore
Interests: computer vision; video understanding; vision and language
Special Issues, Collections and Topics in MDPI journals
School of EIE, The University of Sydney, Sydney, NSW 2006, Australia
Interests: computer vision; machine learning; vision and language

Special Issue Information

Dear Colleagues,

With the rapid development of deep learning technology, intelligent cross-modal systems have garnered a great deal interest from academics and industry alike. Accordingly, we have witnessed the recent dramatic emergence of various AI-based vision–language applications in various fields. This Special Issues invites original research addressing important, innovative and timely challenges in the community. Potential topics include, but are not limited to:

  • visual captioning (image, video);
  • visual question answering (image, video);
  • visual text retrieval (image, video);
  • storytelling; dense visual captioning;
  • visual dialog (image, video);
  • visual grounding;
  • scene graph generation.

Dr. Wei Ji
Dr. Yiming Wu
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Electronics is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2400 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • computer vision
  • natural language processing
  • machine learning
  • artificial intelligence
  • visual understanding and recognition
  • deep learning

Published Papers (1 paper)

Order results
Result details
Select all
Export citation of selected articles as:

Research

13 pages, 1724 KiB  
Article
Context-Dependent Multimodal Sentiment Analysis Based on a Complex Attention Mechanism
by Lujuan Deng, Boyi Liu, Zuhe Li, Jiangtao Ma and Hanbing Li
Electronics 2023, 12(16), 3516; https://doi.org/10.3390/electronics12163516 - 20 Aug 2023
Cited by 1 | Viewed by 1238
Abstract
Multimodal sentiment analysis aims to understand people’s attitudes and opinions from different data forms. Traditional modality fusion methods for multimodal sentiment analysis con-catenate or multiply various modalities without fully utilizing context information and the correlation between modalities. To solve this problem, this article [...] Read more.
Multimodal sentiment analysis aims to understand people’s attitudes and opinions from different data forms. Traditional modality fusion methods for multimodal sentiment analysis con-catenate or multiply various modalities without fully utilizing context information and the correlation between modalities. To solve this problem, this article provides a new model based on a multimodal sentiment analysis framework based on a recurrent neural network with a complex attention mechanism. First, after the raw data is preprocessed, the numerical feature representation is obtained using feature extraction. Next, the numerical features are input into the recurrent neural network, and the output results are multimodally fused using a complex attention mechanism layer. The objective of the complex attention mechanism is to leverage enhanced non-linearity to more effectively capture the inter-modal correlations, thereby improving the performance of multimodal sentiment analysis. Finally, the processed results are fed into the classification layer and the sentiment output is obtained using the classification layer. This process can effectively capture the semantic information and contextual relationship of the input sequence and fuse different pieces of modal information. Our model was tested on the CMU-MOSEI datasets, achieving an accuracy of 82.04%. Full article
(This article belongs to the Special Issue Natural Language Processing Method: Deep Learning and Deep Semantics)
Show Figures

Figure 1

Back to TopTop