Next Article in Journal
A Simultaneous Wireless Information and Power Transfer-Based Multi-Hop Uneven Clustering Routing Protocol for EH-Cognitive Radio Sensor Networks
Previous Article in Journal
Data-Driven Short-Term Load Forecasting for Multiple Locations: An Integrated Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mixture of Attention Variants for Modal Fusion in Multi-Modal Sentiment Analysis

1
School of Computer Science, South China Normal University, Guangzhou 510631, China
2
Aberdeen Institute of Data Science and Artificial Intelligence, South China Normal University, Guangzhou 528225, China
3
International United College, South China Normal University, Guangzhou 528225, China
4
School of Software, South China Normal University, Guangzhou 528225, China
*
Authors to whom correspondence should be addressed.
Big Data Cogn. Comput. 2024, 8(2), 14; https://doi.org/10.3390/bdcc8020014
Submission received: 5 December 2023 / Revised: 18 January 2024 / Accepted: 20 January 2024 / Published: 29 January 2024

Abstract

:
With the popularization of better network access and the penetration of personal smartphones in today’s world, the explosion of multi-modal data, particularly opinionated video messages, has created urgent demands and immense opportunities for Multi-Modal Sentiment Analysis (MSA). Deep learning with the attention mechanism has served as the foundation technique for most state-of-the-art MSA models due to its ability to learn complex inter- and intra-relationships among different modalities embedded in video messages, both temporally and spatially. However, modal fusion is still a major challenge due to the vast feature space created by the interactions among different data modalities. To address the modal fusion challenge, we propose an MSA algorithm based on deep learning and the attention mechanism, namely the Mixture of Attention Variants for Modal Fusion (MAVMF). The MAVMF algorithm includes a two-stage process: in stage one, self-attention is applied to effectively extract image and text features, and the dependency relationships in the context of video discourse are captured by a bidirectional gated recurrent neural module; in stage two, four multi-modal attention variants are leveraged to learn the emotional contributions of important features from different modalities. Our proposed approach is end-to-end and has been shown to achieve a superior performance to the state-of-the-art algorithms when tested with two largest public datasets, CMU-MOSI and CMU-MOSEI.

1. Introduction

The accessibility of 5G networks and the popularity of social media have given rise to the ubiquity of opinionated video messages in our cyber world. Amateur users are producing large numbers of videos on social media platforms such as TikTok and on the internet to share their sentiments and emotions toward all aspects of their daily lives [1,2]. These multi-modal data consist of at least two modalities, commonly including texts, acoustics, and images [3]. For instance, videos on TikTok not only contain the creator’s spoken language but also their body movements and facial expressions with pleasing background music and special animation effects.
The rich information contained in multi-modal data creates immense opportunities for different organizations. The ability to digest multi-modal data is beneficial for a variety of application scenarios that can greatly enhance user experiences and utility. For example, in autonomous driving, the vehicular control unit can leverage cameras to monitor drivers’ emotions, driving behaviors, and fatigue conditions in real-time and provide necessary feedback based on the multi-modal signal [4]. This can effectively enhance driving safety and reduce accidents. In the field of medical and health services, multi-modal information (e.g., patient counseling recordings) can be applied for emotion tendency assessments to assist doctors in their decisions on patient treatments [5]. Social media platforms can adopt machine learning techniques to automatically poll collective sentiment tendencies on videos of a given topic, which can then be relayed to other relevant organizations for decision making [6]. However, with greater possibilities come greater challenges with leveraging multi-modal data. Unlike single-modality data (e.g., texts) for sentiment analysis, the diverse modalities contained in multi-modal data both complement and interfere with each other, making information fusion extremely challenging.
The challenges with fusing multi-modal data mainly lie in the encoding both single-modality and cross-modality information and the modeling of contextual relationships among the targeted units of analysis (e.g., utterance). Both tasks require significant computing in an enormous feature space, which invalidates manual feature engineering as a solution option. Fortunately, deep learning has shown promising potential for modeling multi-modal data, particularly for Multi-Modal Sentiment Analysis (MSA) tasks [7]. Various deep learning methods, including CNN, LSTM/GRU, (Self-)Attention, and BERT, have been leveraged to learn encoding from single and cross modalities [7] and have been shown to achieve state-of-the-art performances in MSA tasks [7]. However, there is still no consensus on how to efficiently fuse multi-modal information to remove existing noise while taking contexts into considerations for optimizing the performance of MSA.
In this work, we propose a novel algorithm based on a mixture of attention variants for multi-modal fusion in MSA tasks. Our approach divides the multi-modal fusion problem into two stages: (1) we leverage the self-attention module to maximize the intra-modality information value, and we use the BiGRU module to maintain inter-utterance contexts within each modality; (2) after compressing the feature space using a fully connected module, the tensors resulting from each modality are fed into four different attention variants for multi-modal fusion. In both stages, we apply the attention mechanism to distinguish the contributions from each modality, assigning higher weights to useful features, while reducing irrelevant background interference.
The main contributions of the current work can be summarized as follows: (1) We present a comprehensive literature review on both single and Multi-Modal Sentiment Analysis. (2) A novel MSA method, namely the Mixture of Attention Variants for Modal Fusion (MAVMF), is proposed to solve the multi-modal fusion challenge. (3) Experimental data on the two largest benchmark public datasets show that our proposed MAVMF algorithm can effectively extract multi-modal information, and it is shown to demonstrate improvements when compared with other baseline methods.
In the remaining sections, we provide a comprehensive summary of the sentiment analysis and new developments in Section 2, introduce our problem definition and proposed algorithm in Section 3, provide the details of the experiments in Section 4, and present the results in Section 5, followed by a discussion in Section 6 and a conclusion in Section 7.

2. Related Work

Before we entered into the video age in social media, text data dominated the Sentiment Analysis (SA) sphere and were considered the default data modality in SA. However, research on SA using only text data can suffer from issues like having an “emotional gap” and “subjective perception” [3], leading to unsatisfactory results in SA tasks. Compared with text-only data, user-recorded videos convey emotional information through subscripts (e.g., texts), images and the acoustic signals embedded in them. The popularity of video data gave rise to MSA, in which various modalities are leveraged to corroborate each other and provide a better recognition performance [8]. Since the effective extraction of features within single modalities serves as the foundation and prerequisite for MSA tasks, it is necessary to use feature extractors to extract internal features within each modality. In addition, since the fusion of multi-modal features is critical to the success of MSA tasks, effective fusion algorithms are at the core of MSA research. From these two perspectives, we cover both single- and multi-modality SA in this section.

2.1. Single-Modality Sentiment Analysis

2.1.1. Text Sentiment Analysis

When conducting a text sentiment analysis (TSA), we aim to discover the emotional attitudes expressed by the authors by analyzing the emotions within the text. Before the advent of text analysis technology, people had to manually read and analyze the emotions conveyed in text, resulting in a significant increase in their workload. In addition, manual classifications are prone to human error. Therefore, the use of automation technology to infer text sentiments can significantly improve the label efficiency in TSA.
TSA tasks can be categorized into word-level, sentence-level, and document-level inferences [9]. Document-level SA focuses on the overall emotional tendency, which is obtained by assigning different sentiment contributions to different sentences and aggregating the sentiment tendencies of all sentences in the document. Sentence-level SA focuses on each individual sentence in a document and studies the emotional polarity of the sentence based on the sentiment contributions of the words within the sentence. Word-level SA focuses on each word that appears in a sentence and directly determines its emotional polarity through sentiment lexicons.
All TSA methods can be simply divided into rule-based and machine-learning-based methods [10]. The rule-based approaches use predesigned rules, such as sentiment lexicons, to determine text sentiments. For example, sentiment lexicons may define the polarity scores of emotion words, and the overall sentiment polarity is determined by aggregating the positive and negative scores of the words. The one with the higher score is selected as the final sentiment polarity. The performance of rule-based SA methods largely depends on the accuracy of the scoring for each word and the comprehensiveness of the lexicon set. Due to its simplicity in implementation, rule-based SA methods are widely adopted by researchers and practitioners [11]. For instance, Thelwall et al. [12] proposed the SentiStrength algorithm; Saif et al. [13] developed the SentiCircles platform for SA on Twitter; Li et al. [14] constructed a lexicon to effectively enhance the sentiment analysis performance; Kanayama et al. [15] proposed a syntactic-based method for detecting the sentiment polarity; and Rao et al. [16] conducted a sentiment analysis based on the document topic classification.
The machine-learning-based SA methods aim to automate SA tasks by using supervised models. For example, Chen et al. [17] proposed a novel SA algorithm to extract sentiment features from mobile app reviews and used Support Vector Machines (SVMs) for sentiment classification. Zhao et al. [18] used supervised algorithms to perform a binary classification on product review data, classifying the comments into positive and negative categories. Kiritchenko et al. [19] proposed a SVM algorithm for short and informal texts. Silva et al. [20] applied ensemble learning by combining various classifiers such as random forests and SVMs. More recently, deep learning approaches have emerged as a new avenue of research in TSA. Kim et al. [21] used Convolutional Neural Networks (CNNs) for SA and was able to demonstrate an excellent performance. Makoto et al. [22] combined spatial pyramid pooling with max pooling and used gated neural networks to classify user review texts. Meng et al. [23] proposed a multi-layer CNN algorithm and was able to prove its superiority through experiments. Jiang et al. [24] combined long short-term memory (LSTM) networks [25] with CNNs to handle the dependency on distant sentences. Luo et al. [26] introduced a gated recurrent neural network (RNN) to enhance the contextual relationships between words and texts. Minh et al. [27] proposed three variants of neural networks [28] to capture the long-term dependencies of information.

2.1.2. Image Sentiment Analysis

Image Sentiment Analysis (ISA) mainly focuses on the modeling of users’ facial expressions and postures contained in an image to infer their emotional tendencies. Colombo et al. [29] first proposed an automatic emotion retrieval system that effectively extracts image features and performs emotion classification. Singh et al. [30] applied the CNN with domain specific fine tuning to classify sentiments on Flickr images. Yang et al. [31] created a learning framework that explores only the affective regions in an image and combined it with a CNN to classify the sentiment for an image. Yang et al. [32] proposed a weakly supervised coupled CNN with two branches to leverage localized information from an image for ISA. Kumar et al. [33] constructed a visual emotion framework for emotion feature extraction using the Flickr dataset. Truong et al. [34] developed item-oriented and user-oriented CNNs to better capture the interaction of image features with the specific expressions of users or items for the inference of user review sentiments. You et al. [35] extracted features from local image regions and conducted an ISA by incorporating an attention mechanism into the proposed network. Wu et al. [36] proposed a scheme for ISA that leverages both the inference on the whole image and subimages that contain salient objects. Zheng et al. [37] introduced an “Emotion Region Attention” module, while Li et al. [38] proposed a novel SentiNet model for ISA.

2.1.3. Speech Emotion Analysis

Compared to text and image SA tasks, the development of Speech Emotion Analysis (SEA) has been relatively slow. SEA focuses on analyzing emotions based on factors such as the tone, bandwidth, pitch, and duration of user speech [39]. Since deep learning techniques have been shown to improve the speech recognition performance [40], researchers have proposed various neural-network-based models to enhance the accuracy of speech emotion recognition [41,42,43].

2.2. Multi-Modal Sentiment Analysis

Existing and emergent social media platforms have enabled common users to post self-recorded videos to share their day-to-day living experiences and sentiments on any subject. This led to an explosion of multi-modal information on the internet and created tremendous opportunities for MSA [44]. Morency et al. [1] created the YouTube dataset and constructed a joint model to extract multi-modal features for SA. Poria et al. [45] applied single-modality feature extractors (e.g., CNN on text embeddings and Part-of-Speech taggings) on the visual, audio, and textual channels and trained a multi-kernel learning classifier for MSA. Zadeh et al. [2] introduced a multi-modal lexicon to better capture the interactions between facial gestures and speech. They also published the CMU-MOSI [46] dataset, which became the first benchmark dataset to support research in MSA. Zadeh et al. [47] presented a tensor fusion network model, which learns the interactions within and between text, vision, and acoustic channels. Chen et al. [48] proposed a novel SA model that comprises a gated multi-modal embedding module for information fusion in noisy environments and an LSTM module with temporal attention for higher-resolution word-level fusion.
The aforementioned works only considered the fusion of information between modalities without considering the dependencies between contexts. In order to improve the performance of MSA, Poria et al. [49] introduced an LSTM-based framework that captures the mutual dependencies between utterances using contextual information. In another study, Poria et al. [50] proposed a user-opinion-based model that combines the three modality inputs using a multi-modal learning approach. Zadeh et al. [51] proposed multiple attention blocks to capture information from the three modalities.
More recently, Wang et al. [52] proposed a novel Text Enhanced Transformer Fusion Network (TETFN) method that learns text-oriented pairwise cross-modal mappings to obtain effective unified multi-modal representations. Yang et al. [53] applied BERT to translate visual and audio features into text features to enhance the quality of both visual and audio features. Wu et al. [54] extracted bi-modal features from the acoustic–visual, acoustic–textual, and visual–textual pairs with a multi-head attention module to improve video sentiment analysis tasks. Wang et al. [55] created a lightweight Hierarchical Cross-Modal Transformer with Modality Gating (HCT-MG) for MSA through primary modality identification. He et al. [56] proposed the Multi-Modal Temporal Attention (MMTA) algorithm, which considers the temporal effects of all modalities on each uni-modal branch to balance the interactions between unimodal branches and the adaptive inter-modal balance. Mai et al. [57] leveraged the contrastive learning framework both within modalities and between modalities for the MSA tasks.
Although these scholars have achieved promising results, there is still room for improvement. In SA tasks, within-modality representations are as important as inter-modality information fusion. The aforementioned research methods do not fully address the extraction of modality-specific features or the fusion of information between modalities. Particularly, the interaction and fusion of multi-modal features may lead to the presence of redundant information in the target network, making it challenging to focus on important information. Therefore, it is critical to identify the contributions of features from different modalities to SA at each stage of the deep learning network, which is the goal of our proposed MAVMF algorithm.

3. Method

3.1. Problem Definition

In this work, videos are considered the source of multi-modal data for sentiment analysis. A video usually contains a series of consecutive image frames, and a user’s emotional tendency can be different or related in consecutive frames. Because of this, the video is processed into video utterances, each of which contains the same emotional tendency of the user, as shown in Figure 1. We aim to perform a sentiment analysis on the utterance level in a given video.
Assuming a dataset contains m videos, D = [ V 1 , V 2 , , V m ] . For the i-th video, V i , the video is composed of n i video segments or utterances, V i = [ u i 1 , u i 2 , , u i n i ] , where u i 1 denotes the first utterance in V i , and n i denotes the total number of utterances in V i . For each utterance u i j , 1 i m , and 1 j n i , contains a feature vector u i j = [ t i j , v i j , a i j ] , representing three modalities, i.e., t i j is the text feature representation, v i j is the visual feature representation, and a i j is the audio feature representation.
Assuming there are C classes of emotional categories for the user’s video, the goal is to label the emotional category of each video utterance. In order to perform sentiment classification, the utterances in the video, except for u i j , are considered to be the context of u i j , and the accuracy and F-1 score are used as the evaluation metrics for the model.

3.2. Model

The MAVMF algorithm can be divided into five steps from an end-to-end pipeline as shown in Figure 2: (A) single-modal feature representation, (B) single-modal attention, (C) contextual feature extraction, (D) multi-modal feature fusion, and (E) sentiment classification. The concrete architecture of MAVMF is illustrated in Algorithm 1. We employ ⊙ for the dot product, ⊗ for element-wise multiplication, and ⊕ for feature concatenation in all formulas in the below sections.

3.2.1. Single-Modal Feature Representation

Due to the distinct semantic spaces of the text, image, and audio, different feature extractors should be used to extract the features within each modality. We adopt the following single-modal feature extractors that were chosen by other studies for the CMU-MOSI and CMU-MOSEI datasets. Specifically, for the CMU-MOSI dataset, we use the utterance-level features provided by Poria et al. [49] as inputs to the MAVMF model; for the CMU-MOSEI dataset, we employ the CMU-multi-modal data SDK [51] tool to extract the corresponding text, audio, and video features as inputs to the MAVMF model.
Algorithm 1: MAVMF architecture
Bdcc 08 00014 i001

3.2.2. Single-Modal Attention

The attention mechanism enables the target network to prioritize high-contributing features and disregard interference from the background information. This section primarily focuses on the single-modal attention modules, which employ distinct attention mechanisms to encode image and text features independently, enhancing the extraction of single-modal features. The single-modal attention module comprises two components: the reduced-attention ( R A t t e n ) block for the visual modality and the self-attention ( S e l f A t t e n ) block for the text modality. The structure of the single-modal attention module is illustrated in Figure 2B.
The RAtten Block. The RAtten block employs channel attention and spatial attention to encode the internal features of the image, enhancing features that significantly contribute to emotions while suppressing background features. The image’s feature vector sequentially passes through the channel attention and spatial attention modules, capturing the importance of each channel and feature map in the image. The specific process can be expressed as follows:
F a = F σ ( D e n s e ( A v g P o o l ( F ) ) )
F v = F a σ ( C o n v ( F a ) )
F is the image information extracted by the preprocessing method described in Section 3.2.1, A v g P o o l is one-dimensional global average pooling, D e n s e represents the fully connected layer, σ represents the S i g m o i d function, F a represents the output of the image features F after passing through the channel attention, C o n v represents a one-dimensional convolution operation, and F v represents the output of the spatial attention.
The Self_Atten Block. The Self_Atten block applies the self-attention mechanism to encode text features, which can take into account the mutual influences among video segments. Since words in the same sentence in a video have different semantic associations, the self-attention mechanism can calculate the semantic associations between a word in a sentence and other words in the same sentence, providing them with different weights. This process can be expressed as follows:
m i = x i x i T
n i = s o f t m a x ( m i )
o i = n i x i
a i = o i x i
t i = x i a i
x i represents the text feature vectors extracted by the preprocessing method described in Section 3.2.1, a i is the output after self-attention and represents the importance of the different words in each utterance, t i is the output of the text features after passing through the Self_Atten block.

3.2.3. Contextual Feature Extraction

In order to capture the contexts and dependencies between utterances in each modality, we feed the single-modal attention features extracted in Section 3.2.2 for both the visual and text modalities and the preprocessed acoustic features to a B i G R U module separately. The B i G R U module consists of two Gated Recurrent Units (GRUs) with opposite directions, which can effectively capture the spatio-temporal information between video clip sequences and can also capture the forward and backward long-term dependencies between video clip sequences. The working principles of B i G R U can be expressed as follows:
r t = σ ( W r x t + U r h t 1 + b r )
z t = σ ( W z x t + U z h t 1 + b z )
h t ˜ = t a n h ( r t U h t 1 + W x t + b h )
h t = ( 1 z t ) h t 1 + z t h t
x t is the input feature sequence of the current node, h t 1 is the hidden layer state of the previous GRU unit of the current node; r t and z t are the reset gate and update gate of the GRU unit; W r , b r , W z , b z , U r , and U z are the weight parameters of the target network; σ is the corresponding s i g m o i d function of the network; and ∗ represents the multiplication of the corresponding feature vectors.
To better handle the heterogeneity within each modality and fully explore the internal correlations of single-modal features, the representation of each modality obtained from the above BiGRU module is unfolded temporally and fused through fully connected dense layer(s). This can be expressed as follows:
D t = t a n h ( W t B t + b t )
D a = t a n h ( W a B a + b a )
D v = t a n h ( W v B v + b v )
B t , B a , and B v are the output features of the text, audio, and video after going through the B i G R U module; W t , b t , W a , b a , W v , and b v are the weight parameters of the target network; D t R u × d , D a R u × d , D v R u × d are the fully connected layer’s text, acoustic, and visual information; u represents the total number of sentences; and d represents the number of neurons in the fully connected layer.

3.2.4. Multi-Modal Feature Fusion

This module comprises attention modules at four different levels and dimensions: (1) the sentence-level self-attention module; (2) the bi-modality attention module; (3) the tri-modality attention module; and (4) the self-attention module on (3). It is based on the outputs from Section 3.2.3.
(1) SL_SAtten Module. The contribution of internal features in each modality to users’ emotional tendencies is often different. For example, in the sentence, “The weather is really good today! I really like this kind of weather”, the word “like” contributes more to the users’ emotions than the word “weather”. Therefore, we propose a sentence-level self-attention mechanism, referred to as S L S A t t e n , to select the emotional contributions of words within the modality at the sentence level. Taking the text modality as an example, we assume there are u sentences in the text modality in total. For each sentence x i , where 1 i u , the working principle of the SL_SAtten module is as follows:
m i = x i x i T
n i = s o f t m a x ( m i )
o i = n i x i
a i = o i x i
T = a 1 a 2 a u
O S L S A t t e n = T V A
Here, a i is the output of the discourse x i after self-attention, indicating the importance of different words in each utterance. Then, we concatenate all of the outputs to obtain the corresponding text feature T. In the same vein, you can obtain the corresponding visual feature V and the corresponding audio feature A. By concatenating the text, visual, and audio features, you can obtain the final output of the SL_SAtten module O S L _ S A t t e n .
(2) The Bi_Atten Module. In order to improve the interactions between pairs of modalities in video data, a bi-modality attention module is proposed. This module aims to integrate two different modalities from different semantic spaces by enhancing the connections between them while eliminating the interference from background information in them, and thus the learning process will be able to focus on the associations between them. Taking text and visual modalities as an example, we suppose D t and D v are the output feature vectors of the text and visual modalities after going through the context feature extraction module, so its working principle is as follows:
m 1 = D t D v T
n 1 = s o f t m a x ( m 1 )
o 1 = n 1 D v
a 1 = o 1 D t
m 2 = D v D t T
n 2 = s o f t m a x ( m 2 )
o 2 = m 2 D t
a 2 = o 2 D v
O B C A t t e n ( v t ) = a 1 a 2
O B C _ A t t e n ( v t ) is the output feature vector fused from the video and text modalities, which can be used for subsequent emotion classification. In a similar way, we can obtain O B C _ A t t e n ( a v ) for acoustic and visual modalities and O B C _ A t t e n ( a t ) for acoustic and text modalities.
(3) The Tri_Atten Module. In order to model the interactions among all three modalities in a video, a tri-modal attention is proposed. Assuming the text, visual, and acoustic feature vectors after the context feature extraction module are D t , D v , and D a , respectively, we first concatenate and fuse the text and image, text and acoustic, and image and acoustic modality information. Then, we use a fully connected network to map the information to the same semantic space, thereby initially fusing the information between different modality pairs. The process is shown in the following formulas:
F T V = t a n h ( ( D T D V ) W t v + b t v )
F T A = t a n h ( ( D T D A ) W t a + b t a )
F A V = t a n h ( ( D A D V ) W a v + b a v )
W t v , W t a , W a v , b t v , and b t a , b a v are the weights and biases of the fully connected layer, and F T V , F T V , F T V R u × d represents the pairwise fused feature vectors, where d is the number of neurons in the fully connected layer.
In order to further extract effective features, the feature vector of the third modality is multiplied by the results of the pairwise fused feature vectors obtained in Equations (30)–(32) to produce the matrix C k ( k = 1 , 2 , 3 ) . Then, the s o f t m a x function is used to calculate the attention distribution of the feature vector fusion results P k ( k = 1 , 2 , 3 ) , forming the tri-modal attention module T k ( k = 1 , 2 , 3 ) , and finally, we obtain the tri-modal fusion information T r i A T V , T r i V T A , T r i T A V through matrix multiplication operations. This is then concatenated to form a feature vector O T r i _ A t t e n , which is the output of the Tri_Atten module. The process is shown in the following formulas:
C 1 = F A F T V T
C 2 = F V F T A T
C 3 = F T F A V T
P 1 = s o f t m a x ( C 1 )
P 2 = s o f t m a x ( C 2 )
P 3 = s o f t m a x ( C 3 )
T 1 = P 1 F A
T 2 = P 2 F V
T 3 = P 3 F T
T r i A T V = T 1 F T V
T r i V T A = T 2 F T A
T r i T A V = T 3 F A V
O T r i _ A t t e n = T r i A T V T r i V T A T r i T A V
C 1 , C 2 , C 3 R u × u , T 1 , T 2 , T 3 R u × d , and O T r i _ A t t e n R u × 3 d .
(4) Self_Atten Module. The output of the Tri_Atten module may carry redundant features. To filter out redundant information, we apply a self-attention module for feature selection. This process is shown in the following formulas:
m i = O T r i _ A t t e n O T r i _ A t t e n T
n i = s o f t m a x ( m i )
o i = n i x i
a i = o i x i
O S e l f _ A t t e n = x i a i
O S e l f _ A t t e n is the output feature vector of the Self_Atten module.

3.2.5. Multi-Modal Sentiment Classification

Finally, the multi-modal sentiment classification module concatenates and combines the feature vectors obtained above and uses a fully connected layer to integrate and classify sentiments based on both inter-modal and intra-modal information. It is shown as follows:
o u t = D v D t D a O S L S A t t e n O C S S A t t e n O B C A t t e n ( v t ) O B C A t t e n ( a t ) O B C A t t e n ( a v ) O S e l f A t t e n
o u t p u t = s o f t m a x ( o u t )
o u t p u t is the final output information for the MAVMF model.

4. Experiments

The experiments were conducted on a Windows system using an NVIDIA GeForce RTX 2060 graphics card with 8G running memory. Python was used as the programming language with the Keras framework. The effectiveness of the MAVMF model was validated on the two benchmark datasets, CMU-MOSI and CMU-MOSEI.

4.1. Data

(1) The CMU-MOSI dataset includes 93 videos sourced from YouTube, covering topics such as movies, products, and books. There are a total of 2199 utterances, each of which is labeled as positive or negative. In the experiment, we used training and test sets containing 62 and 31 videos, respectively.
(2) The CMU-MOSEI dataset includes 3229 videos with a total of 22,676 utterances, each with an emotional score in the range of [ 3 , + 3 ] . For the purpose of sentiment classification, utterances with a score of greater than or equal to 0 were labeled as positive, while those with scores less than 0 were labeled as negative. In the experiment, we used training, test, and validation sets containing 2250, 679, and 300 videos, respectively. Detailed information about the CMU-MOSI and CMU-MOSEI datasets is shown in Table 1:
From the detailed information about the CMU-MOSE and CMU-MOSEI datasets presented in Table 1, it is apparent that the number of positive utterance samples in these two datasets is greater than the number of negative utterance samples, leading to an imbalance in the distribution of positive and negative samples. Therefore, the accuracy and F-1 scores are used as evaluation metrics for the models.

4.2. Parameter Tuning

During the experiments, we investigated the impacts of different learning rates and batch sizes on the model performance. The learning rates chosen were 0.05 , 0.01 , 0.005 , and 0.001 , and the batch sizes chosen were 32 and 64. The parameters that yielded the best results were used for the final model. The final parameter settings used for the MAVMF model are shown in Table 2:

4.3. Baseline Models

To compare the performances of the MAVMF model in the MSA tasks, for the CMU-MOSI dataset, we used the following baseline methods:
(1)
GME-LSTM [48]: This model is composed of two modules. One is the gated multi-modal embedding module, which can perform information fusion in noisy environments. The other is an LSTM module with temporal attention, which can perform word-level fusion with a higher fusion resolution.
(2)
MARN [51]: This model captures the inter-relationships among text, images, and speech in a time series through multiple attention modules and stores the information in a long short-term hybrid memory.
(3)
TFN [47]: This model encodes intra-modal and inter-modal information by embedding subnetworks within a single modality and tensor fusion strategy.
(4)
MFRN [58]: This model first stores the modality information through a long short-term fusion memory network and fully considers the information of other modalities when encoding a certain modality, thereby enhancing the modality interactivity. Then, it further considers the information of other modalities when encoding a single modality through a modality fusion recurrent network. Finally, further information fusion is achieved through the attention mechanism.
(5)
Multilogue-Net [59]: Based on a recurrent neural network, this model captures the context of utterances and the relevance of the current speaker and listener in the utterance through multi-modal information.
(6)
DialogueRNN [60]: This model tracks the states of independent parties throughout the dialogue process and processes the information through a global GRU, party GRU, and emotion GRU units and uses it for emotion classification.
(7)
AMF-BiGRU [61]: This model first extracts the connections between contexts in each modality through the BiGRU, merges information through cross-modal attention, and finally uses multi-modal attention to select contributions from the merged information.
For the CMU-MOSEI dataset, we have the following baseline methods:
(1)
MFRN [58]: As described above.
(2)
Graph-MFN [62]: This model’s concept is similar to that of the MFN model, except that Graph-MFN uses a dynamic fusion graph to replace the fusion block in the MFN model.
(3)
CIM-Att [63]: This model first uses the BiGRU to extract the intra-modal context features, inputs these context features into the CIM attention module to capture the associations between pairwise modalities, and then concatenates the context features and CIM module features for sentiment classification.
(4)
AMF-BiGRU [61]: As described above.
(5)
MAM [64]: This model first uses the CNN and BiGRU to extract features from the text, speech, and image signals and then applies cross-modal attention and self-attention for information fusion and contribution selection.

5. Results

5.1. CMU-MOSI

Table 3 presents a comparison of the experiments between the MAVMF model and the chosen baseline models on the CMU-MOSI dataset. The MAVMF model shows some improvements in both the classification accuracy and F-1 score. Specifically, the accuracy of the MAVMF model is increased by 5.81 % , 5.21 % , 5.21 % , 4.21 % , 1.12 % , and 2.51 % , 0.26 % when compared to the GME-LSTM, MARN, TFN, MFRN, Multilogue-Net, DialogueRNN, and AMF-BiGRU models, respectively. The F-1 score of the MAVMF model is increased by 8.8 % , 5.2 % , 4.3 % , 4.3 % , 2.1 % , 2.36 % , and 0.18 % when compared to the GME-LSTM, MARN, TFN, MFRN, Multilogue-Net, DialogueRNN, and AMF-BiGRU models, respectively.

5.2. CMU-MOSEI

Table 4 presents a comparison of the experiments conducted between the MAVMF model and the chosen baseline models on the CMU-MOSEI dataset. The MAVMF model shows some improvements in both the classification accuracy and F-1 score. Specifically, the accuracy of the MAVMF model is increased by 3.2%, 4.2%, 1.3%, 2.26%, and 0.1% when compared to the MFRN, Graph-MFN, CIM-Att, AMF-BiGRU, and MAM models, respectively. The F-1 score of the MAVMF model is increased by 2.08%, 2.48%, 1.88%, 1.3%, and 0.58% when compared to the MFRN, Graph-MFN, CIM-Att, AMF-BiGRU, and MAM models, respectively.

5.3. Modality Analysis

In order to further analyze the impacts of features from different modalities on the classification performance of the MAVMF model, experiments were conducted on the CMU-MOSI dataset for both bi-modal and tri-modal feature sets. The experimental results are shown in Figure 3. We use T, V, and A to represent the text, visual, and acoustic modalities, respectively.
From the above results, when compared to the selected baseline models, the classification accuracy of the text plus acoustic modalities is increased by 0.53–1.09%, the classification accuracy of the text plus visual modalities is increased by 0.27–2.35%, and the classification accuracy of the acoustic plus visual plus text modalities is increased by 0.4–2.51%. Apart from the fusion results for the video and acoustic modalities, the MAVMF model achieves the best performance for all other modality fusion methods. The fusion result for the acoustic and visual features is the worst, reflecting that the emotional expression polarity of the acoustic and visual modalities is weaker than that of text and that these modalities may be affected by background noise. This is consistent with the experimental results presented in the literature [65]. In addition, in MSA tasks, the classification performance obtained by fusing all three modalities is the best. This proves the necessity of leveraging multi-modal information in SA.

5.4. Ablation Study

To understand the impacts of the different modules applied in the MAVMF model, we conducted experiments on variants of the MAVMF model using the CMU-MOSI dataset and analyzed the experimental results. We included the following MAVMF variant models:
(1) MAVMF_Concat: This includes Modules A and E, as shown in Figure 2. (2) MAVMF_SAtten: This includes Module A, the self-attention module in Module B for the text modality, and Module E, as shown in Figure 2. (3) MAVMF_RAtten: This includes Module A, the reduced-attention module in Module B for the visual modality, and Module E, as shown in Figure 2. (4) MAVMF_RAtten_SAtten: This includes Module A, Module B, and Module E, as shown in Figure 2. (5) MAVMF_BiGRU: This includes Module A, Module B, Module C, and Module E, as shown in Figure 2. (6) MAVMF_SL-SAtten: This includes Module A, Module B, and Module C, a sentence level self-attention module, and Module E, as shown in Figure 2. (7) MAVMF_Bi-Atten: This includes Module A, Module B, and Module C, a sentence level self-attention module, a dual modality cross modal attention module, and Module E, as shown in Figure 2. (8) MAVMF_Tri-Atten: This includes Module A, Module B, Module C, a sentence level self-attention module, a dual-modality cross-modal attention module, a tri-modality cross-modal attention module, and Module E, as shown in Figure 2. (9) MAVMF_Self-Atten: This includes Module A, Module B, Module C, Module D, and Module E, as shown in Figure 2.
Table 5 compares the proposed MAVMF model with its variant on the CMU-MOSI dataset. From the experiments, we see that the multi-modal sentiment classification accuracy of the MAVMF model gradually improves after adding each module. The text self-attention module, visual reduced-attention module, single-modality attention module, bidirectional gated recurrent unit module, sentence-level self-attention module, dual-modality cross-modal attention module, tri-modality cross-modal attention module, and self-attention module contribute 2.27 % , 2.53 % , 4.26 % , 5.32 % , 0.4 % , 0.4 % , 0.26 % , 0.4 % , and 0.53 % to the classification accuracy, respectively. The marginal improvements become smaller as the complexity of the model increases.

6. Discussion

Multi-Modal Sentiment Analysis tasks are commonplace in a diverse array of application scenarios. Video-based social media platforms, in particular, have empowered general users to generate an unprecedented amount of multi-modal data such as texts, audios, images, and the various combinations of them, which have enabled developers and practitioners to create multi-modal artificial intelligence systems that have already transformed our lives and work, as has been witnessed in the current wave of generative AI applications.
The work in this paper provides new insights into the fusion of multi-modal data for more general tasks beyond sentiment analysis. Our proposed MAVMF algorithm systematically explore the vast feature spaces that are generated by different modalities and their inherent spatial and temporal relationships. Unfortunately, when compared with other AI tasks such as image recognition and machine translation, the task of fusing multi-modal information for sentiment analysis remains an unsolved problem.
Currently, the underpinning theory regarding how different text, audio, and visual modalities complement or interfere with each other has not been formalized. However, one future direction in terms of solving the multi-modal fusion challenge could be relying on computation power, similar to what we have observed in large language models (LLMs). Doing this without leveraging a larger network with higher computation power may pose limits on the current performance of our proposed method. In other words, we did not consider constraints on the algorithm speed given a limited computation capacity, as the full MAVMF model combines features from all proposed modules shown in Figure 2. However, we believe that, as computation is becoming cheaper and more accessible over time, there should be a focus on network design and modal fusion that encompasses all possible interactions among different modalities.
In our next step, we plan to investigate the performance of large foundation models in MSA tasks. Given that the current state-of-the-art performance in the existing literature on MSA tasks is still in the 80% range, we will first focus on improving the accuracy and robustness of novel algorithms for MSA tasks. Through pretraining with large amounts of unlabeled text and image data, foundation models have implicitly encoded large amounts of human knowledge in their weight parameters. It may be possible to adopt them when addressing MSA tasks. Certainly, we can also explore knowledge distillation techniques on successful foundation-model-based algorithms to obtain compact apprentice models for more resource constraint scenarios in MSA tasks.

7. Conclusions

To address the effective extraction of single-modal features and the efficient integration of multi-modal information, we propose an MSA algorithm, MAVMF. First, feature extractors are used to capture single-modal information. Then, for single-modal features, a reduced-attention module is used to encode the image, while a self-attention module is used for text. Subsequently, a bidirectional GRU and a fully connected network are applied to extract context-aware discourse features, capturing the context information between discourses in each modality. A sentence-level self-attention module is then used to model different types of modality information. At the modality level, dual-modality and tri-modality attention modules are applied to merge information, and a self-attention module is used to select features with significant contributions to reveal sentiment tendencies. Experiments on public datasets prove that, when compared to other deep learning algorithms, the MAVMF model has a better or comparable classification performance.

Author Contributions

Conceptualization, C.H., L.C. and D.Z.; methodology, C.H. and D.S.; software, C.H. and D.S.; validation, C.H. and D.S.; formal analysis, C.H. and D.S.; investigation, C.H.; resources, L.C., D.Z., X.Z., Y.S., C.M. and H.W.; data curation, C.H.; writing—original draft preparation, C.H., D.S., L.C., X.Z., Y.S., C.M. and H.W.; writing—review and editing, L.C., D.Z. and X.Z.; visualization, C.H., D.S. and X.Z.; supervision, L.C. and D.Z.; project administration, L.C. and D.Z.; funding acquisition, D.Z. Special thanks to Xinghua Zhang for valuable contributions to the graphical work. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding. The APC was funded by Prof. Dingju Zhu.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Both the datasets applied in the current research are open datasets and can be found at the given link: http://multicomp.cs.cmu.edu/resources/ accessed on 1 January 2020. The datasets are available for downloading through the CMU Multimodal Data SDK GitHub: https://github.com/CMU-MultiComp-Lab/CMU-MultimodalSDK accessed on 1 January 2020.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Morency, L.P.; Mihalcea, R.; Doshi, P. Towards multimodal sentiment analysis: Harvesting opinions from the web. In Proceedings of the 13th International Conference on Multimodal Interfaces, Alicante, Spain, 14–18 November 2011; pp. 169–176. [Google Scholar]
  2. Zadeh, A.; Zellers, R.; Pincus, E.; Morency, L.P. Multimodal sentiment intensity analysis in videos: Facial gestures and verbal messages. IEEE Intell. Syst. 2016, 31, 82–88. [Google Scholar] [CrossRef]
  3. Poria, S.; Cambria, E.; Bajpai, R.; Hussain, A. A review of affective computing: From unimodal analysis to multimodal fusion. Inf. Fusion 2017, 37, 98–125. [Google Scholar] [CrossRef]
  4. Prakash, A.; Chitta, K.; Geiger, A. Multi-modal fusion transformer for end-to-end autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7077–7087. [Google Scholar]
  5. Li, X.; Lu, G.M.; Yan, J.J.; Zhang, Z.Y. A Comprehensive Review on Multimodal Dimensional Emotion Prediction. Acta Autom. Sin. 2018, 44, 2142–2159. [Google Scholar]
  6. Grimaldo, F.; Lozano, M.; Barber, F. MADeM: A multi-modal decision making for social MAS. In Proceedings of the AAMAS (1), Estoril, Portugal, 12–16 May 2008; pp. 183–190. [Google Scholar]
  7. Zhu, L.; Zhu, Z.; Zhang, C.; Xu, Y.; Kong, X. Multimodal sentiment analysis based on fusion methods: A survey. Inf. Fusion 2023, 95, 306–325. [Google Scholar] [CrossRef]
  8. Song, X.J. A Study on Multimodal Emotion Recognition Based on Text, Speech, and Video. Master’s Thesis, Shandong University, Shandong, China, 2019. [Google Scholar]
  9. Liu, B. Sentiment Analysis and Opinion Mining; Springer Nature: Berlin/Heidelberg, Germany, 2022. [Google Scholar]
  10. Ting, W.; Yang, W.Z. A Review of Text Sentiment Analysis Methods. J. Comput. Eng. Appl. 2021, 57, 11–24. [Google Scholar]
  11. Lin, J.H.; Gu, Y.L.; Zhou, Y.M.; Yang, A.M.; Chen, J. A Study on Constructing an Emotion Dictionary Based on Emoji. Comput. Technol. Dev. 2019, 29, 181–185. [Google Scholar]
  12. Mike, T.; Kevan, B.; Georgios, P.; Di, C.; Arvid, K. Sentiment in short strength detection informal text. JASIST 2010, 61, 2544–2558. [Google Scholar]
  13. Saif, H.; He, Y.; Fernandez, M.; Alani, H. Contextual semantics for sentiment analysis of Twitter. Inf. Process. Manag. 2016, 52, 5–19. [Google Scholar] [CrossRef]
  14. Li, Y.S.; Wang, L.M.; Chai, Y.M.; Liu, Z. A Study on Dynamic Emotion Dictionary Construction Method Based on Bidirectional LSTM. Microcomput. Syst. 2019, 40, 503–509. [Google Scholar]
  15. Kanayama, H.; Nasukawa, T. Fully automatic lexicon expansion for domain-oriented sentiment analysis. In Proceedings of the 2006 Conference on Empirical Methods in Natural Language Processing, Sydney, Australia, 22–23 July 2006; pp. 355–363. [Google Scholar]
  16. Rao, Y.; Lei, J.; Wenyin, L.; Li, Q.; Chen, M. Building emotional dictionary for sentiment analysis of online news. World Wide Web 2014, 17, 723–742. [Google Scholar] [CrossRef]
  17. Qi, C.; Li, Z.; Jing, J.; Huang, X.Y. A Review Analysis Method Based on Support Vector Machine and Topic Model. J. Softw. 2019, 30, 1547–1560. [Google Scholar]
  18. Gang, Z.; Zan, X. A Study on Sentiment Analysis Model of Product Reviews Based on Machine Learning. Res. Inf. Secur. 2017, 3, 166–170. [Google Scholar]
  19. Kiritchenko, S.; Zhu, X.; Mohammad, S.M. Sentiment analysis of short informal texts. J. Artif. Intell. Res. 2014, 50, 723–762. [Google Scholar] [CrossRef]
  20. Da Silva, N.F.; Hruschka, E.R.; Hruschka, E.R., Jr. Tweet sentiment analysis with classifier ensembles. Decis. Support Syst. 2014, 66, 170–179. [Google Scholar] [CrossRef]
  21. Kim, Y. Convolutional neural networks for sentence classification. arXiv 2014, arXiv:1408.5882. [Google Scholar]
  22. Okada, M.; Yanagimoto, H.; Hashimoto, K. Sentiment Classification with Gated CNN and Spatial Pyramid Pooling. In Proceedings of the 2018 7th International Congress on Advanced Applied Informatics (IIAI-AAI), Yonago, Japan, 8–13 July 2018; pp. 133–138. [Google Scholar]
  23. Meng, J.; Long, Y.; Yu, Y.; Zhao, D.; Liu, S. Cross-domain text sentiment analysis based on CNN_FT method. Information 2019, 10, 162. [Google Scholar] [CrossRef]
  24. Jiang, M.; Zhang, W.; Zhang, M.; Wu, J.; Wen, T. An LSTM-CNN attention approach for aspect-level sentiment classification. J. Comput. Methods Sci. Eng. 2019, 19, 859–868. [Google Scholar] [CrossRef]
  25. Zhou, H.; Yu, Y.; Jia, Y.Y.; Zhao, W.L. A Study on Sentiment Classification of Online Consumer Reviews Based on Deep LSTM Neural Network. Chin. J. Med Libr. Inf. 2018, 27, 23–29. [Google Scholar]
  26. Luo, L.X. Network text sentiment analysis method combining LDA text representation and GRU-CNN. Pers. Ubiquitous Comput. 2019, 23, 405–412. [Google Scholar] [CrossRef]
  27. Minh, D.L.; Sadeghi-Niaraki, A.; Huy, H.D.; Min, K.; Moon, H. Deep learning approach for short-term stock trends prediction based on two-stream gated recurrent unit network. IEEE Access 2018, 6, 55392–55404. [Google Scholar] [CrossRef]
  28. Zhang, Y.; Jiang, Y.; Tong, Y. Study of sentiment classification for Chinese microblog based on recurrent neural network. Chin. J. Electron. 2016, 25, 601–607. [Google Scholar] [CrossRef]
  29. Colombo, C.; Del Bimbo, A.; Pala, P. Semantics in visual information retrieval. IEEE Multimed. 1999, 6, 38–53. [Google Scholar] [CrossRef]
  30. Jindal, S.; Singh, S. Image sentiment analysis using deep convolutional neural networks with domain specific fine tuning. In Proceedings of the 2015 International Conference on Information Processing (ICIP), Pune, India, 16–19 December 2015; pp. 447–451. [Google Scholar]
  31. Yang, J.; She, D.; Sun, M.; Cheng, M.M.; Rosin, P.L.; Wang, L. Visual sentiment prediction based on automatic discovery of affective regions. IEEE Trans. Multimed. 2018, 20, 2513–2525. [Google Scholar] [CrossRef]
  32. Yang, J.; She, D.; Lai, Y.K.; Rosin, P.L.; Yang, M.H. Weakly supervised coupled networks for visual sentiment analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Lake Salt City, UT, USA, 18–22 June 2018; pp. 7584–7592. [Google Scholar]
  33. Kumar, A.; Jaiswal, A. Image sentiment analysis using convolutional neural network. In Proceedings of the Intelligent Systems Design and Applications: 17th International Conference on Intelligent Systems Design and Applications (ISDA 2017), Delhi, India, 14–16 December 2017; Springer: Berlin/Heidelberg, Germany, 2018; pp. 464–473. [Google Scholar]
  34. Truong, Q.T.; Lauw, H.W. Visual sentiment analysis for review images with item-oriented and user-oriented CNN. In Proceedings of the 25th ACM International Conference on Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 1274–1282. [Google Scholar]
  35. You, Q.; Jin, H.; Luo, J. Visual sentiment analysis by attending on local image regions. In Proceedings of the AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; Volume 31. [Google Scholar]
  36. Wu, L.; Qi, M.; Jian, M.; Zhang, H. Visual sentiment analysis by combining global and local information. Neural Process. Lett. 2020, 51, 2063–2075. [Google Scholar] [CrossRef]
  37. Zheng, R.; Li, W.; Wang, Y. Visual sentiment analysis by leveraging local regions and human faces. In Proceedings of the MultiMedia Modeling: 26th International Conference, MMM 2020, Daejeon, Republic of Korea, 5–8 January 2020; Part I 26. Springer: Berlin/Heidelberg, Germany, 2020; pp. 303–314. [Google Scholar]
  38. Li, L.; Li, S.; Cao, D.; Lin, D. SentiNet: Mining visual sentiment from scratch. In Proceedings of the Advances in Computational Intelligence Systems: Contributions Presented at the 16th UK Workshop on Computational Intelligence, Lancaster, UK, 7–9 September 2016; Springer: Berlin/Heidelberg, Germany, 2017; pp. 309–317. [Google Scholar]
  39. Li, W.F. A Study on Social Emotion Classification Based on Multimodal Fusion. Master’s Thesis, Chongqing University of Posts and Telecommunications, Chongqing, China, 2019. [Google Scholar]
  40. Navas, E.; Hernáez, I.; Luengo, I. An objective and subjective study of the role of semantics and prosodic features in building corpora for emotional TTS. IEEE Trans. Audio Speech Lang. Process. 2006, 14, 1117–1127. [Google Scholar] [CrossRef]
  41. Xu, X.; Hu, Y.C.; Wang, Q.M. Speech Emotion Recognition System and Method Based on Machine Learning. Available online: https://wenku.baidu.com/view/8469574cb2717fd5360cba1aa8114431b80d8ed4?fr=xueshu_top&_wkts_=1706505509577&needWelcomeRecommand=1 (accessed on 1 January 2020).
  42. Li, B.; Dimitriadis, D.; Stolcke, A. Acoustic and lexical sentiment analysis for customer service calls. In Proceedings of the ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 5876–5880. [Google Scholar]
  43. Li, W.Q. A Comparative Study of Speech Enhancement Algorithms and Their Applications in Feature Extraction. Master’s Thesis, Shandong University, Shandong, China, 2020. [Google Scholar]
  44. He, J.; Liu, Y.; He, Z.W. Advances in Multimodal Emotion Recognition. Appl. Res. Comput. Jisuanji Yingyong Yanjiu 2018, 35, 3201–3205. [Google Scholar]
  45. Poria, S.; Cambria, E.; Gelbukh, A. Deep convolutional neural network textual features and multiple kernel learning for utterance-level multimodal sentiment analysis. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 2539–2544. [Google Scholar]
  46. Zadeh, A.; Zellers, R.; Pincus, E.; Morency, L.P. Mosi: Multimodal corpus of sentiment intensity and subjectivity analysis in online opinion videos. arXiv 2016, arXiv:1606.06259. [Google Scholar]
  47. Zadeh, A.; Chen, M.; Poria, S.; Cambria, E.; Morency, L.P. Tensor fusion network for multimodal sentiment analysis. arXiv 2017, arXiv:1707.07250. [Google Scholar]
  48. Chen, M.; Wang, S.; Liang, P.P.; Baltrušaitis, T.; Zadeh, A.; Morency, L.P. Multimodal sentiment analysis with word-level fusion and reinforcement learning. In Proceedings of the 19th ACM International Conference on Multimodal Interaction, Glasgow, UK, 13–17 November 2017; pp. 163–171. [Google Scholar]
  49. Poria, S.; Cambria, E.; Hazarika, D.; Majumder, N.; Zadeh, A.; Morency, L.P. Context-dependent sentiment analysis in user-generated videos. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long papers), Vancouver, Canada, 30 July–4 August 2017; pp. 873–883. [Google Scholar]
  50. Poria, S.; Peng, H.; Hussain, A.; Howard, N.; Cambria, E. Ensemble application of convolutional neural networks and multiple kernel learning for multimodal sentiment analysis. Neurocomputing 2017, 261, 217–230. [Google Scholar] [CrossRef]
  51. Zadeh, A.; Liang, P.P.; Poria, S.; Vij, P.; Cambria, E.; Morency, L.P. Multi-attention recurrent network for human communication comprehension. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  52. Wang, D.; Guo, X.; Tian, Y.; Liu, J.; He, L.; Luo, X. TETFN: A text enhanced transformer fusion network for multimodal sentiment analysis. Pattern Recognit. 2023, 136, 109259. [Google Scholar] [CrossRef]
  53. Yang, B.; Wu, L.; Zhu, J.; Shao, B.; Lin, X.; Liu, T.Y. Multimodal sentiment analysis with two-phase multi-task learning. IEEE/ACM Trans. Audio Speech Lang. Process. 2022, 30, 2015–2024. [Google Scholar] [CrossRef]
  54. Wu, T.; Peng, J.; Zhang, W.; Zhang, H.; Tan, S.; Yi, F.; Ma, C.; Huang, Y. Video sentiment analysis with bimodal information-augmented multi-head attention. Knowl. Based Syst. 2022, 235, 107676. [Google Scholar] [CrossRef]
  55. Wang, Y.; Li, Y.; Bell, P.; Lai, C. Cross-Attention is Not Enough: Incongruity-Aware Multimodal Sentiment Analysis and Emotion Recognition. arXiv 2023, arXiv:2305.13583. [Google Scholar]
  56. He, Y.; Sun, L.; Lian, Z.; Liu, B.; Tao, J.; Wang, M.; Cheng, Y. Multimodal Temporal Attention in Sentiment Analysis. In Proceedings of the 3rd International on Multimodal Sentiment Analysis Workshop and Challenge, Lisboa, Portugal, 10 October 2022; pp. 61–66. [Google Scholar]
  57. Mai, S.; Zeng, Y.; Zheng, S.; Hu, H. Hybrid contrastive learning of tri-modal representation for multimodal sentiment analysis. IEEE Trans. Affect. Comput. 2022, 14, 2276–2289. [Google Scholar] [CrossRef]
  58. Liu, Q. Study on Emotion Analysis Method Based on Multimodal Information Fusion. 2019. Available online: https://kns.cnki.net/kcms2/article/abstract?v=dFlgZ3unFPiOMAnTaqVHESvFy9yv01Hjk3IFI1xuIzn5BPQf5EAkVa1UDnqDorqJ7K6I8-P3WE6Wl9Yof-6g4u6lJPeMLt7zn8B0OubMKMFrUmR95rJDYPPvuBYrEVSuNDZcWhNIViNlwlDmr0ElJA==&uniplatform=NZKPT&language=CHS (accessed on 1 January 2020).
  59. Shenoy, A.; Sardana, A. Multilogue-net: A context aware rnn for multi-modal emotion detection and sentiment analysis in conversation. arXiv 2020, arXiv:2002.08267. [Google Scholar]
  60. Majumder, N.; Poria, S.; Hazarika, D.; Mihalcea, R.; Gelbukh, A.; Cambria, E. Dialoguernn: An attentive rnn for emotion detection in conversations. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 6818–6825. [Google Scholar]
  61. Lai, X.M.; Tang, H.; Chen, H.Y.; Li, S.S. Feature Fusion Based on Attention Mechanism - Multi-modal Emotion Analysis Using Bidirectional Gated Recurrent Unit. Comput. Appl. 2021, 41, 1268. [Google Scholar]
  62. Zadeh, A.B.; Liang, P.P.; Poria, S.; Cambria, E.; Morency, L.P. Multimodal language analysis in the wild: Cmu-mosei dataset and interpretable dynamic fusion graph. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Melbourne, Australia, 15–20 July 2018; pp. 2236–2246. [Google Scholar]
  63. Akhtar, M.S.; Chauhan, D.S.; Ghosal, D.; Poria, S.; Ekbal, A.; Bhattacharyya, P. Multi-task learning for multi-modal emotion recognition and sentiment analysis. arXiv 2019, arXiv:1905.05812. [Google Scholar]
  64. Song, Y.F.; Ren, G.; Yang, Y.; Fan, X.C. Multi-task Multi-modal Emotion Analysis Based on Attention-driven Multilevel Hybrid Fusion. Appl. Res. Comput. Jisuanji Yingyong Yanjiu 2022, 39. [Google Scholar]
  65. Bao, G.B.; Li, G.L.; Wang, G.X. Bimodal Interaction Attention for Multi-modal Emotion Analysis. Comput. Sci. Explor. 2022, 16, 909. [Google Scholar]
Figure 1. Selected samples from the CMU-MOSI dataset. Each sample represents an utterance from the given video. The output labels were obtained by using our proposed MAVMF algorithm.
Figure 1. Selected samples from the CMU-MOSI dataset. Each sample represents an utterance from the given video. The output labels were obtained by using our proposed MAVMF algorithm.
Bdcc 08 00014 g001
Figure 2. MAVMF architecture. (A) Preprocessing the input multi-modal data. (B) Single-modality self-attention and the reduced-attention module for the text and visual modalities. (C) BiGRU + fully connected dense layer(s). (D) Four attention and self-attention modules for modal fusion. (E) Concatenation + fully connected dense layer(s) with the softmax activation function for predictions.
Figure 2. MAVMF architecture. (A) Preprocessing the input multi-modal data. (B) Single-modality self-attention and the reduced-attention module for the text and visual modalities. (C) BiGRU + fully connected dense layer(s). (D) Four attention and self-attention modules for modal fusion. (E) Concatenation + fully connected dense layer(s) with the softmax activation function for predictions.
Bdcc 08 00014 g002
Figure 3. Visualization of the impacts of the modalities on the performances of the baseline algorithms and MAVMF.
Figure 3. Visualization of the impacts of the modalities on the performances of the baseline algorithms and MAVMF.
Bdcc 08 00014 g003
Table 1. The details of the CMU-MOSE and CMU-MOSEI datasets.
Table 1. The details of the CMU-MOSE and CMU-MOSEI datasets.
DescriptionCMU-MOSICMU-MOSEI
Training Set Test Set Training Set Test Set
# Video62312250679
# Utterance1447752162164625
# Pos Utterance709467114983281
# Neg Utterance73828547181344
Table 2. Experimental parameter settings.
Table 2. Experimental parameter settings.
ParameterValue
BiGRU unit300
BiGRU dropout0.5
fully connected unit100
fully connected dropout0.5
activation functiontanh
learning rate0.001
batch processing32
number of iterations64
optimization functionAdam
loss functioncategorical cross-entropy
Table 3. Comparison of the performance with different models.
Table 3. Comparison of the performance with different models.
Network ModelCMU-MOSI
Accuracy (%) F-1
GME-LSTM [48]76.5073.40
MARN [51]77.1077.00
TFN [47]77.1077.90
MFRN [58]78.1077.90
Multilogue-Net [59]81.1980.10
DialogueRNN [60]79.8079.48
AMF-BiGRU [61]82.0582.02
MAVMF82.3182.20
Table 4. Comparison of the performance on different models.
Table 4. Comparison of the performance on different models.
Network ModelCMU-MOSEI
Accuracy (%) F-1
MFRN [58]77.9077.40
Graph-MFN [62]76.9077.00
CIM-Att [63]79.8077.60
AMF-BiGRU [61]78.4878.18
MAM [64]81.0078.90
MAVMF81.1079.48
Table 5. Comparison of the performance on different models.
Table 5. Comparison of the performance on different models.
Network ModelCMU-MOSI
Accuracy (%) F-1
MAVMF_Concat70.7471.01
MAVMF_SAtten73.0173.05
MAVMF_RAtten73.2773.24
MAVMF_RAtten_SAtten75.0074.95
MAVMF_BiGRU80.3280.21
MAVMF_SL-SAtten80.7281.09
MAVMF_Bi-Atten81.1281.10
MAVMF_Tri-Atten81.3881.41
MAVMF_Self-Atten81.7882.06
MAVMF82.3182.20
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

He, C.; Zhang, X.; Song, D.; Shen, Y.; Mao, C.; Wen, H.; Zhu , D.; Cai, L. Mixture of Attention Variants for Modal Fusion in Multi-Modal Sentiment Analysis. Big Data Cogn. Comput. 2024, 8, 14. https://doi.org/10.3390/bdcc8020014

AMA Style

He C, Zhang X, Song D, Shen Y, Mao C, Wen H, Zhu  D, Cai L. Mixture of Attention Variants for Modal Fusion in Multi-Modal Sentiment Analysis. Big Data and Cognitive Computing. 2024; 8(2):14. https://doi.org/10.3390/bdcc8020014

Chicago/Turabian Style

He, Chao, Xinghua Zhang, Dongqing Song, Yingshan Shen, Chengjie Mao, Huosheng Wen, Dingju Zhu , and Lihua Cai. 2024. "Mixture of Attention Variants for Modal Fusion in Multi-Modal Sentiment Analysis" Big Data and Cognitive Computing 8, no. 2: 14. https://doi.org/10.3390/bdcc8020014

Article Metrics

Back to TopTop