Previous Article in Journal
AI-Enhanced Non-Intrusive Load Monitoring for Smart Home Energy Optimization and User-Centric Interaction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

IER-SMCEM: An Implicit Expression Recognition Model of Emojis in Social Media Comments Based on Prompt Learning

1
School of Computer Science, China West Normal University, Nanchong 637000, China
2
Education Information Technology Center, China West Normal University, Nanchong 637000, China
3
School of Electronic Information, China West Normal University, Nanchong 637000, China
*
Authors to whom correspondence should be addressed.
Informatics 2025, 12(2), 56; https://doi.org/10.3390/informatics12020056
Submission received: 5 April 2025 / Revised: 10 June 2025 / Accepted: 13 June 2025 / Published: 18 June 2025
(This article belongs to the Special Issue Practical Applications of Sentiment Analysis)

Abstract

:
Financial text analytics methods are employed to examine social media comments, allowing investors to gain insights and make informed financial decisions. Some emojis within these comments often convey diverse semantics, emotions, or intentions depending on the context. However, traditional financial text analysis methods relying on public annotations struggle to identify implicit expressions, leading to suboptimal performance. To address this challenge, this paper proposes an implicit expression recognition model of emojis in social media comments (IER-SMCEM). Firstly, IER-SMCEM innovative designs a data enhancement method based on the implicit expression of emoji. This method expands the pure text financial sentiment analysis dataset into the implicit expression dataset of emoji by homophonic replacement. Secondly, IER-SMCEM designs a prompt learning template to identify the implicit expression of emoji. Through hand-designed templates, large-scale language models can predict the true meaning that emojis are most likely to express. Finally, IER-SMCEM recovers implicit expression by choosing the predictions of models. Thus, the downstream financial sentiment analysis model can more precisely realize the sentiment recognition of the text with emoji by the recovered text. The experimental results indicate that IER-SMCEM achieves a 98.03% accuracy in semantically recovering implicit expressions within financial texts. In the task of financial sentiment analysis, the sentiment analysis model achieves the highest accuracy of 3.99% after restoring the true implied expression of the texts. Therefore, the model can be effectively applied to sentiment analysis or quantitative analysis.

1. Introduction

In recent years, financial text analytics has been extensively applied in the financial sector, including risk prediction, investment decision-making, and stock market trading [1,2,3,4]. As illustrated in Figure 1, the Stocktwits platform shows that on 19 September 2024, the share price of Avalanche Corporation (AVAX.X) experienced a significant increase [2]. Notably, the platform data indicate a strong positive correlation between stock price changes, investor sentiment, and overall chatter [3]. Consequently, financial text analytics can assist investors in more accurately predicting stock price trends by analyzing sentiment, offering substantial support for investment decisions.
Social media platforms have given rise to a new pattern of communication where market participants increasingly share their views and sentiments through emojis, as illustrated in Table 1. Emojis amplify market volatility through sentiment contagion and algorithmic interference, which may lead to follow-up trading or manipulation risk. Elon Musk’s emoji (Shiba Inu) posted on Twitter has repeatedly triggered violent fluctuations in the price of the Dogecoin. The most typical event occurred on 3 April 2023, when he changed the official DOGE of Twitter from Bluebird to Shiba Inu, which caused the price of Dogecoin to soar by over 30% in a short time. Financial researchers are highly concerned about the significant impact of emojis on the market [4,5,6]. The research found that the sentiment analysis model can associate emojis with sentiment. The association can improve the accuracy of sentiment analysis [6]. Not only that, emojis can also soften the views of non-professional investors on bad news or enhance their views on good news [7]. To determine the meaning, sentiment, or intent of emojis in financial social media posts, researchers typically rely on human annotators. This process is often time-consuming and not scalable [8]. In social media communications, emojis can convey different meanings across trading communities, influenced by varying cultural backgrounds and market contexts [4]. In Section 3, we explore how these context-specific meanings of emojis vary across different market scenarios and communities. Due to the complex nature of these community-driven interpretations, traditional analysis methods based on manual annotations may misinterpret the actual meaning of these emojis in financial contexts [9]. This misinterpretation can hinder the effectiveness of financial text analysis models (e.g., financial sentiment analysis models and investment decision-making models).
To address the problem that the implicit expression of emojis can amplify the expression of sentiments, this paper proposes an implicit expression recognition model of emojis in social media comments (IER-SMCEM). First, the model is designed with the Implicit Emoji Enhancement Algorithm (IEEA) for data enhancement. IEEA is based on the homophone of the original text between the words and the emoji substitution, allowing the emoji to express different implicit expressions beyond public annotations in the comments. Then, this paper utilizes the Implicit Emoji Prediction Module (IEPM) to integrate the content of comments with implicit expressions into the prompt template. This data is then input into a pre-trained masked language model to predict the implicit expressions conveyed by emojis. The experimental results demonstrate that predicting implicit emoji expressions can significantly enhance the performance of traditional financial sentiment analysis models when dealing with implicit expressions. The experimental results indicate that IER-SMCEM achieves a 98.03% accuracy in semantically recovering implicit expressions within financial texts. The primary contributions of this paper are outlined below. Data and codes are available at https://github.com/junzhangNLP/MIDF, accessed on 24 April 2024.
  • By applying IEEA to an existing publicly accessible dataset, we have created a sentiment analysis dataset for Chinese and English social comments. Because it is difficult to collect and process data in other languages, the target language of this study is based on Chinese and English.
  • In this paper, IEPM is designed to address the challenge of recognizing implicit expressions of emojis in financial text analysis models.
  • In this paper, we validate the positive impact of recognizing implicit expressions of emojis on financial text analysis tasks using a financial sentiment analysis model.

2. Related Work

From the perspective of semantic analysis technology, this section introduces the development process of technology and discusses the related research of emoji semantics. Then, this section discusses the influence of emojis on financial decision-making from the perspective of financial text analysis technology. In addition, this section discusses the situation of the financial dataset in detail. Based on the related research of datasets, we put forward the idea of constructing a new dataset.

2.1. Semantic Analysis Technology

Semantic analysis technology has primarily focused on document-level [10], sentence-level [11], and aspect-level [12] studies, with applications spanning social media, restaurant reviews, and product evaluations. In early research around 2003, systems based on logical expressions and syntactic rules were widely adopted. For instance, Alquézar et al. [13] proposed a method that employed handcrafted rules to structurally process text for identifying and learning context-sensitive language patterns. Park et al. [14] utilized logical expressions and syntactic rules to describe complex contextual relationships and introduced memory-based learning for more precise text chunking. Recently, researchers have explored a novel technique known as “prompt learning” within the text-mining domain for accurate textual information extraction. Liu et al. [15] introduced prompt learning as a groundbreaking paradigm in natural language processing. Central to prompt learning is the pivotal challenge of constructing prompt templates, as underscored by Liu X et al. [16], who proposed a P-tuning method that automatically searches for prompts in a continuous space. However, empirical evidence suggests that the effectiveness of this method may lag behind that of manual prompting [15,17]. In-depth investigations by numerous scholars have focused on developing superior manual prompt templates [18]. Schick et al. [19] emphasized the necessity of domain expertise and an understanding of language modeling capabilities for manually defining the mapping between words and tags. Similarly, Hu et al. [20] highlighted the importance of constructing a projection between the tag space and the tagged word space, known as verbalization, as the core step in prompt tuning. Building on these insights, this study adopts a manual approach to devise prompt templates.
With the development of social tools, emojis and pictures have become tools to express semantics. Many researchers have made contributions to the semantic extraction of emojis. The first emoji can be traced back to Japanese interface designer Kurita Satoshi [21], who created 176 unique graphic symbols for NTT DoCoMo devices in 1999. Today, 92% of the online population in the world uses emojis in their communication. With the popularity of emojis on social networks and other platforms, there is increasing interest in studying the functions of emojis [22]. Researchers have explored emojis with the function of expressing emotions, highlighting themes, decorating texts, adjusting mood, indicating identity, and attracting audiences [23]. Using these emoji functions, the researchers summarized the intention of using emoji on different platforms. In addition to these functions and intentions, researchers have also noticed the differences in the application scenarios of emojis, such as applications, culture, gender, labels, and platforms [24]. With the rise in popularity of emojis on social networks and other platforms, researchers have conducted extensive studies on their functions [25]. Lu et al. [26] explored how emojis express emotions, highlight themes, decorate text, adjust tone, indicate identity, and engage the audience. Hu et al. [22] summarized the intentions behind using emojis on various platforms based on these features. In addition to these functions and intentions, Chen et al. [27] highlighted differences in emoji usage across various application scenarios (e.g., apps, cultures, genders, hashtags, and platforms). However, much of the research on emoji functionality has relied on manual annotation for emoji interpretation. Given the richness and diversity of emoji semantics, there is an urgent need for automated tools to interpret these meanings. Zhou et al. [8] explored the use of ChatGPT (GPT-4o) for automated semantic annotation of emojis, as opposed to manual methods. They found that the semantic annotation of emojis is challenging due to their susceptibility to cultural differences, which complicates the interpretation process. In addition to semantic annotation of emojis, researchers have applied emoji information to downstream tasks such as emoji prediction and enhanced sentiment analysis [28]. Most existing emoji annotation methods rely on manual annotation. Although the emergence of LLMs has effectively alleviated this issue, its practical application remains limited due to the need for vast amounts of data and computational resources. In addition to the emoji function, some researchers have applied emoji information to downstream tasks, such as emoji prediction, improving sentiment analysis and predicting the loss of developers [29]. Recent evidence shows that emojis are also very useful for sentiment analysis [27]. Therefore, this paper considers emojis to be beneficial to financial sentiment analysis.

2.2. Financial Text Analysis and Dataset Challenge

Based on the research of semantic analysis, many researchers apply it to the financial field [30]. In the early stages of the research, sentences were first converted into vectors using embedding models or bag-of-word techniques such as GloVe, Word2Vec, and sentiment dictionaries [31]. These vectors were then classified using memory-based (recurrent) neural networks like RNNs, LSTMs, and GRUs [31,32]. In the subsequent phase, marked by the development of pre-trained models like BERT, the model achieved state-of-the-art performance in various tasks, including sentiment analysis, entity recognition, and text generation [33]. The textual data used for financial sentiment analysis include email communications, social media posts (e.g., tweets), corporate reports, and daily news [34]. Financial corpora are labeled through either manual annotation [7] or based on stock price [35]. Overall, there is one document-level, four sentence-level, two target-level, and one targeted aspect-level dataset. It is also observed that each public release of financial sentiment analysis benchmark dataset, particularly in open challenges such as FiQA Task 1 [36] and SemEval 2017 Task 5 [37] organized more recently, has promoted the research in FSA techniques. However, these mainstream datasets do not contain emojis. In fact, there are often many emojis in social media comments.
Due to the lack of labeled training datasets and domain adaptation, the sentiment analysis of financial reviews with emojis became more challenging. They considered sentiment analysis technology easy to fail for six types of reasons, such as subjunctive mood, rhetoric, dependence on opinions, etc. [38]. When studying the financial sentiment analysis dataset StockSen deeply, many emojis in the byte code were not converted correctly. In the original preprocessing step of the dataset, these codes are simply deleted from the corpus to fit into the machine learning model [39]. Emojis are not considered and used in general fields in financial sentiment analysis. To better realize financial sentiment analysis, this paper attempts to solve the problem of sentiment recognition of emojis in the financial field. Therefore, this paper first constructs the first implicit expression dataset of emojis by homophonic replacement in the aspect of the dataset. Secondly, this paper effectively enhances the task of financial sentiment analysis by restoring the true semantics of emojis.

3. Implicit Expression of Emojis

Social comments differ from formally regulated financial documents, as they often include semantically ambiguous emojis influenced by factors like cultural context [6,7,8,25,27]. Figure 2 illustrates a user comment about GME (GameStop Corp., an entertainment software company, Grapevine, TX, USA) on StockTwits. In early 2021, shares in the American video game retailer GME surged more than 700% in one week following the speculative involvement of individual investors, a move touted as investors piling in to buy stock they like [40]. This section explores how emojis are implicitly expressed in social media comments by analyzing recent posts about GME on StockTwits.
In Figure 2, the author demonstrates how implicit expressions are conveyed through emojis, capturing a range of semantics across different cultural contexts and application scenarios. The semantics of the overt annotations of the emojis in Figure 2 are compared with those expressed in the text, as shown in Table 2.
The implicit expression of emojis effectively conveys the author’s emotional tendency, but recognizing these expressions presents significant challenges [6,7,8,9,25,27]. Implicit expressions can take various forms, such as homophones (Informatics 12 00056 i010). Different expressions can arise depending on the language, cultural background, or domain, making the automated annotation of implicit expressions in emojis particularly difficult [8]. Given the significant variations in cultural contexts across different domains, this chapter focuses on the implicit expressions communicated through emoji harmonics. In Table 2, emojis utilize harmonics (e.g., Informatics 12 00056 i008Informatics 12 00056 i009, Informatics 12 00056 i010) to achieve implicit expressions.
Referring to the previous research on emojis, this paper defines the implicit expression of emojis as follows [7,8,9,27,41,42]. Emojis are defined as the implicit expression when emojis express a different meaning from the official expression in the text through homophonic or combination. Table 3 offers illustrative examples of these implicit expressions enabled by harmonics. For more scientific purposes, this paper gives the following mathematical definition of the implicit expression of emojis. M M e a n i n g = { m 1 , m 2 , m 3 , , m n } represents all possible expressions of an emoji e , where m denotes a semantic expression of the emoji. These public annotations form the set of shallow expressions, denoted as M S h a l l o w   = m 1 , m 2 , m 3 , , m i , where M S h a l l o w   M M e a n i n g . However, emojis may also convey implicit expression diverging from the public annotation. These implicit expressions constitute the set of implicit expression M D e e p = m 1 , m 2 , m 3 , , m j , where M D e e p M M e a n i n g and MDeepMShallow = ∅. The traditional model lacks the capacity to discern these deep semantics. e can convey M D e e p through a specific expression relation f R e l a t i o n s h i p , as shown in Equation (1).
M D e e p = f R e l a t i o n s h i p m , m M S h a l l o w
This paper summarizes the process of implicit expression of emojis. Based on the process shown in Figure 3, humans can interpret the implicit expressions of emojis through harmonics. Financial professionals do have a different view on emojis compared to common people. However, traditional financial text analysis models typically focus only on the explicit annotations of emojis and struggle to grasp these implicit expressions [1,7]. This limitation adversely impacts the model’s performance when dealing with implicit emoji expressions.

4. Methods

This paper proposes IER-SMCEM, a model for recognizing implicit emoji expressions in social media comments based on prompt learning. IER-SMCEM comprises two main components: IEEA and IEPM. Firstly, IEEA leverages the implicit expressions of harmonics to map textual content into emojis. The input for IEPM consists of the IEEA-enhanced dataset X = x 1 , x 2 , x 3 , , x n . The text data x n from dataset X is populated into the predefined template T . The prediction process in the model for each online financial text can be represented by Equation (2). M denotes the pre-trained language model, while θ represents the parameter set of the model. The model’s output is the probability P .
P = M ( T x n , θ )
For example, the model puts the input “I do not Informatics 12 00056 i013Informatics 12 00056 i014 in the FINRA.” in the template “Related: [symbol]. Complete this sentence: [input] [mask].” Then, the input of the model is “ Related: bee leaf. Complete this sentence: I do not [mask] in the FINRA.” Finally, the prediction result of the model is “believe.” This paper introduces Bert’s workflow to explain how the model is predicted.
The pre-trained language model (PLM) receives input in the form of a sequence of tokens W = w 1 , w 2 , w 3 , , w n , where n represents the length of the text. This step converts the text into a vector for input into the model. Usually, it consists of three kinds of word embeddings: token embedding, segment embedding, and position embedding. Firstly, the token embedding layer of the model converts the tokens into token embedding, denoted as E W , as shown in Equation (3). f t o k e n indicates the process of conversion.
E W = f t o k e n W
Secondly, the segment embedding layer encodes the sentence to distinguish the segment embedding of the two sentences, as shown in Equation (4), where E A represents the segment embedding. The segment embedding process is denoted as f S e g m e n t .
E A = f S e g m e n t E w
Finally, the position embedding layer encodes the positional information of words into position embedding, as shown in Equation (5), where E n represents the position embedding. The position embedding process is denoted as f P o s i t i o n .
E n = f P o s i t i o n E A
Input the above vectors into the main model frame in the form of superposition and the coding and decoding structure based on Bi-transformer [33] can obtain all the predicted values of the possible words in the [mask] position. The vectors are entered into the Bi-transformer for calculation, as shown in Equation (6), where f B i T represents the Bi-transformer. E B i represents the output of the Bi-transformer.
E B i = f B i T E W , E A , E n
Finally, the model transforms the predicted value into a probability value of 0 to 1 through normalization operation. By choosing the prediction with a high probability value, the model can predict the true meaning of the emoji. The output is then normalized by the Softmax layer, as shown in Equation (7), where f s o f t m a x represents the Softmax layer. W f d represents the output of the Softmax layer.
W f d = f s o f t m a x E B i
Finally, the output marker layer converts the output to the corresponding encoding. Figure 4 shows the general structure of the model of the IER-SMCEM.

4.1. Implicit Emoji Enhancement Algorithm (IEEA)

The proposed IEEA is a method for embedding the implicit expressions of emojis into text. For the input text, IEEA first matches emojis based on harmonics using an emoji library, replacing well-matched words with corresponding emojis and masking the original word’s position. Then, the public annotation of the emoji, the masked text, and the actual expression are combined into the input format for the IEPM. It leverages harmonics to integrate emojis into an existing corpus. Figure 5 presents the principle of IEEA.
This paper employs homophones as the replacement logic for IEEA. The function f c l i p determines the position for replacing the emoji. By comparing the Levenshtein Distance between words and emojis one by one, f c l i p replaces the words with suitable emojis. Only a ratio of words in the sentence will be replaced by f c l i p . As shown in Figure 5, “I believe in the FINRA.” is clipped into three parts by f c l i p . w t e x t is the remaining text, w m a s k is the pronunciation of the replaced word, and w l a b e l is the replaced word. IEEA retains the phonological features of the original text as w m a s k . The expression of the original text is retained as the data label, w l a b e l , and the remaining part of the original text is w t e x t . The process can be represented by Equation (8).
w t e x t , w m a s k , w l a b e l = f c l i p ( x i )
The function f t o n e maps w m a s k within the data x i = [ w t e x t , w m a s k , w l a b e l ] to the corresponding emoji from set E . Subsequently, w m a s k is replaced with the expression of the mapped emoji w s y m b o l . The position of w s y m b o l is marked as [MASK]. This replacement process is delineated by Equation (9). As shown in Figure 5, IEEA selects the replacement emoji by Levenshtein Distance in the emoji set. Replace w m a s k with the emoji’s annotation. Thus, enhanced x i is obtained.
w e m o j i = f t o n e w m a s k , E
Through Equation (9), IEEA achieved the construction of the dataset. Each data in the dataset consists of three parts: the marked text ( w t e x t ), the emojis in the text ( w s y m b o l ), and the true meaning of the emoji ( w l a b e l ). IEEA reconstructed the content of x i . Thus, the output of x i through the IEEA module is x i = [ w t e x t , w s y m b o l , w l a b e l ] .

4.2. Implicit Emoji Prediction Module (IEPM)

For the input text generated by IEEA with emojis, IEPM first embeds the text along with its corresponding public annotations into a predefined template. Next, IEPM leverages the mask prediction function of the pre-trained model to infer the expression of the masked token. The model computes the probability distribution at the mask position and selects the prediction with the highest probability of replacing the emoji in the original text. This process effectively identifies and interprets the implicit expression of the emojis, thereby improving the text’s accurate representation. Specifically, IEPM is utilized to extract sample features from the IEEA-enhanced dataset X . IEPM fine-tunes the model’s learning focus through the design of the input text. Figure 6 presents the model’s structure.
This paper takes one piece of data x i in X as an example. The content in x i is spliced with the prompt template T = R e l a t e d   t o   [ s y m b o l ] .   C o m p l e t e   t h i s   s e n t e n c e : [ i n p u t ] . The [ i n p u t ] position in template T is embedded by w t e x t . The [ s y m b o l ] position is embedded by w s y m b o l . Thus, the final output, x T , is obtained, as shown in Equation (10).
x T = f ( x i , T )
For input x T , the word embedding model will predict the mask position in x T . Therefore, this paper takes the output of the word embedding model as the prediction vector. The word embedding layer converts the x T into the corresponding word vector E T R N , N = 768 . N is the word vector dimension in the Bert model [33], as shown in Equation (11).
E T = E m b e d d i n g ( x T )
This work inputs E T into the pre-trained model M with the model parameter θ . Subsequently, the output of the model is transformed into the probability vector of the corresponding position using the function f s o f t m a x . The final answer list L A n s w e r of the model is obtained, as shown in Equation (12).
L A n s w e r = f s o f t m a x ( M E T , θ )
This work employs a stricter training requirement when training model M to pay attention to the expression of deep semantics. Therefore, in this paper, the model parameters are adjusted using the L1 loss. Equation (13) shows the loss function.
L o s s ( X , Y ) = i = 1 n y i m a x ( L A n s w e r ) n

5. Results

5.1. Dataset

In the realm of emoji semantic recognition, there is no dataset directly suited to the task addressed in this paper. To address it, we construct two datasets: IEEA-C for Chinese and IEEA-E for English. These datasets are created by enhancing the implicit expression of emojis using publicly available datasets, including SST-2 [43], IMDB [44], Weibo-100k [45], and ChnSentiCorp [46].
This paper adopts the Levenshtein Distance (LD) as the criterion for replacement. Specifically, emoji with an LD greater than 0.7 are targeted for replacement. Section 5.2 elaborates on the text’s ambiguity level under this condition. Table 4 presents the distribution of the IEEA-E dataset post-replacement.
This paper combines Chinese financial commentaries sourced from online platforms with publicly available datasets, Weibo_100k and ChnSentiCorp, to formulate the Chinese dataset IEEA-C. Table 5 presents the example data from IEEA-C. Table 6 illustrates the dataset’s distribution.

5.2. Ambiguity Analysis

To assess the level of ambiguity [47] in sentences before and after data enhancement, this paper adopts the evaluation methodology outlined in Basile et al.’s [48] study. The evaluation metrics employed are detailed in Table 7, where a and b represent the sentences before and after enhancement, respectively.
Because the range of values for the four indicators is not uniform, this paper uses Equation (14) to map the range of variation from 0 to 1. The genuine value is V . The Normalization is M a p .
M a p = V m i n ( V ) m a x ( V ) m i n ( V )
This paper investigates the relationship between the replacement ratio and the similarity. The replacement ratio ranges from 10% to 70%. Figure 7 and Figure 8 show the experimental results.
As shown in Figure 7 and Figure 8, all similarity metrics (InDel, Jaro–Winkler, Jaccard) exhibit a significant increase when substitutions are made using higher values of Levenshtein Distance (LD). Conversely, the distance metric (Levenshtein) decreases rapidly. The experiments demonstrated that higher LD values correlate with improved sentence similarity. With an increasing replacement ratio, metrics associated with sentence similarity decrease. Table 8 presents the text before and after the replacement. Guided by these experimental findings, this paper chooses an LD threshold of 0.7 and a replacement ratio of 0.4 to enhance the data. Therefore, in the example shown in Table 8, this paper replaces 40% of the words in the sentence with emojis that have an LD higher than 0.7 by comparing the LD between each word and the corresponding emoji.

5.3. Pre-Trained Models

In IEPM, IER-SMCEM utilizes the mask prediction function of pre-trained models. To evaluate how different pre-trained models affect the performance of IER-SMCEM, this paper employs BERT-base, BERT-large, and RoBERTa as the pre-trained language models (PLMs). The relevant parameters for these models are detailed in Table 9.
During the fine-tuning phase of the pre-trained model, we utilize its mask prediction function (MLM). The data is first tokenized to break the text into units that the model can process. In the process of tokenization, data is represented as vectors by the embedded layer. The text with [mask] tokens is then fed into the pre-trained model, where the mask prediction function predicts the masked positions and generates the most contextually appropriate emoticon or word. The hyperparameters for this phase are optimized based on the validation set, with the specific parameters detailed in Table 10.

5.4. Prompt Template

IER-SMCEM utilizes the T E M J P L M template, which incorporates the relationship between public annotations and the implicit expressions of emojis. We devise template T P L M , which is devoid of semantic relation information of emojis. Additionally, we introduce an unprompted template, T L M , to evaluate the role of prompts. Table 11 and Table 12 present the templates employed for the experiments.

5.5. Evaluation

Several similarities exist between our work and that of XMTC: (1) both necessitate retrieving the accurate label from a broad spectrum of potential labels and (2) both demonstrate a long-tail impact on the labels. The task addressed in this paper involves recalling a single label, whereas the XMTC task involves recalling multiple labels. The goal of this study is to recover the implicit expression of emojis through a special case of extreme multi-label text classification (XMTC) [49]. We use p r e c i s i o n @ k to evaluate the performance at the top of the result, a metric that has been widely used to evaluate XMTC methods [49]. Equation (15) shows the calculation for P @ k .
P @ k = 1 k l = 1 k y r a n k ( l )
where y { 0,1 } L , L R is the true binary vector, and r a n k ( l ) is the index of the l -th highest predicted label. We propose Equations (16) and (17) as an evaluation indicator based on Equation (18).
P = l = 1 r y r a n k ( l )
A c c u r a c y = n = 1 N D P T n = 1 N D P T + n = 1 N D P F , w h e r e P T = 1 , P F = 0   P = 1 P T = 0 , P F = 1   P = 0
where r , r R , represents the size of the answer space when the answer space threshold is set to ω , ω R . N D , N D R represents the size of the dataset. To evaluate the size of r , this paper introduces a metric for assessing answer correlation, as shown in Equation (18).
σ A R = n = 1 N D P T r
In the task of XMTC, the precise retrieval of target labels presents significant challenges due to the enormous scale of the label space. Similar to the P @ k evaluation metric, the P metric adopted in this study is also used to assess the accuracy of the model’s predictions within a given range. However, unlike P @ k , which limits the evaluation scope to the top k predictions, the P metric extends the evaluation range to r . This adjustment holds substantial theoretical and practical significance. Specifically, in traditional XMTC tasks, each instance typically corresponds to multiple relevant labels, whereas the single-label prediction task focused on in this study exhibits notable differences. In XMTC, several labels need to be predicted from hundreds of labels. In the task, this paper needs to predict a word from tens of thousands of words. Moreover, some words have synonymous or similar expressions, which makes the task difficulty inconsistent. Given the inherent polysemy and context-dependency of emojis, employing the P @ k metric would introduce overly stringent evaluation criteria. This not only misaligns with the actual semantic expression characteristics of emojis but may also lead to a misjudgment of model performance. Therefore, the P metric proposed in this study better accommodates the characteristics of emoji prediction tasks, providing a more reasonable standard for evaluating model performance.
This paper uses the commonly binary sentiment analysis metrics in the downstream sentiment analysis tasks. Metrics are shown by Equation (19). True Positive (TP) refers to the number of samples correctly predicted as positive, while True Negative (TN) represents the number of samples correctly predicted as negative. False Positive (FP) indicates the number of negative samples incorrectly predicted as positive, while False Negative (FN) denotes the number of positive samples incorrectly predicted as negative.
A c c = T P + T N T P + T N + F P + F N

5.6. Selection of Answer Space

To determine the most appropriate answer space size r across different datasets, this paper investigates the accuracy and answer relevance of various models on their respective language datasets. Table 13 shows the answer space thresholds represented by the answer space steps across different models. Due to the inconsistent size of the answer space for each model, this paper uniformly divides the answer space into five intervals to fairly evaluate model performance. Each interval represents an answer step. Since different models require varying thresholds for the answer space, as an optional parameter, this paper divides the answer space into five segments (answer space steps) for comparison purposes.
Figure 9 depicts the accuracy and answer correlation of the three models at the identical answer space step.
Based on the definition of answer correlation, a higher correlation indicates better model performance. IER-SMCEM-base-E demonstrates similar accuracy to the other two models while surpassing them in terms of answer correlation. Regarding accuracy, IER-SMCEM-large-C shows comparable performance to the other two models. When it comes to answer relevance, IER-SMCEM-large-C outperforms the other two models.
According to the definition of answer space, we select an answer space threshold with an answer space step of 3 to participate in subsequent model performance comparisons and application experiments.

5.7. The Semantic Recovery of Traditional Models

We utilize the T L M template to assess traditional models’ capacity in discerning the deep semantics of non-textual symbols. Figure 10 shows the experimental results on the IEEA-E.
The accuracy of Bert-large-E surpasses that of Bert-base-E, suggesting that having more deep layers is advantageous for the task of recognizing the implicit expression of emojis as it enables the model to more accurately learn sample features and achieve improved results. This conclusion also holds true for IEEA-C.

5.8. Experimental Results of Prompt Templates

We conduct two experiments using two different prompt templates: T P L M and T E M J P L M . Figure 11 shows the experimental results.
When using the T P L M template, IER-SMCEM-base-C demonstrates similar accuracy to that of IER-SMCEM-large-C. However, in the IEEA-E dataset, IER-SMCEM-base-E does not follow this trend because the English content in this dataset is less challenging. Compared with English, Chinese is richer in form. There is a similar S (Subject)-V (Verb)-O (Object) structure in Chinese. Generally, models with more complex structures perform better. Therefore, models with more intricate architectures tend to excel when handling simpler content.
After using the T E M J P L M template, each model shows some improvement. This proves that embedding emoji information into the model’s input through templates can effectively improve the mode’s performance.

6. Case Study

The IER-SMCEM proposed in this paper is a semantic-driven model. Although the performance index of automation reflects the performance of the model to a certain extent, the performance of the model can be understood more deeply through intuitive manual analysis. Therefore, in this section, this paper selects two samples to evaluate the effectiveness of the model by comparing the semantic recovery effect. The case is shown in Table 14.
In Table 14, this paper selected two cases and analyzed their sentiment. In the experimental setup, DistilRoBERTa [50] was used to respectively analyze the sentiment of the cases directly or after semantic recovery. By comparing the results, it can be seen that IER-SMCEM can effectively recover implicit expression. In the results of sentiment analysis, because emojis hide some positive sentiment words, direct sentiment analysis leads to errors. After semantic recovery by IER-SMCEM, the semantic expression of emoji is correct, and the correct sentiment analysis results are achieved. The experimental results show that it is very important to restore the implicit expression of emojis for downstream text analysis tasks.

7. Application

Previous research in finance has highlighted numerous links between investor sentiment and market [51,52]. To accurately identify investor sentiment, it is essential to extract the true content. The proposed IER-SMCEM focuses on recovering the true expressions of emojis. Therefore, this study employed common sentiment analysis tasks to validate the applicability of the recovered text for subsequent research tasks.
The performance of financial markets is significantly influenced by sentiment. Therefore, effectively extracting and quantifying sentiment from users and investors is a pressing issue in the financial field. Vamossy’s [53] research team has developed a tool that aids researchers in exploring the impact of sentiment on financial markets. By extracting sentiment from social media texts specifically related to finance, their research indicates that sentiment expressed by investors regarding specific firms can forecast daily price movements, highlighting the strong correlation between sentiment and market dynamics.
When facing implicit expressions, as shown in Figure 12 the performance of these models is inevitably affected to some extent.
The IER-SMCEM introduced in this paper helps sentiment analysis models manage the implicit expressions of emojis. To illustrate that IER-SMCEM is effective, this paper conducted experiments using standard datasets and widely used sentiment analysis models. After balancing the number of labels, we verified the robustness of the results with a k-fold CV (k = 5). Table 15 shows the distribution of labels in the two datasets.
This paper selected five sentiment analysis models representing deep learning (i.e., TextCNN, BiLSTM, Bert, Bert-large, and DistilRoBERTa) [50] and two traditional machine learning models (i.e., CART [54], and XGB [55]). Table 16 and Table 17 present experimental results.
Based on these experimental results, the IER-SMCEM proposed in this paper effectively enhanced the performance of downstream sentiment mining models. When mining social media information, models using the data recovered from this paper demonstrated excellent results. Experiments demonstrate that IER-SMCEM significantly enhances the model’s ability to analyze sentiment in the presence of implicit emoji expressions. Financial fraud can be effectively detected by analyzing corporate events and investor sentiment on social media. In future work, the accuracy of fraud detection can be enhanced by leveraging investor sentiment, with corporate events serving as the primary driver. In future works, we plan to explore applying financial sentiment analysis to detect potential financial fraud.

8. Conclusions and Future Works

This paper introduces an implicit expression recognition model for implicit expressions, addressing the challenge of accurately extracting valuable information. The model integrates relational information between public annotations and the implicit expressions of emojis, enhancing the model’s ability to recognize the implicit expressions.
Comparing the IER-SMCEM-base and IER-SMCEM-large models reveals that the more complex model structure yields better performance without prompts. For example, in Figure 10, the IER-SMCEM-large-E model achieves an accuracy rate of 79.29%, indicating that the increase in the number of hidden layers can improve the model’s learning ability. With the inclusion of prompt templates, the performance of a simple model can approach that of a complex model. For example, in Figure 11, the accuracy of IER-SMCEM-base-C with prompt templates reaches 97.10%. In Figure 11, IER-SMCEM-base-C achieves the highest accuracy of 98.03% in semantically recovering the implicit expressions within financial texts. Semantic reduction using this framework also significantly improves subsequent text-mining tasks such as sentiment analysis.
The semantics of the emoji recovery technology proposed in this paper, combined with downstream market analysis technology such as sentiment analysis, can be used as an early warning signal/indicator in the stock market (assuming the comments are reliable). Moreover, in other application fields that need social comment analysis, the method in this paper can be used to enhance the ability of other methods to recognize emojis. Although this study has achieved some progress in the semantic recovery of emojis, there are still some limitations. Firstly, the experiment relies on quantitative indicators (such as accuracy) and does not include human subjective evaluation (such as usability and manual evaluation). Therefore, different people will have different understandings of the true meaning of emojis. Secondly, data collection and experimental design are limited to specific languages (such as Chinese and English). Therefore, the universality of the conclusion in other cultures needs further verification. Finally, the language representation of the dataset is insufficient. This may introduce potential linguistic and socio-cultural deviations. To summarize, in future works, we will combine multilingual datasets and introduce human experts to enhance the reliability and generalization of the results.

Author Contributions

J.Z. wrote the main manuscript text and prepared Figure 1, Figure 2 and Figure 3. C.W. contributed to the methodology and data analysis. Z.L. assisted with the literature review and editing. H.D. provided the resources and supported the data collection. Q.L. contributed to the data interpretation and figure preparation. B.Z. supervised the project and provided critical revisions. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Chinese Government Guidance Fund on Local Science and Technology Development of Sichuan Province (2024ZYD0272).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used and analyzed in the current study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Du, K.; Xing, F.; Mao, R.; Cambria, E. Financial Sentiment Analysis: Techniques and Applications. ACM Comput. Surv. 2024, 56, 1–42. Available online: https://stocktwits.com/ (accessed on 24 April 2024). [CrossRef]
  2. Zhou, Z.; Ma, L.; Liu, H. Trade the event: Corporate events detection for news-based event-driven trading. arXiv 2021, arXiv:2105.12825. [Google Scholar]
  3. Lin, W.; Liao, L.C. Lexicon-based prompt for financial dimensional sentiment analysis. Expert Syst. Appl. 2024, 244, 122936. [Google Scholar] [CrossRef]
  4. Padmanayana, V.; Bhavya, K. Stock market prediction using Twitter sentiment analysis. Int. J. Sci. Res. Comput. Sci. Eng. Inf. Technol. 2021, 7, 265–270. [Google Scholar] [CrossRef]
  5. Singla, A.; Gupta, M. Exploring the potential of dogecoin promoted by elon musk. J. Metaverse Blockchain Technol. 2024, 2, 35–43. [Google Scholar] [CrossRef]
  6. Carrillo, H.E.; Pennington, R.; Zhang, Y. Is an Emoji Worth a Thousand Words? The Effect of Emoji Usage on Nonprofessional Investors’ Perceptions. J. Inf. Syst. 2022, 36, 1–15. [Google Scholar] [CrossRef]
  7. Chen, S.; Xing, F. Understanding Emojis for Financial Sentiment Analysis. In Proceedings of the ICIS 2023, Hyderabad, India, 10–13 December 2023; Available online: https://aisel.aisnet.org/icis2023/socmedia_digcollab/socmedia_digcollab/3 (accessed on 24 April 2024).
  8. Lu, X.; Cao, Y.; Chen, Z.; Liu, X. A first look at emoji usage on github: An empirical study. arXiv 2018, arXiv:1812.04863. [Google Scholar]
  9. Zhou, Y.; Xu, P.; Wang, X.; Lu, X.; Gao, G.; Ai, W. Emojis decoded: Leveraging chatgpt for enhanced understanding in social media communications. arXiv 2024, arXiv:2402.01681. [Google Scholar] [CrossRef]
  10. Zhao, Z.; Rao, G.; Feng, Z. DFDS: A domain-independent framework for document-level sentiment analysis based on RST. In Proceedings of the Web and Big Data: First International Joint Conference, APWeb-WAIM 2017, Beijing, China, 7–9 July 2017; Proceedings, Part I 1. Springer International Publishing: Berlin/Heidelberg, Germany, 2017; pp. 297–310. [Google Scholar]
  11. Zhang, Y.; Zhang, Y. Tree communication models for sentiment analysis. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 3518–3527. [Google Scholar]
  12. Liu, J.; Zhang, Y. Attention modeling for targeted sentiment. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 2, Short Papers, Valencia, Spain, 3–7 April 2017; pp. 572–577. [Google Scholar]
  13. Alquézar, R.; Sanfeliu, A. Recognition and learning of a class of context-sensitive languages described by augmented regular expressions. Pattern Recognit. 1997, 30, 163–182. [Google Scholar] [CrossRef]
  14. Park, S.B.; Zhang, B.T. Text chunking by combining hand-crafted rules and memory-based learning. In Proceedings of the 41st Annual Meeting of the Association for Computational Linguistics, Sapporo, Japan, 7–12 July 2003; pp. 497–504. [Google Scholar]
  15. Liu, P.; Yuan, W.; Fu, J.; Jiang, Z.; Hayashi, H.; Neubig, G. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Comput. Surv. 2023, 55, 1–35. [Google Scholar] [CrossRef]
  16. Liu, X.; Zheng, Y.; Du, Z.; Ding, M.; Qian, Y.; Yang, Z.; Tang, J. GPT understands, too. arXiv 2021, arXiv:2103.10385. [Google Scholar] [CrossRef]
  17. Ding, N.; Hu, S.; Zhao, W.; Chen, Y.; Liu, Z.; Zheng, H.T.; Sun, M. Openprompt: An open-source framework for prompt-learning. arXiv 2021, arXiv:2111.01998. [Google Scholar]
  18. Schulhoff, S.; Ilie, M.; Balepur, N.; Kahadze, K.; Liu, A.; Si, C.; Li, Y.; Gupta, A.; Han, H.; Schulhoff, S.; et al. The prompt report: A systematic survey of prompting techniques. arXiv 2024, arXiv:2406.06608. [Google Scholar]
  19. Schick, T.; Schütze, H. Exploiting cloze questions for few shot text classification and natural language inference. arXiv 2020, arXiv:2001.07676. [Google Scholar]
  20. Hu, S.; Ding, N.; Wang, H.; Liu, Z.; Li, J.; Sun, M. Knowledgeable prompt-tuning: Incorporating knowledge into prompt verbalizer for text classification. arXiv 2021, arXiv:2108.02035. [Google Scholar]
  21. Aliprand, J.M. The unicode standard. Libr. Resour. Tech. Serv. 2000, 44, 160–167. [Google Scholar] [CrossRef]
  22. Hu, T.; Guo, H.; Sun, H.; Nguyen, T.V.; Luo, J. Spice up your chat: The intentions and sentiment effects of using emojis. In Proceedings of the International AAAI Conference on Web and Social Media, Montreal, QC, Canada, 15–18 May 2017; Volume 11, pp. 102–111. [Google Scholar]
  23. Cramer, H.; De Juan, P.; Tetreault, J. Sender-intended functions of emojis in US messaging. In Proceedings of the 18th International Conference on Human-Computer Interaction with Mobile Devices and Services, Florence, Italy, 6–9 September 2016; pp. 504–509. [Google Scholar]
  24. Qiu, L.; Wang, W.; Pang, J. The persuasive power of emoticons in electronic word-of-mouth communication on social networking services. MIS Q. 2023, 47, 511. [Google Scholar] [CrossRef]
  25. Ai, W.; Lu, X.; Liu, X.; Wang, N.; Huang, G.; Mei, Q. Untangling emoji popularity through semantic embeddings. In Proceedings of the International AAAI Conference on Web and Social Media, Montreal, QC, Canada, 15–18 May 2017; Volume 11, pp. 2–11. [Google Scholar]
  26. Lu, X.; Ai, W.; Liu, X.; Li, Q.; Wang, N.; Huang, G.; Mei, Q. Learning from the ubiquitous language: An empirical analysis of emoji usage of smartphone users. In Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Heidelberg, Germany, 12–16 September 2016; pp. 770–780. [Google Scholar]
  27. Chen, Z.; Shen, S.; Hu, Z.; Lu, X.; Mei, Q.; Liu, X. Emoji-powered representation learning for cross-lingual sentiment classification. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 251–262. [Google Scholar]
  28. Felbo, B.; Mislove, A.; Søgaard, A.; Rahwan, I.; Lehmann, S. Using millions of emoji occurrences to learn any-domain representations for detecting sentiment, emotion and sarcasm. arXiv 2017, arXiv:1708.00524. [Google Scholar]
  29. Barbieri, F.; Kruszewski, G.; Ronzano, F.; Saggion, H. How cosmopolitan are emojis? Exploring emojis usage and meaning over different languages with distributional semantics. In Proceedings of the 24th ACM International Conference on Multimedia, Amsterdam, The Netherlands, 15–19 October 2016; pp. 531–535. [Google Scholar]
  30. Mikolov, T. Efficient estimation of word representations in vector space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
  31. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2013; Volume 26. [Google Scholar]
  32. Tino, P.; Schittenkopf, C.; Dorffner, G. Financial volatility trading using recurrent neural networks. IEEE Trans. Neural Netw. 2001, 12, 865–874. [Google Scholar] [CrossRef]
  33. Kenton, J.D.M.W.C.; Toutanova, L.K. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the NAACL-HLT, Minneapolis, Minnesota, USA, 2–7 June 2019; Volume 1, p. 2. [Google Scholar]
  34. Chan, S.W.K.; Chong, M.W.C. Sentiment analysis in financial texts. Decis. Support Syst. 2017, 94, 53–64. [Google Scholar] [CrossRef]
  35. Deveikyte, J.; Geman, H.; Piccari, C.; Provetti, A. A sentiment analysis approach to the prediction of market volatility. Front. Artif. Intell. 2022, 5, 836809. [Google Scholar] [CrossRef] [PubMed]
  36. de França Costa, D.; da Silva, N.F.F. INF-UFG at FiQA 2018 Task 1: Predicting sentiments and aspects on financial tweets and news headlines. In Proceedings of Companion Proceedings of the The Web Conference 2018; International World Wide Web Conferences Steering Committee: Republic and Canton of Geneva, Switzerland, 2018; pp. 1967–1971. [Google Scholar] [CrossRef]
  37. Cortis, K.; Freitas, A.; Daudert, T.; Huerlimann, M.; Zarrouk, M.; Handschuh, S.; Davis, B. Semeval-2017 task 5: Fine-grained sentiment analysis on financial microblogs and news. In Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017), Vancouver, BC, Canada, 3–4 August 2017; pp. 519–535. [Google Scholar]
  38. Xing, F.; Malandri, L.; Zhang, Y.; Cambria, E. Financial sentiment analysis: An investigation into common mistakes and silver bullets. In Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain, 8–13 December 2020; pp. 978–987. [Google Scholar]
  39. Deng, S.; Huang, Z.; Sinha, A.P.; Zhao, H. The interaction between microblog sentiment snd stock returns. MIS Q. 2018, 42, 895-A13. [Google Scholar] [CrossRef]
  40. Long, S.; Lucey, B.; Xie, Y.; Yarovaya, L. “I just like the stock”: The role of Reddit sentiment in the GameStop share rally. Financ. Rev. 2023, 58, 19–37. [Google Scholar] [CrossRef]
  41. Sampietro, A. Emojis and the performance of humour in everyday electronically-mediated conversation: A corpus study of WhatsApp chats. Internet Pragmat. 2021, 4, 87–110. [Google Scholar] [CrossRef]
  42. Cui, J.; Dandan, Y.R.; Jiang, G. Judging emoji by occupation: A case of emoji-based sarcasm interpretation. Acta Psychol. 2023, 234, 103870. [Google Scholar] [CrossRef]
  43. Munikar, M.; Shakya, S.; Shrestha, A. Fine-grained sentiment classification using BERT. In Proceedings of the 2019 Artificial Intelligence for Transforming Business and Society (AITB), Kathmandu, Nepal, 5 November 2019; Volume 1, pp. 1–5. [Google Scholar]
  44. Oghina, A.; Breuss, M.; Tsagkias, M.; De Rijke, M. Predicting imdb movie ratings using social media. In Proceedings of the Advances in Information Retrieval: 34th European Conference on IR Research, ECIR 2012, Barcelona, Spain, 1–5 April 2012; Proceedings 34. Springer: Berlin/Heidelberg, Germany, 2012; pp. 503–507. [Google Scholar]
  45. Li, Y.; Yu, B.; Xue, M.; Liu, T. Enhancing pre-trained Chinese character representation with word-aligned attention. arXiv 2019, arXiv:1911.02821. [Google Scholar]
  46. Liu, B.; Xu, W.; Xiang, Y.; Wu, X.; He, L.; Zhang, B.; Zhu, L. Noise learning for text classification: A benchmark. In Proceedings of the 29th International Conference on Computational Linguistics, Gyeongju, Republic of Korea, 12–17 October 2022; pp. 4557–4567. [Google Scholar]
  47. Page, B.I. The theory of political ambiguity. Am. Political Sci. Rev. 1976, 70, 742–752. [Google Scholar] [CrossRef]
  48. Basile, A.; Crupi, R.; Grasso, M.; Mercanti, A.; Regoli, D.; Scarsi, S.; Yang, S.; Cosentini, A.C. Disambiguation of company names via deep recurrent networks. Expert Syst. Appl. 2024, 238, 122035. [Google Scholar] [CrossRef]
  49. Huang, Y.; Giledereli, B.; Köksal, A.; Özgür, A.; Ozkirimli, E. Balancing methods for multi-label text classification with long-tailed class distribution. arXiv 2021, arXiv:2109.04712. [Google Scholar]
  50. Hartmann, J. Emotion English Distilroberta-Base. 2022. Available online: https://huggingface.co/j-hartmann/emotion-english-distilroberta-base/ (accessed on 24 April 2024).
  51. Ahmadi, J. Investigating the relationship between investor sentiments and fraudulent financial reporting with emphasis on political connections. J. Account. Manag. Vis. 2023, 6, 285–302. [Google Scholar]
  52. Ramsheh, M.; Arefmanesh, Z.; Rostami, R.; Khastar, Z. Investor Sentiment and the Likelihood of Fraudulent Financial Reporting in Petroleum and Petrochemical Industries: The Moderating Role of Risk Disclosure. Pet. Bus. Rev. 2023, 7, 1–14. [Google Scholar]
  53. Vamossy, D.F.; Skog, R.P. EmTract: Extracting emotions from social media. Int. Rev. Financ. Anal. 2025, 97, 103769. [Google Scholar] [CrossRef]
  54. Lewis, R.J. An introduction to classification and regression tree (CART) analysis. In Proceedings of the Annual Meeting of the Society for Academic Emergency Medicine, San Francisco, CA, USA, 22–25 May 2000; p. 14. [Google Scholar]
  55. Pathy, A.; Meher, S. Predicting algal biochar yield using eXtreme Gradient Boosting (XGB) algorithm of machine learning methods. Algal Res. 2020, 50, 102006. [Google Scholar] [CrossRef]
Figure 1. Relationship between stock price and investor sentiment (https://stocktwits.com/ (accessed on 24 April 2024)).
Figure 1. Relationship between stock price and investor sentiment (https://stocktwits.com/ (accessed on 24 April 2024)).
Informatics 12 00056 g001
Figure 2. An example taken from StockTwits (https://stocktwits.com/ (accessed on 24 April 2024)).
Figure 2. An example taken from StockTwits (https://stocktwits.com/ (accessed on 24 April 2024)).
Informatics 12 00056 g002
Figure 3. The process of identifying the implicit expression of emojis.
Figure 3. The process of identifying the implicit expression of emojis.
Informatics 12 00056 g003
Figure 4. IER-SMCEM framework structure.
Figure 4. IER-SMCEM framework structure.
Informatics 12 00056 g004
Figure 5. Detailed processes within IEEA.
Figure 5. Detailed processes within IEEA.
Informatics 12 00056 g005
Figure 6. Principle of operation of IEPM.
Figure 6. Principle of operation of IEPM.
Informatics 12 00056 g006
Figure 7. Evolution of evaluation indicators in IEEA-E.
Figure 7. Evolution of evaluation indicators in IEEA-E.
Informatics 12 00056 g007
Figure 8. Evolution of evaluation indicators in IEEA-C.
Figure 8. Evolution of evaluation indicators in IEEA-C.
Informatics 12 00056 g008
Figure 9. Accuracy and answer correlation for different models at the same answer space step.
Figure 9. Accuracy and answer correlation for different models at the same answer space step.
Informatics 12 00056 g009
Figure 10. Comparing the accuracy of different models using T L M -E and T L M -C.
Figure 10. Comparing the accuracy of different models using T L M -E and T L M -C.
Informatics 12 00056 g010
Figure 11. Accuracy of different models on multiple templates.
Figure 11. Accuracy of different models on multiple templates.
Informatics 12 00056 g011
Figure 12. Accuracy of different models on multiple templates.
Figure 12. Accuracy of different models on multiple templates.
Informatics 12 00056 g012
Table 1. Implicit expressions of emojis in financial commentary.
Table 1. Implicit expressions of emojis in financial commentary.
ContentEmojisSemantics
Just bought the dip Informatics 12 00056 i032Informatics 12 00056 i003. Time to HODL Informatics 12 00056 i008Informatics 12 00056 i009 and watch it go Informatics 12 00056 i001Informatics 12 00056 i002. Not financial advice, just pure Informatics 12 00056 i001Informatics 12 00056 i001Informatics 12 00056 i001 vibes![Informatics 12 00056 i032Informatics 12 00056 i003: Hole; Downward]
[Informatics 12 00056 i008Informatics 12 00056 i009: Diamond; Hands]
[Informatics 12 00056 i001Informatics 12 00056 i002: Rocket; Moon]
[Informatics 12 00056 i001Informatics 12 00056 i001Informatics 12 00056 i001: Rocket]
Endless Downwards;
Diamond Hands;
Soaring Stock Prices;
Soaring Stock Prices;
Market be like: Informatics 12 00056 i003Informatics 12 00056 i003Informatics 12 00056 i003, but my portfolio’s like: Informatics 12 00056 i004Informatics 12 00056 i005. Guess we’re all just clowning around until the next bull run Informatics 12 00056 i006Informatics 12 00056 i007[Informatics 12 00056 i003Informatics 12 00056 i003Informatics 12 00056 i003: Downwards]
[Informatics 12 00056 i004Informatics 12 00056 i005: Clown; Flying Money]
[Informatics 12 00056 i006Informatics 12 00056 i007: Bull; Flexed Biceps]
Downwards Trend;
Juggled by the Market; Expecting the Market to Rebound
Table 2. Public annotations and semantics of emojis.
Table 2. Public annotations and semantics of emojis.
EmojisPublic AnnotationsSemantics
Informatics 12 00056 i001Informatics 12 00056 i030[Rocket] [Chart Increasing]Rapid Increase in Share Price
Informatics 12 00056 i008Informatics 12 00056 i009[Gem Stone] [Raising Hands]Diamond Hands
Informatics 12 00056 i010[Banana]Contempt
Informatics 12 00056 i001Informatics 12 00056 i001Informatics 12 00056 i001Informatics 12 00056 i001[Rocket]Significant Increase
Informatics 12 00056 i011Informatics 12 00056 i012[Bar Chart] [Bullseye]Precise Stock Price Forecast
Table 3. The implicit expressions of emojis.
Table 3. The implicit expressions of emojis.
ContentContent with Emojis
I do not believe in the FINRA.I do not Informatics 12 00056 i013Informatics 12 00056 i014 in the FINRA.
It seemed like a bad idea.It seemed like a bad Informatics 12 00056 i015Informatics 12 00056 i016.
Table 4. Distribution of IEEA-E samples.
Table 4. Distribution of IEEA-E samples.
DatasetTrain SetValidation SetTest Set
IEEA-E14,39436764410
Table 5. Example of IEEA-C sample table.
Table 5. Example of IEEA-C sample table.
ContentExample1Example2
TextInformatics 12 00056 i006动资金 1财务欺Informatics 12 00056 i017 2
SemanticsWorking capitalFinancial fraud
1 Working fund. 2 Financial fraud.
Table 6. Distribution of IEEA-C samples.
Table 6. Distribution of IEEA-C samples.
DatasetTrain SetValidation SetTest Set
IEEA-C16,01051135191
Table 7. Indicators of the degree of sentence ambiguity.
Table 7. Indicators of the degree of sentence ambiguity.
ScoreTypeRange
Levenshtein Distance (↓) 1distance D L e v ( a , b ) [ 0 , m a x ( a , b ) ]
InDel (↑)similarity R I D ( a , b ) [ 0,1 ]
Jaro–Winkler (↑)similarity S J W a , b 0,1
Jaccard (↑)similarity S J a c ( a , b ) [ 0,1 ]
1 ↓ means the lower the value, the better. ↑ means the higher the value, the better.
Table 8. Comparison of text before and after replacement.
Table 8. Comparison of text before and after replacement.
Before ReplacementAfter Replacement
This show really is the Broadway American Idol. It has singing, the British Guy, A guy who’s sometimes nice, and a super-nice woman.This show really is the Broadway American Idol. It has Informatics 12 00056 i018, the Informatics 12 00056 i031Informatics 12 00056 i019, a Informatics 12 00056 i019 who’s sometimes Informatics 12 00056 i020, and a Informatics 12 00056 i020 Informatics 12 00056 i021. 1
1 Informatics 12 00056 i018 is “singing”. Informatics 12 00056 i031Informatics 12 00056 i019 is “British Guy”. Informatics 12 00056 i019 is “guy”. Informatics 12 00056 i020 is “nice”. Informatics 12 00056 i020 Informatics 12 00056 i021 is “super-nice woman”.
Table 9. Parameters of the pre-trained model based on the corpus. (The static mask is a pattern of generating masks in the data preprocessing stage, while the dynamic mask is a pattern of generating masks randomly in real time during model training.).
Table 9. Parameters of the pre-trained model based on the corpus. (The static mask is a pattern of generating masks in the data preprocessing stage, while the dynamic mask is a pattern of generating masks randomly in real time during model training.).
ModelModel-LayerPre-Training Tasks
Bert-base-E12-layerMLM (static mask), NSP
Bert-large-E24-layerMLM (static mask), NSP
Roberta-E12-layerMLM (dynamic mask)
Bert-base-C12-layerMLM (static mask), NSP
Bert-large-C24-layerMLM (static mask), NSP
Roberta-C12-layerMLM (dynamic mask)
Table 10. The hyper-parameter setting of the model.
Table 10. The hyper-parameter setting of the model.
Hyper-ParametersIEEA-CIEEA-E
learning rate1 × 10−52 × 10−5
training batch size3232
maximum context length128128
number of training epochs2020
Table 11. English prompt template.
Table 11. English prompt template.
Type T L M -E T P L M -E T E M J P L M -E
Prompt Template[input] [mask]Complete this sentence: [input] [mask]Related: [symbol]. Complete this sentence: [input] [mask].
ExampleI do not [mask] in the FINRA.Complete this sentence: I do not [mask] in the FINRA.Related: bee leaf. Complete this sentence: I do not [mask] in the FINRA.
Table 12. Chinese prompt template.
Table 12. Chinese prompt template.
Type T L M -C T P L M -C T E M J P L M -C
Prompt Template[input] [mask]请补全这个成语:[input] [mask]。 1请补全这个成语:[input] [mask]。提示:与[symbol]有关。 2
Translation[input] [mask]Please complete this idiom: [input] [mask].Please complete this idiom: [input] [mask]. Tip: It is related to [symbol].
Example[mask]动资金请补全这个成语: [mask]动资金请补全这个成语:[mask]动资金。提示:与牛有关。
1 “请补全这个成语:” means “Please complete this idiom:”. 2 “提示:与[symbol]有关。” means “Tip: It is related to [symbol].”
Table 13. Answer space thresholds corresponding to answer space step.
Table 13. Answer space thresholds corresponding to answer space step.
Answer Space Step12345
IER-SMCEM-base-E44.555.56
IER-SMCEM-Roberta-E99.51010.511
IER-SMCEM-large-E77.588.59
IER-SMCEM-base-C6.577.588.5
IER-SMCEM-Roberta-C6.577.588.5
IER-SMCEM-large-C6.577.588.5
Table 14. Example of case.
Table 14. Example of case.
Original DataSentiment LabelImplicit
Expressions
Analysis ResultsIER-SMCEMAnalysis Results
I believe in the FINRA. Well actually, I forgot to mention that ‘diamond hands’ is typically ‘paper hands’ when faced with market fluctuations.PositiveI Informatics 12 00056 i013Informatics 12 00056 i014 in the FINRA.
Informatics 12 00056 i022Informatics 12 00056 i023, I forgot to mention that ‘Informatics 12 00056 i008Informatics 12 00056 i024’ is typically Informatics 12 00056 i025Informatics 12 00056 i024 when faced with market fluctuations.
NegativeI believe in the FINRA. Well actually, I forgot to mention that ‘diamond hands’ is typically ‘paper hands’ when faced with market fluctuations.Positive
For the most part, I spy was an amusing lark that will probably rank as one of Murphy’s better performances in one of his lesser-praised movies.PositiveFor the most part, I Informatics 12 00056 i026 was an amusing Informatics 12 00056 i027 that will probably Informatics 12 00056 i028 as one of Murphy’s Informatics 12 00056 i029 performances in one of his lesser-praised movies.NeutralFor the most part, I spy was an amusing lark that will probably rank as one of Murphy’s better performances in one of his lesser-praised movies.Positive
Table 15. Data distribution in the two datasets used for sentiment analysis.
Table 15. Data distribution in the two datasets used for sentiment analysis.
Sentiment AnalysisIEEA-EIEEA-C
Positive11,24013,157
Negative11,24013,157
Table 16. Impact of recovered text on traditional sentiment analysis models (English).
Table 16. Impact of recovered text on traditional sentiment analysis models (English).
IER-SMCEM-Base-EIER-SMCEM-Large-EIER-SMCEM-Roberta-E
RecoverFalseTrueFalseTrueFalseTrue
TextCNN70.37%73.65%69.66%73.65%70.37%72.51%
BiLSTM77.95%80.04%79.16%80.39%77.81%79.47%
Bert81.18%81.85%80.97%82.10%81.46%81.68%
Bert-large83.45%85.79%84.01%86.50%82.10%83.02%
CART61.87%63.49%61.87%63.27%61.87%62.50%
XGB69.52%72.37%69.52%72.23%69.52%71.16%
DistilRoBERTa84.23%87.42%84.73%87.00%87.28%87.99%
Table 17. Impact of recovered text on traditional sentiment analysis models (Chinese).
Table 17. Impact of recovered text on traditional sentiment analysis models (Chinese).
IER-SMCEM-Base-CIER-SMCEM-Large-CIER-SMCEM-Roberta-C
RecoverFalseTrueFalseTrueFalseTrue
TextCNN82.24%84.61%82.85%83.69%82.76%83.59%
BiLSTM81.07%81.62%83.65%83.74%83.46%84.86%
Bert81.39%81.48%80.96%81.29%81.62%81.91%
Bert-large79.41%80.31%78.28%80.92%80.26%80.45%
CART69.66%71.05%69.65%70.91%69.65%71.19%
XGB62.56%64.14%62.56%63.95%62.56%64.19%
DistilRoBERTa86.74%87.07%84.16%87.92%85.66%86.46%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Wang, C.; Liu, Z.; Deng, H.; Li, Q.; Zheng, B. IER-SMCEM: An Implicit Expression Recognition Model of Emojis in Social Media Comments Based on Prompt Learning. Informatics 2025, 12, 56. https://doi.org/10.3390/informatics12020056

AMA Style

Zhang J, Wang C, Liu Z, Deng H, Li Q, Zheng B. IER-SMCEM: An Implicit Expression Recognition Model of Emojis in Social Media Comments Based on Prompt Learning. Informatics. 2025; 12(2):56. https://doi.org/10.3390/informatics12020056

Chicago/Turabian Style

Zhang, Jun, Chaobin Wang, Ziyu Liu, Hongli Deng, Qinru Li, and Bochuan Zheng. 2025. "IER-SMCEM: An Implicit Expression Recognition Model of Emojis in Social Media Comments Based on Prompt Learning" Informatics 12, no. 2: 56. https://doi.org/10.3390/informatics12020056

APA Style

Zhang, J., Wang, C., Liu, Z., Deng, H., Li, Q., & Zheng, B. (2025). IER-SMCEM: An Implicit Expression Recognition Model of Emojis in Social Media Comments Based on Prompt Learning. Informatics, 12(2), 56. https://doi.org/10.3390/informatics12020056

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop