Next Article in Journal
Maximum Power Efficiency
Previous Article in Journal
Experimental Advances in Phase Estimation with Photonic Quantum States
Previous Article in Special Issue
A Hybrid Quantum-Classical Model for Stock Price Prediction Using Quantum-Enhanced Long Short-Term Memory
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Contrastive Learning Pre-Training and Quantum Theory for Cross-Lingual Aspect-Based Sentiment Analysis

by
Xun Li
and
Kun Zhang
*
School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(7), 713; https://doi.org/10.3390/e27070713
Submission received: 18 June 2025 / Accepted: 30 June 2025 / Published: 1 July 2025
(This article belongs to the Special Issue The Future of Quantum Machine Learning and Quantum AI, 2nd Edition)

Abstract

The cross-lingual aspect-based sentiment analysis (ABSA) task continues to pose a significant challenge, as it involves training a classifier on high-resource source languages and then applying it to classify texts in low-resource target languages, thereby bridging linguistic gaps while preserving accuracy. Most existing methods achieve exceptional performance by relying on multilingual pre-trained language models (mPLM) and translation systems to transfer knowledge across languages. However, little attention has been paid to factors beyond semantic similarity, which ultimately hinders classification performance in target languages. To address this challenge, we propose CLQT, a novel framework that combines contrastive learning pre-training with quantum theory to address the cross-lingual ABSA task. Firstly, we develop a contrastive learning strategy to align data between the source and target languages. Subsequently, we incorporate a quantum network that employs quantum projection and quantum entanglement to facilitate effective knowledge transfer across languages. Extensive experiments reveal that the novel CLQT framework both achieves strong results and has a beneficial overall influence on the cross-lingual ABSA task.

1. Introduction

Aspect-based sentiment analysis (ABSA) is defined as extracting aspect information from text and determining the corresponding sentiment polarity [1,2,3]. For example, in a restaurant review, “The food is great, but the service was awful” in Figure 1. In this sentence, there are two aspect terms, “food” and “service,” respectively. The polarity of two aspects is food (positive) and service (negative). The vast availability of public opinion data allows stakeholders to formulate strategic plans by analyzing the ratio of favorable feedback. However, manually examining and interpreting a large number of user comments is both labor-intensive and costly, emphasizing the necessity for an automated approach.
Cross-lingual sentiment analysis (CLSA) involves training a classifier using source languages, typically resource-rich languages, and applying it to the target languages [4,5,6], as illustrated in Figure 1b. In contrast to monolingual sentiment analysis, CLSA accounts for structural and meaning variations across different linguistic systems, facilitates cross-linguistic knowledge transfer, and enables sentiment analysis models trained on data-rich languages to be effectively adapted for studying low-resource languages. This approach effectively tackles the challenge of limited sentiment resources in many non-English languages.
In recent years, multilingual BERT(mBERT) models and XLM-Roberta (XLM-R) models have become essential tools for addressing cross-lingual aspect-based sentiment analysis (ABSA) [7,8]. These pre-trained models acquire multilingual knowledge during their initial training phase and are subsequently adapted using annotated data from a high-resource language (typically English), allowing them to be applied to other target languages [9]. To further strengthen the transfer of language-specific information for cross-lingual aspect-based sentiment analysis, Zhang et al. [10] proposed a method that incorporates lexical alternation alongside a knowledge compression technique, aiming to enhance the alignment of these pre-trained models across different linguistic systems.
Despite substantial advancements in multilingual pre-trained models, their effectiveness in ABSA across different languages remains challenging, mainly due to the inadequate representation of low-resource languages during the pre-training process. One usual method for transferring language-specific knowledge is to use translated target language data with corresponding labels [11]. However, the effectiveness of methods relying on translated text is heavily dependent on translation accuracy and precise label alignment. Therefore, obtaining high-quality parallel translation data and eliminating translation errors remains a long-standing challenge.
Quantum mechanisms offer distinct advantages for cross-lingual ABSA [12]. In particular, quantum entanglement allows the model to encode correlations among various linguistic elements, such as aspect terms, opinion words, and sentiment polarities, simultaneously, thereby capturing intricate dependencies that traditional approaches might overlook. These entangled representations facilitate a more precise interpretation and analysis of sentiment information. Furthermore, quantum superposition enables a quantum system to exist in multiple states concurrently, a property that is particularly beneficial for NLP. Within a Hilbert space, superposition can represent several semantic layers at once, accommodating the inherent ambiguity and richness of language. Since words and phrases can convey different meanings depending on context, leveraging superposition allows the model to encode multiple semantic interpretations concurrently, thereby enhancing its understanding of complex linguistic structures. This capability renders the model more flexible and efficient in managing semantic variations and sentiment expressions across diverse languages.
Zhao et al. [12] applied quantum entanglement to connect subsystems in quantum systems and resolving cross-lingual ABSA challenges through integration of quantum components within multilingual pre-trained models. However, this approach primarily relies on traditional multilingual models and does not effectively leverage the annotated data available in the high-resource domain. To address this limitation, we design a contrastive learning pre-training strategy that incorporates sentence translation to facilitate cross-lingual alignment and knowledge transfer. Contrastive learning is then employed to bring semantically similar sentences closer together while pushing semantically dissimilar sentences apart. This method not only maximizes the use of source language label knowledge but also reduces errors caused by translation quality. In contrast to [12], our method devises a dynamically adjustable entanglement state, which is governed by a linguistic similarity-weighted gating mechanism embedded in the quantum entanglement framework. The structural and grammatical differences between languages result in varying degrees of similarity across languages. For example, French and English are more similar than English and Russian. The quantum entanglement module proposed in [12] assumes equal influence across all languages, which is inconsistent with the nature of languages. To overcome this limitation, we introduce a dynamic entanglement module based on word vectors. Our proposed model, CLQT, combines contrastive learning pre-training with quantum mechanism networks, as shown in Figure 2.
We present CLQT, a novel framework combining quantum and contrastive learning pre-training, designed to enhance aspect sentiment analysis across multiple languages. This model consists of two main components: contrastive learning pre-training and quantum mechanism modules. The quantum mechanism module is further divided into the quantum projection module and the entanglement module. Specifically, to capture general knowledge across languages, the quantum projection module maps the text features obtained through contrastive learning pre-training into Hilbert space, where each aspect term is represented as a quantum superposition state. Additionally, to obtain language-specific knowledge, we design a quantum dynamic entanglement module, enabling the sharing of specific knowledge between languages through entangled representations. Finally, we connect the complex representations from both the quantum projection and entanglement modules to perform sequence labeling classification.
Our contributions are as follows:
  • This work proposes an innovative approach to cross-lingual ABSA that integrates contrastive learning pre-training with quantum mechanisms to enhance cross-lingual adaptability. To the best of our knowledge, this study is the earliest attempt to integrate these methodologies into aspect-based sentiment classification across multiple languages.
  • In the quantum mechanism module, we employ a dynamic gate based on language similarity to establish quantum entanglement states, which more effectively capture the varying degrees of similarity between languages.
  • We evaluated our CLQT model on the SemEval-2016 dataset [1]. The experimental results demonstrate the robustness of our proposed framework. Furthermore, a comprehensive evaluation was performed to analyze the contribution of each key module. The analysis further demonstrates the framework’s strong capability in transferring knowledge across languages.

2. Related Work

This section provides a concise overview of related research, encompassing aspect-based sentiment analysis, contrastive learning, and quantum networks.

2.1. Cross-Lingual Aspect-Based Sentiment Analysis

Conventional sentiment analysis is generally conducted at either the sentence or document level [13], while ABSA targets more fine-grained sentiment classification at the entity level [14]. The objective of cross-lingual ABSA is to develop an aspect-based sentiment classification model in one language and subsequently transfer it for use in another. Recent studies on cross-lingual ABSA have predominantly concentrated on subtasks involving the identification of aspect terms and the determination of their sentiment across different languages [15]. Most approaches utilize parallel corpora derived from machine translation, which help capture the semantic and syntactic knowledge of the target language [16]. Word or phrase alignment algorithms are then utilized to map label information between sentences in different languages [17], the effectiveness of these approaches largely relies on the accuracy of both translation and alignment. To enhance this approach, a technique leveraging multilingual word embeddings learned from extensive bilingual datasets has been introduced [18], allowing word embeddings to transfer across languages. Recently, transformer-based models pre-trained on extensive multilingual datasets, including multilingual BERT [7], XLM-Roberta [8], and mT5 [19], have achieved remarkable results across a range of cross-lingual NLP tasks. These models acquire multilingual knowledge during the pre-training phase from large corpora. These models are initially adapted using annotated examples in a high-resource language and are then directly utilized for predictions in the target language, exhibiting strong performance. To tackle the challenge of cross-lingual aspect-based sentiment classification, numerous studies have explored the integration of translation systems with pre-trained models, leveraging their ability to capture multilingual representations effectively [20,21]. To investigate the role of language-specific knowledge in cross-lingual aspect-based sentiment classification, Zhang et al. [10] proposed an aspect term code-switching (ACS) model.This architecture utilizes unsupervised learning mechanisms for aspect term extraction and is later adapted using annotated examples from a high-resource corpus, achieving results comparable to those of supervised approaches.

2.2. Contrastive Learning

The core idea behind contrastive learning is to create paired samples that enable the model to learn more distinct feature representations. Recently, this technique has achieved significant improvements across multiple natural language processing applications, mainly due to its ability to develop robust text embeddings [22]. Mitra et al. introduced a method combining memory networks and contrastive learning for cross-lingual stance detection [23]. Guo et al. applied contrastive learning to detect Alzheimer’s disease by bringing average vectors of confirmed cases closer together and pushing those of non-diagnosed cases further apart [24]. Li et al. proposed a representation learning strategy based on contrastive objectives to identify implicit sentiment in aspect-level sentiment classification [25]. Luo et al. enhanced the performance of contrastive learning by incorporating in-batch negative samples [26]. Zhou et al. integrated an emotion-aware thinking chain prompt module with contrastive learning to capture label relevance [27].
Contrastive learning has demonstrated its effectiveness across diverse natural language processing applications, particularly in learning semantic representations of labels within monolingual settings. However, traditional contrastive learning is constrained to within-language feature comparisons, often disregarding semantic relationships between different languages and sentences. Conventional methods primarily emphasize feature alignment within a single language, lacking the ability to fully leverage the rich cross-linguistic semantic information. This limitation becomes especially apparent in tasks that demand a deeper integration and comprehension of multiple semantic layers [28]. To address this issue, this paper strengthens representation learning by integrating sentence translations to create both similar and contrasting sample pairs, thereby enhancing overall performance.

2.3. Quantum Neural Network

Quantum neural networks (QNN) extend the functionality of traditional neural networks by leveraging quantum properties such as superposition and entanglement. Li et al. introduced an innovative fusion mechanism that incorporates quantum cognition into neural networks for emotion prediction. This method employs the quantum superposition state of judgments within a complex Hilbert space, utilizing positive operator-valued measures to classify samples as expressing either positive or negative emotions [29]. Yan et al. introduced an innovative quantum-based model for solving optimization challenges, employing a supervised learning framework to achieve enhanced performance [30]. Additionally, Zhou et al. proposed a generative adversarial network that integrates quantum and classical components to synthesize images by modeling discrete distributions [31]. Gong et al. construct a variational quantum circuit that incorporates quantum principles into convolutional neural networks to enhance model performance [32].
Building on recent progress in quantum-enhanced learning, we propose a novel framework for cross-lingual ABSA tasks. This framework combines quantum mechanisms with contrastive learning pre-training methods, aiming to improve performance in sentiment analysis across languages.

3. Method

This part presents the architecture of the CLQT model, outlining its key components and design principles. The model includes two main modules: contrastive learning pre-training strategy and quantum networks. In the pre-training phase, we apply a contrastive learning approach to reduce the semantic gap between source instances and their respective translated counterparts. Inspired by [12], in the fine-tuning phase, we employ quantum projection, quantum measurement, and quantum entanglement techniques to project both source and target languages into Hilbert space. We construct entangled states across multiple languages by the fundamental properties of quantum systems and using quantum measurements to compute the probability distributions for each aspect. Finally, we propose a parameterized rotation gate to replace the static Controlled-NOT (CNOT) gate, which enables dynamic adjustment of entanglement states based on the linguistic similarity between languages.

3.1. Problem Definition

Cross-linguistic sentiment analysis at the aspect level requires detailed text prediction and can be represented using a sequence labeling framework [33,34]. x = { x i } i = l N represents a sentence, and the sentence length is N; d  x i represents the i-th word; the model predicts the label sequence { y i } i = 1 L , where y i Y = { B , I , E , S } { P O S , N E U , N E G } { O } represents the label of word x i , where B denotes the beginning of an aspect term, I represents the intermediate word, E denotes the end of the aspect term, and S represents a single aspect term. { P O S , N E U , N E G } represent the polarity of the aspect term, and O represents non-aspect terms. For example, y i = B N E G means the word x i is the beginning of a sentiment-negative aspect term. In this study, we only use the label from the source language ( x S , y S ) D S to predict the y T for the target language text x T .

3.2. Contrastive Learning Pre-Training

As a self-supervised approach, this technique aims to acquire meaningful data representations by analyzing similarities and distinctions among samples. Specifically, an anchor sample is used as a reference, positive samples are semantically similar to the anchor. In the training phase, the feature representations of positive samples are brought closer to the anchor. In contrast, negative samples, which are semantically dissimilar to the anchor, have their representations pushed further apart. In this study, we construct a contrastive learning module by representing source languages and their translated samples in a batch as anchor and positive samples, respectively, and all other samples within the batch are treated as negative instances. Contrastive learning pre-training strategy is described in Algorithm 1. Let x = x i N denote a sentence, where N is the sentence length and x i represents the i-th word. We apply the multilingual pre-trained model (mPLM) to encode the sequence x, resulting in:
H S = { h 1 , h 2 , , h n } = m P L M ( x 1 , x 2 , , x n )
H T i represents the feature representation of the translated target language encoded through m P L M , where T i denotes the target language, P represents the total count of target languages, which we set to four in this study. H N represents the feature representation of non-translated samples. The pre-training contrastive learning loss function is defined as
l c l = i = 1 B l o g 1 P j = 1 P e x p ( S i m ( H S i , H T j ) / τ ) k = 1 , k i B e x p ( S i m ( H S i , H N k ) / τ )
Here B represents the number of samples in a batch; τ represents the temperature coefficient.
Algorithm 1 Contrastive learning Pre-training
  Input: 
H S , H T , H N , P, temperature coefficient τ , batch-size B, epochs e.
  Output: 
The model M
  1:
Initialization parameter M
  2:
for  e p o c h in range ( 1 , e + 1 )  do
  3:
    for i in range ( 1 , B + 1 )  do
  4:
        for j in range ( 1 , P + 1 )  do
  5:
           # Compute the similarity between languages data
  6:
            S i m ( H S , H T ) = H S · H T | | H S | | | | H T | |
  7:
           # define contrastive learning l o s s
  8:
            l c l = log e x p ( S i m ( H S , H T ) / τ ) e x p ( S i m ( H S , H N ) / τ )
  9:
        end for
10:
    end for
11:
    # Back propagation and optimization
12:
    Update model M using gradient descent to minimize l c l
13:
end for
14:
return M

3.3. Quantum Theory

This subsection focuses on how to apply quantum theory for cross-lingual ABSA, including the application of quantum projection, quantum measurement, and quantum entanglement. We first introduce the foundational concepts of quantum theory.

3.3.1. The Fundamental Knowledge of Quantum Theory

Following Busemeyer et al. [35] and Fell et al. [36], we introduce key foundational concepts in quantum theory.
Quantum Projection
In quantum cognition, an infinite-dimensional complex vector space known as a Hilbert space H plays a pivotal role. A Hilbert space is endowed with an inner product that allows quantum states to be represented as unit vectors. Unlike classical probability, quantum probabilities derive from orthogonal basis states, with the relationship between a state vector and these bases governed by projection geometry. Notably, a single Hilbert space can be described by multiple orthogonal basis sets, which form the mathematical foundation for quantum operations such as projection and measurement. This theoretical framework is essential for understanding the application of quantum principles in our model.
Quantum Superposition
Quantum superposition is a core principle of quantum mechanics, suggesting that multiple states coexist within a quantum system, persisting in this state of overlap until a measurement forces a collapse into a single, well-defined outcome. A quantum system composed of an electron or a photon can remain in a superposition, where it occupies multiple possible states simultaneously. A pure state | ψ is represented as a vector on the unit sphere and is defined as follows:
| ψ = ω 1 | e 1 + ω 2 | e 2 + + ω n | e n
Here, { | e 1 , | e 2 , , | e n } represents the orthogonal basis that defines the Hilbert space, while the probability amplitudes { ω 1 , ω 2 , , ω n } are complex scalars satisfying the normalization condition i = 1 n | ω i | 2 = 1 , where | · | denotes the modulus of a complex number. The superposition state | ψ does not coincide with any basis state | e i . For instance, in the two-dimensional Hilbert space H 2 , a pure state | ψ formed by the basis states | 0 and | 1 can be defined as:
| ψ = c o s θ 2 | 0 + e i ϕ s i n θ 2 | 1
where i 2 = 1 , θ [ 0 , 2 π ] and ϕ [ 0 , 2 π ] .
Quantum Measurement
The process of measuring a quantum system plays a crucial role in quantum theory, describing how a system initially existing in a superposition of multiple states collapses into a single, well-defined state upon observation. This process is essential for defining quantum probabilities within the framework of quantum cognition. The projection-valued measure (PVM) maps a system’s state from uncertainty to a distinct outcome by aligning it with its corresponding eigenstate. The system remains in a superposition, encompassing all possible outcomes before a measurement t is made. However, the system collapses into a particular eigenstate, when the measurement t is made.
Quantum Entanglement
When multiple quantum entities interact and become entangled, their individual states can no longer be fully described independently. Instead, these entities collectively form a unified quantum system represented by a pure entangled state, denoted as
| C l H 1 H 2 H N
within the composite Hilbert space. This global pure state encapsulates non-classical correlations between subsystems, crucial for our cross-lingual ABSA framework. Importantly, while the joint entangled state | Ψ remains pure, the state of any individual subsystem i, considered separately, is described by a mixed quantum state, namely the reduced density matrix obtained by tracing out the other subsystems:
ρ i = Tr j i | C l C l | .
This mixed subsystem state reflects both local properties and the global correlations arising from entanglement. Consequently, observing or measuring one subsystem instantaneously influences the statistical description of the others, capturing precisely the deep interdependencies across languages utilized in our framework.

3.3.2. Quantum Module Design

Inspired by [12], we apply quantum mechanics to solve the cross-lingual ABSA task. First, the feature representations obtained from contrastive learning pre-training serve as inputs to the quantum projection module. These features are then linearly projected into a three-dimensional Hilbert space, with measurement probabilities derived through observable operators to compute the sentiment polarity probabilities. Specifically, feature representations obtained from contrastive learning pre-training are linearly transformed into a three-dimensional Hilbert space using a projection matrix. This transformation encodes each aspect term as a quantum superposition state, enabling the model to capture multiple semantic layers simultaneously. The formal framework underlying this projection is critical for elucidating how quantum states encode linguistic features and facilitate cross-lingual knowledge transfer.
In cross-lingual aspect-based sentiment analysis, the quantum module offers two decisive benefits. Superposition naturally captures sentiment ambiguity. A three-level quantum state encodes positive, neutral, and negative polarities in a single complex vector, allowing the model to postpone a hard decision until sufficient context is observed. Classical probability vectors, by contrast, must be normalized into mutually exclusive states before inference and therefore cannot express such delayed collapse. Entanglement enables cross-lingual knowledge transfer. By preparing a global pure state, the source- and target-language subsystems are bound in a non-separable way; a measurement on the source instantly updates the target’s reduced density matrix, achieving efficient knowledge migration. Thanks to the expressive power of superposition for ambiguous sentiment and the global correlations afforded by entanglement, the quantum module delivers advantages that a purely classical probabilistic framework cannot replicate.
Quantum Projection Module
Each aspect word is represented as a mutually exclusive observable value within a complex-valued Hilbert space. Additionally, the feature representation h of words is projected into a three-dimensional Hilbert space after contrastive learning pre-training, obtaining the quantum state:
| A i S = W p h i
where W P is the projection matrix. The Hilbert space is a three-dimensional vector space defined by the basis states { | + , | , | 0 } . Here, the basis states { | + , | , | 0 } represent positive, neutral, and negative sentiments, respectively. The aspect word A i S is represented as a pure state within the three-dimensional Hilbert space, denoted as | A i S , Figure 3 shows the representation of aspects in Hilbert space, where each aspect word is represented by the pure state | A i S in the three-dimensional Hilbert space with positive, neutral, and negative sentiments.
| A i S = α | + + β | 0 + γ |
where | α | 2 + | β | 2 + | γ | 2 = 1 . Here, the coefficients α , β and γ indicate the contribution of each sentiment state to the overall representation.
A projection matrix W p is applied to the feature representation h i of the aspect word, yielding an unnormalized vector:
W p h i = [ α i , β i , γ i ]
which is then normalized to produce the coefficients:
α i = α i | α i | 2 + | β i | 2 + | γ 2 |
β i = β i | α i | 2 + | β i | 2 + | γ 2 |
γ i = γ i | α i | 2 + | β i | 2 + | γ 2 |
Defining the sentiment observable operator in the three-dimensional Hilbert space:
S ^ i A = ( + 1 ) | + + | + ( 0 ) | 0 0 | + ( 1 ) | |
The eigenvalues + 1 , 0 , 1 represent positive, neutral, and negative sentiments, respectively. The probability of sentiment category is calculated through quantum measurement:
P ( k ) i = | k | A i S | 2
where k { | + , | 0 , | } , P ( + ) i = | α i | 2 is the probability of positive sentiment, P ( 0 ) i = | β i | 2 is the probability of neutral sentiment, and P ( ) i = | γ i | 2 is the probability of negative sentiment. This equation quantifies the likelihood of observing each sentiment by taking the squared magnitude of the projection of the state | A i S onto the corresponding basis state. These equations formalize how our model encodes linguistic sentiment information within a quantum framework, leveraging the principles of quantum superposition and measurement to capture complex semantic interdependencies. Consequently, the measured sentiment probability vector is defined as follows:
p i = [ P ( + ) i , P ( 0 ) i , P ( ) i ]
where P ( + ) i , P ( 0 ) i , P ( ) i represent positive, neutral, and negative probability, respectively.
Quantum Entanglement Module
Cross-lingual ABSA tasks rely heavily on language-specific knowledge. In quantum mechanics, when multiple particles interact, each particle can acquire information about the entire quantum system through the observation of its individual state. Similarly, we create quantum entangled states between different languages to facilitate the sharing of specific linguistic knowledge. Quantum entanglement distinguishes itself from classical attention mechanisms by enabling the simultaneous representation and interaction of multilingual semantic elements within a unified quantum framework. Unlike classical attention methods, which typically aggregate information through weighted summation, quantum entanglement encodes cross-lingual semantic interactions in a high-dimensional complex vector space, allowing multiple linguistic states to coexist in superposition. This coexistence inherently accommodates linguistic ambiguity, facilitating a richer and more nuanced representation of sentiment-related information.
Therefore, Zhao et al. [12] designed a five-qubit entangled state; however, this design is fixed and assumes symmetric relationships between the languages. In reality, the relationships between languages are asymmetric. For example, the semantic overlap between English and French is more significant than that between English and Russian. Therefore, we propose an adaptive quantum entanglement module that dynamically adjusts the entanglement strength according to the similarity between languages. Additionally, we introduce quantum gates with variable parameters to ensure that the entangled state captures the specific feature across languages.
Language Similarity Calculation
We extract the top 100 high-frequency words V i for each language L i from the dataset and encode them as word vectors e v i . The center of language embedding can be calculated as
E i = 1 | V i | v V i e v i
where | V i | = 100 . The similarity between any two languages can be computed as:
S i j = E i · E j | | E i | | | | E j | |
where S i j represents the cosine similarity between the average embeddings E i and E j of languages L i and L j . The entanglement weight is formulated by
w i j = s i j k , l = 1 5 s i k
where w i j is the normalized entanglement weight between languages. This normalization ensures that the weights w i j reflect the relative similarity of language L i to each target language. Since the similarity between languages varies, we define the dynamic entangled state as
| C l = w i j | ψ i | ψ j
where | ψ i = a i | 0 + b i | 1 is the quantum state of language L i , where | 0 represents spans the language-invariant subspace, where | 1 represents spans the language-specific subspace, and the tensor product | ψ i | ψ j represents the joint state of language. The weights w i j modulate the contribution of each language pair based on their computed similarity, allowing the entangled state to dynamically reflect the asymmetric relationships between languages. This equation defines the dynamic entangled state | C l as a weighted sum of the tensor products of quantum states from different languages. Measurement probability of | ψ i :
P ( | ψ i ) = T r j i ( | C l C l | )
where T r j i represents the partial trace operation, summing over all qubits except the i-th one ( i j ), yielding the reduced density matrix of the i-th subsystem. | C l C l | represents the density matrix of the entanglement state, a projection operator of | C l . This function extracts the state of an individual language via the trace operation, enabling the computation of language-specific knowledge. Thus, language-specific knowledge representation can be defined as
e i = ψ i | E ^ | ψ i
where ψ i | represents the conjugate transpose (bra vector) of | ψ i . E ^ represents the observable operator (Hermitian operator), defined as
E ^ = k λ k | k k |
where k represents the basis state index ( k = 0 , 1 ), corresponding to | 0 and | 1 . λ k is the eigenvalue, representing the physical quantity measured in state | k . | k k | represents the projection operator.
Consequently, language knowledge vectors spanning multiple languages can be derived:
e = [ e 1 , e 2 , e 3 , e 4 , e 5 ]
where e represents a vector of knowledge representations for the five languages, each component e i corresponds to a language. The application circuit of entanglement state is shown in Figure 4.
By measuring its own quantum state, each particle can obtain language-specific knowledge from other particles within the system, enabling the exchange of language-specific information.
Fusion of Quantum Entanglement and Quantum Projection
To adapt global language knowledge to the sentiment distribution of each word, the language knowledge is mapped into the emotion space.
e i = W e e
where e i represents the adjusted language knowledge vector, aligned with the sentiment space. W e represents the mapping matrix. Integrating the knowledge of quantum measurement and quantum entanglement:
q i = ( 1 μ ) p i + μ e i
where q i represents fused sentiment probability distribution, μ represents fusion weight, μ [ 0 , 1 ] , p i represents sentiment probability from the measure module. The normalized representation of q i is given by:
q i = q i k q i , k
where q i , k represents the k-th component of q i . To map the fused probabilities to the label space for sequence tagging, the final probability is expressed as follows:
y ^ i = W o q i + b o
where W o and b o represent the weight matrix and bias vector, respectively. We optimize the system using the cross-entropy function:
L c e = a D t r a i n c C y · log s o f t m a x ( y ^ i )
where a represents aspect words, and C represents the number of categories. The model parameters are optimized through gradient descent by minimizing the cost function.

3.3.3. Resource Estimation for the Quantum Circuit

To precisely assess the computational requirements for our proposed quantum algorithm, we provide a detailed resource estimation, including circuit width, depth, measurement strategy, and the number of shots for sufficient measurement precision. Table 1 presents the circuit width.
The circuit consists primarily of an entanglement layer that encodes cross-lingual correlations. For each token, the operation involves one single-qubit Hadamard gate to prepare superposition on the reference language qubit and four CRY gates to establish entanglement between the reference and the other four language qubits. Thus, the single-layer entanglement structure has a gate depth of approximately 5 layers.
Measurement Strategy and Shots Recommendation. In state-vector simulation, we obtain exact outcome probabilities by directly computing the squared amplitudes, so no shot-based sampling is required. On real quantum hardware, to ensure that the statistical uncertainty in the estimated probabilities 95% confidence half-width remains below 0.03 , we recommend performing approximately 1024 measurement shots on all five language qubits and aggregating the observed frequencies.

4. Experiments

Quantum mechanics is typically represented through transformation matrices in mathematics [37]. Therefore, we represent quantum states in a linear algebraic form to evaluate the effectiveness of CLQT. A comprehensive evaluation is then performed utilizing the SemEval-2016 dataset [1] and the Amazon Reviews Corpus [38].

4.1. Dataset

The SemEval-2016 dataset comprises consumer opinion text in five different languages, including English, French, Spanish, Dutch, and Russian. For each language, we divide the dataset into training and test datasets. Additionally, 20% of the training samples are randomly chosen to serve as a validation dataset. Table 2 presents the SemEval-2016 dataset statistics, where “No.Sen” denotes the overall sentence count, and “No.Asp” denotes the total count of aspect terms.
The Amazon Reviews Corpus is a large-scale multilingual text classification dataset, widely used for ABSA. It includes consumer product reviews in six languages (English, German, French, Spanish, Japanese, and Chinese), annotated with star ratings ranging from 1 to 5, reflecting reviewer satisfaction. For our analysis, we categorize reviews into negative (ratings below 3), neutral (ratings equal to 3), and positive (ratings above 3). Crucially, the dataset maintains class balance, ensuring equal representation of sentiment classes. Adopting a zero-shot learning scenario, we utilize English as the source language by randomly selecting 512 reviews per sentiment category, thus compiling a total of 2560 samples for training. For validation, we randomly select 128 reviews per category from each of the remaining five languages, assembling a validation set of 3200 instances. The test set remains unchanged, comprising 5000 samples.

4.2. Experiment Setting

First, we translate the source language training data from the dataset into target languages to create a contrastive training dataset. The multilingual BERT and XLM-R pre-trained models are employed to encode aspects and sentences. For the contrastive learning pre-training phase, the learning rate is set to 2 × 10 4 , the batch size is 16, and the temperature coefficient is 0.1. In the quantum mechanism module, the learning rate is set to 5 × 10 5 , with a batch size of 16. To ensure the robustness and statistical reliability of our experimental results, we repeated each experiment five times using different random seeds. The average F1-scores, along with the corresponding standard deviations, were computed and reported. All experimental evaluations utilize the Adam optimizer.

4.3. Baseline Model

A comparative analysis is conducted between the introduced framework and multiple baseline approaches, along with state-of-the-art models, as follows:
  • Zero-shot model [39]: This method employs annotated examples from English and utilizes them directly for predictions in another language.
  • Translation-TA and Bilingual-TA [10]: These approaches leverage pseudo-labeled samples obtained via translation from a high-resource language and integrate both original and translated text for model training.
  • Translation-AF and Bilingual-AF [10]: These approaches emphasize transferring labels across languages without requiring precise word-level alignment.
  • ACS, ACS-DISTILL-S, and ACS-DISTILL-M [10]: These models propose a non-aligned label projection method, outperforming translation-based methods in performance. They introduce a code-switching mechanism to enrich knowledge transfer between languages using bilingual sentences.
  • XLM-RoBERTa [8]: This is a transformer-based multilingual pre-trained language model that extends RoBERTa to over 100 languages, trained with a masked language modeling objective on a large-scale CommonCrawl corpus. It serves as a robust encoder for cross-lingual transfer tasks.
  • CL-XABSA [20]: This model proposes two contrastive learning methods based on token labels and sentiment polarity, respectively, and integrates knowledge distillation with attention-based multilingual models to achieve cross-lingual aspect-based sentiment classification.
  • CAPIT [40]: The model achieves knowledge transfer from one language to another by integrating contrastive learning with a generative prompt-based large language model.
  • QPEN [12]: This quantum-enhanced network model employs quantum mechanics to solve the cross-lingual ABSA task and achieves the best performance.

4.4. Experimental Results

There are many non-aspect terms in sentences for cross-lingual ABSA tasks, causing a substantial imbalance in label distribution. Therefore, we use the F1 score to assess the CLQT framework. The results of the F1 score on the SemEval-2016 dataset and Amazon Reviews Corpus are shown in Table 3, Table 4 and Table 5.
Table 3 and Table 4 present F1 scores for the SemEval-2016 dataset, the proposed model achieves excellent performance across four languages, outperforming not only the baselines but also state-of-the-art models. We encode texts with different multilingual pre-trained models, including mBERT and XLM-R. We observed that CLQT outperforms both the mBERT and XLM-R models, with the improvement being more pronounced in the XLM-R model compared to mBERT. CLQT achieves better performance compared to QPEN methods, with an average absolute improvement of 1.07% in F1-score using mBERT and 2.74% using XLM-R. The results suggest that modeling aspects as quantum superposition states within a complex-valued Hilbert space, combined with facilitating the exchange of language-specific knowledge through a similarity-weighted gate quantum entanglement mechanism, can substantially improve the performance of cross-lingual ABSA tasks.
Table 5 summarizes the F1-score results of our experiments on the Amazon Reviews Corpus across five languages: German, Spanish, French, Japanese, and Chinese. Our method, denoted as CLQT, is compared against several baselines, including XLM-RoBERTa, ACS-Distill, CAPIT-base, CL-XABSA, CAPIT-large, and QPEN. Notably, CAPIT employs the mT5 [19] pre-trained encoder, which is optimally designed for generative tasks, whereas other models utilize the XLM-R pre-trained encoder.
The XLM-RoBERTa baseline achieves an average F1-score of only 50.86%, indicating that without specialized cross-lingual adaptation techniques, performance remains suboptimal. CAPIT integrates contrastive learning with large language models. In contrast, our proposed model achieves higher F1-score across all languages than CAPIT, demonstrating the benefits of combining contrastive learning with quantum networks. Similarly, the QPEN method employs the XLM-R pre-trained encoder and leverages quantum networks for performance enhancement; however, in its quantum entanglement module, different target languages are assigned the same entanglement state. By comparison, our approach significantly outperforms QPEN across all languages, yielding an average F1 improvement of 3.28%. This advantage arises not only from our dynamic entanglement module, which better adapts to real-world textual characteristics, but also from our pre-training strategy that effectively aligns semantically equivalent information across languages, thereby facilitating more effective knowledge transfer from the source to the target languages.

5. Analysis and Discussion

In this section, the performance of the proposed CLQT model is evaluated by verifying the effectiveness of each module, as well as ablation study, error analysis, case study, and visualization.

5.1. The Effectiveness of the Contrastive Learning Pre-Training

To validate the effectiveness of the contrastive learning approach in cross-lingual aspect-based sentiment analysis, we conduct both F1-score evaluations and visualization experiments on the SemEval-2016 dataset. The F1-score results are presented in Figure 5, while the visualization results are shown in Figure 6.
In Figure 5, blue bars represent the baseline pre-trained models, while orange bars indicate the models augmented with contrastive learning (mBERT-CL and XLM-R-CL). The x-axis denotes the four evaluated languages, French, Spanish, Dutch, and Russian, while the y-axis reflects F1 scores for the cross-lingual aspect-based sentiment classification task.
Across all languages, models enhanced with contrastive learning consistently outperform their baseline counterparts. For instance, in Spanish, mBERT-CL achieves an F1 score of nearly 58%, compared to approximately 56% for the original mBERT. A similar trend is observed with XLM-R, particularly in lower-resource languages such as Russian and Dutch, where contrastive learning leads to notable improvements. Even in Spanish, the best-performing language overall, contrastive learning yields consistent gains.
These F1 score results strongly validate the effectiveness of contrastive learning pre-training, demonstrating its ability to effectively reduce cross-lingual semantic gaps and significantly enhance the generalization capabilities of pre-trained language models across diverse linguistic settings.
Figure 6 presents the t-SNE visualizations of multilingual feature representations. These visualizations are generated based on the training dataset, which includes English (EN) samples and their corresponding translations in French (FR), Spanish (ES), Dutch (NL), and Russian (RU), encoded using the pre-trained XLM-R model.
(a) displays the feature distribution without the application of contrastive learning. Although the sentences across different languages are semantically equivalent, their encoded representations exhibit clear language-specific clustering, where samples from each language form separate clusters. This language-dependent distribution suggests that, without contrastive learning, the model struggles to align cross-lingual semantics effectively, potentially introducing noise during knowledge transfer and negatively impacting cross-lingual aspect-based sentiment classification. In contrast, (b) presents the feature distribution after applying contrastive pre-training. Representations of semantically equivalent samples across different languages are noticeably closer in the feature space, forming well-aligned, cross-lingual clusters. This indicates that contrastive learning facilitates the model in capturing language-invariant semantic information, thereby enhancing cross-lingual alignment.
These findings corroborate the quantitative improvements observed in Figure 5, confirming the effectiveness of contrastive learning in improving cross-lingual semantic consistency. By reducing semantic gaps between translated instances, contrastive learning helps mitigate translation-induced noise and enhances the generalization of the model and accuracy in multilingual aspect-based sentiment classification.

5.2. The Effectiveness of Quantum Mechanism Dynamic Entanglement

To assess the effectiveness of the dynamic entanglement mechanism proposed in our quantum framework, we devise a series of experiments on the SemEval-2016 dataset. These experiments examine the impact of integrating the dynamic similarity weighting strategy across various model architectures. The corresponding F1 scores are reported in Figure 7, providing empirical evidence of the utility of this component in enhancing cross-lingual aspect-based sentiment classification performance.
Subgraph (a) presents the experimental results of integrating the widely used multilingual model XLM-RoBERTa with the proposed Dynamic Similarity Weights (DS) module. XLM-RoBERTa captures contextual semantic representations via a multilingual attention mechanism. When enhanced with the DS module, it dynamically adjusts the fusion weights between source and target language representations based on their similarity, resulting in improved F1 performance across various languages. However, since the fusion is still performed via linear weighting, the overall performance gain remains relatively limited.
Subgraphs (c) and (d) further explore the impact of the DS module on two attention-based baseline models, CL-XABSA-TL and CL-XABSA-SL. In both cases, the inclusion of DS leads to measurable performance gains in French, Spanish, Dutch, and Russian, though the improvements are constrained by the limited capacity of attention mechanisms to capture complex semantic dependencies.
Subgraph (b) illustrates the performance of the proposed CLQT model and its ablated variants. Removing the quantum module entirely (CLQT w/o QM) results in the most pronounced performance degradation, underscoring the pivotal role of quantum components in capturing semantic relationships. A partial ablation, excluding only the dynamic entanglement module (CLQT w/o DE), still yields competitive results, exceeding those of XLM-R and CL-XABSA, but falls short of the full CLQT configuration. These findings highlight the unique contribution of dynamic quantum entanglement in modeling semantic alignment across languages. The complete CLQT model achieves the highest F1 scores across all languages, confirming the synergistic benefit of combining quantum projections, entanglement, and adaptive similarity weighting.
While conventional approaches such as attention or linear weighting can integrate cross-lingual features, they often rely on local information and struggle to fully model nonlinear semantic mappings across languages. In contrast, the proposed dynamic quantum entanglement mechanism constructs global entangled states in Hilbert space guided by dynamic similarity, enabling multilingual features to reside in a unified, cooperative quantum representation. This facilitates a more precise modeling of deep semantic dependencies during knowledge transfer. Experimental results confirm that the quantum entanglement mechanism, driven by dynamic similarity, significantly enhances the expressive power and generalization performance of the model in cross-lingual aspect-based sentiment classification tasks.

5.3. Ablation Study

To further explore the contribution of each module in the model, we designed several variants. The experimental results for each variant are presented in Table 6 and Table 7. Specifically, Variant 1, only containing the multilingual pre-training module, exhibits the poorest performance, whereas the inclusion of additional modules generally results in improved performance. Variant 2, which incorporates a contrastive learning pre-training module, demonstrates a significant improvement in experimental performance compared to Variant 1. This indicates that the proposed contrastive learning pre-training model effectively reduces the semantic distance between languages through knowledge transfer and thereby improves the performance of the model. Variant 3 removes the contrastive learning pre training and dynamic entanglement modules from the proposed CLQT model. Variant 4 adds a contrastive learning pre-training module on the basis of Variant 3. The results of variant 3 indicate that our proposed quantum mechanism module successfully integrates syntactic and semantic information from five languages. Comparing variant 4 with the CLQT model, it was found that applying the similarity rotation gate to quantum dynamic entanglement is effective and improves the performance of the model, which could better recognize aspect words and emotions.

5.4. Case Study

Table 8 presents three examples from the QPEN and our proposed CLQT models. −1, 0, and 1 represent negative, neutral, and positive sentiment, respectively. The results in brackets are aspect terms and sentiment polarities. From the first two examples, it is evident that both the QPEN and CLQT models effectively identify aspect terms in target language sentences, even when multiple aspect terms are present. Example 3 is a Dutch sentence that contains two aspects wijnkaart and wijnen per glass. It is observed that our proposed CLQT model successfully identifies both aspect terms and sentiment polarities, whereas the QPEN model only identifies the wine list aspect term. The enhanced performance of the CLQT model can be attributed to two primary factors. First, dynamic entanglement captures the relationships between languages more effectively than static entanglement states. Second, the contrastive learning pre-training module substantially improves the transfer of knowledge from the source language to the target language.

5.5. Error Analysis

An error analysis of CLQT performance is conducted by selecting 20 failed instances from each language, resulting in a total of 100 samples. The distribution of different error types is depicted in Figure 8. A manual review indicates that the majority of errors fell into the following categories:
  • Aspect item missing: Such errors are common and may arise from variations in language distribution, particularly due to structural inconsistencies across languages. It leads to a loss of crucial knowledge during the transfer from the source language to the target language, which impacts performance.
  • Wrong aspect item prediction: Incorrect aspect word predictions indicate a deviation in cross-lingual knowledge transfer, and the use of quantum projection and superposition introduces noise that further compromises the accuracy of this process.
  • Wrong prediction of emotional polarity: Incorrect predictions of emotional polarity are frequently observed, particularly in texts that are challenging to interpret, such as those containing satirical expressions.
  • Other errors
According to error analysis for our model, we plan to pursue several research directions to further mitigate error propagation and enhance the robustness of our framework. We will refine our feature extraction techniques by incorporating more advanced syntactic and semantic representations that capture subtle linguistic cues more effectively. This may involve leveraging state-of-the-art contextual embeddings models to reduce the incidence of missing aspect items. We will develop advanced disambiguation strategies to enhance sentiment polarity prediction, particularly in challenging cases such as texts with ambiguous or satirical expressions. These strategies may combine rule-based approaches with quantum techniques and integrate external knowledge sources to provide richer contextual cues.
These enhancements are expected to significantly reduce error and further improve the overall performance and robustness of our cross-lingual ABSA framework.

5.6. Visualization

To systematically analyze the behavior of our framework, we utilize t-SNE to visualize the feature representations, as presented in Figure 9. Panel (a) shows the embedding distribution from the XLM-R model, while panel (b) depicts the visualization produced by CLQT. Compared with the baseline XLM-R, CLQT yields a more uniform distribution representation between across languages. This indicates that the proposed contrastive learning pre-training and quantum modules facilitate the sharing of language-specific knowledge, ultimately leading to more accurate predictions for the cross-lingual ABSA task.
Specifically, the t-SNE visualizations provide a qualitative insight into the embedding space and demonstrate the effectiveness of our cross-lingual alignment strategy. In our experiments, we compare the t-SNE plots of token embeddings generated by a baseline XLM-R with those produced by our proposed CLQT model. The baseline embeddings exhibit distinct clusters corresponding to individual languages, reflecting the persistence of language-specific features and limited cross-lingual transfer. In contrast, the embeddings derived from CLQT show a much more uniform distribution across languages, with clusters that are more intermixed and aligned. Moreover, the visualization highlights that the utilization of contrastive learning pre-training and dynamic quantum entanglement in the model contributes to a more uniform distribution of representations across languages. This uniformity indicates that the framework successfully bridges the semantic gap between source and target languages, ensuring that similar sentiment expressions are aligned regardless of language.
Overall, the t-SNE visualizations reinforce the quantitative improvements reported in our experimental results by offering a visual confirmation that our integrated contrastive learning and quantum mechanism approach effectively bridges the gap between languages.

6. Conclusions

In this study, we introduced CLQT, an innovative framework designed to enhance cross-lingual ABSA through the integration of contrastive learning pre-training and quantum mechanisms. Specifically, our framework leverages contrastive learning to align representations of source language data and its translations effectively, thereby facilitating more robust semantic knowledge transfer across languages. Additionally, we incorporated quantum projection and dynamic quantum entanglement modules, enabling the model to automatically adjust entanglement strength based on linguistic similarities. This approach allows for more precise encoding of linguistic interactions and nuanced sentiment dependencies that classical methods may overlook. Our extensive experiments conducted on two benchmark datasets, SemEval-2016 and Amazon Reviews Corpus, demonstrate the effectiveness of CLQT. The model consistently outperformed existing state-of-the-art methods, showcasing marked improvements in F1 scores across multiple languages. Ablation studies further highlighted the critical role of the dynamic quantum entanglement module, affirming its significant contribution to performance enhancement.
However, our analysis also revealed several limitations. Errors in aspect extraction and sentiment classification often arise from structural and semantic divergences across languages, as well as from the inherent ambiguities of natural text. To address these challenges, we will refine our feature-extraction pipeline by incorporating advanced syntactic and semantic representations, such as graph-based linguistic structures and deep contextual embeddings, and by developing translation-noise mitigation strategies, including iterative back-translation and adversarial data augmentation. We also plan to explore deeper, error-mitigated entanglement circuits and to extend CLQT into a fully quantum, end-to-end sequence-labeling framework that jointly encodes aspect extraction and sentiment prediction. Finally, we will investigate the broader applicability of quantum-inspired representations across other cross-lingual NLP tasks, aiming to harness their power for modeling complex linguistic phenomena.

Author Contributions

Conceptualization, X.L. and K.Z.; methodology, X.L.; software, X.L.; validation, K.Z.; formal analysis, K.Z.; data curation, K.Z.; writing—original draft preparation, X.L.; writing—review and editing, K.Z.; visualization, X.L.; supervision, K.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

All original contributions of this study are incorporated in this article; any further inquiries may be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Pontiki, M.; Galanis, D.; Papageorgiou, H.; Androutsopoulos, I.; Manandhar, S.; Al-Smadi, M.; Al-Ayyoub, M.; Zhao, Y.; Qin, B.; De Clercq, O.; et al. Semeval-2016 task 5: Aspect based sentiment analysis. In Proceedings of the 10th International Workshop on Semantic Evaluation (SemEval-2016), San Diego, CA, USA, 16–17 June 2016; pp. 19–30. [Google Scholar]
  2. Zhang, W.; Li, X.; Deng, Y.; Bing, L.; Lam, W. A survey on aspect-based sentiment analysis: Tasks, methods, and challenges. IEEE Trans. Knowl. Data Eng. 2022, 35, 11019–11038. [Google Scholar] [CrossRef]
  3. Zhao, Z.; Tang, M.; Zhao, F.; Zhang, Z.; Chen, X. Incorporating semantics, syntax and knowledge for aspect based sentiment analysis. Appl. Intell. 2023, 53, 16138–16150. [Google Scholar] [CrossRef]
  4. Zhao, C.; Wu, M.; Yang, X.; Zhang, W.; Zhang, S.; Wang, S.; Li, D. A systematic review of cross-lingual sentiment analysis: Tasks, strategies, and prospects. ACM Comput. Surv. 2024, 56, 1–37. [Google Scholar] [CrossRef]
  5. Panchendrarajan, R.; Zubiaga, A. Claim detection for automated fact-checking: A survey on monolingual, multilingual and cross-lingual research. Nat. Lang. Process. J. 2024, 7, 100066. [Google Scholar] [CrossRef]
  6. Přibáň, P.; Šmíd, J.; Steinberger, J.; Mištera, A. A comparative study of cross-lingual sentiment analysis. Expert Syst. Appl. 2024, 247, 123247. [Google Scholar] [CrossRef]
  7. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–7 June 2019; Volume 1, pp. 4171–4186. [Google Scholar]
  8. Conneau, A.; Khandelwal, K.; Goyal, N.; Chaudhary, V.; Wenzek, G.; Guzmán, F.; Grave, E.; Ott, M.; Zettlemoyer, L.; Stoyanov, V. Unsupervised Cross-lingual Representation Learning at Scale. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 8440–8451. [Google Scholar]
  9. Wu, S.; Dredze, M. Beto, Bentz, Becas: The surprising cross-lingual effectiveness of Bert. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP, Hong Kong, China, 3–7 November 2019; pp. 833–844. [Google Scholar]
  10. Zhang, W.; He, R.; Peng, H.; Bing, L.; Lam, W. Cross-lingual aspect-based sentiment analysis with aspect term code-switching. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Online, 7–11 November 2021; pp. 9220–9230. [Google Scholar]
  11. Hämmerl, K.; Libovický, J.; Fraser, A. Understanding Cross-Lingual Alignment—A Survey. In Proceedings of the Findings of the Association for Computational Linguistics ACL, Bangkok, Thailand, 11–16 August 2024; pp. 10922–10943. [Google Scholar]
  12. Zhao, X.; Wan, H.; Qi, K. QPEN: Quantum projection and quantum entanglement enhanced network for cross-lingual aspect-based sentiment analysis. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–28 February 2024. [Google Scholar]
  13. Liu, B. Sentiment Analysis and Opinion Mining; Synthesis Lectures on Human Language Technologies; Springer: Cham, Switzerland, 2012; Volume 5, p. 1. [Google Scholar]
  14. D’Aniello, G.; Gaeta, M.; La Rocca, I. KnowMIS-ABSA: An overview and a reference model for applications of sentiment analysis and aspect-based sentiment analysis. Artif. Intell. Rev. 2022, 55, 5543–5574. [Google Scholar] [CrossRef]
  15. Hazem, A.; Bouhandi, M.; Boudin, F.; Daille, B. Cross-lingual and cross-domain transfer learning for automatic term extraction from low resource data. In Proceedings of the Thirteenth Language Resources and Evaluation Conference, Marseille, France, 20–25 June 2022. [Google Scholar]
  16. Zhou, X.; Wan, X.; Xiao, J. CLOpinionMiner: Opinion target extraction in a cross-language scenario. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 619–630. [Google Scholar] [CrossRef]
  17. Zhou, H.; Chen, L.; Shi, F.; Huang, D. Learning bilingual sentiment word embeddings for cross-language sentiment classification. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing, Beijing, China, 26–31 July 2015. [Google Scholar]
  18. Ruder, S.; Vulić, I.; Søgaard, A. A survey of cross-lingual word embedding models. J. Artif. Intell. Res. 2019, 65, 569–631. [Google Scholar] [CrossRef]
  19. Xue, L.; Constant, N.; Roberts, A.; Kale, M.; Al-Rfou, R.; Siddhant, A.; Barua, A.; Raffel, C. mT5: A Massively Multilingual Pre-trained Text-to-Text Transformer. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online, 6–11 June 2021. [Google Scholar]
  20. Lin, N.; Fu, Y.; Lin, X.; Zhou, D.; Yang, A.; Jiang, S. Cl-xabsa: Contrastive learning for cross-lingual aspect-based sentiment analysis. IEEE/ACM Trans. Audio Speech Lang. Process. 2023, 31, 2935–2946. [Google Scholar] [CrossRef]
  21. Sattar, K.; Umer, Q.; Vasbieva, D.G.; Chung, S.; Latif, Z.; Lee, C. A multi-layer network for aspect-based cross-lingual sentiment classification. IEEE Access 2021, 9, 133961–133973. [Google Scholar] [CrossRef]
  22. Hu, H.; Wang, X.; Zhang, Y.; Chen, Q.; Guan, Q. A comprehensive survey on contrastive learning. Neurocomputing 2024, 610, 128645. [Google Scholar] [CrossRef]
  23. Mohtarami, M.; Glass, J.; Nakov, P. Contrastive Language Adaptation for Cross-Lingual Stance Detection. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019. [Google Scholar]
  24. Guo, Z.; Liu, Z.; Ling, Z.; Wang, S.; Jin, L.; Li, Y. Text classification by contrastive learning and cross-lingual data augmentation for alzheimer’s disease detection. In Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain, 8–13 December 2020. [Google Scholar]
  25. Li, Z.; Zou, Y.; Zhang, C.; Zhang, Q.; Wei, Z. Learning Implicit Sentiment in Aspect-based Sentiment Analysis with Supervised Contrastive Pre-Training. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Online, 7–11 November 2021. [Google Scholar]
  26. Luo, Y.; Guo, F.; Liu, Z.; Zhang, Y. Mere Contrastive Learning for Cross-Domain Sentiment Analysis. In Proceedings of the 29th International Conference on Computational Linguistics, Gyeongju, Republic of Korea, 12–17 October 2022. [Google Scholar]
  27. Zhou, J.; Zhou, J.; Zhao, J.; Wang, S.; Shan, H.; Gui, T.; Zhang, Q.; Huang, X. A soft contrastive learning-based prompt model for few-shot sentiment analysis. In Proceedings of the ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024. [Google Scholar]
  28. Qin, L.; Chen, Q.; Xie, T.; Li, Q.; Lou, J.G.; Che, W.; Kan, M.Y. GL-CLeF: A Global–Local Contrastive Learning Framework for Cross-lingual Spoken Language Understanding. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Dublin, Ireland, 22–27 May 2022. [Google Scholar]
  29. Li, Q.; Gkoumas, D.; Sordoni, A.; Nie, J.Y.; Melucci, M. Quantum-inspired neural network for conversational emotion recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 19–21 May 2021; Volume 35, No. 15. pp. 13270–13278. [Google Scholar]
  30. Yan, G.; Wu, H.; Yan, J. Quantum 3D graph learning with applications to molecule embedding. In Proceedings of the 40th International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; pp. 2677–2686. [Google Scholar]
  31. Zhou, N.R.; Zhang, T.F.; Xie, X.W.; Wu, J.Y. Hybrid quantum–classical generative adversarial networks for image generation via learning discrete distribution. Signal Process. Image Commun. 2023, 110, 116891. [Google Scholar] [CrossRef]
  32. Gong, L.H.; Pei, J.J.; Zhang, T.F.; Zhou, N.R. Quantum convolutional neural network based on variational quantum circuits. Opt. Commun. 2024, 550, 129993. [Google Scholar] [CrossRef]
  33. He, R.; Lee, W.S.; Ng, H.T.; Dahlmeier, D. An Interactive Multi-Task Learning Network for End-to-End Aspect-Based Sentiment Analysis. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019. [Google Scholar]
  34. Li, X.; Bing, L.; Zhang, W.; Lam, W. Exploiting BERT for End-to-End Aspect-based Sentiment Analysis. In Proceedings of the 5th Workshop on Noisy User-generated Text (W-NUT 2019), Hong Kong, China, 4 November 2019. [Google Scholar]
  35. Busemeyer, J.R.; Bruza, P.D. Quantum Models of Cognition and Decision; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
  36. Fell, L.; Dehdashti, S.; Bruza, P.; Moreira, C. An Experimental Protocol to Derive and Validate a Quantum Model of Decision-Making. In Proceedings of the Annual Meeting of the Cognitive Science Society, Montreal, QC, Canada, 24–27 July 2019. [Google Scholar]
  37. Miller, D.M.; Thornton, M.A. QMDD: A decision diagram structure for reversible and quantum circuits. In Proceedings of the 36th International Symposium on Multiple-Valued Logic (ISMVL’06), Singapore, 17–20 May 2006. [Google Scholar]
  38. Keung, P.; Lu, Y.; Szarvas, G.S.; Smith, N.A. The Multilingual Amazon Reviews Corpus. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 16–20 November 2020. [Google Scholar]
  39. Li, X.; Bing, L.; Zhang, W.; Li, Z.; Lam, W. Unsupervised cross-lingual adaptation for sequence tagging and beyond. arXiv 2020, arXiv:2010.12405. [Google Scholar]
  40. Zhao, W.; Yang, Z.; Yu, S.; Zhu, S.; Li, L. Contrastive pre-training and instruction tuning for cross-lingual aspect-based sentiment analysis. Appl. Intell. 2025, 55, 358. [Google Scholar] [CrossRef]
Figure 1. An example of cross-lingual aspect-based sentiment analysis. (a) is an English sample and (b) is a Spanish sample.
Figure 1. An example of cross-lingual aspect-based sentiment analysis. (a) is an English sample and (b) is a Spanish sample.
Entropy 27 00713 g001
Figure 2. The CLQT framework is structured around three key components: contrastive learning, quantum projection, and a quantum entanglement module.
Figure 2. The CLQT framework is structured around three key components: contrastive learning, quantum projection, and a quantum entanglement module.
Entropy 27 00713 g002
Figure 3. Representations of the aspect in a Hilbert space.
Figure 3. Representations of the aspect in a Hilbert space.
Entropy 27 00713 g003
Figure 4. Implementation of the quantum entanglement module using a quantum circuit. H is the Hadamard gate, and CRY is the weighted revolving door.
Figure 4. Implementation of the quantum entanglement module using a quantum circuit. H is the Hadamard gate, and CRY is the weighted revolving door.
Entropy 27 00713 g004
Figure 5. Comparison of F1 score in pre-trained language models. (a) presents the F1 score for mBERT and its contrastive learning-enhanced variant (mBERT-CL), while subfigure (b) displays the corresponding results for XLM-R and XLM-R-CL.
Figure 5. Comparison of F1 score in pre-trained language models. (a) presents the F1 score for mBERT and its contrastive learning-enhanced variant (mBERT-CL), while subfigure (b) displays the corresponding results for XLM-R and XLM-R-CL.
Entropy 27 00713 g005
Figure 6. Visualization of contrastive learning pre-training. (a) illustrates the feature distribution in the absence of contrastive learning. (b) shows the feature distribution following contrastive learning-based pre-training.
Figure 6. Visualization of contrastive learning pre-training. (a) illustrates the feature distribution in the absence of contrastive learning. (b) shows the feature distribution following contrastive learning-based pre-training.
Entropy 27 00713 g006
Figure 7. Performance comparison of different models with the addition of the dynamic similarity module. (a) XLM-RoBERTa and its derivative models; (b) CLQT together with its companion approaches; (c) CL-XABSA-TL and its associated variants; and (d) CL-XABSA-SL along with its related methods.
Figure 7. Performance comparison of different models with the addition of the dynamic similarity module. (a) XLM-RoBERTa and its derivative models; (b) CLQT together with its companion approaches; (c) CL-XABSA-TL and its associated variants; and (d) CL-XABSA-SL along with its related methods.
Entropy 27 00713 g007
Figure 8. Error type distribution heatmap.
Figure 8. Error type distribution heatmap.
Entropy 27 00713 g008
Figure 9. We visualize the test set by applying t-SNE to the sentence embeddings generated by CLQT. In this visualization, panel (a) illustrates the embedding plot produced by the XLM-R model, while panel (b) depicts the plot obtained from CLQT.
Figure 9. We visualize the test set by applying t-SNE to the sentence embeddings generated by CLQT. In this visualization, panel (a) illustrates the embedding plot produced by the XLM-R model, while panel (b) depicts the plot obtained from CLQT.
Entropy 27 00713 g009
Table 1. Circuit width.
Table 1. Circuit width.
Sub-RegisterLogical ObjectsPhysical EncodingQubit Count
Language register L = 5 languages1 qubit per language5
Sentiment register3-level ( + , 0 , )mapped post-entanglement2
controlled rotations and read-outtransientreused and reset≤1
These two qubits are not part of the core entanglement network but are used later for sentiment measurement.
Table 2. SemEval-2016 consumer reviews dataset.
Table 2. SemEval-2016 consumer reviews dataset.
EnglishFrenchSpanishDutchRussian
TrainNo.Sen20001664207017223655
No.Asp17431641185612313077
TestNo.Sen6766688815751209
No.Asp612650713373949
Table 3. CLQT uses mBERT pre-trained model in SemEval-2016 experimental results.
Table 3. CLQT uses mBERT pre-trained model in SemEval-2016 experimental results.
MethodsmBert
French Spanish Dutch Russian Avgerage
Zero-shot45.6057.3242.6836.0145.40
Translation-TA40.7650.7447.1341.6745.08
Bilingual-TA41.0051.2349.7243.6746.41
Translation-AF48.0359.7449.7350.1751.92
Binglingual-AF48.0560.2349.8351.2452.34
ACS49.6559.9951.1952.0953.23
ACS-Distill-S52.2362.0452.7253.0055.00
ACS-Distill-M52.2562.9153.4054.5855.79
CL-XABSA-TL48.5360.6450.9650.7752.73
CL-XABSA-SL49.5061.6250.6450.6553.10
QPEN53.2763.8454.6155.3656.77
CLQT (ours)54.3465.0155.7356.2757.84
Note: Values in bold represent the optimal results.
Table 4. CLQT uses XLM-R pre-trained model in SemEval-2016 experimental results.
Table 4. CLQT uses XLM-R pre-trained model in SemEval-2016 experimental results.
MethodsXLM-R
French Spanish Dutch Russian Avgerage
Zero-shot56.4367.1059.0356.8059.84
Translation-TA47.0058.1056.1950.3452.91
Bilingual-TA49.3461.8758.6452.8955.69
Translation-AF57.0766.6161.2659.5561.12
Binglingual-AF57.9168.0460.8060.8161.89
ACS59.3967.3262.8360.9762.63
ACS-Distill-S61.0068.9362.8960.9763.45
ACS-Distill-M59.9069.2463.7462.0263.73
CL-XABSA-TL60.4169.8761.3058.8262.60
CL-XABSA-SL61.8770.9562.0358.1863.26
QPEN63.2171.5966.1664.5265.79
CLQT (ours)66.1072.9868.3266.7368.53
Note: Values in bold represent the optimal results.
Table 5. The results of experiment on Amazon Reviews Corpus dataset.
Table 5. The results of experiment on Amazon Reviews Corpus dataset.
MethodsGermanSpanishFrenchJapaneseChineseAvgerage
CAPIT-base76.9975.6475.3673.4868.7874.09
CAPIT-large78.3475.5676.9073.5470.6075.00
XLM-RoBERTa55.5251.9652.4251.3048.8650.86
ACS-Distill-S68.7167.2964.0261.8558.6764.05
ACS-Distill-M70.5868.9766.3563.1460.7265.95
CL-XABSA-TL68.3267.7063.4761.1657.8963.71
CL-XABSA-SL69.5969.0464.7062.6358.0764.80
QPEN78.9878.8475.9372.1968.2074.83
CLQT82.0281.8579.1975.7771.9378.15
Note: Values in bold represent the optimal results.
Table 6. The results of ablation study on mBERT pre-training model.
Table 6. The results of ablation study on mBERT pre-training model.
MethodsmBert
French Spanish Dutch Russian Average
Variant146.01 ± 2.23 57.19 ± 1.67 43.05 ± 2.44 36.24 ± 1.89 45.62 ± 2.06
Variant246.82 ± 1.94 59.71 ± 1.42 43.94 ± 2.17 37.65 ± 1.63 47.03 ± 1.79
Variant353.27 ± 1.88 63.84 ± 1.39 54.61 ± 2.12 55.36 ± 1.58 56.97 ± 1.74
Variant453.80 ± 1.49 64.27 ± 1.08 55.14 ± 1.66 55.92 ± 1.27 57.28 ± 1.38
CLQT (ours)54.34 ± 0.9565.01 ± 1.3055.73 ± 1.3556.27 ± 0.5257.84 ± 1.03
Table 7. The results of ablation study on XLM-R pre-training model.
Table 7. The results of ablation study on XLM-R pre-training model.
MethodsXLM-R
French Spanish Dutch Russian Average
Variant155.92 ± 2.16 67.61 ± 1.33 59.27 ± 2.23 57.35 ± 1.87 60.04 ± 1.90
Variant256.88 ± 1.82 68.10 ± 1.09 60.86 ± 1.94 58.34 ± 1.61 61.05 ± 1.62
Variant363.21 ± 1.78 71.59 ± 1.06 66.16 ± 1.91 64.52 ± 1.58 66.37 ± 1.58
Variant464.36 ± 1.47 71.93 ± 0.88 67.25 ± 1.62 65.49 ± 1.36 67.26 ± 1.33
CLQT (ours)66.10 ± 1.4772.98 ± 1.1468.32 ± 1.8466.73 ± 1.2368.53 ± 1.42
Table 8. Case study.
Table 8. Case study.
LanguagesSentencesQPENCLQT
SPLa comida estuvo muy sabrosa.(comida 1)(comida 1)
FRLe cadre et le personnel sont agréables.(cadre 1, personnel, 1)(cadre 1, personnel, 1)
DUGeen kennis van de wijnkaart laat staan van de wijnen per glass.(wijnkaart −1)(wijnkaart −1, wijnen per glass −1)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, X.; Zhang, K. Contrastive Learning Pre-Training and Quantum Theory for Cross-Lingual Aspect-Based Sentiment Analysis. Entropy 2025, 27, 713. https://doi.org/10.3390/e27070713

AMA Style

Li X, Zhang K. Contrastive Learning Pre-Training and Quantum Theory for Cross-Lingual Aspect-Based Sentiment Analysis. Entropy. 2025; 27(7):713. https://doi.org/10.3390/e27070713

Chicago/Turabian Style

Li, Xun, and Kun Zhang. 2025. "Contrastive Learning Pre-Training and Quantum Theory for Cross-Lingual Aspect-Based Sentiment Analysis" Entropy 27, no. 7: 713. https://doi.org/10.3390/e27070713

APA Style

Li, X., & Zhang, K. (2025). Contrastive Learning Pre-Training and Quantum Theory for Cross-Lingual Aspect-Based Sentiment Analysis. Entropy, 27(7), 713. https://doi.org/10.3390/e27070713

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop